TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Centipede

Atari Games on Atari 2600 Centipede

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1Go-Explore1422628NoFirst return, then explore2020-04-27Code
2GDI-H3(1B frames)1359533No---
3MuZero1159049.27NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
4MuZero (Res2 Adam)874301.64NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
5R2D2599140.3No--Code
6Agent57412847.86NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
7GDI-H3195630NoGeneralized Data Distribution Iteration2022-06-07-
8GDI-I3155830NoGeneralized Data Distribution Iteration2022-06-07-
9GDI-I3155830NoGeneralized Data Distribution Iteration2022-06-07-
10Full Tree125123NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
11DNA100194NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
12DDQN+Pop-Art noop49065.8NoLearning values across many orders of magnitude2016-02-24-
13CGP24708NoEvolving simple programs for playing Atari games2018-06-14Code
14Ape-X12974NoDistributed Prioritized Experience Replay2018-03-02Code
15QR-DQN-112447NoDistributional Reinforcement Learning with Quant...2017-10-27Code
16DreamerV211883NoMastering Atari with Discrete World Models2020-10-05Code
17IQN11561NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
18IMPALA (deep)11049.75NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
19C51 noop9646NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
20Best Learner8803.8NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
21Nature DQN8309No--Code
22ES FF (1 hour) noop7783.9NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
23Prior+Duel noop7687.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
24NoisyNet-Dueling7596NoNoisy Networks for Exploration2017-06-30Code
25Duel noop7561.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
26A2C + SIL7559.5NoSelf-Imitation Learning2018-06-14Code
27Gorila6296.9NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
28Prior+Duel hs5570.2NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
29DDQN (tuned) noop5409.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
30Duel hs4881NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
31DQN noop4657.7NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
32SARSA4647No---
33Bootstrapped DQN4553.5NoDeep Exploration via Bootstrapped DQN2016-02-15Code
34Persistent AL4539.55NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
35Prior noop4463.2NoPrioritized Experience Replay2015-11-18Code
36Advantage Learning4225.18NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
37DQN hs3973.9NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
38ASL DDQN3899.8NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
39DDQN (tuned) hs3853.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
40A3C FF hs3755.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
41Prior hs3489.1NoPrioritized Experience Replay2015-11-18Code
42Reactor 500M3422NoThe Reactor: A fast and sample-efficient Actor-C...2017-04-15-
43POP3D3315.44NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
44A3C FF (1 day) hs3306.5NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
45A3C LSTM hs1997NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code