TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 Gravitar

Video Games on Atari 2600 Gravitar

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1Agent5719213.96NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
2R2D215680.7No--Code
3MuZero (Res2 Adam)8006.93NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
4Go-Explore7588NoFirst return, then explore2020-04-27Code
5SND-VIC6712NoSelf-supervised network distillation: an effecti...2023-02-22Code
6MuZero6682.7NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
7GDI-H35915NoGeneralized Data Distribution Iteration2022-06-07-
8GDI-I35905NoGeneralized Data Distribution Iteration2022-06-07-
9GDI-I35905NoGeneralized Data Distribution Iteration2022-06-07-
10SND-STD4643NoSelf-supervised network distillation: an effecti...2023-02-22Code
11RND3906NoExploration by Random Network Distillation2018-10-30Code
12DreamerV23789NoMastering Atari with Discrete World Models2020-10-05Code
13UCT2850NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
14SND-V2741NoSelf-supervised network distillation: an effecti...2023-02-22Code
15CGP2350NoEvolving simple programs for playing Atari games2018-06-14Code
16NoisyNet-Dueling2209NoNoisy Networks for Exploration2017-06-30Code
17DNA2190NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
18A2C + SIL1874.2NoSelf-Imitation Learning2018-06-14Code
19Ape-X1598.5NoDistributed Prioritized Experience Replay2018-03-02Code
20FQF1406NoFully Parameterized Quantile Function for Distri...2019-11-05Code
21Intrinsic Reward Agent1165.1NoLarge-Scale Study of Curiosity-Driven Learning2018-08-13Code
22DQNMMCe1078.3NoCount-Based Exploration with the Successor Repre...2018-07-31Code
23QR-DQN-1995NoDistributional Reinforcement Learning with Quant...2017-10-27Code
24IQN911NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
25ES FF (1 hour) noop805NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
26ASL DDQN760NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
27Duel noop588NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
28POP3D557.17NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
29Prior noop548.5NoPrioritized Experience Replay2015-11-18Code
30Gorila538.4NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
31DQN-PixelCNN498.3NoCount-Based Exploration with Neural Density Models2017-03-03Code
32DDQN+Pop-Art noop483.5NoLearning values across many orders of magnitude2016-02-24-
33DQN noop473NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
34Persistent AL446.92NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
35C51 noop440NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
36SARSA429No---
37Advantage Learning417.65NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
38DDQN (tuned) noop412NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
39Best Learner387.7NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
40IMPALA (deep)359.5NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
41A3C LSTM hs320NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
42Nature DQN306.7No--Code
43A3C FF hs303.5NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
44DQN hs298NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
45Duel hs297NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
46Bootstrapped DQN286.1NoDeep Exploration via Bootstrapped DQN2016-02-15Code
47Prior hs269.5NoPrioritized Experience Replay2015-11-18Code
48A3C FF (1 day) hs269.5NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
49A3C-CTS238.68NoUnifying Count-Based Exploration and Intrinsic M...2016-06-06Code
50Prior+Duel noop238NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
51DQN-CTS238NoCount-Based Exploration with Neural Density Models2017-03-03Code
52DDQN (tuned) hs200.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
53Prior+Duel hs167NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code