TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Tutankham

Atari Games on Atari 2600 Tutankham

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1Agent572354.91NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
2MuZero491.48NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
3GDI-I3423.9NoGeneralized Data Distribution Iteration2022-06-07-
4GDI-I3423.9NoGeneralized Data Distribution Iteration2022-06-07-
5GDI-H3418.2NoGeneralized Data Distribution Iteration2022-06-07-
6R2D2395.3No--Code
7MuZero (Res2 Adam)347.99NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
8A2C + SIL340.5NoSelf-Imitation Learning2018-06-14Code
9QR-DQN-1297NoDistributional Reinforcement Learning with Quant...2017-10-27Code
10IQN293NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
11IMPALA (deep)292.11NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
12C51 noop280NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
13Ape-X272.6NoDistributed Prioritized Experience Replay2018-03-02Code
14NoisyNet-Dueling269NoNoisy Networks for Exploration2017-06-30Code
15DreamerV2264NoMastering Atari with Discrete World Models2020-10-05Code
16ASL DDQN252.9NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
17Prior+Duel noop245.9NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
18Advantage Learning245.22NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
19POP3D241.21NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
20UCT225.5NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
21DDQN (tuned) noop218.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
22Bootstrapped DQN214.8NoDeep Exploration via Bootstrapped DQN2016-02-15Code
23Duel noop211.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
24Prior noop204.6NoPrioritized Experience Replay2015-11-18Code
25DARQN soft197NoDeep Attention Recurrent Q-Network2015-12-05Code
26Nature DQN186.7No--Code
27Recurrent Rational DQN Average184NoAdaptive Rational Activations to Boost Deep Rein...2021-02-18Code
28DDQN+Pop-Art noop183.9NoLearning values across many orders of magnitude2016-02-24-
29Rational DQN Average179NoAdaptive Rational Activations to Boost Deep Rein...2021-02-18Code
30A3C FF hs156.3NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
31A3C LSTM hs144.2NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
32ES FF (1 hour) noop130.3NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
33DNA127NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
34Gorila118.5NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
35Best Learner114.3NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
36Prior+Duel hs108.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
37SARSA98.2No---
38DDQN (tuned) hs92.2NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
39DQN noop68.1NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
40Prior hs56.9NoPrioritized Experience Replay2015-11-18Code
41Duel hs48NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
42DQN hs45.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
43A3C FF (1 day) hs26.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
44CGP0No--Code