TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 Gopher

Video Games on Atari 2600 Gopher

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1GDI-I3488830NoGeneralized Data Distribution Iteration2022-06-07-
2GDI-H3473560NoGeneralized Data Distribution Iteration2022-06-07-
3MuZero130345.58NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
4R2D2124776.3No--Code
5MuZero (Res2 Adam)122882.5NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
6Ape-X120500.9NoDistributed Prioritized Experience Replay2018-03-02Code
7IQN118365NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
8Agent57117777.08NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
9QR-DQN-1113585NoDistributional Reinforcement Learning with Quant...2017-10-27Code
10Prior+Duel hs105148.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
11Prior+Duel noop104368.2NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
12ASL DDQN103514.4NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
13DreamerV292282NoMastering Atari with Discrete World Models2020-10-05Code
14DNA80104NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
15IMPALA (deep)66782.3NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
16DDQN+Pop-Art noop56218.2NoLearning values across many orders of magnitude2016-02-24-
17NoisyNet-Dueling38909NoNoisy Networks for Exploration2017-06-30Code
18Prior hs34858.8NoPrioritized Experience Replay2015-11-18Code
19C51 noop33641NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
20Prior noop32487.2NoPrioritized Experience Replay2015-11-18Code
21A2C + SIL23304.2NoSelf-Imitation Learning2018-06-14Code
22UCT20560NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
23Duel hs20051.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
24Bootstrapped DQN17438.4NoDeep Exploration via Bootstrapped DQN2016-02-15Code
25A3C LSTM hs17106.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
26Duel noop15718.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
27DDQN (tuned) hs15253NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
28DDQN (tuned) noop14840.8NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
29Advantage Learning11912.68NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
30Persistent AL10611.81NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
31A3C FF hs10022.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
32DQN noop8777.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
33Nature DQN8520No--Code
34A3C FF (1 day) hs8442.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
35DQN hs8190.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
36POP3D6207NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
37DARQN soft5356NoDeep Attention Recurrent Q-Network2015-12-05Code
38Gorila4373NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
39SARSA2368No---
40CGP1696NoEvolving simple programs for playing Atari games2018-06-14Code
41Best Learner1288.3NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
42CURL801.4NoCURL: Contrastive Unsupervised Representations f...2020-04-08Code
43ES FF (1 hour) noop582NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code