TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Montezuma's Revenge

Atari Games on Atari 2600 Montezuma's Revenge

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1Go-Explore43791NoFirst return, then explore2020-04-27Code
2Go-Explore43763NoGo-Explore: a New Approach for Hard-Exploration ...2019-01-30Code
3SND-V21565NoSelf-supervised network distillation: an effecti...2023-02-22Code
4Agent579352.01NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
5RND8152NoExploration by Random Network Distillation2018-10-30Code
6SND-VIC7838NoSelf-supervised network distillation: an effecti...2023-02-22Code
7SND-STD7212NoSelf-supervised network distillation: an effecti...2023-02-22Code
8A2C+CoEX6635NoContingency-Aware Exploration in Reinforcement L...2018-11-05-
9DQN-PixelCNN3705.5NoCount-Based Exploration with Neural Density Models2017-03-03Code
10DDQN-PC3459NoUnifying Count-Based Exploration and Intrinsic M...2016-06-06Code
11GDI-I33000NoGeneralized Data Distribution Iteration2022-06-07-
12GDI-I33000NoGeneralized Data Distribution Iteration2022-06-07-
13Sarsa-φ-EB2745.4NoCount-Based Exploration in Feature Space for Rei...2017-06-25Code
14Intrinsic Reward Agent2504.6NoLarge-Scale Study of Curiosity-Driven Learning2018-08-13Code
15Ape-X2500NoDistributed Prioritized Experience Replay2018-03-02Code
16MuZero (Res2 Adam)2500NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
17GDI-H32500NoGeneralized Data Distribution Iteration2022-06-07-
18R2D22061.3No--Code
19DQN+SR1778.8NoCount-Based Exploration with the Successor Repre...2018-07-31Code
20DQNMMCe+SR1778.6NoCount-Based Exploration with the Successor Repre...2018-07-31Code
21A2C + SIL1100NoSelf-Imitation Learning2018-06-14Code
22Sarsa-ε399.5NoCount-Based Exploration in Feature Space for Rei...2017-06-25Code
23A3C-CTS273.7NoUnifying Count-Based Exploration and Intrinsic M...2016-06-06Code
24SARSA259No---
25MP-EB142NoIncentivizing Exploration In Reinforcement Learn...2015-07-03Code
26Bootstrapped DQN100NoDeep Exploration via Bootstrapped DQN2016-02-15Code
27Gorila84NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
28DreamerV281NoMastering Atari with Discrete World Models2020-10-05Code
29TRPO-hash75No#Exploration: A Study of Count-Based Exploration...2016-11-15Code
30A3C FF hs67NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
31NoisyNet-Dueling57NoNoisy Networks for Exploration2017-06-30Code
32A3C FF (1 day) hs53NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
33Prior hs51NoPrioritized Experience Replay2015-11-18Code
34DQN hs47NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
35DDQN (tuned) hs42NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
36A3C LSTM hs41NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
37Prior+Duel hs24NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
38Duel hs22NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
39Best Learner10.7NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
40Persistent AL1.72NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
41Advantage Learning0.42NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
42IQN0No--Code
43MuZero0No--Code
44IMPALA (deep)0No--Code
45CGP0No--Code
46POP3D0No--Code
47QR-DQN-10No--Code
48DNA0No--Code
49ASL DDQN0No--Code
50Nature DQN0No--Code