TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 James Bond

Video Games on Atari 2600 James Bond

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1GDI-H3620780NoGeneralized Data Distribution Iteration2022-06-07-
2GDI-I3594500NoGeneralized Data Distribution Iteration2022-06-07-
3GDI-I3594500NoGeneralized Data Distribution Iteration2022-06-07-
4Agent57135784.96NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
5FQF87291.7NoFully Parameterized Quantile Function for Distri...2019-11-05Code
6MuZero41063.25NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
7DreamerV240445NoMastering Atari with Discrete World Models2020-10-05Code
8IQN35108NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
9MuZero (Res2 Adam)28626.23NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
10R2D225354No--Code
11Ape-X21322.5NoDistributed Prioritized Experience Replay2018-03-02Code
12DNA14102NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
13CGP6130NoEvolving simple programs for playing Atari games2018-06-14Code
14Prior noop5148NoPrioritized Experience Replay2015-11-18Code
15QR-DQN-14703NoDistributional Reinforcement Learning with Quant...2017-10-27Code
16Prior hs3961NoPrioritized Experience Replay2015-11-18Code
17ASL DDQN2237NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
18C51 noop1909NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
19Bootstrapped DQN1663.5NoDeep Exploration via Bootstrapped DQN2016-02-15Code
20DDQN (tuned) noop1358NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
21Duel noop1312.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
22Recurrent Rational DQN Average1137NoAdaptive Rational Activations to Boost Deep Rein...2021-02-18Code
23Rational DQN Average1122NoAdaptive Rational Activations to Boost Deep Rein...2021-02-18Code
24Advantage Learning848.46NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
25Duel hs835.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
26Prior+Duel noop812NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
27Persistent AL772.09NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
28DQN noop768.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
29DQN hs697.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
30A3C LSTM hs613NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
31IMPALA (deep)601.5NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
32Prior+Duel hs585NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
33Nature DQN576.7No--Code
34DDQN (tuned) hs573NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
35A3C FF hs541NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
36DDQN+Pop-Art noop507.5NoLearning values across many orders of magnitude2016-02-24-
37Gorila444NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
38POP3D358.54NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
39SARSA354.1No---
40A3C FF (1 day) hs351.5NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
41UCT330NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
42A2C + SIL310.8NoSelf-Imitation Learning2018-06-14Code
43Best Learner202.8NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
44SAC68.3NoSoft Actor-Critic for Discrete Action Settings2019-10-16Code