TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Double Dunk

Atari Games on Atari 2600 Double Dunk

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1UCT24NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
2GDI-H324NoGeneralized Data Distribution Iteration2022-06-07-
3GDI-I324NoGeneralized Data Distribution Iteration2022-06-07-
4GDI-H324NoGeneralized Data Distribution Iteration2022-06-07-
5MuZero23.94NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
6Agent5723.93NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
7MuZero (Res2 Adam)23.91NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
8R2D223.7No--Code
9Ape-X23.5NoDistributed Prioritized Experience Replay2018-03-02Code
10Reactor 500M23NoThe Reactor: A fast and sample-efficient Actor-C...2017-04-15-
11QR-DQN-121.9NoDistributional Reinforcement Learning with Quant...2017-10-27Code
12A2C + SIL21.5NoSelf-Imitation Learning2018-06-14Code
13Prior noop18.5NoPrioritized Experience Replay2015-11-18Code
14DreamerV217NoMastering Atari with Discrete World Models2020-10-05Code
15Prior hs16NoPrioritized Experience Replay2015-11-18Code
16IQN5.6NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
17Bootstrapped DQN3NoDeep Exploration via Bootstrapped DQN2016-02-15Code
18C51 noop2.5NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
19CGP2NoEvolving simple programs for playing Atari games2018-06-14Code
20NoisyNet-Dueling1NoNoisy Networks for Exploration2017-06-30Code
21ES FF (1 hour) noop0.2NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
22Duel noop0.1NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
23A3C FF (1 day) hs0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
24A3C LSTM hs0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
25ASL DDQN0.1NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
26A3C FF hs-0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
27Advantage Learning-0.15NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
28DDQN (tuned) hs-0.3NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
29IMPALA (deep)-0.33NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
30Duel hs-0.8NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
31DNA-1.3NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
32Persistent AL-2.51NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
33DDQN (tuned) noop-5.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
34DQN hs-6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
35DQN noop-6.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
36POP3D-7.89NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
37Prior+Duel hs-10.7NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
38Gorila-11.3NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
39DDQN+Pop-Art noop-11.5NoLearning values across many orders of magnitude2016-02-24-
40Prior+Duel noop-12.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
41Best Learner-13.1NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
42SARSA-16No---
43Nature DQN-18.1No--Code