TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 Berzerk

Video Games on Atari 2600 Berzerk

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1Go-Explore197376NoFirst return, then explore2020-04-27Code
2MuZero85932.6NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
3Agent5761507.83NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
4Ape-X57196.7NoDistributed Prioritized Experience Replay2018-03-02Code
5R2D253318.7No--Code
6DNA19789NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
7GDI-H314649NoGeneralized Data Distribution Iteration2022-06-07-
8FQF12422.2NoFully Parameterized Quantile Function for Distri...2019-11-05Code
9GDI-I37607NoGeneralized Data Distribution Iteration2022-06-07-
10GDI-I37607NoGeneralized Data Distribution Iteration2022-06-07-
11Prior+Duel noop3409NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
12QR-DQN-13117NoDistributional Reinforcement Learning with Quant...2017-10-27Code
13MuZero (Res2 Adam)2705.82NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
14ASL DDQN2597.2NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
15Reactor 500M2303.1NoThe Reactor: A fast and sample-efficient Actor-C...2017-04-15-
16Prior+Duel hs2178.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
17NoisyNet-Dueling1896NoNoisy Networks for Exploration2017-06-30Code
18IMPALA (deep)1852.7NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
19C51 noop1645NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
20Duel noop1472.6NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
21A3C FF (1 day) hs1433.4NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
22Persistent AL1328.25NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
23Prior noop1305.6NoPrioritized Experience Replay2015-11-18Code
24DDQN (tuned) noop1225.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
25DDQN+Pop-Art noop1199.6NoLearning values across many orders of magnitude2016-02-24-
26CGP1138NoEvolving simple programs for playing Atari games2018-06-14Code
27IQN1053NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
28DDQN (tuned) hs1011.1NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
29Duel hs910.6NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
30Prior hs865.9NoPrioritized Experience Replay2015-11-18Code
31A3C LSTM hs862.2NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
32A3C FF hs817.9NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
33DreamerV2810NoMastering Atari with Discrete World Models2020-10-05Code
34Advantage Learning747.26NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
35ES FF (1 hour) noop686NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
36Best Baseline670NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
37DQN noop585.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
38Best Learner501.3NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
39DQN hs493.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code