TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Bowling

Atari Games on Atari 2600 Bowling

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1MuZero260.13NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
2Go-Explore260NoFirst return, then explore2020-04-27Code
3Agent57251.18NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
4R2D2219.5No--Code
5GDI-H3205.2NoGeneralized Data Distribution Iteration2022-06-07-
6GDI-I3201.9NoGeneralized Data Distribution Iteration2022-06-07-
7GDI-I3201.9NoGeneralized Data Distribution Iteration2022-06-07-
8DNA181NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
9RUDDER179NoRUDDER: Return Decomposition for Delayed Rewards2018-06-20Code
10MuZero (Res2 Adam)131.65NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
11FQF102.3NoFully Parameterized Quantile Function for Distri...2019-11-05Code
12DDQN+Pop-Art noop102.1NoLearning values across many orders of magnitude2016-02-24-
13IQN86.5NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
14CGP85.8NoEvolving simple programs for playing Atari games2018-06-14Code
15C51 noop81.8NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
16Reactor 500M81NoThe Reactor: A fast and sample-efficient Actor-C...2017-04-15-
17QR-DQN-177.2NoDistributional Reinforcement Learning with Quant...2017-10-27Code
18Persistent AL71.59NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
19DDQN (tuned) hs69.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
20DDQN (tuned) noop68.1NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
21Duel hs65.7NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
22Duel noop65.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
23ASL DDQN62.4NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
24Bootstrapped DQN60.2NoDeep Exploration via Bootstrapped DQN2016-02-15Code
25IMPALA (deep)59.92NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
26Advantage Learning57.41NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
27DQN hs56.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
28Gorila54NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
29Prior hs52NoPrioritized Experience Replay2015-11-18Code
30DQN noop50.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
31Prior+Duel hs50.4NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
32DreamerV249NoMastering Atari with Discrete World Models2020-10-05Code
33Prior noop47.9NoPrioritized Experience Replay2015-11-18Code
34Prior+Duel noop46.7NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
35Best Learner43.9NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
36Nature DQN42.4No--Code
37A3C LSTM hs41.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
38POP3D38.99NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
39SARSA36.4No---
40A3C FF (1 day) hs36.2NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
41A3C FF hs35.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
42A2C + SIL31.1NoSelf-Imitation Learning2018-06-14Code
43ES FF (1 hour) noop30NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
44Ape-X17.6NoDistributed Prioritized Experience Replay2018-03-02Code