TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 Robotank

Video Games on Atari 2600 Robotank

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1MuZero131.13NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
2Agent57127.32NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
3GDI-H3113.4NoGeneralized Data Distribution Iteration2022-06-07-
4GDI-I3108.2NoGeneralized Data Distribution Iteration2022-06-07-
5GDI-I3108.2NoGeneralized Data Distribution Iteration2022-06-07-
6MuZero (Res2 Adam)100.59NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
7R2D2100.4No--Code
8DreamerV278NoMastering Atari with Discrete World Models2020-10-05Code
9FQF75.7NoFully Parameterized Quantile Function for Distri...2019-11-05Code
10Ape-X73.8NoDistributed Prioritized Experience Replay2018-03-02Code
11Advantage Learning69.31NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
12Bootstrapped DQN66.6NoDeep Exploration via Bootstrapped DQN2016-02-15Code
13ASL DDQN65.8NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
14Duel noop65.3NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
15DDQN (tuned) noop65.1NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
16DNA64.8NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
17DDQN+Pop-Art noop64.3NoLearning values across many orders of magnitude2016-02-24-
18NoisyNet-Dueling64NoNoisy Networks for Exploration2017-06-30Code
19DQN noop63.9NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
20Prior noop62.6NoPrioritized Experience Replay2015-11-18Code
21IQN62.5NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
22Duel hs62NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
23Gorila61.8NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
24QR-DQN-159.4NoDistributional Reinforcement Learning with Quant...2017-10-27Code
25DDQN (tuned) hs59.1NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
26DQN hs58.7NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
27Prior hs56.2NoPrioritized Experience Replay2015-11-18Code
28C51 noop52.3NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
29Nature DQN51.6No--Code
30UCT50.4NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
31A3C FF hs32.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
32Best Learner28.7NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
33Prior+Duel noop27.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
34Prior+Duel hs24.7NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
35CGP24.2NoEvolving simple programs for playing Atari games2018-06-14Code
36IMPALA (deep)12.96NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
37SARSA12.4No---
38ES FF (1 hour) noop11.9NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
39A2C + SIL10.5NoSelf-Imitation Learning2018-06-14Code
40POP3D4.6NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
41A3C LSTM hs2.6NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
42A3C FF (1 day) hs2.3NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code