TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Ice Hockey

Atari Games on Atari 2600 Ice Hockey

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1GDI-H3481.9NoGeneralized Data Distribution Iteration2022-06-07-
2R2D279.3No--Code
3MuZero67.04NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
4Agent5763.64NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
5GDI-I344.94NoGeneralized Data Distribution Iteration2022-06-07-
6GDI-I344.94NoGeneralized Data Distribution Iteration2022-06-07-
7MuZero (Res2 Adam)41.66NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
8UCT39.4NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
9Ape-X33NoDistributed Prioritized Experience Replay2018-03-02Code
10DreamerV226NoMastering Atari with Discrete World Models2020-10-05Code
11FQF17.3NoFully Parameterized Quantile Function for Distri...2019-11-05Code
12DNA7.2NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
13CGP4NoEvolving simple programs for playing Atari games2018-06-14Code
14IMPALA (deep)3.48NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
15NoisyNet-Dueling3NoNoisy Networks for Exploration2017-06-30Code
16Prior noop1.3NoPrioritized Experience Replay2015-11-18Code
17Duel noop0.5NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
18Prior+Duel hs0.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
19IQN0.2NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
20Prior hs-0.2NoPrioritized Experience Replay2015-11-18Code
21Persistent AL-0.25NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
22Prior+Duel noop-0.4NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
23Advantage Learning-1.24NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
24Duel hs-1.3NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
25Bootstrapped DQN-1.3NoDeep Exploration via Bootstrapped DQN2016-02-15Code
26Nature DQN-1.6No--Code
27DQN hs-1.6NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
28Gorila-1.7NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
29A3C LSTM hs-1.7NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
30QR-DQN-1-1.7NoDistributional Reinforcement Learning with Quant...2017-10-27Code
31DQN noop-1.9NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
32A2C + SIL-2.4NoSelf-Imitation Learning2018-06-14Code
33DDQN (tuned) hs-2.5NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
34DDQN (tuned) noop-2.7NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
35A3C FF hs-2.8NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
36SARSA-3.2No---
37C51 noop-3.5NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
38ASL DDQN-3.6NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
39DDQN+Pop-Art noop-4.1NoLearning values across many orders of magnitude2016-02-24-
40ES FF (1 hour) noop-4.1NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
41POP3D-4.12NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
42A3C FF (1 day) hs-4.7NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
43Best Learner-9.5NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code