TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Atari Games/Atari 2600 Freeway

Atari Games on Atari 2600 Freeway

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1TRPO-hash34No#Exploration: A Study of Count-Based Exploration...2016-11-15Code
2IQN34NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
3NoisyNet-Dueling34NoNoisy Networks for Exploration2017-06-30Code
4QR-DQN-134NoDistributional Reinforcement Learning with Quant...2017-10-27Code
5Go-Explore34NoFirst return, then explore2020-04-27Code
6GDI-I334NoGDI: Rethinking What Makes Reinforcement Learnin...2021-06-11-
7GDI-I334NoGDI: Rethinking What Makes Reinforcement Learnin...2021-06-11-
8GDI-H3(200M frames)34NoGeneralized Data Distribution Iteration2022-06-07-
9GDI-H334NoGeneralized Data Distribution Iteration2022-06-07-
10C51 noop33.9NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
11Bootstrapped DQN33.9NoDeep Exploration via Bootstrapped DQN2016-02-15Code
12ASL DDQN33.9NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
13MuZero (Res2 Adam)33.87NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
14Prior noop33.7NoPrioritized Experience Replay2015-11-18Code
15Ape-X33.7NoDistributed Prioritized Experience Replay2018-03-02Code
16DDQN+Pop-Art noop33.4NoLearning values across many orders of magnitude2016-02-24-
17DDQN (tuned) noop33.3NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
18MuZero33.03NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
19Prior+Duel noop33NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
20DQN-CTS33NoCount-Based Exploration with Neural Density Models2017-03-03Code
21DreamerV233NoMastering Atari with Discrete World Models2020-10-05Code
22DNA33NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
23Intrinsic Reward Agent32.8NoLarge-Scale Study of Curiosity-Driven Learning2018-08-13Code
24Agent5732.59NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
25R2D232.5No--Code
26Persistent AL32.3NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
27A2C + SIL32.2NoSelf-Imitation Learning2018-06-14Code
28Advantage Learning31.72NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
29DQN-PixelCNN31.7NoCount-Based Exploration with Neural Density Models2017-03-03Code
30ES FF (1 hour) noop31NoEvolution Strategies as a Scalable Alternative t...2017-03-10Code
31DQN noop30.8NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
32A3C-CTS30.48NoUnifying Count-Based Exploration and Intrinsic M...2016-06-06Code
33Nature DQN30.3No--Code
34Sarsa-ε29.9NoCount-Based Exploration in Feature Space for Rei...2017-06-25Code
35DQNMMCe29.5NoCount-Based Exploration with the Successor Repre...2018-07-31Code
36Discrete Latent Space World Model (VQ-VAE)29NoSmaller World Models for Reinforcement Learning2020-10-12-
37Prior hs28.9NoPrioritized Experience Replay2015-11-18Code
38DDQN (tuned) hs28.8NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
39Prior+Duel hs28.2NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
40CGP28.2NoEvolving simple programs for playing Atari games2018-06-14Code
41CURL27.9NoCURL: Contrastive Unsupervised Representations f...2020-04-08Code
42MP-EB27NoIncentivizing Exploration In Reinforcement Learn...2015-07-03Code
43DQN hs26.9NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
44Best Baseline22.5NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
45ENAS22NoOptimizing the Neural Architecture of Reinforcem...2020-11-30Code
46SPOS22NoOptimizing the Neural Architecture of Reinforcem...2020-11-30Code
47POP3D21.21NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
48SARSA19.7No---
49Best Learner19.1NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
50Gorila10.2NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
51SAC4.4NoSoft Actor-Critic for Discrete Action Settings2019-10-16Code
52UCT0.4NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
53Duel hs0.2NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
54A3C FF (1 day) hs0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
55A3C FF hs0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
56A3C LSTM hs0.1NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
57Duel noop0No--Code
58Sarsa-φ-EB0No--Code
59IMPALA (deep)0No--Code