TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Playing Games/Video Games/Atari 2600 Kung-Fu Master

Video Games on Atari 2600 Kung-Fu Master

Metric: Score (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Score▼Extra DataPaperDate↕Code
1GDI-H31666665NoGeneralized Data Distribution Iteration2022-06-07-
2GDI-H3 (200M)1666000No---
3R2D2233413.3No--Code
4Agent57206845.82NoAgent57: Outperforming the Atari Human Benchmark2020-03-30Code
5MuZero204824NoMastering Atari, Go, Chess and Shogi by Planning...2019-11-19Code
6GDI-I3140440NoGeneralized Data Distribution Iteration2022-06-07-
7MuZero (Res2 Adam)116726.96NoOnline and Offline Reinforcement Learning by Pla...2021-04-13Code
8FQF111138.5NoFully Parameterized Quantile Function for Distri...2019-11-05Code
9DNA110962NoDNA: Proximal Policy Optimization with a Dual Ne...2022-06-20Code
10Ape-X97829.5NoDistributed Prioritized Experience Replay2018-03-02Code
11ASL DDQN85182NoTrain a Real-world Local Path Planner in One Hou...2023-05-07Code
12QR-DQN-176642NoDistributional Reinforcement Learning with Quant...2017-10-27Code
13IQN73512NoImplicit Quantile Networks for Distributional Re...2018-06-14Code
14DreamerV262741NoMastering Atari with Discrete World Models2020-10-05Code
15CGP57400NoEvolving simple programs for playing Atari games2018-06-14Code
16UCT48854.5NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
17Prior+Duel noop48375NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
18C51 noop48192NoA Distributional Perspective on Reinforcement Le...2017-07-21Code
19IMPALA (deep)43375.5NoIMPALA: Scalable Distributed Deep-RL with Import...2018-02-05Code
20NoisyNet-Dueling41672NoNoisy Networks for Exploration2017-06-30Code
21A3C LSTM hs40835NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
22Prior noop39581NoPrioritized Experience Replay2015-11-18Code
23Prior+Duel hs37484NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
24Bootstrapped DQN36733.3NoDeep Exploration via Bootstrapped DQN2016-02-15Code
25Persistent AL34650.91NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
26A2C + SIL34449.2NoSelf-Imitation Learning2018-06-14Code
27DDQN+Pop-Art noop34393NoLearning values across many orders of magnitude2016-02-24-
28Duel noop34294NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
29POP3D33728NoPolicy Optimization With Penalized Point Probabi...2018-07-02Code
30Advantage Learning32182.99NoIncreasing the Action Gap: New Operators for Rei...2015-12-15Code
31Prior hs31676NoPrioritized Experience Replay2015-11-18Code
32DDQN (tuned) hs30207NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
33DDQN (tuned) noop29710NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
34SARSA29151No---
35A3C FF hs28819NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code
36DQN noop26059NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
37Duel hs24288NoDueling Network Architectures for Deep Reinforce...2015-11-20Code
38Nature DQN23270No--Code
39DQN hs20882NoDeep Reinforcement Learning with Double Q-learning2015-09-22Code
40Gorila20620NoMassively Parallel Methods for Deep Reinforcemen...2015-07-15Code
41Best Learner19544NoThe Arcade Learning Environment: An Evaluation P...2012-07-19Code
42CURL14280NoCURL: Contrastive Unsupervised Representations f...2020-04-08Code
43A3C FF (1 day) hs3046NoAsynchronous Methods for Deep Reinforcement Lear...2016-02-04Code