Cameron Allen, Kavosh Asadi, Melrose Roderick, Abdel-rahman Mohamed, George Konidaris, Michael Littman
We propose a new algorithm, Mean Actor-Critic (MAC), for discrete-action continuous-state reinforcement learning. MAC is a policy gradient algorithm that uses the agent's explicit representation of all action values to estimate the gradient of the policy, rather than using only the actions that were actually executed. We prove that this approach reduces variance in the policy gradient estimate relative to traditional actor-critic methods. We show empirical results on two control domains and on six Atari games, where MAC is competitive with state-of-the-art policy search algorithms.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Continuous Control | Cart Pole (OpenAI Gym) | Score | 178.3 | MAC |
| Continuous Control | Lunar Lander (OpenAI Gym) | Score | 163.5 | MAC |
| Atari Games | Atari 2600 Pong | Score | 10.6 | MAC |
| Atari Games | Atari 2600 Breakout | Score | 372.7 | MAC |
| Atari Games | Atari 2600 Space Invaders | Score | 1173.1 | MAC |
| Atari Games | Atari 2600 Beam Rider | Score | 6072 | MAC |
| Atari Games | Atari 2600 Seaquest | Score | 1703.4 | MAC |
| Atari Games | Atari 2600 Q*Bert | Score | 243.4 | MAC |
| Video Games | Atari 2600 Pong | Score | 10.6 | MAC |
| Video Games | Atari 2600 Breakout | Score | 372.7 | MAC |
| Video Games | Atari 2600 Space Invaders | Score | 1173.1 | MAC |
| Video Games | Atari 2600 Beam Rider | Score | 6072 | MAC |
| Video Games | Atari 2600 Seaquest | Score | 1703.4 | MAC |
| Video Games | Atari 2600 Q*Bert | Score | 243.4 | MAC |
| 3D | Cart Pole (OpenAI Gym) | Score | 178.3 | MAC |
| 3D | Lunar Lander (OpenAI Gym) | Score | 163.5 | MAC |
| 3D Face Modelling | Cart Pole (OpenAI Gym) | Score | 178.3 | MAC |
| 3D Face Modelling | Lunar Lander (OpenAI Gym) | Score | 163.5 | MAC |