Multi-GPU version of the A3C algorithm in Asynchronous Methods for Deep Reinforcement Learning.
Results of the same code trained on 47 different Atari games were uploaded to OpenAI Gym. Most of them were the best reproducible results on gym. However OpenAI has later completely removed leaderboard from their site.
./train-atari.py --env Breakout-v0 --gpu 0
In each iteration it trains on a batch of 128 new states.
The speed is about 6~10 iterations/s on 1 GPU plus 12+ CPU cores.
With 2 TitanX + 20+ CPU cores, by setting SIMULATOR_PROC=240, PREDICT_BATCH_SIZE=30, PREDICTOR_THREAD_PER_GPU=6
, it can improve to 16 it/s (2K images/s).
Note that the network architecture is larger than what's used in the original paper.
The pretrained models are all trained with 4 GPUs for about 2 days. But on simple games like Breakout, you can get good performance within several hours. Also note that multi-GPU doesn't give you obvious speedup here, because the bottleneck in this implementation is not computation but simulation.
Some practicical notes:
- Prefer Python 3; Windows not supported.
- Occasionally, processes may not get terminated completely. It is suggested to use
systemd-run
to run any multiprocess Python program to get a cgroup dedicated for the task. - Training with a significant slower speed (e.g. on CPU) will result in very bad score, probably because of the slightly off-policy implementation.
Download models from model zoo.
Watch the agent play:
./train-atari.py --task play --env Breakout-v0 --load Breakout-v0.npz
Dump some videos:
./train-atari.py --task dump_video --load Breakout-v0.npz --env Breakout-v0 --output output_dir --episode 3
This table lists available pretrained models and scores (average over 100 episodes), with their submission links. The site is not maintained any more so the links might become invalid any time.
AirRaid(2727) | Alien (2611) | Amidar(1376) | Assault(3397) |
Asterix(407432) | Asteroids(1965) | Atlantis(217186) | BankHeist(1274) |
BattleZone(29210) | BeamRider(5972) | Berzerk(2289) | Breakout (667) |
Carnival(5211) | Centipede(2909) | ChopperCommand(6031) | CrazyClimber(105297) |
DemonAttack(33992) | DoubleDunk(23) | ElevatorAction(11377) | FishingDerby(34) |
Frostbite(6824) | Gopher(22595) | Gravitar(2144) | IceHockey(19) |
Jamesbond(640) | JourneyEscape(-407) | Kangaroo(6540) | Krull(6100) |
KungFuMaster(34767) | MsPacman(5738) | NameThisGame(15321) | Phoenix(75312) |
Pong(21) | Pooyan(5607) | Qbert(20182) | Riverraid(14185) |
RoadRunner(60615) | Robotank(60) | Seaquest(46890) | SpaceInvaders(3454) |
StarGunner(93480) | Tennis(23) | Tutankham(275) | UpNDown(92163) |
VideoPinball(140156) | WizardOfWor(3824) | Zaxxon(32894) |
All models above are trained with the -v0
variant of atari games.
Note that this variant is quite different from DeepMind papers, so the scores are not directly comparable.
The most notable differences are:
- Each action is randomly repeated 2~4 times.
- Inputs are RGB instead of greyscale.
- An episode is limited to 10000 steps.
- Lost of live is not end of episode.
Also see the DQN implementation here