Abstract

We present the first massively distributed architecture for deep reinforcement learning. This architecture uses four main components: parallel actors that generate new behaviour; parallel learners that are trained from stored experience; a distributed neural network to represent the value function or behaviour policy; and a distributed store of experience. We used our architecture to implement the Deep Q-Network algorithm (DQN). Our distributed algorithm was applied to 49 games from Atari 2600 games from the Arcade Learning Environment, using identical hyperparameters. Our performance surpassed non-distributed DQN in 41 of the 49 games and also reduced the wall-time required to achieve these results by an order of magnitude on most games.

Keywords

Reinforcement learningMassively parallelComputer scienceArchitectureHyperparameterArtificial neural networkFunction (biology)Distributed computingDeep learningArtificial intelligenceParallel computing

Affiliated Institutions

Related Publications

Publication Info

Year
2015
Type
preprint
Citations
405
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

405
OpenAlex

Cite This

Arun Sukumaran Nair, P. Srinivasan, Sam Blackwell et al. (2015). Massively Parallel Methods for Deep Reinforcement Learning. arXiv (Cornell University) . https://doi.org/10.48550/arxiv.1507.04296

Identifiers

DOI
10.48550/arxiv.1507.04296