Abstract

I present a new way to parallelize the training of convolutional neural networks across multiple GPUs. The method scales significantly better than all alternatives when applied to modern convolutional neural networks.

Keywords

Computer scienceConvolutional neural networkArtificial intelligenceProgramming languageParallel computing

Related Publications

Neural GPUs Learn Algorithms

Abstract: Learning an algorithm from examples is a fundamental problem that has been widely studied. Recently it has been addressed using neural networks, in particular by Neura...

2016 arXiv (Cornell University) 63 citations

Publication Info

Year
2014
Type
preprint
Citations
980
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

980
OpenAlex

Cite This

Alex Krizhevsky (2014). One weird trick for parallelizing convolutional neural networks. arXiv (Cornell University) . https://doi.org/10.48550/arxiv.1404.5997

Identifiers

DOI
10.48550/arxiv.1404.5997