Abstract

Despite their massive size, successful deep artificial neural networks can exhibit a remarkably small gap between training and test performance. Conventional wisdom attributes small generalization error either to properties of the model family or to the regularization techniques used during training. Through extensive systematic experiments, we show how these traditional approaches fail to explain why large neural networks generalize well in practice. Specifically, our experiments establish that state-of-the-art convolutional networks for image classification trained with stochastic gradient methods easily fit a random labeling of the training data. This phenomenon is qualitatively unaffected by explicit regularization and occurs even if we replace the true images by completely unstructured random noise. We corroborate these experimental findings with a theoretical construction showing that simple depth two neural networks already have perfect finite sample expressivity as soon as the number of parameters exceeds the number of data points as it usually does in practice. We interpret our experimental findings by comparison with traditional models. We supplement this republication with a new section at the end summarizing recent progresses in the field since the original version of this paper.

Keywords

Regularization (linguistics)Computer scienceGeneralizationArtificial intelligenceArtificial neural networkDeep neural networksDeep learningEarly stoppingConvolutional neural networkMachine learningStochastic gradient descentAlgorithmMathematics

Affiliated Institutions

Related Publications

Publication Info

Year
2021
Type
article
Volume
64
Issue
3
Pages
107-115
Citations
2043
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

2043
OpenAlex
91
Influential

Cite This

Chiyuan Zhang, Samy Bengio, Moritz Hardt et al. (2021). Understanding deep learning (still) requires rethinking generalization. Communications of the ACM , 64 (3) , 107-115. https://doi.org/10.1145/3446776

Identifiers

DOI
10.1145/3446776

Data Quality

Data completeness: 86%