Abstract

Recently, convolutional neural networks (CNN) have demonstrated impressive performance in various computer vision tasks. However, high performance hardware is typically indispensable for the application of CNN models due to the high computation complexity, which prohibits their further extensions. In this paper, we propose an efficient framework, namely Quantized CNN, to simultaneously speed-up the computation and reduce the storage and memory overhead of CNN models. Both filter kernels in convolutional layers and weighting matrices in fully-connected layers are quantized, aiming at minimizing the estimation error of each layer's response. Extensive experiments on the ILSVRC-12 benchmark demonstrate 4 ~ 6× speed-up and 15 ~ 20× compression with merely one percentage loss of classification accuracy. With our quantized CNN model, even mobile devices can accurately classify images within one second.

Keywords

Computer scienceConvolutional neural networkComputationBenchmark (surveying)Overhead (engineering)WeightingMobile deviceConvolutional codeArtificial intelligenceSpeedupPattern recognition (psychology)Computer engineeringAlgorithmParallel computingDecoding methods

Affiliated Institutions

Related Publications

Publication Info

Year
2016
Type
preprint
Pages
4820-4828
Citations
1228
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1228
OpenAlex

Cite This

Jiaxiang Wu, Cong Leng, Yuhang Wang et al. (2016). Quantized Convolutional Neural Networks for Mobile Devices. , 4820-4828. https://doi.org/10.1109/cvpr.2016.521

Identifiers

DOI
10.1109/cvpr.2016.521