Activation Distribution-based Layer-wise Quantization for Convolutional Neural Networks

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

As convolutional neural network (CNN) research accelerates with advances in GPU s, the accuracy of CNN models has been continuously improved. However, in proportion to the enhancement of model accuracy, the computational amount of the CNN models increases, which causes a problem that it is difficult to practically use the CNN models in mobile/embedded platforms. To address this problem, optimization and weight reduction methods for CNN models have been actively studied. This paper proposes a new scale factor for layer-specific quantization considering the activation distribution of CNN s. The proposed method has the advantage that it is possible to minimize the accuracy drop for each layer and is friendly to hardware accelerator design. As a result, the proposed quantization method achieves much higher accuracy compared to the quantization studies on the conventional accelerator design while maintaining low hardware resources.

Original languageEnglish
Title of host publication2022 International Conference on Electronics, Information, and Communication, ICEIC 2022
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781665409346
DOIs
StatePublished - 2022
Event2022 International Conference on Electronics, Information, and Communication, ICEIC 2022 - Jeju, Korea, Republic of
Duration: 6 Feb 20229 Feb 2022

Publication series

Name2022 International Conference on Electronics, Information, and Communication, ICEIC 2022

Conference

Conference2022 International Conference on Electronics, Information, and Communication, ICEIC 2022
Country/TerritoryKorea, Republic of
CityJeju
Period6/02/229/02/22

Keywords

  • Activation quantization
  • convolutional neural networks
  • layer-wise
  • scale factor

Fingerprint

Dive into the research topics of 'Activation Distribution-based Layer-wise Quantization for Convolutional Neural Networks'. Together they form a unique fingerprint.

Cite this