Feature Map-Aware Activation Quantization for Low-bit Neural Networks

Seungjin Lee, Hyun Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

Quantization, the most popular deep neural network (DNN) compression method, can reduce the computational complexity and save a lot of memory resources by converting the existing 32-bit floating point values to low-bit integer point values. However, as DNNs are widely used in mobile and edge devices, which have relatively less hardware resources, there are demands for more aggressive quantization methods. To meet these needs, this paper introduces a dedicated method that divides activation maps of DNNs into several regions according to the activation size and quantizes them to 4-bit by setting scale factors adaptively for each region. As a result of applying the proposed method to the backbone of YOLACT, a representative instance segmentation model, the proposed method achieves approximately 2% increase in both box and mask mAPs compared to the naive 4-bit activation quantization.

Original languageEnglish
Title of host publication2021 36th International Technical Conference on Circuits/Systems, Computers and Communications, ITC-CSCC 2021
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781665435536
DOIs
StatePublished - 27 Jun 2021
Event36th International Technical Conference on Circuits/Systems, Computers and Communications, ITC-CSCC 2021 - Jeju, Korea, Republic of
Duration: 27 Jun 202130 Jun 2021

Publication series

Name2021 36th International Technical Conference on Circuits/Systems, Computers and Communications, ITC-CSCC 2021

Conference

Conference36th International Technical Conference on Circuits/Systems, Computers and Communications, ITC-CSCC 2021
Country/TerritoryKorea, Republic of
CityJeju
Period27/06/2130/06/21

Keywords

  • Deep learning
  • Low-precision
  • Quantization
  • ResNet50
  • YOLACT

Fingerprint

Dive into the research topics of 'Feature Map-Aware Activation Quantization for Low-bit Neural Networks'. Together they form a unique fingerprint.

Cite this