HLQ: Hardware-Friendly Logarithmic Quantization Aware Training for Power-Efficient Low-Precision CNN Models

Dahun Choi, Juntae Park, Hyun Kim

Research output: Contribution to journalArticlepeer-review

2 Scopus citations

Abstract

With the development of deep learning and graphics processing units (GPUs), various convolutional neural network (CNN)-based computer vision studies have been conducted. Because numerous computations are involved in the inference and training process of CNNs, research on network compression, including quantization, is being actively conducted along with the use of CNNs. Unlike the existing linear quantization, logarithmic quantization has the advantage that the multiply-accumulate (MAC) operation in the convolution (CONV) operation, which occupies most of the CNNs, can be replaced with the addition operation and is suitable for low-precision quantization. In this paper, we propose a logarithmic quantization aware training technique that effectively reduces quantization loss while maximizing the effect of reducing hardware resources and power consumption in the forward and backward propagation processes of the CNN. The proposed method minimizes the accuracy drop by allocating the rounding point with the least quantization loss for each specific training in the forward pass and propagates the optimized gradient by scaling the gradient of parameters with a high quantization loss in the backward pass. As a result of scratch training on the Tiny-ImageNet dataset using ResNet-18, 34, and 50, where both weights and activations are quantized to 4-bits through the proposed method, an improvement in accuracy of 0.88%, 0.48%, and 1.72%, respectively, can be achieved compared to that of the baseline (i.e., full-precision). In addition, as a result of synthesizing the CONV acceleration unit of ResNet-18 through RTL implementation, the proposed 4-bit quantization can achieve a power saving of 82.3% compared to the baseline (i.e., full-precision) when computing ResNet-18.

Original languageEnglish
Pages (from-to)159611-159621
Number of pages11
JournalIEEE Access
Volume12
DOIs
StatePublished - 2024

Keywords

  • Logarithmic quantization
  • convolutional neural network
  • low-power
  • network compression

Fingerprint

Dive into the research topics of 'HLQ: Hardware-Friendly Logarithmic Quantization Aware Training for Power-Efficient Low-Precision CNN Models'. Together they form a unique fingerprint.

Cite this