TY - GEN
T1 - Hardware-friendly Log-scale Quantization for CNNs with Activation Functions Containing Negative Values
AU - Choi, Dahun
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Recently, with the development of deep learning and GPU, various convolutional neural network (CNN)-based object detection studies are accelerating, and accordingly, research on network compression including quantization is being actively conducted. Due to the characteristics of CNN that accompanies multiple multiplication operations, it is difficult for the existing linear scale quantization method to maximize the benefits of network compression in the accelerator implementation process. On the other hand, log scale quantization has a good quantization effect when implementing an accelerator, but causes a relatively large decrease in accuracy. To address this problem, this paper proposes a technique that minimizes the accuracy degradation of CNNs due to quantization by re-scaling the distribution in consideration of the activation function with negative values in the log scale quantization process. As a result of the experiment, the proposed quantization technique can achieve an accuracy improvement of 1.6% compared to the existing log scale quantization, which can reduce hardware resources by more than 87.52% while maintaining a similar accuracy degradation compared to the existing linear scale quantization.
AB - Recently, with the development of deep learning and GPU, various convolutional neural network (CNN)-based object detection studies are accelerating, and accordingly, research on network compression including quantization is being actively conducted. Due to the characteristics of CNN that accompanies multiple multiplication operations, it is difficult for the existing linear scale quantization method to maximize the benefits of network compression in the accelerator implementation process. On the other hand, log scale quantization has a good quantization effect when implementing an accelerator, but causes a relatively large decrease in accuracy. To address this problem, this paper proposes a technique that minimizes the accuracy degradation of CNNs due to quantization by re-scaling the distribution in consideration of the activation function with negative values in the log scale quantization process. As a result of the experiment, the proposed quantization technique can achieve an accuracy improvement of 1.6% compared to the existing log scale quantization, which can reduce hardware resources by more than 87.52% while maintaining a similar accuracy degradation compared to the existing linear scale quantization.
KW - activation function
KW - Convolutional neural networks
KW - Deep learning
KW - object detection
KW - quantization
UR - http://www.scopus.com/inward/record.url?scp=85123367493&partnerID=8YFLogxK
U2 - 10.1109/ISOCC53507.2021.9613921
DO - 10.1109/ISOCC53507.2021.9613921
M3 - Conference contribution
AN - SCOPUS:85123367493
T3 - Proceedings - International SoC Design Conference 2021, ISOCC 2021
SP - 415
EP - 416
BT - Proceedings - International SoC Design Conference 2021, ISOCC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 18th International System-on-Chip Design Conference, ISOCC 2021
Y2 - 6 October 2021 through 9 October 2021
ER -