TY - GEN
T1 - DC-MPQ
T2 - 4th IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022
AU - Lee, Seungjin
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Quantization is a representative network compression technique that reduces the number of computational operations and memory accesses in the computation process of convolutional neural networks (CNNs). The existing naïve quantization method has a problem in that the quantization point corresponding to the near-zero value decreases as the precision decreases; as a result, the quantization error increases. Recent quantization-related studies have suggested various solutions to this problem. Nevertheless, studies that suggest a method to solve this problem by considering the characteristics of hardware accelerator implementation have not been actively conducted. To address this problem, this study proposes a method of using standard deviation values, which are simple statistical values of distribution for each layer, as clipping points and setting a scale factor with a clipping point as the base to quantize the weights into a mixed-precision integer format of 4-bit/8-bit. The proposed technique can be applied to any network without additional training, and only biasing and mapping are performed based on the pre-stored standard deviation values; thus, the computational complexity is low, rendering it hardware-friendly. Experimental results indicate that the proposed mixed-precision quantization of the weights of ResNet-18 on ImageNet achieved an effect of reducing the weight capacity by 84% with a 0.34% Top-1 accuracy drop compared to full precision. In YOLACT, an instance segmentation model using a ResNet-50 backbone, on MS COCO, a weight capacity reduction of 81.7% was achieved with only 0.27% and 0.19% drops in box mean average precision (mAP) and mask mAP, respectively.
AB - Quantization is a representative network compression technique that reduces the number of computational operations and memory accesses in the computation process of convolutional neural networks (CNNs). The existing naïve quantization method has a problem in that the quantization point corresponding to the near-zero value decreases as the precision decreases; as a result, the quantization error increases. Recent quantization-related studies have suggested various solutions to this problem. Nevertheless, studies that suggest a method to solve this problem by considering the characteristics of hardware accelerator implementation have not been actively conducted. To address this problem, this study proposes a method of using standard deviation values, which are simple statistical values of distribution for each layer, as clipping points and setting a scale factor with a clipping point as the base to quantize the weights into a mixed-precision integer format of 4-bit/8-bit. The proposed technique can be applied to any network without additional training, and only biasing and mapping are performed based on the pre-stored standard deviation values; thus, the computational complexity is low, rendering it hardware-friendly. Experimental results indicate that the proposed mixed-precision quantization of the weights of ResNet-18 on ImageNet achieved an effect of reducing the weight capacity by 84% with a 0.34% Top-1 accuracy drop compared to full precision. In YOLACT, an instance segmentation model using a ResNet-50 backbone, on MS COCO, a weight capacity reduction of 81.7% was achieved with only 0.27% and 0.19% drops in box mean average precision (mAP) and mask mAP, respectively.
KW - convolutional neural network
KW - distributional clipping
KW - hardware-aware quantization
KW - mixed-precision quantization
KW - segmentation
UR - http://www.scopus.com/inward/record.url?scp=85139020082&partnerID=8YFLogxK
U2 - 10.1109/AICAS54282.2022.9869959
DO - 10.1109/AICAS54282.2022.9869959
M3 - Conference contribution
AN - SCOPUS:85139020082
T3 - Proceeding - IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022
SP - 130
EP - 133
BT - Proceeding - IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 13 June 2022 through 15 June 2022
ER -