TY - GEN
T1 - LAMP-Q
T2 - 2025 International Conference on Electronics, Information, and Communication, ICEIC 2025
AU - Yoon, Seokkyu
AU - Kim, Namjoon
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Quantization is an effective technique for reducing memory usage and power consumption in deep neural networks (DNNs) by decreasing parameter size. However, conventional quantization methods often lead to significant accuracy loss when applied to compact architectures such as MobileNet. In particular, quantizing MobileNetV3 causes accuracy degradation due to the presence of large outliers. To address this challenge, we propose a hardware-friendly mixed-precision quantization approach. Unlike existing methods, which suffer from low memory and computational efficiency due to the use of diverse bit-widths that do not align with memory address space sizes, our approach applies 8-bit quantization to activations and selectively quantizes weights to 4-, 8-, or 16-bit, depending on the sensitivity of each layer. This strategy not only enhances memory and computational efficiency but also minimizes accuracy degradation. When evaluated on the ImageNet-1k dataset, our proposed method reduces the parameter count of MobileNetV3-small and MobileNetV3-large by 78.31% and 75.61%, respectively, while achieving accuracy drops of only 0.90% and 0.81%.
AB - Quantization is an effective technique for reducing memory usage and power consumption in deep neural networks (DNNs) by decreasing parameter size. However, conventional quantization methods often lead to significant accuracy loss when applied to compact architectures such as MobileNet. In particular, quantizing MobileNetV3 causes accuracy degradation due to the presence of large outliers. To address this challenge, we propose a hardware-friendly mixed-precision quantization approach. Unlike existing methods, which suffer from low memory and computational efficiency due to the use of diverse bit-widths that do not align with memory address space sizes, our approach applies 8-bit quantization to activations and selectively quantizes weights to 4-, 8-, or 16-bit, depending on the sensitivity of each layer. This strategy not only enhances memory and computational efficiency but also minimizes accuracy degradation. When evaluated on the ImageNet-1k dataset, our proposed method reduces the parameter count of MobileNetV3-small and MobileNetV3-large by 78.31% and 75.61%, respectively, while achieving accuracy drops of only 0.90% and 0.81%.
KW - Convolutional Neural Network
KW - Mixed Precision Quantization
KW - MobileNetV3
KW - Sensitivity
UR - http://www.scopus.com/inward/record.url?scp=86000008535&partnerID=8YFLogxK
U2 - 10.1109/ICEIC64972.2025.10879604
DO - 10.1109/ICEIC64972.2025.10879604
M3 - Conference contribution
AN - SCOPUS:86000008535
T3 - 2025 International Conference on Electronics, Information, and Communication, ICEIC 2025
BT - 2025 International Conference on Electronics, Information, and Communication, ICEIC 2025
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 19 January 2025 through 22 January 2025
ER -