TY - GEN
T1 - Hardware-Friendly Quantization via Outlier Scaling in Convolution-Attention-Based Hybrid Networks
AU - Kim, Nam Joon
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Hybrid networks that combine convolution and attention have achieved state-of-the-art performance in computer vision tasks. Quantization is a promising compression method to efficiently utilize these hybrid networks in resource-constrained consumer electronics like IoT devices. However, although many hybrid networks have been proposed, research on quantization dedicated to hybrid networks has not yet been actively conducted. To bridge this gap, we propose a novel hardware-friendly post-training quantization method. Initially, we observe significant outliers in the bottleneck blocks of hybrid networks, which result in severe accuracy degradation due to quantization. To effectively address these outliers, we propose not only a novel outlier scaling method but also an objective function-based power-of-two approximation method that replaces conventional floating-point multiplication with hardware-friendly shift operations. To demonstrate the effectiveness of the proposed method, experiments were conducted on the ImageNet-1k dataset using a representative hybrid network, MobileViT. Our proposed method significantly mitigated the accuracy drop with a small parameter increase at the same model size compared to existing quantization methods.
AB - Hybrid networks that combine convolution and attention have achieved state-of-the-art performance in computer vision tasks. Quantization is a promising compression method to efficiently utilize these hybrid networks in resource-constrained consumer electronics like IoT devices. However, although many hybrid networks have been proposed, research on quantization dedicated to hybrid networks has not yet been actively conducted. To bridge this gap, we propose a novel hardware-friendly post-training quantization method. Initially, we observe significant outliers in the bottleneck blocks of hybrid networks, which result in severe accuracy degradation due to quantization. To effectively address these outliers, we propose not only a novel outlier scaling method but also an objective function-based power-of-two approximation method that replaces conventional floating-point multiplication with hardware-friendly shift operations. To demonstrate the effectiveness of the proposed method, experiments were conducted on the ImageNet-1k dataset using a representative hybrid network, MobileViT. Our proposed method significantly mitigated the accuracy drop with a small parameter increase at the same model size compared to existing quantization methods.
KW - Convolution neural network
KW - Hybrid network
KW - IoT devices
KW - Quantization
KW - Transformer
UR - https://www.scopus.com/pages/publications/105006561767
U2 - 10.1109/ICCE63647.2025.10930053
DO - 10.1109/ICCE63647.2025.10930053
M3 - Conference contribution
AN - SCOPUS:105006561767
T3 - Digest of Technical Papers - IEEE International Conference on Consumer Electronics
BT - 2025 IEEE International Conference on Consumer Electronics, ICCE 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2025 IEEE International Conference on Consumer Electronics, ICCE 2025
Y2 - 11 January 2025 through 14 January 2025
ER -