TY - GEN
T1 - GDS
T2 - 2024 International Conference on Electronics, Information, and Communication, ICEIC 2024
AU - Choi, Da Hun
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Although recent research in the field of instance segmentation has achieved high accuracy, it suffers from the cost of increased parameters and computation due to the increased model size. To address this issue, various compression techniques are being explored to reduce hardware resources. In particular, quantization is gaining attention as it can significantly reduce hardware resources and power. Recently, quantization studies have expanded from the quantization of the forward pass to also include the backward pass. However, in segmentation tasks, gradient quantization is challenging due to the dynamic distribution across channels, leading to significant quantization errors and making it difficult to achieve high accuracy. In this paper, we propose a technique that applies scaling to layers with a dynamic distribution across channels to address this issue. Additionally, for stability in the initial training, we adopt mixed-precision quantization based on quantization errors. Experimental results show that the proposed method achieves higher accuracy in YOLACT than other quantization methods.
AB - Although recent research in the field of instance segmentation has achieved high accuracy, it suffers from the cost of increased parameters and computation due to the increased model size. To address this issue, various compression techniques are being explored to reduce hardware resources. In particular, quantization is gaining attention as it can significantly reduce hardware resources and power. Recently, quantization studies have expanded from the quantization of the forward pass to also include the backward pass. However, in segmentation tasks, gradient quantization is challenging due to the dynamic distribution across channels, leading to significant quantization errors and making it difficult to achieve high accuracy. In this paper, we propose a technique that applies scaling to layers with a dynamic distribution across channels to address this issue. Additionally, for stability in the initial training, we adopt mixed-precision quantization based on quantization errors. Experimental results show that the proposed method achieves higher accuracy in YOLACT than other quantization methods.
KW - Convolution neural network
KW - Deep learning
KW - Gradient quantization
KW - Instance segmentation
UR - https://www.scopus.com/pages/publications/85189245876
U2 - 10.1109/ICEIC61013.2024.10457260
DO - 10.1109/ICEIC61013.2024.10457260
M3 - Conference contribution
AN - SCOPUS:85189245876
T3 - 2024 International Conference on Electronics, Information, and Communication, ICEIC 2024
BT - 2024 International Conference on Electronics, Information, and Communication, ICEIC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 28 January 2024 through 31 January 2024
ER -