TY - GEN
T1 - Analysis for Implementing Power-Efficient Convolutional Operators on FPGA Platforms
AU - Park, Juntae
AU - Kim, Hyun
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Recently, research on FPGA-based accelerators for deep learning models has been actively conducted to overcome problems of GPUs such as power consumption, size, and price. FPGAs have the characteristic that hardware resource and power consumption vary depending on how the target logic is configured. This paper proposes a more power-efficient FPGA implementation method by analyzing the hardware resource and power consumption according to the implementation method and precision of the convolution operator, which is the core operation of CNN. As a result, proper utilization of DSP can increase the power efficiency of the convolution operator in FPGA design, and optimal CNN accelerator design is possible through a well-balanced implementation that considers the hardware resources required for the implementation of various operations other than the convolution operation (e.g., batch-normalization, activation function).
AB - Recently, research on FPGA-based accelerators for deep learning models has been actively conducted to overcome problems of GPUs such as power consumption, size, and price. FPGAs have the characteristic that hardware resource and power consumption vary depending on how the target logic is configured. This paper proposes a more power-efficient FPGA implementation method by analyzing the hardware resource and power consumption according to the implementation method and precision of the convolution operator, which is the core operation of CNN. As a result, proper utilization of DSP can increase the power efficiency of the convolution operator in FPGA design, and optimal CNN accelerator design is possible through a well-balanced implementation that considers the hardware resources required for the implementation of various operations other than the convolution operation (e.g., batch-normalization, activation function).
KW - Convolutional neural network (CNN)
KW - Digital signal processing (DSP) block
KW - Field programmable gate array (FPGA)
UR - http://www.scopus.com/inward/record.url?scp=85123806165&partnerID=8YFLogxK
U2 - 10.1109/ICCE-Asia53811.2021.9641973
DO - 10.1109/ICCE-Asia53811.2021.9641973
M3 - Conference contribution
AN - SCOPUS:85123806165
T3 - 2021 IEEE International Conference on Consumer Electronics-Asia, ICCE-Asia 2021
BT - 2021 IEEE International Conference on Consumer Electronics-Asia, ICCE-Asia 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 IEEE International Conference on Consumer Electronics-Asia, ICCE-Asia 2021
Y2 - 1 November 2021 through 3 November 2021
ER -