TY - JOUR
T1 - Survey of convolutional neural network accelerators on field-programmable gate array platforms
T2 - architectures and optimization techniques
AU - Hong, Hyeonseok
AU - Choi, Dahun
AU - Kim, Namjoon
AU - Lee, Haein
AU - Kang, Beomjin
AU - Kang, Huibeom
AU - Kim, Hyun
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2024.
PY - 2024/5
Y1 - 2024/5
N2 - With the recent advancements in high-performance computing, convolutional neural networks (CNNs) have achieved remarkable success in various vision tasks. However, along with improvements in model accuracy, the size and computational complexity of the models have significantly increased with the increasing number of parameters. Although graphics processing unit (GPU) platforms equipped with high-performance memory and specialized in parallel processing are commonly used for CNN processing, the significant power consumption presents challenges in their utilization on edge devices. To address these issues, research is underway to design CNN models using field-programmable gate arrays (FPGAs) as accelerators. FPGAs provide a high level of flexibility, allowing efficient optimization of convolution operations, which account for a significant portion of the CNN computations. Additionally, FPGAs are known for their low power consumption compared to GPUs, making them a promising energy-efficient platform. In this paper, we review and summarize various approaches and techniques related to the design of FPGA-based CNN accelerators. Specifically, to comprehensively study CNN accelerators, we investigate the advantages and disadvantages of various methods for optimizing CNN accelerators and previously designed efficient accelerator architectures. We expect this paper to serve as an important guideline for future hardware research in artificial intelligence.
AB - With the recent advancements in high-performance computing, convolutional neural networks (CNNs) have achieved remarkable success in various vision tasks. However, along with improvements in model accuracy, the size and computational complexity of the models have significantly increased with the increasing number of parameters. Although graphics processing unit (GPU) platforms equipped with high-performance memory and specialized in parallel processing are commonly used for CNN processing, the significant power consumption presents challenges in their utilization on edge devices. To address these issues, research is underway to design CNN models using field-programmable gate arrays (FPGAs) as accelerators. FPGAs provide a high level of flexibility, allowing efficient optimization of convolution operations, which account for a significant portion of the CNN computations. Additionally, FPGAs are known for their low power consumption compared to GPUs, making them a promising energy-efficient platform. In this paper, we review and summarize various approaches and techniques related to the design of FPGA-based CNN accelerators. Specifically, to comprehensively study CNN accelerators, we investigate the advantages and disadvantages of various methods for optimizing CNN accelerators and previously designed efficient accelerator architectures. We expect this paper to serve as an important guideline for future hardware research in artificial intelligence.
KW - Accelerator
KW - Convolutional neural network
KW - Data flow
KW - Design optimization
KW - Field-programmable gate array (FPGA)
UR - http://www.scopus.com/inward/record.url?scp=85189504190&partnerID=8YFLogxK
U2 - 10.1007/s11554-024-01442-8
DO - 10.1007/s11554-024-01442-8
M3 - Article
AN - SCOPUS:85189504190
SN - 1861-8200
VL - 21
JO - Journal of Real-Time Image Processing
JF - Journal of Real-Time Image Processing
IS - 3
M1 - 64
ER -