SRU-Q: Hardware-friendly Stochastic Rounding Unit-based Gradient Quantization for CNN Training

Sangbeom Jeong, Dahun Choi, Hyun Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Scopus citations

Abstract

Quantization in convolutional neural networks (CNNs) involves using a low precision to decrease convolution operation costs, resulting in reduced power consumption and faster network performance. Notably, gradient quantization plays a crucial role in CNN training accelerators, as backpropagation typically incurs higher computational costs for calculating weight gradients than forward propagation does. Stochastic rounding (SR), which employs random number generation, is recognized as an effective method for stabilizing backpropagation quantization. However, the process of generating random numbers in hardware has significant drawbacks - notably high computational costs and substantial difficulties in implementation. This paper introduces a technique for efficient SR using a hardware-optimized random number generator, termed linear feedback shift register-bitwise-stochastic rounding unit (LBSRU). The LBSRU efficiently conducts SR with a small amount of random number generation and adapts to various network types by altering the random number generation approach for different batch sizes. Specifically, we designed and synthesized our method on the FPGA platform to create a prototype. A comparison with previous studies revealed that our method requires significantly fewer resources: 98.19% fewer lookup tables (LUTs) and 98.38% fewer flip-flops (FFs).

Original languageEnglish
Title of host publication2024 IEEE 6th International Conference on AI Circuits and Systems, AICAS 2024 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages457-461
Number of pages5
ISBN (Electronic)9798350383638
DOIs
StatePublished - 2024
Event6th IEEE International Conference on AI Circuits and Systems, AICAS 2024 - Abu Dhabi, United Arab Emirates
Duration: 22 Apr 202425 Apr 2024

Publication series

Name2024 IEEE 6th International Conference on AI Circuits and Systems, AICAS 2024 - Proceedings

Conference

Conference6th IEEE International Conference on AI Circuits and Systems, AICAS 2024
Country/TerritoryUnited Arab Emirates
CityAbu Dhabi
Period22/04/2425/04/24

Keywords

  • convolutional neural networks
  • FPGA
  • gradient quantization
  • random number generator
  • stochastic rounding

Fingerprint

Dive into the research topics of 'SRU-Q: Hardware-friendly Stochastic Rounding Unit-based Gradient Quantization for CNN Training'. Together they form a unique fingerprint.

Cite this