Extreme Pruning Technique Based on Filter Deactivation Using Sparsity Training for Deep Convolutional Neural Networks

Kwanghyun Koo, Hyun Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Pruning represents a significant compression performance in deep convolutional neural networks and has been a subject of research in two major directions: filter pruning and weight pruning. Filter pruning, as a method for eliminating 3-dimensional parameters known as filters, exhibits notable advantages in terms of high acceleration performance. However, it presents challenges when applied to complex network structures and may yield relatively lower compression performance. On the other hand, weight pruning, which involves the removal of 1-dimensional parameters, proves to be a viable option for complex network structures due to its ability to achieve high compression performance. Nevertheless, it does have limitations when it comes to achieving acceleration in a general GPU environment. To solve this problem, a kernel pruning method was proposed that removes the 2-dimensional parameter kernels. Kernel pruning has the advantage of being easier to accelerate in hardware than weight pruning while achieving a relatively high pruning rate. Nonetheless, in the process of employing kernel pruning, a phenomenon known as filter deactivation emerges, where certain filters generate outputs but remain entirely unused in the subsequent layer. In this paper, we present a novel approach to create more deactivated filters through sparsity training by taking advantage of the fact that deactivated filters do not affect the performance of the network even if they are removed, and consequently achieve a higher pruning rate. By applying the proposed method for kernel pruning, we achieve a performance enhancement of 0.22%, while successfully removing an additional 2.58% of parameters in ResNet-110 when evaluated on the CIFAR-10 dataset.

Original languageEnglish
Title of host publication2024 International Conference on Electronics, Information, and Communication, ICEIC 2024
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350371888
DOIs
StatePublished - 2024
Event2024 International Conference on Electronics, Information, and Communication, ICEIC 2024 - Taipei, Taiwan, Province of China
Duration: 28 Jan 202431 Jan 2024

Publication series

Name2024 International Conference on Electronics, Information, and Communication, ICEIC 2024

Conference

Conference2024 International Conference on Electronics, Information, and Communication, ICEIC 2024
Country/TerritoryTaiwan, Province of China
CityTaipei
Period28/01/2431/01/24

Keywords

  • Convolutional neural network
  • Kernel pruning
  • Network compression
  • Sparsity training

Fingerprint

Dive into the research topics of 'Extreme Pruning Technique Based on Filter Deactivation Using Sparsity Training for Deep Convolutional Neural Networks'. Together they form a unique fingerprint.

Cite this