ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

This works explores the benefits of structured parameter pruning in the framework of the MPEG standardization efforts for neural network compression. First less relevant parameters are pruned from the network, then remaining parameters are quantized and finally quantized parameters are entropy coded. We consider an unstructured pruning strategy that maximizes the number of pruned parameters at the price of randomly sparse tensors and a structured strategy that prunes fewer parameters yet yields regularly sparse tensors. We show that structured pruning enables better end-to-end compression despite lower pruning ratio because it boosts the efficiency of the arithmetic coder. As a bonus, once decompressed, the network memory footprint is lower as well as its inference time.

Original languageEnglish
Title of host publication2021 IEEE International Conference on Image Processing, ICIP 2021 - Proceedings
PublisherIEEE Computer Society
Pages3527-3531
Number of pages5
ISBN (Electronic)9781665441155
DOIs
Publication statusPublished - 1 Jan 2021
Externally publishedYes
Event28th IEEE International Conference on Image Processing, ICIP 2021 - Anchorage, United States
Duration: 19 Sept 202122 Sept 2021

Publication series

NameProceedings - International Conference on Image Processing, ICIP
Volume2021-September
ISSN (Print)1522-4880

Conference

Conference28th IEEE International Conference on Image Processing, ICIP 2021
Country/TerritoryUnited States
CityAnchorage
Period19/09/2122/09/21

Keywords

  • Compression
  • Deep learning
  • MPEG-7
  • Pruning

Fingerprint

Dive into the research topics of 'ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION'. Together they form a unique fingerprint.

Cite this