A round-trip journey in pruned artificial neural networks

Research output: Contribution to journalConference articlepeer-review

Abstract

In the last decade, deep learning models competed for performance at the price of tremendous computational costs. Such a critical aspect recently attracted more attention for both the training and inference phases. The latter is obviously orders of magnitude lower than the training complexity, but on the other hand, it contributes many times, which impacts efficiency on edge or embedded devices. Inference can be made efficient through neural network pruning, which consists of parameters and neurons' removal from the model's topology while maintaining the model's accuracy. This results in reduced resource and energy requirements for the models. This paper describes two pruning procedures for lowering the operations required during the inference phase and a method to exploit the resulting sparsity. The same cannot be applied at training time: we show it is possible to borrow similar ideas to reduce the cost of gradient backpropagation by disabling the computation for selected neurons.

Original languageEnglish
Pages (from-to)561-566
Number of pages6
JournalCEUR Workshop Proceedings
Volume3486
Publication statusPublished - 1 Jan 2023
Event2023 Italia Intelligenza Artificiale - Thematic Workshops, Ital-IA 2023 - Pisa, Italy
Duration: 29 May 202330 May 2023

Keywords

  • Deep Learning
  • Efficiency
  • Pruning

Fingerprint

Dive into the research topics of 'A round-trip journey in pruned artificial neural networks'. Together they form a unique fingerprint.

Cite this