LOss-Based SensiTivity rEgulaRization: Towards deep sparse neural networks

Research output: Contribution to journalArticlepeer-review

Abstract

LOBSTER (LOss-Based SensiTivity rEgulaRization) is a method for training neural networks having a sparse topology. Let the sensitivity of a network parameter be the variation of the loss function with respect to the variation of the parameter. Parameters with low sensitivity, i.e. having little impact on the loss when perturbed, are shrunk and then pruned to sparsify the network. Our method allows to train a network from scratch, i.e. without preliminary learning or rewinding. Experiments on multiple architectures and datasets show competitive compression ratios with minimal computational overhead.

Original languageEnglish
Pages (from-to)230-237
Number of pages8
JournalNeural Networks
Volume146
DOIs
Publication statusPublished - 1 Feb 2022

Keywords

  • Deep learning
  • Pruning
  • Regularization
  • Sparsity

Fingerprint

Dive into the research topics of 'LOss-Based SensiTivity rEgulaRization: Towards deep sparse neural networks'. Together they form a unique fingerprint.

Cite this