Passer à la navigation principale Passer à la recherche Passer au contenu principal

TEP-ones: A simple yet effective approach for transferability estimation of pruned backbones

  • University of Turin
  • Institut Polytechnique de Paris
  • Links Foundation

Résultats de recherche: Contribution à un journalArticleRevue par des pairs

Résumé

In deep learning, the conventional transfer learning paradigm involves fine-tuning a model pre-trained on a complex source task to adapt it to a simpler target task, capitalizing on abundant training data. Concurrently, the paradigm of neural network pruning has emerged as a powerful strategy for enhancing model efficiency, reducing complexity, and optimizing resource utilization. This paper focuses on pruned model transferability estimation for resource-constraint scenarios, where the goal is to rank the performance of pruned pre-trained models on a downstream task without fine-tuning. To this end, from a formal analysis of the intra-class mutual information between samples belonging to the same target class, we observe that, as pruning increases, a sweet phase naturally arises, where the model benefits from better features at the encoder's output. From this, we derive a Transferability Estimation for Pruned Backbones (TEP-ones) that eases the choice of which pruned model (without the need to train the classifier) is the best candidate for transfer learning. We publicly released the code and pre-trained pruned models at https://github.com/EIDOSLAB/TEP-ones.

langue originaleAnglais
Numéro d'article132209
journalNeurocomputing
Volume668
Les DOIs
étatPublié - 1 mars 2026

Empreinte digitale

Examiner les sujets de recherche de « TEP-ones: A simple yet effective approach for transferability estimation of pruned backbones ». Ensemble, ils forment une empreinte digitale unique.

Contient cette citation