Annealed Multiple Choice Learning: Overcoming limitations of Winner-takes-all with annealing

  • David Perera
  • , Victor Letzelter
  • , Théo Mariotte
  • , Adrien Cortés
  • , Mickael Chen
  • , Slim Essid
  • , Gaël Richard

Research output: Contribution to journalConference articlepeer-review

Abstract

We introduce Annealed Multiple Choice Learning (aMCL) which combines simulated annealing with MCL. MCL is a learning framework handling ambiguous tasks by predicting a small set of plausible hypotheses. These hypotheses are trained using the Winner-takes-all (WTA) scheme, which promotes the diversity of the predictions. However, this scheme may converge toward an arbitrarily suboptimal local minimum, due to the greedy nature of WTA. We overcome this limitation using annealing, which enhances the exploration of the hypothesis space during training. We leverage insights from statistical physics and information theory to provide a detailed description of the model training trajectory. Additionally, we validate our algorithm by extensive experiments on synthetic datasets, on the standard UCI benchmark, and on speech separation.

Original languageEnglish
JournalAdvances in Neural Information Processing Systems
Volume37
Publication statusPublished - 1 Jan 2024
Event38th Conference on Neural Information Processing Systems, NeurIPS 2024 - Vancouver, Canada
Duration: 9 Dec 202415 Dec 2024

Fingerprint

Dive into the research topics of 'Annealed Multiple Choice Learning: Overcoming limitations of Winner-takes-all with annealing'. Together they form a unique fingerprint.

Cite this