GAP safe screening rules for sparse multi-task and multi-class models

Research output: Contribution to journalConference articlepeer-review

Abstract

High dimensional regression benefits from sparsity promoting regularizations. Screening rules leverage the known sparsity of the solution by ignoring some variables in the optimization, hence speeding up solvers. When the procedure is proven not to discard features wrongly the rules are said to be safe. In this paper we derive new safe rules for generalized linear models regularized with ℓ1 and ℓ1/ℓ2 norms. The rules are based on duality gap computations and spherical safe regions whose diameters converge to zero. This allows to discard safely more variables, in particular for low regularization parameters. The GAP Safe rule can cope with any iterative solver and we illustrate its performance on coordinate descent for multi-task Lasso, binary and multinomial logistic regression, demonstrating significant speed ups on all tested datasets with respect to previous safe rules.

Original languageEnglish
Pages (from-to)811-819
Number of pages9
JournalAdvances in Neural Information Processing Systems
Volume2015-January
Publication statusPublished - 1 Jan 2015
Externally publishedYes
Event29th Annual Conference on Neural Information Processing Systems, NIPS 2015 - Montreal, Canada
Duration: 7 Dec 201512 Dec 2015

Fingerprint

Dive into the research topics of 'GAP safe screening rules for sparse multi-task and multi-class models'. Together they form a unique fingerprint.

Cite this