Convergent Working Set Algorithm for Lasso with Non-Convex Sparse Regularizers

Research output: Contribution to journalConference articlepeer-review

Abstract

Non-convex sparse regularizers are common tools for learning with high-dimensional data. For accelerating convergence of a Lasso problem using those regularizers, a working set strategy addresses the optimization problem through an iterative algorithm by gradually incrementing the number of variables to optimize until the identification of the solution support. We propose in this paper the first Lasso working set algorithm for non-convex sparse regularizers with convergence guarantees. The algorithm, named FireWorks, is based on a non-convex reformulation of a recent duality-based approach and leverages on the geometry of the residuals. We provide theoretical guarantees showing that convergence is preserved even when the inner solver is inexact, under sufficient decay of the error across iterations. Experimental results demonstrate strong computational gain when using our working set strategy compared to full problem solvers for both block-coordinate descent or a proximal gradient solver.

Original languageEnglish
Pages (from-to)5196-5211
Number of pages16
JournalProceedings of Machine Learning Research
Volume151
Publication statusPublished - 1 Jan 2022
Event25th International Conference on Artificial Intelligence and Statistics, AISTATS 2022 - Virtual, Online, Spain
Duration: 28 Mar 202230 Mar 2022

Fingerprint

Dive into the research topics of 'Convergent Working Set Algorithm for Lasso with Non-Convex Sparse Regularizers'. Together they form a unique fingerprint.

Cite this