Random extrapolation for primal-dual coordinate descent

Research output: Contribution to journalConference articlepeer-review

Abstract

We introduce a randomly extrapolated primal-dual coordinate descent method that adapts to sparsity of the data matrix and the favorable structures of the objective function. Our method updates only a subset of primal and dual variables with sparse data, and it uses large step sizes with dense data, retaining the benefits of the specific methods designed for each case. In addition to adapting to sparsity, our method attains fast convergence guarantees in favorable cases without any modifications. In particular, we prove linear convergence under metric subregularity, which applies to strongly convex-strongly concave problems and piecewise linear quadratic functions. We show almost sure convergence of the sequence and optimal sublinear convergence rates for the primal-dual gap and objective values, in the general convex-concave case. Numerical evidence demonstrates the state-of-the-art empirical performance of our method in sparse and dense settings, matching and improving the existing methods.

Original languageEnglish
Pages (from-to)191-201
Number of pages11
JournalProceedings of Machine Learning Research
Volume119
Publication statusPublished - 1 Jan 2020
Event37th International Conference on Machine Learning, ICML 2020 - Virtual, Online
Duration: 13 Jul 202018 Jul 2020

Fingerprint

Dive into the research topics of 'Random extrapolation for primal-dual coordinate descent'. Together they form a unique fingerprint.

Cite this