Smooth primal-dual coordinate descent algorithms for nonsmooth convex optimization

Ahmet Alacaoglu, Quoc Tran-Dinh, Olivier Fercoq, Volkan Cevher

Research output: Contribution to journalConference articlepeer-review

Abstract

We propose a new randomized coordinate descent method for a convex optimization template with broad applications. Our analysis relies on a novel combination of four ideas applied to the primal-dual gap function: smoothing, acceleration, homotopy, and coordinate descent with non-uniform sampling. As a result, our method features the first convergence rate guarantees among the coordinate descent methods, that are the best-known under a variety of common structure assumptions on the template. We provide numerical evidence to support the theoretical results with a comparison to state-of-the-art algorithms.

Original languageEnglish
Pages (from-to)5853-5862
Number of pages10
JournalAdvances in Neural Information Processing Systems
Volume2017-December
Publication statusPublished - 1 Jan 2017
Externally publishedYes
Event31st Annual Conference on Neural Information Processing Systems, NIPS 2017 - Long Beach, United States
Duration: 4 Dec 20179 Dec 2017

Fingerprint

Dive into the research topics of 'Smooth primal-dual coordinate descent algorithms for nonsmooth convex optimization'. Together they form a unique fingerprint.

Cite this