Quadratic error bound of the smoothed gap and the restarted averaged primal-dual hybrid gradient

Research output: Contribution to journalArticlepeer-review

Abstract

We study the linear convergence of the primal-dual hybrid gradient method. After a review of current analyses, we show that they do not explain properly the behavior of the algorithm, even on the most simple problems. We thus introduce the quadratic error bound of the smoothed gap, a new regularity assumption that holds for a wide class of optimization problems. Equipped with this tool, we manage to prove tighter convergence rates. Then, we show that averaging and restarting the primal-dual hybrid gradient allows us to leverage better the regularity constant. Numerical experiments on linear and quadratic programs, ridge regression and image denoising illustrate the findings of the paper.

Original languageEnglish
Article number6
JournalOpen Journal of Mathematical Optimization
Volume4
DOIs
Publication statusPublished - 1 Jan 2023

Keywords

  • error bound
  • linear convergence
  • primal-dual algorithm
  • restart

Fingerprint

Dive into the research topics of 'Quadratic error bound of the smoothed gap and the restarted averaged primal-dual hybrid gradient'. Together they form a unique fingerprint.

Cite this