The Perturbed Prox-Preconditioned Spider Algorithm: Non-Asymptotic Convergence Bounds

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

A novel algorithm named PerturbedProx-Preconditioned SPIDER (3P-SPIDER) is introduced. It is a stochastic variancereduced proximal-gradient type algorithm built on Stochastic Path Integral Differential EstimatoR (SPIDER), an algorithm known to achieve near-optimal first-order oracle inequality for nonconvex and nonsmooth optimization. Compared to the vanilla prox-SPIDER, 3P-SPIDER uses preconditioned gradient estimators. Preconditioning can either be applied explicitly to a gradient estimator or be introduced implicitly as in applications to the EM algorithm. 3P-SPIDER also assumes that the preconditioned gradients may (possibly) be not known in closed analytical form and therefore must be approximated which adds an additional degree of perturbation. Studying the convergence in expectation, we show that 3P-SPIDER achieves a near-optimal oracle inequality O(n1/2/ϵ) where n is the number of observations and ϵ the target precision even when the gradient is estimated by Monte Carlo methods. We illustrate the algorithm on an application to the minimization of a penalized empirical loss.

Original languageEnglish
Title of host publication2021 IEEE Statistical Signal Processing Workshop, SSP 2021
PublisherIEEE Computer Society
Pages96-100
Number of pages5
ISBN (Electronic)9781728157672
DOIs
Publication statusPublished - 11 Jul 2021
Externally publishedYes
Event21st IEEE Statistical Signal Processing Workshop, SSP 2021 - Virtual, Rio de Janeiro, Brazil
Duration: 11 Jul 202114 Jul 2021

Publication series

NameIEEE Workshop on Statistical Signal Processing Proceedings
Volume2021-July

Conference

Conference21st IEEE Statistical Signal Processing Workshop, SSP 2021
Country/TerritoryBrazil
CityVirtual, Rio de Janeiro
Period11/07/2114/07/21

Keywords

  • Control Variates
  • Finite sum optimization
  • Large Scale Learning
  • Statistical Learning
  • Variance reduced Stochastic gradient

Fingerprint

Dive into the research topics of 'The Perturbed Prox-Preconditioned Spider Algorithm: Non-Asymptotic Convergence Bounds'. Together they form a unique fingerprint.

Cite this