On the convergence of stochastic approximations under a subgeometric ergodic markov dynamic

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, we extend the framework of the convergence of stochastic approximations. Such a procedure is used in many methods such as parameters estimation inside a Metropolis Hastings algorithm, stochastic gradient descent or stochastic Expectation Maximization algorithm. It is given by θn+1 = θnn+1Hθn (Xn+1), where (Xn)n∈N is a sequence of random variables following a parametric distribution which depends on (θn)n∈N, and (Δn)n∈N is a step sequence. The convergence of such a stochastic approximation has already been proved under an assumption of geometric ergodicity of the Markov dynamic. However, in many practical situations this hypothesis is not satisfied, for instance for any heavy tail target distribution in a Monte Carlo Metropolis Hastings algorithm. In this paper, we relax this hypothesis and prove the convergence of the stochastic approximation by only assuming a subgeometric ergodicity of the Markov dynamic. This result opens up the possibility to derive more generic algorithms with proven convergence. As an example, we first study an adaptive Markov Chain Monte Carlo algorithm where the proposal distribution is adapted by learning the variance of a heavy tail target distribution. We then apply our work to the Independent Component Analysis when a positive heavy tail noise leads to a subgeometric dynamic in an Expectation Maximization algorithm.

Original languageEnglish
Pages (from-to)1583-1609
Number of pages27
JournalElectronic Journal of Statistics
Volume15
Issue number1
DOIs
Publication statusPublished - 1 Jan 2021
Externally publishedYes

Keywords

  • Markovian dynamic
  • Stochastic approximation
  • Subgeometric ergodicity

Fingerprint

Dive into the research topics of 'On the convergence of stochastic approximations under a subgeometric ergodic markov dynamic'. Together they form a unique fingerprint.

Cite this