Stochastic Gradient Langevin Dynamics for (weakly) log-concave posterior distributions

Marelys Crespo Navas, Sébastien Gadat, Xavier Gendre

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, we investigate a continuous time version of the Stochastic Gradient Langevin Dynamics, introduced in [53], that incorporates a stochastic sampling step inside the traditional over-damped Langevin diffusion. This method is popular in machine learning for sampling a posterior distribution. We will pay specific attention to the computational cost in terms of n (the number of observations that produces the posterior distribution), and d (the dimension of the ambient space where the parameter of interest is living). We derive our analysis in the weakly convex framework, which is parameterized with the help of the Kurdyka-Łojasiewicz (KL) inequality, that permits to handle a vanishing curvature settings, which is far less restrictive when compared to the simple strongly convex case. We establish that the final horizon of simulation to obtain an ε approximation (in terms of entropy) is of the order (d log2(n))(1+r)2 [log2−1) + n2d2(1+r) log4(1+r) (n)] with a Poissonian sub-sampling of parameter (d log2(n))−(1+r)2, where the parameter r is involved in the KL inequality and varies between 0 (strongly log-concave case) and 1 (limiting Laplace situation).

Original languageEnglish
Article number179
JournalElectronic Journal of Probability
Volume29
DOIs
Publication statusPublished - 1 Jan 2024
Externally publishedYes

Keywords

  • Stochastic Gradient Langevin Dynamics
  • log-concave models
  • weak convexity

Fingerprint

Dive into the research topics of 'Stochastic Gradient Langevin Dynamics for (weakly) log-concave posterior distributions'. Together they form a unique fingerprint.

Cite this