A global stochastic optimization particle filter algorithm

Research output: Contribution to journalArticlepeer-review

Abstract

We introduce a new online algorithm for expected loglikelihood maximization in situations where the objective function is multimodal or has saddle points. The key element underpinning the algorithm is a probability distribution that concentrates on the target parameter value as the sample size increases and can be efficiently estimated by means of a standard particle filter algorithm. This distribution depends on a learning rate, such that the faster the learning rate the quicker the distribution concentrates on the desired element of the search space, but the less likely the algorithm is to escape from a local optimum of the objective function. In order to achieve a fast convergence rate with a slow learning rate, our algorithm exploits the acceleration property of averaging, which is well known from the stochastic gradient literature. Considering several challenging estimation problems, our numerical experiments show that with high probability, the algorithm successfully finds the highest mode of the objective function and converges to the global maximizer at the optimal rate. While the focus of this work is expected loglikelihood maximization, the proposed methodology and its theory apply more generally to optimization of a function defined through an expectation.

Original languageEnglish
Pages (from-to)937-955
Number of pages19
JournalBiometrika
Volume109
Issue number4
DOIs
Publication statusPublished - 1 Dec 2022

Keywords

  • Global optimization
  • Inference in large datasets
  • Particle filtering
  • Stochastic optimization

Fingerprint

Dive into the research topics of 'A global stochastic optimization particle filter algorithm'. Together they form a unique fingerprint.

Cite this