TY - JOUR
T1 - A global stochastic optimization particle filter algorithm
AU - Gerber, M.
AU - Douc, R.
N1 - Publisher Copyright:
© 2021 The Author(s). Published by Oxford University Press on behalf of the Biometrika Trust. All rights reserved.
PY - 2022/12/1
Y1 - 2022/12/1
N2 - We introduce a new online algorithm for expected loglikelihood maximization in situations where the objective function is multimodal or has saddle points. The key element underpinning the algorithm is a probability distribution that concentrates on the target parameter value as the sample size increases and can be efficiently estimated by means of a standard particle filter algorithm. This distribution depends on a learning rate, such that the faster the learning rate the quicker the distribution concentrates on the desired element of the search space, but the less likely the algorithm is to escape from a local optimum of the objective function. In order to achieve a fast convergence rate with a slow learning rate, our algorithm exploits the acceleration property of averaging, which is well known from the stochastic gradient literature. Considering several challenging estimation problems, our numerical experiments show that with high probability, the algorithm successfully finds the highest mode of the objective function and converges to the global maximizer at the optimal rate. While the focus of this work is expected loglikelihood maximization, the proposed methodology and its theory apply more generally to optimization of a function defined through an expectation.
AB - We introduce a new online algorithm for expected loglikelihood maximization in situations where the objective function is multimodal or has saddle points. The key element underpinning the algorithm is a probability distribution that concentrates on the target parameter value as the sample size increases and can be efficiently estimated by means of a standard particle filter algorithm. This distribution depends on a learning rate, such that the faster the learning rate the quicker the distribution concentrates on the desired element of the search space, but the less likely the algorithm is to escape from a local optimum of the objective function. In order to achieve a fast convergence rate with a slow learning rate, our algorithm exploits the acceleration property of averaging, which is well known from the stochastic gradient literature. Considering several challenging estimation problems, our numerical experiments show that with high probability, the algorithm successfully finds the highest mode of the objective function and converges to the global maximizer at the optimal rate. While the focus of this work is expected loglikelihood maximization, the proposed methodology and its theory apply more generally to optimization of a function defined through an expectation.
KW - Global optimization
KW - Inference in large datasets
KW - Particle filtering
KW - Stochastic optimization
U2 - 10.1093/biomet/asab067
DO - 10.1093/biomet/asab067
M3 - Article
AN - SCOPUS:85148236621
SN - 0006-3444
VL - 109
SP - 937
EP - 955
JO - Biometrika
JF - Biometrika
IS - 4
ER -