Asymptotic Bias of Stochastic Gradient Search

08/30/2017
by   Vladislav B. Tadic, et al.
0

The asymptotic behavior of the stochastic gradient algorithm with a biased gradient estimator is analyzed. Relying on arguments based on the dynamic system theory (chain-recurrence) and the differential geometry (Yomdin theorem and Lojasiewicz inequality), tight bounds on the asymptotic bias of the iterates generated by such an algorithm are derived. The obtained results hold under mild conditions and cover a broad class of high-dimensional nonlinear algorithms. Using these results, the asymptotic properties of the policy-gradient (reinforcement) learning and adaptive population Monte Carlo sampling are studied. Relying on the same results, the asymptotic behavior of the recursive maximum split-likelihood estimation in hidden Markov models is analyzed, too.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/25/2018

Asymptotic Properties of Recursive Maximum Likelihood Estimation in Non-Linear State-Space Models

Using stochastic gradient search and the optimal filter derivative, it i...
research
09/20/2020

Stochastic Gradient Langevin Dynamics Algorithms with Adaptive Drifts

Bayesian deep learning offers a principled way to address many issues co...
research
05/22/2018

Langevin Markov Chain Monte Carlo with stochastic gradients

Monte Carlo sampling techniques have broad applications in machine learn...
research
04/18/2023

Finite-Sample Bounds for Adaptive Inverse Reinforcement Learning using Passive Langevin Dynamics

Stochastic gradient Langevin dynamics (SGLD) are a useful methodology fo...
research
01/02/2023

State and parameter learning with PaRIS particle Gibbs

Non-linear state-space models, also known as general hidden Markov model...

Please sign up or login with your details

Forgot password? Click here to reset