On Riemannian Stochastic Approximation Schemes with Fixed Step-Size

02/15/2021
by   Alain Durmus, et al.
0

This paper studies fixed step-size stochastic approximation (SA) schemes, including stochastic gradient schemes, in a Riemannian framework. It is motivated by several applications, where geodesics can be computed explicitly, and their use accelerates crude Euclidean methods. A fixed step-size scheme defines a family of time-homogeneous Markov chains, parametrized by the step-size. Here, using this formulation, non-asymptotic performance bounds are derived, under Lyapunov conditions. Then, for any step-size, the corresponding Markov chain is proved to admit a unique stationary distribution, and to be geometrically ergodic. This result gives rise to a family of stationary distributions indexed by the step-size, which is further shown to converge to a Dirac measure, concentrated at the solution of the problem at hand, as the step-size goes to 0. Finally, the asymptotic rate of this convergence is established, through an asymptotic expansion of the bias, and a central limit theorem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/02/2015

(Non-) asymptotic properties of Stochastic Gradient Langevin Dynamics

Applying standard Markov chain Monte Carlo (MCMC) algorithms to large da...
research
04/09/2020

On Linear Stochastic Approximation: Fine-grained Polyak-Ruppert and Non-Asymptotic Concentration

We undertake a precise study of the asymptotic and non-asymptotic proper...
research
06/28/2023

Stochastic Methods in Variational Inequalities: Ergodicity, Bias and Refinements

For min-max optimization and variational inequalities problems (VIP) enc...
research
02/17/2010

Iterative exact global histogram specification and SSIM gradient ascent: a proof of convergence, step size and parameter selection

The SSIM-optimized exact global histogram specification (EGHS) is shown ...
research
09/05/2022

The Proxy Step-size Technique for Regularized Optimization on the Sphere Manifold

We give an effective solution to the regularized optimization problem g ...
research
11/06/2020

On the Ergodicity, Bias and Asymptotic Normality of Randomized Midpoint Sampling Method

The randomized midpoint method, proposed by [SL19], has emerged as an op...
research
10/27/2021

The ODE Method for Asymptotic Statistics in Stochastic Approximation and Reinforcement Learning

The paper concerns convergence and asymptotic statistics for stochastic ...

Please sign up or login with your details

Forgot password? Click here to reset