Almost Sure Convergence Rates of Stochastic Zeroth-order Gradient Descent for Łojasiewicz Functions

10/31/2022
by   Tianyu Wang, et al.
0

We prove almost sure convergence rates of Zeroth-order Gradient Descent (SZGD) algorithms for Łojasiewicz functions. The SZGD algorithm iterates as x_t+1 = x_t - η_t ∇ f (x_t), t = 0,1,2,3,⋯ , where f is the objective function that satisfies the Łojasiewicz inequality with Łojasiewicz exponent θ, η_t is the step size (learning rate), and ∇ f (x_t) is the approximate gradient estimated using zeroth-order information. We show that, for smooth Łojasiewicz functions, the sequence { x_t }_t∈ℕ governed by SZGD converges to a bounded point x_∞ almost surely, and x_∞ is a critical point of f. If θ∈ (0,1/2], f (x_t) - f (x_∞), ∑_s=t^∞ x_s - x_∞^2 and x_t - x_∞ (· is the Euclidean norm) converge to zero linearly almost surely. If θ∈ (1/2, 1), then f (x_t) - f (x_∞) (and ∑_s=t^∞ x_s+1 - x_s ^2) converges to zero at rate o ( t^1/1 - 2θlog t ) almost surely; x_t - x_∞ converges to zero at rate o ( t^1-θ/1-2θlog t ) almost surely. To the best of our knowledge, this paper provides the first almost sure convergence rate guarantee for stochastic zeroth order algorithms for Łojasiewicz functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2021

Accelerated Almost-Sure Convergence Rates for Nonconvex Stochastic Gradient Descent using Stochastic Learning Rates

Large-scale optimization problems require algorithms both effective and ...
research
09/04/2021

On Faster Convergence of Scaled Sign Gradient Descent

Communication has been seen as a significant bottleneck in industrial ap...
research
07/29/2023

A new Gradient TD Algorithm with only One Step-size: Convergence Rate Analysis using L-λ Smoothness

Gradient Temporal Difference (GTD) algorithms (Sutton et al., 2008, 2009...
research
07/19/2019

On Linear Convergence of Weighted Kernel Herding

We provide a novel convergence analysis of two popular sampling algorith...
research
07/05/2021

The Last-Iterate Convergence Rate of Optimistic Mirror Descent in Stochastic Variational Inequalities

In this paper, we analyze the local convergence rate of optimistic mirro...
research
01/07/2020

Backtracking Gradient Descent allowing unbounded learning rates

In unconstrained optimisation on an Euclidean space, to prove convergenc...
research
01/31/2020

Learning Unitaries by Gradient Descent

We study the hardness of learning unitary transformations by performing ...

Please sign up or login with your details

Forgot password? Click here to reset