Convergence rates for momentum stochastic gradient descent with noise of machine learning type

02/07/2023
by   Benjamin Gess, et al.
0

We consider the momentum stochastic gradient descent scheme (MSGD) and its continuous-in-time counterpart in the context of non-convex optimization. We show almost sure exponential convergence of the objective function value for target functions that are Lipschitz continuous and satisfy the Polyak-Lojasiewicz inequality on the relevant domain, and under assumptions on the stochastic noise that are motivated by overparameterized supervised learning applications. Moreover, we optimize the convergence rate over the set of friction parameters and show that the MSGD process almost surely converges.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset