Online Learning Under A Separable Stochastic Approximation Framework

05/12/2023
by   Min Gan, et al.
0

We propose an online learning algorithm for a class of machine learning models under a separable stochastic approximation framework. The essence of our idea lies in the observation that certain parameters in the models are easier to optimize than others. In this paper, we focus on models where some parameters have a linear nature, which is common in machine learning. In one routine of the proposed algorithm, the linear parameters are updated by the recursive least squares (RLS) algorithm, which is equivalent to a stochastic Newton method; then, based on the updated linear parameters, the nonlinear parameters are updated by the stochastic gradient method (SGD). The proposed algorithm can be understood as a stochastic approximation version of block coordinate gradient descent approach in which one part of the parameters is updated by a second-order SGD method while the other part is updated by a first-order SGD. Global convergence of the proposed online algorithm for non-convex cases is established in terms of the expected violation of a first-order optimality condition. Numerical experiments have shown that the proposed method accelerates convergence significantly and produces more robust training and test performance when compared to other popular learning algorithms. Moreover, our algorithm is less sensitive to the learning rate and outperforms the recently proposed slimTrain algorithm. The code has been uploaded to GitHub for validation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2018

Online Newton Step Algorithm with Estimated Gradient

Online learning with limited information feedback (bandit) tries to solv...
research
06/12/2018

Convergence of SGD in Learning ReLU Models with Separable Data

We consider the binary classification problem in which the objective fun...
research
01/16/2015

Stochastic Gradient Based Extreme Learning Machines For Online Learning of Advanced Combustion Engines

In this article, a stochastic gradient based online learning algorithm f...
research
07/21/2019

Alice's Adventures in the Markovian World

This paper proposes an algorithm Alice having no access to the physics l...
research
07/20/2020

Iterative Method for Tuning Complex Simulation Code

Tuning a complex simulation code refers to the process of improving the ...
research
07/23/2019

Mix and Match: An Optimistic Tree-Search Approach for Learning Models from Mixture Distributions

We consider a co-variate shift problem where one has access to several m...
research
10/19/2018

Exchangeability and Kernel Invariance in Trained MLPs

In the analysis of machine learning models, it is often convenient to as...

Please sign up or login with your details

Forgot password? Click here to reset