Nonlinear Least Squares for Large-Scale Machine Learning using Stochastic Jacobian Estimates

07/12/2021
by   Johannes J. Brust, et al.
0

For large nonlinear least squares loss functions in machine learning we exploit the property that the number of model parameters typically exceeds the data in one batch. This implies a low-rank structure in the Hessian of the loss, which enables effective means to compute search directions. Using this property, we develop two algorithms that estimate Jacobian matrices and perform well when compared to state-of-the-art methods.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset