Nonlinear Least Squares for Large-Scale Machine Learning using Stochastic Jacobian Estimates

by   Johannes J. Brust, et al.

For large nonlinear least squares loss functions in machine learning we exploit the property that the number of model parameters typically exceeds the data in one batch. This implies a low-rank structure in the Hessian of the loss, which enables effective means to compute search directions. Using this property, we develop two algorithms that estimate Jacobian matrices and perform well when compared to state-of-the-art methods.


page 1

page 2

page 3

page 4


Stochastic Mirror Descent for Low-Rank Tensor Decomposition Under Non-Euclidean Losses

This work considers low-rank canonical polyadic decomposition (CPD) unde...

PNKH-B: A Projected Newton-Krylov Method for Large-Scale Bound-Constrained Optimization

We present PNKH-B, a projected Newton-Krylov method with a low-rank appr...

Adaptive Graduated Nonconvexity Loss

Many problems in robotics, such as estimating the state from noisy senso...

Scalable Derivative-Free Optimization for Nonlinear Least-Squares Problems

Derivative-free - or zeroth-order - optimization (DFO) has gained recent...

A Note On Nonlinear Regression Under L2 Loss

We investigate the nonlinear regression problem under L2 loss (square lo...

Search Algorithms and Loss Functions for Bayesian Clustering

We propose a randomized greedy search algorithm to find a point estimate...

Please sign up or login with your details

Forgot password? Click here to reset