Debiasing Stochastic Gradient Descent to handle missing values

02/21/2020
by   Aude Sportisse, et al.
0

A major caveat of large scale data is their incom-pleteness. We propose an averaged stochastic gradient algorithm handling missing values in linear models. This approach has the merit to be free from the need of any data distribution modeling and to account for heterogeneous missing proportion. In both streaming and finite-sample settings, we prove that this algorithm achieves convergence rate of O(1 n) at the iteration n, the same as without missing values. We show the convergence behavior and the relevance of the algorithm not only on synthetic data but also on real data sets, including those collected from medical register.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2016

Trading-off variance and complexity in stochastic gradient descent

Stochastic gradient descent is the method of choice for large-scale mach...
research
10/19/2018

A Model Parallel Proximal Stochastic Gradient Algorithm for Partially Asynchronous Systems

Large models are prevalent in modern machine learning scenarios, includi...
research
11/06/2015

Streaming regularization parameter selection via stochastic gradient descent

We propose a framework to perform streaming covariance selection. Our ap...
research
02/24/2016

Online Dual Coordinate Ascent Learning

The stochastic dual coordinate-ascent (S-DCA) technique is a useful alte...
research
02/22/2022

Convergence of online k-means

We prove asymptotic convergence for a general class of k-means algorithm...
research
02/03/2022

Minimax rate of consistency for linear models with missing values

Missing values arise in most real-world data sets due to the aggregation...
research
11/12/2021

Solving A System Of Linear Equations By Randomized Orthogonal Projections

Randomization has shown catalyzing effects in linear algebra with promis...

Please sign up or login with your details

Forgot password? Click here to reset