Non-Convex SGD Learns Halfspaces with Adversarial Label Noise

06/11/2020
by   Ilias Diakonikolas, et al.
0

We study the problem of agnostically learning homogeneous halfspaces in the distribution-specific PAC model. For a broad family of structured distributions, including log-concave distributions, we show that non-convex SGD efficiently converges to a solution with misclassification error O()+, where is the misclassification error of the best-fitting halfspace. In sharp contrast, we show that optimizing any convex surrogate inherently leads to misclassification error of ω(), even under Gaussian marginals.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2020

Learning Halfspaces with Massart Noise Under Structured Distributions

We study the problem of learning halfspaces with Massart noise in the di...
research
06/11/2020

Learning Halfspaces with Tsybakov Noise

We study the efficient PAC learnability of halfspaces in the presence of...
research
12/10/2018

Why Does Stagewise Training Accelerate Convergence of Testing Error Over SGD?

Stagewise training strategy is commonly used for learning neural network...
research
07/07/2019

Quantitative W_1 Convergence of Langevin-Like Stochastic Processes with Non-Convex Potential State-Dependent Noise

We prove quantitative convergence rates at which discrete Langevin-like ...
research
05/21/2018

Learning with Non-Convex Truncated Losses by SGD

Learning with a convex loss function has been a dominating paradigm for...
research
11/13/2018

A unified algorithm for the non-convex penalized estimation: The ncpen package

Various R packages have been developed for the non-convex penalized esti...
research
03/09/2020

Revisiting SGD with Increasingly Weighted Averaging: Optimization and Generalization Perspectives

Stochastic gradient descent (SGD) has been widely studied in the literat...

Please sign up or login with your details

Forgot password? Click here to reset