Scaling-up Empirical Risk Minimization: Optimization of Incomplete U-statistics

01/12/2015
by   Stephan Clémençon, et al.
0

In a wide range of statistical learning problems such as ranking, clustering or metric learning among others, the risk is accurately estimated by U-statistics of degree d≥ 1, i.e. functionals of the training data with low variance that take the form of averages over k-tuples. From a computational perspective, the calculation of such statistics is highly expensive even for a moderate sample size n, as it requires averaging O(n^d) terms. This makes learning procedures relying on the optimization of such data functionals hardly feasible in practice. It is the major goal of this paper to show that, strikingly, such empirical risks can be replaced by drastically computationally simpler Monte-Carlo estimates based on O(n) terms only, usually referred to as incomplete U-statistics, without damaging the O_P(1/√(n)) learning rate of Empirical Risk Minimization (ERM) procedures. For this purpose, we establish uniform deviation results describing the error made when approximating a U-process by its incomplete version under appropriate complexity assumptions. Extensions to model selection, fast rate situations and various sampling techniques are also considered, as well as an application to stochastic gradient descent for ERM. Finally, numerical examples are displayed in order to provide strong empirical evidence that the approach we promote largely surpasses more naive subsampling techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/12/2020

Weighted Empirical Risk Minimization: Sample Selection Bias Correction based on Importance Sampling

We consider statistical learning problems, when the distribution P' of t...
research
11/01/2022

On Medians of (Randomized) Pairwise Means

Tournament procedures, recently introduced in Lugosi Mendelson (2016...
research
10/08/2016

Variance-based regularization with convex objectives

We develop an approach to risk minimization and stochastic optimization ...
research
01/09/2015

Survey schemes for stochastic gradient descent with applications to M-estimation

In certain situations that shall be undoubtedly more and more common in ...
research
11/02/2018

Non-Asymptotic Guarantees For Sampling by Stochastic Gradient Descent

Sampling from various kinds of distributions is an issue of paramount im...
research
09/07/2023

Empirical Risk Minimization for Losses without Variance

This paper considers an empirical risk minimization problem under heavy-...
research
11/17/2015

Extending Gossip Algorithms to Distributed Estimation of U-Statistics

Efficient and robust algorithms for decentralized estimation in networks...

Please sign up or login with your details

Forgot password? Click here to reset