Uniform Generalization, Concentration, and Adaptive Learning

08/22/2016
by   Ibrahim Alabdulmohsin, et al.
0

One fundamental goal in any learning algorithm is to mitigate its risk for overfitting. Mathematically, this requires that the learning algorithm enjoys a small generalization risk, which is defined either in expectation or in probability. Both types of generalization are commonly used in the literature. For instance, generalization in expectation has been used to analyze algorithms, such as ridge regression and SGD, whereas generalization in probability is used in the VC theory, among others. Recently, a third notion of generalization has been studied, called uniform generalization, which requires that the generalization risk vanishes uniformly in expectation across all bounded parametric losses. It has been shown that uniform generalization is, in fact, equivalent to an information-theoretic stability constraint, and that it recovers classical results in learning theory. It is achievable under various settings, such as sample compression schemes, finite hypothesis spaces, finite domains, and differential privacy. However, the relationship between uniform generalization and concentration remained unknown. In this paper, we answer this question by proving that, while a generalization in expectation does not imply a generalization in probability, a uniform generalization in expectation does imply concentration. We establish a chain rule for the uniform generalization risk of the composition of hypotheses and use it to derive a large deviation bound. Finally, we prove that the bound is tight.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/24/2018

Generalization Bounds for Uniformly Stable Algorithms

Uniform stability of a learning algorithm is a classical notion of algor...
research
03/12/2019

An Exponential Efron-Stein Inequality for Lq Stable Learning Rules

There is accumulating evidence in the literature that stability of learn...
research
06/08/2022

Boosting the Confidence of Generalization for L_2-Stable Randomized Learning Algorithms

Exponential generalization bounds with near-tight rates have recently be...
research
12/09/2019

In Defense of Uniform Convergence: Generalization via derandomization with an application to interpolating predictors

We propose to study the generalization error of a learned predictor ĥ in...
research
05/07/2014

A Mathematical Theory of Learning

In this paper, a mathematical theory of learning is proposed that has ma...
research
07/01/2020

The Restricted Isometry of ReLU Networks: Generalization through Norm Concentration

While regression tasks aim at interpolating a relation on the entire inp...
research
11/08/2021

Realizable Learning is All You Need

The equivalence of realizable and agnostic learnability is a fundamental...

Please sign up or login with your details

Forgot password? Click here to reset