Rate-Distortion Theoretic Generalization Bounds for Stochastic Learning Algorithms

03/04/2022
by   Milad Sefidgaran, et al.
0

Understanding generalization in modern machine learning settings has been one of the major challenges in statistical learning theory. In this context, recent years have witnessed the development of various generalization bounds suggesting different complexity notions such as the mutual information between the data sample and the algorithm output, compressibility of the hypothesis space, and the fractal dimension of the hypothesis space. While these bounds have illuminated the problem at hand from different angles, their suggested complexity notions might appear seemingly unrelated, thereby restricting their high-level impact. In this study, we prove novel generalization bounds through the lens of rate-distortion theory, and explicitly relate the concepts of mutual information, compressibility, and fractal dimensions in a single mathematical framework. Our approach consists of (i) defining a generalized notion of compressibility by using source coding concepts, and (ii) showing that the `compression error rate' can be linked to the generalization error both in expectation and with high probability. We show that in the `lossless compression' setting, we recover and improve existing mutual information-based bounds, whereas a `lossy compression' scheme allows us to link generalization to the rate-distortion dimension – a particular notion of fractal dimension. Our results bring a more unified perspective on generalization and open up several future research directions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/04/2023

Generalization Guarantees via Algorithm-dependent Rademacher Complexity

Algorithm- and data-dependent generalization bounds are required to expl...
research
03/09/2023

Data-dependent Generalization Bounds via Variable-Size Compressibility

In this paper, we establish novel data-dependent upper bounds on the gen...
research
06/21/2022

Supermodular f-divergences and bounds on lossy compression and generalization error with mutual f-information

In this paper, we introduce super-modular -divergences and provide three...
research
06/11/2018

Chaining Mutual Information and Tightening Generalization Bounds

Bounding the generalization error of learning algorithms has a long hist...
research
10/13/2021

Lossy Compression with Universal Distortion

A novel variant of lossy coding is considered in which the distortion me...
research
02/06/2023

Generalization Bounds with Data-dependent Fractal Dimensions

Providing generalization guarantees for modern neural networks has been ...
research
05/28/2019

Understanding the Behaviour of the Empirical Cross-Entropy Beyond the Training Distribution

Machine learning theory has mostly focused on generalization to samples ...

Please sign up or login with your details

Forgot password? Click here to reset