Distributional Robustness Bounds Generalization Errors

12/20/2022
by   Shixiong Wang, et al.
0

Bayesian methods, distributionally robust optimization methods, and regularization methods are three pillars of trustworthy machine learning hedging against distributional uncertainty, e.g., the uncertainty of an empirical distribution compared to the true underlying distribution. This paper investigates the connections among the three frameworks and, in particular, explores why these frameworks tend to have smaller generalization errors. Specifically, first, we suggest a quantitative definition for "distributional robustness", propose the concept of "robustness measure", and formalize several philosophical concepts in distributionally robust optimization. Second, we show that Bayesian methods are distributionally robust in the probably approximately correct (PAC) sense; In addition, by constructing a Dirichlet-process-like prior in Bayesian nonparametrics, it can be proven that any regularized empirical risk minimization method is equivalent to a Bayesian method. Third, we show that generalization errors of machine learning models can be characterized using the distributional uncertainty of the nominal distribution and the robustness measures of these machine learning models, which is a new perspective to bound generalization errors, and therefore, explain the reason why distributionally robust machine learning models, Bayesian models, and regularization models tend to have smaller generalization errors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/31/2023

Learning Against Distributional Uncertainty: On the Trade-off Between Robustness and Specificity

Trustworthy machine learning aims at combating distributional uncertaint...
research
05/19/2017

Doubly Robust Data-Driven Distributionally Robust Optimization

Data-driven Distributionally Robust Optimization (DD-DRO) via optimal tr...
research
05/27/2019

Distributionally Robust Optimization and Generalization in Kernel Methods

Distributionally robust optimization (DRO) has attracted attention in ma...
research
09/17/2020

Distributional Generalization: A New Kind of Generalization

We introduce a new notion of generalization – Distributional Generalizat...
research
06/04/2020

Robust Sampling in Deep Learning

Deep learning requires regularization mechanisms to reduce overfitting a...
research
11/05/2021

Toward Learning Human-aligned Cross-domain Robust Models by Countering Misaligned Features

Machine learning has demonstrated remarkable prediction accuracy over i....
research
07/26/2023

Topology-aware Robust Optimization for Out-of-distribution Generalization

Out-of-distribution (OOD) generalization is a challenging machine learni...

Please sign up or login with your details

Forgot password? Click here to reset