Robust PAC^m: Training Ensemble Models Under Model Misspecification and Outliers

03/03/2022
by   Matteo Zecchin, et al.
0

Standard Bayesian learning is known to have suboptimal generalization capabilities under model misspecification and in the presence of outliers. PAC-Bayes theory demonstrates that the free energy criterion minimized by Bayesian learning is a bound on the generalization error for Gibbs predictors (i.e., for single models drawn at random from the posterior) under the assumption of sampling distributions uncontaminated by outliers. This viewpoint provides a justification for the limitations of Bayesian learning when the model is misspecified, requiring ensembling, and when data is affected by outliers. In recent work, PAC-Bayes bounds - referred to as PAC^m - were derived to introduce free energy metrics that account for the performance of ensemble predictors, obtaining enhanced performance under misspecification. This work presents a novel robust free energy criterion that combines the generalized logarithm score function with PAC^m ensemble bounds. The proposed free energy training criterion produces predictive distributions that are able to concurrently counteract the detrimental effects of model misspecification and outliers.

READ FULL TEXT
research
06/17/2021

PAC-Bayes, MAC-Bayes and Conditional Mutual Information: Fast rate bounds that handle general VC classes

We give a novel, unified derivation of conditional PAC-Bayesian and mutu...
research
12/18/2019

Learning under Model Misspecification: Applications to Variational and Ensemble methods

This paper provides a novel theoretical analysis of the problem of learn...
research
10/21/2021

User-friendly introduction to PAC-Bayes bounds

Aggregated predictors are obtained by making a set of basic predictors v...
research
02/28/2012

PAC-Bayesian Generalization Bound on Confusion Matrix for Multi-Class Classification

In this work, we propose a PAC-Bayes bound for the generalization risk o...
research
10/10/2019

Still no free lunches: the price to pay for tighter PAC-Bayes bounds

"No free lunch" results state the impossibility of obtaining meaningful ...
research
12/18/2019

Learning from i.i.d. data under model miss-specification

This paper introduces a new approach to learning from i.i.d. data under ...
research
06/01/2021

A unified PAC-Bayesian framework for machine unlearning via information risk minimization

Machine unlearning refers to mechanisms that can remove the influence of...

Please sign up or login with your details

Forgot password? Click here to reset