Abstract
Standard Bayesian learning is known to have suboptimal generalization capabilities under misspecification and in the presence of outliers. Probably approximately correct (PAC)-Bayes theory demonstrates that the free energy criterion minimized by Bayesian learning is a bound on the generalization error for Gibbs predictors (i.e., for single models drawn at random from the posterior) under the assumption of sampling distributions uncontaminated by outliers. This viewpoint provides a justification for the limitations of Bayesian learning when the model is misspecified, requiring ensembling, and when data are affected by outliers. In recent work, PAC-Bayes bounds—referred to as PAC<inline-formula> <tex-math notation="LaTeX">$^m$</tex-math> </inline-formula>—were derived to introduce free energy metrics that account for the performance of ensemble predictors, obtaining enhanced performance under misspecification. This work presents a novel robust free energy criterion that combines the generalized logarithm score function with PAC<inline-formula> <tex-math notation="LaTeX">$^m$</tex-math> </inline-formula> ensemble bounds. The proposed free energy training criterion produces predictive distributions that are able to concurrently counteract the detrimental effects of misspecification—with respect to both likelihood and prior distribution—and outliers.
Original language | English (US) |
---|---|
Pages (from-to) | 1-15 |
Number of pages | 15 |
Journal | IEEE Transactions on Neural Networks and Learning Systems |
DOIs | |
State | Accepted/In press - 2023 |
All Science Journal Classification (ASJC) codes
- Software
- Computer Science Applications
- Computer Networks and Communications
- Artificial Intelligence
Keywords
- Bayes methods
- Bayesian learning
- Europe
- Pollution measurement
- Predictive models
- Robustness
- Standards
- Training
- ensemble models
- machine learning
- misspecification
- outliers
- robustness