Using bagged posteriors for robust inference and model criticism

12/15/2019
by   Jonathan H. Huggins, et al.
1

Standard Bayesian inference is known to be sensitive to model misspecification, leading to unreliable uncertainty quantification and poor predictive performance. However, finding generally applicable and computationally feasible methods for robust Bayesian inference under misspecification has proven to be a difficult challenge. An intriguing, easy-to-use, and widely applicable approach is to use bagging on the Bayesian posterior ("BayesBag"); that is, to use the average of posterior distributions conditioned on bootstrapped datasets. In this paper, we comprehensively develop the asymptotic theory of BayesBag, propose a model–data mismatch index for model criticism using BayesBag, and empirically validate our theory and methodology on synthetic and real-world data in linear regression (both feature selection and parameter inference), sparse logistic regression, insurance loss prediction, and phylogenetic tree reconstruction. We find that in the presence of significant misspecification, BayesBag yields more reproducible inferences, has better predictive accuracy, and selects correct models more often than the standard Bayesian posterior; meanwhile, when the model is correctly specified, BayesBag produces superior or equally good results for parameter inference and prediction, while being slightly more conservative for model selection. Overall, our results demonstrate that BayesBag combines the attractive modeling features of standard Bayesian inference with the distributional robustness properties of frequentist methods, providing benefits over both Bayes alone and the bootstrap alone.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/24/2020

Robust and Reproducible Model Selection Using Bagged Posteriors

Bayesian model selection is premised on the assumption that the data are...
research
05/20/2016

Coresets for Scalable Bayesian Logistic Regression

The use of Bayesian methods in large-scale data settings is attractive b...
research
06/06/2018

Doubly Robust Bayesian Inference for Non-Stationary Streaming Data with β-Divergences

We present the very first robust Bayesian Online Changepoint Detection a...
research
01/31/2023

On the Stability of General Bayesian Inference

We study the stability of posterior predictive inferences to the specifi...
research
08/31/2020

β-Cores: Robust Large-Scale Bayesian Data Summarization in the Presence of Outliers

Modern machine learning applications should be able to address the intri...
research
10/21/2019

Safe-Bayesian Generalized Linear Regression

We study generalized Bayesian inference under misspecification, i.e. whe...
research
06/25/2021

Robust Real-Time Delay Predictions in a Network of High-Frequency Urban Buses

Providing transport users and operators with accurate forecasts on trave...

Please sign up or login with your details

Forgot password? Click here to reset