Bayesian model selection in the ℳ-open setting – Approximate posterior inference and probability-proportional-to-size subsampling for efficient large-scale leave-one-out cr

05/27/2020
by   Riko Kelter, et al.
0

Comparison of competing statistical models is an essential part of psychological research. From a Bayesian perspective, various approaches to model comparison and selection have been proposed in the literature. However, the applicability of these approaches strongly depends on the assumptions about the model space ℳ, the so-called model view. Furthermore, traditional methods like leave-one-out cross-validation (LOO-CV) estimate the expected log predictive density (ELPD) of a model to investigate how the model generalises out-of-sample, which quickly becomes computationally inefficient when sample size becomes large. Here, we provide a tutorial on approximate Pareto-smoothed importance sampling leave-one-out cross-validation (PSIS-LOO), a computationally efficient method for Bayesian model comparison. First, we discuss several model views and the available Bayesian model comparison methods in each. We then use Bayesian logistic regression as a running example how to apply the method in practice, and show that it outperforms other methods like LOO-CV or information criteria in terms of computational effort while providing similarly accurate ELPD estimates. In a second step, we show how even large-scale models can be compared efficiently by using posterior approximations in combination with probability-proportional-to-size subsampling. We show how to compare competing models based on the ELPD estimates provided, and how to conduct posterior predictive checks to safeguard against overconfidence in one of the models under consideration. We conclude that the method is attractive for mathematical psychologists who aim at comparing several competing statistical models, which are possibly high-dimensional and in the big-data regime.

READ FULL TEXT
research
04/24/2019

Bayesian leave-one-out cross-validation for large data

Model inference, such as model comparison, model checking, and model sel...
research
02/17/2019

Approximate leave-future-out cross-validation for Bayesian time series models

One of the common goals of time series analysis is to use the observed s...
research
01/03/2020

Leave-One-Out Cross-Validation for Bayesian Model Comparison in Large Data

Recently, new methods for model assessment, based on subsampling and pos...
research
02/12/2021

Efficient Selection Between Hierarchical Cognitive Models: Cross-validation With Variational Bayes

Model comparison is the cornerstone of theoretical progress in psycholog...
research
09/05/2022

Using leave-one-out cross-validation (LOO) in a multilevel regression and poststratification (MRP) workflow: A cautionary tale

In recent decades, multilevel regression and poststratification (MRP) ha...
research
12/01/2020

BayesTime: Bayesian Functional Principal Components for Sparse Longitudinal Data

Modeling non-linear temporal trajectories is of fundamental interest in ...
research
11/29/2020

Approximate Cross-validated Mean Estimates for Bayesian Hierarchical Regression Models

We introduce a novel procedure for obtaining cross-validated predictive ...

Please sign up or login with your details

Forgot password? Click here to reset