Combining predictions from linear models when training and test inputs differ

06/24/2014
by   Thijs van Ommen, et al.
0

Methods for combining predictions from different models in a supervised learning setting must somehow estimate/predict the quality of a model's predictions at unknown future inputs. Many of these methods (often implicitly) make the assumption that the test inputs are identical to the training inputs, which is seldom reasonable. By failing to take into account that prediction will generally be harder for test inputs that did not occur in the training set, this leads to the selection of too complex models. Based on a novel, unbiased expression for KL divergence, we propose XAIC and its special case FAIC as versions of AIC intended for prediction that use different degrees of knowledge of the test inputs. Both methods substantially differ from and may outperform all the known versions of AIC even when the training and test inputs are iid, and are especially useful for deterministic inputs and under covariate shift. Our experiments on linear models suggest that if the test and training inputs differ substantially, then XAIC and FAIC predictively outperform AIC, BIC and several other methods including Bayesian model averaging.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2019

Conformal Prediction Under Covariate Shift

We extend conformal prediction methodology beyond the case of exchangeab...
research
08/24/2018

Unknown Examples & Machine Learning Model Generalization

Over the past decades, researchers and ML practitioners have come up wit...
research
04/20/2023

The Dataset Multiplicity Problem: How Unreliable Data Impacts Predictions

We introduce dataset multiplicity, a way to study how inaccuracies, unce...
research
11/29/2017

Dimension Reduction for Robust Covariate Shift Correction

In the covariate shift learning scenario, the training and test covariat...
research
06/01/2020

Nonlinear observability algorithms with known and unknown inputs: analysis and implementation

The observability of a dynamical system is affected by the presence of e...
research
01/22/2021

Bayesian hierarchical stacking

Stacking is a widely used model averaging technique that yields asymptot...
research
07/20/2021

Evaluating Probabilistic Inference in Deep Learning: Beyond Marginal Predictions

A fundamental challenge for any intelligent system is prediction: given ...

Please sign up or login with your details

Forgot password? Click here to reset