General Bayesian Loss Function Selection and the use of Improper Models

06/02/2021
by   Jack Jewson, et al.
0

Statisticians often face the choice between using probability models or a paradigm defined by minimising a loss function. Both approaches are useful and, if the loss can be re-cast into a proper probability model, there are many tools to decide which model or loss is more appropriate for the observed data, in the sense of explaining the data's nature. However, when the loss leads to an improper model, there are no principled ways to guide this choice. We address this task by combining the Hyvärinen score, which naturally targets infinitesimal relative probabilities, and general Bayesian updating, which provides a unifying framework for inference on losses and models. Specifically we propose the H-score, a general Bayesian selection criterion and prove that it consistently selects the (possibly improper) model closest to the data-generating truth in Fisher's divergence. We also prove that an associated H-posterior consistently learns optimal hyper-parameters featuring in loss functions, including a challenging tempering parameter in generalised Bayesian inference. As salient examples, we consider robust regression and non-parametric density estimation where popular loss functions define improper models for the data and hence cannot be dealt with using standard model selection tools. These examples illustrate advantages in robustness-efficiency trade-offs and provide a Bayesian implementation for kernel density estimation, opening a new avenue for Bayesian non-parametrics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/26/2021

Optimal Bayesian Estimation of a Regression Curve, a Conditional Density and a Conditional Distribution

In this paper several related estimation problems are addressed from a B...
research
01/30/2013

Decision Theoretic Foundations of Graphical Model Selection

This paper describes a decision theoretic formulation of learning the gr...
research
05/18/2023

Evidence Networks: simple losses for fast, amortized, neural Bayesian model comparison

Evidence Networks can enable Bayesian model comparison when state-of-the...
research
05/10/2018

On the Universality of the Logistic Loss Function

A loss function measures the discrepancy between the true values (observ...
research
07/11/2023

Bayesian taut splines for estimating the number of modes

The number of modes in a probability density function is representative ...
research
07/08/2020

Generalised Bayes Updates with f-divergences through Probabilistic Classifiers

A stream of algorithmic advances has steadily increased the popularity o...

Please sign up or login with your details

Forgot password? Click here to reset