Interpreting a Penalty as the Influence of a Bayesian Prior

02/01/2020
by   Pierre Wolinski, et al.
0

In machine learning, it is common to optimize the parameters of a probabilistic model, modulated by a somewhat ad hoc regularization term that penalizes some values of the parameters. Regularization terms appear naturally in Variational Inference (VI), a tractable way to approximate Bayesian posteriors: the loss to optimize contains a Kullback–Leibler divergence term between the approximate posterior and a Bayesian prior. We fully characterize which regularizers can arise this way, and provide a systematic way to compute the corresponding prior. This viewpoint also provides a prediction for useful values of the regularization factor in neural networks. We apply this framework to regularizers such as L1 or group-Lasso.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/26/2019

Variational Inference to Measure Model Uncertainty in Deep Neural Networks

We present a novel approach for training deep neural networks in a Bayes...
research
11/06/2018

Deep Probabilistic Ensembles: Approximate Variational Inference through KL Regularization

In this paper, we introduce Deep Probabilistic Ensembles (DPEs), a scala...
research
02/17/2019

Bayesian Regularization: From Tikhonov to Horseshoe

Bayesian regularization is a central tool in modern-day statistical and ...
research
07/01/2019

Radial Bayesian Neural Networks: Robust Variational Inference In Big Models

We propose Radial Bayesian Neural Networks: a variational distribution f...
research
02/07/2020

Random weighting to approximate posterior inference in LASSO regression

We consider a general-purpose approximation approach to Bayesian inferen...
research
10/14/2022

Bayesian Regularization on Function Spaces via Q-Exponential Process

Regularization is one of the most important topics in optimization, stat...

Please sign up or login with your details

Forgot password? Click here to reset