Bayesian Update with Importance Sampling: Required Sample Size

09/22/2020
by   Daniel Sanz-Alonso, et al.
0

Importance sampling is used to approximate Bayes' rule in many computational approaches to Bayesian inverse problems, data assimilation and machine learning. This paper reviews and further investigates the required sample size for importance sampling in terms of the χ^2-divergence between target and proposal. We develop general abstract theory and illustrate through numerous examples the roles that dimension, noise-level and other model parameters play in approximating the Bayesian update with importance sampling. Our examples also facilitate a new direct comparison of standard and optimal proposals for particle filtering.

READ FULL TEXT
research
10/08/2019

Distilling importance sampling

The two main approaches to Bayesian inference are sampling and optimisat...
research
03/02/2018

Gradient-based Sampling: An Adaptive Importance Sampling for Least-squares

In modern data analysis, random sampling is an efficient and widely-used...
research
11/04/2019

Importance Sampling via Local Sensitivity

Given a loss function F:X→R^+ that can be written as the sum of losses o...
research
03/26/2019

A layered multiple importance sampling scheme for focused optimal Bayesian experimental design

We develop a new computational approach for "focused" optimal Bayesian e...
research
10/27/2021

How Important is Importance Sampling for Deep Budgeted Training?

Long iterative training processes for Deep Neural Networks (DNNs) are co...
research
04/04/2022

Test Against High-Dimensional Uncertainties: Accelerated Evaluation of Autonomous Vehicles with Deep Importance Sampling

Evaluating the performance of autonomous vehicles (AV) and their complex...
research
12/18/2019

Tree pyramidal adaptive importance sampling

This paper introduces Tree-Pyramidal Adaptive Importance Sampling (TP-AI...

Please sign up or login with your details

Forgot password? Click here to reset