Bregman Deviations of Generic Exponential Families

01/18/2022
by   Sayak Ray Chowdhury, et al.
0

We revisit the method of mixture technique, also known as the Laplace method, to study the concentration phenomenon in generic exponential families. Combining the properties of Bregman divergence associated with log-partition function of the family with the method of mixtures for super-martingales, we establish a generic bound controlling the Bregman divergence between the parameter of the family and a finite sample estimate of the parameter. Our bound is time-uniform and makes appear a quantity extending the classical information gain to exponential families, which we call the Bregman information gain. For the practitioner, we instantiate this novel bound to several classical families, e.g., Gaussian, Bernoulli, Exponential and Chi-square yielding explicit forms of the confidence sets and the Bregman information gain. We further numerically compare the resulting confidence bounds to state-of-the-art alternatives for time-uniform concentration and show that this novel method yields competitive results. Finally, we highlight how our results can be applied in a linear contextual multi-armed bandit problem.

READ FULL TEXT

page 9

page 29

page 35

research
05/24/2017

Boundary Crossing Probabilities for General Exponential Families

We consider parametric exponential families of dimension K on the real l...
research
07/12/2013

Thompson Sampling for 1-Dimensional Exponential Family Bandits

Thompson Sampling has been demonstrated in many complex bandit models, h...
research
01/30/2020

Finite-time Analysis of Kullback-Leibler Upper Confidence Bounds for Optimal Adaptive Allocation with Multiple Plays and Markovian Rewards

We study an extension of the classic stochastic multi-armed bandit probl...
research
12/02/2021

Indexed Minimum Empirical Divergence for Unimodal Bandits

We consider a multi-armed bandit problem specified by a set of one-dimen...
research
06/29/2023

Poisson and Gaussian approximations of the power divergence family of statistics

Consider the family of power divergence statistics based on n trials, ea...
research
03/28/2016

Generalized Exponential Concentration Inequality for Rényi Divergence Estimation

Estimating divergences in a consistent way is of great importance in man...
research
10/09/2014

Distributed Estimation, Information Loss and Exponential Families

Distributed learning of probabilistic models from multiple data reposito...

Please sign up or login with your details

Forgot password? Click here to reset