Causal Bandits: Learning Good Interventions via Causal Inference

06/10/2016
by   Finnian Lattimore, et al.
0

We study the problem of using causal models to improve the rate at which good interventions can be learned online in a stochastic environment. Our formalism combines multi-arm bandits and causal inference to model a novel type of bandit feedback that is not exploited by existing approaches. We propose a new algorithm that exploits the causal feedback and prove a bound on its simple regret that is strictly better (in all quantities) than algorithms that do not use the additional causal information.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/11/2019

Regret Analysis of Causal Bandit Problems

We study how to learn optimal interventions sequentially given causal in...
research
01/16/2020

Combining Offline Causal Inference and Online Bandit Learning for Data Driven Decisions

A fundamental question for companies is: How to make good decisions with...
research
10/30/2019

Bayesian causal inference via probabilistic program synthesis

Causal inference can be formalized as Bayesian inference that combines a...
research
06/04/2022

Combinatorial Causal Bandits

In combinatorial causal bandits (CCB), the learning agent chooses at mos...
research
10/23/2019

Strategic Adaptation to Classifiers: A Causal Perspective

Consequential decision-making incentivizes individuals to adapt their be...
research
09/15/2022

Stochastic Tree Ensembles for Estimating Heterogeneous Effects

Determining subgroups that respond especially well (or poorly) to specif...
research
05/30/2019

Rarely-switching linear bandits: optimization of causal effects for the real world

Exploring the effect of policies in many real world scenarios is difficu...

Please sign up or login with your details

Forgot password? Click here to reset