Identifying Best Interventions through Online Importance Sampling

01/10/2017
by   Rajat Sen, et al.
0

Motivated by applications in computational advertising and systems biology, we consider the problem of identifying the best out of several possible soft interventions at a source node V in an acyclic causal directed graph, to maximize the expected value of a target node Y (located downstream of V). Our setting imposes a fixed total budget for sampling under various interventions, along with cost constraints on different types of interventions. We pose this as a best arm identification bandit problem with K arms where each arm is a soft intervention at V, and leverage the information leakage among the arms to provide the first gap dependent error and simple regret bounds for this problem. Our results are a significant improvement over the traditional best arm identification results. We empirically show that our algorithms outperform the state of the art in the Flow Cytometry data-set, and also apply our algorithm for model interpretation of the Inception-v3 deep net that classifies images.

READ FULL TEXT
research
06/06/2018

Causal Bandits with Propagating Inference

Bandit is a framework for designing sequential experiments. In each expe...
research
12/13/2020

Budgeted and Non-budgeted Causal Bandits

Learning good interventions in a causal graph can be modelled as a stoch...
research
11/08/2021

Identifying Best Fair Intervention

We study the problem of best arm identification with a fairness constrai...
research
02/10/2022

Bayes Optimal Algorithm is Suboptimal in Frequentist Best Arm Identification

We consider the fixed-budget best arm identification problem with Normal...
research
08/26/2022

Causal Bandits for Linear Structural Equation Models

This paper studies the problem of designing an optimal sequence of inter...
research
10/09/2018

Bridging the gap between regret minimization and best arm identification, with application to A/B tests

State of the art online learning procedures focus either on selecting th...
research
06/14/2021

Planning to Fairly Allocate: Probabilistic Fairness in the Restless Bandit Setting

Restless and collapsing bandits are commonly used to model constrained r...

Please sign up or login with your details

Forgot password? Click here to reset