Combinatorial Causal Bandits

06/04/2022
by   Shi Feng, et al.
0

In combinatorial causal bandits (CCB), the learning agent chooses at most K variables in each round to intervene, collects feedback from the observed variables, with the goal of minimizing expected regret on the target variable Y. Different from all prior studies on causal bandits, CCB needs to deal with exponentially large action space. We study under the context of binary generalized linear models (BGLMs) with a succinct parametric representation of the causal models. We present the algorithm BGLM-OFU for Markovian BGLMs (i.e. no hidden variables) based on the maximum likelihood estimation method, and show that it achieves O(√(T)log T) regret, where T is the time horizon. For the special case of linear models with hidden variables, we apply causal inference techniques such as the do-calculus to convert the original model into a Markovian model, and then show that our BGLM-OFU algorithm and another algorithm based on the linear regression both solve such linear models with hidden variables. Our novelty includes (a) considering the combinatorial intervention action space, (b) considering general causal models including ones with hidden variables, (c) integrating and adapting techniques from diverse studies such as generalized linear bandits and online influence maximization, and (d) not relying on unrealistic assumptions such as knowing the joint distribution of the parents of Y under all interventions used in some prior studies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/31/2023

Combinatorial Causal Bandits without Graph Skeleton

In combinatorial causal bandits (CCB), the learning agent chooses a subs...
research
06/16/2022

Pure Exploration of Causal Bandits

Causal bandit problem integrates causal inference with multi-armed bandi...
research
06/10/2016

Causal Bandits: Learning Good Interventions via Causal Inference

We study the problem of using causal models to improve the rate at which...
research
03/07/2021

Hierarchical Causal Bandit

Causal bandit is a nascent learning model where an agent sequentially ex...
research
06/13/2023

Additive Causal Bandits with Unknown Graph

We explore algorithms to select actions in the causal bandit setting whe...
research
03/03/2021

Regularizing towards Causal Invariance: Linear Models with Proxies

We propose a method for learning linear models whose predictive performa...
research
03/02/2018

Detecting non-causal artifacts in multivariate linear regression models

We consider linear models where d potential causes X_1,...,X_d are corre...

Please sign up or login with your details

Forgot password? Click here to reset