Causal Transfer Random Forest: Combining Logged Data and Randomized Experiments for Robust Prediction

10/17/2020
by   Shuxi Zeng, et al.
0

It is often critical for prediction models to be robust to distributional shifts between training and testing data. Viewed from a causal perspective, the challenge is to distinguish the stable causal relationships from the unstable spurious correlations across shifts. We describe a causal transfer random forest (CTRF) that combines existing training data with a small amount of data from a randomized experiment to train a model which is robust to the feature shifts and therefore transfers to a new targeting distribution. Theoretically, we justify the robustness of the approach against feature shifts with the knowledge from causal learning. Empirically, we evaluate the CTRF using both synthetic data experiments and real-world experiments in the Bing Ads platform, including a click prediction task and in the context of an end-to-end counterfactual optimization system. The proposed CTRF produces robust predictions and outperforms most baseline methods compared in the presence of feature shifts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/18/2023

Causality-oriented robustness: exploiting general additive interventions

Since distribution shifts are common in real-world applications, there i...
research
02/20/2020

I-SPEC: An End-to-End Framework for Learning Transportable, Shift-Stable Models

Shifts in environment between development and deployment cause classical...
research
05/10/2023

Causal Information Splitting: Engineering Proxy Features for Robustness to Distribution Shifts

Statistical prediction models are often trained on data that is drawn fr...
research
09/21/2021

Beyond Discriminant Patterns: On the Robustness of Decision Rule Ensembles

Local decision rules are commonly understood to be more explainable, due...
research
02/13/2023

Causal Strategic Classification: A Tale of Two Shifts

When users can benefit from certain predictive outcomes, they may be pro...
research
05/18/2021

Distributionally Robust Learning in Heterogeneous Contexts

We consider the problem of learning from training data obtained in diffe...
research
01/20/2021

Distributional Anchor Regression

Prediction models often fail if train and test data do not stem from the...

Please sign up or login with your details

Forgot password? Click here to reset