DeepAI AI Chat
Log In Sign Up

Testing Conditional Independence on Discrete Data using Stochastic Complexity

by   Alexander Marx, et al.
Max Planck Society

Testing for conditional independence is a core aspect of constraint-based causal discovery. Although commonly used tests are perfect in theory, they often fail to reject independence in practice, especially when conditioning on multiple variables. We focus on discrete data and propose a new test based on the notion of algorithmic independence that we instantiate using stochastic complexity. Amongst others, we show that our proposed test, SCI, is an asymptotically unbiased as well as L_2 consistent estimator for conditional mutual information (CMI). Further, we show that SCI can be reformulated to find a sensible threshold for CMI that works well on limited samples. Empirical evaluation shows that SCI has a lower type II error than commonly used tests. As a result, we obtain a higher recall when we use SCI in causal discovery algorithms, without compromising the precision.


page 1

page 2

page 3

page 4


Kernel-based Conditional Independence Test and Application in Causal Discovery

Conditional independence testing is an important problem, especially in ...

Empirical Bayesian Approaches for Robust Constraint-based Causal Discovery under Insufficient Data

Causal discovery is to learn cause-effect relationships among variables ...

Approximate Kernel-based Conditional Independence Tests for Fast Non-Parametric Causal Discovery

Constraint-based causal discovery (CCD) algorithms require fast and accu...

A Robust Independence Test for Constraint-Based Learning of Causal Structure

Constraint-based (CB) learning is a formalism for learning a causal netw...

Is Information Theory Inherently a Theory of Causation?

Information theory gives rise to a novel method for causal skeleton disc...

Causal Discovery by Telling Apart Parents and Children

We consider the problem of inferring the directed, causal graph from obs...