A Simple and General Debiased Machine Learning Theorem with Finite Sample Guarantees

by   Victor Chernozhukov, et al.

Debiased machine learning is a meta algorithm based on bias correction and sample splitting to calculate confidence intervals for functionals (i.e. scalar summaries) of machine learning algorithms. For example, an analyst may desire the confidence interval for a treatment effect estimated with a neural network. We provide a nonasymptotic debiased machine learning theorem that encompasses any global or local functional of any machine learning algorithm that satisfies a few simple, interpretable conditions. Formally, we prove consistency, Gaussian approximation, and semiparametric efficiency by finite sample arguments. The rate of convergence is root-n for global functionals, and it degrades gracefully for local functionals. Our results culminate in a simple set of conditions that an analyst can use to translate modern learning theory rates into traditional statistical inference. The conditions reveal a new double robustness property for ill posed inverse problems.


page 1

page 2

page 3

page 4


A Finite Sample Theorem for Longitudinal Causal Inference with Machine Learning: Long Term, Dynamic, and Mediated Effects

I construct and justify confidence intervals for longitudinal causal par...

Debiased Kernel Methods

I propose a practical procedure based on bias correction and sample spli...

Geometric Inference for General High-Dimensional Linear Inverse Problems

This paper presents a unified geometric framework for the statistical an...

Orthogonal Statistical Learning

We provide excess risk guarantees for statistical learning in the presen...

Nonparametric identification is not enough, but randomized controlled trials are

We argue that randomized controlled trials (RCTs) are special even among...

Meta-Learners for Estimation of Causal Effects: Finite Sample Cross-Fit Performance

Estimation of causal effects using machine learning methods has become a...

Decontamination of Mutual Contamination Models

Many machine learning problems can be characterized by mutual contaminat...