Metric Entropy Duality and the Sample Complexity of Outcome Indistinguishability

03/09/2022
by   Lunjia Hu, et al.
4

We give the first sample complexity characterizations for outcome indistinguishability, a theoretical framework of machine learning recently introduced by Dwork, Kim, Reingold, Rothblum, and Yona (STOC 2021). In outcome indistinguishability, the goal of the learner is to output a predictor that cannot be distinguished from the target predictor by a class D of distinguishers examining the outcomes generated according to the predictors' predictions. In the distribution-specific and realizable setting where the learner is given the data distribution together with a predictor class P containing the target predictor, we show that the sample complexity of outcome indistinguishability is characterized by the metric entropy of P w.r.t. the dual Minkowski norm defined by D, and equivalently by the metric entropy of D w.r.t. the dual Minkowski norm defined by P. This equivalence makes an intriguing connection to the long-standing metric entropy duality conjecture in convex geometry. Our sample complexity characterization implies a variant of metric entropy duality, which we show is nearly tight. In the distribution-free setting, we focus on the case considered by Dwork et al. where P contains all possible predictors, hence the sample complexity only depends on D. In this setting, we show that the sample complexity of outcome indistinguishability is characterized by the fat-shattering dimension of D. We also show a strong sample complexity separation between realizable and agnostic outcome indistinguishability in both the distribution-free and the distribution-specific settings. This is in contrast to distribution-free (resp. distribution-specific) PAC learning where the sample complexity in both the realizable and the agnostic settings can be characterized by the VC dimension (resp. metric entropy).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2022

Estimation of Entropy in Constant Space with Improved Sample Complexity

Recent work of Acharya et al. (NeurIPS 2019) showed how to estimate the ...
research
03/29/2021

The Sample Complexity of Distribution-Free Parity Learning in the Robust Shuffle Model

We provide a lowerbound on the sample complexity of distribution-free pa...
research
05/11/2015

Sample complexity of learning Mahalanobis distance metrics

Metric learning seeks a transformation of the feature space that enhance...
research
11/23/2022

A Moment-Matching Approach to Testable Learning and a New Characterization of Rademacher Complexity

A remarkable recent paper by Rubinfeld and Vasilyan (2022) initiated the...
research
11/16/2022

Comparative Learning: A Sample Complexity Theory for Two Hypothesis Classes

In many learning theory problems, a central role is played by a hypothes...
research
10/26/2022

Learning versus Refutation in Noninteractive Local Differential Privacy

We study two basic statistical tasks in non-interactive local differenti...
research
05/16/2023

Private Everlasting Prediction

A private learner is trained on a sample of labeled points and generates...

Please sign up or login with your details

Forgot password? Click here to reset