Improved generalization bounds for robust learning

10/04/2018
by   Idan Attias, et al.
0

We consider a model of robust learning in an adversarial environment. The learner gets uncorrupted training data with access to possible corruptions that may be used by the adversary during testing. Their aim is to build a robust classifier that would be tested on future adversarially corrupted data. We use a zero-sum game between the learner and the adversary as our game theoretic framework. The adversary is limited to k possible corruptions for each input. Our model is closely related to the adversarial examples model of one of Schmidt et al. (2018) and Madry et al. (2017). We refer to binary and multi-class classification settings, and regression setting. Our main results are generalization bounds for all settings. For the binary classification setting, we improve a generalization bound previously found in Feige et al. (2015). We generalize to the case of weighted average of hypotheses from H that is not limited to be finite. The sample complexity has been improved from Ø(1/ϵ^4(|H|/δ)) to Ø(1/ϵ^2(k(k)VC(H)+1/δ)). The core of all is proofs based on bounds of the empirical Rademacher complexity. For the binary classification, we use a known regret minimization algorithm of Feige et al. that uses an ERM oracle as a blackbox and we expand on the multi-class and regression settings. The algorithm provides us near optimal policies for the players on a given training sample. The learner starts with a fixed hypothesis class H and chooses a convex combination of hypotheses from H. The learner's loss is measured on adversarial corrupted inputs. Along the way, we obtain results on fat-shattering dimension and Rademacher complexity of k-fold maxima over function classes; these may be of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/07/2020

Sample-efficient proper PAC learning with approximate differential privacy

In this paper we prove that the sample complexity of properly learning a...
research
05/28/2021

Towards optimally abstaining from prediction

A common challenge across all areas of machine learning is that training...
research
10/06/2022

On Optimal Learning Under Targeted Data Poisoning

Consider the task of learning a hypothesis class ℋ in the presence of an...
research
10/10/2019

Learning from Multiple Corrupted Sources, with Application to Learning from Label Proportions

We study binary classification in the setting where the learner is prese...
research
10/12/2022

When are Local Queries Useful for Robust Learning?

Distributional assumptions have been shown to be necessary for the robus...
research
05/26/2015

Some Open Problems in Optimal AdaBoost and Decision Stumps

The significance of the study of the theoretical and practical propertie...
research
06/03/2022

Generalization for multiclass classification with overparameterized linear models

Via an overparameterized linear model with Gaussian features, we provide...

Please sign up or login with your details

Forgot password? Click here to reset