On Optimal Learning Under Targeted Data Poisoning

10/06/2022
by   Steve Hanneke, et al.
0

Consider the task of learning a hypothesis class ℋ in the presence of an adversary that can replace up to an η fraction of the examples in the training set with arbitrary adversarial examples. The adversary aims to fail the learner on a particular target test point x which is known to the adversary but not to the learner. In this work we aim to characterize the smallest achievable error ϵ=ϵ(η) by the learner in the presence of such an adversary in both realizable and agnostic settings. We fully achieve this in the realizable setting, proving that ϵ=Θ(𝚅𝙲(ℋ)·η), where 𝚅𝙲(ℋ) is the VC dimension of ℋ. Remarkably, we show that the upper bound can be attained by a deterministic learner. In the agnostic setting we reveal a more elaborate landscape: we devise a deterministic learner with a multiplicative regret guarantee of ϵ≤ C·𝙾𝙿𝚃 + O(𝚅𝙲(ℋ)·η), where C > 1 is a universal numerical constant. We complement this by showing that for any deterministic learner there is an attack which worsens its error to at least 2·𝙾𝙿𝚃. This implies that a multiplicative deterioration in the regret is unavoidable in this case. Finally, the algorithms we develop for achieving the optimal rates are inherently improper. Nevertheless, we show that for a variety of natural concept classes, such as linear classifiers, it is possible to retain the dependence ϵ=Θ_ℋ(η) by a proper algorithm in the realizable setting. Here Θ_ℋ conceals a polynomial dependence on 𝚅𝙲(ℋ).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2023

Adversarial Resilience in Sequential Prediction via Abstention

We study the problem of sequential prediction in the stochastic setting ...
research
10/04/2018

Improved generalization bounds for robust learning

We consider a model of robust learning in an adversarial environment. Th...
research
10/20/2021

Transductive Robust Learning Guarantees

We study the problem of adversarially robust learning in the transductiv...
research
03/02/2022

Adversarially Robust Learning with Tolerance

We study the problem of tolerant adversarial PAC learning with respect t...
research
10/21/2010

On the Foundations of Adversarial Single-Class Classification

Motivated by authentication, intrusion and spam detection applications w...
research
03/08/2022

Robustly-reliable learners under poisoning attacks

Data poisoning attacks, in which an adversary corrupts a training set wi...
research
11/15/2019

Query Complexity of Bayesian Private Learning

We study the query complexity of Bayesian Private Learning: a learner wi...

Please sign up or login with your details

Forgot password? Click here to reset