Reliable Learning for Test-time Attacks and Distribution Shift

04/06/2023
by   Maria-Florina Balcan, et al.
0

Machine learning algorithms are often used in environments which are not captured accurately even by the most carefully obtained training data, either due to the possibility of `adversarial' test-time attacks, or on account of `natural' distribution shift. For test-time attacks, we introduce and analyze a novel robust reliability guarantee, which requires a learner to output predictions along with a reliability radius η, with the meaning that its prediction is guaranteed to be correct as long as the adversary has not perturbed the test point farther than a distance η. We provide learners that are optimal in the sense that they always output the best possible reliability radius on any test point, and we characterize the reliable region, i.e. the set of points where a given reliability radius is attainable. We additionally analyze reliable learners under distribution shift, where the test points may come from an arbitrary distribution Q different from the training distribution P. For both cases, we bound the probability mass of the reliable region for several interesting examples, for linear separators under nearly log-concave and s-concave distributions, as well as for smooth boundary classifiers under smooth probability distributions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/09/2022

Test-time Recalibration of Conformal Predictors Under Distribution Shift Based on Unlabeled Examples

Modern image classifiers achieve high predictive accuracy, but the predi...
research
03/08/2022

Robustly-reliable learners under poisoning attacks

Data poisoning attacks, in which an adversary corrupts a training set wi...
research
08/23/2023

Sample Complexity of Robust Learning against Evasion Attacks

It is becoming increasingly important to understand the vulnerability of...
research
12/12/2011

Robust Learning via Cause-Effect Models

We consider the problem of function estimation in the case where the dat...
research
06/06/2018

Adversarial Regression with Multiple Learners

Despite the considerable success enjoyed by machine learning techniques ...
research
04/04/2022

DAD: Data-free Adversarial Defense at Test Time

Deep models are highly susceptible to adversarial attacks. Such attacks ...
research
10/12/2022

Double Bubble, Toil and Trouble: Enhancing Certified Robustness through Transitivity

In response to subtle adversarial examples flipping classifications of n...

Please sign up or login with your details

Forgot password? Click here to reset