Reducing Certified Regression to Certified Classification

08/29/2022
by   Zayd Hammoudeh, et al.
0

Adversarial training instances can severely distort a model's behavior. This work investigates certified regression defenses, which provide guaranteed limits on how much a regressor's prediction may change under a training-set attack. Our key insight is that certified regression reduces to certified classification when using median as a model's primary decision function. Coupling our reduction with existing certified classifiers, we propose six new provably-robust regressors. To the extent of our knowledge, this is the first work that certifies the robustness of individual regression predictions without any assumptions about the data distribution and model architecture. We also show that existing state-of-the-art certified classifiers often make overly-pessimistic assumptions that can degrade their provable guarantees. We introduce a tighter analysis of model robustness, which in many cases results in significantly improved certified guarantees. Lastly, we empirically demonstrate our approaches' effectiveness on both regression and classification data, where the accuracy of up to 50 under 1 corruption. Our source code is available at https://github.com/ZaydH/certified-regression.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/09/2019

Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers

Recent works have shown the effectiveness of randomized smoothing as a s...
research
06/14/2021

PopSkipJump: Decision-Based Attack for Probabilistic Classifiers

Most current classifiers are vulnerable to adversarial examples, small i...
research
10/03/2022

MultiGuard: Provably Robust Multi-label Classification against Adversarial Examples

Multi-label classification, which predicts a set of labels for an input,...
research
12/11/2022

Corruption-tolerant Algorithms for Generalized Linear Models

This paper presents SVAM (Sequential Variance-Altered MLE), a unified fr...
research
09/11/2022

Learning When to Say "I Don't Know"

We propose a new Reject Option Classification technique to identify and ...
research
02/05/2022

LyaNet: A Lyapunov Framework for Training Neural ODEs

We propose a method for training ordinary differential equations by usin...
research
01/17/2022

Cyberbullying Classifiers are Sensitive to Model-Agnostic Perturbations

A limited amount of studies investigates the role of model-agnostic adve...

Please sign up or login with your details

Forgot password? Click here to reset