Asymmetric Certified Robustness via Feature-Convex Neural Networks

02/03/2023
by   Samuel Pfrommer, et al.
0

Recent works have introduced input-convex neural networks (ICNNs) as learning models with advantageous training, inference, and generalization properties linked to their convex structure. In this paper, we propose a novel feature-convex neural network architecture as the composition of an ICNN with a Lipschitz feature map in order to achieve adversarial robustness. We consider the asymmetric binary classification setting with one "sensitive" class, and for this class we prove deterministic, closed-form, and easily-computable certified robust radii for arbitrary ℓ_p-norms. We theoretically justify the use of these models by characterizing their decision region geometry, extending the universal approximation theorem for ICNN regression to the classification setting, and proving a lower bound on the probability that such models perfectly fit even unstructured uniformly distributed data in sufficiently high dimensions. Experiments on Malimg malware classification and subsets of MNIST, Fashion-MNIST, and CIFAR-10 datasets show that feature-convex classifiers attain state-of-the-art certified ℓ_1-radii as well as substantial ℓ_2- and ℓ_∞-radii while being far more computationally efficient than any competitive baseline.

READ FULL TEXT

page 16

page 23

page 29

research
06/01/2020

Second-Order Provable Defenses against Adversarial Attacks

A robustness certificate is the minimum distance of a given input to the...
research
05/29/2023

Explainability in Simplicial Map Neural Networks

Simplicial map neural networks (SMNNs) are topology-based neural network...
research
09/30/2019

Universal Approximation with Certified Networks

Training neural networks to be certifiably robust is a powerful defense ...
research
12/18/2018

PROVEN: Certifying Robustness of Neural Networks with a Probabilistic Approach

With deep neural networks providing state-of-the-art machine learning mo...
research
07/24/2022

Can we achieve robustness from data alone?

Adversarial training and its variants have come to be the prevailing met...
research
02/04/2019

Deep One-Class Classification Using Data Splitting

This paper introduces a generic method which enables to use conventional...
research
02/09/2020

Robust binary classification with the 01 loss

The 01 loss is robust to outliers and tolerant to noisy data compared to...

Please sign up or login with your details

Forgot password? Click here to reset