Improved techniques for deterministic l2 robustness

11/15/2022
by   Sahil Singla, et al.
0

Training convolutional neural networks (CNNs) with a strict 1-Lipschitz constraint under the l_2 norm is useful for adversarial robustness, interpretable gradients and stable training. 1-Lipschitz CNNs are usually designed by enforcing each layer to have an orthogonal Jacobian matrix (for all inputs) to prevent the gradients from vanishing during backpropagation. However, their performance often significantly lags behind that of heuristic methods to enforce Lipschitz constraints where the resulting CNN is not provably 1-Lipschitz. In this work, we reduce this gap by introducing (a) a procedure to certify robustness of 1-Lipschitz CNNs by replacing the last linear layer with a 1-hidden layer MLP that significantly improves their performance for both standard and provably robust accuracy, (b) a method to significantly reduce the training time per epoch for Skew Orthogonal Convolution (SOC) layers (>30% reduction for deeper networks) and (c) a class of pooling layers using the mathematical property that the l_2 distance of an input to a manifold is 1-Lipschitz. Using these methods, we significantly advance the state-of-the-art for standard and provable robust accuracies on CIFAR-10 (gains of +1.79% and +3.82%) and similarly on CIFAR-100 (+3.78% and +4.75%) across all networks. Code is available at <https://github.com/singlasahil14/improved_l2_robustness>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2021

Skew Orthogonal Convolutions

Training convolutional neural networks with a Lipschitz constraint under...
research
10/20/2022

LOT: Layer-wise Orthogonal Training on Improving l2 Certified Robustness

Recent studies show that training deep neural networks (DNNs) with Lipsc...
research
08/05/2021

Householder Activations for Provable Robustness against Adversarial Attacks

Training convolutional neural networks (CNNs) with a strict Lipschitz co...
research
08/05/2022

Almost-Orthogonal Layers for Efficient General-Purpose Lipschitz Networks

It is a highly desirable property for deep networks to be robust against...
research
03/06/2023

A Unified Algebraic Perspective on Lipschitz Neural Networks

Important research efforts have focused on the design and training of ne...
research
07/01/2021

Boosting Certified ℓ_∞ Robustness with EMA Method and Ensemble Model

The neural network with 1-Lipschitz property based on ℓ_∞-dist neuron ha...
research
11/24/2022

Towards Practical Control of Singular Values of Convolutional Layers

In general, convolutional neural networks (CNNs) are easy to train, but ...

Please sign up or login with your details

Forgot password? Click here to reset