Skew Orthogonal Convolutions

05/24/2021
by   Sahil Singla, et al.
0

Training convolutional neural networks with a Lipschitz constraint under the l_2 norm is useful for provable adversarial robustness, interpretable gradients, stable training, etc. While 1-Lipschitz networks can be designed by imposing a 1-Lipschitz constraint on each layer, training such networks requires each layer to be gradient norm preserving (GNP) to prevent gradients from vanishing. However, existing GNP convolutions suffer from slow training, lead to significant reduction in accuracy and provide no guarantees on their approximations. In this work, we propose a GNP convolution layer called () that uses the following mathematical property: when a matrix is Skew-Symmetric, its exponential function is an orthogonal matrix. To use this property, we first construct a convolution filter whose Jacobian is Skew-Symmetric. Then, we use the Taylor series expansion of the Jacobian exponential to construct the layer that is orthogonal. To efficiently implement , we keep a finite number of terms from the Taylor series and provide a provable guarantee on the approximation error. Our experiments on CIFAR-10 and CIFAR-100 show that allows us to train provably Lipschitz, large convolutional neural networks significantly faster than prior works while achieving significant improvements for both standard and certified robust accuracies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/15/2022

Improved techniques for deterministic l2 robustness

Training convolutional neural networks (CNNs) with a strict 1-Lipschitz ...
research
08/05/2021

Householder Activations for Provable Robustness against Adversarial Attacks

Training convolutional neural networks (CNNs) with a strict Lipschitz co...
research
11/03/2019

Preventing Gradient Attenuation in Lipschitz Constrained Convolutional Networks

Lipschitz constraints under L2 norm on deep neural networks are useful f...
research
10/20/2022

LOT: Layer-wise Orthogonal Training on Improving l2 Certified Robustness

Recent studies show that training deep neural networks (DNNs) with Lipsc...
research
08/10/2020

Intelligent Matrix Exponentiation

We present a novel machine learning architecture that uses the exponenti...
research
08/05/2022

Almost-Orthogonal Layers for Efficient General-Purpose Lipschitz Networks

It is a highly desirable property for deep networks to be robust against...
research
11/13/2018

Sorting out Lipschitz function approximation

Training neural networks subject to a Lipschitz constraint is useful for...

Please sign up or login with your details

Forgot password? Click here to reset