On genuine invariance learning without weight-tying

08/07/2023
by   Artem Moskalev, et al.
0

In this paper, we investigate properties and limitations of invariance learned by neural networks from the data compared to the genuine invariance achieved through invariant weight-tying. To do so, we adopt a group theoretical perspective and analyze invariance learning in neural networks without weight-tying constraints. We demonstrate that even when a network learns to correctly classify samples on a group orbit, the underlying decision-making in such a model does not attain genuine invariance. Instead, learned invariance is strongly conditioned on the input data, rendering it unreliable if the input distribution shifts. We next demonstrate how to guide invariance learning toward genuine invariance by regularizing the invariance of a model at the training. To this end, we propose several metrics to quantify learned invariance: (i) predictive distribution invariance, (ii) logit invariance, and (iii) saliency invariance similarity. We show that the invariance learned with the invariance error regularization closely reassembles the genuine invariance of weight-tying models and reliably holds even under a severe input distribution shift. Closer analysis of the learned invariance also reveals the spectral decay phenomenon, when a network chooses to achieve the invariance to a specific transformation group by reducing the sensitivity to any input perturbation.

READ FULL TEXT

page 1

page 5

page 6

research
10/07/2022

In What Ways Are Deep Neural Networks Invariant and How Should We Measure This?

It is often said that a deep learning model is "invariant" to some speci...
research
11/28/2022

Malign Overfitting: Interpolation Can Provably Preclude Invariance

Learned classifiers should often possess certain invariance properties m...
research
04/20/2021

Neural Networks for Learning Counterfactual G-Invariances from Single Environments

Despite – or maybe because of – their astonishing capacity to fit data, ...
research
07/03/2017

Appearance invariance in convolutional networks with neighborhood similarity

We present a neighborhood similarity layer (NSL) which induces appearanc...
research
06/12/2020

Traversal-invariant characterizations of logarithmic space

We give a novel descriptive-complexity theoretic characterization of L a...
research
11/02/2017

The (Un)reliability of saliency methods

Saliency methods aim to explain the predictions of deep neural networks....
research
11/08/2019

Discovering Invariances in Healthcare Neural Networks

We study the invariance characteristics of pre-trained predictive models...

Please sign up or login with your details

Forgot password? Click here to reset