Neural Networks for Learning Counterfactual G-Invariances from Single Environments

04/20/2021
by   S Chandra Mouli, et al.
0

Despite – or maybe because of – their astonishing capacity to fit data, neural networks are believed to have difficulties extrapolating beyond training data distribution. This work shows that, for extrapolations based on finite transformation groups, a model's inability to extrapolate is unrelated to its capacity. Rather, the shortcoming is inherited from a learning hypothesis: Examples not explicitly observed with infinitely many training examples have underspecified outcomes in the learner's model. In order to endow neural networks with the ability to extrapolate over group transformations, we introduce a learning framework counterfactually-guided by the learning hypothesis that any group invariance to (known) transformation groups is mandatory even without evidence, unless the learner deems it inconsistent with the training data. Unlike existing invariance-driven methods for (counterfactual) extrapolations, this framework allows extrapolations from a single environment. Finally, we introduce sequence and image extrapolation tasks that validate our framework and showcase the shortcomings of traditional approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/07/2023

On genuine invariance learning without weight-tying

In this paper, we investigate properties and limitations of invariance l...
research
10/12/2022

How Much Data Are Augmentations Worth? An Investigation into Scaling Laws, Invariance, and Implicit Regularization

Despite the clear performance benefits of data augmentations, little is ...
research
06/18/2021

Training or Architecture? How to Incorporate Invariance in Neural Networks

Many applications require the robustness, or ideally the invariance, of ...
research
11/11/2022

Equivariance with Learned Canonicalization Functions

Symmetry-based neural networks often constrain the architecture in order...
research
07/12/2018

HyperNets and their application to learning spatial transformations

In this paper we propose a conceptual framework for higher-order artific...
research
06/13/2018

Group Equivariant Capsule Networks

We present group equivariant capsule networks, a framework to introduce ...
research
02/06/2023

A Toy Model of Universality: Reverse Engineering How Networks Learn Group Operations

Universality is a key hypothesis in mechanistic interpretability – that ...

Please sign up or login with your details

Forgot password? Click here to reset