Generalisation in Neural Networks Does not Require Feature Overlap

07/04/2021
by   Jeff Mitchell, et al.
0

That shared features between train and test data are required for generalisation in artificial neural networks has been a common assumption of both proponents and critics of these models. Here, we show that convolutional architectures avoid this limitation by applying them to two well known challenges, based on learning the identity function and learning rules governing sequences of words. In each case, successful performance on the test set requires generalising to features that were not present in the training data, which is typically not feasible for standard connectionist models. However, our experiments demonstrate that neural networks can succeed on such problems when they incorporate the weight sharing employed by convolutional architectures. In the image processing domain, such architectures are intended to reflect the symmetry under spatial translations of the natural world that such images depict. We discuss the role of symmetry in the two tasks and its connection to generalisation.

READ FULL TEXT
research
06/10/2019

SymNet: Symmetrical Filters in Convolutional Neural Networks

Symmetry is present in nature and science. In image processing, kernels ...
research
11/11/2022

Artificial neural networks for predicting the viscosity of lead-containing glasses

The viscosity of lead-containing glasses is of fundamental importance fo...
research
10/22/2020

Learning Invariances in Neural Networks

Invariances to translations have imbued convolutional neural networks wi...
research
08/25/2017

Identifying Mirror Symmetry Density with Delay in Spiking Neural Networks

The ability to rapidly identify symmetry and anti-symmetry is an essenti...
research
12/08/2017

Artificial Neural Networks that Learn to Satisfy Logic Constraints

Logic-based problems such as planning, theorem proving, or puzzles, typi...
research
04/14/2022

Relaxing Equivariance Constraints with Non-stationary Continuous Filters

Equivariances provide useful inductive biases in neural network modeling...
research
03/03/2022

Symmetry Structured Convolutional Neural Networks

We consider Convolutional Neural Networks (CNNs) with 2D structured feat...

Please sign up or login with your details

Forgot password? Click here to reset