Affine Symmetries and Neural Network Identifiability

06/21/2020
by   Verner Vlačić, et al.
0

We address the following question of neural network identifiability: Suppose we are given a function f:ℝ^m→ℝ^n and a nonlinearity ρ. Can we specify the architecture, weights, and biases of all feed-forward neural networks with respect to ρ giving rise to f? Existing literature on the subject suggests that the answer should be yes, provided we are only concerned with finding networks that satisfy certain "genericity conditions". Moreover, the identified networks are mutually related by symmetries of the nonlinearity. For instance, the tanh function is odd, and so flipping the signs of the incoming and outgoing weights of a neuron does not change the output map of the network. The results known hitherto, however, apply either to single-layer networks, or to networks satisfying specific structural assumptions (such as full connectivity), as well as to specific nonlinearities. In an effort to answer the identifiability question in greater generality, we consider arbitrary nonlinearities with potentially complicated affine symmetries, and we show that the symmetries can be used to find a rich set of networks giving rise to the same function f. The set obtained in this manner is, in fact, exhaustive (i.e., it contains all networks giving rise to f) unless there exists a network 𝒜 "with no internal symmetries" giving rise to the identically zero function. This result can thus be interpreted as an analog of the rank-nullity theorem for linear operators. We furthermore exhibit a class of "tanh-type" nonlinearities (including the tanh function itself) for which such a network 𝒜 does not exist, thereby solving the identifiability question for these nonlinearities in full generality. Finally, we show that this class contains nonlinearities with arbitrarily complicated symmetries.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/11/2019

Neural network identifiability for a family of sigmoidal nonlinearities

This paper addresses the following question of neural network identifiab...
research
09/07/2021

On the space of coefficients of a Feed Forward Neural Network

We define and establish the conditions for `equivalent neural networks' ...
research
08/17/2001

Artificial Neurons with Arbitrarily Complex Internal Structures

Artificial neurons with arbitrarily complex internal structure are intro...
research
05/03/2022

Optimal (r,δ)-LRCs from zero-dimensional affine variety codes and their subfield-subcodes

We introduce zero-dimensional affine variety codes (ZAVCs) which can be ...
research
03/04/2021

Clusterability in Neural Networks

The learned weights of a neural network have often been considered devoi...
research
11/27/2018

Knots in random neural networks

The weights of a neural network are typically initialized at random, and...
research
03/23/2020

Safe Crossover of Neural Networks Through Neuron Alignment

One of the main and largely unexplored challenges in evolvingthe weights...

Please sign up or login with your details

Forgot password? Click here to reset