It's Hard for Neural Networks To Learn the Game of Life

09/03/2020
by   Jacob M. Springer, et al.
0

Efforts to improve the learning abilities of neural networks have focused mostly on the role of optimization methods rather than on weight initializations. Recent findings, however, suggest that neural networks rely on lucky random initial weights of subnetworks called "lottery tickets" that converge quickly to a solution. To investigate how weight initializations affect performance, we examine small convolutional networks that are trained to predict n steps of the two-dimensional cellular automaton Conway's Game of Life, the update rules of which can be implemented efficiently in a 2n+1 layer convolutional network. We find that networks of this architecture trained on this task rarely converge. Rather, networks require substantially more parameters to consistently converge. In addition, near-minimal architectures are sensitive to tiny changes in parameters: changing the sign of a single weight can cause the network to fail to learn. Finally, we observe a critical value d_0 such that training minimal networks with examples in which cells are alive with probability d_0 dramatically increases the chance of convergence to a solution. We conclude that training convolutional neural networks to learn the input/output function represented by n steps of Game of Life exhibits many characteristics predicted by the lottery ticket hypothesis, namely, that the size of the networks required to learn this function are often significantly larger than the minimal network required to implement the function.

READ FULL TEXT

page 3

page 7

research
08/08/2022

Understanding Weight Similarity of Neural Networks via Chain Normalization Rule and Hypothesis-Training-Testing

We present a weight similarity measure method that can quantify the weig...
research
06/29/2023

Weight Compander: A Simple Weight Reparameterization for Regularization

Regularization is a set of techniques that are used to improve the gener...
research
03/09/2018

The Lottery Ticket Hypothesis: Training Pruned Neural Networks

Recent work on neural network pruning indicates that, at training time, ...
research
05/17/2019

Sequential training algorithm for neural networks

A sequential training method for large-scale feedforward neural networks...
research
04/23/2018

N-fold Superposition: Improving Neural Networks by Reducing the Noise in Feature Maps

Considering the use of Fully Connected (FC) layer limits the performance...
research
06/12/2020

How many winning tickets are there in one DNN?

The recent lottery ticket hypothesis proposes that there is one sub-netw...
research
03/29/2019

A Local Approach to Forward Model Learning: Results on the Game of Life Game

This paper investigates the effect of learning a forward model on the pe...

Please sign up or login with your details

Forgot password? Click here to reset