Neural networks with differentiable structure

06/20/2016
by   Thomas Miconi, et al.
0

While gradient descent has proven highly successful in learning connection weights for neural networks, the actual structure of these networks is usually determined by hand, or by other optimization algorithms. Here we describe a simple method to make network structure differentiable, and therefore accessible to gradient descent. We test this method on recurrent neural networks applied to simple sequence prediction problems. Starting with initial networks containing only one node, the method automatically builds networks that successfully solve the tasks. The number of nodes in the final network correlates with task difficulty. The method can dynamically increase network size in response to an abrupt complexification in the task; however, reduction in network size in response to task simplification is not evident for reasonable meta-parameters. The method does not penalize network performance for these test tasks: variable-size networks actually reach better performance than fixed-size networks of higher, lower or identical size. We conclude by discussing how this method could be applied to more complex networks, such as feedforward layered networks, or multiple-area networks of arbitrary shape.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2016

Learning to learn by gradient descent by gradient descent

The move from hand-designed features to learned features in machine lear...
research
10/03/2018

Optimization Algorithm Inspired Deep Neural Network Structure Design

Deep neural networks have been one of the dominant machine learning appr...
research
06/22/2020

Neural networks adapting to datasets: learning network size and topology

We introduce a flexible setup allowing for a neural network to learn bot...
research
10/08/2019

Differentiable Sparsification for Deep Neural Networks

A deep neural network has relieved the burden of feature engineering by ...
research
04/06/2018

Differentiable plasticity: training plastic neural networks with backpropagation

How can we build agents that keep learning from experience, quickly and ...
research
10/11/2021

Mining the Weights Knowledge for Optimizing Neural Network Structures

Knowledge embedded in the weights of the artificial neural network can b...
research
06/21/2012

A biological gradient descent for prediction through a combination of STDP and homeostatic plasticity

Identifying, formalizing and combining biological mechanisms which imple...

Please sign up or login with your details

Forgot password? Click here to reset