Continuously Constructive Deep Neural Networks

04/07/2018
by   Ozan Irsoy, et al.
0

Traditionally, deep learning algorithms update the network weights whereas the network architecture is chosen manually, using a process of trial and error. In this work, we propose two novel approaches that automatically update the network structure while also learning its weights. The novelty of our approach lies in our parameterization where the depth, or additional complexity, is encapsulated continuously in the parameter space through control parameters that add additional complexity. We propose two methods: In tunnel networks, this selection is done at the level of a hidden unit, and in budding perceptrons, this is done at the level of a network layer; updating this control parameter introduces either another hidden unit or another hidden layer. We show the effectiveness of our methods on the synthetic two-spirals data and on two real data sets of MNIST and MIRFLICKR, where we see that our proposed methods, with the same set of hyperparameters, can correctly adjust the network complexity to the task complexity.

READ FULL TEXT

page 4

page 5

page 6

page 8

research
06/09/2023

Hidden symmetries of ReLU networks

The parameter space for any fixed architecture of feedforward ReLU neura...
research
02/07/2022

Approximation error of single hidden layer neural networks with fixed weights

This paper provides an explicit formula for the approximation error of s...
research
05/06/2014

Pulling back error to the hidden-node parameter technology: Single-hidden-layer feedforward network without output weight

According to conventional neural network theories, the feature of single...
research
09/21/2022

Variational Inference for Infinitely Deep Neural Networks

We introduce the unbounded depth neural network (UDN), an infinitely dee...
research
05/10/2018

Monotone Learning with Rectified Wire Networks

We introduce a new neural network model, together with a tractable and m...
research
01/09/2019

A Constructive Approach for One-Shot Training of Neural Networks Using Hypercube-Based Topological Coverings

In this paper we presented a novel constructive approach for training de...
research
11/23/2021

Critical initialization of wide and deep neural networks through partial Jacobians: general theory and applications to LayerNorm

Deep neural networks are notorious for defying theoretical treatment. Ho...

Please sign up or login with your details

Forgot password? Click here to reset