Self Configuration in Machine Learning

09/17/2018
by   Eugene Wong, et al.
0

In this paper we first present a class of algorithms for training multi-level neural networks with a quadratic cost function one layer at a time starting from the input layer. The algorithm is based on the fact that for any layer to be trained, the effect of a direct connection to an optimized linear output layer can be computed without the connection being made. Thus, starting from the input layer, we can train each layer in succession in isolation from the other layers. Once trained, the weights are kept fixed and the outputs of the trained layer then serve as the inputs to the next layer to be trained. The result is a very fast algorithm. The simplicity of this training arrangement allows the activation function and step size in weight adjustment to be adaptive and self-adjusting. Furthermore, the stability of the training process allows relatively large steps to be taken and thereby achieving in even greater speeds. Finally, in our context configuring the network means determining the number of outputs for each layer. By decomposing the overall cost function into separate components related to approximation and estimation, we obtain an optimization formula for determining the number of outputs for each layer. With the ability to self-configure and set parameters, we now have more than a fast training algorithm, but the ability to build automatically a fully trained deep neural network starting with nothing more than data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2019

On the Approximation Properties of Neural Networks

We prove two new results concerning the approximation properties of neur...
research
07/26/2023

Understanding Deep Neural Networks via Linear Separability of Hidden Layers

In this paper, we measure the linear separability of hidden layer output...
research
04/16/2020

Machine-learning-based methods for output only structural modal identification

In this study, we propose a machine-learning-based approach to identify ...
research
02/26/2018

A representer theorem for deep neural networks

We propose to optimize the activation functions of a deep neural network...
research
10/02/2004

Applying Policy Iteration for Training Recurrent Neural Networks

Recurrent neural networks are often used for learning time-series data. ...
research
06/17/2018

Fast Convex Pruning of Deep Neural Networks

We develop a fast, tractable technique called Net-Trim for simplifying a...
research
11/04/2020

Kernel Dependence Network

We propose a greedy strategy to spectrally train a deep network for mult...

Please sign up or login with your details

Forgot password? Click here to reset