Avoiding overfitting of multilayer perceptrons by training derivatives

02/28/2018
by   V. I. Avrutskiy, et al.
0

Resistance to overfitting is observed for neural networks trained with extended backpropagation algorithm. In addition to target values, its cost function uses derivatives of those up to the 4^th order. For common applications of neural networks, high order derivatives are not readily available, so simpler cases are considered: training network to approximate analytical function inside 2D and 5D domains and solving Poisson equation inside a 2D circle. For function approximation, the cost is a sum of squared differences between output and target as well as their derivatives with respect to the input. Differential equations are usually solved by putting a multilayer perceptron in place of unknown function and training its weights, so that equation holds within some margin of error. Commonly used cost is the equation's residual squared. Added terms are squared derivatives of said residual with respect to the independent variables. To investigate overfitting, the cost is minimized for points of regular grids with various spacing, and its root mean is compared with its value on much denser test set. Fully connected perceptrons with six hidden layers and 2·10^4, 1·10^6 and 5·10^6 weights in total are trained with Rprop until cost changes by less than 10 reached. Training the network with 5·10^6 weights to represent simple 2D function using 10 points with 8 extra derivatives in each produces cost test to train ratio of 1.5, whereas for classical backpropagation in comparable conditions this ratio is 2·10^4.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/12/2017

Enhancing approximation abilities of neural networks by training derivatives

Method for increasing precision of feedforward networks is presented. Wi...
research
12/14/2017

Neural networks catching up with finite differences in solving partial differential equations in higher dimensions

Fully connected multilayer perceptrons are used for obtaining numerical ...
research
12/12/2017

Backpropagation generalized for output derivatives

Backpropagation algorithm is the cornerstone for neural network analysis...
research
03/12/2015

Training Binary Multilayer Neural Networks for Image Classification using Expectation Backpropagation

Compared to Multilayer Neural Networks with real weights, Binary Multila...
research
07/21/2022

Sobolev Training for Implicit Neural Representations with Approximated Image Derivatives

Recently, Implicit Neural Representations (INRs) parameterized by neural...
research
12/26/2022

Quaternion Backpropagation

Quaternion valued neural networks experienced rising popularity and inte...
research
12/27/2022

The Forward-Forward Algorithm: Some Preliminary Investigations

The aim of this paper is to introduce a new learning procedure for neura...

Please sign up or login with your details

Forgot password? Click here to reset