Enhancing approximation abilities of neural networks by training derivatives

12/12/2017
by   V. I. Avrutskiy, et al.
0

Method for increasing precision of feedforward networks is presented. With the aid of it they can serve as a better tool for describing smooth functions. Namely, it is shown that when training uses derivatives of target function up to the fourth order, approximation can be nearly machine precise. It is demonstrated in a number of cases: 2D function approximation, training autoencoder to compress 3D spiral into 1D, and solving 2D boundary value problem for Poisson equation with nonlinear source. In the first case cost function in addition to squared difference between output and target contains squared differences between their derivatives with respect to input variables. Training autoencoder is similar, but differentiation is done with respect to parameter that generates the spiral. Supplied with derivatives up to the fourth the method is found to be 30-200 times more accurate than regular training provided networks are of sufficient size and depth. Solving PDE is more practical since higher derivatives are not calculated beforehand, but information about them is extracted from the equation itself. Classical approach is to put perceptron in place of unknown function, choose the cost as squared residual and to minimize it with respect to weights. This would ensure that equation holds within some margin of error. Additional terms used in cost function are squared derivatives of the residual with respect to independent variables. Supplied with terms up to the second order the method is found to be 5 times more accurate. Efficient GPU version of algorithm is proposed.

READ FULL TEXT
research
02/28/2018

Avoiding overfitting of multilayer perceptrons by training derivatives

Resistance to overfitting is observed for neural networks trained with e...
research
12/14/2017

Neural networks catching up with finite differences in solving partial differential equations in higher dimensions

Fully connected multilayer perceptrons are used for obtaining numerical ...
research
12/12/2017

Backpropagation generalized for output derivatives

Backpropagation algorithm is the cornerstone for neural network analysis...
research
05/19/2022

Accelerated Training of Physics Informed Neural Networks (PINNs) using Meshless Discretizations

We present a new technique for the accelerated training of physics-infor...
research
07/26/2018

Advanced iterative procedures for solving the implicit Colebrook equation for fluid flow friction

Empirical Colebrook equation from 1939 is still accepted as an informal ...
research
01/01/2022

On automatic differentiation for the Matérn covariance

To target challenges in differentiable optimization we analyze and propo...
research
02/04/2023

Model-based Smoothing with Integrated Wiener Processes and Overlapping Splines

In many applications that involve the inference of an unknown smooth fun...

Please sign up or login with your details

Forgot password? Click here to reset