Neural network with optimal neuron activation functions based on additive Gaussian process regression

01/13/2023
by   Sergei Manzhos, et al.
0

Feed-forward neural networks (NN) are a staple machine learning method widely used in many areas of science and technology. While even a single-hidden layer NN is a universal approximator, its expressive power is limited by the use of simple neuron activation functions (such as sigmoid functions) that are typically the same for all neurons. More flexible neuron activation functions would allow using fewer neurons and layers and thereby save computational cost and improve expressive power. We show that additive Gaussian process regression (GPR) can be used to construct optimal neuron activation functions that are individual to each neuron. An approach is also introduced that avoids non-linear fitting of neural network parameters. The resulting method combines the advantage of robustness of a linear regression with the higher expressive power of a NN. We demonstrate the approach by fitting the potential energy surfaces of the water molecule and formaldehyde. Without requiring any non-linear optimization, the additive GPR based approach outperforms a conventional NN in the high accuracy regime, where a conventional NN suffers more from overfitting.

READ FULL TEXT

page 12

page 17

page 18

page 20

research
02/24/2017

Activation Ensembles for Deep Neural Networks

Many activation functions have been proposed in the past, but selecting ...
research
11/29/2017

Gaussian Process Neurons Learn Stochastic Activation Functions

We propose stochastic, non-parametric activation functions that are full...
research
08/08/2022

Neural Optimization Machine: A Neural Network Approach for Optimization

A novel neural network (NN) approach is proposed for constrained optimiz...
research
07/11/2023

Using Linear Regression for Iteratively Training Neural Networks

We present a simple linear regression based approach for learning the we...
research
06/10/2020

Scalable Partial Explainability in Neural Networks via Flexible Activation Functions

Achieving transparency in black-box deep learning algorithms is still an...
research
09/29/2012

Self-Delimiting Neural Networks

Self-delimiting (SLIM) programs are a central concept of theoretical com...
research
06/25/2021

Tensor-based framework for training flexible neural networks

Activation functions (AFs) are an important part of the design of neural...

Please sign up or login with your details

Forgot password? Click here to reset