A Use of Even Activation Functions in Neural Networks

11/23/2020
by   Fuchang Gao, et al.
0

Despite broad interest in applying deep learning techniques to scientific discovery, learning interpretable formulas that accurately describe scientific data is very challenging because of the vast landscape of possible functions and the "black box" nature of deep neural networks. The key to success is to effectively integrate existing knowledge or hypotheses about the underlying structure of the data into the architecture of deep learning models to guide machine learning. Currently, such integration is commonly done through customization of the loss functions. Here we propose an alternative approach to integrate existing knowledge or hypotheses of data structure by constructing custom activation functions that reflect this structure. Specifically, we study a common case when the multivariate target function f to be learned from the data is partially exchangeable, i.e. f(u,v,w)=f(v,u,w) for u,v∈ℝ^d. For instance, these conditions are satisfied for the classification of images that is invariant under left-right flipping. Through theoretical proof and experimental verification, we show that using an even activation function in one of the fully connected layers improves neural network performance. In our experimental 9-dimensional regression problems, replacing one of the non-symmetric activation functions with the designated "Seagull" activation function log(1+x^2) results in substantial improvement in network performance. Surprisingly, even activation functions are seldom used in neural networks. Our results suggest that customized activation functions have great potential in neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/16/2023

Data-aware customization of activation functions reduces neural network error

Activation functions play critical roles in neural networks, yet current...
research
01/13/2021

Reproducing Activation Function for Deep Learning

In this paper, we propose the reproducing activation function to improve...
research
11/12/2020

Empirical Performance Analysis of Conventional Deep Learning Models for Recognition of Objects in 2-D Images

Artificial Neural Networks, an essential part of Deep Learning, are deri...
research
09/07/2022

Parallel and Streaming Wavelet Neural Networks for Classification and Regression under Apache Spark

Wavelet neural networks (WNN) have been applied in many fields to solve ...
research
10/22/2019

Improving Siamese Networks for One Shot Learning using Kernel Based Activation functions

The lack of a large amount of training data has always been the constrai...
research
01/23/2023

Topological Understanding of Neural Networks, a survey

We look at the internal structure of neural networks which is usually tr...
research
08/02/2022

Lossy compression of multidimensional medical images using sinusoidal activation networks: an evaluation study

In this work, we evaluate how neural networks with periodic activation f...

Please sign up or login with your details

Forgot password? Click here to reset