Consistency of Neural Networks with Regularization

06/22/2022
by   Xiaoxi Shen, et al.
0

Neural networks have attracted a lot of attention due to its success in applications such as natural language processing and computer vision. For large scale data, due to the tremendous number of parameters in neural networks, overfitting is an issue in training neural networks. To avoid overfitting, one common approach is to penalize the parameters especially the weights in neural networks. Although neural networks has demonstrated its advantages in many applications, the theoretical foundation of penalized neural networks has not been well-established. Our goal of this paper is to propose the general framework of neural networks with regularization and prove its consistency. Under certain conditions, the estimated neural network will converge to true underlying function as the sample size increases. The method of sieves and the theory on minimal neural networks are used to overcome the issue of unidentifiability for the parameters. Two types of activation functions: hyperbolic tangent function(Tanh) and rectified linear unit(ReLU) have been taken into consideration. Simulations have been conducted to verify the validation of theorem of consistency.

READ FULL TEXT
research
05/13/2022

Convergence of Deep Neural Networks with General Activation Functions and Pooling

Deep neural networks, as a powerful system to represent high dimensional...
research
12/10/2020

The Representation Power of Neural Networks: Breaking the Curse of Dimensionality

In this paper, we analyze the number of neurons and training parameters ...
research
10/05/2014

On the Computational Efficiency of Training Neural Networks

It is well-known that neural networks are computationally hard to train....
research
10/31/2020

Asymptotic Theory of Expectile Neural Networks

Neural networks are becoming an increasingly important tool in applicati...
research
03/07/2020

AL2: Progressive Activation Loss for Learning General Representations in Classification Neural Networks

The large capacity of neural networks enables them to learn complex func...
research
01/07/2023

ExcelFormer: A Neural Network Surpassing GBDTs on Tabular Data

Though neural networks have achieved enormous breakthroughs on various f...
research
06/07/2013

Non-constant bounded holomorphic functions of hyperbolic numbers - Candidates for hyperbolic activation functions

The Liouville theorem states that bounded holomorphic complex functions ...

Please sign up or login with your details

Forgot password? Click here to reset