The Efficacy of L_1 Regularization in Two-Layer Neural Networks

10/02/2020
by   Gen Li, et al.
0

A crucial problem in neural networks is to select the most appropriate number of hidden neurons and obtain tight statistical risk bounds. In this work, we present a new perspective towards the bias-variance tradeoff in neural networks. As an alternative to selecting the number of neurons, we theoretically show that L_1 regularization can control the generalization error and sparsify the input dimension. In particular, with an appropriate L_1 regularization on the output layer, the network can produce a statistical risk that is near minimax optimal. Moreover, an appropriate L_1 regularization on the input layer leads to a risk bound that does not involve the input data dimension. Our analysis is based on a new amalgamation of dimension-based and norm-based complexity analysis to bound the generalization error. A consequent observation from our results is that an excessively large number of neurons do not necessarily inflate generalization errors under a suitable regularization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/04/2022

On the Generalization Power of the Overfitted Three-Layer Neural Tangent Kernel Model

In this paper, we study the generalization performance of overparameteri...
research
10/06/2021

VC dimension of partially quantized neural networks in the overparametrized regime

Vapnik-Chervonenkis (VC) theory has so far been unable to explain the sm...
research
03/02/2021

Learning with Hyperspherical Uniformity

Due to the over-parameterization nature, neural networks are a powerful ...
research
09/04/2018

Understanding Regularization in Batch Normalization

Batch Normalization (BN) makes output of hidden neuron had zero mean and...
research
06/11/2020

Deep Learning Requires Explicit Regularization for Reliable Predictive Probability

From the statistical learning perspective, complexity control via explic...
research
09/04/2019

Empirical Hypothesis Space Reduction

Selecting appropriate regularization coefficients is critical to perform...
research
06/29/2020

A Bayesian regularization-backpropagation neural network model for peeling computations

Bayesian regularization-backpropagation neural network (BR-BPNN), a mach...

Please sign up or login with your details

Forgot password? Click here to reset