Generalization and Expressivity for Deep Nets

03/10/2018
by   Shao-Bo Lin, et al.
0

Along with the rapid development of deep learning in practice, the theoretical explanations for its success become urgent. Generalization and expressivity are two widely used measurements to quantify theoretical behaviors of deep learning. The expressivity focuses on finding functions expressible by deep nets but cannot be approximated by shallow nets with the similar number of neurons. It usually implies the large capacity. The generalization aims at deriving fast learning rate for deep nets. It usually requires small capacity to reduce the variance. Different from previous studies on deep learning, pursuing either expressivity or generalization, we take both factors into account to explore the theoretical advantages of deep nets. For this purpose, we construct a deep net with two hidden layers possessing excellent expressivity in terms of localized and sparse approximation. Then, utilizing the well known covering number to measure the capacity, we find that deep nets possess excellent expressive power (measured by localized and sparse approximation) without enlarging the capacity of shallow nets. As a consequence, we derive near optimal learning rates for implementing empirical risk minimization (ERM) on the constructed deep nets. These results theoretically exhibit the advantage of deep nets from learning theory viewpoints.

READ FULL TEXT
research
01/01/2019

Realizing data features by deep nets

This paper considers the power of deep neural networks (deep nets for sh...
research
01/13/2020

Approximation smooth and sparse functions by deep neural networks without saturation

Constructing neural networks for function approximation is a classical a...
research
03/09/2018

Construction of neural networks for realization of localized deep learning

The subject of deep learning has recently attracted users of machine lea...
research
04/03/2019

Deep Neural Networks for Rotation-Invariance Approximation and Learning

Based on the tree architecture, the objective of this paper is to design...
research
12/16/2019

Realization of spatial sparseness by deep ReLU nets with massive data

The great success of deep learning poses urgent challenges for understan...
research
06/02/2022

Deep Learning on Implicit Neural Datasets

Implicit neural representations (INRs) have become fast, lightweight too...
research
09/29/2020

Deep discriminant analysis for task-dependent compact network search

Most of today's popular deep architectures are hand-engineered for gener...

Please sign up or login with your details

Forgot password? Click here to reset