Recent advances in deep learning theory

12/20/2020
by   Fengxiang He, et al.
0

Deep learning is usually described as an experiment-driven field under continuous criticizes of lacking theoretical foundations. This problem has been partially fixed by a large volume of literature which has so far not been well organized. This paper reviews and organizes the recent advances in deep learning theory. The literature is categorized in six groups: (1) complexity and capacity-based approaches for analyzing the generalizability of deep learning; (2) stochastic differential equations and their dynamic systems for modelling stochastic gradient descent and its variants, which characterize the optimization and generalization of deep learning, partially inspired by Bayesian inference; (3) the geometrical structures of the loss landscape that drives the trajectories of the dynamic systems; (4) the roles of over-parameterization of deep neural networks from both positive and negative perspectives; (5) theoretical foundations of several special structures in network architectures; and (6) the increasingly intensive concerns in ethics and security and their relationships with generalizability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/16/2020

Hausdorff Dimension, Stochastic Differential Equations, and Generalization in Neural Networks

Despite its success in a wide range of applications, characterizing the ...
research
07/14/2021

Continuous vs. Discrete Optimization of Deep Neural Networks

Existing analyses of optimization in deep learning are either continuous...
research
12/15/2022

Neuroevolution Surpasses Stochastic Gradient Descent for Physics-Informed Neural Networks

The potential of learned models for fundamental scientific research and ...
research
12/01/2020

Deep dynamic modeling with just two time points: Can we still allow for individual trajectories?

Longitudinal biomedical data are often characterized by a sparse time gr...
research
04/10/2019

A Selective Overview of Deep Learning

Deep learning has arguably achieved tremendous success in recent years. ...
research
07/16/2020

Data-driven effective model shows a liquid-like deep learning

Geometric structure of an optimization landscape is argued to be fundame...
research
04/06/2021

Proof of the Theory-to-Practice Gap in Deep Learning via Sampling Complexity bounds for Neural Network Approximation Spaces

We study the computational complexity of (deterministic or randomized) a...

Please sign up or login with your details

Forgot password? Click here to reset