Neurashed: A Phenomenological Model for Imitating Deep Learning Training

12/17/2021
by   Weijie J. Su, et al.
0

To advance deep learning methodologies in the next decade, a theoretical framework for reasoning about modern neural networks is needed. While efforts are increasing toward demystifying why deep learning is so effective, a comprehensive picture remains lacking, suggesting that a better theory is possible. We argue that a future deep learning theory should inherit three characteristics: a hierarchically structured network architecture, parameters iteratively optimized using stochastic gradient-based methods, and information from the data that evolves compressively. As an instantiation, we integrate these characteristics into a graphical model called neurashed. This model effectively explains some common empirical patterns in deep learning. In particular, neurashed enables insights into implicit regularization, information bottleneck, and local elasticity. Finally, we discuss how neurashed can guide the development of deep learning theories.

READ FULL TEXT
research
05/07/2021

A Critical Review of Information Bottleneck Theory and its Applications to Deep Learning

In the past decade, deep neural networks have seen unparalleled improvem...
research
08/29/2023

Gradient-based methods for spiking physical systems

Recent efforts have fostered significant progress towards deep learning ...
research
05/17/2019

An Essay on Optimization Mystery of Deep Learning

Despite the huge empirical success of deep learning, theoretical underst...
research
04/10/2019

A Selective Overview of Deep Learning

Deep learning has arguably achieved tremendous success in recent years. ...
research
06/12/2019

Is Deep Learning an RG Flow?

Although there has been a rapid development of practical applications, t...
research
02/12/2020

The Unreasonable Effectiveness of Deep Learning in Artificial Intelligence

Deep learning networks have been trained to recognize speech, caption ph...
research
10/03/2018

Theory of Generative Deep Learning : Probe Landscape of Empirical Error via Norm Based Capacity Control

Despite its remarkable empirical success as a highly competitive branch ...

Please sign up or login with your details

Forgot password? Click here to reset