Failures of Gradient-Based Deep Learning

03/23/2017
by   Shai Shalev-Shwartz, et al.
0

In recent years, Deep Learning has become the go-to solution for a broad range of applications, often outperforming state-of-the-art. However, it is important, for both theoreticians and practitioners, to gain a deeper understanding of the difficulties and limitations associated with common approaches and algorithms. We describe four types of simple problems, for which the gradient-based algorithms commonly used in deep learning either fail or suffer from significant difficulties. We illustrate the failures through practical experiments, and provide theoretical insights explaining their source, and how they might be remedied.

READ FULL TEXT
research
06/24/2012

Practical recommendations for gradient-based training of deep architectures

Learning algorithms related to artificial neural networks and in particu...
research
01/02/2020

Thresholds of descending algorithms in inference problems

We review recent works on analyzing the dynamics of gradient-based algor...
research
03/08/2019

Is Deeper Better only when Shallow is Good?

Understanding the power of depth in feed-forward neural networks is an o...
research
08/26/2022

On the Implicit Bias in Deep-Learning Algorithms

Gradient-based deep-learning algorithms exhibit remarkable performance i...
research
09/09/2023

A Gentle Introduction to Gradient-Based Optimization and Variational Inequalities for Machine Learning

The rapid progress in machine learning in recent years has been based on...
research
08/29/2023

Gradient-based methods for spiking physical systems

Recent efforts have fostered significant progress towards deep learning ...
research
10/26/2017

Maximum Principle Based Algorithms for Deep Learning

The continuous dynamical system approach to deep learning is explored in...

Please sign up or login with your details

Forgot password? Click here to reset