Distribution-Specific Hardness of Learning Neural Networks

09/05/2016
by   Ohad Shamir, et al.
0

Although neural networks are routinely and successfully trained in practice using simple gradient-based methods, most existing theoretical results are negative, showing that learning such networks is difficult, in a worst-case sense over all data distributions. In this paper, we take a more nuanced view, and consider whether specific assumptions on the "niceness" of the input distribution, or "niceness" of the target function (e.g. in terms of smoothness, non-degeneracy, incoherence, random choice of parameters etc.), are sufficient to guarantee learnability using gradient-based methods. We provide evidence that neither class of assumptions alone is sufficient: On the one hand, for any member of a class of "nice" target functions, there are difficult input distributions. On the other hand, we identify a family of simple target functions, which are difficult to learn even if the input distribution is "nice". To prove our results, we develop some tools which may be of independent interest, such as extending Fourier-based hardness techniques developed in the context of statistical queries blum1994weakly, from the Boolean cube to Euclidean space and to more general classes of functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/15/2023

Efficiently Learning Neural Networks: What Assumptions May Suffice?

Understanding when neural networks can be learned efficiently is a funda...
research
08/24/2021

The staircase property: How hierarchical structure can guide deep learning

This paper identifies a structural property of data distributions that e...
research
10/25/2019

Learning Boolean Circuits with Neural Networks

Training neural-networks is computationally hard. However, in practice t...
research
04/01/2019

On the Power and Limitations of Random Features for Understanding Neural Networks

Recently, a spate of papers have provided positive theoretical results f...
research
06/05/2020

Hardness of Learning Neural Networks with Natural Weights

Neural networks are nowadays highly successful despite strong hardness r...
research
06/22/2017

A Note on Learning Algorithms for Quadratic Assignment with Graph Neural Networks

Many inverse problems are formulated as optimization problems over certa...
research
06/21/2018

Learning Graph Weighted Models on Pictures

Graph Weighted Models (GWMs) have recently been proposed as a natural ge...

Please sign up or login with your details

Forgot password? Click here to reset