Are Efficient Deep Representations Learnable?

07/17/2018
by   Maxwell Nye, et al.
0

Many theories of deep learning have shown that a deep network can require dramatically fewer resources to represent a given function compared to a shallow network. But a question remains: can these efficient representations be learned using current deep learning techniques? In this work, we test whether standard deep learning methods can in fact find the efficient representations posited by several theories of deep representation. Specifically, we train deep neural networks to learn two simple functions with known efficient solutions: the parity function and the fast Fourier transform. We find that using gradient-based optimization, a deep network does not learn the parity function, unless initialized very close to a hand-coded exact solution. We also find that a deep linear neural network does not learn the fast Fourier transform, even in the best-case scenario of infinite training data, unless the weights are initialized very close to the exact hand-coded solution. Our results suggest that not every element of the class of compositional functions can be learned efficiently by a deep network, and further restrictions are necessary to understand what functions are both efficiently representable and learnable.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/31/2021

The Connection Between Approximation, Depth Separation and Learnability in Neural Networks

Several recent works have shown separation results between deep neural n...
research
06/10/2022

Training Neural Networks using SAT solvers

We propose an algorithm to explore the global optimization method, using...
research
02/01/2023

Multi-Grade Deep Learning

The current deep learning model is of a single-grade, that is, it learns...
research
12/16/2018

Provable limitations of deep learning

As the success of deep learning reaches more grounds, one would like to ...
research
08/24/2021

The staircase property: How hierarchical structure can guide deep learning

This paper identifies a structural property of data distributions that e...
research
01/04/2023

Learning Decorrelated Representations Efficiently Using Fast Fourier Transform

Barlow Twins and VICReg are self-supervised representation learning mode...
research
06/18/2021

An Empirical Investigation into Deep and Shallow Rule Learning

Inductive rule learning is arguably among the most traditional paradigms...

Please sign up or login with your details

Forgot password? Click here to reset