Can stable and accurate neural networks be computed? – On the barriers of deep learning and Smale's 18th problem

01/20/2021
by   Vegard Antun, et al.
20

Deep learning (DL) has had unprecedented success and is now entering scientific computing with full force. However, DL suffers from a universal phenomenon: instability, despite universal approximating properties that often guarantee the existence of stable neural networks (NNs). We show the following paradox. There are basic well-conditioned problems in scientific computing where one can prove the existence of NNs with great approximation qualities, however, there does not exist any algorithm, even randomised, that can train (or compute) such a NN. Indeed, for any positive integers K > 2 and L, there are cases where simultaneously: (a) no randomised algorithm can compute a NN correct to K digits with probability greater than 1/2, (b) there exists a deterministic algorithm that computes a NN with K-1 correct digits, but any such (even randomised) algorithm needs arbitrarily many training data, (c) there exists a deterministic algorithm that computes a NN with K-2 correct digits using no more than L training samples. These results provide basic foundations for Smale's 18th problem and imply a potentially vast, and crucial, classification theory describing conditions under which (stable) NNs with a given accuracy can be computed by an algorithm. We begin this theory by initiating a unified theory for compressed sensing and DL, leading to sufficient conditions for the existence of algorithms that compute stable NNs in inverse problems. We introduce Fast Iterative REstarted NETworks (FIRENETs), which we prove and numerically verify are stable. Moreover, we prove that only 𝒪(|log(ϵ)|) layers are needed for an ϵ accurate solution to the inverse problem (exponential convergence), and that the inner dimensions in the layers do not exceed the dimension of the inverse problem. Thus, FIRENETs are computationally very efficient.

READ FULL TEXT

page 3

page 4

page 13

page 14

page 15

page 22

page 24

research
09/13/2021

The mathematics of adversarial attacks in AI – Why deep learning is unstable despite the existence of stable neural networks

The unprecedented success of deep learning (DL) makes it unchallenged wh...
research
03/02/2022

Stable, accurate and efficient deep neural networks for inverse problems with analysis-sparse models

Solving inverse problems is a fundamental component of science, engineer...
research
12/30/2021

Two Instances of Interpretable Neural Network for Universal Approximations

This paper proposes two bottom-up interpretable neural network (NN) cons...
research
01/05/2020

The troublesome kernel: why deep learning for inverse problems is typically unstable

There is overwhelming empirical evidence that Deep Learning (DL) leads t...
research
02/28/2022

Limitations of Deep Learning for Inverse Problems on Digital Hardware

Deep neural networks have seen tremendous success over the last years. S...
research
05/01/2023

Towards a Phenomenological Understanding of Neural Networks: Data

A theory of neural networks (NNs) built upon collective variables would ...
research
10/16/2018

From Deep to Physics-Informed Learning of Turbulence: Diagnostics

We describe physical tests validating progress made toward acceleration ...

Please sign up or login with your details

Forgot password? Click here to reset