Convex Geometry and Duality of Over-parameterized Neural Networks

02/25/2020
by   Tolga Ergen, et al.
3

We develop a convex analytic framework for ReLU neural networks which elucidates the inner workings of hidden neurons and their function space characteristics. We show that neural networks with rectified linear units act as convex regularizers, where simple solutions are encouraged via extreme points of a certain convex set. For one dimensional regression and classification, as well as rank-one data matrices, we prove that finite two-layer ReLU networks with norm regularization yield linear spline interpolation. We characterize the classification decision regions in terms of a closed form kernel matrix and minimum L1 norm solutions. This is in contrast to Neural Tangent Kernel which is unable to explain neural network predictions with finitely many neurons. Our convex geometric description also provides intuitive explanations of hidden neurons as auto-encoders. In higher dimensions, we show that the training problem for two-layer networks can be cast as a convex optimization problem with infinitely many constraints. We then provide a family of convex relaxations to approximate the solution, and a cutting-plane algorithm to improve the relaxations. We derive conditions for the exactness of the relaxations and provide simple closed form formulas for the optimal neural network weights in certain cases. We also establish a connection to ℓ_0-ℓ_1 equivalence for neural networks analogous to the minimal cardinality solutions in compressed sensing. Extensive experimental results show that the proposed approach yields interpretable and accurate models.

READ FULL TEXT

page 3

page 6

page 7

page 8

page 15

page 18

page 19

page 23

research
02/22/2020

Convex Duality of Deep Neural Networks

We study regularized deep neural networks and introduce an analytic fram...
research
09/30/2022

Overparameterized ReLU Neural Networks Learn the Simplest Models: Neural Isometry and Exact Recovery

The practice of deep learning has shown that neural networks generalize ...
research
08/16/2022

Universal Solutions of Feedforward ReLU Networks for Interpolations

This paper provides a theoretical framework on the solution of feedforwa...
research
06/24/2020

Neural Splines: Fitting 3D Surfaces with Infinitely-Wide Neural Networks

We present Neural Splines, a technique for 3D surface reconstruction tha...
research
07/15/2020

Convexifying Sparse Interpolation with Infinitely Wide Neural Networks: An Atomic Norm Approach

This work examines the problem of exact data interpolation via sparse (n...
research
08/19/2022

Kernel Memory Networks: A Unifying Framework for Memory Modeling

We consider the problem of training a neural network to store a set of p...
research
02/01/2022

Deep Layer-wise Networks Have Closed-Form Weights

There is currently a debate within the neuroscience community over the l...

Please sign up or login with your details

Forgot password? Click here to reset