Convex Duality of Deep Neural Networks

02/22/2020
by   Tolga Ergen, et al.
23

We study regularized deep neural networks and introduce an analytic framework to characterize the structure of the hidden layers. We show that a set of optimal hidden layer weight matrices for a norm regularized deep neural network training problem can be explicitly found as the extreme points of a convex set. For two-layer linear networks, we first formulate a convex dual program and prove that strong duality holds. We then extend our derivations to prove that strong duality also holds for certain deep networks. In particular, for linear deep networks, we show that each optimal layer weight matrix is rank-one and aligns with the previous layers when the network output is scalar. We also extend our analysis to the vector outputs and other convex loss functions. More importantly, we show that the same characterization can also be applied to deep ReLU networks with rank-one inputs, where we prove that strong duality still holds and optimal layer weight matrices are rank-one for scalar output networks. As a corollary, we prove that norm regularized deep ReLU networks yield spline interpolation for one-dimensional datasets which was previously known only for two-layer networks. We then verify our theoretical results via several numerical experiments.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/13/2021

Parallel Deep Neural Networks Have Zero Duality Gap

Training deep neural networks is a well-known highly non-convex problem....
research
02/25/2020

Convex Geometry and Duality of Over-parameterized Neural Networks

We develop a convex analytic framework for ReLU neural networks which el...
research
05/18/2018

Reconstruction of training samples from loss functions

This paper presents a new mathematical framework to analyze the loss fun...
research
02/09/2022

A Local Geometric Interpretation of Feature Extraction in Deep Feedforward Neural Networks

In this paper, we present a local geometric analysis to interpret how de...
research
04/14/2023

Convex Dual Theory Analysis of Two-Layer Convolutional Neural Networks with Soft-Thresholding

Soft-thresholding has been widely used in neural networks. Its basic net...
research
08/16/2022

Universal Solutions of Feedforward ReLU Networks for Interpolations

This paper provides a theoretical framework on the solution of feedforwa...
research
07/18/2023

Convex Geometry of ReLU-layers, Injectivity on the Ball and Local Reconstruction

The paper uses a frame-theoretic setting to study the injectivity of a R...

Please sign up or login with your details

Forgot password? Click here to reset