Fenchel Lifted Networks: A Lagrange Relaxation of Neural Network Training

11/20/2018
by   Fangda Gu, et al.
0

Despite the recent successes of deep neural networks, the corresponding training problem remains highly non-convex and difficult to optimize. Classes of models have been proposed that introduce greater structure to the objective function at the cost of lifting the dimension of the problem. However, these lifted methods sometimes perform poorly compared to traditional neural networks. In this paper, we introduce a new class of lifted models, Fenchel lifted networks, that enjoy the same benefits as previous lifted models, without suffering a degradation in performance over classical networks. Our model represents activation functions as equivalent biconvex constraints and uses Lagrange Multipliers to arrive at a rigorous lower bound of the traditional neural network training problem. This model is efficiently trained using block-coordinate descent and is parallelizable across data points and/or layers. We compare our model against standard fully connected and convolutional networks and show that we are able to match or beat their performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/22/2023

Fixed points of arbitrarily deep 1-dimensional neural networks

In this paper, we introduce a new class of functions on ℝ that is closed...
research
05/03/2018

Lifted Neural Networks

We describe a novel family of models of multi- layer feedforward neural ...
research
06/09/2019

Stochastic In-Face Frank-Wolfe Methods for Non-Convex Optimization and Sparse Neural Network Training

The Frank-Wolfe method and its extensions are well-suited for delivering...
research
10/12/2020

Activation function impact on Sparse Neural Networks

While the concept of a Sparse Neural Network has been researched for som...
research
02/11/2018

Optimizing Neural Networks in the Equivalent Class Space

It has been widely observed that many activation functions and pooling m...
research
02/17/2022

Training neural networks using monotone variational inequality

Despite the vast empirical success of neural networks, theoretical under...
research
06/09/2020

The Curious Case of Convex Networks

In this paper, we investigate a constrained formulation of neural networ...

Please sign up or login with your details

Forgot password? Click here to reset