CDiNN -Convex Difference Neural Networks

Neural networks with ReLU activation function have been shown to be universal function approximators and learn function mapping as non-smooth functions. Recently, there is considerable interest in the use of neural networks in applications such as optimal control. It is well-known that optimization involving non-convex, non-smooth functions are computationally intensive and have limited convergence guarantees. Moreover, the choice of optimization hyper-parameters used in gradient descent/ascent significantly affect the quality of the obtained solutions. A new neural network architecture called the Input Convex Neural Networks (ICNNs) learn the output as a convex function of inputs thereby allowing the use of efficient convex optimization methods. Use of ICNNs for determining the input for minimizing output has two major problems: learning of a non-convex function as a convex mapping could result in significant function approximation error, and we also note that the existing representations cannot capture simple dynamic structures like linear time delay systems. We attempt to address the above problems by introduction of a new neural network architecture, which we call the CDiNN, which learns the function as a difference of polyhedral convex functions from data. We also discuss that, in some cases, the optimal input can be obtained from CDiNN through difference of convex optimization with convergence guarantees and that at each iteration, the problem is reduced to a linear programming problem.

READ FULL TEXT

page 14

page 16

page 17

page 18

research
10/15/2021

Gradient Descent on Infinitely Wide Neural Networks: Global Convergence and Generalization

Many supervised machine learning methods are naturally cast as optimizat...
research
12/02/2022

On Solution Functions of Optimization: Universal Approximation and Covering Number Bounds

We study the expressibility and learnability of convex optimization solu...
research
05/21/2019

A Universal Approximation Result for Difference of log-sum-exp Neural Networks

We show that a neural network whose output is obtained as the difference...
research
05/23/2019

How degenerate is the parametrization of neural networks with the ReLU activation function?

Neural network training is usually accomplished by solving a non-convex ...
research
06/26/2023

GloptiNets: Scalable Non-Convex Optimization with Certificates

We present a novel approach to non-convex optimization with certificates...
research
05/13/2023

Successive Affine Learning for Deep Neural Networks

This paper introduces a successive affine learning (SAL) model for const...
research
02/27/2018

Identification of LTV Dynamical Models with Smooth or Discontinuous Time Evolution by means of Convex Optimization

We establish a connection between trend filtering and system identificat...

Please sign up or login with your details

Forgot password? Click here to reset