Input Invex Neural Network

06/16/2021
by   Suman Sapkota, et al.
14

In this paper, we present a novel method to constrain invexity on Neural Networks (NN). Invex functions ensure every stationary point is global minima. Hence, gradient descent commenced from any point will lead to the global minima. Another advantage of invexity on NN is to divide data space locally into two connected sets with a highly non-linear decision boundary by simply thresholding the output. To this end, we formulate a universal invex function approximator and employ it to enforce invexity in NN. We call it Input Invex Neural Networks (II-NN). We first fit data with a known invex function, followed by modification with a NN, compare the direction of the gradient and penalize the direction of gradient on NN if it contradicts with the direction of reference invex function. In order to penalize the direction of the gradient we perform Gradient Clipped Gradient Penalty (GC-GP). We applied our method to the existing NNs for both image classification and regression tasks. From the extensive empirical and qualitative experiments, we observe that our method gives the performance similar to ordinary NN yet having invexity. Our method outperforms linear NN and Input Convex Neural Network (ICNN) with a large margin. We publish our code and implementation details at github.

READ FULL TEXT

page 3

page 12

page 13

page 14

page 17

page 20

research
10/19/2019

Neural Spectrum Alignment

Expressiveness of deep models was recently addressed via the connection ...
research
09/26/2017

Output Range Analysis for Deep Neural Networks

Deep neural networks (NN) are extensively used for machine learning task...
research
06/24/2019

The NN-Stacking: Feature weighted linear stacking through neural networks

Stacking methods improve the prediction performance of regression models...
research
11/04/2020

Which Minimizer Does My Neural Network Converge To?

The loss surface of an overparameterized neural network (NN) possesses m...
research
10/13/2022

Wasserstein Barycenter-based Model Fusion and Linear Mode Connectivity of Neural Networks

Based on the concepts of Wasserstein barycenter (WB) and Gromov-Wasserst...
research
05/22/2020

Premium Access to Convolutional Neural Networks

Neural Networks (NNs) are today used for all our daily tasks; for instan...
research
02/16/2021

Message Passing Descent for Efficient Machine Learning

We propose a new iterative optimization method for the Data-Fitting (DF)...

Please sign up or login with your details

Forgot password? Click here to reset