Complexity of Training ReLU Neural Network

09/27/2018
by   Digvijay Boob, et al.
0

In this paper, we explore some basic questions on the complexity of training Neural networks with ReLU activation function. We show that it is NP-hard to train a two- hidden layer feedforward ReLU neural network. If dimension d of the data is fixed then we show that there exists a polynomial time algorithm for the same training problem. We also show that if sufficient over-parameterization is provided in the first hidden layer of ReLU neural network then there is a polynomial time algorithm which finds weights such that output of the over-parameterized ReLU neural network matches with the output of the given data

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/29/2023

Training Neural Networks is NP-Hard in Fixed Dimension

We study the parameterized complexity of training two-layer neural netwo...
research
11/08/2021

Efficiently Learning Any One Hidden Layer ReLU Network From Queries

Model extraction attacks have renewed interest in the classic problem of...
research
02/19/2021

Training Neural Networks is ER-complete

Given a neural network, training data, and a threshold, it was known tha...
research
04/24/2021

Achieving Small Test Error in Mildly Overparameterized Neural Networks

Recent theoretical works on over-parameterized neural nets have focused ...
research
10/01/2022

A Combinatorial Perspective on the Optimization of Shallow ReLU Networks

The NP-hard problem of optimizing a shallow ReLU network can be characte...
research
05/20/2021

An Exact Poly-Time Membership-Queries Algorithm for Extraction a three-Layer ReLU Network

As machine learning increasingly becomes more prevalent in our everyday ...
research
05/17/2022

Sharp asymptotics on the compression of two-layer neural networks

In this paper, we study the compression of a target two-layer neural net...

Please sign up or login with your details

Forgot password? Click here to reset