Sharp asymptotics on the compression of two-layer neural networks

05/17/2022
by   Mohammad Hossein Amani, et al.
0

In this paper, we study the compression of a target two-layer neural network with N nodes into a compressed network with M < N nodes. More precisely, we consider the setting in which the weights of the target network are i.i.d. sub-Gaussian, and we minimize the population L2 loss between the outputs of the target and of the compressed network, under the assumption of Gaussian inputs. By using tools from high-dimensional probability, we show that this non-convex problem can be simplified when the target network is sufficiently over-parameterized, and provide the error rate of this approximation as a function of the input dimension and N . For a ReLU activation function, we conjecture that the optimum of the simplified optimization problem is achieved by taking weights on the Equiangular Tight Frame (ETF), while the scaling of the weights and the orientation of the ETF depend on the parameters of the target network. Numerical evidence is provided to support this conjecture.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2019

On the Approximation Properties of Neural Networks

We prove two new results concerning the approximation properties of neur...
research
09/27/2018

Complexity of Training ReLU Neural Network

In this paper, we explore some basic questions on the complexity of trai...
research
06/30/2023

Efficient uniform approximation using Random Vector Functional Link networks

A Random Vector Functional Link (RVFL) network is a depth-2 neural netwo...
research
02/01/2023

Deterministic equivalent and error universality of deep random features learning

This manuscript considers the problem of learning a random Gaussian netw...
research
02/20/2018

On the Connection Between Learning Two-Layers Neural Networks and Tensor Decomposition

We establish connections between the problem of learning a two-layers ne...
research
06/12/2022

Universality and approximation bounds for echo state networks with random weights

We study the uniform approximation of echo state networks with randomly ...
research
04/20/2018

A Simple Quantum Neural Net with a Periodic Activation Function

In this paper, we propose a simple neural net that requires only O(nlog_...

Please sign up or login with your details

Forgot password? Click here to reset