Limitations of Lazy Training of Two-layers Neural Networks

06/21/2019
by   Behrooz Ghorbani, et al.
0

We study the supervised learning problem under either of the following two models: (1) Feature vectors x_i are d-dimensional Gaussians and responses are y_i = f_*( x_i) for f_* an unknown quadratic function; (2) Feature vectors x_i are distributed as a mixture of two d-dimensional centered Gaussians, and y_i's are the corresponding class labels. We use two-layers neural networks with quadratic activations, and compare three different learning regimes: the random features (RF) regime in which we only train the second-layer weights; the neural tangent (NT) regime in which we train a linearization of the neural network around its initialization; the fully trained neural network (NN) regime in which we train all the weights in the network. We prove that, even for the simple quadratic model of point (1), there is a potentially unbounded gap between the prediction risk achieved in these three training regimes, when the number of neurons is smaller than the ambient dimension. When the number of neurons is larger than the number of dimensions, the problem is significantly easier and both NT and NN learning achieve zero risk.

READ FULL TEXT
research
09/26/2019

Mildly Overparametrized Neural Nets can Memorize Training Data Efficiently

It has been observed zhang2016understanding that deep neural networks ca...
research
02/17/2022

Universality of empirical risk minimization

Consider supervised learning from i.i.d. samples { x_i,y_i}_i≤ n where x...
research
07/25/2020

The Interpolation Phase Transition in Neural Networks: Memorization and Generalization under Lazy Training

Modern neural networks are often operated in a strongly overparametrized...
research
10/26/2020

The estimation of training accuracy for two-layer neural networks on random datasets without training

Although the neural network (NN) technique plays an important role in ma...
research
11/02/2016

Natural-Parameter Networks: A Class of Probabilistic Neural Networks

Neural networks (NN) have achieved state-of-the-art performance in vario...
research
03/22/2022

On the (Non-)Robustness of Two-Layer Neural Networks in Different Learning Regimes

Neural networks are known to be highly sensitive to adversarial examples...
research
04/27/2019

Linearized two-layers neural networks in high dimension

We consider the problem of learning an unknown function f_ on the d-dime...

Please sign up or login with your details

Forgot password? Click here to reset