Implicit Bias of Gradient Descent on Linear Convolutional Networks

06/01/2018
by   Suriya Gunasekar, et al.
0

We show that gradient descent on full-width linear convolutional networks of depth L converges to a linear predictor related to the ℓ_2/L bridge penalty in the frequency domain. This is in contrast to linearly fully connected networks, where gradient descent converges to the hard margin linear support vector machine solution, regardless of depth.

READ FULL TEXT
research
02/16/2016

Gradient Descent Converges to Minimizers

We show that gradient descent converges to a local minimizer, almost sur...
research
10/03/2020

Computational Separation Between Convolutional and Fully-Connected Networks

Convolutional neural networks (CNN) exhibit unmatched performance in a m...
research
10/06/2020

A Unifying View on Implicit Bias in Training Linear Neural Networks

We study the implicit bias of gradient flow (i.e., gradient descent with...
research
07/03/2018

On the Computational Power of Online Gradient Descent

We prove that the evolution of weight vectors in online gradient descent...
research
10/08/2021

On the Implicit Biases of Architecture Gradient Descent

Do neural networks generalise because of bias in the functions returned ...
research
08/03/2021

Geometry of Linear Convolutional Networks

We study the family of functions that are represented by a linear convol...
research
03/13/2020

Balancedness and Alignment are Unlikely in Linear Neural Networks

We study the invariance properties of alignment in linear neural network...

Please sign up or login with your details

Forgot password? Click here to reset