Deep linear neural networks with arbitrary loss: All local minima are global

12/05/2017
by   Thomas Laurent, et al.
0

We consider deep linear networks with arbitrary differentiable loss. We provide a short and elementary proof of the following fact: all local minima are global minima if each hidden layer is wider than either the input or output layer.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset