Numerical Exploration of Training Loss Level-Sets in Deep Neural Networks

11/09/2020
by   Naveed Tahir, et al.
10

We present a computational method for empirically characterizing the training loss level-sets of deep neural networks. Our method numerically constructs a path in parameter space that is constrained to a set with a fixed near-zero training loss. By measuring regularization functions and test loss at different points within this path, we examine how different points in the parameter space with the same fixed training loss compare in terms of generalization ability. We also compare this method for finding regularized points with the more typical method, that uses objective functions which are weighted sums of training loss and regularization terms. We apply dimensionality reduction to the traversed paths in order to visualize the loss level sets in a well-regularized region of parameter space. Our results provide new information about the loss landscape of deep neural networks, as well as a new strategy for reducing test loss.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2018

On the loss landscape of a class of deep neural networks with no bad local valleys

We identify a class of over-parameterized deep neural networks with stan...
research
06/20/2021

Better Training using Weight-Constrained Stochastic Dynamics

We employ constraints to control the parameter space of deep neural netw...
research
05/24/2022

Linear Connectivity Reveals Generalization Strategies

It is widely accepted in the mode connectivity literature that when two ...
research
03/20/2021

Train Deep Neural Networks in 40-D Subspaces

Although there are massive parameters in deep neural networks, the train...
research
03/08/2023

Loss-Curvature Matching for Dataset Selection and Condensation

Training neural networks on a large dataset requires substantial computa...
research
10/31/2019

Dynamic Regularizer with an Informative Prior

Regularization methods, specifically those which directly alter weights ...
research
03/04/2020

Rethinking Parameter Counting in Deep Models: Effective Dimensionality Revisited

Neural networks appear to have mysterious generalization properties when...

Please sign up or login with your details

Forgot password? Click here to reset