Visualizing the Loss Landscape of Neural Nets

12/28/2017
by   Hao Li, et al.
0

Neural network training relies on our ability to find "good" minimizers of highly non-convex loss functions. It is well known that certain network architecture designs (e.g., skip connections) produce loss functions that train easier, and well-chosen training parameters (batch size, learning rate, optimizer) produce minimizers that generalize better. However, the reasons for these differences, and their effect on the underlying loss landscape, is not well understood. In this paper, we explore the structure of neural loss functions, and the effect of loss landscapes on generalization, using a range of visualization methods. First, we introduce a simple "filter normalization" method that helps us visualize loss function curvature, and make meaningful side-by-side comparisons between loss functions. Then, using a variety of visualizations, we explore how network architecture affects the loss landscape, and how training parameters affect the shape of minimizers.

READ FULL TEXT
research
02/07/2021

Tilting the playing field: Dynamical loss functions for machine learning

We show that learning can be improved by using loss functions that evolv...
research
01/31/2021

Visualizing High-Dimensional Trajectories on the Loss-Landscape of ANNs

Training artificial neural networks requires the optimization of highly ...
research
09/30/2018

Nth Absolute Root Mean Error

Neural network training process takes long time when the size of trainin...
research
04/24/2022

The Multiscale Structure of Neural Network Loss Functions: The Effect on Optimization and Origin

Local quadratic approximation has been extensively used to study the opt...
research
04/14/2023

The R-mAtrIx Net

We provide a novel Neural Network architecture that can: i) output R-mat...
research
06/11/2019

Large Scale Structure of Neural Network Loss Landscapes

There are many surprising and perhaps counter-intuitive properties of op...
research
06/05/2019

A Tunable Loss Function for Classification

Recently, a parametrized class of loss functions called α-loss, α∈ [1,∞]...

Please sign up or login with your details

Forgot password? Click here to reset