Loss Surface Modality of Feed-Forward Neural Network Architectures

05/24/2019
by   Anna Sergeevna Bosman, et al.
0

It has been argued in the past that high-dimensional neural networks do not exhibit local minima capable of trapping an optimisation algorithm. However, the relationship between loss surface modality and the neural architecture parameters, such as the number of hidden neurons per layer and the number of hidden layers, remains poorly understood. This study employs fitness landscape analysis to study the modality of neural network loss surfaces under various feed-forward architecture settings. An increase in the problem dimensionality is shown to yield a more searchable and more exploitable loss surface. An increase in the hidden layer width is shown to effectively reduce the number of local minima, and simplify the shape of the global attractor. An increase in the architecture depth is shown to sharpen the global attractor, thus making it more exploitable.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/08/2019

Visualising Basins of Attraction for the Cross-Entropy and the Squared Error Neural Network Loss Functions

Quantification of the stationary points and the associated basins of att...
research
02/28/2011

Improving the character recognition efficiency of feed forward BP neural network

This work is focused on improving the character recognition capability o...
research
06/02/2022

A Local Optima Network Analysis of the Feedforward Neural Architecture Space

This study investigates the use of local optima network (LON) analysis, ...
research
11/12/2022

On the High Symmetry of Neural Network Functions

Training neural networks means solving a high-dimensional optimization p...
research
06/11/2018

When and where do feed-forward neural networks learn localist representations?

According to parallel distributed processing (PDP) theory in psychology,...
research
02/10/2022

Exact Solutions of a Deep Linear Network

This work finds the exact solutions to a deep linear network with weight...
research
01/27/2023

Fine-tuning Neural-Operator architectures for training and generalization

In this work, we present an analysis of the generalization of Neural Ope...

Please sign up or login with your details

Forgot password? Click here to reset