The Effects of Hyperparameters on SGD Training of Neural Networks

08/12/2015
by   Thomas M. Breuel, et al.
0

The performance of neural network classifiers is determined by a number of hyperparameters, including learning rate, batch size, and depth. A number of attempts have been made to explore these parameters in the literature, and at times, to develop methods for optimizing them. However, exploration of parameter spaces has often been limited. In this note, I report the results of large scale experiments exploring these different parameters and their interactions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/20/2020

How do SGD hyperparameters in natural training affect adversarial robustness?

Learning rate, batch size and momentum are three important hyperparamete...
research
11/13/2017

Three Factors Influencing Minima in SGD

We study the properties of the endpoint of stochastic gradient descent (...
research
05/26/2023

Rotational Optimizers: Simple Robust DNN Training

The training dynamics of modern deep neural networks depend on complex i...
research
01/12/2019

Recombination of Artificial Neural Networks

We propose a genetic algorithm (GA) for hyperparameter optimization of a...
research
08/03/2022

Empirical Study of Overfitting in Deep FNN Prediction Models for Breast Cancer Metastasis

Overfitting is defined as the fact that the current model fits a specifi...
research
04/21/2020

Hyperparameter Optimization in Binary Communication Networks for Neuromorphic Deployment

Training neural networks for neuromorphic deployment is non-trivial. The...
research
08/17/2019

A Sensitivity Analysis of Attention-Gated Convolutional Neural Networks for Sentence Classification

Recently, Attention-Gated Convolutional Neural Networks (AGCNNs) perform...

Please sign up or login with your details

Forgot password? Click here to reset