Population-Based Training for Loss Function Optimization

02/11/2020
by   Jason Liang, et al.
0

Metalearning of deep neural network (DNN) architectures and hyperparameters has become an increasingly important area of research. Loss functions are a type of metaknowledge that is crucial to effective training of DNNs and their potential role in metalearning has not yet been fully explored. This paper presents an algorithm called Enhanced Population-Based Training (EPBT) that interleaves the training of a DNN's weights with the metalearning of optimal hyperparameters and loss functions. Loss functions use a TaylorGLO parameterization, based on multivariate Taylor expansions, that EPBT can directly optimize. On the CIFAR-10 and SVHN image classification benchmarks, EPBT discovers loss function schedules that enable faster, more accurate learning. The discovered functions adapt to the training process and serve to regularize the learning task by discouraging overfitting to the labels. EPBT thus demonstrates a promising synergy of simultaneous training and metalearning.

READ FULL TEXT
research
05/27/2019

Improved Training Speed, Accuracy, and Data Utilization Through Loss Function Optimization

As the complexity of neural network models has grown, it has become incr...
research
06/28/2019

Learning Effective Loss Functions Efficiently

We consider the problem of learning a loss function which, when minimize...
research
02/06/2019

The role of a layer in deep neural networks: a Gaussian Process perspective

A fundamental question in deep learning concerns the role played by indi...
research
07/24/2019

Learning Embedding of 3D models with Quadric Loss

Sharp features such as edges and corners play an important role in the p...
research
06/21/2020

Applications of Koopman Mode Analysis to Neural Networks

We consider the training process of a neural network as a dynamical syst...
research
08/01/2022

Backdoor Watermarking Deep Learning Classification Models With Deep Fidelity

Backdoor Watermarking is a promising paradigm to protect the copyright o...
research
04/03/2020

Gradient Centralization: A New Optimization Technique for Deep Neural Networks

Optimization techniques are of great importance to effectively and effic...

Please sign up or login with your details

Forgot password? Click here to reset