Tune smarter not harder: A principled approach to tuning learning rates for shallow nets

03/22/2020
by   Thulasi Tholeti, et al.
0

Effective hyper-parameter tuning is essential to guarantee the performance that neural networks have come to be known for. In this work, a principled approach to choosing the learning rate is proposed for shallow feedforward neural networks. We associate the learning rate with the gradient Lipschitz constant of the objective to be minimized while training. An upper bound on the mentioned constant is derived and a search algorithm, which always results in non-divergent traces, is proposed to exploit the derived bound. It is shown through simulations that the proposed search method significantly outperforms the existing tuning methods such as Tree Parzen Estimators (TPE). The proposed method is applied to two different existing applications, namely, channel estimation in a wireless communication system and prediction of the exchange currency rates, and it is shown to pick better learning rates than the existing methods using the same or lesser compute power.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/18/2020

Parsimonious Computing: A Minority Training Regime for Effective Prediction in Large Microarray Expression Data Sets

Rigorous mathematical investigation of learning rates used in back-propa...
research
06/03/2015

Cyclical Learning Rates for Training Neural Networks

It is known that the learning rate is the most important hyper-parameter...
research
08/18/2019

Demystifying Learning Rate Polices for High Accuracy Training of Deep Neural Networks

Learning Rate (LR) is an important hyper-parameter to tune for effective...
research
05/06/2023

Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization

Prompt tuning is one of the successful approaches for parameter-efficien...
research
11/22/2021

Towards a Principled Learning Rate Adaptation for Natural Evolution Strategies

Natural Evolution Strategies (NES) is a promising framework for black-bo...
research
01/31/2023

Mind the (optimality) Gap: A Gap-Aware Learning Rate Scheduler for Adversarial Nets

Adversarial nets have proved to be powerful in various domains including...

Please sign up or login with your details

Forgot password? Click here to reset