Variants of RMSProp and Adagrad with Logarithmic Regret Bounds

06/17/2017
by   Mahesh Chandra Mukkamala, et al.
0

Adaptive gradient methods have become recently very popular, in particular as they have been shown to be useful in the training of deep neural networks. In this paper we have analyzed RMSProp, originally proposed for the training of deep neural networks, in the context of online convex optimization and show √(T)-type regret bounds. Moreover, we propose two variants SC-Adagrad and SC-RMSProp for which we show logarithmic regret bounds for strongly convex functions. Finally, we demonstrate in the experiments that these new variants outperform other adaptive gradient techniques or stochastic gradient descent in the optimization of strongly convex functions as well as in training of deep neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

Fast Rates for Online Gradient Descent Without Strong Convexity via Hoffman's Bound

Hoffman's classical result gives a bound on the distance of a point from...
research
05/08/2019

SAdam: A Variant of Adam for Strongly Convex Functions

The Adam algorithm has become extremely popular for large-scale machine ...
research
06/08/2015

Adaptive Normalized Risk-Averting Training For Deep Neural Networks

This paper proposes a set of new error criteria and learning approaches,...
research
09/04/2022

Dynamic Regret of Adaptive Gradient Methods for Strongly Convex Problems

Adaptive gradient algorithms such as ADAGRAD and its variants have gaine...
research
09/26/2019

GradVis: Visualization and Second Order Analysis of Optimization Surfaces during the Training of Deep Neural Networks

Current training methods for deep neural networks boil down to very high...
research
03/04/2019

Optimistic Adaptive Acceleration for Optimization

We consider a new variant of AMSGrad. AMSGrad RKK18 is a popular adaptiv...
research
05/21/2018

Never look back - A modified EnKF method and its application to the training of neural networks without back propagation

In this work, we present a new derivative-free optimization method and i...

Please sign up or login with your details

Forgot password? Click here to reset