Conjugate-gradient-based Adam for stochastic optimization and its application to deep learning

02/29/2020
by   Yu Kobayashi, et al.
0

This paper proposes a conjugate-gradient-based Adam algorithm blending Adam with nonlinear conjugate gradient methods and shows its convergence analysis. Numerical experiments on text classification and image classification show that the proposed algorithm can train deep neural network models in fewer epochs than the existing adaptive stochastic optimization algorithms can.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/12/2018

Maximizing Invariant Data Perturbation with Stochastic Optimization

Feature attribution methods, or saliency maps, are one of the most popul...
research
07/27/2020

Binary Search and First Order Gradient Based Method for Stochastic Optimization

In this paper, we present a novel stochastic optimization method, which ...
research
06/15/2016

Optimization Methods for Large-Scale Machine Learning

This paper provides a review and commentary on the past, present, and fu...
research
10/20/2020

Multi-objective free-form shape optimization of a synchronous reluctance machine

This paper deals with the design optimization of a synchronous reluctanc...
research
01/26/2021

Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization

We introduce MADGRAD, a novel optimization method in the family of AdaGr...
research
04/14/2020

Stochastic batch size for adaptive regularization in deep network optimization

We propose a first-order stochastic optimization algorithm incorporating...
research
01/19/2015

Microscopic Advances with Large-Scale Learning: Stochastic Optimization for Cryo-EM

Determining the 3D structures of biological molecules is a key problem f...

Please sign up or login with your details

Forgot password? Click here to reset