Tangent-Space Gradient Optimization of Tensor Network for Machine Learning

01/10/2020
by   Zheng-Zhi Sun, et al.
0

The gradient-based optimization method for deep machine learning models suffers from gradient vanishing and exploding problems, particularly when the computational graph becomes deep. In this work, we propose the tangent-space gradient optimization (TSGO) for the probabilistic models to keep the gradients from vanishing or exploding. The central idea is to guarantee the orthogonality between the variational parameters and the gradients. The optimization is then implemented by rotating parameter vector towards the direction of gradient. We explain and testify TSGO in tensor network (TN) machine learning, where the TN describes the joint probability distribution as a normalized state | ψ〉 in Hilbert space. We show that the gradient can be restricted in the tangent space of 〈ψ.| ψ〉 = 1 hyper-sphere. Instead of additional adaptive methods to control the learning rate in deep learning, the learning rate of TSGO is naturally determined by the angle θ as η = tanθ. Our numerical results reveal better convergence of TSGO in comparison to the off-the-shelf Adam.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2023

Angle based dynamic learning rate for gradient descent

In our work, we propose a novel yet simple approach to obtain an adaptiv...
research
04/07/2020

Automatic, Dynamic, and Nearly Optimal Learning Rate Specification by Local Quadratic Approximation

In deep learning tasks, the learning rate determines the update step siz...
research
09/12/2023

ELRA: Exponential learning rate adaption gradient descent optimization method

We present a novel, fast (exponential rate adaption), ab initio (hyper-p...
research
03/03/2022

Improvements to Gradient Descent Methods for Quantum Tensor Network Machine Learning

Tensor networks have demonstrated significant value for machine learning...
research
11/22/2021

Towards a Principled Learning Rate Adaptation for Natural Evolution Strategies

Natural Evolution Strategies (NES) is a promising framework for black-bo...
research
12/30/2019

Bayesian Tensor Network and Optimization Algorithm for Probabilistic Machine Learning

Describing or calculating the conditional probabilities of multiple even...
research
06/05/2018

On layer-level control of DNN training and its impact on generalization

The generalization ability of a neural network depends on the optimizati...

Please sign up or login with your details

Forgot password? Click here to reset