Sensitivity – Local Index to Control Chaoticity or Gradient Globally

12/24/2020
by   Katsunari Shibata, et al.
19

In this paper, we propose a fully local index named "sensitivity" for each neuron to control chaoticity or gradient globally in a neural network (NN), and also propose a learning method to adjust it named "sensitivity adjustment learning (SAL)". The index is the gradient magnitude of its output with respect to its inputs. By adjusting it around 1.0, information transmission in the neuron changes to moderate without shrinking or expanding for both forward and backward computations, and the information transmission through a layer of neurons also moderate when the weights and inputs are random. Therefore, it can be used in a recurrent NN (RNN) to control chaoticity of its global network dynamics, and also can be used to solve the vanishing gradient problem in error back propagation (BP) learning in a deep feedforward NN (DFNN) or an RNN with long-term dependency. We demonstrated that when SAL is applied to an RNN with small random weights, the sum of log-sensitivities is almost equivalent to the maximum Lyapunov exponent until it reaches 0.0 regardless of the network architecture. We also show that SAL works with BP or BPTT to avoid the vanishing gradient problem in a 300-layer DFNN or an RNN solving a problem with 300-step lag between the first input and the output. Compared with the fine manual tuning of the spectral radius of weight matrix before learning, the learning performance was quite better due to the continuous nonlinear learning nature of SAL, which prevented the loss of sensitivity.

READ FULL TEXT

page 8

page 10

page 11

page 12

page 13

page 18

page 19

page 20

research
10/03/2022

Random orthogonal additive filters: a solution to the vanishing/exploding gradient of deep neural networks

Since the recognition in the early nineties of the vanishing/exploding (...
research
02/07/2021

SeReNe: Sensitivity based Regularization of Neurons for Structured Sparsity in Neural Networks

Deep neural networks include millions of learnable parameters, making th...
research
09/14/2021

Oscillatory Fourier Neural Network: A Compact and Efficient Architecture for Sequential Processing

Tremendous progress has been made in sequential processing with the rece...
research
03/30/2023

Optimal Input Gain: All You Need to Supercharge a Feed-Forward Neural Network

Linear transformation of the inputs alters the training performance of f...
research
04/21/2014

Influence of the learning method in the performance of feedforward neural networks when the activity of neurons is modified

A method that allows us to give a different treatment to any neuron insi...
research
06/10/2020

ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach

It is hard to train Recurrent Neural Network (RNN) with stable convergen...
research
11/17/2021

Random Graph-Based Neuromorphic Learning with a Layer-Weaken Structure

Unified understanding of neuro networks (NNs) gets the users into great ...

Please sign up or login with your details

Forgot password? Click here to reset