Residual Memory Networks: Feed-forward approach to learn long temporal dependencies

08/06/2018
by   Murali Karthick Baskar, et al.
0

Training deep recurrent neural network (RNN) architectures is complicated due to the increased network complexity. This disrupts the learning of higher order abstracts using deep RNN. In case of feed-forward networks training deep structures is simple and faster while learning long-term temporal information is not possible. In this paper we propose a residual memory neural network (RMN) architecture to model short-time dependencies using deep feed-forward layers having residual and time delayed connections. The residual connection paves way to construct deeper networks by enabling unhindered flow of gradients and the time delay units capture temporal information with shared weights. The number of layers in RMN signifies both the hierarchical processing depth and temporal depth. The computational complexity in training RMN is significantly less when compared to deep recurrent networks. RMN is further extended as bi-directional RMN (BRMN) to capture both past and future information. Experimental analysis is done on AMI corpus to substantiate the capability of RMN in learning long-term information and hierarchical information. Recognition performance of RMN trained with 300 hours of Switchboard corpus is compared with various state-of-the-art LVCSR systems. The results indicate that RMN and BRMN gains 6

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/16/2016

Recurrent Dropout without Memory Loss

This paper presents a novel approach to recurrent neural network (RNN) r...
research
04/24/2017

k-FFNN: A priori knowledge infused Feed-forward Neural Networks

Recurrent neural network (RNN) are being extensively used over feed-forw...
research
07/18/2018

General Value Function Networks

In this paper we show that restricting the representation-layer of a Rec...
research
07/27/2023

Fading memory as inductive bias in residual recurrent networks

Residual connections have been proposed as architecture-based inductive ...
research
01/10/2017

Residual LSTM: Design of a Deep Recurrent Architecture for Distant Speech Recognition

In this paper, a novel architecture for a deep recurrent neural network,...
research
07/08/2018

Learning The Sequential Temporal Information with Recurrent Neural Networks

Recurrent Networks are one of the most powerful and promising artificial...
research
10/30/2021

Neural Network based on Automatic Differentiation Transformation of Numeric Iterate-to-Fixedpoint

This work proposes a Neural Network model that can control its depth usi...

Please sign up or login with your details

Forgot password? Click here to reset