An Improved Residual LSTM Architecture for Acoustic Modeling

08/17/2017
by   Lu Huang, et al.
0

Long Short-Term Memory (LSTM) is the primary recurrent neural networks architecture for acoustic modeling in automatic speech recognition systems. Residual learning is an efficient method to help neural networks converge easier and faster. In this paper, we propose several types of residual LSTM methods for our acoustic modeling. Our experiments indicate that, compared with classic LSTM, our architecture shows more than 8 Error Rate (PER) on TIMIT tasks. At the same time, our residual fast LSTM approach shows 4 that all this architecture could have good results on THCHS-30, Librispeech and Switchboard corpora.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/05/2014

Long Short-Term Memory Based Recurrent Neural Network Architectures for Large Vocabulary Speech Recognition

Long Short-Term Memory (LSTM) is a recurrent neural network (RNN) archit...
research
11/21/2017

Deep Long Short-Term Memory Adaptive Beamforming Networks For Multichannel Robust Speech Recognition

Far-field speech recognition in noisy and reverberant conditions remains...
research
01/10/2017

Residual LSTM: Design of a Deep Recurrent Architecture for Distant Speech Recognition

In this paper, a novel architecture for a deep recurrent neural network,...
research
03/13/2015

LSTM: A Search Space Odyssey

Several variants of the Long Short-Term Memory (LSTM) architecture for r...
research
06/13/2017

Modelling prosodic structure using Artificial Neural Networks

The ability to accurately perceive whether a speaker is asking a questio...
research
02/13/2019

Modeling default rate in P2P lending via LSTM

With the fast development of peer to peer (P2P) lending, financial insti...
research
01/30/2019

Hardware-Guided Symbiotic Training for Compact, Accurate, yet Execution-Efficient LSTM

Many long short-term memory (LSTM) applications need fast yet compact mo...

Please sign up or login with your details

Forgot password? Click here to reset