Highway Long Short-Term Memory RNNs for Distant Speech Recognition

10/30/2015
by   Yu Zhang, et al.
0

In this paper, we extend the deep long short-term memory (DLSTM) recurrent neural networks by introducing gated direct connections between memory cells in adjacent layers. These direct links, called highway connections, enable unimpeded information flow across different layers and thus alleviate the gradient vanishing problem when building deeper LSTMs. We further introduce the latency-controlled bidirectional LSTMs (BLSTMs) which can exploit the whole history while keeping the latency under control. Efficient algorithms are proposed to train these novel networks using both frame and sequence discriminative criteria. Experiments on the AMI distant speech recognition (DSR) task indicate that we can train deeper LSTMs and achieve better improvement from sequence training with highway LSTMs (HLSTMs). Our novel model obtains 43.9/47.7% WER on AMI (SDM) dev and eval sets, outperforming all previous works. It beats the strong DNN and DLSTM baselines with 15.7% and 5.3% relative improvement respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2018

Bidirectional Quaternion Long-Short Term Memory Recurrent Neural Networks for Speech Recognition

Recurrent neural networks (RNN) are at the core of modern automatic spee...
research
09/29/2017

Improving speech recognition by revising gated recurrent units

Speech recognition is largely taking advantage of deep learning, showing...
research
03/04/2018

Deep-FSMN for Large Vocabulary Continuous Speech Recognition

In this paper, we present an improved feedforward sequential memory netw...
research
03/19/2019

NeuralHydrology - Interpreting LSTMs in Hydrology

Despite the huge success of Long Short-Term Memory networks, their appli...
research
03/12/2017

Combining Residual Networks with LSTMs for Lipreading

We propose an end-to-end deep learning architecture for word-level visua...
research
09/11/2018

Response Characterization for Auditing Cell Dynamics in Long Short-term Memory Networks

In this paper, we introduce a novel method to interpret recurrent neural...
research
10/17/2014

Learning to Execute

Recurrent Neural Networks (RNNs) with Long Short-Term Memory units (LSTM...

Please sign up or login with your details

Forgot password? Click here to reset