Improved training of end-to-end attention models for speech recognition

05/08/2018
by   Albert Zeyer, et al.
0

Sequence-to-sequence attention-based models on subword units allow simple open-vocabulary end-to-end speech recognition. In this work, we show that such models can achieve competitive results on the Switchboard 300h and LibriSpeech 1000h tasks. In particular, we report the state-of-the-art word error rates (WER) of 3.54 of LibriSpeech. We introduce a new pretraining scheme by starting with a high time reduction factor and lowering it during training, which is crucial both for convergence and final performance. In some experiments, we also use an auxiliary CTC loss function to help the convergence. In addition, we train long short-term memory (LSTM) language models on subword units. By shallow fusion, we report up to 27 without a language model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2016

End-to-end attention-based distant speech recognition with Highway LSTM

End-to-end attention-based models have been shown to be competitive alte...
research
11/20/2019

On using 2D sequence-to-sequence models for speech recognition

Attention-based sequence-to-sequence models have shown promising results...
research
11/08/2018

Few-shot learning with attention-based sequence-to-sequence models

End-to-end approaches have recently become popular as a means of simplif...
research
05/19/2020

Investigations on Phoneme-Based End-To-End Speech Recognition

Common end-to-end models like CTC or encoder-decoder-attention models us...
research
11/06/2018

Language model integration based on memory control for sequence to sequence speech recognition

In this paper, we explore several new schemes to train a seq2seq model t...
research
03/12/2019

End-To-End Speech Recognition Using A High Rank LSTM-CTC Based Model

Long Short Term Memory Connectionist Temporal Classification (LSTM-CTC) ...
research
07/31/2020

Future Vector Enhanced LSTM Language Model for LVCSR

Language models (LM) play an important role in large vocabulary continuo...

Please sign up or login with your details

Forgot password? Click here to reset