Improved Regularization Techniques for End-to-End Speech Recognition

12/19/2017
by   Yingbo Zhou, et al.
0

Regularization is important for end-to-end speech models, since the models are highly flexible and easy to overfit. Data augmentation and dropout has been important for improving end-to-end models in other domains. However, they are relatively under explored for end-to-end speech models. Therefore, we investigate the effectiveness of both methods for end-to-end trainable, deep speech recognition models. We augment audio data through random perturbations of tempo, pitch, volume, temporal alignment, and adding random noise.We further investigate the effect of dropout when applied to the inputs of all layers of the network. We show that the combination of data augmentation and dropout give a relative performance improvement on both Wall Street Journal (WSJ) and LibriSpeech dataset of over 20 other end-to-end speech models on both datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/08/2022

Auditory-Based Data Augmentation for End-to-End Automatic Speech Recognition

End-to-end models have achieved significant improvement on automatic spe...
research
10/08/2020

Population Based Training for Data Augmentation and Regularization in Speech Recognition

Varying data augmentation policies and regularization over the course of...
research
03/12/2019

End-To-End Speech Recognition Using A High Rank LSTM-CTC Based Model

Long Short Term Memory Connectionist Temporal Classification (LSTM-CTC) ...
research
12/29/2022

Macro-block dropout for improved regularization in training end-to-end speech recognition models

This paper proposes a new regularization algorithm referred to as macro-...
research
05/23/2023

BM25 Query Augmentation Learned End-to-End

Given BM25's enduring competitiveness as an information retrieval baseli...
research
04/05/2019

Jasper: An End-to-End Convolutional Neural Acoustic Model

In this paper, we report state-of-the-art results on LibriSpeech among e...
research
06/08/2020

The Penalty Imposed by Ablated Data Augmentation

There is a set of data augmentation techniques that ablate parts of the ...

Please sign up or login with your details

Forgot password? Click here to reset