Reducing Exposure Bias in Training Recurrent Neural Network Transducers

08/24/2021
by   Xiaodong Cui, et al.
0

When recurrent neural network transducers (RNNTs) are trained using the typical maximum likelihood criterion, the prediction network is trained only on ground truth label sequences. This leads to a mismatch during inference, known as exposure bias, when the model must deal with label sequences containing errors. In this paper we investigate approaches to reducing exposure bias in training to improve the generalization of RNNT models for automatic speech recognition (ASR). A label-preserving input perturbation to the prediction network is introduced. The input token sequences are perturbed using SwitchOut and scheduled sampling based on an additional token language model. Experiments conducted on the 300-hour Switchboard dataset demonstrate their effectiveness. By reducing the exposure bias, we show that we can further improve the accuracy of a high-performance RNNT ASR model and obtain state-of-the-art results on the 300-hour Switchboard dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2019

Quantifying Exposure Bias for Neural Language Generation

The exposure bias problem refers to the training-inference discrepancy c...
research
03/29/2022

Improving Generalization of Deep Neural Network Acoustic Models with Length Perturbation and N-best Based Label Smoothing

We introduce two techniques, length perturbation and n-best based label ...
research
05/14/2018

Token-level and sequence-level loss smoothing for RNN language models

Despite the effectiveness of recurrent neural network language models, t...
research
01/27/2023

Input Perturbation Reduces Exposure Bias in Diffusion Models

Denoising Diffusion Probabilistic Models have shown an impressive genera...
research
08/29/2019

Translating Mathematical Formula Images to LaTeX Sequences Using Deep Neural Networks with Sequence-level Training

In this paper we propose a deep neural network model with an encoder-dec...
research
08/29/2023

Elucidating the Exposure Bias in Diffusion Models

Diffusion models have demonstrated impressive generative capabilities, b...
research
11/06/2018

Language GANs Falling Short

Generating high-quality text with sufficient diversity is essential for ...

Please sign up or login with your details

Forgot password? Click here to reset