Scalable Multi Corpora Neural Language Models for ASR

07/02/2019
by   Anirudh Raju, et al.
0

Neural language models (NLM) have been shown to outperform conventional n-gram language models by a substantial margin in Automatic Speech Recognition (ASR) and other tasks. There are, however, a number of challenges that need to be addressed for an NLM to be used in a practical large-scale ASR system. In this paper, we present solutions to some of the challenges, including training NLM from heterogenous corpora, limiting latency impact and handling personalized bias in the second-pass rescorer. Overall, we show that we can achieve a 6.2 rescoring framework with a minimal increase in latency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/12/2023

On the N-gram Approximation of Pre-trained Language Models

Large pre-trained language models (PLMs) have shown remarkable performan...
research
04/13/2022

Scalable Training of Language Models using JAX pjit and TPUv4

Modern large language models require distributed training strategies due...
research
06/29/2022

Space-Efficient Representation of Entity-centric Query Language Models

Virtual assistants make use of automatic speech recognition (ASR) to hel...
research
07/04/2020

Robust Prediction of Punctuation and Truecasing for Medical ASR

Automatic speech recognition (ASR) systems in the medical domain that fo...
research
01/08/2022

Defining maximum acceptable latency of AI-enhanced CAI tools

Recent years have seen an increasing number of studies around the design...
research
07/04/2020

Robust Prediction of Punctuation and Truecasingfor Medical ASR

Automatic speech recognition (ASR) systems in the medical domain that fo...
research
11/18/2022

A Persian ASR-based SER: Modification of Sharif Emotional Speech Database and Investigation of Persian Text Corpora

Speech Emotion Recognition (SER) is one of the essential perceptual meth...

Please sign up or login with your details

Forgot password? Click here to reset