Efficient Dynamic WFST Decoding for Personalized Language Models

10/23/2019
by   Jun Liu, et al.
0

We propose a two-layer cache mechanism to speed up dynamic WFST decoding with personalized language models. The first layer is a public cache that stores most of the static part of the graph. This is shared globally among all users. A second layer is a private cache that caches the graph that represents the personalized language model, which is only shared by the utterances from a particular user. We also propose two simple yet effective pre-initialization methods, one based on breadth-first search, and another based on a data-driven exploration of decoder states using previous utterances. Experiments with a calling speech recognition task using a personalized contact list demonstrate that the proposed public cache reduces decoding time by factor of three compared to decoding without pre-initialization. Using the private cache provides additional efficiency gains, reducing the decoding time by a factor of five.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2021

Back from the future: bidirectional CTC decoding using future information in speech recognition

In this paper, we propose a simple but effective method to decode the ou...
research
07/07/2021

Improving Speech Recognition Accuracy of Local POI Using Geographical Models

Nowadays voice search for points of interest (POI) is becoming increasin...
research
10/31/2019

Learning to Customize Language Model for Generation-based dialog systems

Personalized conversation systems have received increasing attention rec...
research
03/16/2021

An Asynchronous WFST-Based Decoder For Automatic Speech Recognition

We introduce asynchronous dynamic decoder, which adopts an efficient A* ...
research
09/24/2018

Information-Weighted Neural Cache Language Models for ASR

Neural cache language models (LMs) extend the idea of regular cache lang...
research
04/02/2020

Full-Sum Decoding for Hybrid HMM based Speech Recognition using LSTM Language Model

In hybrid HMM based speech recognition, LSTM language models have been w...
research
10/01/2021

UserIdentifier: Implicit User Representations for Simple and Effective Personalized Sentiment Analysis

Global models are trained to be as generalizable as possible, with user ...

Please sign up or login with your details

Forgot password? Click here to reset