Integrating Discrete and Neural Features via Mixed-feature Trans-dimensional Random Field Language Models

02/14/2020
by   Silin Gao, et al.
0

There has been a long recognition that discrete features (n-gram features) and neural network based features have complementary strengths for language models (LMs). Improved performance can be obtained by model interpolation, which is, however, a suboptimal two-step integration of discrete and neural features. The trans-dimensional random field (TRF) framework has the potential advantage of being able to flexibly integrate a richer set of features. However, either discrete or neural features are used alone in previous TRF LMs. This paper develops a mixed-feature TRF LM and demonstrates its advantage in integrating discrete and neural features. Various LMs are trained over PTB and Google one-billion-word datasets, and evaluated in N-best list rescoring experiments for speech recognition. Among all single LMs (i.e. without model interpolation), the mixed-feature TRF LMs perform the best, improving over both discrete TRF LMs and neural TRF LMs alone, and also being significantly better than LSTM LMs. Compared to interpolating two separately trained models with discrete and neural features respectively, the performance of mixed-feature TRF LMs matches the best interpolated model, and with simplified one-step training process and reduced training time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2016

Model Interpolation with Trans-dimensional Random Field Language Models for Speech Recognition

The dominant language models (LMs) such as n-gram and neural network (NN...
research
07/23/2017

Language modeling with Neural trans-dimensional random fields

Trans-dimensional random field language models (TRF LMs) have recently b...
research
10/30/2017

Learning neural trans-dimensional random field language models with noise-contrastive estimation

Trans-dimensional random field language models (TRF LMs) where sentences...
research
07/03/2018

Improved training of neural trans-dimensional random field language models with dynamic noise-contrastive estimation

A new whole-sentence language model - neural trans-dimensional random fi...
research
06/23/2016

NN-grams: Unifying neural network and n-gram language models for Speech Recognition

We present NN-grams, a novel, hybrid language model integrating n-grams ...
research
05/09/2013

Inferring Team Strengths Using a Discrete Markov Random Field

We propose an original model for inferring team strengths using a Markov...
research
05/03/2016

TheanoLM - An Extensible Toolkit for Neural Network Language Modeling

We present a new tool for training neural network language models (NNLMs...

Please sign up or login with your details

Forgot password? Click here to reset