On the long-term learning ability of LSTM LMs

06/16/2021
by   Wim Boes, et al.
0

We inspect the long-term learning ability of Long Short-Term Memory language models (LSTM LMs) by evaluating a contextual extension based on the Continuous Bag-of-Words (CBOW) model for both sentence- and discourse-level LSTM LMs and by analyzing its performance. We evaluate on text and speech. Sentence-level models using the long-term contextual module perform comparably to vanilla discourse-level LSTM LMs. On the other hand, the extension does not provide gains for discourse-level models. These findings indicate that discourse-level LSTM LMs already rely on contextual information to perform long-term learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/31/2020

Future Vector Enhanced LSTM Language Model for LVCSR

Language models (LM) play an important role in large vocabulary continuo...
research
02/19/2016

Contextual LSTM (CLSTM) models for Large scale NLP tasks

Documents exhibit sequential structure at multiple levels of abstraction...
research
08/11/2017

Argument Labeling of Explicit Discourse Relations using LSTM Neural Networks

Argument labeling of explicit discourse relations is a challenging task....
research
11/11/2015

Larger-Context Language Modelling

In this work, we propose a novel method to incorporate corpus-level disc...
research
10/21/2020

LSTM-LM with Long-Term History for First-Pass Decoding in Conversational Speech Recognition

LSTM language models (LSTM-LMs) have been proven to be powerful and yiel...
research
06/06/2020

Do RNN and LSTM have Long Memory?

The LSTM network was proposed to overcome the difficulty in learning lon...
research
06/14/2016

Shallow Discourse Parsing Using Distributed Argument Representations and Bayesian Optimization

This paper describes the Georgia Tech team's approach to the CoNLL-2016 ...

Please sign up or login with your details

Forgot password? Click here to reset