Meta-Learning a Dynamical Language Model

03/28/2018
by   Thomas Wolf, et al.
0

We consider the task of word-level language modeling and study the possibility of combining hidden-states-based short-term representations with medium-term representations encoded in dynamical weights of a language model. Our work extends recent experiments on language models with dynamically evolving weights by casting the language modeling problem into an online learning-to-learn framework in which a meta-learner is trained by gradient-descent to continuously update a language model weights.

READ FULL TEXT
research
05/15/2018

Continuous Learning in a Hierarchical Multiscale Neural Network

We reformulate the problem of encoding a multi-scale representation of a...
research
09/04/2019

PaLM: A Hybrid Parser and Language Model

We present PaLM, a hybrid parser and neural language model. Building on ...
research
05/01/2020

Selecting Informative Contexts Improves Language Model Finetuning

We present a general finetuning meta-method that we call information gai...
research
09/08/2020

Probabilistic Predictions of People Perusing: Evaluating Metrics of Language Model Performance for Psycholinguistic Modeling

By positing a relationship between naturalistic reading times and inform...
research
04/08/2021

Revisiting Simple Neural Probabilistic Language Models

Recent progress in language modeling has been driven not only by advance...
research
05/29/2021

Predictive Representation Learning for Language Modeling

To effectively perform the task of next-word prediction, long short-term...
research
11/02/2022

Audio Language Modeling using Perceptually-Guided Discrete Representations

In this work, we study the task of Audio Language Modeling, in which we ...

Please sign up or login with your details

Forgot password? Click here to reset