Examination and Extension of Strategies for Improving Personalized Language Modeling via Interpolation

06/09/2020
by   Liqun Shao, et al.
0

In this paper, we detail novel strategies for interpolating personalized language models and methods to handle out-of-vocabulary (OOV) tokens to improve personalized language models. Using publicly available data from Reddit, we demonstrate improvements in offline metrics at the user level by interpolating a global LSTM-based authoring model with a user-personalized n-gram model. By optimizing this approach with a back-off to uniform OOV penalty and the interpolation coefficient, we observe that over 80 perplexity, with an average of 5.2 research we extend previous work in building NLIs and improve the robustness of metrics for downstream tasks.

READ FULL TEXT
research
06/01/2022

What Changed? Investigating Debiasing Methods using Causal Mediation Analysis

Previous work has examined how debiasing language models affect downstre...
research
02/27/2023

LLaMA: Open and Efficient Foundation Language Models

We introduce LLaMA, a collection of foundation language models ranging f...
research
04/04/2022

"This is my unicorn, Fluffy": Personalizing frozen vision-language representations

Large Vision Language models pretrained on web-scale data provide re...
research
04/22/2023

LaMP: When Large Language Models Meet Personalization

This paper highlights the importance of personalization in the current s...
research
05/26/2023

Tokenization Impacts Multilingual Language Modeling: Assessing Vocabulary Allocation and Overlap Across Languages

Multilingual language models have recently gained attention as a promisi...
research
07/10/2020

Neural Composition: Learning to Generate from Multiple Models

Decomposing models into multiple components is critically important in m...
research
05/24/2023

KNN-LM Does Not Improve Open-ended Text Generation

In this paper, we study the generation quality of interpolation-based re...

Please sign up or login with your details

Forgot password? Click here to reset