Spell Once, Summon Anywhere: A Two-Level Open-Vocabulary Language Model

04/23/2018
by   Sebastian J. Mielke, et al.
0

We show how to deploy recurrent neural networks within a hierarchical Bayesian language model. Our generative story combines a standard RNN language model (generating the word tokens in each sentence) with an RNN-based spelling model (generating the letters in each word type). These two RNNs respectively capture sentence structure and word structure, and are kept separate as in linguistics. The model can generate spellings for novel words in context and thus serves as an open-vocabulary language model. For known words, embeddings are naturally inferred by combining evidence from type spelling and token context. We compare to a number of baselines and previous work, establishing state-of-the-art results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2023

From Characters to Words: Hierarchical Pre-trained Language Model for Open-vocabulary Language Understanding

Current state-of-the-art models for natural language understanding requi...
research
05/11/2020

Toward Better Storylines with Sentence-Level Language Models

We propose a sentence-level language model which selects the next senten...
research
05/17/2023

A Better Way to Do Masked Language Model Scoring

Estimating the log-likelihood of a given sentence under an autoregressiv...
research
10/09/2020

Solving Historical Dictionary Codes with a Neural Language Model

We solve difficult word-based substitution codes by constructing a decod...
research
02/29/2016

Representation of linguistic form and function in recurrent neural networks

We present novel methods for analyzing the activation patterns of RNNs f...
research
01/03/2016

Contrastive Entropy: A new evaluation metric for unnormalized language models

Perplexity (per word) is the most widely used metric for evaluating lang...

Please sign up or login with your details

Forgot password? Click here to reset