Reusing Weights in Subword-aware Neural Language Models

02/23/2018
by   Zhenisbek Assylbekov, et al.
0

We propose several ways of reusing subword embeddings and other weights in subword-aware neural language models. The proposed techniques do not benefit a competitive character-aware model, but some of them improve the performance of syllable- and morpheme-aware models while showing significant reductions in model sizes. We discover a simple hands-on principle: in a multi-layer input embedding model, layers should be tied consecutively bottom-up if reused at output. Our best morpheme-aware model with properly reused weights beats the competitive word-level model by a large margin across multiple languages and has 20

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2017

Syllable-aware Neural Language Models: A Failure to Beat Character-aware Ones

Syllabification does not seem to improve word-level RNN language modelin...
research
10/25/2019

Stem-driven Language Models for Morphologically Rich Languages

Neural language models (LMs) have shown to benefit significantly from en...
research
01/10/2019

Context Aware Machine Learning

We propose a principle for exploring context in machine learning models....
research
04/21/2017

Improving Context Aware Language Models

Increased adaptability of RNN language models leads to improved predicti...
research
08/18/2017

Syllable-level Neural Language Model for Agglutinative Language

Language models for agglutinative languages have always been hindered in...
research
11/23/2018

Unsupervised Word Discovery with Segmental Neural Language Models

We propose a segmental neural language model that combines the represent...
research
08/27/2019

Bridging the Gap for Tokenizer-Free Language Models

Purely character-based language models (LMs) have been lagging in qualit...

Please sign up or login with your details

Forgot password? Click here to reset