Syllable-level Neural Language Model for Agglutinative Language

08/18/2017
by   Seunghak Yu, et al.
0

Language models for agglutinative languages have always been hindered in past due to myriad of agglutinations possible to any given word through various affixes. We propose a method to diminish the problem of out-of-vocabulary words by introducing an embedding derived from syllables and morphemes which leverages the agglutinative property. Our model outperforms character-level embedding in perplexity by 16.87 with 9.50M parameters. Proposed method achieves state of the art performance over existing input prediction methods in terms of Key Stroke Saving and has been commercialized.

READ FULL TEXT
research
04/10/2017

Character-Word LSTM Language Models

We present a Character-Word Long Short-Term Memory Language Model which ...
research
12/10/2016

A Character-Word Compositional Neural Language Model for Finnish

Inspired by recent research, we explore ways to model the highly morphol...
research
08/26/2015

Character-Aware Neural Language Models

We describe a simple neural language model that relies only on character...
research
11/23/2022

Word-Level Representation From Bytes For Language Modeling

Modern language models mostly take sub-words as input, a design that bal...
research
09/24/2020

Grounded Compositional Outputs for Adaptive Language Modeling

Language models have emerged as a central component across NLP, and a gr...
research
04/08/2019

Crosslingual Document Embedding as Reduced-Rank Ridge Regression

There has recently been much interest in extending vector-based word rep...
research
02/23/2018

Reusing Weights in Subword-aware Neural Language Models

We propose several ways of reusing subword embeddings and other weights ...

Please sign up or login with your details

Forgot password? Click here to reset