A Data-Oriented Model of Literary Language

01/12/2017
by   Andreas van Cranenburgh, et al.
0

We consider the task of predicting how literary a text is, with a gold standard from human ratings. Aside from a standard bigram baseline, we apply rich syntactic tree fragments, mined from the training set, and a series of hand-picked features. Our model is the first to distinguish degrees of highly and less literary novels using a variety of lexical and syntactic features, and explains 76.0

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/02/2018

Syntax-Aware Language Modeling with Recurrent Neural Networks

Neural language models (LMs) are typically trained using only lexical fe...
research
09/30/2019

Lexical Features Are More Vulnerable, Syntactic Features Have More Predictive Power

Understanding the vulnerability of linguistic features extracted from no...
research
04/06/2018

A Categorical Approach to Syntactic Monoids

The syntactic monoid of a language is generalized to the level of a symm...
research
04/23/2018

Detecting Syntactic Features of Translated Chinese

We present a machine learning approach to distinguish texts translated t...
research
11/05/2018

Do RNNs learn human-like abstract word order preferences?

RNN language models have achieved state-of-the-art results on various ta...
research
01/11/2017

Distinguishing Antonyms and Synonyms in a Pattern-based Neural Network

Distinguishing between antonyms and synonyms is a key task to achieve hi...
research
05/24/2022

How Human is Human Evaluation? Improving the Gold Standard for NLG with Utility Theory

Human ratings are treated as the gold standard in NLG evaluation. The st...

Please sign up or login with your details

Forgot password? Click here to reset