ChapterBreak: A Challenge Dataset for Long-Range Language Models

04/22/2022
by   Simeng Sun, et al.
0

While numerous architectures for long-range language models (LRLMs) have recently been proposed, a meaningful evaluation of their discourse-level language understanding capabilities has not yet followed. To this end, we introduce ChapterBreak, a challenge dataset that provides an LRLM with a long segment from a narrative that ends at a chapter boundary and asks it to distinguish the beginning of the ground-truth next chapter from a set of negative segments sampled from the same narrative. A fine-grained human annotation reveals that our dataset contains many complex types of chapter transitions (e.g., parallel narratives, cliffhanger endings) that require processing global context to comprehend. Experiments on ChapterBreak show that existing LRLMs fail to effectively leverage long-range context, substantially underperforming a segment-level model trained directly for this task. We publicly release our ChapterBreak dataset to spur more principled future research into LRLMs.

READ FULL TEXT
research
09/19/2021

Do Long-Range Language Models Actually Use Long-Range Context?

Language models are generally trained on short, truncated input sequence...
research
04/03/2020

TimeGate: Conditional Gating of Segments in Long-range Activities

When recognizing a long-range activity, exploring the entire video is ex...
research
12/21/2022

Training language models for deeper understanding improves brain alignment

Building systems that achieve a deeper understanding of language is one ...
research
10/12/2021

DiscoDVT: Generating Long Text with Discourse-Aware Discrete Variational Transformer

Despite the recent advances in applying pre-trained language models to g...
research
04/06/2023

Large language models effectively leverage document-level context for literary translation, but critical errors persist

Large language models (LLMs) are competitive with the state of the art o...
research
08/16/2023

MemoChat: Tuning LLMs to Use Memos for Consistent Long-Range Open-Domain Conversation

We propose MemoChat, a pipeline for refining instructions that enables l...
research
12/27/2016

The ontogeny of discourse structure mimics the development of literature

Discourse varies with age, education, psychiatric state and historical e...

Please sign up or login with your details

Forgot password? Click here to reset