Discourse Probing of Pretrained Language Models

04/13/2021
by   Fajri Koto, et al.
4

Existing work on probing of pretrained language models (LMs) has predominantly focused on sentence-level syntactic tasks. In this paper, we introduce document-level discourse probing to evaluate the ability of pretrained LMs to capture document-level relations. We experiment with 7 pretrained LMs, 4 languages, and 7 discourse probing tasks, and find BART to be overall the best model at capturing discourse – but only in its encoder, with BERT performing surprisingly well as the baseline model. Across the different models, there are substantial differences in which layers best capture discourse information, and large disparities between models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/18/2021

Evaluating Document Coherence Modelling

While pretrained language models ("LM") have driven impressive gains ove...
research
05/31/2023

How Does Pretraining Improve Discourse-Aware Translation?

Pretrained language models (PLMs) have produced substantial improvements...
research
10/18/2022

Less is More: Simplifying Feature Extractors Prevents Overfitting for Neural Discourse Parsing Models

Complex feature extractors are widely employed for text representation b...
research
11/12/2015

Document Context Language Models

Text documents are structured on multiple levels of detail: individual w...
research
08/31/2019

Evaluation Benchmarks and Learning Criteriafor Discourse-Aware Sentence Representations

Prior work on pretrained sentence embeddings and benchmarks focus on the...
research
11/28/2022

Scientific and Creative Analogies in Pretrained Language Models

This paper examines the encoding of analogy in large-scale pretrained la...
research
07/16/2023

Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language Modelling

Modeling discourse – the linguistic phenomena that go beyond individual ...

Please sign up or login with your details

Forgot password? Click here to reset