A Cross-Task Analysis of Text Span Representations

06/06/2020
by   Shubham Toshniwal, et al.
0

Many natural language processing (NLP) tasks involve reasoning with textual spans, including question answering, entity recognition, and coreference resolution. While extensive research has focused on functional architectures for representing words and sentences, there is less work on representing arbitrary spans of text within sentences. In this paper, we conduct a comprehensive empirical evaluation of six span representation methods using eight pretrained language representation models across six tasks, including two tasks that we introduce. We find that, although some simple span representations are fairly reliable across tasks, in general the optimal span representation varies by task, and can also vary within different facets of individual tasks. We also find that the choice of span representation has a bigger impact with a fixed pretrained encoder than with a fine-tuned encoder.

READ FULL TEXT

page 5

page 6

page 7

research
05/07/2022

Number Entity Recognition

Numbers are essential components of text, like any other word tokens, fr...
research
01/02/2021

Coreference Resolution without Span Representations

Since the introduction of deep pretrained language models, most task-spe...
research
03/22/2023

Salient Span Masking for Temporal Understanding

Salient Span Masking (SSM) has shown itself to be an effective strategy ...
research
07/24/2019

SpanBERT: Improving Pre-training by Representing and Predicting Spans

We present SpanBERT, a pre-training method that is designed to better re...
research
01/10/2022

SCROLLS: Standardized CompaRison Over Long Language Sequences

NLP benchmarks have largely focused on short texts, such as sentences an...
research
05/08/2022

A Structured Span Selector

Many natural language processing tasks, e.g., coreference resolution and...

Please sign up or login with your details

Forgot password? Click here to reset