SpanBERT: Improving Pre-training by Representing and Predicting Spans

07/24/2019
by   Mandar Joshi, et al.
0

We present SpanBERT, a pre-training method that is designed to better represent and predict spans of text. Our approach extends BERT by (1) masking contiguous random spans, rather than random tokens, and (2) training the span boundary representations to predict the entire content of the masked span, without relying on the individual token representations within it. SpanBERT consistently outperforms BERT and our better-tuned baselines, with substantial gains on span selection tasks such as question answering and coreference resolution. In particular, with the same training data and model size as BERT-large, our single model obtains 94.6 respectively. We also achieve a new state of the art on the OntoNotes coreference resolution task (79.6 benchmark (70.8

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2019

Span Selection Pre-training for Question Answering

BERT (Bidirectional Encoder Representations from Transformers) and relat...
research
01/02/2021

Improving Sequence-to-Sequence Pre-training via Sequence Span Rewriting

In this paper, we generalize text infilling (e.g., masked language model...
research
04/07/2020

Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-based Question Answering

We introduce a novel approach to transformers that learns hierarchical r...
research
02/24/2022

BERTVision – A Parameter-Efficient Approach for Question Answering

We present a highly parameter efficient approach for Question Answering ...
research
12/09/2021

Semantic Search as Extractive Paraphrase Span Detection

In this paper, we approach the problem of semantic search by framing the...
research
10/13/2022

An Empirical Study on Finding Spans

We present an empirical study on methods for span finding, the selection...
research
06/06/2020

A Cross-Task Analysis of Text Span Representations

Many natural language processing (NLP) tasks involve reasoning with text...

Please sign up or login with your details

Forgot password? Click here to reset