Efficient Attentions for Long Document Summarization

04/05/2021
by   Luyang Huang, et al.
0

The quadratic computational and memory complexities of large Transformers have limited their scalability for long document summarization. In this paper, we propose Hepos, a novel efficient encoder-decoder attention with head-wise positional strides to effectively pinpoint salient information from the source. We further conduct a systematic study of existing efficient self-attentions. Combined with Hepos, we are able to process ten times more tokens than existing models that use full attentions. For evaluation, we present a new dataset, GovReport, with significantly longer documents and summaries. Results show that our models produce significantly higher ROUGE scores than competitive comparisons, including new state-of-the-art results on PubMed. Human evaluation also shows that our models generate more informative summaries with fewer unfaithful errors.

READ FULL TEXT
research
06/18/2020

SEAL: Segment-wise Extractive-Abstractive Long-form Text Summarization

Most prior work in the sequence-to-sequence paradigm focused on datasets...
research
05/11/2017

A Deep Reinforced Model for Abstractive Summarization

Attentional, RNN-based encoder-decoder models for abstractive summarizat...
research
12/03/2020

Bengali Abstractive News Summarization(BANS): A Neural Attention Approach

Abstractive summarization is the process of generating novel sentences b...
research
05/28/2018

Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting

Inspired by how humans summarize long documents, we propose an accurate ...
research
05/03/2020

Knowledge Graph-Augmented Abstractive Summarization with Semantic-Driven Cloze Reward

Sequence-to-sequence models for abstractive summarization have been stud...
research
05/24/2023

AWESOME: GPU Memory-constrained Long Document Summarization using Memory Mechanism and Global Salient Content

Long document summarization systems are critical for domains with length...
research
06/04/2019

How Large Are Lions? Inducing Distributions over Quantitative Attributes

Most current NLP systems have little knowledge about quantitative attrib...

Please sign up or login with your details

Forgot password? Click here to reset