The Summary Loop: Learning to Write Abstractive Summaries Without Examples

05/11/2021
by   Philippe Laban, et al.
8

This work presents a new approach to unsupervised abstractive summarization based on maximizing a combination of coverage and fluency for a given length constraint. It introduces a novel method that encourages the inclusion of key terms from the original document into the summary: key terms are masked out of the original document and must be filled in by a coverage model using the current generated summary. A novel unsupervised training procedure leverages this coverage model along with a fluency model to generate and score summaries. When tested on popular news summarization datasets, the method outperforms previous unsupervised methods by more than 2 R-1 points, and approaches results of competitive supervised methods. Our model attains higher levels of abstraction with copied passages roughly two times shorter than prior work, and learns to compress and merge sentences without supervision.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2020

Unsupervised Reference-Free Summary Quality Evaluation via Contrastive Learning

Evaluation of a document summarization system has been a critical factor...
research
07/17/2020

SummPip: Unsupervised Multi-Document Summarization with Sentence Graph Compression

Obtaining training data for multi-document summarization (MDS) is time c...
research
01/07/2022

An Unsupervised Masking Objective for Abstractive Multi-Document News Summarization

We show that a simple unsupervised masking objective can approach near s...
research
12/28/2020

On Generating Extended Summaries of Long Documents

Prior work in document summarization has mainly focused on generating sh...
research
06/07/2021

Neural Abstractive Unsupervised Summarization of Online News Discussions

Summarization has usually relied on gold standard summaries to train ext...
research
12/07/2020

An Enhanced MeanSum Method For Generating Hotel Multi-Review Summarizations

Multi-document summaritazion is the process of taking multiple texts as ...
research
09/16/2019

BottleSum: Unsupervised and Self-supervised Sentence Summarization using the Information Bottleneck Principle

The principle of the Information Bottleneck (Tishby et al. 1999) is to p...

Please sign up or login with your details

Forgot password? Click here to reset