On Faithfulness and Factuality in Abstractive Summarization

05/02/2020
by   Joshua Maynez, et al.
0

It is well known that the standard likelihood training and approximate decoding objectives in neural text generation models lead to less human-like responses for open-ended tasks such as language modeling and story generation. In this paper we have analyzed limitations of these models for abstractive document summarization and found that these models are highly prone to hallucinate content that is unfaithful to the input document. We conducted a large scale human evaluation of several neural abstractive summarization systems to better understand the types of hallucinations they produce. Our human annotators found substantial amounts of hallucinated content in all model generated summaries. However, our analysis does show that pretrained models are better summarizers not only in terms of raw metrics, i.e., ROUGE, but also in generating faithful and factual summaries as evaluated by humans. Furthermore, we show that textual entailment measures better correlate with faithfulness than standard metrics, potentially leading the way to automatic evaluation metrics as well as training and decoding criteria.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2023

Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback

Despite the seeming success of contemporary grounded text generation sys...
research
07/20/2018

Abstractive and Extractive Text Summarization using Document Context Vector and Recurrent Neural Networks

Sequence to sequence (Seq2Seq) learning has recently been used for abstr...
research
10/30/2022

How Far are We from Robust Long Abstractive Summarization?

Abstractive summarization has made tremendous progress in recent years. ...
research
11/03/2020

Topic-Centric Unsupervised Multi-Document Summarization of Scientific and News Articles

Recent advances in natural language processing have enabled automation o...
research
04/15/2021

SummVis: Interactive Visual Analysis of Models, Data, and Evaluation for Text Summarization

Novel neural architectures, training strategies, and the availability of...
research
10/14/2021

MoFE: Mixture of Factual Experts for Controlling Hallucinations in Abstractive Summarization

Neural abstractive summarization models are susceptible to generating fa...
research
04/14/2021

SummScreen: A Dataset for Abstractive Screenplay Summarization

We introduce SummScreen, a summarization dataset comprised of pairs of T...

Please sign up or login with your details

Forgot password? Click here to reset