MoFE: Mixture of Factual Experts for Controlling Hallucinations in Abstractive Summarization

10/14/2021
by   Prafulla Kumar Choubey, et al.
5

Neural abstractive summarization models are susceptible to generating factually inconsistent content, a phenomenon known as hallucination. This limits the usability and adoption of these systems in real-world applications. To reduce the presence of hallucination, we propose the Mixture of Factual Experts (MoFE) model, which combines multiple summarization experts that each target a specific type of error. We train our experts using reinforcement learning (RL) to minimize the error defined by two factual consistency metrics: entity overlap and dependency arc entailment. We construct MoFE by combining the experts using two ensembling strategies (weights and logits) and evaluate them on two summarization datasets (XSUM and CNN/DM). Our experiments on BART models show that the MoFE improves performance according to both entity overlap and dependency arc entailment, without a significant performance drop on standard ROUGE metrics. The performance improvement also transfers to unseen factual consistency metrics, such as question answer-based factuality evaluation metric and BERTScore precision with respect to the source document.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2023

Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback

Despite the seeming success of contemporary grounded text generation sys...
research
02/18/2021

Entity-level Factual Consistency of Abstractive Text Summarization

A key challenge for abstractive summarization is ensuring factual consis...
research
12/16/2021

QAFactEval: Improved QA-Based Factual Consistency Evaluation for Summarization

Factual consistency is an essential quality of text summarization models...
research
05/02/2020

On Faithfulness and Factuality in Abstractive Summarization

It is well known that the standard likelihood training and approximate d...
research
09/04/2019

Answers Unite! Unsupervised Metrics for Reinforced Summarization Models

Abstractive summarization approaches based on Reinforcement Learning (RL...
research
04/17/2018

Multi-Reward Reinforced Summarization with Saliency and Entailment

Abstractive text summarization is the task of compressing and rewriting ...
research
05/26/2023

With a Little Push, NLI Models can Robustly and Efficiently Predict Faithfulness

Conditional language models still generate unfaithful output that is not...

Please sign up or login with your details

Forgot password? Click here to reset