Dissecting Generation Modes for Abstractive Summarization Models via Ablation and Attribution

06/03/2021
by   Jiacheng Xu, et al.
0

Despite the prominence of neural abstractive summarization models, we know little about how they actually form summaries and how to understand where their decisions come from. We propose a two-step method to interpret summarization model decisions. We first analyze the model's behavior by ablating the full model to categorize each decoder decision into one of several generation modes: roughly, is the model behaving like a language model, is it relying heavily on the input, or is it somewhere in between? After isolating decisions that do depend on the input, we explore interpreting these decisions using several different attribution methods. We compare these techniques based on their ability to select content and reconstruct the model's predicted token from perturbations of the input, thus revealing whether highlighted attributions are truly important for the generation of the next token. While this machinery can be broadly useful even beyond summarization, we specifically demonstrate its capability to identify phrases the summarization model has memorized and determine where in the training pipeline this memorization happened, as well as study complex generation phenomena like sentence fusion on a per-instance basis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/15/2020

Understanding Neural Abstractive Summarization Models via Uncertainty

An advantage of seq2seq abstractive summarization models is that they ge...
research
06/25/2016

Summarizing Decisions in Spoken Meetings

This paper addresses the problem of summarizing decisions in spoken meet...
research
10/04/2022

Towards Improving Faithfulness in Abstractive Summarization

Despite the success achieved in neural abstractive summarization based o...
research
04/30/2020

How do Decisions Emerge across Layers in Neural Models? Interpretation with Differentiable Masking

Attribution methods assess the contribution of inputs (e.g., words) to t...
research
03/16/2022

Don't Say What You Don't Know: Improving the Consistency of Abstractive Summarization by Constraining Beam Search

Abstractive summarization systems today produce fluent and relevant outp...
research
04/07/2020

Windowing Models for Abstractive Summarization of Long Texts

Neural summarization models suffer from the fixed-size input limitation:...
research
07/15/2020

Learning Invariances for Interpretability using Supervised VAE

We propose to learn model invariances as a means of interpreting a model...

Please sign up or login with your details

Forgot password? Click here to reset