Deep Reinforced Self-Attention Masks for Abstractive Summarization (DR.SAS)

12/30/2019
by   Ankit Chadha, et al.
0

We present a novel architectural scheme to tackle the abstractive summarization problem based on the CNN/DMdataset which fuses Reinforcement Learning (RL) withUniLM, which is a pre-trained Deep Learning Model, to solve various natural language tasks. We have tested the limits of learning fine-grained attention in Transformers to improve the summarization quality. UniLM applies attention to the entire token space in a global fashion. We propose DR.SAS which applies the Actor-Critic (AC) algorithm to learn a dynamic self-attention distribution over the tokens to reduce redundancy and generate factual and coherent summaries to improve the quality of summarization. After performing hyperparameter tuning, we achievedbetter ROUGE results compared to the baseline. Our model tends to be more extractive/factual yet coherent in detail because of optimization over ROUGE rewards. We present detailed error analysis with examples of the strengths and limitations of our model. Our codebase will be publicly available on our GitHub.

READ FULL TEXT

page 4

page 6

page 7

research
03/28/2018

Actor-Critic based Training Framework for Abstractive Summarization

We present a training framework for neural abstractive summarization bas...
research
10/20/2018

Abstractive Summarization Using Attentive Neural Techniques

In a world of proliferating data, the ability to rapidly summarize text ...
research
03/15/2022

Long Document Summarization with Top-down and Bottom-up Inference

Text summarization aims to condense long documents and retain key inform...
research
04/12/2022

RL-CoSeg : A Novel Image Co-Segmentation Algorithm with Deep Reinforcement Learning

This paper proposes an automatic image co-segmentation algorithm based o...
research
05/17/2022

Unraveling Attention via Convex Duality: Analysis and Interpretations of Vision Transformers

Vision transformers using self-attention or its proposed alternatives ha...
research
05/25/2022

Leveraging Locality in Abstractive Text Summarization

Despite the successes of neural attention models for natural language ge...
research
10/15/2020

Understanding Neural Abstractive Summarization Models via Uncertainty

An advantage of seq2seq abstractive summarization models is that they ge...

Please sign up or login with your details

Forgot password? Click here to reset