Abstractive Text Summarization based on Language Model Conditioning and Locality Modeling

03/29/2020
by   Dmitrii Aksenov, et al.
0

We explore to what extent knowledge about the pre-trained language model that is used is beneficial for the task of abstractive summarization. To this end, we experiment with conditioning the encoder and decoder of a Transformer-based neural model on the BERT language model. In addition, we propose a new method of BERT-windowing, which allows chunk-wise processing of texts longer than the BERT window size. We also explore how locality modelling, i.e., the explicit restriction of calculations to the local context, can affect the summarization ability of the Transformer. This is done by introducing 2-dimensional convolutional self-attention into the first layers of the encoder. The results of our models are compared to a baseline and the state-of-the-art models on the CNN/Daily Mail dataset. We additionally train our model on the SwissText dataset to demonstrate usability on German. Both models outperform the baseline in ROUGE scores on two datasets and show its superiority in a manual qualitative analysis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/21/2019

Sample Efficient Text Summarization Using a Single Pre-Trained Transformer

Language model (LM) pre-training has resulted in impressive performance ...
research
04/05/2022

Abstractive summarization of hospitalisation histories with transformer networks

In this paper we present a novel approach to abstractive summarization o...
research
02/25/2019

Pretraining-Based Natural Language Generation for Text Summarization

In this paper, we propose a novel pretraining-based encoder-decoder fram...
research
11/16/2021

Meeting Summarization with Pre-training and Clustering Methods

Automatic meeting summarization is becoming increasingly popular these d...
research
10/10/2021

On Automatic Text Extractive Summarization Based on Graph and pre-trained Language Model Attention

Representing text as graph to solve the summarization task has been disc...
research
08/03/2022

Introducing BEREL: BERT Embeddings for Rabbinic-Encoded Language

We present a new pre-trained language model (PLM) for Rabbinic Hebrew, t...
research
10/09/2020

Q-learning with Language Model for Edit-based Unsupervised Summarization

Unsupervised methods are promising for abstractive text summarization in...

Please sign up or login with your details

Forgot password? Click here to reset