Enhancing Semantic Understanding with Self-supervised Methods for Abstractive Dialogue Summarization

09/01/2022
by   Hyunjae Lee, et al.
0

Contextualized word embeddings can lead to state-of-the-art performances in natural language understanding. Recently, a pre-trained deep contextualized text encoder such as BERT has shown its potential in improving natural language tasks including abstractive summarization. Existing approaches in dialogue summarization focus on incorporating a large language model into summarization task trained on large-scale corpora consisting of news articles rather than dialogues of multiple speakers. In this paper, we introduce self-supervised methods to compensate shortcomings to train a dialogue summarization model. Our principle is to detect incoherent information flows using pretext dialogue text to enhance BERT's ability to contextualize the dialogue text representations. We build and fine-tune an abstractive dialogue summarization model on a shared encoder-decoder architecture using the enhanced BERT. We empirically evaluate our abstractive dialogue summarizer with the SAMSum corpus, a recently introduced dataset with abstractive dialogue summaries. All of our methods have contributed improvements to abstractive summary measured in ROUGE scores. Through an extensive ablation study, we also present a sensitivity analysis to critical model hyperparameters, probabilities of switching utterances and masking interlocutors.

READ FULL TEXT
research
12/20/2022

DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization

Dialogue summarization has recently garnered significant attention due t...
research
11/02/2020

Liputan6: A Large-scale Indonesian Dataset for Text Summarization

In this paper, we introduce a large-scale Indonesian summarization datas...
research
09/16/2021

A Bag of Tricks for Dialogue Summarization

Dialogue summarization comes with its own peculiar challenges as opposed...
research
05/13/2023

Self-Supervised Sentence Compression for Meeting Summarization

The conventional summarization model often fails to capture critical inf...
research
09/22/2021

DialogueBERT: A Self-Supervised Learning based Dialogue Pre-training Encoder

With the rapid development of artificial intelligence, conversational bo...
research
05/12/2015

Turn Segmentation into Utterances for Arabic Spontaneous Dialogues and Instance Messages

Text segmentation task is an essential processing task for many of Natur...
research
02/08/2023

Leveraging Summary Guidance on Medical Report Summarization

This study presents three deidentified large medical text datasets, name...

Please sign up or login with your details

Forgot password? Click here to reset