TANet: Thread-Aware Pretraining for Abstractive Conversational Summarization

04/09/2022
by   Liran Wang, et al.
0

Although pre-trained language models (PLMs) have achieved great success and become a milestone in NLP, abstractive conversational summarization remains a challenging but less studied task. The difficulty lies in two aspects. One is the lack of large-scale conversational summary data. Another is that applying the existing pre-trained models to this task is tricky because of the structural dependence within the conversation and its informal expression, etc. In this work, we first build a large-scale (11M) pretraining dataset called RCS, based on the multi-person discussions in the Reddit community. We then present TANet, a thread-aware Transformer-based network. Unlike the existing pre-trained models that treat a conversation as a sequence of sentences, we argue that the inherent contextual dependency among the utterances plays an essential role in understanding the entire conversation and thus propose two new techniques to incorporate the structural information into our model. The first is thread-aware attention which is computed by taking into account the contextual dependency within utterances. Second, we apply thread prediction loss to predict the relations between utterances. We evaluate our model on four datasets of real conversations, covering types of meeting transcripts, customer-service records, and forum threads. Experimental results demonstrate that TANET achieves a new state-of-the-art in terms of both automatic evaluation and human judgment.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/04/2021

Conversations Are Not Flat: Modeling the Dynamic Information Flow across Dialogue Utterances

Nowadays, open-domain dialogue models can generate acceptable responses ...
research
12/16/2021

Call for Customized Conversation: Customized Conversation Grounding Persona and Knowledge

Humans usually have conversations by making use of prior knowledge about...
research
07/17/2022

Effectiveness of French Language Models on Abstractive Dialogue Summarization Task

Pre-trained language models have established the state-of-the-art on var...
research
09/09/2022

Enhancing Pre-trained Models with Text Structure Knowledge for Question Generation

Today the pre-trained language models achieve great success for question...
research
04/16/2021

Structure-Aware Abstractive Conversation Summarization via Discourse and Action Graphs

Abstractive conversation summarization has received much attention recen...
research
05/16/2023

GIFT: Graph-Induced Fine-Tuning for Multi-Party Conversation Understanding

Addressing the issues of who saying what to whom in multi-party conversa...

Please sign up or login with your details

Forgot password? Click here to reset