Abstractive Summarization for Low Resource Data using Domain Transfer and Data Synthesis

02/09/2020
by   Ahmed Magooda, et al.
0

Training abstractive summarization models typically requires large amounts of data, which can be a limitation for many domains. In this paper we explore using domain transfer and data synthesis to improve the performance of recent abstractive summarization methods when applied to small corpora of student reflections. First, we explored whether tuning state of the art model trained on newspaper data could boost performance on student reflection data. Evaluations demonstrated that summaries produced by the tuned model achieved higher ROUGE scores compared to model trained on just student reflection data or just newspaper data. The tuned model also achieved higher scores compared to extractive summarization baselines, and additionally was judged to produce more coherent and readable summaries in human evaluations. Second, we explored whether synthesizing summaries of student data could additionally boost performance. We proposed a template-based model to synthesize new data, which when incorporated into training further increased ROUGE scores. Finally, we showed that combining data synthesis with domain transfer achieved higher ROUGE scores compared to only using one of the two approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2023

How Ready are Pre-trained Abstractive Models and LLMs for Legal Case Judgement Summarization?

Automatic summarization of legal case judgements has traditionally been ...
research
06/20/2023

GUMSum: Multi-Genre Data and Evaluation for English Abstractive Summarization

Automatic summarization with pre-trained language models has led to impr...
research
12/20/2022

DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization

Dialogue summarization has recently garnered significant attention due t...
research
09/17/2021

Exploring Multitask Learning for Low-Resource AbstractiveSummarization

This paper explores the effect of using multitask learning for abstracti...
research
02/09/2020

Attend to the beginning: A study on using bidirectional attention for extractive summarization

Forum discussion data differ in both structure and properties from gener...
research
10/22/2022

Extractive Summarization of Legal Decisions using Multi-task Learning and Maximal Marginal Relevance

Summarizing legal decisions requires the expertise of law practitioners,...
research
09/24/2019

In Conclusion Not Repetition: Comprehensive Abstractive Summarization With Diversified Attention Based On Determinantal Point Processes

Various Seq2Seq learning models designed for machine translation were ap...

Please sign up or login with your details

Forgot password? Click here to reset