At Which Level Should We Extract? An Empirical Study on Extractive Document Summarization

04/06/2020
by   Qingyu Zhou, et al.
0

Extractive methods have proven to be very effective in automatic document summarization. Previous works perform this task by identifying informative contents at sentence level. However, it is unclear whether performing extraction at sentence level is the best solution. In this work, we show that unnecessity and redundancy issues exist when extracting full sentences, and extracting sub-sentential units is a promising alternative. Specifically, we propose extracting sub-sentential units on the corresponding constituency parsing tree. A neural extractive model which leverages the sub-sentential information and extracts them is presented. Extensive experiments and analyses show that extracting sub-sentential units performs competitively comparing to full sentence extraction under the evaluation of both automatic and human evaluations. Hopefully, our work could provide some inspiration of the basic extraction units in extractive summarization for future research.

READ FULL TEXT
research
03/23/2016

Neural Summarization by Extracting Sentences and Words

Traditional approaches to extractive summarization rely heavily on human...
research
01/20/2016

Improved Spoken Document Summarization with Coverage Modeling Techniques

Extractive summarization aims at selecting a set of indicative sentences...
research
06/14/2019

DocRED: A Large-Scale Document-Level Relation Extraction Dataset

Multiple entities in a document generally exhibit complex inter-sentence...
research
04/16/2021

Unsupervised Extractive Summarization by Human Memory Simulation

Summarization systems face the core challenge of identifying and selecti...
research
03/07/2023

Towards Interpretable and Efficient Automatic Reference-Based Summarization Evaluation

Interpretability and efficiency are two important considerations for the...
research
09/14/2023

Less is More for Long Document Summary Evaluation by LLMs

Large Language Models (LLMs) have shown promising performance in summary...
research
11/09/2022

Novel Chapter Abstractive Summarization using Spinal Tree Aware Sub-Sentential Content Selection

Summarizing novel chapters is a difficult task due to the input length a...

Please sign up or login with your details

Forgot password? Click here to reset