A Hierarchical Structured Self-Attentive Model for Extractive Document Summarization (HSSAS)

05/20/2018
by   Kamal Al-Sabahi, et al.
0

The recent advance in neural network architecture and training algorithms have shown the effectiveness of representation learning. The neural network-based models generate better representation than the traditional ones. They have the ability to automatically learn the distributed representation for sentences and documents. To this end, we proposed a novel model that addresses several issues that are not adequately modeled by the previously proposed models, such as the memory problem and incorporating the knowledge of document structure. Our model uses a hierarchical structured self-attention mechanism to create the sentence and document embeddings. This architecture mirrors the hierarchical structure of the document and in turn enables us to obtain better feature representation. The attention mechanism provides extra source of information to guide the summary extraction. The new model treated the summarization task as a classification problem in which the model computes the respective probabilities of sentence-summary membership. The model predictions are broken up by several features such as information content, salience, novelty and positional representation. The proposed model was evaluated on two well-known datasets, the CNN / Daily Mail, and DUC 2002. The experimental results show that our model outperforms the current extractive state-of-the-art by a considerable margin.

READ FULL TEXT
research
11/14/2016

Classify or Select: Neural Architectures for Extractive Document Summarization

We present two novel and contrasting Recurrent Neural Network (RNN) base...
research
01/15/2023

Hawk: An Industrial-strength Multi-label Document Classifier

There are a plethora of methods and algorithms that solve the classical ...
research
03/01/2020

StructSum: Incorporating Latent and Explicit Sentence Dependencies for Single Document Summarization

Traditional preneural approaches to single document summarization relied...
research
09/23/2021

Incorporating Linguistic Knowledge for Abstractive Multi-document Summarization

Within natural language processing tasks, linguistic knowledge can alway...
research
09/27/2018

Iterative Document Representation Learning Towards Summarization with Polishing

In this paper, we introduce Iterative Text Summarization (ITS), an itera...
research
05/10/2018

Obligation and Prohibition Extraction Using Hierarchical RNNs

We consider the task of detecting contractual obligations and prohibitio...
research
05/24/2019

Outline Generation: Understanding the Inherent Content Structure of Documents

In this paper, we introduce and tackle the Outline Generation (OG) task,...

Please sign up or login with your details

Forgot password? Click here to reset