On the impressive performance of randomly weighted encoders in summarization tasks

02/21/2020
by   Jonathan Pilault, et al.
0

In this work, we investigate the performance of untrained randomly initialized encoders in a general class of sequence to sequence models and compare their performance with that of fully-trained encoders on the task of abstractive summarization. We hypothesize that random projections of an input text have enough representational power to encode the hierarchical structure of sentences and semantics of documents. Using a trained decoder to produce abstractive text summaries, we empirically demonstrate that architectures with untrained randomly initialized encoders perform competitively with respect to the equivalent architectures with fully-trained encoders. We further find that the capacity of the encoder not only improves overall model generalization but also closes the performance gap between untrained randomly initialized and full-trained encoders. To our knowledge, it is the first time that general sequence to sequence models with attention are assessed for trained and randomly projected representations on abstractive summarization.

READ FULL TEXT

page 10

page 11

research
10/02/2019

SummAE: Zero-Shot Abstractive Text Summarization using Length-Agnostic Auto-Encoders

We propose an end-to-end neural model for zero-shot abstractive text sum...
research
10/03/2022

Probing of Quantitative Values in Abstractive Summarization Models

Abstractive text summarization has recently become a popular approach, b...
research
03/27/2018

Deep Communicating Agents for Abstractive Summarization

We present deep communicating agents in an encoder-decoder architecture ...
research
12/24/2019

Improving Abstractive Text Summarization with History Aggregation

Recent neural sequence to sequence models have provided feasible solutio...
research
06/18/2020

SEAL: Segment-wise Extractive-Abstractive Long-form Text Summarization

Most prior work in the sequence-to-sequence paradigm focused on datasets...
research
06/07/2021

Attention Temperature Matters in Abstractive Summarization Distillation

Recent progress of abstractive text summarization largely relies on larg...
research
11/06/2019

Enriching Conversation Context in Retrieval-based Chatbots

Work on retrieval-based chatbots, like most sequence pair matching tasks...

Please sign up or login with your details

Forgot password? Click here to reset