Experiments with LVT and FRE for Transformer model

04/26/2020
by   Ilshat Gibadullin, et al.
0

In this paper, we experiment with Large Vocabulary Trick and Feature-rich encoding applied to the Transformer model for Text Summarization. We could not achieve better results, than the analogous RNN-based sequence-to-sequence model, so we tried more models to find out, what improves the results and what deteriorates them.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/24/2019

Improving Abstractive Text Summarization with History Aggregation

Recent neural sequence to sequence models have provided feasible solutio...
research
08/28/2019

Solving Math Word Problems with Double-Decoder Transformer

This paper proposes a Transformer-based model to generate equations for ...
research
06/18/2019

Scheduled Sampling for Transformers

Scheduled sampling is a technique for avoiding one of the known problems...
research
06/18/2020

SEAL: Segment-wise Extractive-Abstractive Long-form Text Summarization

Most prior work in the sequence-to-sequence paradigm focused on datasets...
research
06/07/2021

Attention Temperature Matters in Abstractive Summarization Distillation

Recent progress of abstractive text summarization largely relies on larg...
research
04/01/2018

Training Tips for the Transformer Model

This article describes our experiments in neural machine translation usi...
research
03/25/2023

Indian Language Summarization using Pretrained Sequence-to-Sequence Models

The ILSUM shared task focuses on text summarization for two major Indian...

Please sign up or login with your details

Forgot password? Click here to reset