Continuous Decomposition of Granularity for Neural Paraphrase Generation

09/05/2022
by   Xiaodong Gu, et al.
1

While Transformers have had significant success in paragraph generation, they treat sentences as linear sequences of tokens and often neglect their hierarchical information. Prior work has shown that decomposing the levels of granularity (e.g., word, phrase, or sentence) for input tokens has produced substantial improvements, suggesting the possibility of enhancing Transformers via more fine-grained modeling of granularity. In this work, we propose a continuous decomposition of granularity for neural paraphrase generation (C-DNPG). In order to efficiently incorporate granularity into sentence encoding, C-DNPG introduces a granularity-aware attention (GA-Attention) mechanism which extends the multi-head self-attention with: 1) a granularity head that automatically infers the hierarchical structure of a sentence by neurally estimating the granularity level of each input token; and 2) two novel attention masks, namely, granularity resonance and granularity scope, to efficiently encode granularity into attention. Experiments on two benchmarks, including Quora question pairs and Twitter URLs have shown that C-DNPG outperforms baseline models by a remarkable margin and achieves state-of-the-art results in terms of many metrics. Qualitative analysis reveals that C-DNPG indeed captures fine-grained levels of granularity with effectiveness.

READ FULL TEXT
research
09/05/2019

Multi-Granularity Self-Attention for Neural Machine Translation

Current state-of-the-art neural machine translation (NMT) uses a deep mu...
research
11/29/2018

Multi-granularity hierarchical attention fusion networks for reading comprehension and question answering

This paper describes a novel hierarchical attention network for reading ...
research
04/23/2016

Why and How to Pay Different Attention to Phrase Alignments of Different Intensities

This work studies comparatively two typical sentence pair classification...
research
06/24/2019

Decomposable Neural Paraphrase Generation

Paraphrasing exists at different granularity levels, such as lexical lev...
research
12/21/2019

Measuring Dataset Granularity

Despite the increasing visibility of fine-grained recognition in our fie...
research
10/11/2020

SDMTL: Semi-Decoupled Multi-grained Trajectory Learning for 3D human motion prediction

Predicting future human motion is critical for intelligent robots to int...
research
11/15/2022

Hierarchical Pronunciation Assessment with Multi-Aspect Attention

Automatic pronunciation assessment is a major component of a computer-as...

Please sign up or login with your details

Forgot password? Click here to reset