Multi-level Head-wise Match and Aggregation in Transformer for Textual Sequence Matching

01/20/2020
by   Shuohang Wang, et al.
0

Transformer has been successfully applied to many natural language processing tasks. However, for textual sequence matching, simple matching between the representation of a pair of sequences might bring in unnecessary noise. In this paper, we propose a new approach to sequence pair matching with Transformer, by learning head-wise matching representations on multiple levels. Experiments show that our proposed approach can achieve new state-of-the-art performance on multiple tasks that rely only on pre-computed sequence-vector-representation, such as SNLI, MNLI-match, MNLI-mismatch, QQP, and SQuAD-binary.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/30/2019

An Augmented Transformer Architecture for Natural Language Generation Tasks

The Transformer based neural networks have been showing significant adva...
research
06/18/2020

Multi-branch Attentive Transformer

While the multi-branch architecture is one of the key ingredients to the...
research
10/30/2020

SLM: Learning a Discourse Language Representation with Sentence Unshuffling

We introduce Sentence-level Language Modeling, a new pre-training object...
research
06/16/2019

Multi-Level Matching and Aggregation Network for Few-Shot Relation Classification

This paper presents a multi-level matching and aggregation network (MLMA...
research
07/29/2019

Leveraging Pre-trained Checkpoints for Sequence Generation Tasks

Unsupervised pre-training of large neural models has recently revolution...
research
07/23/2023

Transformer-based Joint Source Channel Coding for Textual Semantic Communication

The Space-Air-Ground-Sea integrated network calls for more robust and se...
research
11/06/2019

Enriching Conversation Context in Retrieval-based Chatbots

Work on retrieval-based chatbots, like most sequence pair matching tasks...

Please sign up or login with your details

Forgot password? Click here to reset