LadRa-Net: Locally-Aware Dynamic Re-read Attention Net for Sentence Semantic Matching

08/06/2021
by   Kun Zhang, et al.
6

Sentence semantic matching requires an agent to determine the semantic relation between two sentences, which is widely used in various natural language tasks, such as Natural Language Inference (NLI), Paraphrase Identification (PI), and so on. Much recent progress has been made in this area, especially attention-based methods and pre-trained language model based methods. However, most of these methods focus on all the important parts in sentences in a static way and only emphasize how important the words are to the query, inhibiting the ability of attention mechanism. In order to overcome this problem and boost the performance of attention mechanism, we propose a novel dynamic re-read attention, which can pay close attention to one small region of sentences at each step and re-read the important parts for better sentence representations. Based on this attention variation, we develop a novel Dynamic Re-read Network (DRr-Net) for sentence semantic matching. Moreover, selecting one small region in dynamic re-read attention seems insufficient for sentence semantics, and employing pre-trained language models as input encoders will introduce incomplete and fragile representation problems. To this end, we extend DRrNet to Locally-Aware Dynamic Re-read Attention Net (LadRa-Net), in which local structure of sentences is employed to alleviate the shortcoming of Byte-Pair Encoding (BPE) in pre-trained language models and boost the performance of dynamic reread attention. Extensive experiments on two popular sentence semantic matching tasks demonstrate that DRr-Net can significantly improve the performance of sentence semantic matching. Meanwhile, LadRa-Net is able to achieve better performance by considering the local structures of sentences. In addition, it is exceedingly interesting that some discoveries in our experiments are consistent with some findings of psychological research.

READ FULL TEXT

page 1

page 14

research
06/09/2021

DGA-Net Dynamic Gaussian Attention Network for Sentence Semantic Matching

Sentence semantic matching requires an agent to determine the semantic r...
research
12/16/2020

R^2-Net: Relation of Relation Learning Network for Sentence Semantic Matching

Sentence semantic matching is one of the fundamental tasks in natural la...
research
03/30/2022

Incorporating Dynamic Semantics into Pre-Trained Language Model for Aspect-based Sentiment Analysis

Aspect-based sentiment analysis (ABSA) predicts sentiment polarity towar...
research
10/07/2022

DABERT: Dual Attention Enhanced BERT for Semantic Matching

Transformer-based pre-trained language models such as BERT have achieved...
research
10/15/2020

Inducing Alignment Structure with Gated Graph Attention Networks for Sentence Matching

Sentence matching is a fundamental task of natural language processing w...
research
10/31/2019

Positional Attention-based Frame Identification with BERT: A Deep Learning Approach to Target Disambiguation and Semantic Frame Selection

Semantic parsing is the task of transforming sentences from natural lang...
research
07/23/2022

Chunk-aware Alignment and Lexical Constraint for Visual Entailment with Natural Language Explanations

Visual Entailment with natural language explanations aims to infer the r...

Please sign up or login with your details

Forgot password? Click here to reset