VIRT: Improving Representation-based Models for Text Matching through Virtual Interaction

12/08/2021
by   Dan Li, et al.
0

With the booming of pre-trained transformers, remarkable progress has been made on textual pair modeling to support relevant natural language applications. Two lines of approaches are developed for text matching: interaction-based models performing full interactions over the textual pair, and representation-based models encoding the pair independently with siamese encoders. The former achieves compelling performance due to its deep interaction modeling ability, yet with a sacrifice in inference latency. The latter is efficient and widely adopted for practical use, however, suffers from severe performance degradation due to the lack of interactions. Though some prior works attempt to integrate interactive knowledge into representation-based models, considering the computational cost, they only perform late interaction or knowledge transferring at the top layers. Interactive information in the lower layers is still missing, which limits the performance of representation-based solutions. To remedy this, we propose a novel Virtual InteRacTion mechanism, termed as VIRT, to enable full and deep interaction modeling in representation-based models without actual inference computations. Concretely, VIRT asks representation-based encoders to conduct virtual interactions to mimic the behaviors as interaction-based models do. In addition, the knowledge distilled from interaction-based encoders is taken as supervised signals to promise the effectiveness of virtual interactions. Since virtual interactions only happen at the training stage, VIRT would not increase the inference cost. Furthermore, we design a VIRT-adapted late interaction strategy to fully utilize the learned virtual interactive knowledge.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2019

Enriching Conversation Context in Retrieval-based Chatbots

Work on retrieval-based chatbots, like most sequence pair matching tasks...
research
09/08/2021

Unsupervised Pre-training with Structured Knowledge for Improving Natural Language Inference

While recent research on natural language inference has considerably ben...
research
06/04/2023

I^3 Retriever: Incorporating Implicit Interaction in Pre-trained Language Models for Passage Retrieval

Passage retrieval is a fundamental task in many information systems, suc...
research
10/11/2022

Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair Modeling

Transformer-based models have achieved great success on sentence pair mo...
research
08/22/2022

Improving Knowledge-aware Recommendation with Multi-level Interactive Contrastive Learning

Incorporating Knowledge Graphs (KG) into recommeder system has attracted...
research
09/08/2021

Machine learning modeling of family wide enzyme-substrate specificity screens

Biocatalysis is a promising approach to sustainably synthesize pharmaceu...
research
04/09/2020

Interactions in information spread: quantification and interpretation using stochastic block models

In most real-world applications, it is seldom the case that a given obse...

Please sign up or login with your details

Forgot password? Click here to reset