Learning for Expressive Task-Related Sentence Representations

05/24/2022
by   Xueying Bai, et al.
0

NLP models learn sentence representations for downstream tasks by tuning a model which is pre-trained by masked language modeling. However, after tuning, the learned sentence representations may be skewed heavily toward label space and thus are not expressive enough to represent whole samples, which should contain task-related information of both sentence inputs and labels. In this work, we learn expressive sentence representations for supervised tasks which (1). contain task-related information in the sentence inputs, and (2). enable correct label predictions. To achieve this goal, we first propose a new objective which explicitly points out the label token space in the input, and predicts categories of labels via an added [MASK] token. This objective encourages fusing the semantic information of both the label and sentence. Then we develop a neighbor attention module, added on a frozen pre-trained model, to build connections between label/sentence tokens via their neighbors. The propagation can be further guided by the regularization on neighborhood representations to encourage expressiveness. Experimental results show that, despite tuning only 5 our model can achieve classification results comparable to the SOTA while maintaining strong expressiveness as well.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/07/2022

Pre-trained Token-replaced Detection Model as Few-shot Learner

Pre-trained masked language models have demonstrated remarkable ability ...
research
11/14/2018

Jointly Learning to Label Sentences and Tokens

Learning to construct text representations in end-to-end systems can be ...
research
12/21/2020

Narrative Incoherence Detection

Motivated by the increasing popularity of intelligent editing assistant,...
research
06/14/2021

Exploiting Sentence-Level Representations for Passage Ranking

Recently, pre-trained contextual models, such as BERT, have shown to per...
research
10/28/2022

DiMBERT: Learning Vision-Language Grounded Representations with Disentangled Multimodal-Attention

Vision-and-language (V-L) tasks require the system to understand both vi...
research
05/15/2019

What do you learn from context? Probing for sentence structure in contextualized word representations

Contextualized representation models such as ELMo (Peters et al., 2018a)...
research
10/15/2021

Probing as Quantifying the Inductive Bias of Pre-trained Representations

Pre-trained contextual representations have led to dramatic performance ...

Please sign up or login with your details

Forgot password? Click here to reset