Prior Omission of Dissimilar Source Domain(s) for Cost-Effective Few-Shot Learning

09/11/2021
by   Zezhong Wang, et al.
0

Few-shot slot tagging is an emerging research topic in the field of Natural Language Understanding (NLU). With sufficient annotated data from source domains, the key challenge is how to train and adapt the model to another target domain which only has few labels. Conventional few-shot approaches use all the data from the source domains without considering inter-domain relations and implicitly assume each sample in the domain contributes equally. However, our experiments show that the data distribution bias among different domains will significantly affect the adaption performance. Moreover, transferring knowledge from dissimilar domains will even introduce some extra noises so that affect the performance of models. To tackle this problem, we propose an effective similarity-based method to select data from the source domains. In addition, we propose a Shared-Private Network (SP-Net) for the few-shot slot tagging task. The words from the same class would have some shared features. We extract those shared features from the limited annotated data on the target domain and merge them together as the label embedding to help us predict other unlabelled data on the target domain. The experiment shows that our method outperforms the state-of-the-art approaches with fewer source data. The result also proves that some training data from dissimilar sources are redundant and even negative for the adaption.

READ FULL TEXT
research
10/07/2021

Bridge to Target Domain by Prototypical Contrastive Learning and Label Confusion: Re-explore Zero-Shot Learning for Slot Filling

Zero-shot cross-domain slot filling alleviates the data dependence in th...
research
05/17/2018

Cross-domain attribute representation based on convolutional neural network

In the problem of domain transfer learning, we learn a model for the pre...
research
09/21/2020

Vector Projection Network for Few-shot Slot Tagging in Natural Language Understanding

Few-shot slot tagging becomes appealing for rapid domain transfer and ad...
research
10/11/2022

TGDM: Target Guided Dynamic Mixup for Cross-Domain Few-Shot Learning

Given sufficient training data on the source domain, cross-domain few-sh...
research
12/09/2021

Few-Shot NLU with Vector Projection Distance and Abstract Triangular CRF

Data sparsity problem is a key challenge of Natural Language Understandi...
research
04/02/2022

Inverse is Better! Fast and Accurate Prompt for Few-shot Slot Tagging

Prompting methods recently achieve impressive success in few-shot learni...
research
03/03/2021

Few-shot Learning for Slot Tagging with Attentive Relational Network

Metric-based learning is a well-known family of methods for few-shot lea...

Please sign up or login with your details

Forgot password? Click here to reset