Improving Meta-learning for Low-resource Text Classification and Generation via Memory Imitation

03/22/2022
by   Yingxiu Zhao, et al.
5

Building models of natural language processing (NLP) is challenging in low-resource scenarios where only limited data are available. Optimization-based meta-learning algorithms achieve promising results in low-resource scenarios by adapting a well-generalized model initialization to handle new tasks. Nonetheless, these approaches suffer from the memorization overfitting issue, where the model tends to memorize the meta-training tasks while ignoring support sets when adapting to new tasks. To address this issue, we propose a memory imitation meta-learning (MemIML) method that enhances the model's reliance on support sets for task adaptation. Specifically, we introduce a task-specific memory module to store support set information and construct an imitation module to force query sets to imitate the behaviors of some representative support-set samples stored in the memory. A theoretical analysis is provided to prove the effectiveness of our method, and empirical results also demonstrate that our method outperforms competitive baselines on both text classification and generation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2019

Meta-Learning for Low-resource Natural Language Generation in Task-oriented Dialogue Systems

Natural language generation (NLG) is an essential component of task-orie...
research
07/26/2020

Don't Overlook the Support Set: Towards Improving Generalization in Meta-learning

Meta-learning has proven to be a powerful paradigm for transferring the ...
research
09/10/2023

Retrieval-Augmented Meta Learning for Low-Resource Text Classification

Meta learning have achieved promising performance in low-resource text c...
research
10/25/2022

Leveraging Open Data and Task Augmentation to Automated Behavioral Coding of Psychotherapy Conversations in Low-Resource Scenarios

In psychotherapy interactions, the quality of a session is assessed by c...
research
05/24/2020

When does MAML Work the Best? An Empirical Study on Model-Agnostic Meta-Learning in NLP Applications

Model-Agnostic Meta-Learning (MAML), a model-agnostic meta-learning meth...
research
06/03/2018

On the Importance of Attention in Meta-Learning for Few-Shot Text Classification

Current deep learning based text classification methods are limited by t...
research
06/25/2020

Stochastic Subset Selection

Current machine learning algorithms are designed to work with huge volum...

Please sign up or login with your details

Forgot password? Click here to reset