Meta-Tsallis-Entropy Minimization: A New Self-Training Approach for Domain Adaptation on Text Classification

08/04/2023
by   Menglong Lu, et al.
0

Text classification is a fundamental task for natural language processing, and adapting text classification models across domains has broad applications. Self-training generates pseudo-examples from the model's predictions and iteratively trains on the pseudo-examples, i.e., minimizes the loss on the source domain and the Gibbs entropy on the target domain. However, Gibbs entropy is sensitive to prediction errors, and thus, self-training tends to fail when the domain shift is large. In this paper, we propose Meta-Tsallis Entropy minimization (MTEM), which applies a meta-learning algorithm to optimize the instance adaptive Tsallis entropy on the target domain. To reduce the computation cost of MTEM, we propose an approximation technique to approximate the Second-order derivation involved in the meta-learning. To efficiently generate pseudo labels, we propose an annealing sampling mechanism for exploring the model's prediction probability. Theoretically, we prove the convergence of the meta-learning algorithm in MTEM and analyze the effectiveness of MTEM in achieving domain adaptation. Experimentally, MTEM improves the adaptation performance of BERT with an average of 4 percent on the benchmark dataset.

READ FULL TEXT

page 4

page 8

research
07/26/2021

Meta-Learning Adversarial Domain Adaptation Network for Few-Shot Text Classification

Meta-learning has emerged as a trending technique to tackle few-shot tex...
research
12/07/2022

Reconciling a Centroid-Hypothesis Conflict in Source-Free Domain Adaptation

Source-free domain adaptation (SFDA) aims to transfer knowledge learned ...
research
09/24/2020

Feature Adaptation of Pre-Trained Language Models across Languages and Domains for Text Classification

Adapting pre-trained language models (PrLMs) (e.g., BERT) to new domains...
research
08/26/2022

Constraining Pseudo-label in Self-training Unsupervised Domain Adaptation with Energy-based Model

Deep learning is usually data starved, and the unsupervised domain adapt...
research
01/18/2021

Model Compression for Domain Adaptation through Causal Effect Estimation

Recent improvements in the predictive quality of natural language proces...
research
08/24/2021

Meta Self-Learning for Multi-Source Domain Adaptation: A Benchmark

In recent years, deep learning-based methods have shown promising result...
research
12/12/2022

Selective classification using a robust meta-learning approach

Selective classification involves identifying the subset of test samples...

Please sign up or login with your details

Forgot password? Click here to reset