On Dataset Transferability in Active Learning for Transformers

05/16/2023
by   Fran Jelenić, et al.
0

Active learning (AL) aims to reduce labeling costs by querying the examples most beneficial for model learning. While the effectiveness of AL for fine-tuning transformer-based pre-trained language models (PLMs) has been demonstrated, it is less clear to what extent the AL gains obtained with one model transfer to others. We consider the problem of transferability of actively acquired datasets in text classification and investigate whether AL gains persist when a dataset built using AL coupled with a specific PLM is used to train a different PLM. We link the AL dataset transferability to the similarity of instances queried by the different PLMs and show that AL methods with similar acquisition sequences produce highly transferable datasets regardless of the models used. Additionally, we show that the similarity of acquisition sequences is influenced more by the choice of the AL method than the choice of the model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/20/2022

Smooth Sailing: Improving Active Learning for Pre-trained Language Models with Representation Smoothness Analysis

Developed as a solution to a practical need, active learning (AL) method...
research
05/23/2023

Parameter-Efficient Language Model Tuning with Active Learning in Low-Resource Settings

Pre-trained language models (PLMs) have ignited a surge in demand for ef...
research
07/12/2018

How transferable are the datasets collected by active learners?

Active learning is a widely-used training strategy for maximizing predic...
research
09/12/2023

Annotating Data for Fine-Tuning a Neural Ranker? Current Active Learning Strategies are not Better than Random Selection

Search methods based on Pretrained Language Models (PLM) have demonstrat...
research
05/07/2022

Towards Computationally Feasible Deep Active Learning

Active learning (AL) is a prominent technique for reducing the annotatio...
research
10/06/2022

To Softmax, or not to Softmax: that is the question when applying Active Learning for Transformer Models

Despite achieving state-of-the-art results in nearly all Natural Languag...
research
11/08/2022

Active Learning with Tabular Language Models

Despite recent advancements in tabular language model research, real-wor...

Please sign up or login with your details

Forgot password? Click here to reset