Efficiently Tuned Parameters are Task Embeddings

10/21/2022
by   Wangchunshu Zhou, et al.
4

Intermediate-task transfer can benefit a wide range of NLP tasks with properly selected source datasets. However, it is computationally infeasible to experiment with all intermediate transfer combinations, making choosing a useful source task a challenging problem. In this paper, we anticipate that task-specific parameters updated in parameter-efficient tuning methods are likely to encode task-specific information. Therefore, such parameters can be predictive for inter-task transferability. Thus, we propose to exploit these efficiently tuned parameters as off-the-shelf task embeddings for the efficient selection of source datasets for intermediate-task transfer. We experiment with 11 text classification tasks and 11 question answering tasks. Experimental results show that our approach can consistently outperform existing inter-task transferability prediction methods while being conceptually simple and computationally efficient. Our analysis also reveals that the ability of efficiently tuned parameters on transferability prediction is disentangled with their in-task performance. This allows us to use parameters from early checkpoints as task embeddings to further improve efficiency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/15/2021

SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer

As pre-trained language models have gotten larger, there has been growin...
research
04/16/2021

What to Pre-Train on? Efficient Intermediate Task Selection

Intermediate task fine-tuning has been shown to culminate in large trans...
research
05/02/2020

Exploring and Predicting Transferability across NLP Tasks

Recent advances in NLP demonstrate the effectiveness of training large-s...
research
12/10/2019

Unsupervised Transfer Learning via BERT Neuron Selection

Recent advancements in language representation models such as BERT have ...
research
07/06/2023

Text Alignment Is An Efficient Unified Model for Massive NLP Tasks

Large language models (LLMs), typically designed as a function of next-w...
research
05/12/2023

To transfer or not transfer: Unified transferability metric and analysis

In transfer learning, transferability is one of the most fundamental pro...
research
09/24/2018

Joint Multitask Learning for Community Question Answering Using Task-Specific Embeddings

We address jointly two important tasks for Question Answering in communi...

Please sign up or login with your details

Forgot password? Click here to reset