SubTuning: Efficient Finetuning for Multi-Task Learning

02/13/2023
by   Gal Kaplun, et al.
0

Finetuning a pretrained model has become a standard approach for training neural networks on novel tasks, resulting in fast convergence and improved performance. In this work, we study an alternative finetuning method, where instead of finetuning all the weights of the network, we only train a carefully chosen subset of layers, keeping the rest of the weights frozen at their initial (pretrained) values. We demonstrate that subset finetuning (or SubTuning) often achieves accuracy comparable to full finetuning of the model, and even surpasses the performance of full finetuning when training data is scarce. Therefore, SubTuning allows deploying new tasks at minimal computational cost, while enjoying the benefits of finetuning the entire model. This yields a simple and effective method for multi-task learning, where different tasks do not interfere with one another, and yet share most of the resources at inference time. We demonstrate the efficiency of SubTuning across multiple tasks, using different network architectures and pretraining methods.

READ FULL TEXT

page 1

page 15

research
04/16/2022

Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners

Traditional multi-task learning (MTL) methods use dense networks that us...
research
11/08/2019

Dynamic Multi-Task Learning for Face Recognition with Facial Expression

Benefiting from the joint learning of the multiple tasks in the deep mul...
research
05/03/2022

Adaptable Adapters

State-of-the-art pretrained NLP models contain a hundred million to tril...
research
07/21/2022

UFO: Unified Feature Optimization

This paper proposes a novel Unified Feature Optimization (UFO) paradigm ...
research
09/27/2021

Consistency Training of Multi-exit Architectures for Sensor Data

Deep neural networks have become larger over the years with increasing d...
research
03/10/2023

Adaptive Weight Assignment Scheme For Multi-task Learning

Deep learning based models are used regularly in every applications nowa...
research
05/28/2021

Efficient and robust multi-task learning in the brain with modular task primitives

In a real-world setting biological agents do not have infinite resources...

Please sign up or login with your details

Forgot password? Click here to reset