Vision Transformer Adapters for Generalizable Multitask Learning

08/23/2023
by   Deblina Bhattacharjee, et al.
0

We introduce the first multitasking vision transformer adapters that learn generalizable task affinities which can be applied to novel tasks and domains. Integrated into an off-the-shelf vision transformer backbone, our adapters can simultaneously solve multiple dense vision tasks in a parameter-efficient manner, unlike existing multitasking transformers that are parametrically expensive. In contrast to concurrent methods, we do not require retraining or fine-tuning whenever a new task or domain is added. We introduce a task-adapted attention mechanism within our adapter framework that combines gradient-based task similarities with attention-based ones. The learned task affinities generalize to the following settings: zero-shot task transfer, unsupervised domain adaptation, and generalization without fine-tuning to novel domains. We demonstrate that our approach outperforms not only the existing convolutional neural network-based multitasking methods but also the vision transformer-based ones. Our project page is at <https://ivrl.github.io/VTAGML>.

READ FULL TEXT

page 4

page 7

page 15

page 17

page 18

page 19

page 20

page 21

research
12/21/2022

SPT: Semi-Parametric Prompt Tuning for Multitask Prompted Learning

Pre-trained large language models can efficiently interpolate human-writ...
research
10/07/2022

Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks

Adapting large-scale pretrained models to various downstream tasks via f...
research
01/06/2023

Exploring Efficient Few-shot Adaptation for Vision Transformers

The task of Few-shot Learning (FSL) aims to do the inference on novel ca...
research
07/25/2023

E^2VPT: An Effective and Efficient Approach for Visual Prompt Tuning

As the size of transformer-based models continues to grow, fine-tuning t...
research
07/16/2023

Dense Multitask Learning to Reconfigure Comics

In this paper, we develop a MultiTask Learning (MTL) model to achieve de...
research
04/17/2023

Hyper-Decision Transformer for Efficient Online Policy Adaptation

Decision Transformers (DT) have demonstrated strong performances in offl...
research
06/23/2021

IA-RED^2: Interpretability-Aware Redundancy Reduction for Vision Transformers

The self-attention-based model, transformer, is recently becoming the le...

Please sign up or login with your details

Forgot password? Click here to reset