Prompt Guided Transformer for Multi-Task Dense Prediction

07/28/2023
by   Yuxiang Lu, et al.
0

Task-conditional architecture offers advantage in parameter efficiency but falls short in performance compared to state-of-the-art multi-decoder methods. How to trade off performance and model parameters is an important and difficult problem. In this paper, we introduce a simple and lightweight task-conditional model called Prompt Guided Transformer (PGT) to optimize this challenge. Our approach designs a Prompt-conditioned Transformer block, which incorporates task-specific prompts in the self-attention mechanism to achieve global dependency modeling and parameter-efficient feature adaptation across multiple tasks. This block is integrated into both the shared encoder and decoder, enhancing the capture of intra- and inter-task features. Moreover, we design a lightweight decoder to further reduce parameter usage, which accounts for only 2.7 dense prediction benchmarks, PASCAL-Context and NYUD-v2, demonstrate that our approach achieves state-of-the-art results among task-conditional methods while using fewer parameters, and maintains a significant balance between performance and parameter size.

READ FULL TEXT

page 3

page 7

page 8

page 10

research
08/10/2023

Deformable Mixer Transformer with Gating for Multi-Task Learning of Dense Prediction

CNNs and Transformers have their own advantages and both have been widel...
research
01/09/2023

DeMT: Deformable Mixer Transformer for Multi-Task Learning of Dense Prediction

Convolution neural networks (CNNs) and Transformers have their own advan...
research
03/15/2022

Inverted Pyramid Multi-task Transformer for Dense Scene Understanding

Multi-task dense scene understanding is a thriving research domain that ...
research
06/12/2023

CD-CTFM: A Lightweight CNN-Transformer Network for Remote Sensing Cloud Detection Fusing Multiscale Features

Clouds in remote sensing images inevitably affect information extraction...
research
06/01/2020

Residual Squeeze-and-Excitation Network for Fast Image Deraining

Image deraining is an important image processing task as rain streaks no...
research
03/21/2023

LiDARFormer: A Unified Transformer-based Multi-task Network for LiDAR Perception

There is a recent trend in the LiDAR perception field towards unifying m...
research
10/10/2022

LMQFormer: A Laplace-Prior-Guided Mask Query Transformer for Lightweight Snow Removal

Snow removal aims to locate snow areas and recover clean images without ...

Please sign up or login with your details

Forgot password? Click here to reset