Scalable Weight Reparametrization for Efficient Transfer Learning

02/26/2023
by   Byeonggeun Kim, et al.
0

This paper proposes a novel, efficient transfer learning method, called Scalable Weight Reparametrization (SWR) that is efficient and effective for multiple downstream tasks. Efficient transfer learning involves utilizing a pre-trained model trained on a larger dataset and repurposing it for downstream tasks with the aim of maximizing the reuse of the pre-trained model. However, previous works have led to an increase in updated parameters and task-specific modules, resulting in more computations, especially for tiny models. Additionally, there has been no practical consideration for controlling the number of updated parameters. To address these issues, we suggest learning a policy network that can decide where to reparametrize the pre-trained model, while adhering to a given constraint for the number of updated parameters. The policy network is only used during the transfer learning process and not afterward. As a result, our approach attains state-of-the-art performance in a proposed multi-lingual keyword spotting and a standard benchmark, ImageNet-to-Sketch, while requiring zero additional computations and significantly fewer additional parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/14/2023

Revisit Parameter-Efficient Transfer Learning: A Two-Stage Paradigm

Parameter-Efficient Transfer Learning (PETL) aims at efficiently adaptin...
research
06/27/2022

ST-Adapter: Parameter-Efficient Image-to-Video Transfer Learning for Action Recognition

Capitalizing on large pre-trained models for various downstream tasks of...
research
09/04/2023

Parameter and Computation Efficient Transfer Learning for Vision-Language Pre-trained Models

With ever increasing parameters and computation, vision-language pre-tra...
research
03/26/2023

Δ-Networks for Efficient Model Patching

Models pre-trained on large-scale datasets are often finetuned to suppor...
research
08/19/2023

Disposable Transfer Learning for Selective Source Task Unlearning

Transfer learning is widely used for training deep neural networks (DNN)...
research
03/17/2023

A Unified Continual Learning Framework with General Parameter-Efficient Tuning

The "pre-training → downstream adaptation" presents both new opportuniti...
research
07/14/2023

MGit: A Model Versioning and Management System

Models derived from other models are extremely common in machine learnin...

Please sign up or login with your details

Forgot password? Click here to reset