AutoPEFT: Automatic Configuration Search for Parameter-Efficient Fine-Tuning

01/28/2023
by   Han Zhou, et al.
0

Large pretrained language models have been widely used in downstream NLP tasks via task-specific fine-tuning. Recently, an array of Parameter-Efficient Fine-Tuning (PEFT) methods have also achieved strong task performance while updating a much smaller number of parameters compared to full model tuning. However, it is non-trivial to make informed per-task design choices (i.e., to create PEFT configurations) concerning the selection of PEFT architectures and modules, the number of tunable parameters, and even the layers in which the PEFT modules are inserted. Consequently, it is highly likely that the current, manually set PEFT configurations might be suboptimal for many tasks from the perspective of the performance-to-efficiency trade-off. To address the core question of the PEFT configuration selection that aims to control and maximise the balance between performance and parameter efficiency, we first define a rich configuration search space spanning multiple representative PEFT modules along with finer-grained configuration decisions over the modules (e.g., parameter budget, insertion layer). We then propose AutoPEFT, a novel framework to traverse this configuration space: it automatically configures multiple PEFT modules via high-dimensional Bayesian optimisation. We show the resource scalability and task transferability of AutoPEFT-found configurations, outperforming existing PEFT methods on average on the standard GLUE benchmark while conducting the configuration search on a single task. The per-task AutoPEFT-based configuration search even outperforms full-model fine-tuning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/06/2021

On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation

Adapter-based tuning has recently arisen as an alternative to fine-tunin...
research
10/29/2019

Real-time Bidding campaigns optimization using attribute selection

Real-Time Bidding is nowadays one of the most promising systems in the o...
research
06/05/2019

Revisiting Hyper-Parameter Tuning for Search-based Test Data Generation

Search-based software testing (SBST) has been studied a lot in the liter...
research
08/29/2022

Resolving inconsistencies of runtime configuration changes through change propagation and adjustments

A system configuration may be modified at runtime to adapt the system to...
research
05/20/2022

AutoFedNLP: An efficient FedNLP framework

Transformer-based pre-trained models have revolutionized NLP for superio...
research
07/01/2020

Few-shots Parameter Tuning via Co-evolution

Generalization, i.e., the ability of addressing problem instances that a...
research
06/15/2022

Sparse Structure Search for Parameter-Efficient Tuning

Adapting large pre-trained models (PTMs) through fine-tuning imposes pro...

Please sign up or login with your details

Forgot password? Click here to reset