One Adapter for All Programming Languages? Adapter Tuning for Code Search and Summarization

03/28/2023
by   Deze Wang, et al.
0

As pre-trained models automate many code intelligence tasks, a widely used paradigm is to fine-tune a model on the task dataset for each programming language. A recent study reported that multilingual fine-tuning benefits a range of tasks and models. However, we find that multilingual fine-tuning leads to performance degradation on recent models UniXcoder and CodeT5. To alleviate the potentially catastrophic forgetting issue in multilingual models, we fix all pre-trained model parameters, insert the parameter-efficient structure adapter, and fine-tune it. Updating only 0.6% of the overall parameters compared to full-model fine-tuning for each programming language, adapter tuning yields consistent improvements on code search and summarization tasks, achieving state-of-the-art results. In addition, we experimentally show its effectiveness in cross-lingual and low-resource scenarios. Multilingual fine-tuning with 200 samples per programming language approaches the results fine-tuned with the entire dataset on code summarization. Our experiments on three probing tasks show that adapter tuning significantly outperforms full-model fine-tuning and effectively overcomes catastrophic forgetting.

READ FULL TEXT

page 1

page 5

page 6

research
04/05/2022

On the Transferability of Pre-trained Language Models for Low-Resource Programming Languages

A recent study by Ahmed and Devanbu reported that using a corpus of code...
research
05/30/2023

AdapterEM: Pre-trained Language Model Adaptation for Generalized Entity Matching using Adapter-tuning

Entity Matching (EM) involves identifying different data representations...
research
05/19/2023

Analyzing and Reducing the Performance Gap in Cross-Lingual Transfer with Fine-tuning Slow and Fast

Existing research has shown that a multilingual pre-trained language mod...
research
12/12/2022

Searching for Effective Multilingual Fine-Tuning Methods: A Case Study in Summarization

Recently, a large number of tuning strategies have been proposed to adap...
research
05/12/2023

Model-based Programming: Redefining the Atomic Unit of Programming for the Deep Learning Era

This paper introduces and explores a new programming paradigm, Model-bas...
research
07/15/2023

Multilingual Adapter-based Knowledge Aggregation on Code Summarization for Low-Resource Languages

Multilingual fine-tuning (of a multilingual Pre-trained Language Model) ...
research
11/30/2022

MSV Challenge 2022: NPU-HC Speaker Verification System for Low-resource Indian Languages

This report describes the NPU-HC speaker verification system submitted t...

Please sign up or login with your details

Forgot password? Click here to reset