To Tune or Not To Tune? How About the Best of Both Worlds?

07/09/2019
by   Ran Wang, et al.
0

The introduction of pre-trained language models has revolutionized natural language research communities. However, researchers still know relatively little regarding their theoretical and empirical properties. In this regard, Peters et al. perform several experiments which demonstrate that it is better to adapt BERT with a light-weight task-specific head, rather than building a complex one on top of the pre-trained language model, and freeze the parameters in the said language model. However, there is another option to adopt. In this paper, we propose a new adaptation method which we first train the task model with the BERT parameters frozen and then fine-tune the entire model together. Our experimental results show that our model adaptation method can achieve 4.7 accuracy improvement in semantic similarity task, 0.99 sequence labeling task and 0.72 classification task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/06/2023

Spanish Pre-trained BERT Model and Evaluation Data

The Spanish language is one of the top 5 spoken languages in the world. ...
research
05/20/2019

Enriching Pre-trained Language Model with Entity Information for Relation Classification

Relation classification is an important NLP task to extract relations be...
research
02/24/2021

From Universal Language Model to Downstream Task: Improving RoBERTa-Based Vietnamese Hate Speech Detection

Natural language processing is a fast-growing field of artificial intell...
research
10/14/2021

Plug-Tagger: A Pluggable Sequence Labeling Framework Using Language Models

Plug-and-play functionality allows deep learning models to adapt well to...
research
12/21/2021

DB-BERT: a Database Tuning Tool that "Reads the Manual"

DB-BERT is a database tuning tool that exploits information gained via n...
research
06/30/2023

Ticket-BERT: Labeling Incident Management Tickets with Language Models

An essential aspect of prioritizing incident tickets for resolution is e...
research
07/06/2019

Applying a Pre-trained Language Model to Spanish Twitter Humor Prediction

Our entry into the HAHA 2019 Challenge placed 3^rd in the classification...

Please sign up or login with your details

Forgot password? Click here to reset