Unsupervised Transfer Learning via BERT Neuron Selection

12/10/2019
by   Mehrdad Valipour, et al.
0

Recent advancements in language representation models such as BERT have led to a rapid improvement in numerous natural language processing tasks. However, language models usually consist of a few hundred million trainable parameters with embedding space distributed across multiple layers, thus making them challenging to be fine-tuned for a specific task or to be transferred to a new domain. To determine whether there are task-specific neurons that can be exploited for unsupervised transfer learning, we introduce a method for selecting the most important neurons to solve a specific classification task. This algorithm is further extended to multi-source transfer learning by computing the importance of neurons for several single-source transfer learning scenarios between different subsets of data sources. Besides, a task-specific fingerprint for each data source is obtained based on the percentage of the selected neurons in each layer. We perform extensive experiments in unsupervised transfer learning for sentiment analysis, natural language inference and sentence similarity, and compare our results with the existing literature and baselines. Significantly, we found that the source and target data sources with higher degrees of similarity between their task-specific fingerprints demonstrate a better transferability property. We conclude that our method can lead to better performance using just a few hundred task-specific and interpretable neurons.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2019

An Investigation of Transfer Learning-Based Sentiment Analysis in Japanese

Text classification approaches have usually required task-specific model...
research
10/15/2021

SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer

As pre-trained language models have gotten larger, there has been growin...
research
04/26/2021

Evaluating the Values of Sources in Transfer Learning

Transfer learning that adapts a model trained on data-rich sources to lo...
research
04/20/2022

Deep transfer learning for partial differential equations under conditional shift with DeepONet

Traditional machine learning algorithms are designed to learn in isolati...
research
01/10/2023

Cross-Model Comparative Loss for Enhancing Neuronal Utility in Language Understanding

Current natural language understanding (NLU) models have been continuous...
research
10/21/2022

Efficiently Tuned Parameters are Task Embeddings

Intermediate-task transfer can benefit a wide range of NLP tasks with pr...
research
01/05/2021

COVID-19: Comparative Analysis of Methods for Identifying Articles Related to Therapeutics and Vaccines without Using Labeled Data

Here we proposed an approach to analyze text classification methods base...

Please sign up or login with your details

Forgot password? Click here to reset