Sensitivity Analysis on Transferred Neural Architectures of BERT and GPT-2 for Financial Sentiment Analysis

07/07/2022
by   Tracy Qian, et al.
0

The explosion in novel NLP word embedding and deep learning techniques has induced significant endeavors into potential applications. One of these directions is in the financial sector. Although there is a lot of work done in state-of-the-art models like GPT and BERT, there are relatively few works on how well these methods perform through fine-tuning after being pre-trained, as well as info on how sensitive their parameters are. We investigate the performance and sensitivity of transferred neural architectures from pre-trained GPT-2 and BERT models. We test the fine-tuning performance based on freezing transformer layers, batch size, and learning rate. We find the parameters of BERT are hypersensitive to stochasticity in fine-tuning and that GPT-2 is more stable in such practice. It is also clear that the earlier layers of GPT-2 and BERT contain essential word pattern information that should be maintained.

READ FULL TEXT

page 7

page 8

page 9

research
11/20/2020

Fine-Tuning BERT for Sentiment Analysis of Vietnamese Reviews

Sentiment analysis is an important task in the field ofNature Language P...
research
10/26/2022

Exploring Robustness of Prefix Tuning in Noisy Data: A Case Study in Financial Sentiment Analysis

The invention of transformer-based models such as BERT, GPT, and RoBERTa...
research
07/25/2022

Fine-Tuning BERT for Automatic ADME Semantic Labeling in FDA Drug Labeling to Enhance Product-Specific Guidance Assessment

Product-specific guidances (PSGs) recommended by the United States Food ...
research
09/12/2020

Fine-tuning Pre-trained Contextual Embeddings for Citation Content Analysis in Scholarly Publication

Citation function and citation sentiment are two essential aspects of ci...
research
01/10/2022

TiltedBERT: Resource Adjustable Version of BERT

In this paper, we proposed a novel adjustable finetuning method that imp...
research
03/25/2022

A Comparative Evaluation Of Transformer Models For De-Identification Of Clinical Text Data

Objective: To comparatively evaluate several transformer model architect...
research
02/09/2020

Segmented Graph-Bert for Graph Instance Modeling

In graph instance representation learning, both the diverse graph instan...

Please sign up or login with your details

Forgot password? Click here to reset