BERT Fine-Tuning for Sentiment Analysis on Indonesian Mobile Apps Reviews

07/14/2021
by   Kuncahyo Setyo Nugroho, et al.
0

User reviews have an essential role in the success of the developed mobile apps. User reviews in the textual form are unstructured data, creating a very high complexity when processed for sentiment analysis. Previous approaches that have been used often ignore the context of reviews. In addition, the relatively small data makes the model overfitting. A new approach, BERT, has been introduced as a transfer learning model with a pre-trained model that has previously been trained to have a better context representation. This study examines the effectiveness of fine-tuning BERT for sentiment analysis using two different pre-trained models. Besides the multilingual pre-trained model, we use the pre-trained model that only has been trained in Indonesian. The dataset used is Indonesian user reviews of the ten best apps in 2020 in Google Play sites. We also perform hyper-parameter tuning to find the optimum trained model. Two training data labeling approaches were also tested to determine the effectiveness of the model, which is score-based and lexicon-based. The experimental results show that pre-trained models trained in Indonesian have better average accuracy on lexicon-based data. The pre-trained Indonesian model highest accuracy is 84 These results are better than all of the machine learning and multilingual pre-trained models.

READ FULL TEXT
research
10/11/2022

Transfer Learning with Joint Fine-Tuning for Multimodal Sentiment Analysis

Most existing methods focus on sentiment analysis of textual data. Howev...
research
10/31/2020

Understanding Pre-trained BERT for Aspect-based Sentiment Analysis

This paper analyzes the pre-trained hidden representations learned from ...
research
05/13/2022

Revisiting the Updates of a Pre-trained Model for Few-shot Learning

Most of the recent few-shot learning algorithms are based on transfer le...
research
08/11/2023

Tweet Sentiment Extraction using Viterbi Algorithm with Transfer Learning

Tweet sentiment extraction extracts the most significant portion of the ...
research
09/11/2020

A Comparison of LSTM and BERT for Small Corpus

Recent advancements in the NLP field showed that transfer learning helps...
research
10/21/2021

Fast Model Editing at Scale

While large pre-trained models have enabled impressive results on a vari...
research
03/22/2023

Generate labeled training data using Prompt Programming and GPT-3. An example of Big Five Personality Classification

We generated 25000 conversations labeled with Big Five Personality trait...

Please sign up or login with your details

Forgot password? Click here to reset