Universal Spam Detection using Transfer Learning of BERT Model

02/07/2022
by   Vijay Srinivas Tida, et al.
0

Deep learning transformer models become important by training on text data based on self-attention mechanisms. This manuscript demonstrated a novel universal spam detection model using pre-trained Google's Bidirectional Encoder Representations from Transformers (BERT) base uncased models with four datasets by efficiently classifying ham or spam emails in real-time scenarios. Different methods for Enron, Spamassain, Lingspam, and Spamtext message classification datasets, were used to train models individually in which a single model was obtained with acceptable performance on four datasets. The Universal Spam Detection Model (USDM) was trained with four datasets and leveraged hyperparameters from each model. The combined model was finetuned with the same hyperparameters from these four models separately. When each model using its corresponding dataset, an F1-score is at and above 0.9 in individual models. An overall accuracy reached 97 implications were discussed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2019

Multi-Task Bidirectional Transformer Representations for Irony Detection

Supervised deep learning requires large amounts of training data. In the...
research
03/13/2023

Transformer-based approaches to Sentiment Detection

The use of transfer learning methods is largely responsible for the pres...
research
07/02/2020

Bidirectional Encoder Representations from Transformers (BERT): A sentiment analysis odyssey

The purpose of the study is to investigate the relative effectiveness of...
research
01/18/2022

Hierarchical Neural Network Approaches for Long Document Classification

Text classification algorithms investigate the intricate relationships b...
research
08/06/2021

Offensive Language and Hate Speech Detection with Deep Learning and Transfer Learning

Toxic online speech has become a crucial problem nowadays due to an expo...
research
12/14/2019

BERTQA – Attention on Steroids

In this work, we extend the Bidirectional Encoder Representations from T...
research
02/25/2019

BUT-FIT at SemEval-2019 Task 7: Determining the Rumour Stance with Pre-Trained Deep Bidirectional Transformers

This paper describes our system submitted to SemEval 2019 Task 7: Rumour...

Please sign up or login with your details

Forgot password? Click here to reset