BERT-Based Arabic Social Media Author Profiling

09/09/2019
by   Chiyu Zhang, et al.
0

We report our models for detecting age, language variety, and gender from social media data in the context of the Arabic author profiling and deception detection shared task (APDA). We build simple models based on pre-trained bidirectional encoders from transformers (BERT). We first fine-tune the pre-trained BERT model on each of the three datasets with shared task released data. Then we augment shared task data with in-house data for gender and dialect, showing the utility of augmenting training data. Our best models on the shared task test data are acquired with a majority voting of various BERT models trained under different data conditions. We acquire 54.72 age, 93.75 the three tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2019

BERT-Based Arabic Social Media AuthorProfiling

We report our models for detecting age, language variety, and gender fro...
research
11/02/2019

Sentence-Level BERT and Multi-Task Learning of Age and Gender in Social Media

Social media currently provide a window on our lives, making it possible...
research
09/08/2019

Multi-Task Bidirectional Transformer Representations for Irony Detection

Supervised deep learning requires large amounts of training data. In the...
research
12/30/2019

AraNet: A Deep Learning Toolkit for Arabic Social Media

We describe AraNet, a collection of deep learning Arabic social media pr...
research
02/21/2021

Pre-Training BERT on Arabic Tweets: Practical Considerations

Pretraining Bidirectional Encoder Representations from Transformers (BER...
research
09/29/2020

Gender prediction using limited Twitter Data

Transformer models have shown impressive performance on a variety of NLP...
research
11/13/2021

SocialBERT – Transformers for Online SocialNetwork Language Modelling

The ubiquity of the contemporary language understanding tasks gives rele...

Please sign up or login with your details

Forgot password? Click here to reset