KUISAIL at SemEval-2020 Task 12: BERT-CNN for Offensive Speech Identification in Social Media

07/26/2020
by   Ali Safaya, et al.
0

In this paper, we describe our approach to utilize pre-trained BERT models with Convolutional Neural Networks for sub-task A of the Multilingual Offensive Language Identification shared task (OffensEval 2020), which is a part of the SemEval 2020. We show that combining CNN with BERT is better than using BERT on its own, and we emphasize the importance of utilizing pre-trained language models for downstream tasks. Our system, ranked 4th with macro averaged F1-Score of 0.897 in Arabic, 4th with score of 0.843 in Greek, and 3rd with score of 0.814 in Turkish. Additionally, we present ArabicBERT, a set of pre-trained transformer language models for Arabic that we share with the community.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/12/2020

CIA_NITT at WNUT-2020 Task 2: Classification of COVID-19 Tweets Using Pre-trained Language Models

This paper presents our models for WNUT 2020 shared task2. The shared ta...
research
07/10/2020

Multi-Dialect Arabic BERT for Country-Level Dialect Identification

Arabic dialect identification is a complex problem for a number of inher...
research
10/07/2020

Galileo at SemEval-2020 Task 12: Multi-lingual Learning for Offensive Language Identification using Pre-trained Language Models

This paper describes Galileo's performance in SemEval-2020 Task 12 on de...
research
06/16/2022

Deep Multi-Task Models for Misogyny Identification and Categorization on Arabic Social Media

The prevalence of toxic content on social media platforms, such as hate ...
research
11/04/2020

MTLB-STRUCT @PARSEME 2020: Capturing Unseen Multiword Expressions Using Multi-task Learning and Pre-trained Masked Language Models

This paper describes a semi-supervised system that jointly learns verbal...
research
03/01/2022

BERT-LID: Leveraging BERT to Improve Spoken Language Identification

Language identification is a task of automatically determining the ident...
research
02/21/2021

Pre-Training BERT on Arabic Tweets: Practical Considerations

Pretraining Bidirectional Encoder Representations from Transformers (BER...

Please sign up or login with your details

Forgot password? Click here to reset