ArabGlossBERT: Fine-Tuning BERT on Context-Gloss Pairs for WSD

05/19/2022
by   Moustafa Al-Hajj, et al.
0

Using pre-trained transformer models such as BERT has proven to be effective in many NLP tasks. This paper presents our work to fine-tune BERT models for Arabic Word Sense Disambiguation (WSD). We treated the WSD task as a sentence-pair binary classification task. First, we constructed a dataset of labeled Arabic context-gloss pairs ( 167k pairs) we extracted from the Arabic Ontology and the large lexicographic database available at Birzeit University. Each pair was labeled as True or False and target words in each context were identified and annotated. Second, we used this dataset for fine-tuning three pre-trained Arabic BERT models. Third, we experimented the use of different supervised signals used to emphasize target words in context. Our experiments achieved promising results (accuracy of 84 senses in the experiment.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/16/2021

LU-BZU at SemEval-2021 Task 2: Word2Vec and Lemma2Vec performance in Arabic Word-in-Context disambiguation

This paper presents a set of experiments to evaluate and compare between...
research
02/06/2023

Context-Gloss Augmentation for Improving Arabic Target Sense Verification

Arabic language lacks semantic datasets and sense inventories. The most ...
research
06/29/2021

New Arabic Medical Dataset for Diseases Classification

The Arabic language suffers from a great shortage of datasets suitable f...
research
05/25/2020

Pointwise Paraphrase Appraisal is Potentially Problematic

The prevailing approach for training and evaluating paraphrase identific...
research
08/07/2023

Detecting Spells in Fantasy Literature with a Transformer Based Artificial Intelligence

Transformer architectures and models have made significant progress in l...
research
05/04/2023

Leveraging BERT Language Model for Arabic Long Document Classification

Given the number of Arabic speakers worldwide and the notably large amou...
research
02/20/2023

Boosting classification reliability of NLP transformer models in the long run

Transformer-based machine learning models have become an essential tool ...

Please sign up or login with your details

Forgot password? Click here to reset