Wiki to Automotive: Understanding the Distribution Shift and its impact on Named Entity Recognition

12/01/2021
by   Anmol Nayak, et al.
0

While transfer learning has become a ubiquitous technique used across Natural Language Processing (NLP) tasks, it is often unable to replicate the performance of pre-trained models on text of niche domains like Automotive. In this paper we aim to understand the main characteristics of the distribution shift with automotive domain text (describing technical functionalities such as Cruise Control) and attempt to explain the potential reasons for the gap in performance. We focus on performing the Named Entity Recognition (NER) task as it requires strong lexical, syntactic and semantic understanding by the model. Our experiments with 2 different encoders, namely BERT-Base-Uncased and SciBERT-Base-Scivocab-Uncased have lead to interesting findings that showed: 1) The performance of SciBERT is better than BERT when used for automotive domain, 2) Fine-tuning the language models with automotive domain text did not make significant improvements to the NER performance, 3) The distribution shift is challenging as it is characterized by lack of repeating contexts, sparseness of entities, large number of Out-Of-Vocabulary (OOV) words and class overlap due to domain specific nuances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/07/2023

German BERT Model for Legal Named Entity Recognition

The use of BERT, one of the most popular language models, has led to imp...
research
06/14/2021

Can BERT Dig It? – Named Entity Recognition for Information Retrieval in the Archaeology Domain

The amount of archaeological literature is growing rapidly. Until recent...
research
12/14/2018

Few-shot classification in Named Entity Recognition Task

For many natural language processing (NLP) tasks the amount of annotated...
research
05/03/2021

Switching Contexts: Transportability Measures for NLP

This paper explores the topic of transportability, as a sub-area of gene...
research
02/25/2021

PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using Transfer Learning

The challenge of recognizing named entities in a given text has been a v...
research
04/26/2023

Impact of Position Bias on Language Models in Token Classification

Language Models (LMs) have shown state-of-the-art performance in Natural...
research
12/23/2021

Distilling the Knowledge of Romanian BERTs Using Multiple Teachers

Running large-scale pre-trained language models in computationally const...

Please sign up or login with your details

Forgot password? Click here to reset