Training Compact Models for Low Resource Entity Tagging using Pre-trained Language Models

10/14/2019
by   Peter Izsak, et al.
0

Training models on low-resource named entity recognition tasks has been shown to be a challenge, especially in industrial applications where deploying updated models a continuous effort and crucial for business operations. Often in such cases, there is abundance of unlabeled data, however, labeled data is scarce or unavailable. Pre-trained language models trained to extract contextual features from text were shown to improve many natural language processing (NLP) tasks, including scarcely labeled tasks, by leveraging on transfer learning. However, such models impose a heavy memory and computational burden, making it a challenge to train and deploy such model for inference use. In this work-in-progress we combined the effectiveness of transfer learning provided by pre-trained masked language models and use a semi-supervised approach to train a fast and compact model using labeled and unlabeled examples. Preliminary evaluations show that the compact models achieve competitive accuracy compared to a state-of-art pre-trained language models with up to 36x compression rate and run significantly faster in inference, thus, allowing deployment of such models in production environments or on edge devices.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2017

Semi-supervised sequence tagging with bidirectional language models

Pre-trained word embeddings learned from unlabeled text have become a st...
research
03/31/2022

Domain Adaptation for Sparse-Data Settings: What Do We Gain by Not Using Bert?

The practical success of much of NLP depends on the availability of trai...
research
06/18/2019

Towards Robust Named Entity Recognition for Historic German

Recent advances in language modeling using deep neural networks have sho...
research
11/06/2022

Prompt-based Text Entailment for Low-Resource Named Entity Recognition

Pre-trained Language Models (PLMs) have been applied in NLP tasks and ac...
research
12/09/2022

From Clozing to Comprehending: Retrofitting Pre-trained Language Model to Pre-trained Machine Reader

We present Pre-trained Machine Reader (PMR), a novel method to retrofit ...
research
02/09/2023

Lightweight Transformers for Clinical Natural Language Processing

Specialised pre-trained language models are becoming more frequent in NL...
research
02/25/2021

PharmKE: Knowledge Extraction Platform for Pharmaceutical Texts using Transfer Learning

The challenge of recognizing named entities in a given text has been a v...

Please sign up or login with your details

Forgot password? Click here to reset