MicroBERT: Effective Training of Low-resource Monolingual BERTs through Parameter Reduction and Multitask Learning

12/23/2022
by   Luke Gessler, et al.
0

Transformer language models (TLMs) are critical for most NLP tasks, but they are difficult to create for low-resource languages because of how much pretraining data they require. In this work, we investigate two techniques for training monolingual TLMs in a low-resource setting: greatly reducing TLM size, and complementing the masked language modeling objective with two linguistically rich supervised tasks (part-of-speech tagging and dependency parsing). Results from 7 diverse languages indicate that our model, MicroBERT, is able to produce marked improvements in downstream task evaluations relative to a typical monolingual TLM pretraining approach. Specifically, we find that monolingual MicroBERT models achieve gains of up to 18 for NER F1 compared to a multilingual baseline, mBERT, while having less than 1 labeled data for pretraining low-resource TLMs can yield large quality benefits and in some cases produce models that outperform multilingual approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2022

Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages

It is challenging to train and deploy Transformer LMs for hybrid speech ...
research
09/29/2020

Parsing with Multilingual BERT, a Small Corpus, and a Small Treebank

Pretrained multilingual contextual representations have shown great succ...
research
05/10/2022

The Importance of Context in Very Low Resource Language Modeling

This paper investigates very low resource language model pretraining, wh...
research
02/12/2021

A Little Pretraining Goes a Long Way: A Case Study on Dependency Parsing Task for Low-resource Morphologically Rich Languages

Neural dependency parsing has achieved remarkable performance for many d...
research
04/16/2023

Sabiá: Portuguese Large Language Models

As the capabilities of language models continue to advance, it is concei...
research
02/15/2023

Meeting the Needs of Low-Resource Languages: The Value of Automatic Alignments via Pretrained Models

Large multilingual models have inspired a new class of word alignment me...
research
09/17/2021

Boosting Transformers for Job Expression Extraction and Classification in a Low-Resource Setting

In this paper, we explore possible improvements of transformer models in...

Please sign up or login with your details

Forgot password? Click here to reset