Adapting to the Low-Resource Double-Bind: Investigating Low-Compute Methods on Low-Resource African Languages

03/29/2023
by   Colin Leong, et al.
2

Many natural language processing (NLP) tasks make use of massively pre-trained language models, which are computationally expensive. However, access to high computational resources added to the issue of data scarcity of African languages constitutes a real barrier to research experiments on these languages. In this work, we explore the applicability of low-compute approaches such as language adapters in the context of this low-resource double-bind. We intend to answer the following question: do language adapters allow those who are doubly bound by data and compute to practically build useful models? Through fine-tuning experiments on African languages, we evaluate their effectiveness as cost-effective approaches to low-resource African NLP. Using solely free compute resources, our results show that language adapters achieve comparable performances to massive pre-trained language models which are heavy on computational resources. This opens the door to further experimentation and exploration on full-extent of language adapters capacities.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/21/2023

PEFTT: Parameter-Efficient Fine-Tuning for low-resource Tibetan pre-trained language models

In this era of large language models (LLMs), the traditional training of...
research
03/07/2023

A Challenging Benchmark for Low-Resource Learning

With promising yet saturated results in high-resource settings, low-reso...
research
04/29/2021

MOROCCO: Model Resource Comparison Framework

The new generation of pre-trained NLP models push the SOTA to the new li...
research
04/06/2021

Efficient transfer learning for NLP with ELECTRA

Clark et al. [2020] claims that the ELECTRA approach is highly efficient...
research
09/11/2022

Detecting Suicide Risk in Online Counseling Services: A Study in a Low-Resource Language

With the increased awareness of situations of mental crisis and their so...
research
09/18/2023

Not Enough Labeled Data? Just Add Semantics: A Data-Efficient Method for Inferring Online Health Texts

User-generated texts available on the web and social platforms are often...
research
06/22/2021

Information Retrieval for ZeroSpeech 2021: The Submission by University of Wroclaw

We present a number of low-resource approaches to the tasks of the Zero ...

Please sign up or login with your details

Forgot password? Click here to reset