Resources and Few-shot Learners for In-context Learning in Slavic Languages

04/04/2023
by   Michal Štefánik, et al.
0

Despite the rapid recent progress in creating accurate and compact in-context learners, most recent work focuses on in-context learning (ICL) for tasks in English. However, the ability to interact with users of languages outside English presents a great potential for broadening the applicability of language technologies to non-English speakers. In this work, we collect the infrastructure necessary for training and evaluation of ICL in a selection of Slavic languages: Czech, Polish, and Russian. We link a diverse set of datasets and cast these into a unified instructional format through a set of transformations and newly-crafted templates written purely in target languages. Using the newly-curated dataset, we evaluate a set of the most recent in-context learners and compare their results to the supervised baselines. Finally, we train, evaluate and publish a set of in-context learning models that we train on the collected resources and compare their performance to previous work. We find that ICL models tuned in English are also able to learn some tasks from non-English contexts, but multilingual instruction fine-tuning consistently improves the ICL ability. We also find that the massive multitask training can be outperformed by single-task training in the target language, uncovering the potential for specializing in-context learners to the language(s) of their application.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/03/2022

Crosslingual Generalization through Multitask Finetuning

Multitask prompted finetuning (MTF) has been shown to help large languag...
research
11/01/2022

Preserving In-Context Learning ability in Large Language Model Fine-tuning

Pretrained large language models (LLMs) are strong in-context learners t...
research
06/20/2023

Democratizing LLMs for Low-Resource Languages by Leveraging their English Dominant Abilities with Linguistically-Diverse Prompts

Large language models (LLMs) are known to effectively perform tasks by s...
research
02/16/2023

Reanalyzing L2 Preposition Learning with Bayesian Mixed Effects and a Large Language Model

We use both Bayesian and neural models to dissect a data set of Chinese ...
research
05/23/2023

Concept-aware Training Improves In-context Learning Ability of Language Models

Many recent language models (LMs) of Transformers family exhibit so-call...
research
05/23/2023

LLM-powered Data Augmentation for Enhanced Crosslingual Performance

This paper aims to explore the potential of leveraging Large Language Mo...
research
09/17/2019

Say Anything: Automatic Semantic Infelicity Detection in L2 English Indefinite Pronouns

Computational research on error detection in second language speakers ha...

Please sign up or login with your details

Forgot password? Click here to reset