Parameter Space Factorization for Zero-Shot Learning across Tasks and Languages

01/30/2020
by   Edoardo M. Ponti, et al.
0

Most combinations of NLP tasks and language varieties lack in-domain examples for supervised training because of the paucity of annotated data. How can neural models make sample-efficient generalizations from task-language combinations with available data to low-resource ones? In this work, we propose a Bayesian generative model for the space of neural parameters. We assume that this space can be factorized into latent variables for each language and each task. We infer the posteriors over such latent variables based on data from seen task-language combinations through variational inference. This enables zero-shot classification on unseen combinations at prediction time. For instance, given training data for named entity recognition (NER) in Vietnamese and for part-of-speech (POS) tagging in Wolof, our model can perform accurate predictions for NER in Wolof. In particular, we experiment with a typologically diverse sample of 33 languages from 4 continents and 11 families, and show that our model yields comparable or better results than state-of-the-art, zero-shot cross-lingual transfer methods; it increases performance by 4.49 points for POS tagging and 7.73 points for NER on average compared to the strongest baseline.

READ FULL TEXT
research
09/11/2021

Learning from Language Description: Low-shot Named Entity Recognition via Decomposed Framework

In this work, we study the problem of named entity recognition (NER) in ...
research
04/28/2020

MultiMix: A Robust Data Augmentation Strategy for Cross-Lingual NLP

Transfer learning has yielded state-of-the-art results in many supervise...
research
01/20/2019

Mixed Formal Learning: A Path to Transparent Machine Learning

This paper presents Mixed Formal Learning, a new architecture that learn...
research
04/19/2023

MasakhaNEWS: News Topic Classification for African languages

African languages are severely under-represented in NLP research due to ...
research
12/07/2022

WIDER CLOSER: Mixture of Short-channel Distillers for Zero-shot Cross-lingual Named Entity Recognition

Zero-shot cross-lingual named entity recognition (NER) aims at transferr...
research
04/14/2021

Zero-Resource Multi-Dialectal Arabic Natural Language Understanding

A reasonable amount of annotated data is required for fine-tuning pre-tr...
research
06/02/2021

Minimax and Neyman-Pearson Meta-Learning for Outlier Languages

Model-agnostic meta-learning (MAML) has been recently put forth as a str...

Please sign up or login with your details

Forgot password? Click here to reset