Non-Parametric Few-Shot Learning for Word Sense Disambiguation

04/26/2021
by   Howard Chen, et al.
0

Word sense disambiguation (WSD) is a long-standing problem in natural language processing. One significant challenge in supervised all-words WSD is to classify among senses for a majority of words that lie in the long-tail distribution. For instance, 84 examples in the SemCor training data. This issue is more pronounced as the imbalance occurs in both word and sense distributions. In this work, we propose MetricWSD, a non-parametric few-shot learning approach to mitigate this data imbalance issue. By learning to compute distances among the senses of a given word through episodic training, MetricWSD transfers knowledge (a learned metric space) from high-frequency words to infrequent ones. MetricWSD constructs the training episodes tailored to word frequencies and explicitly addresses the problem of the skewed distribution, as opposed to mixing all the words trained with parametric models in previous work. Without resorting to any lexical resources, MetricWSD obtains strong performance against parametric alternatives, achieving a 75.1 F1 score on the unified WSD evaluation benchmark (Raganato et al., 2017b). Our analysis further validates that infrequent words and senses enjoy significant improvement.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/19/2020

One-Shot Learning for Language Modelling

Humans can infer a great deal about the meaning of a word, using the syn...
research
10/27/2017

One-shot and few-shot learning of word embeddings

Standard deep learning systems require thousands or millions of examples...
research
06/09/2023

Word sense extension

Humans often make creative use of words to express novel senses. A long-...
research
10/02/2020

Cross-Lingual Transfer Learning for Complex Word Identification

Complex Word Identification (CWI) is a task centered on detecting hard-t...
research
10/01/2019

Bad Form: Comparing Context-Based and Form-Based Few-Shot Learning in Distributional Semantic Models

Word embeddings are an essential component in a wide range of natural la...
research
02/24/2023

NoPPA: Non-Parametric Pairwise Attention Random Walk Model for Sentence Representation

We propose a novel non-parametric/un-trainable language model, named Non...
research
06/20/2023

Visually grounded few-shot word learning in low-resource settings

We propose a visually grounded speech model that learns new words and th...

Please sign up or login with your details

Forgot password? Click here to reset