Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings

09/23/2019
by   Gregor Wiedemann, et al.
0

Contextualized word embeddings (CWE) such as provided by ELMo (Peters et al., 2018), flair NLP (Akbik et al., 2018), or BERT (Devlin et al., 2019) are a major recent innovation in NLP. CWEs provide semantic vector representations of words depending on their respective context. The advantage compared to static word embeddings has been shown for a number of tasks, such as text classification, sequence tagging, or machine translation. Since vectors of the same word can vary due to different contexts, they implicitly provide a model for word sense disambiguation (WSD). We introduce a simple but effective approach to WSD using a nearest neighbor classification on CWEs. We compare the performance of different CWE models for the task and can report improvements above the current state of the art for one standard WSD benchmark dataset. We further show that the pre-trained BERT model is able to place polysemic words into distinct 'sense' regions of the embedding space, while ELMo and flair NLP do not indicate this ability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2021

Learning Sense-Specific Static Embeddings using Contextualised Word Embeddings as a Proxy

Contextualised word embeddings generated from Neural Language Models (NL...
research
11/30/2021

A Comparative Study of Transformers on Word Sense Disambiguation

Recent years of research in Natural Language Processing (NLP) have witne...
research
12/14/2021

Building on Huang et al. GlossBERT for Word Sense Disambiguation

We propose to take on the problem ofWord Sense Disambiguation (WSD). In ...
research
04/13/2021

Semantic maps and metrics for science Semantic maps and metrics for science using deep transformer encoders

The growing deluge of scientific publications demands text analysis tool...
research
08/20/2022

Lost in Context? On the Sense-wise Variance of Contextualized Word Embeddings

Contextualized word embeddings in language models have given much advanc...
research
04/16/2021

Are Classes Clusters?

Sentence embedding models aim to provide general purpose embeddings for ...
research
09/23/2021

Putting Words in BERT's Mouth: Navigating Contextualized Vector Spaces with Pseudowords

We present a method for exploring regions around individual points in a ...

Please sign up or login with your details

Forgot password? Click here to reset