SenseBERT: Driving Some Sense into BERT

08/15/2019
by   Yoav Levine, et al.
0

Self-supervision techniques have allowed neural language models to advance the frontier in Natural Language Understanding. However, existing self-supervision techniques operate at the word-form level, which serves as a surrogate for the underlying semantic content. This paper proposes a method to employ self-supervision directly at the word-sense level. Our model, named SenseBERT, is pre-trained to predict not only the masked words but also their WordNet supersenses. Accordingly, we attain a lexical-semantic level language model, without the use of human annotation. SenseBERT achieves significantly improved lexical understanding, as we demonstrate by experimenting on SemEval, and by attaining a state of the art result on the Word in Context (WiC) task. Our approach is extendable to other linguistic signals, which can be similarly integrated into the pre-training process, leading to increasingly semantically informed language models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2021

Incorporating Word Sense Disambiguation in Neural Language Models

We present two supervised (pre-)training methods to incorporate gloss de...
research
11/13/2019

Word-level Lexical Normalisation using Context-Dependent Embeddings

Lexical normalisation (LN) is the process of correcting each word in a d...
research
01/12/2023

A Cohesive Distillation Architecture for Neural Language Models

A recent trend in Natural Language Processing is the exponential growth ...
research
05/29/2020

A Comparative Study of Lexical Substitution Approaches based on Neural Language Models

Lexical substitution in context is an extremely powerful technology that...
research
09/06/2022

Transfer Learning of Lexical Semantic Families for Argumentative Discourse Units Identification

Argument mining tasks require an informed range of low to high complexit...
research
08/26/2020

Language Models and Word Sense Disambiguation: An Overview and Analysis

Transformer-based language models have taken many fields in NLP by storm...
research
06/10/2023

Universal Language Modelling agent

Large Language Models are designed to understand complex Human Language....

Please sign up or login with your details

Forgot password? Click here to reset