BERT for Monolingual and Cross-Lingual Reverse Dictionary

09/30/2020
by   Hang Yan, et al.
0

Reverse dictionary is the task to find the proper target word given the word description. In this paper, we tried to incorporate BERT into this task. However, since BERT is based on the byte-pair-encoding (BPE) subword encoding, it is nontrivial to make BERT generate a word given the description. We propose a simple but effective method to make BERT generate the target word for this specific task. Besides, the cross-lingual reverse dictionary is the task to find the proper target word described in another language. Previous models have to keep two different word embeddings and learn to align these embeddings. Nevertheless, by using the Multilingual BERT (mBERT), we can efficiently conduct the cross-lingual reverse dictionary with one subword embedding, and the alignment between languages is not necessary. More importantly, mBERT can achieve remarkable cross-lingual reverse dictionary performance even without the parallel corpus, which means it can conduct the cross-lingual reverse dictionary with only corresponding monolingual data. Code is publicly available at https://github.com/yhcc/BertForRD.git.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2020

A Study of Cross-Lingual Ability and Language-specific Information in Multilingual BERT

Recently, multilingual BERT works remarkably well on cross-lingual trans...
research
04/16/2022

WordAlchemy: A transformer-based Reverse Dictionary

A reverse dictionary takes a target word's description as input and retu...
research
08/10/2018

Learning to Represent Bilingual Dictionaries

Bilingual word embeddings have been widely used to capture the similarit...
research
09/12/2019

Lost in Evaluation: Misleading Benchmarks for Bilingual Dictionary Induction

The task of bilingual dictionary induction (BDI) is commonly used for in...
research
05/04/2022

AmbiPun: Generating Humorous Puns with Ambiguous Context

In this paper, we propose a simple yet effective way to generate pun sen...
research
07/25/2023

Combating the Curse of Multilinguality in Cross-Lingual WSD by Aligning Sparse Contextualized Word Representations

In this paper, we advocate for using large pre-trained monolingual langu...

Please sign up or login with your details

Forgot password? Click here to reset