Learning Dense Representations for Entity Retrieval

09/23/2019
by   Daniel Gillick, et al.
0

We show that it is feasible to perform entity linking by training a dual encoder (two-tower) model that encodes mentions and entities in the same dense vector space, where candidate entities are retrieved by approximate nearest neighbor search. Unlike prior work, this setup does not rely on an alias table followed by a re-ranker, and is thus the first fully learned entity retrieval model. We show that our dual encoder, trained using only anchor-text links in Wikipedia, outperforms discrete alias table and BM25 baselines, and is competitive with the best comparable results on the standard TACKBP-2010 dataset. In addition, it can retrieve candidates extremely fast, and generalizes well to a new dataset derived from Wikinews. On the modeling side, we demonstrate the dramatic value of an unsupervised negative mining algorithm for this task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2021

MOLEMAN: Mention-Only Linking of Entities with a Mention Annotation Network

We present an instance-based nearest neighbor approach to entity linking...
research
04/07/2020

Entity Linking via Dual and Cross-Attention Encoders

Entity Linking has two main open areas of research: 1) generate candidat...
research
03/08/2021

Fast and Effective Biomedical Entity Linking Using a Dual Encoder

Biomedical entity linking is the task of identifying mentions of biomedi...
research
03/27/2023

Improving Dual-Encoder Training through Dynamic Indexes for Negative Mining

Dual encoder models are ubiquitous in modern classification and retrieva...
research
10/24/2018

Discovering Entities with Just a Little Help from You

Linking entities like people, organizations, books, music groups and the...
research
09/13/2021

MuVER: Improving First-Stage Entity Retrieval with Multi-View Entity Representations

Entity retrieval, which aims at disambiguating mentions to canonical ent...
research
10/02/2020

Autoregressive Entity Retrieval

Entities are at the center of how we represent and aggregate knowledge. ...

Please sign up or login with your details

Forgot password? Click here to reset