Deeper Text Understanding for IR with Contextual Neural Language Modeling

05/22/2019
by   Zhuyun Dai, et al.
0

Neural networks provide new possibilities to automatically learn complex language patterns and query-document relations. Neural IR models have achieved promising results in learning query-document relevance patterns, but few explorations have been done on understanding the text content of a query or a document. This paper studies leveraging a recently-proposed contextual neural language model, BERT, to provide deeper text understanding for IR. Experimental results demonstrate that the contextual text representations from BERT are more effective than traditional word embeddings. Compared to bag-of-words retrieval models, the contextual language model can better leverage language structures, bringing large improvements on queries written in natural languages. Combining the text understanding ability with search knowledge leads to an enhanced pre-trained BERT model that can benefit related search tasks where training data are limited.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/05/2017

Rhetorical relations for information retrieval

Typically, every part in most coherent text has some plausible reason fo...
research
07/24/2020

IR-BERT: Leveraging BERT for Semantic Search in Background Linking for News Articles

This work describes our two approaches for the background linking task o...
research
07/15/2023

Intuitive Access to Smartphone Settings Using Relevance Model Trained by Contrastive Learning

The more new features that are being added to smartphones, the harder it...
research
09/27/2018

Consistency and Variation in Kernel Neural Ranking Model

This paper studies the consistency of the kernel-based neural ranking mo...
research
09/21/2020

Latin BERT: A Contextual Language Model for Classical Philology

We present Latin BERT, a contextual language model for the Latin languag...
research
02/14/2020

TwinBERT: Distilling Knowledge to Twin-Structured BERT Models for Efficient Retrieval

Pre-trained language models like BERT have achieved great success in a w...
research
05/24/2022

GraphQ IR: Unifying Semantic Parsing of Graph Query Language with Intermediate Representation

Subject to the semantic gap lying between natural and formal language, n...

Please sign up or login with your details

Forgot password? Click here to reset