LoL: A Comparative Regularization Loss over Query Reformulation Losses for Pseudo-Relevance Feedback

04/25/2022
by   Yunchang Zhu, et al.
0

Pseudo-relevance feedback (PRF) has proven to be an effective query reformulation technique to improve retrieval accuracy. It aims to alleviate the mismatch of linguistic expressions between a query and its potential relevant documents. Existing PRF methods independently treat revised queries originating from the same query but using different numbers of feedback documents, resulting in severe query drift. Without comparing the effects of two different revisions from the same query, a PRF model may incorrectly focus on the additional irrelevant information increased in the more feedback, and thus reformulate a query that is less effective than the revision using the less feedback. Ideally, if a PRF model can distinguish between irrelevant and relevant information in the feedback, the more feedback documents there are, the better the revised query will be. To bridge this gap, we propose the Loss-over-Loss (LoL) framework to compare the reformulation losses between different revisions of the same query during training. Concretely, we revise an original query multiple times in parallel using different amounts of feedback and compute their reformulation losses. Then, we introduce an additional regularization loss on these reformulation losses to penalize revisions that use more feedback but gain larger losses. With such comparative regularization, the PRF model is expected to learn to suppress the extra increased irrelevant information by comparing the effects of different revised queries. Further, we present a differentiable query reformulation method to implement this framework. This method revises queries in the vector space and directly optimizes the retrieval performance of query vectors, applicable for both sparse and dense retrieval models. Empirical evaluation demonstrates the effectiveness and robustness of our method for two typical sparse and dense retrieval models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/21/2021

Pseudo-Relevance Feedback for Multiple Representation Dense Retrieval

Pseudo-relevance feedback mechanisms, from Rocchio to the relevance mode...
research
05/12/2023

Generative and Pseudo-Relevant Feedback for Sparse, Dense and Learned Sparse Retrieval

Pseudo-relevance feedback (PRF) is a classical approach to address lexic...
research
05/12/2022

How does Feedback Signal Quality Impact Effectiveness of Pseudo Relevance Feedback for Passage Retrieval?

Pseudo-Relevance Feedback (PRF) assumes that the top results retrieved b...
research
08/01/2023

Generative Query Reformulation for Effective Adhoc Search

Performing automatic reformulations of a user's query is a popular parad...
research
12/18/2022

Curriculum Sampling for Dense Retrieval with Document Expansion

The dual-encoder has become the de facto architecture for dense retrieva...
research
08/25/2021

Pseudo Relevance Feedback with Deep Language Models and Dense Retrievers: Successes and Pitfalls

Pseudo Relevance Feedback (PRF) is known to improve the effectiveness of...
research
12/20/2018

Iterative Relevance Feedback for Answer Passage Retrieval with Passage-level Semantic Match

Relevance feedback techniques assume that users provide relevance judgme...

Please sign up or login with your details

Forgot password? Click here to reset