Iterative Relevance Feedback for Answer Passage Retrieval with Passage-level Semantic Match

12/20/2018 ∙ by Keping Bi, et al. ∙ University of Massachusetts Amherst 0

Relevance feedback techniques assume that users provide relevance judgments for the top k (usually 10) documents and then re-rank using a new query model based on those judgments. Even though this is effective, there has been little research recently on this topic because requiring users to provide substantial feedback on a result list is impractical in a typical web search scenario. In new environments such as voice-based search with smart home devices, however, feedback about result quality can potentially be obtained during users' interactions with the system. Since there are severe limitations on the length and number of results that can be presented in a single interaction in this environment, the focus should move from browsing result lists to iterative retrieval and from retrieving documents to retrieving answers. In this paper, we study iterative relevance feedback techniques with a focus on retrieving answer passages. We first show that iterative feedback is more effective than the top-k approach for answer retrieval. Then we propose an iterative feedback model based on passage-level semantic match and show that it can produce significant improvements compared to both word-based iterative feedback models and those based on term-level semantic similarity.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

In typical relevance feedback (RF) techniques, users are provided with a list of top-ranked documents and asked to assess their relevance. The judged documents, together with the original query, are used to estimate a new query model using an RF model, which further acts as a basis for re-ranking. There were extensive studies of RF

[1, 4, 9, 27, 25, 29, 13, 2, 28, 16, 37]

based on the vector space model (VSM)

[30], the probabilistic model [18] and, more recently, on the language model (LM) for Information Retrieval (IR) approach [23]. Despite the effectiveness of RF, the overhead involved in obtaining user relevance judgments has meant that it is not used in typical search scenarios.

With mobile and voice-based search becoming more popular, it becomes feasible to obtain feedback about result quality during users’ interactions with the system. In these scenarios, the display space or voice bandwidth leads to severe limitations on the length and number of results shown in a single interaction. Thus, instead of providing a list of results, an iterative approach to feedback may be more effective. There has been some work in the past on iterative relevance feedback (IRF) with only a few results in each interaction using the VSM [1, 13, 2], but this has not been looked at for a long time. In addition, the space and bandwidth limitations make the retrieval of longer documents less desirable than shorter answer passages. Motivated by these reasons, in this paper, we present a detailed study of methods for IRF focused on answer passage retrieval.

Although they could be applied to any text retrieval scenario, most existing RF algorithms use word-based models originally designed for document retrieval. Answer passages, however, are much shorter than documents, which could potentially present problems for accurate estimation of word weights in the existing word-based RF methods. Moreover, the limitations on the length and number of results in IRF mean that there is even less relevant text available at every iteration. Given these issues, introducing complementary information from semantic space may help to estimate a more accurate RF model. Dense vector representations of words and paragraphs in distributed semantic space, called embeddings, [21, 17, 32, 8, 5]

, have been effectively applied to many natural language processing (NLP) tasks. Embeddings have also been used in pseudo relevance feedback based on documents

[35, 24], but their impact in iterative and passage-based feedback is not known. Besides, these previous work use semantic similarity at the term level and does not consider semantic match at larger granularity. This had led us to incorporate passage-level semantic match in IRF for answer passage retrieval to improve upon word-based IRF and other embedding-based IRF using term-level semantic similarity.

In the paper, we first investigate whether iterative feedback based on different frameworks is effective relative to RF with a list of top k (k=10) results on answer passage retrieval. The results indicate that IRF is significantly more effective on answer passage collections. In addition, we propose an embedding-based IRF method using passage-level similarity for answer passage retrieval. This method incorporates the similarity scores computed with different types of answer passage embeddings and fuses them with other types of IRF models. The model we propose significantly outperforms IRF baselines based on words or semantic matches between terms. Combining both term-level and passage-level semantic match information leads to additional gains in performance.

2 Related Work

In this section, we first review previous approaches to RF and IRF. We then discuss related work on embeddings of words and paragraphs applied to IR and some previous studies on answer passage retrieval.

Relevance Feedback. In general, there are mainly three types of relevance feedback methods for ad-hoc retrieval, which are based on the vector space model (VSM) [30], the probabilistic model [18] and the language model (LM) [23]. Basically, they all extract expansion terms from annotated relevant documents and re-weight the original query terms so as to estimate a more accurate query model to retrieve better results.

Rocchio [27]

is generally credited as the first RF technique, developed on the VSM. It refines the vector of a user query by bringing it closer to the average vector of relevant documents and further from the average vector of non-relevant documents. In the probabilistic model, expansion terms are scored according to the probability they occur in relevant documents compared to non-relevant documents

[25, 12]. Salton et al. [29] studied various RF techniques based on the VSM and probabilistic model and showed that the probabilistic RF models are in general not as effective as the methods in the VSM.

More recently, feedback techniques have been investigated extensively based on LM, among which, the relevance model [16] and the mixture model [37] are two well-known examples that empirically perform well. In the third version of the relevance model (RM3) [16], the probabilities of expansion terms are estimated with occurrences of the terms in feedback documents. The mixture model [37] considers a feedback document to be generated from a mixture of a corpus language model, and a query topic model, which is estimated with the EM algorithm. Some recent work [4, 9] extend the mixture model by considering additional or different language models as components of the mixture.

Iterative Relevance Feedback. In contrast to most RF systems that ask users to give relevance assessments on a batch of documents, Aalsberg et al. [1] proposed the alternative technique of incremental RF based on Rocchio. Users are asked to judge a single result shown in each interaction, then the query model can be modified iteratively through feedback. This approach showed higher retrieval quality compared with standard batch feedback. Later, Lwayama et al. [13] showed that the incremental relevance feedback used by Aalsberg et al. works better for documents with similar topics, while not as well for documents spanning several topics. In this paper, we investigate how IRF performs on retrieval of answer passages instead of documents using more recent retrieval models.

Some recent TREC tracks [33, 10] have made use of iterative and passage-level feedback, but they focus on document retrieval with different objectives and require a large amount of user feedback. The Total Recall track [33] aims at high recall, where the goal is to promote all of the relevant documents before non-relevant ones. The target of the Dynamic Domain track [10] is to identify documents satisfying all the aspects of the users’ information need with passage-level feedback. In contrast, we focus on iterative feedback for the task of answer passage retrieval and investigate IRF with a fixed small amount of feedback.

Word and Paragraph Embeddings for RF. Dense representations, called embeddings, of words and paragraphs, have become popular and been used [21, 36, 32, 5] to abstract the meaning from a piece of text in semantic space. Two well-known techniques to train word and paragraph embeddings are Word2Vec [20] and Paragraph Vectors (PV) [17] respectively. The similarity of word embeddings can be used to compute the transition probabilities between words [24, 35] and incorporated with the VSM or relevance model [16] to solve problems of term mismatch. Basically, these approaches use semantic match at word level and are in the form of query expansion. In contrast, our approach uses semantic match at passage level and is not based on query expansion.

3 Word-based IRF Models

In IRF, the topic model of users’ intent can be refined each iteration after a small number of results are assessed. Therefore, re-ranking is triggered earlier in IRF than in standard top-k RF methods. On the one hand, earlier re-ranking may produce better results with fewer iterations, which essentially reduces the user efforts in search interactions. On the other hand, having only a small amount of feedback information in each iteration may hurt the accuracy of model estimation and cause topic drift in the iterative process.

We convert several representative models to iterative versions and investigate the performance of the IRF models on answer passage retrieval. Since LM and VSM are the two most effective frameworks for RF, we study iterative feedback under these two frameworks. We use RM3 [16] and the Distillation (or Distill) model [4] to represent the LM framework, and Rocchio [27] for VSM. RM3 is a common baseline for pseudo RF that has also been used for RF. Distillation is one of the most recent RF methods, which is an extension of the mixture model by incorporating a query specific non-relevant topic model. Rocchio [27] is the standard feedback model in VSM. As for the retrieval models for initial ranking, we use Query Likelihood (QL) for LM, and BM25 [26] for VSM respectively.

To keep the query model from diverging to non-relevant topics, we maintain two pools for relevant and non-relevant results, which accumulate all the judgments until the th iteration. During the th iteration, judged relevant results and non-relevant results are added to the corresponding pool. Expanded query models are then estimated from the relevant document pool by RM3 and from both relevant and non-relevant pools by Distillation and Rocchio. Detailed introduction about the IRF models and the experiments can be found in [3].

4 Passage Embedding based IRF Models

Word-based RF methods were initially designed for document retrieval and usually based on query expansion. In contrast to documents, answer passages do not have sufficient text to estimate the probabilities or weights of the expansion terms accurately, especially for IRF when fewer results are available in each iteration. To alleviate the problem of text insufficiency in IRF, we incorporate semantic information about paragraphs to the IRF models. Paragraph embeddings are shown to be capable of capturing the semantic meanings of passages [17, 32, 5], which could potentially help us build more robust IRF models by supporting semantic matching between passages.

In this section, we propose to use paragraph embeddings to improve the performance of IRF for answer passage retrieval. In contrast to existing word-based and embedding-based RF methods, this approach does not extract expansion terms to update the query model. Instead, it represents the relevance topic from feedback passages with embeddings. Similar to Rocchio, we assume a relevant passage should be near the centroid of other relevant passages in the embedding space. Also, we only focus on positive feedback as negative feedback has been shown to have little benefit for RF when positive feedback is available in previous studies [1]. Therefore, our model can be viewed as an embedding version of Rocchio with only positive feedback.

We first describe the methods we use to obtain the semantic representations for answer passages. Then we will introduce the passage embedding based iterative feedback model.

4.1 Passage Embeddings

Aggregated Word Embeddings. One common way of representing passages is to use aggregated embeddings of words in the paragraph. Word2Vec is a well-known method of training word embeddings [20, 21]. It projects words to dense vector space and uses a word to predict its context or predicts a word by its context. In our experiments, we also use average word embeddings trained from Word2Vec both with and without IDF weighting as passage representations.

Figure 1: HDC models used in our experiments. Red words are local context, and blue words are global context.

Paragraph Vectors. The other way of representing passages is using specially designed paragraph vectors models as in [17, 5, 32]. The models we use are PV-HDC [32] with or without corruption, shown in Figure 1. PV-HDC is an extension of the initially proposed paragraph vector model [17], where a document vector is first used to predict an observed word, and afterward, the observed word is used to predict its context words. The recent work of training paragraph representation through corruption [5]

shows advantages in many tasks such as sentiment analysis. It replaces the original part of paragraph representation with a corruption module, where the global context

is generated through an unbiased dropout corruption at each update and the paragraph representation is calculated as the average embeddings of the words in . The final representation is simply the average of the embeddings of all the words in the paragraph. We also investigate other models such as the original PV models, DM, DBOW, [17], and the Parallel Document Context Model (PDC) [32], both with and without corruption, but HDC is better in most cases. So we exclude the other models in the paper.

4.2 IRF with Passage-level Semantic Similarity

As an alternative to query expansion based RF methods, we propose to represent the whole semantic meaning of a passage and a passage set with vectors in the embedding space and measure the similarity between them without explicitly extracting any expansion terms. Specifically, we represent the relevance topic in the th iteration as the embedding of the relevant passage pool and fuse the similarity between a passage with the relevance topic with other RF methods. Thus the score function is shown as follows,


is the expanded query model estimated by iterative version of RF models such as RM3, Distillation and Rocchio; is the candidate passage; denotes the relevant passage pool in the th iteration; denotes the score calculated from other RF models;

is the semantic match score between passages, which is the commonly used cosine similarity in the paper;

is the coefficient of incorporating the passage embedding based similarity; Similar to Rocchio, we assume the topic of a passage set is the centroid of these passages and we consider a relevant passage pool can be represented by average vectors of the passages in it. Thus the similarity between a passage and the pool is


where and is the vector representation of and in the embedding space.

Our method has two advantages over existing RF methods. One is that compared to expansion term based methods that only alleviate word-level mismatch, semantic similarity of larger granularity is captured in our method. The other is the flexibility of combining this semantic match signal with different types of approaches such as RM3, Distillation, Mixture, Rocchio, and other embedding-based feedback approaches.

5 Experiments of Word-based IRF

In this section, we introduce the experimental setup and results of word-based IRF on answer passage retrieval.

5.1 Experimental Setup

Data. In our experiments, we used WebAP and PsgRobust for answer passage retrieval. Statistics of the datasets are summarized in Table 1. WebAP [34] is a web answer passage collection built on Gov2. It uses a subset of queries that are likely to have passage-level answers from Gov2 and retrieved the top 50 documents with the Sequential Dependency Model (SDM) [19]. After that, relevant documents were annotated for relevant answer passages. Overall, 3843 passages from 1200 documents are annotated as relevant. In our experiments, we split the rest of the documents into non-overlapping 2 or 3 (randomly chosen) contiguous sentences as non-relevant passages and used topic descriptions as questions.

PsgRobust 111This dataset is publicly available at is a new collection we created for answer passage retrieval. It is based on the TREC Robust collection following a similar approach as WebAP but without manual annotation. In PsgRobust, we assume that top-ranked passages in relevant documents can be considered as relevant and all passages in non-relevant documents are irrelevant. We first retrieved the top 100 documents for each title query in Robust with SDM [19] and generated answer passages from them with a sliding window of random lengths (2 or 3 sentences) and no overlap. After that, we retrieved top 100 passages with SDM again and treated those from relevant documents as the relevant passages. Similar to WebAP, we used the descriptions of Robust topics as questions and have 246 queries with non-zero relevant answer passages in total. The Recall@100 in the initial retrieval process is 0.43, which means that 43% of relevant documents for all queries were included in the passage collection on average. By manually checking some randomly sampled passages marked as relevant, we found most of them are indeed relevant passages for the questions. There are 6589 relevant passages from 3544 documents for the 246 queries in total.

We also considered other collections that have passage-level annotation such as the DIP2016Corpus [11] and the dataset from the Dynamic Domain track [33]. However, they either are trivial for RF tasks (almost all top 10 results retrieved by BM25 are relevant in DIP2016Corpus) or have few queries (only 26 and 27 for the two domains of the Dynamic Domain track). Other popular question answering datasets usually only have one relevant answer for each query and thus are not suitable for our RF task either. Therefore, we only report the results of WebAP and PsgRobust in this paper.

Dataset #Psg PsgLen Vocab #Query #Psg/D #RelPsg/Q #RelPsg/D
WebAP 379k 45 59k 80 114.3 48.0 3.2
PsgRobust 383k 46 64k 246 17.1 26.8 1.9
Table 1: Statistics of experimental datasets.

System Settings. All the methods we implemented are based on the Galago toolkit [7] 222 Stopwords were removed from all collections using the standard INQUERY stopword list and words were stemmed with Krovetz Stemmer [15]. To compare iterative feedback with typical top-k feedback in a fair manner, we fixed the total number of judged results as 10 and experimented with 1, 2, 5, and 10 iterations, where 10, 5, 2, 1 results were judged during each iteration respectively. Then 10Doc-1Iter is exactly the top-k feedback. We pay more attention to the settings of one or two results per iteration which are more likely to be in a real interactive search scenario considering the limitation of presenting results. True labels of results were used to simulate users’ judgments.

All the parameters were set using 5-fold cross-validation over all the queries in each collection with grid search. For WebAP and PsgRobust, we tuned of QL in and of BM25 from , b set to 0.75 as suggested by [22]. The number of expansion terms is from . The range to scan parameters for RM3, Distillation and Rocchio is similar as the corresponding original paper. They are not shown here due to space limits.

Evaluation. The evaluation should only focus on the ranking of unseen results. So we use freezing ranking [6, 28], as in [1, 14], to evaluate the performance of IRF. The freezing ranking paradigm freezes the ranks of all results presented to the user in the earlier feedback iterations and assigns the first result retrieved in the th iteration rank , where is the number of results shown in each iteration. Note that all the previously shown results are filtered out in the following retrieval to remove duplicates and the final result list concatenates freezing results with the rest candidates ranked in the last iteration, where is the total number of iterations. Then we use mean average precision at cutoff 100 () and to measure the performance of results overall and on the top. As suggested by Smucker et al. [31], statistical significance is calculated with Fisher randomization test with threshold 0.05.

5.2 Results and Discussion

The performance of the initial retrieval with QL and BM25 and the IRF experimental results are shown in Table 2. All the feedback methods are significantly better than their retrieval baselines, i.e. RM3 and Distillation compared with QL, Rocchio compared with BM25, in terms of both and . 333 On PsgRobust, BM25 and Rocchio underperform QL, RM3 and Distillation respectively by a large margin. Because its labels are generated based on retrieval with SDM, this collection favors approaches in the framework of LM more than VSM.

In addition, on both WebAP and PsgRobust, the and of RM3, Distillation and Rocchio increase as the ten results are judged in more iterations. In other words, IRF is much more effective for answer passage retrieval compared with top-k feedback. Performance goes up when re-ranking is done earlier even when we have only a small number of passages, probably because answer passages are usually focused on a single topic and less likely to cause topic drift. Since and show similar trends using IRF under different settings, we only show in Section 6.2 due to the space limitations.

Dataset Method of freezing rank lists of freezing rank lists
(DI) Initial (101) (52) (25) (110) Initial (101) (52) (25) (110)
WebAP RM3 0.076 0.100 0.107* 0.113* 0.113* 0.143 0.170 0.180* 0.185* 0.187*
Distill 0.076 0.099 0.104 0.109 0.111 0.143 0.166 0.177* 0.185* 0.187*
Rocchio 0.081 0.106 0.112 0.118 0.119 0.150 0.169 0.181* 0.190* 0.191*
PsgRobust RM3 0.248 0.293 0.299* 0.306* 0.308* 0.319 0.356 0.363* 0.372* 0.373*
Distill 0.248 0.292 0.299 0.311 0.313 0.319 0.354 0.362* 0.375* 0.379*
Rocchio 0.191 0.268 0.280 0.285 0.286 0.292 0.341 0.356* 0.361* 0.364*
Table 2: Performance of IRF on answer passage collections. stands for . ‘*’ denotes significant improvements over the standard top 10 feedback model ().

6 Experiments of Passage Embedding Based IRF

We compare our method with word-based and embedding-based RF baselines in two groups of experiments. One is the same as in Section 5, i.e. retrieval with a different number of iterations and 10 results judged in total. The other focuses on identifying more relevant passages given only one relevant answer passage. We first describe the experimental setup and then introduce the two groups of experiments in Section 6.2 and Section 6.3.

6.1 Experimental Setup

In this part, we again use WebAP and PsgRobust for experiments. All comparisons are based on LM (RM3, Distillation, and Rocchio) and VSM (Rocchio) to see whether the complementary semantic match benefits in both frameworks. We also include the Embedding-based Relevance Model (ERM) [35] as a baseline. ERM revises in the original RM3 as a linear combination of computed from exact term match and

, which takes the semantic relationship between words into account. The translation probability between words is computed with the cosine similarity of their embeddings transformed with the sigmoid function.

444We also tried the true RF version of BM25-PRF-GT [24], which is a generalized translation model of BM25 based on word embeddings and Rocchio. Due to its inferior performance on our dataset, we did not include the experiments here. Statistical significance in all the result tables is calculated with Fisher randomization test with threshold 0.05.

Embeddings Training. Four paragraph representations are tested in the four groups of experiments, where the base models () can be RM3, ERM, Distillation (or Distill) and Rocchio:

/: uniformly or idf-weighted average word vectors trained with the skip-gram model [20].

/: paragraph vectors trained with the HDC structure with or without corruption [32, 5].

Embeddings of words or paragraphs were trained with each local corpus respectively. Words with the frequency less than 5 were removed. No stemming was done across the collections. 10 negative samples were used for each target word. The learning rate and batch size were 0.05 and 256. The dimension of embedding vectors was set to 100. We also tried other hyper-parameters for training embeddings, and the results were similar under different settings. For PVC, corruption rate [5]

was set to 0.9. All the neural networks of training embeddings were implemented using TensorFlow


Parameter Settings. We used the best settings of the baseline models and tuned the parameters of the semantic signals with 5-fold cross-validation for different paragraph embeddings. All the parameters of ERM are tuned in the same range as [35] suggests. in equation 1 is selected from for WebAP, and for PsgRobust respectively.

6.2 Iterative Feedback with Embeddings

First, we conducted IRF experiments with different number of iterations and 10 results judged in total, as described in Section 5. We use

at cutoff 100 of freezing rank lists as the evaluation metric, which is described in section


Method on WebAP on PsgRobust
(DocIter) (101) (52) (25) (110) (101) (52) (25) (110)
RM3 0.100 0.107 0.113 0.113 0.293 0.299 0.306 0.308
ERM 0.101 0.107 0.113 0.116 0.294 0.301 0.310 0.310
RM3+W2V 0.107 0.115 0.117 0.116 0.298 0.303 0.312 0.312
RM3+idfW2V 0.106 0.113 0.121 0.119 0.298 0.303 0.313 0.313
RM3+PV 0.102 0.113 0.123 0.123 0.298 0.305 0.313 0.314
RM3+PVC 0.107 0.114 0.120 0.114 0.297 0.303 0.308 0.311
ERM+W2V 0.107 0.116 0.119 0.118 0.299 0.304 0.313 0.312
ERM+idfW2V 0.106 0.114 0.121 0.118 0.299 0.304 0.314 0.314
ERM+PV 0.103 0.115 0.122 0.121 0.299 0.307 0.314 0.313
ERM+PVC 0.107 0.114 0.121 0.114 0.298 0.304 0.312 0.313
Distillation 0.099 0.104 0.109 0.111 0.292 0.299 0.311 0.313
Distill+W2V 0.106 0.114 0.120 0.113 0.297 0.304 0.314 0.319
Distill+idfW2V 0.106 0.113 0.116 0.115 0.297 0.306 0.316 0.319
Distill+PV 0.103 0.110 0.118 0.116 0.298 0.306 0.317 0.320
Distill+PVC 0.105 0.112 0.120 0.120 0.297 0.304 0.315 0.317
Table 3: Performance of different IRF models. ‘’ and ‘’ denote significant improvements over word-based (RM3, Distillation) and embedding-based (ERM) baselines respectively. () represents standard top-k feedback.
Figure 2: Performance of our method with different paragraph representations compared with Rocchio. ’+’ means significant difference.

Results and Discussion. We show the experimental results of using language model baselines (RM3, ERM, Distillation) in Table 3 and include Rocchio as a baseline in Figure 2. We can see in general the four representations of paragraphs all can improve performance significantly over the word-based and embedding-based baselines under most iteration settings. ERM performs similar to RM3 on WebAP, and our method based on RM3 and ERM also perform similarly. On PsgRobust, ERM performs slightly better than RM3 and our method also performs slightly better combined with ERM than RM3. 666The reason why ERM does not perform well will be shown in Section 6.3 where we discuss the performance difference of ERM on the two tasks. This shows that incorporating passage-level semantic similarity in embedding space produces improvements to both the word-based RF models and the embedding-based RF model using semantic similarity at term level.

The conclusion that IRF shows advantages over top-k feedback still holds when we incorporate word-based RF models with passage-level semantic match. In addition, there is no one representation better than the others all the time, which implies for different datasets, with different baselines, some representations show their advantages fitting the specific property underlying the setting.

6.3 Retrieval Given One Relevant Passage

As we mentioned in Section 4, the small amount of text in answer passages during each iteration may not be enough to build word-based RF models. The extreme case is when we have only one short passage as positive feedback. Effective re-ranking after the first positive feedback will show the user a second relevant answer in fewer iterations and make users less likely to leave after several interactions. Therefore, it is particularly important to perform well given the first positive feedback from users. We designed the second type of experiment to be answer retrieval given one relevant passage.

For each query, we randomly assign a relevant passage to the model as positive feedback and then retrieve from the remaining results. To make the results more reliable, we randomly draw a relevant passage for each query ten times and do ten retrievals. Then we evaluate the performance of each model based on the overall rank lists from the ten retrievals. We take QL and BM25 as baseline retrieval models that do not consider feedback. Similar to the first group of experiments, we use RM3, Distillation, Rocchio as word-based RF baselines in the framework of LM and VSM, and ERM as the embedding-based RF baseline. We use (precision@1), (mean reciprocal rank) to evaluate the ability of a model to identify a second relevant passage in the next interaction given only one positive feedback. MAP at cutoff 100 measures the ability of the model to identify all the other relevant answers.

Dataset WebAP PsgRobust
QL 0.259 0.373 0.071 0.367 0.486 0.231
RM3 0.498 0.602 0.116 0.515 0.634 0.299
ERM 0.516 0.615 0.125 0.513 0.634 0.307
RM3+W2V 0.488 0.598 0.120 0.524 0.643 0.304
RM3+idfW2V 0.488 0.597 0.120 0.525 0.643 0.304
RM3+PV 0.525* 0.625 0.122 0.521 0.641 0.301
RM3+PVC 0.524 0.635 0.123 0.526 0.644 0.303
ERM+W2V 0.513 0.622 0.131 0.529 0.648 0.312
ERM+idfW2V 0.525 0.627 0.130 0.534 0.650 0.312
ERM+PV 0.556 0.648 0.131 0.531 0.649 0.311
ERM+PVC 0.556 0.658 0.134 0.534 0.653 0.313
Distillation 0.494 0.597 0.113 0.516 0.635 0.299
Distill+W2V 0.489 0.593 0.117 0.528* 0.645* 0.304*
Distill+idfW2V 0.489 0.595 0.117 0.525* 0.643* 0.304*
Distill+PV 0.519 0.621 0.120 0.514 0.638 0.297
Distill+PVC 0.534 0.638 0.123 0.524* 0.643* 0.303*
BM25 0.298 0.399 0.072 0.35 0.479 0.176
Rocchio 0.516 0.616 0.121 0.522 0.641 0.279
Rocchio+W2V 0.531 0.640 0.140 0.526 0.645* 0.282*
Rocchio+idfW2V 0.536 0.642 0.139 0.526 0.644 0.282*
Rocchio+PV 0.576 0.668 0.138 0.518 0.642 0.280*
Rocchio+PVC 0.560 0.668 0.143 0.528* 0.647* 0.281*
Table 4: Performance of different IRF methods on finding other relevant answers given one relevant answer. ’’ and ’’ denote significant improvements over word-based (RM3, Distillation, Rocchio) or embedding-based (ERM) baselines respectively.

Results and Discussion. In Table 4, feedback methods are always better than their base retrieval models, i.e. QL, BM25. In general, with the four paragraph representations, the improvements of over the baselines are always significant; , can also be improved significantly in many cases. This shows that incorporating the passage semantic similarity can improve significantly over both the word-based RF baselines and the embedding-based RF baseline with only term-level semantic match information.

In contrast to the IRF experiments, ERM performs much better than RM3 in this task. The reason may be that in the IRF experiments, there are more relevant passages for RM3 to extract expansion terms and alleviate the term mismatch problem, which makes the term-level semantic match from ERM less helpful. In this task, the text for RM3 is not enough to estimate an accurate model and ERM is effective with semantic match. Since our method considers semantic match at passage level, its benefit does not overlap with that from term-level semantic match.

On WebAP, our method combined with RM3 performs similarly to ERM when using PV and PVC and worse than ERM using W2V and idfW2V. On PsgRobust, incorporating our method to RM3 performs better than ERM in terms of and , but worse than ERM with . This shows that incorporating embedding similarity to do RF at passage level or term level alone with little information are comparable to each other. When we combine these two ways of doing RF together, the performance can be further improved, which is shown from the significant improvements upon ERM when we add the passage similarity signal to ERM on both datasets. This is consistent with our claim that the semantic similarity of the passage level is complementary to the term level when combined with word-based RF models since they capture two different granularities of semantic match.

Different from the IRF experiments, the performance of paragraph vectors are better than W2V and idfW2V. This indicates that when there is little feedback information, more accurate representations lead to better performance. In addition, with scarce user feedback, PVC is more effective than PV, probably because it is less susceptible to overfitting a small dataset due to many fewer parameters, i.e. vocabulary size versus corpus size.

7 Conclusion and Future Work

We first showed that IRF is effective on answer passage retrieval. Then we showed that, with passage-level semantic match the performance of iterative feedback and retrieval given one relevant passage can produce significant improvements compared with word-based RF models in the framework of both LM and VSM. The IRF experiments also show our method is better than the embedding-based baseline using term-level similarity. The retrieval experiment based on one relevance passage shows that combining the word and passage level granularities leads to the best performance.

Our method focuses more on user requests of “more like this”. We know diversity is also very important to provide users more informative results and we will take it into account in our future work. In addition, we will consider IRF on answer passage retrieval with end-to-end neural models.

8 Acknowledgments

This work was supported in part by the Center for Intelligent Information Retrieval and in part by NSF IIS-1715095. Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect those of the sponsor.


  • [1] Aalbersberg, I.J.: Incremental relevance feedback. In: Proceedings of the 15th annual international ACM SIGIR conference. pp. 11–22. ACM (1992)
  • [2] Allan, J.: Incremental relevance feedback for information filtering. In: Proceedings of the 19th annual international ACM SIGIR conference. pp. 270–278. ACM (1996)
  • [3] Bi, K., Ai, Q., Croft, W.B.: Revisiting iterative relevance feedback for document and passage retrieval. arXiv preprint arXiv:1812.05731 (2018)
  • [4] Brondwine, E., Shtok, A., Kurland, O.: Utilizing focused relevance feedback. In: Proceedings of the 39th International ACM SIGIR conference. pp. 1061–1064. ACM (2016)
  • [5] Chen, M.: Efficient vector representation for documents through corruption. arXiv preprint arXiv:1707.02377 (2017)
  • [6] Cirillo, C., Chang, Y., Razon, J.: Evaluation of feedback retrieval using modified freezing, residual collection, and test and control groups. Scientific Report No. ISR-16 to the National Science Foundation (1969)
  • [7] Croft, W.B., Metzler, D., Strohman, T.: Search engines: Information retrieval in practice, vol. 283. Addison-Wesley Reading (2010)
  • [8]

    Dai, A.M., Olah, C., Le, Q.V.: Document embedding with paragraph vectors. In: NIPS Deep Learning Workshop (2015)

  • [9] Dehghani, M., Azarbonyad, H., Kamps, J., Hiemstra, D., Marx, M.: Luhn revisited: Significant words language models. In: Proceedings of the 25th ACM International on Conference on Information and Knowledge Management. pp. 1301–1310. ACM (2016)
  • [10] Grossman, M.R., Cormack, G.V., Roegiest, A.: Trec 2016 total recall track overview. In: TREC (2016)
  • [11] Habernal, I., Sukhareva, M., Raiber, F., Shtok, A., Kurland, O., Ronen, H., Bar-Ilan, J., Gurevych, I.: New collection announcement: Focused retrieval over the web. In: Proceedings of the 39th International ACM SIGIR conference. pp. 701–704. ACM (2016)
  • [12] Harman, D.: Relevance feedback revisited. In: Proceedings of the 15th annual international ACM SIGIR conference. pp. 1–10. ACM (1992)
  • [13] Iwayama, M.: Relevance feedback with a small number of relevance judgements: incremental relevance feedback vs. document clustering. In: Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval. pp. 10–16. ACM (2000)
  • [14] Jones, G., Sakai, T., Kajiura, M., Sumita, K.: Incremental relevance feedback in japanese text retrieval. Information Retrieval 2(4), 361–384 (2000)
  • [15] Krovetz, R.: Viewing morphology as an inference process. In: Proceedings of the 16th annual international ACM SIGIR conference. pp. 191–202. ACM (1993)
  • [16] Lavrenko, V., Croft, W.B.: Relevance-based language models. In: ACM SIGIR Forum. vol. 51, pp. 260–267. ACM (2017)
  • [17]

    Le, Q., Mikolov, T.: Distributed representations of sentences and documents. In: Proceedings of the 31st International Conference on Machine Learning (ICML-14). pp. 1188–1196 (2014)

  • [18] Maron, M.E., Kuhns, J.L.: On relevance, probabilistic indexing and information retrieval. Journal of the ACM (JACM) 7(3), 216–244 (1960)
  • [19] Metzler, D., Croft, W.B.: A markov random field model for term dependencies. In: Proceedings of the 28th annual international ACM SIGIR conference. pp. 472–479. ACM (2005)
  • [20] Mikolov, T., Chen, K., Corrado, G., Dean, J.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013)
  • [21] Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: Advances in neural information processing systems. pp. 3111–3119 (2013)
  • [22] Mogotsi, I.: Christopher d. manning, prabhakar raghavan, and hinrich schütze: Introduction to information retrieval (2010)
  • [23] Ponte, J.M., Croft, W.B.: A language modeling approach to information retrieval. In: Proceedings of the 21st annual international ACM SIGIR conference. pp. 275–281. ACM (1998)
  • [24] Rekabsaz, N., Lupu, M., Hanbury, A., Zuccon, G.: Generalizing translation models in the probabilistic relevance framework. In: Proceedings of the 25th ACM CIKM conference. pp. 711–720. ACM (2016)
  • [25] Robertson, S.E., Jones, K.S.: Relevance weighting of search terms. Journal of the Association for Information Science and Technology 27(3), 129–146 (1976)
  • [26] Robertson, S.E., Walker, S., Jones, S., Hancock-Beaulieu, M.M., Gatford, M., et al.: Okapi at trec-3. Nist Special Publication Sp 109,  109 (1995)
  • [27] Rocchio, J.J.: Relevance feedback in information retrieval. The Smart retrieval system-experiments in automatic document processing (1971)
  • [28]

    Ruthven, I., Lalmas, M.: A survey on the use of relevance feedback for information access systems. The Knowledge Engineering Review

    18(2), 95–145 (2003)
  • [29] Salton, G., Buckley, C.: Improving retrieval performance by relevance feedback. Journal of the American Society for Information Science 41, 288–297 (1990)
  • [30] Salton, G., Wong, A., Yang, C.S.: A vector space model for automatic indexing. Communications of the ACM 18(11), 613–620 (1975)
  • [31] Smucker, M.D., Allan, J., Carterette, B.: A comparison of statistical significance tests for information retrieval evaluation. In: Proceedings of the 16th ACM CIKM conference. pp. 623–632. ACM (2007)
  • [32] Sun, F., Guo, J., Lan, Y., Xu, J., Cheng, X.: Learning word representations by jointly modeling syntagmatic and paradigmatic relations. In: ACL (1). pp. 136–145 (2015)
  • [33] Yang, G.H., Soboroff, I.: Trec 2016 dynamic domain track overview. In: TREC (2016)
  • [34] Yang, L., Ai, Q., Spina, D., Chen, R.C., Pang, L., Croft, W.B., Guo, J., Scholer, F.: Beyond factoid qa: Effective methods for non-factoid answer sentence retrieval. In: ECIR (2016)
  • [35] Zamani, H., Croft, W.B.: Embedding-based query language models. In: Proceedings of the 2016 ACM ICTIR. pp. 147–156. ACM (2016)
  • [36] Zamani, H., Croft, W.B.: Relevance-based word embedding. In: Proceedings of the 40th International ACM SIGIR Conference. SIGIR ’17 (2017)
  • [37] Zhai, C., Lafferty, J.: Model-based feedback in the language modeling approach to information retrieval. In: Proceedings of the tenth CIKM conference. pp. 403–410. ACM (2001)