A split-list approach for relevance feedback in information retrieval

https://doi.org/10.1016/j.ipm.2012.03.007Get rights and content

Abstract

In this paper we present a new algorithm for relevance feedback (RF) in information retrieval. Unlike conventional RF algorithms which use the top ranked documents for feedback, our proposed algorithm is a kind of active feedback algorithm which actively chooses documents for the user to judge. The objectives are (a) to increase the number of judged relevant documents and (b) to increase the diversity of judged documents during the RF process. The algorithm uses document-contexts by splitting the retrieval list into sub-lists according to the query term patterns that exist in the top ranked documents. Query term patterns include a single query term, a pair of query terms that occur in a phrase and query terms that occur in proximity. The algorithm is an iterative algorithm which takes one document for feedback in each of the iterations. We experiment with the algorithm using the TREC-6, -7, -8, -2005 and GOV2 data collections and we simulate user feedback using the TREC relevance judgements. From the experimental results, we show that our proposed split-list algorithm is better than the conventional RF algorithm and that our algorithm is more reliable than a similar algorithm using maximal marginal relevance.

Highlights

► Active feedback algorithms can increase the diversity of feedback documents. ► Active feedback algorithms with a re-ranking step perform better than those without a re-ranking step. ► Active feedback algorithm using document-contexts performs better than conventional relevance feedback algorithm. ► Active feedback algorithm using document-contexts performs more reliably than algorithms using whole document.

Introduction

Relevance feedback (RF) is known to be effective for improving retrieval effectiveness (Harman, 1992, Rocchio, 1971, Salton and Buckley, 1990). RF requires effort and time from users to judge whether a document is relevant to the user’s information need. When a user judges a particular document to be non-relevant, in the point of view from users, some effort and time is wasted because the document provides no relevant information to the user. As a result, it is better to have more relevant documents to be judged by the user in the RF process. However, judging two very similar relevant documents also wastes effort and time from users because the information contained in the two documents is nearly the same. Thus, no additional relevant information is provided to the user. Therefore, two main factors would affect the user’s satisfaction in the RF process:

  • (a)

    the number of relevant documents (the more the better), and

  • (b)

    the diversity of the documents (the more diverse the better).

In conventional RF, the user would judge documents from the top ranked ones in the result list by assuming that the top ranked documents contain more relevant information. In some cases, the top ranked documents are very similar to each other or even identical. Judging the relevance of the nearly identical documents provides no additional useful information to both the user and the retrieval system. Therefore, the set of documents used for RF may not be the top ranked ones. This is called active feedback (Chen and Lu, 2010, Shen and Zhai, 2005, SIA et al., 2007, Xu and Akella, 2008a, Xu and Akella, 2008b, Xu et al., 2007, Zhang and Zhang, 2010) in which the retrieval system actively chooses suitable documents for the user to judge for relevance.

Our proposed algorithm is a kind of active feedback algorithm which uses document-contexts by splitting the retrieval list into sub-lists according to the query term patterns that exist in the top ranked documents. Fig. 1 shows an example of the lists of document-contexts for the query “Hubble Telescope Achievements”. Note that only lists for single query terms are shown. Document-contexts are used because we believe that relevant information is located around query terms (Wu et al., 2007, Wu et al., 2008). By splitting the retrieval list into sub-lists, we hope that the proportion of relevant documents in a particular sub-list will be higher than that of the others. Therefore, the scores of document-contexts in the particular sub-list will be higher. By that we can increase the number of relevant documents judged by the user in the RF process. Also, the set of documents being judged by the user in the split-list approach is different from the set of top ranked documents. Hence, it can increase the diversity of the documents being judged.

The rest of the paper is organized as follows. In Section 2, we describe some active feedback algorithms including the gapped method and cluster method in Shen and Zhai (2005) and the maximal marginal relevance method (Carbonell & Goldstein, 1998). Section 3 describes our split-list approach using document-contexts. We show the experiment results in Sections 4 Experiments, 5 Experimental results concludes the paper.

Section snippets

Related work

In this section we review some active feedback algorithms which do not use the top Nrf ranked documents for feedback. In 2007, Xu et al. (2007) proposed a method to incorporate diversity and density measures in performing relevance feedback. They showed improvements in retrieval effectiveness over other active feedback methods using cross-validation. In our experiments, instead of using cross-validation, we calibrate the parameters using one collection and use the fixed parameter values for

Split-list approach to relevance feedback

In this section we describe the proposed split-list approach to RF. The algorithm uses document-contexts by splitting the retrieval list into sub-lists according to the query term patterns that exist in the top ranked documents. Query term patterns include single query term, a pair of query terms that occur in a phrase and in proximity. Each sub-list contains the document-contexts having the same query term pattern (see Fig. 1 for an example). The document-contexts are then ranked in each of

Set up

We use the TREC-2005 data collection for parameter calibration and perform tests on TREC-6, -7, -8, -2005 and GOV2 data collections. Table 1 shows the statistics of the collections. Title queries are used in all the experiments. The longer the query, the more the processing time is needed. This is because we have to extract the document-contexts of individual query terms and pairs of query terms during retrieval. The number of document-contexts increases with the number of query terms. Since we

Experimental results

The results of the initial retrieval and PRF in TREC-2005 are shown in Table 3. We optimize the P@20 of the PRF in order to have more relevant documents for the subsequent RF algorithms.

Table 4 shows the results of different algorithms using rank freeze in the five tested TREC collections. For completeness and reference, the results without rank freeze are also shown in Table 5. Note that we do not perform direct comparisons on Table 5 because the results do not reflect the real utility

Summary

We have applied the notion of document-context using an iterative process in RF. We split the retrieval list into sub-lists of document-contexts for different query term patterns including single query term occurrence, a pair of query terms that occur in a phrase, and a pair of query terms that occur in proximity. Our objectives are (a) finding more relevant feedback documents and (b) increasing the diversity of the feedback documents in order to enhance user satisfaction during RF. We have

Acknowledgement

This work is supported by a grant from the Research Grants Council of the Hong Kong Special Administrative Region, China (Project No. PolyU 5259/09E). We thank Prof. W. Gao and the anonymous reviewers for providing helpful comments to improve the paper.

References (22)

  • H.C. Wu et al.

    A retrospective study of a hybrid document-context based retrieval model

    Information Processing & Management

    (2007)
  • Buckley, C., Salton, G., Allan, J., & Singhal, A. (1994). Automatic query expansion using SMART. In Proceedings of...
  • Carbonell, J., & Goldstein, J. (1998). The use of MMR, diversity-based re-ranking for reordering documents and...
  • Y.K. Chang et al.

    Evaluation of feedback retrieval using modified freezing, residual collection and test and control groups

  • Z. Chen et al.

    Using text classification method in relevance feedback

  • W. Croft et al.

    Using probabilistic models of document retrieval without relevance information

    Journal of Documentation

    (1979)
  • Harman, D. (1992). Relevance feedback revisited. In Proceedings of ACM SIGIR 98 (pp....
  • L. Kaufman et al.

    Finding groups in data: An introduction to cluster analysis

    (1990)
  • Lafferty, J., & Zhai, C. (2001). Document language models, query models, and risk minimization for information...
  • J. Lin

    Divergence measures based on the shannon entropy

    IEEE Transactions on Information Theory

    (1991)
  • Ponte, J., & Croft, W. B. (1998). A language modeling approach in information retrieval. In Proceedings of ACM SIGIR 98...
  • Cited by (3)

    View full text