skip to main content
10.1145/3331184.3331355acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Unbiased Low-Variance Estimators for Precision and Related Information Retrieval Effectiveness Measures

Published:18 July 2019Publication History

ABSTRACT

This work describes an estimator from which unbiased measurements of precision, rank-biased precision, and cumulative gain may be derived from a uniform or non-uniform sample of relevance assessments. Adversarial testing supports the theory that our estimator yields unbiased low-variance measurements from sparse samples, even when used to measure results that are qualitatively different from those returned by known information retrieval methods. Our results suggest that test collections using sampling to select documents for relevance assessment yield more accurate measurements than test collections using pooling, especially for the results of retrieval methods not contributing to the pool.

References

  1. Aslam, J. A., Pavlu, V., and Savell, R. A unified model for metasearch and the efficient evaluation of retrieval systems via the hedge algorithm. In SIGIR 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Cormack, G. V., and Grossman, M. R. Beyond pooling. In SIGIR 2018. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Horvitz, D. G., and Thompson, D. J. A generalization of sampling without replacement from a finite universe. Journal of the American Statistical Association 47, 260 (1952), 663--685.Google ScholarGoogle ScholarCross RefCross Ref
  4. Pavlu, V., and Aslam, J. A practical sampling strategy for efficient retrieval evaluation. Northeastern University (2007).Google ScholarGoogle Scholar
  5. Sanderson, M., et al. Test collection based evaluation of information retrieval systems. Foundations and Trends in Information Retrieval 4, 4 (2010), 247--375.Google ScholarGoogle ScholarCross RefCross Ref
  6. Voorhees, E., and Harman, D. Overview of the eighth text retrieval conference. In TREC 8 (1999).Google ScholarGoogle Scholar
  7. Voorhees, E. M. The effect of sampling strategy on inferred measures. In SIGIR 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Yilmaz, E., Kanoulas, E., and Aslam, J. A. A simple and efficient sampling method for estimating AP and NDCG. In SIGIR 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Unbiased Low-Variance Estimators for Precision and Related Information Retrieval Effectiveness Measures

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        SIGIR'19: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval
        July 2019
        1512 pages
        ISBN:9781450361729
        DOI:10.1145/3331184

        Copyright © 2019 Owner/Author

        Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 18 July 2019

        Check for updates

        Qualifiers

        • short-paper

        Acceptance Rates

        SIGIR'19 Paper Acceptance Rate84of426submissions,20%Overall Acceptance Rate792of3,983submissions,20%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader