skip to main content
10.1145/3020165.3022141acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
short-paper

Using Topic Models to Assess Document Relevance in Exploratory Search User Studies

Published:07 March 2017Publication History

ABSTRACT

Evaluation is crucial in assessing the effectiveness of new information retrieval and human computer interaction techniques and systems. Relevance judgements are often performed by humans, which makes obtaining them expensive and time consuming. Consequently, relevance judgements are usually performed only on a subset of a given collection of data or experimental results with a focus on the top ranked documents. However, when assessing the performance of exploratory search systems, the diversity or subjective relevance of documents that the user was presented with over a search session are often of more importance than the relative ranking of top documents. In order to perform these types of assessment, all the documents in a given collection need to be judged for relevance. In this paper, we propose an approach based on topic modeling that can greatly accelerate document relevance judgment of an entire document collection with an expert assessor needing to mark only a small subset of documents from a given collection. Experimental results show a substantial overlap between relevance judgments compared to a human assessor.

References

  1. K. Ahukorala, A. Medlar, K. Ilves, and D. Glowacka. Balancing exploration and exploitation: Empirical parameterization of exploratory search systems. In Proc. CIKM, pages 1703--1706. ACM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. D. Andrzejewski and D. Buttler. Latent topic feedback for information retrieval. In Proc. SIGKDD, pages 600--608. ACM, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. K. Athukorala, D. Glowacka, G. Jacucci, A. Oulasvirta, and J. Vreeken. Is exploratory search different? a comparison of information search behavior for exploratory and lookup tasks. Journal of the Association for Information Science and Technology, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. K. Athukorala, A. Medlar, A. Oulasvirta, G. Jacucci, and D. Glowacka. Beyond relevance: Adapting exploration/exploitation in information retrieval. In Proc. IUI, pages 359--369. ACM, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. D. M. Blei. Probabilistic topic models. Communications of the ACM, 55(4):77--84, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent dirichlet allocation. JLMR, 3:993--1022, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. C. Buckley, D. Dimmick, I. Soboroff, and E. Voorhees. Bias and the limits of pooling for large collections. Information retrieval, 10(6):491--508, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. G. V. Cormack, C. R. Palmer, and C. L. Clarke. Efficient construction of large test collections. In Proc. SIGIR, pages 282--289. ACM, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. D. Glowacka, T. Ruotsalo, K. Konyushkova, K. Athukorala, S. Kaski, and G. Jacucci. Directing exploratory search: Reinforcement learning from user interactions with keywords. In Proc. IUI, pages 117--128. ACM, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. D. Greene, D. O'Callaghan, and P. Cunningham. How many topics? stability analysis for topic models. In Proc. ECML PKDD, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  11. K. L. Gwet. Handbook of inter-rater reliability: The definitive guide to measuring the extent of agreement among raters. Advanced Analytics, LLC, 2014.Google ScholarGoogle Scholar
  12. A. Kangasraasio, D. Glowacka, and S. Kaski. Improving controllability and predictability of interactive recommendation interfaces for exploratory search. In Proc. IUI, pages 247--251. ACM, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. D. E. Losada, J. Parapar, and A. Barreiro. Feeling lucky?: multi-armed bandits for ordering judgements in pooling-based evaluation. In Proc. SAC, pages 1027--1034. ACM, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. G. Marchionini. Exploratory search: from finding to understanding. Communications of the ACM, 49(4):41--46, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. A. K. McCallum. Mallet: A machine learning for language toolkit. http://mallet.cs.umass.edu, 2002.Google ScholarGoogle Scholar
  16. A. Medlar, K. Ilves, P. Wang, W. Buntine, and D. Glowacka. PULP: A system for exploratory search of scientific literature. In Proc. SIGIR, pages 1133--1136. ACM, 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. A. Moffat, W. Webber, and J. Zobel. Strategic system comparisons via targeted relevance judgments. In Proc. SIGIR, pages 375--382. ACM, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. M. Sanderson and J. Zobel. Information retrieval system evaluation: effort, sensitivity, and reliability. In Proc. SIGIR, pages 162--169. ACM, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. E. M. Voorhees, D. K. Harman, et al. TREC: Experiment and evaluation in information retrieval. MIT Press, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. X. Wei and W. B. Croft. LDA-based document models for ad-hoc retrieval. In Proc. SIGIR. ACM, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. R. W. White, G. Marchionini, and G. Muresan. Evaluating exploratory search systems: Introduction to special topic issue of information processing and management. Information Processing & Management, 44(2):433--436, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. X. Yi and J. Allan. A comparative study of utilizing topic models for information retrieval. In Proc. ECIR, pages 29--41, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Using Topic Models to Assess Document Relevance in Exploratory Search User Studies

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          CHIIR '17: Proceedings of the 2017 Conference on Conference Human Information Interaction and Retrieval
          March 2017
          454 pages
          ISBN:9781450346771
          DOI:10.1145/3020165
          • Conference Chairs:
          • Ragnar Nordlie,
          • Nils Pharo,
          • Program Chairs:
          • Luanne Freund,
          • Birger Larsen,
          • Dan Russel

          Copyright © 2017 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 7 March 2017

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • short-paper

          Acceptance Rates

          CHIIR '17 Paper Acceptance Rate10of48submissions,21%Overall Acceptance Rate55of163submissions,34%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader