skip to main content
10.1145/3340531.3412772acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

MIMICS: A Large-Scale Data Collection for Search Clarification

Published:19 October 2020Publication History

ABSTRACT

Search clarification has recently attracted much attention due to its applications in search engines. It has also been recognized as a major component in conversational information seeking systems. Despite its importance, the research community still feels the lack of a large-scale dataset for studying different aspects of search clarification. In this paper, we introduce MIMICS, a collection of search clarification datasets for real web search queries sampled from the Bing query logs. Each clarification in MIMICS is generated by a Bing production algorithm and consists of a clarifying question and up to five candidate answers. MIMICS contains three datasets: (1) MIMICS-Click includes over 400k unique queries, their associated clarification panes, and the corresponding aggregated user interaction signals (i.e., clicks). (2) MIMICS-ClickExplore is an exploration data that includes aggregated user interaction signals for over 60k unique queries, each with multiple clarification panes. (3) MIMICS-Manual includes over 2k unique real search queries. Each query-clarification pair in this dataset has been manually labeled by at least three trained annotators. It contains graded quality labels for the clarifying question, the candidate answer set, and the landing result page for each candidate answer.

MIMICS is publicly available for research purposes, thus enables researchers to study a number of tasks related to search clarification, including clarification generation and selection, user engagement prediction for clarification, click models for clarification, and analyzing user interactions with search clarification. We also release the results returned by the Bing's web search API for all the queries in MIMICS. This would allow researchers to utilize search results for the tasks related to search clarification.

Skip Supplemental Material Section

Supplemental Material

3340531.3412772.mp4

mp4

73.4 MB

References

  1. Mohammad Aliannejadi, Hamed Zamani, Fabio Crestani, and W. Bruce Croft. 2019. Asking Clarifying Questions in Open-Domain Information-Seeking Conversations. In SIGIR '19. 475--484.Google ScholarGoogle Scholar
  2. James Allan. 2004. HARD Track Overview in TREC 2004: High Accuracy Retrieval from Documents. In TREC '04.Google ScholarGoogle ScholarCross RefCross Ref
  3. Avishek Anand, Lawrence Cavedon, Matthias Hagen, Hideo Joho, Mark Sanderson, and Benno Stein. 2020. Conversational Search - A Report from Dagstuhl Seminar 19461. arXiv preprint arXiv:2005.08658 (2020).Google ScholarGoogle Scholar
  4. Marco De Boni and Suresh Manandhar. 2003. An Analysis of Clarification Dialogue for Question Answering. In NAACL '03. 48--55.Google ScholarGoogle Scholar
  5. Pavel Braslavski, Denis Savenkov, Eugene Agichtein, and Alina Dubatovka. 2017. What Do You Mean Exactly?: Analyzing Clarification Questions in CQA. In CHIIR '17. 345--348.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. D. Carmel and E. Yom-Tov. 2010. Estimating the Query Difficulty for Information Retrieval 1st ed.). Morgan and Claypool Publishers.Google ScholarGoogle Scholar
  7. Eunsol Choi, He He, Mohit Iyyer, Mark Yatskar, Wen-tau Yih, Yejin Choi, Percy Liang, and Luke Zettlemoyer. 2018. QuAC: Question Answering in Context. In EMNLP '18. 2174--2184.Google ScholarGoogle Scholar
  8. Anni Coden, Daniel Gruhl, Neal Lewis, and Pablo N. Mendes. 2015. Did you mean A or B? Supporting Clarification Dialog for Entity Disambiguation. In SumPre '15.Google ScholarGoogle Scholar
  9. Nick Craswell, Daniel Campos, Bhaskar Mitra, Emine Yilmaz, and Bodo Billerbeck. 2020. ORCAS: 18 Million Clicked Query-Document Pairs for Analyzing Search. In CIKM '20.Google ScholarGoogle Scholar
  10. Nick Craswell, Onno Zoeter, Michael Taylor, and Bill Ramsey. 2008. An Experimental Comparison of Click Position-Bias Models. In WSDM '08. 87--94.Google ScholarGoogle Scholar
  11. S. Cronen-Townsend, Y. Zhou, and W. B. Croft. 2002. Predicting Query Performance. In SIGIR '02. 299--306.Google ScholarGoogle Scholar
  12. Jeffrey Dalton, Chenyan Xiong, and Jamie Callan. 2019. TREC CAsT 2019: The Conversational Assistance Track Overview. In TREC '19.Google ScholarGoogle Scholar
  13. Marco De Boni and Suresh Manandhar. 2005. Implementing Clarification Dialogues in Open Domain Question Answering. Nat. Lang. Eng., Vol. 11, 4 (2005).Google ScholarGoogle Scholar
  14. Helia Hashemi, Hamed Zamani, and W. Bruce Croft. 2020. Guided Transformer: Leveraging Multiple External Sources for Representation Learning in Conversational Search. In SIGIR '20. 1131--1140.Google ScholarGoogle Scholar
  15. Kalervo J"arvelin and Jaana Kek"al"ainen. 2002. Cumulated Gain-based Evaluation of IR Techniques. ACM Trans. Inf. Syst. , Vol. 20, 4 (2002), 422--446.Google ScholarGoogle Scholar
  16. Johannes Kiesel, Arefeh Bahrami, Benno Stein, Avishek Anand, and Matthias Hagen. 2018. Toward Voice Query Clarification. In SIGIR '18. 1257--1260.Google ScholarGoogle Scholar
  17. Mounia Lalmas, Heather O'Brien, and Elad Yom-Tov. 2014. Measuring User Engagement .Morgan & Claypool Publishers.Google ScholarGoogle Scholar
  18. Chin-Yew Lin and Eduard Hovy. 2003. Automatic Evaluation of Summaries Using N-Gram Co-Occurrence Statistics. In NAACL '03. 71--78.Google ScholarGoogle Scholar
  19. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: A Method for Automatic Evaluation of Machine Translation. In ACL '02. 311--318.Google ScholarGoogle Scholar
  20. Luis Quintano and Irene Pimenta Rodrigues. 2008. Question/Answering Clarification Dialogues. In MICAI '08. 155--164.Google ScholarGoogle Scholar
  21. Filip Radlinski, Krisztian Balog, Bill Byrne, and Karthik Krishnamoorthi. 2019. Coached Conversational Preference Elicitation: A Case Study in Understanding Movie Preferences. In SIGDIAL '19.Google ScholarGoogle Scholar
  22. Filip Radlinski and Nick Craswell. 2017. A Theoretical Framework for Conversational Search. CHIIR '17. 117--126.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. Sudha Rao and Hal Daumé III. 2018. Learning to Ask Good Questions: Ranking Clarification Questions using Neural Expected Value of Perfect Information. ACL '18. 2737--2746.Google ScholarGoogle ScholarCross RefCross Ref
  24. Sudha Rao and Hal Daumé III. 2019. Answer-based Adversarial Training for Generating Clarification Questions. In NAACL '19. 143--155.Google ScholarGoogle Scholar
  25. Siva Reddy, Danqi Chen, and Christopher D. Manning. 2019. CoQA: A Conversational Question Answering Challenge. TACL, Vol. 7 (2019), 249--266.Google ScholarGoogle Scholar
  26. Svetlana Stoyanchev, Alex Liu, and Julia Hirschberg. 2014. Towards Natural Clarification Questions in Dialogue Systems. In AISB '14, Vol. 20.Google ScholarGoogle Scholar
  27. Paul Thomas, Daniel McDuff, Mary Czerwinski, and Nick Craswell. 2017. MISC: A data set of information-seeking conversations. In CAIR '17.Google ScholarGoogle Scholar
  28. Hamed Zamani and Nick Craswell. 2020. Macaw: An Extensible Conversational Information Seeking Platform. In SIGIR '20. 2193--2196.Google ScholarGoogle Scholar
  29. Hamed Zamani, Susan T. Dumais, Nick Craswell, Paul N. Bennett, and Gord Lueck. 2020 a. Generating Clarifying Questions for Information Retrieval. In WWW '20. 418--428.Google ScholarGoogle Scholar
  30. Hamed Zamani, Bhaskar Mitra, Everest Chen, Gord Lueck, Fernando Diaz, Paul N. Bennett, Nick Craswell, and Susan T. Dumais. 2020 b. Analyzing and Learning from User Interactions with Search Clarification. In SIGIR '20. 1181--1190.Google ScholarGoogle Scholar
  31. Hamed Zamani, Pooya Moradi, and Azadeh Shakery. 2015. Adaptive User Engagement Evaluation via Multi-Task Learning. In SIGIR '15. 1011--1014.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Yongfeng Zhang, Xu Chen, Qingyao Ai, Liu Yang, and W. Bruce Croft. 2018. Towards Conversational Search and Recommendation: System Ask, User Respond. In CIKM '18. 177--186.Google ScholarGoogle Scholar

Index Terms

  1. MIMICS: A Large-Scale Data Collection for Search Clarification

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
            October 2020
            3619 pages
            ISBN:9781450368599
            DOI:10.1145/3340531

            Copyright © 2020 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 19 October 2020

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            Overall Acceptance Rate1,861of8,427submissions,22%

            Upcoming Conference

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader