skip to main content
10.1145/2362724.2362757acmotherconferencesArticle/Chapter ViewAbstractPublication PagesiiixConference Proceedingsconference-collections
research-article

Visual interactive failure analysis: supporting users in information retrieval evaluation

Published: 21 August 2012 Publication History

Abstract

Measuring is a key to scientific progress. This is particularly true for research concerning complex systems, whether natural or human-built. Multilingual and multimedia information access systems, such as search engines, are increasingly complex: they need to satisfy diverse user needs and support challenging tasks. Their development calls for proper evaluation methodologies to ensure that they meet the expected user requirements and provide the desired effectiveness. In this context, failure analysis is crucial to understand the behaviour of complex systems. Unfortunately, this is an especially challenging activity, requiring vast amounts of human effort to inspect query-by-query the output of a system in order to understand what went well or bad. It is therefore fundamental to provide automated tools to examine system behaviour, both visually and analytically. Moreover, once you understand the reason behind a failure, you still need to conduct a "what-if" analysis to understand what among the different possible solutions is most promising and effective before actually starting to modify your system. This paper provides an analytical model for examining performances of IR systems, based on the discounted cumulative gain family of metrics, and visualization for interacting and exploring the performances of the system under examination. Moreover, we propose machine learning approach to learn the ranking model of the examined system in order to be able to conduct a "what-if" analysis and visually explore what can happen if you adopt a given solution before having to actually implement it.

References

[1]
{Banks et al., 1999} Banks, D., Over, P., and Zhang, N.-F. (1999). Blind Men and Elephants: Six Approaches to TREC data. Information Retrieval, 1: 7--34.
[2]
{Berkhin, 2006} Berkhin, P. (2006). A Survey of Clustering Data Mining Techniques. In Kogan, J., Nicholas, C., and Teboulle, M., editors, Grouping Multidimensional Data, pages 25--71. Springer-Verlag, Heidelberg, Germany.
[3]
{Derthick et al., 2003a} Derthick, M., Christel, M. G., Hauptmann, A. G., and Wactlar, H. D. (2003a). Constant density displays using diversity sampling. In Proceedings of InfoVis'03, pages 137--144, Washington, DC, USA. IEEE Computer Society.
[4]
{Derthick et al., 2003b} Derthick, M., Christel, M. G., Hauptmann, A. G., and Wactlar, H. D. (2003b). Constant density displays using diversity sampling. In Proceedings of the IEEE Information Visualization, pages 137--144.
[5]
{Freund et al., 2003} Freund, Y., Iyer, R., Schapire, R. E., and Singer, Y. (2003). An Efficient Boosting Algorithm for Combining Preferences. Journal of Machine Learning Research, 4(Nov): 933--969.
[6]
{Geng et al., 2007} Geng, X., Liu, T.-Y., Qin, T., and Li, H. (2007). Feature Selection for Ranking. In Kraaij, W., de Vries, A. P., Clarke, C. L. A., Fuhr, N., and Kando, N., editors, Proc. 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2007), pages 407--414. ACM Press, New York, USA.
[7]
{Harman and Buckley, 2009} Harman, D. and Buckley, C. (2009). Overview of the reliable information access workshop. Information Retrieval, 12(6): 615--641.
[8]
{Järvelin and Kekäläinen, 2002} Järvelin, K. and Kekäläinen, J. (2002). Cumulated Gain-Based Evaluation of IR Techniques. ACM Transactions on Information System (TOIS), 20(4): 422--446.
[9]
{Keskustalo et al., 2008} Keskustalo, H., Järvelin, K., Pirkola, A., and Kekäläinen, J. (2008). Intuition-Supporting Visualization of User's Performance Based on Explicit Negative Higher-Order Relevance. In Chua, T.-S., Leong, M.-K., Oard, D. W., and Sebastiani, F., editors, Proc. 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2008), pages 675--682. ACM Press, New York, USA.
[10]
{Korfhage, 1997} Korfhage, R. R. (1997). Information Storage and Retrieval. Wiley Computer Publishing, John Wiley & Sons, Inc., USA.
[11]
{Liu, 2009} Liu, T.-Y. (2009). Learning to Rank for Information Retrieval. Foundations and Trends in Information Retrieval, 3(3): 225--331.
[12]
{Liu et al., 2007} Liu, T.-Y. Y., Xu, J., Qin, T., Xiong, W., and Li, H. (2007). LETOR: Benchmark Dataset for Research on Learning to Rank for Information Retrieval. In Joachims, T., Li, H., Liu, T.-Y., and Zhai, C., editors, SIGIR 2007 Workshop on Learning to Rank for Information Retrieval.
[13]
{Seo and Shneiderman, 2004} Seo, J. and Shneiderman, B. (2004). A rank-by-feature framework for interactive exploration of multidimensional data. In Proceedings of the IEEE Information Visualization, pages 65--72.
[14]
{Seo and Shneiderman, 2005} Seo, J. and Shneiderman, B. (2005). A rank-by-feature framework for interactive exploration of multidimensional data. Information Visualization, 4: 96--113.
[15]
{Sormunen et al., 2002} Sormunen, E., Hokkanen, S., Kangaslampi, P., Pyy, P., and Sepponen, B. (2002). Query Performance Analyser -- a Web-based tool for IR research and instruction. In Järvelin, K., Beaulieu, M., Baeza-Yates, R., and Hyon Myaeng, S., editors, Proc. 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2002), page 450. ACM Press, New York, USA.
[16]
{Teevan et al., 2010} Teevan, J., Dumais, S. T., and Horvitz, E. (2010). Potential for personalization. ACM Transactions on Computer-Human Interaction (TOCHI), 17(1): 1--31.
[17]
{van Rijsbergen, 1979} van Rijsbergen, C. J. (1979). Information Retrieval. Butterworths, London, England, 2nd edition.
[18]
{Voorhees and Harman, 1999} Voorhees, E. and Harman, D. (1999). Overview of the Seventh Text REtrieval Conference (TREC-7). In NIST Special Publication 500--242: The Seventh Text REtrieval Conference (TREC 7). Springer-Verlag, Heidelberg, Germany.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
IIIX '12: Proceedings of the 4th Information Interaction in Context Symposium
August 2012
347 pages
ISBN:9781450312820
DOI:10.1145/2362724
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

  • University of Amsterdam: The University of Amsterdam

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 21 August 2012

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. best practices
  2. data test collection
  3. evaluation infrastructure
  4. experimental evaluation
  5. scientific data

Qualifiers

  • Research-article

Conference

IIiX'12
Sponsor:
  • University of Amsterdam
IIiX'12: Information Interaction in Context: 2012
August 21 - 24, 2012
Nijmegen, The Netherlands

Acceptance Rates

Overall Acceptance Rate 21 of 45 submissions, 47%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)1
Reflects downloads up to 01 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2024)A Blueprint of IR Evaluation Integrating Task and User CharacteristicsACM Transactions on Information Systems10.1145/367516242:6(1-38)Online publication date: 1-Jul-2024
  • (2018)RETRIEVAL—An Online Performance Evaluation Tool for Information Retrieval MethodsIEEE Transactions on Multimedia10.1109/TMM.2017.271619320:1(119-127)Online publication date: 25-Dec-2018
  • (2018)VIRTUEJournal of Visual Languages and Computing10.1016/j.jvlc.2013.12.00325:4(394-413)Online publication date: 27-Dec-2018
  • (2016)The twist measure for IR evaluationJournal of the Association for Information Science and Technology10.1002/asi.2341667:3(620-648)Online publication date: 1-Mar-2016
  • (2014)A Visual Interactive Environment for Making Sense of Experimental DataProceedings of the 36th European Conference on IR Research on Advances in Information Retrieval - Volume 841610.5555/2964060.2964135(767-770)Online publication date: 13-Apr-2014
  • (2013)Improving Ranking Evaluation Employing Visual AnalyticsProceedings of the 4th International Conference on Information Access Evaluation. Multilinguality, Multimodality, and Visualization - Volume 813810.1007/978-3-642-40802-1_4(29-40)Online publication date: 23-Sep-2013
  • (2012)PROMISE retreat report prospects and opportunities for information access evaluationACM SIGIR Forum10.1145/2422256.242226546:2(60-84)Online publication date: 21-Dec-2012

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media