Abstract
Systems enabling Information Access whether it be; a file browser, a retrieval engine, a mobile device providing content, a personalized agent, etc, need to be evaluated appropriately for the discipline to be considered a science. The problem of how to appropriately evaluate such systems is even more problematic when the evaluation is conducted with human subjects.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Bailey, A., Ruthven, I., Azzopardi, L. (2005). Evaluating User Studies in Information Access. In: Crestani, F., Ruthven, I. (eds) Context: Nature, Impact, and Role. CoLIS 2005. Lecture Notes in Computer Science, vol 3507. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11495222_21
Download citation
DOI: https://doi.org/10.1007/11495222_21
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-26178-0
Online ISBN: 978-3-540-32101-9
eBook Packages: Computer ScienceComputer Science (R0)