Abstract
When we use guidelines or questionnaires to evaluate software, some problems can appear: the number of questions increases and thus the duration of an evaluation becomes too long; sometimes evaluation becomes too subjective to be relevant; the choice of questions becomes complex, and so on. To try to solve these problems, in educational context, we have conceived a method (called EMPI: Evaluation of Multimedia Pedagogical and Interactive software) that dynamically builds an evaluation: precision and depth of analysis, choice and weight of questions, are done during the evaluation. We present in this paper, how the evaluator is free to explore questions and to adapt the base to his case. We also explain how he is guided, helped and confronted to his own subjectivity with an aim of building the most relevant evaluation. At the end, we present our validations and experiments done with a prototype that instruments most of characteristics of EMPI method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Grislin, M., Kolski, C.: Evaluation des interfaces Homme/Machine lors du développe-ment des systèmes interactifs. Technique et Science Informatique 15, 3 (1996)
Scapin, D., Bastien, C.: Ergonomic criteria for evaluating the ergonomic quality of inter-active systems. Behaviour & Information Technology 16 (1997) 226–231
Bastien, C., Scapin, D.: Evaluating a user interface with ergonomic criteria. International Journal of Human Computer Interaction 7 (1995) 105–121
Vanderdonckt, J.: Guide ergonomique de la présentation des applications hautement interactives. Presses Universitairs de Namur, Namur (1994)
Crozat S., Hû O., Trigano, Ph.: EMPI: une méthode informatisé pour l’évaluation des didacticiels multimédias. Revue d’Interaction Homme-Machine 1, 2 (1999)*
Hû O., Trigano, Ph.: Considerating subjectivity in software evaluation — Application for teachware evaluation. In Vanderdonckt, J., Puerta, A. (eds.): Proc. of Computer Aided Design of User Interfaces II CADUI’99. Kluwer Academics Publishers, Dordrecht (1999)*
Vanderdonckt, J.: Development Milestones Towards a Tool for Working with Guidelines. Interacting with Computers 12, 2 (September 1999) 81–118
Linard, M., Zeiliger, R.: Designing an interface in an educational context: first steps to a principle approach. In: Norby, K., Helmersen, P., Gilmore, D., Arnesen, S. (eds.): Proc. of 5’ Int. Conference INTERACT’95 (Lillehammer, June 1995). Chapman & Hall, London (1995) 401–404
Linard M., Belisle C., Zeiliger R.: Reconciling information processing and activity theo-ries: HELICES, a foundation model for creating cognitive artefacts. In: Proc. of ISCRAT98, International Society for Cultural Research and Activity Theory, 4“ Congress, (Aahrus, June 1998)
Lamrous S., Trigano Ph: L’organisation des espaces documentaires: vers une exploitation optimale. Document Electronique 1,4 (December 1997)
MEDA: 61 critères d’évaluation de logiciels de formation-éducation. MédiaScreen, cd-rom. Université de Liège (1997)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag London
About this paper
Cite this paper
Hû, O., Trigano, P. (2001). A Tool for Evaluation Using Dynamic Navigation in a Set of Questions. In: Vanderdonckt, J., Farenc, C. (eds) Tools for Working with Guidelines. Springer, London. https://doi.org/10.1007/978-1-4471-0279-3_26
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0279-3_26
Publisher Name: Springer, London
Print ISBN: 978-1-85233-355-3
Online ISBN: 978-1-4471-0279-3
eBook Packages: Springer Book Archive