Skip to main content

A Tool for Evaluation Using Dynamic Navigation in a Set of Questions

  • Conference paper
Tools for Working with Guidelines
  • 172 Accesses

Abstract

When we use guidelines or questionnaires to evaluate software, some problems can appear: the number of questions increases and thus the duration of an evaluation becomes too long; sometimes evaluation becomes too subjective to be relevant; the choice of questions becomes complex, and so on. To try to solve these problems, in educational context, we have conceived a method (called EMPI: Evaluation of Multimedia Pedagogical and Interactive software) that dynamically builds an evaluation: precision and depth of analysis, choice and weight of questions, are done during the evaluation. We present in this paper, how the evaluator is free to explore questions and to adapt the base to his case. We also explain how he is guided, helped and confronted to his own subjectivity with an aim of building the most relevant evaluation. At the end, we present our validations and experiments done with a prototype that instruments most of characteristics of EMPI method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Similar content being viewed by others

References

  1. Grislin, M., Kolski, C.: Evaluation des interfaces Homme/Machine lors du développe-ment des systèmes interactifs. Technique et Science Informatique 15, 3 (1996)

    Google Scholar 

  2. Scapin, D., Bastien, C.: Ergonomic criteria for evaluating the ergonomic quality of inter-active systems. Behaviour & Information Technology 16 (1997) 226–231

    Google Scholar 

  3. Bastien, C., Scapin, D.: Evaluating a user interface with ergonomic criteria. International Journal of Human Computer Interaction 7 (1995) 105–121

    Article  Google Scholar 

  4. Vanderdonckt, J.: Guide ergonomique de la présentation des applications hautement interactives. Presses Universitairs de Namur, Namur (1994)

    Google Scholar 

  5. Crozat S., Hû O., Trigano, Ph.: EMPI: une méthode informatisé pour l’évaluation des didacticiels multimédias. Revue d’Interaction Homme-Machine 1, 2 (1999)*

    Google Scholar 

  6. Hû O., Trigano, Ph.: Considerating subjectivity in software evaluation — Application for teachware evaluation. In Vanderdonckt, J., Puerta, A. (eds.): Proc. of Computer Aided Design of User Interfaces II CADUI’99. Kluwer Academics Publishers, Dordrecht (1999)*

    Google Scholar 

  7. Vanderdonckt, J.: Development Milestones Towards a Tool for Working with Guidelines. Interacting with Computers 12, 2 (September 1999) 81–118

    Article  Google Scholar 

  8. Linard, M., Zeiliger, R.: Designing an interface in an educational context: first steps to a principle approach. In: Norby, K., Helmersen, P., Gilmore, D., Arnesen, S. (eds.): Proc. of 5’ Int. Conference INTERACT’95 (Lillehammer, June 1995). Chapman & Hall, London (1995) 401–404

    Google Scholar 

  9. Linard M., Belisle C., Zeiliger R.: Reconciling information processing and activity theo-ries: HELICES, a foundation model for creating cognitive artefacts. In: Proc. of ISCRAT98, International Society for Cultural Research and Activity Theory, 4“ Congress, (Aahrus, June 1998)

    Google Scholar 

  10. Lamrous S., Trigano Ph: L’organisation des espaces documentaires: vers une exploitation optimale. Document Electronique 1,4 (December 1997)

    Google Scholar 

  11. MEDA: 61 critères d’évaluation de logiciels de formation-éducation. MédiaScreen, cd-rom. Université de Liège (1997)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag London

About this paper

Cite this paper

Hû, O., Trigano, P. (2001). A Tool for Evaluation Using Dynamic Navigation in a Set of Questions. In: Vanderdonckt, J., Farenc, C. (eds) Tools for Working with Guidelines. Springer, London. https://doi.org/10.1007/978-1-4471-0279-3_26

Download citation

  • DOI: https://doi.org/10.1007/978-1-4471-0279-3_26

  • Publisher Name: Springer, London

  • Print ISBN: 978-1-85233-355-3

  • Online ISBN: 978-1-4471-0279-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics