Abstract
Most of the software maintenance costs come from usability bugs reported after the release and deployment. A usability bug is really subjective, hence there is a large communication overhead between the end user and the developer. Moreover, the reputation of the software development company could be decreased as well. Therefore, proactively testing and maintaining software systems from a usability point of view is unambiguously beneficial.
In this paper we propose a research prototype, the Usability Evaluation Framework. The development of the framework is driven by welldefined requirements. It is built upon a usability model, it calculates usability metrics, it integrates questionnaires and it also ensures several meaningful reports. We have successfully applied the framework to evaluate and to improve the usability of two industrial software systems.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Andreasen, M.S., Nielsen, H.V., Schrøder, S.O., Stage, J.: What happened to remote usability testing?: an empirical study of three methods. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, New York, NY, USA, pp. 1405–1414 (2007)
Au, F.T.W., Baker, S., Warren, I., Dobbie, G.: Automated usability testing framework. In: Proceedings of the Ninth Conference on Australasian User Interface, Darlinghurst, Australia, vol. 76, pp. 55–64 (2008)
Basit, H.A., Jarzabek, S.: Efficient token based clone detection with flexible tokenization. In: Proceedings of the 6th Joint Meeting of the European Software Engineering Conference and the ACM SIGSOFT Symposium on The Foundations of Software Engineering, ESEC-FSE 2007, pp. 513–516. ACM, New York (2007)
Chilana, P.K., Ko, A.J., Wobbrock, J.O., Grossman, T., Fitzmaurice, G.: Post-deployment usability: a survey of current practices. In: Proceedings of the 2011 Annual Conference on Human Factors in Computing Systems, New York, NY, USA, pp. 2243–2246 (2011)
Homepage of Google Analytics (May 7, 2013), http://www.google.com/analytics/
Harty, J.: Finding usability bugs with automated tests. Communications of the ACM 54, 44–49 (2011)
Hasan, L., Morris, A., Probets, S.: Using Google Analytics to Evaluate the Usability of E-Commerce Sites. In: Kurosu, M. (ed.) HCD 2009. LNCS, vol. 5619, pp. 697–706. Springer, Heidelberg (2009)
Hosseini-Khayat, A., Hellmann, T.D., Maurer, F.: Distributed and Automated Usability Testing of Low-Fidelity Prototypes. In: Proceedings of the AGILE Conference, pp. 59–66 (2010)
Ivory, M.Y., Hearst, M.A.: The state of the art in automating usability evaluation of user interfaces. ACM Computing Surveys 33, 470–516 (2001)
Landauer, T.K.: The Trouble with Computers: Usefulness, Usability, and Productivity. A Bradford Book (1995)
LimeSurvey - Open Source Survey Application (May 7, 2013), http://www.limesurvey.org
Manber, U., Myers, G.: Suffix arrays: a new method for on-line string searches. In: Proceedings of the First Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 1990, pp. 319–327. Society for Industrial and Applied Mathematics, Philadelphia (1990)
Propp, S., Forbrig, P.: ViSE – A virtual smart environment for usability evaluation. In: Bernhaupt, R., Forbrig, P., Gulliksen, J., Lárusdóttir, M. (eds.) HCSE 2010. LNCS, vol. 6409, pp. 38–45. Springer, Heidelberg (2010)
Runge, M.: Simulation of Cognitive Processes for automated Usability Testing. Diploma, Deutche Telekom Laboratories, Berlin (2008)
Schrettner, L., Fülöp, L.J., Beszédes, A., Kiss, A., Gyimóthy, T.: Software Quality Model and Framework with Applications in Industrial Context. In: Proceedings of 16th European Conference on Software Maintenance and Reengineering (CSMR 2012) (2012)
Seffah, A., Donyaee, M., Kline, R.B., Padda, H.K.: Usability measurement and metrics: A consolidated model. Software Quality Control 14(2), 159–178 (2006)
Homepage of TrackerBird (May 7, 2013), http://www.trackerbird.com/
Homepage of UEF (May 7, 2013), http://www.inf.u-szeged.hu/~flajos/usability
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Muhi, K., Szőke, G., Fülöp, L.J., Ferenc, R., Berger, Á. (2013). A Semi-automatic Usability Evaluation Framework. In: Murgante, B., et al. Computational Science and Its Applications – ICCSA 2013. ICCSA 2013. Lecture Notes in Computer Science, vol 7972. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-39643-4_38
Download citation
DOI: https://doi.org/10.1007/978-3-642-39643-4_38
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-39642-7
Online ISBN: 978-3-642-39643-4
eBook Packages: Computer ScienceComputer Science (R0)