Abstract
Educational institutions currently favor the adoption and use of modern information and communication technology for teaching and learning. As a result, many Learning Management Systems (LMS), which are online, content-management systems specifically designed to fulfil this function, have been developed over the few years and established in education systems. This paper reports the results of a study that experimentally compared the usability of two LMSs: Blackboard (BB) and Desire2Learn (D2L), at Umm Al-Qura University in the Kingdom of Saudi Arabia. The comparison involved four points: the quantity and quality of usability problems discovered, effectiveness of the targeted systems, their efficiency, and users’ satisfaction. The results of the study revealed that there are significant differences between the two systems. The evaluation discovered a higher number of major usability problems on BB. Participants were also significantly faster in completing their tasks on D2L, made fewer errors and mouse clicks, and visited fewer pages to achieve the task on D2L than on BB. However, interestingly no difference was found between the two systems in terms of users’ satisfaction. Taken together, the findings suggest that appropriate customization of Blackboard should be considered in order to enhance its usability and meet the specific requirements of its end users at the current institution.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
About Us. https://www.blackboard.com/about-us. Accessed 10 July 2019
Adelsberger, H.H., Collis, B., Pawlowski, J.M.: Handbook on Information Technologies for Education and Training. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-540-74155-8
Agariya, A.K., Singh, D.: E-learning quality: scale development and validation in Indian context, knowledge management & e-learning. Int. J. (KM&EL) 4(4), 500–517 (2012)
Albert, W., Tullis, T.: Measuring the User Experience: Collecting, Analyzing, and Presenting Usability Metrics. Newnes, Oxford (2013)
Alhadreti, O., Mayhew, P.: To intervene or not to intervene: an investigation of three think-aloud protocols in usability testing. J. Usability Stud. 12(3), 111–132 (2017)
Alturki, U.T., Aldraiweesh, A.: Evaluating the usability and accessibility of LMS “Blackboard” at King Saud University. Contemp. Issues Educ. Res. (CIER) 9(1), 33–44 (2016)
Aydin, B., Darwish, M.M., Selvi, E.: The state-of-the-art matrix analysis for usability of learning management systems. ASEE Comput. Educ. (CoED) J. 7(4), 48 (2016)
Blandford, A., Hyde, J., Green, T., Connell, I.: Scoping analytical usability evaluation methods: a case study. Hum.-Comput. Interact. 23(3), 278–327 (2008)
Brooke, J.: SUS - a quick and dirty usability scale. Usability Eval. Ind. 189(194), 4–7 (1996)
Brown, M.: Evaluating computer game usability: developing heuristics based on user experience. In: Proceedings of IHCI Conference, pp. 16–21 (2008)
Cakmak, E.K., Gunes, E., Ciftci, S., Ustundag, M.T.: Developing a web site usability scale. The validity and reliability analysis & implementation results. Pegem Eğitim Ve Öğreti I(II), 31–40 (2011)
Chawdhry, A., Paullet, K., Benjamin, D.: Assessing blackboard: improving online instructional delivery. Inf. Syst. Educ. J. 9(4), 20 (2011)
Dillon, A.: The Evaluation of Software Usability. Taylor and Francis, London (2001)
Docebo. https://eclass.teicrete.gr/modules/document/file.php/TP271/Additional%20material/docebo-elearning-trends-report-2017.pdf. Accessed 11 Sept 2019 and 21 Aug 2019
Dumas, J.S., Redish, J.: A Practical Guide to Usability Testing. Intellect, Bristol (1999)
Ebling, M.R., John, B.E.: On the contributions of different empirical data in usability testing. In: Proceedings of the 3rd Conference on Designing Interactive Systems: Processes, Practices, Methods, and Techniques, pp. 289–296. ACM, New York (2000)
Ericsson, K.A., Simon, H.A.: Protocol Analysis: Verbal Reports as Data. Revised edn. MIT Press, Cambridge (1993)
Hall, M., De Jong, M., Steehouder, M.: Cultural differences and usability evaluation: individualistic and collectivistic participants compared. Tech. Commun. 51(4), 489–503 (2004)
Hertzum, M., Hansen, K.D., Andersen, H.H.K.: Scrutinising usability evaluation: does thinking aloud affect behaviour and mental workload? Behav. Inf. Technol. 28(2), 165–181 (2009)
Hertzum, M.: Problem prioritization in usability evaluation: from severity assessments toward impact on design. Int. J. Hum. Comput. Interact. 21(2), 125–146 (2006)
Hornbaek, K.: Dogmas in the assessment of usability evaluation methods. Behav. Inf. Technol. 29(1), 97–111 (2010)
How to Use the System Usability Scale (SUS) to Evaluate the Usability of Your Website. http://usabilitygeek.com/how-to-use-the-system-usability-scale-sus-to-evaluate-theusability. Accessed 11 June 2019
Inversini, A., Botturi, L., Triacca, L.: Evaluating LMS usability for enhanced e-learning experience. In: Inedmedia + Innovate Learning Association for the Advancement of Computing in Education (AACE), pp. 595–601 (2006)
ISO W. 9241-11: Ergonomic Requirements for Office Work with Visual Display Terminals (VDTs). The International Organization for Standardization (1998)
Koutsabasis, P., Spyrou, T., Darzentas, J.: Evaluating usability evaluation methods: criteria, method and a case study. In: Jacko, J.A. (ed.) HCI 2007. LNCS, vol. 4550, pp. 569–578. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-73105-4_63
Kraleva, R., Sabani, M., Kralev, V.: An analysis of some learning management systems. Int. J. Adv. Sci. Eng. Inf. Technol. 9(4), 1190–1198 (2019)
Lalande, N., Grewal, R.: Blackboard vs. Desire2Learn: a system administrator’s perspective on usability. In: International Conference on Education and E-Learning Innovations, pp. 1–4. IEEE (2012)
Lavery, D., Cockton, G., Atkinson, M.P.: Comparison of evaluation methods using structured usability problem reports. Behav. Inf. Technol. 16(4–5), 246–266 (1997)
Lazar, J., Feng, J.H., Hochheiser, H.: Research Methods in Human-Computer Interaction. Wiley, Hoboken (2010)
Lindgaard, G., Chattratichart, J.: Usability testing: what have we overlooked? In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 1415–1424. ACM, New York (2007)
Maguire, M.: Context of use within usability activities. Int. J. Hum. Comput. Stud. 55(4), 453–483 (2001)
Mea. https://www.d2l.com/en-mea/. Accessed 23 Aug 2019
Melton, J.: The LMS moodle: a usability evaluation. Lang. Issues 11/12(1), 1–24 (2006)
Molich, R., Dumas, J.: Comparative usability evaluation (CUE 4). Behav. Inf. Technol. 27(3), 263–281 (2008)
Saudi Arabia. https://gs.statcounter.com/browser-market-share/all/saudi-arabia/2019. Accessed 06 Oct 2019
Sauro, J.: A Practical Guide to Measuring Usability: 72 Answers to the Most Common Questions About Quantifying the Usability of Websites and Software. A Measuring Usability, LLC (2010)
Tips and Tricks for Recruiting Users. http://www.nngroup.com/reports/tips/recruiting/234_recruiting_tips. Accessed 17 Feb 2019
Zhao, T., Mcdonald, S., Edwards, H.M.: The impact of two different think aloud instructions in a usability test: a case of just following orders? Behav. Inf. Technol. 33(2), 163–183 (2012)
Acknowledgements
The author would like to thank all those people who took time to take part in the experiments. Thanks also to the anonymous reviewers for their helpful comments.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Alhadreti, O. (2020). A Comparative Usability Study of Blackboard and Desire2Learn: Students’ Perspective. In: Zaphiris, P., Ioannou, A. (eds) Learning and Collaboration Technologies. Designing, Developing and Deploying Learning Experiences. HCII 2020. Lecture Notes in Computer Science(), vol 12205. Springer, Cham. https://doi.org/10.1007/978-3-030-50513-4_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-50513-4_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50512-7
Online ISBN: 978-3-030-50513-4
eBook Packages: Computer ScienceComputer Science (R0)