Abstract
In the post-COVID-19 era, online learning has changed from an emergency teaching method to a new, normalized one. However, online learning is often plagued by low participation and high dropout compared to offline learning. A critical way to address these issues is the accurate detection of student engagement, which will help teachers promptly assess learners’ status. Image data are one of the most straightforward ways to reflect student engagement levels. However, traditional engagement detection methods with images either rely on manual analysis or interfere with student behavior, which leads to a need for more objectivity in final engagement levels. This paper proposes a system that utilizes images obtained from individual webcams in online classrooms. Based on the techniques of multi-dimensional feature fusion and multimodal analysis, this system can rapidly detect and output students’ classroom engagement levels which provides real-time support for teachers to adjust their teaching methods during the teaching process, aiming to enhance students’ engagement in online courses. In the feature extraction module, VGG16 is utilized to recognize students’ facial expressions, ResNet-101 is designed to estimate head pose in each image, and Mediapipe is applied to estimate facial landmarks that reflect eye–mouth behavior. Subsequently, a BP neural network is constructed to fuse these multi-dimensional features and output the engagement level in each image in the data fusion module. The present method is evaluated on the wacv2016 data set and achieves an accuracy of 62.03%, outperforming the single-dimensional method. It is also applied in online courses to demonstrate its validity in the actual scenario further. Pearson Correlation between engagement levels calculated by our multi-dimensional fusion method and NSSE–China survey scores filled out by students is 0.714. It indicates that the method can enable real-time monitoring of students’ classroom engagement with similar results to traditional questionnaires with little human resources and time.
Similar content being viewed by others
Data availability
The datasets generated and analyzed during the current study are available from the corresponding authors upon reasonable request.
References
Webb, M.E., et al.: “Machine learning for human learners: Opportunities, issues, tensions and threats.” Educ. Technol. Res. Dev. 69, 2109–2130 (2021)
Blankenberger, B., Williams, A.M.: COVID and the impact on higher education: The essential role of integrity and accountability. Adm. Theory Praxis 42(3), 404–423 (2020)
Yates, A., et al.: High school students’ experience of online learning during Covid-19 the influence of technology and pedagogy”. Technol. Pedag. Educ. 30(1), 59–73 (2021)
Muangprathub, J., Boonjing, V., Chamnongthai, K.: Learning recommendation with formal concept analysis for intelligent tutoring system. Heliyon 6(10), e05227 (2020)
Hofer, S.I., Nistor, N., Scheibenzuber, C.: Online teaching and learning in higher education: Lessons learned in crisis situations. Comput. Hum. Behav. 121, 106789 (2021)
Xue, E., Li, J., Liujie, Xu.: Online education action for defeating COVID-19 in China: An analysis of the system, mechanism and mode. Educ. Philos. Theory 54(6), 799–811 (2022)
Rapanta, C., et al.: “Online university teaching during and after the Covid-19 crisis: Refocusing teacher presence and learning activity.” Postdigital Sci. Educ. 2, 923–945 (2020)
Rawashdeh, Al., Zuhir, A., et al.: Advantages and disadvantages of using e-learning in university education: Analyzing students’ perspectives. Electron. J. E-learning 19(3), 107–117 (2021)
Pelletier, K., et al.: EDUCAUSE Horizon report teaching and learning edition. EDUC22, (2022)
Hensley, L.C., Iaconelli, R., Wolters, C.A.: “This weird time we’re in”: How a sudden change to remote education impacted college students’ self-regulated learning. J. Res. Technol. Educ. 54(sup1), S203–S218 (2022)
Bozkurt, A., Sharma, R.C.: Education in normal, new normal, and next normal: Observations from the past, insights from the present and projections for the future. Asian J. Distance Educ. 15(2), i–x (2020)
Fredricks, J.A., Blumenfeld, P.C., Paris, A.H.: School engagement: Potential of the concept, state of the evidence. Rev. Educ. Res. 74(1), 59–109 (2004)
Phan, T., McNeil, S.G., Robin, B.R.: Students’ patterns of engagement and course performance in a Massive Open Online Course. Comput. Educ. 95, 36–44 (2016)
Fisher, R., Perényi, A., Birdthistle, N.: The positive relationship between flipped and blended learning and student engagement, performance and satisfaction. Act. Learn. High. Educ. 22(2), 97–113 (2021)
Hussain, M., et al.: “Student engagement predictions in an e-learning system and their impact on student course assessment scores.” Computational Intell. Neurosci. (2018). https://doi.org/10.1155/2018/6347186
Walker, K.A., Koralesky, K.E.: “Student and instructor perceptions of engagement after the rapid online transition of teaching due to COVID-19.” Nat. Sci. Educ. 50(1), e20038 (2021)
Kim, C., et al.: “From motivation to engagement: The role of effort regulation of virtual high school students in mathematics courses.” J. Educ. Technol. Soc. 18(4), 261–272 (2015)
Martin, F., Borup, J.: “Online learner engagement: Conceptual definitions, research themes, and supportive practices.” Educ. Psychol. 57(3), 162–177 (2022)
Wang, Ze., Bergin, C., Bergin, D.A.: “Measuring engagement in fourth to twelfth grade classrooms: The classroom engagement inventory.” Sch. Psychol. Q. 29(4), 517 (2014)
Özhan, ŞÇ., Kocadere, S.A.: “The effects of flow emotional engagement and motivation on success in a gamified online learning environment.” J. Educ. Comput. Res. 57(8), 2006–2031 (2020)
D'Errico, F., Marinella P., Luca C.: “When emotions enhance students’ engagement in e-learning processes.” J. E-Learning Knowl. Soc. 12(4), (2016).
Anthonysamy, L., Koo, A.C., Hew, S.H.: “Self-regulated learning strategies in higher education: Fostering digital literacy for sustainable lifelong learning.” Educ. Inf. Technol. 25, 2393–2414 (2020)
Dupeyrat, C., Mariné, C.: Implicit theories of intelligence, goal orientation, cognitive engagement, and achievement: A test of Dweck’s model with returning to school adults. Contemp. Educ. Psychol. 30(1), 43–59 (2005)
Alqurashi, E.: What do students engage with the most? A comparative study between high and low achieving students within online learning environments. Open Learning J. Open Distance E-Learning 37(3), 219–234 (2022)
Greene, B.A.: “Measuring cognitive engagement with self-report scales: Reflections from over 20 years of research.” Educ. Psychol. 50(1), 14–30 (2015)
O’Brien, H.L., Toms, E.G.: “The development and evaluation of a survey to measure user engagement.” J. Am. Soc. Inform. Sci. Technol. 61(1), 50–69 (2010)
Ober, T.M., et al.: “Linking self-report and process data to performance as measured by different assessment types.” Comp. Educ. 167, 104188 (2021)
Volpe, R.J., et al.: “Observing students in classroom settings: A review of seven coding schemes.” Sch. Psychol. Rev. 34(4), 454–474 (2005)
Wu, Y., et al.: “Digital twin of intelligent small surface defect detection with cyber-manufacturing systems.” ACM Trans. Internet Technol. (2022). https://doi.org/10.1145/3571734
Wu, Y., et al.: “Edge computing driven low-light image dynamic enhancement for object detection.” IEEE Transactions Netw. Sci. Eng. (2022). https://doi.org/10.1109/TNSE.2022.3151502
Ashwin, T.S., Guddeti, R.M.R.: “Automatic detection of students’ affective states in classroom environment using hybrid convolutional neural networks”. Educ. Inf. Technol. 25(2), 1387–1415 (2020)
Sharma, P., et al.: “Student engagement detection using emotion analysis, eye tracking and head movement with machine learning.” Technology and Innovation in Learning, Teaching and Education: Third International Conference, TECH-EDU 2022, Lisbon, Portugal, August 31–September 2, 2022, Revised Selected Papers. Cham: Springer Nature Switzerland, (2023)
Dewan, M., Murshed, M., Lin, F.: Engagement detection in online learning: A review. Smart Learning Environ. 6(1), 1–20 (2019)
Altuwairqi, K., et al.: Student behavior analysis to measure engagement levels in online learning environments”. Sign. Image Video Process. 15(7), 1387–1395 (2021)
Aluja-Banet, T., Sancho, M.-R., Vukic, I.: Measuring motivation from the virtual learning environment in secondary education. J. Computational Sci. 36, 100629 (2019)
Ghergulescu, I., Muntean, C.H.: “ToTCompute: A novel EEG-based TimeOnTask threshold computation mechanism for engagement modelling and monitoring.” Int. J. Artif. Intell. Educ. 26(3), 821–854 (2016)
KerdawyMohamed, El., et al.: “The automatic detection of cognition using eeg and facial expressions.” Sensors 20(12), 3516 (2020)
Bustos-López, M., et al.: “Wearables for engagement detection in learning environments: A review.” Biosensors 12(7), 509 (2022)
Cacioppo, J.T., Tassinary, L.G.: Inferring psychological significance from physiological signals. Am. Psychol. 45(1), 16 (1990)
Uzelac, A., Gligoric, N., Krco, S.: “A comprehensive study of parameters in physical environment that impact students’ focus during lecture using internet of things”. Comput. Hum. Behav. 53, 427–434 (2015)
Whitehill, J., et al.: “The faces of engagement: Automatic recognition of student engagement from facial expressions.” IEEE Trans. Affect. Comput. 5(1), 86–98 (2014)
Yu, H., et al.: “Measuring and integrating facial expressions and head pose as indicators of engagement and affect in tutoring systems.” International Conference on Human-Computer Interaction. Springer, Cham, (2021)
Buono, P., et al.: “Assessing student engagement from facial behavior in on-line learning.” Multimedia Tools and Applications, 1–19 (2022)
Shen, J., et al.: “Assessing learning engagement based on facial expression recognition in MOOC’s scenario.” Multimed. Syst. 28(2), 469–478 (2022)
Savchenko, A.V., Savchenko, L.V., Makarov, I.: “Classifying emotions and engagement in online learning based on a single facial expression recognition neural network.” IEEE Trans. Affect. Comput. 13(4), 2132–2143 (2022)
Yirui, W., Ma, Y., Wan, S.: “Multi-scale relation reasoning for multi-modal visual question answering.” Sign. Process. Image Commun. (2021). https://doi.org/10.1016/j.image.2021.116319
Zhang, W., et al.: “Transformer-based multimodal information fusion for facial expression analysis.” Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. (2022)
Garrison, D.R., Cleveland-Innes, M.: “Facilitating cognitive presence in online learning: Interaction is not enough.” Am J. Distance Educ. 19(3), 133–148 (2005)
Smilek, D., Carriere, J.S.A., Allan Cheyne, J.: “Out of mind, out of sight: Eye blinking as indicator and embodiment of mind wandering.” Psychol. Sci. 21(6), 786–789 (2010)
Al-Madani, Ali Mansour, et al.: “Real-time driver drowsiness detection based on eye movement and yawning using facial landmark.” 2021 International Conference on Computer Communication and Informatics (ICCCI). IEEE, (2021)
Pabba, C., Kumar, P.: “An intelligent system for monitoring students’ engagement in large classroom teaching through facial expression recognition”. Expert. Syst. 39(1), e12839 (2022)
Li, X., et al.: “Multi-window Transformer parallel fusion feature pyramid network for pedestrian orientation detection.” Multime. Syst. (2022). https://doi.org/10.1007/s00530-022-00993-9
Monkaresi, H., et al.: “Automated detection of engagement using video-based estimation of facial expressions and heart rate.” IEEE Trans. Affect. Comput. 8(1), 15–28 (2016)
Zhang, Z., et al.: “Data-driven online learning engagement detection via facial expression and mouse behavior recognition technology.” J. Educ. Comput. Res. 58(1), 63–86 (2020)
Xiaoming, C., et al.: Research on student engagement recognition method from the perspective of artificial intelligence: Analysis of deep learning experiment based on a multimodal data fusion. J. Dist. Educ. 01, 32–44 (2019)
Lugaresi, C., et al.: “Mediapipe: A framework for building perception pipelines.” arXiv preprint arXiv:1906.08172 (2019)
Cech, J., Tereza S.: “Real-time eye blink detection using facial landmarks.” Cent. Mach. Perception, Dep. Cybern. Fac. Electr. Eng. Czech Tech. Univ. Prague. 1–8 (2016)
Carrier, Pierre-Luc, et al.: “FER-2013 face database.” Universit de Montral 3 (2013)
He, Kaiming, et al.: “Deep residual learning for image recognition.” Proceedings of the IEEE conference on computer vision and pattern recognition. (2016)
Ruiz, N., Chong, E., Rehg, J.M.: Fine-grained head pose estimation without keypoints[C]//Proceedings of the IEEE conference on computer vision and pattern recognition workshops. 2074–2083, 2018
Wang, H., Zhenghua C., Yi Z.: “Hybrid coarse-fine classification for head pose estimation.” arXiv preprint arXiv:1901.06778 (2019)
Wu, Y., Ji, Q.: “Facial landmark detection: A literature survey.” Int. J. Comput. Vision 127, 115–142 (2019)
Kamath, A., Aradhya B., Vineeth B.: “A crowdsourced approach to student engagement recognition in e-learning environments.” 2016 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, (2016)
Gupta, A., et al.: “Daisee: Towards user engagement recognition in the wild.” arXiv preprint arXiv:1609.01885 (2016)
Zheng, X., et al.: “Estimation of learners’ engagement using face and body features by transfer learning.” Artificial Intelligence in HCI: Second International Conference, AI-HCI 2021, Held as Part of the 23rd HCI International Conference, HCII 2021, Virtual Event, July 24–29, 2021, Proceedings. Cham: Springer International Publishing, (2021)
Huang, T., et al.: “Fine-grained engagement recognition in online learning environment.” 2019 IEEE 9th international conference on electronics information and emergency communication (ICEIEC). IEEE, (2019)
Kamath, S., et al.: “Engagement analysis of students in online learning environments.” Machine learning and big data analytics (Proceedings of International Conference on Machine Learning and Big Data Analytics (ICMLBDA) 2021). Springer International Publishing, (2022)
Yonghe, Z., et al.: Learning engagement detection based on face dataset in the mixed scene. Mod. Educ. Technol. 31(10), 84–92 (2021)
Park, Y.-M., Lee, G.-M., Yang, H.-S.: “Deep feature based efficient regularised ensemble for engagement recognition.” Electron. Lett. 55(24), 1281–1283 (2019)
Batra, S., et al.: “DMCNet: Diversified model combination network for understanding engagement from video screengrabs.” Syst. Soft Comput. 4, 200039 (2022)
Liao, J., Liang, Y., Pan, J.: “Deep facial spatiotemporal network for engagement prediction in online learning.” Appl. Intell. 51, 6609–6621 (2021)
Tanwar, S., Vinay K., Shailza S.: “Engagement measurement of a learner during e-learning: A deep learning architecture.” 2022 Seventh International Conference on Parallel, Distributed and Grid Computing (PDGC). IEEE, (2022)
Abedi, A., Shehroz S. K.: “Improving state-of-the-art in detecting student engagement with resnet and tcn hybrid network.” 2021 18th Conference on Robots and Vision (CRV). IEEE, (2021)
Monahan, T., Fisher, J.A.: “Benefits of ‘observer effects’: lessons from the field.” Qual. Res. 10(3), 357–376 (2010)
Goldberg, P., et al.: “Attentive or not? Toward a machine learning approach to assessing students’ visible engagement in classroom instruction”. Educ. Psychol. Rev. 33, 27–49 (2021)
Yan, L., et al.: Higher education measurement in the context of globalization—The development of NSSE-China: cultural adaptation, reliability and validity. Fudan Educ. Forum. 05, 12–18 (2009)
Author information
Authors and Affiliations
Contributions
NX proposed the work concept and designed the manuscript. ZL and ZL collected online videos and processed raw data. WP and BL made important revisions to the manuscript. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest.
Additional information
Communicated by A. Liu.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Xie, N., Liu, Z., Li, Z. et al. Student engagement detection in online environment using computer vision and multi-dimensional feature fusion. Multimedia Systems 29, 3559–3577 (2023). https://doi.org/10.1007/s00530-023-01153-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00530-023-01153-3