Skip to main content
Log in

A computer vision-based perceived attention monitoring technique for smart teaching

  • 1226: Deep-Patterns Emotion Recognition in the Wild
  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

This paper aims to improve the lecture delivery mechanism in real-time in a classroom and remote sessions over web-based applications. In the traditional system, a lecturer observes their students’ attention levels from his/her experience. To date, no system automatically tracks the students’ attention level in a class in real-time (or while the lecturer is delivering his/her lectures remotely over web-based applications). On the other hand, our proposed system periodically will monitor the learning behaviour of the whole class and track the attentiveness of each student. The proposed system is not meant to identify the non-attentive students and punish them. Contrary to the punishment-based mechanism, it introduces a counseling-based mechanism. This deep learning-based real-time face monitoring system will allow lecturers to improvise/her delivery either through bringing diversity in the class contents or personal care to those non-attentive students. The concept of the deep learning technique in an ensemble configuration has been used to predict the likelihood of eyes’ openness. Separately, a student’s facial expressions are also recognized using our Convolutional Neural Network (CNN) model. Finally, the net learning behaviour of a student has been computed by a weighted average of these two features (that is, eyes’ openness and facial expressions). The student learning behaviour is validated twice with Pearson correlation coefficient and Spearman correlation coefficient measures between the openness of eye and facial expressions. Again, the Cosine similarity has been used to further examine the periodical similarity of the student’s learning patterns. The proposed pipeline has performed even better than the state-of-the-art models such as ResNet50, MobileNetV2, and EfficientNet-B0 in terms of accuracy and f1-score.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Algorithm 1
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17

Similar content being viewed by others

Data Availability

The datasets generated during and/or analysed during the current study are available in the following urls: 1. CEW dataset: http://parnec.nuaa.edu.cn/_upload/tpl/02/db/731/template731/pages/xtan/ClosedEyeDatabases.html 2. MRL dataset: http://mrl.cs.vsb.cz/eyedataset. 3. Extended Cohn-Kanade Dataset (CK+) and Perceived Attention GitHub: https://github.com/cserajdeep/Perceived-Attention

Notes

  1. CEW:http://parnec.nuaa.edu.cn/

  2. MRL:http://mrl.cs.vsb.cz/eyedataset

  3. Perceived Attention GitHub: https://github.com/cserajdeep/Perceived-Attention

References

  1. An N, Ding H, Yang J, Au R, Ang TF (2020) Deep ensemble learning for alzheimer’s disease classification. J Biomed Inf 105(103):411

    Google Scholar 

  2. Bdiwi R, de Runz C, Faiz S, Cherif AA (2019) Smart learning environment: teacher’s role in assessing classroom attention. Res Learn Technol :27

  3. Bi AQ, Tian XY, Wang SH, Zhang YD (2022) A dynamic transfer exemplar based facial emotion recognition model towards online video. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM)

  4. Chatterjee R, Maitra T, Islam SH, Hassan MM, Alamri A, Fortino G (2019) A novel machine learning based feature selection for motor imagery eeg signal classification in internet of medical things environment. Futur Gener Comput Syst 98:419–434

    Article  Google Scholar 

  5. Chatterjee R, Roy S, Roy S (2021) International conference on advanced network technologies and intelligent computing, Springer, pp 517–529

  6. Courtad CA (2019) Smart education and e-learning, Springer, pp 501–510

  7. Du S, Martinez AM (2022) Compound facial expressions of emotion: from basic research to clinical applications. Dialogues Clin Neurosci

  8. Francis B, Craig N, Hodgen J, Taylor B, Tereshchenko A, Connolly P, Archer L (2020) The impact of tracking by attainment on pupil self-confidence over time: Demonstrating the accumulative impact of self-fulfilling prophecy. Br J Sociol Educ 41(5):626–642

    Article  Google Scholar 

  9. Ganaie MA, Hu M et al (2021) Ensemble deep learning: a review. arXiv:2104.02395

  10. Jaiswal S, Virmani S, Sethi V, De K, Roy PP (2019) An intelligent recommendation system using gaze and emotion detection. Multimed Tools Appl 78(11):14,231–14,250

    Article  Google Scholar 

  11. Krizhevsky A, Sutskever I, Hinton GE (2012) Advances in neural information processing systems, pp 1097–1105

  12. Kurakin A, Chien S, Song S, Geambasu R, Terzis A, Thakurta A (2022) Toward training at imagenet scale with differential privacy. arXiv:2201.12328

  13. Li TM, Shen WX, Chao HC, Zeadally S (2019) International conference on innovative technologies and learning, Springer, pp 498–504

  14. Liu S, Lu C, Alghowinem S, Gotoh L, Breazeal C, Park HW (2022) International conference on human-computer interaction, Springer, pp 161–178

  15. Lucey P, Cohn JF, Kanade T, Saragih J, Ambadar Z, Matthews I (2010) 2010 IEEE computer society conference on computer vision and pattern recognition-workshops, IEEE, pp 94–101

  16. Mazumdar S, Chatterjee R (2022) Advances in data computing, communication and security, Springer, pp 193–205

  17. Perdiz J, Garrote L, Pires G, Nunes UJ (2021) A reinforcement learning assisted eye-driven computer game employing a decision tree-based approach and cnn classification. IEEE Access 9:46,011–46,021

    Article  Google Scholar 

  18. Riccio A, Simione L, Schettini F, Pizzimenti A, Inghilleri M, Olivetti Belardinelli M, Mattia D, Cincotti F (2013) Attention and P300-based BCI performance in people with amyotrophic lateral sclerosis. Front Hum Neurosci 7:732

    Article  Google Scholar 

  19. Saurav S, Gidde P, Saini R, Singh S (2022) Real-time eye state recognition using dual convolutional neural network ensemble. J Real-Time Image Proc :1–16

  20. Shen F, Ye L, Ma X, Zhong W (2019) 3Rd international seminar on education innovation and economic management (SEIEM 2018) (Atlantis press)

  21. Song F, Tan X, Chen S (2014) Exploiting relationship between attributes for improved face verification. Comput Vis Image Underst 122:143–154

    Article  Google Scholar 

  22. Steil J, Huang MX, Bulling A (2018) Proceedings of the 2018 ACM symposium on eye tracking research & applications, pp 1–9

  23. Wang Y, Huang R, Guo L (2019) Eye gaze pattern analysis for fatigue detection based on gp-bcnn with esm. Pattern Recogn Lett 123:61–74

    Article  Google Scholar 

  24. Wang Y, Lu S, Harter D (2021) Multi-sensor eye-tracking systems and tools for capturing student attention and understanding engagement in learning: a review. IEEE Sensors J

  25. Yoon HS, Baek NR, Truong NQ, Park KR (2019) Driver gaze detection based on deep residual networks using the combined single image of dual near-infrared cameras. IEEE Access 7:93,448–93,461

    Article  Google Scholar 

Download references

Funding

Nil.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tanmoy Maitra.

Ethics declarations

Ethics approval and consent to participate

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards.

Consent for Publication

Consent for Publication Informed consent was obtained from all individual participants included in the study.

Conflict of Interests

The authors declare that they have no conflict of interest.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rajdeep Chatterjee is also affiliated with Amygdala AI, a volunteer-driven global research community www.amygdalaai.org

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chatterjee, R., Halder, R., Maitra, T. et al. A computer vision-based perceived attention monitoring technique for smart teaching. Multimed Tools Appl 82, 11523–11547 (2023). https://doi.org/10.1007/s11042-022-14283-z

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-022-14283-z

Keywords

Navigation