Skip to main content

Evaluation Method of Online Education Learners’ Emotional Input Based on Multimodal Data Fusion

  • Conference paper
  • First Online:
e-Learning, e-Education, and Online Training (eLEOT 2023)

Abstract

In order to better understand learners’ emotional orientation and learning status, an online education learners’ emotional engagement assessment method based on multimodal data fusion was proposed. In order to comprehensively evaluate learners’ emotional engagement, two modal data, comment data and facial expression images, were selected as the evaluation basis. Collect comment data using crawler technology and preprocess word segmentation and part of speech tagging; Extract features from comment data using improved TF-IDF and construct a classifier for comment data using the K-nearest neighbor algorithm. Using a camera to capture facial expression images, and performing lighting transformation, graying, and filtering; Extract HOG features of facial expression images using absolute gradient histogram algorithm, and construct a classifier for facial expression images using Adaboost algorithm. By synthesizing the above two parts of the process, two evaluation results were obtained, with different weights set for each type of single modal data. The weighted sum rule is used to fuse multimodal data at the decision-making level to obtain the final evaluation decision result. The results show that the MSE, MAE and MAPE of the evaluation methods based on review text, facial expression and body movements are relatively smaller than those based on body movements, which indicates that the evaluation accuracy is higher.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Yuan, X., Liu, R., Liu, M.: Multi-feature perception sentiment analysis model. Comput. Simul. 40(4), 509–513 (2023)

    Google Scholar 

  2. Deng, R., Benckendorff, P., Gannaway, D.: Learner engagement in MOOCs: scale development and validation. Br. J. Edu. Technol. 51(1), 245–262 (2020)

    Article  Google Scholar 

  3. Liu, Y., Sun, P., Wergeles, N., et al.: A survey and performance evaluation of deep learning methods for small object detection. Expert Syst. Appl. 172(4), 114602 (2021)

    Article  Google Scholar 

  4. Yu, H., Li, X.: An evaluation model of English teaching effectiveness based on online education. Int. J. Continuing Eng. Educ. Life-Long Learn. 31(2), 218–233 (2021)

    Article  MathSciNet  Google Scholar 

  5. Hu, X.: Analysis and research on the integrated English teaching effectiveness of internet of things based on stochastic forest algorithm. Int. J. Continuing Eng. Educ. Life-Long Learn. 31(1), 1–18 (2022)

    Article  Google Scholar 

  6. John, A., Redmond, S.J., Cardiff, B., et al.: A multimodal data fusion technique for heartbeat detection in wearable IoT sensors. IEEE Internet Things J. 9(3), 2071–2082 (2022)

    Article  Google Scholar 

  7. Bokade, G.U., Kanphade, R.D.: Template security scheme for multimodal biometrics using data fusion technique. Int. J. Biometrics 14(2), 166–190 (2022)

    Article  Google Scholar 

  8. Syed, M., Pirogova, E., Lech, M.: Prediction of public trust in politicians using a multimodal fusion approach. Electronics 10(11), 1259 (2021)

    Article  Google Scholar 

  9. Mou, L., Zhou, C., Zhao, P., et al.: Driver stress detection via multimodal fusion using attention-based CNN-LSTM. Expert Syst. Appl. 173(12), 114693 (2021)

    Article  Google Scholar 

  10. Li, H., Huang, J., Huang, J., et al.: Deep multimodal learning and fusion based intelligent fault diagnosis approach. J. Beijing Inst. Technol. 30(2), 172–185 (2021)

    Google Scholar 

Download references

Aknowledgement

Shanxi Provincial Education Department: The Research and Practice of the Cultivation Model of Chemical Engineering Talents Highlighting the Deep Coordination of Basic Practice and Specialization (J2020060).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yong Zhang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, Y., Ren, E., Song, Y., Chen, F. (2024). Evaluation Method of Online Education Learners’ Emotional Input Based on Multimodal Data Fusion. In: Gui, G., Li, Y., Lin, Y. (eds) e-Learning, e-Education, and Online Training. eLEOT 2023. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 546. Springer, Cham. https://doi.org/10.1007/978-3-031-51503-3_27

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-51503-3_27

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-51502-6

  • Online ISBN: 978-3-031-51503-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics