Skip to main content
Log in

Multi-type factors representation learning for deep learning-based knowledge tracing

  • Published:
World Wide Web Aims and scope Submit manuscript

Abstract

Knowledge Tracing (KT) refers to the problem of predicting future learner performance given their historical interactions with e-learning platforms. Recent years, Deep Learning-based Knowledge Tracing (DLKT) methods show superior performance than traditional methods due to their strong representational ability. However, researchers usually focus on innovations in model structure, while ignoring the importance of Representation Learning (RL) for DLKT. Investigating previous studies, it is found that the mining and integration of learning-related factors can effectively improve the performance of DLKT models. This paper focuses on providing a model embedding interface for DLKT by considering multiple types of learning-related factors. We first explore and analyze four types of learning-related factors: exercise and skill, the attributes of exercise, learners’ historical performance, and learners’ forgetting behavior in the learning process. We then propose an Extensible Representation Learning (ERL) approach for DLKT to extract and integrate the representations of these four types of factors by setting five components: base embedding, auxiliary embedding, performance embedding, forgetting embedding, and embedding integration. Finally, we apply ERL into two mainstream DLKT models and comprehensively evaluate the proposed approach on several real-world benchmark datasets. Results show that ERL can significantly improve the performances of these two network on predicting future learner responses.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

Notes

  1. The corresponding source code and all preprocessed datasets are available at https://github.com/HLBilove/ERL-master

References

  1. Bengio, Y.: Deep learning of representations for unsupervised and transfer learning. In: Proceedings of ICML Workshop on Unsupervised and Transfer Learning, pp 17–36 (2012)

  2. Bengio, Y., Courville, A., Vincent, P.: Representation learning: a review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35(8), 1798–1828 (2013)

    Article  Google Scholar 

  3. Bengio, Y., Ducharme, R., Vincent, P., Janvin, C.: A neural probabilistic language model. J. Mach. Learn. Res. 3, 1137–1155 (2003)

    MATH  Google Scholar 

  4. Cen, H., Koedinger, K., Junker, B.: Learning factors analysis–a general method for cognitive model evaluation and improvement. In: International Conference on Intelligent Tutoring Systems, pp 164–175 (2016)

  5. Chaudhry, R., Singh, H., Dogga, P., Saini, S.K.: Modeling Hint-Taking behavior and knowledge state of students with Multi-Task learning. Int. Educ. Data Mining Soc. (2018)

  6. Cinquin, P.A., Guitton, P., Sauzéon, H.: Online e-learning and cognitive disabilities: a systematic review. Comput. Educ. 130, 152–167 (2019)

    Article  Google Scholar 

  7. Corbett, A.T., Anderson, J.R.: Knowledge tracing: Modeling the acquisition of procedural knowledge. User Model. User-adapted Interact. 4(4), 253–278 (1994)

    Article  Google Scholar 

  8. Dauphin, G.M.Y., Glorot, X., Rifai, S., Bengio, Y., Goodfellow, I., Lavoie, E., Muller, X., Desjardins, G., Warde-Farley, D., Vincent, P., Bergstra, J, et al.: Unsupervised and transfer learning challenge: a deep learning approach. In: Proceedings of ICML Workshop on Unsupervised and Transfer Learning, pp 97–110 (2012)

  9. Dong, X., Gabrilovich, E., Heitz, G., Horn, W., Lao, N., Murphy, K., Murphy, K., Strohmann, T., Sun, S., Zhang, W., Zhang, W.: Knowledge vault: A web-scale approach to probabilistic knowledge fusion. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp 601–610 (2014)

  10. Dong, G., Zhang, X., Lan, L., Wang, S., Luo, Z.: Label guided correlation hashing for large-scale cross-modal retrieval. Multimed. Tools Appl. (2019)

  11. Ghosh, A., Heffernan, N., Lan, A. S.: Context-aware attentive knowledge tracing. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp 2330–2339 (2020)

  12. He, L.: Integrating performance and side factors into embeddings for deep Learning-Based knowledge tracing. In: 2021 IEEE International Conference on Multimedia and Expo (ICME) (2021)

  13. He, L., Tang, J., Li, X., Wang, T.: ADKT: Adaptive deep knowledge tracing. In: International Conference on Web Information Systems Engineering, pp. 302–314 (2020)

  14. Hinton, G.E.: Learning distributed representations of concepts. In: Proceedings of the Eighth Annual Conference of the Cognitive Science Society, vol. 1, p 12 (1986)

  15. Khajah, M.M., Huang, Y., González-Brenes, J.P., Mozer, M.C., Brusilovsky, P.: Integrating knowledge tracing and item response theory: a tale of two frameworks. CEUR Workshop Proc. 1181, 7–15 (2014)

    Google Scholar 

  16. Khajah, M., Lindsey, R.V., Mozer, M.C.: How deep is knowledge tracing?, arXiv:1604.02416 (2016)

  17. Krishnan, R., Singh, J., Sato, M., Zhang, Q., Ohkuma, T: Incorporating wide context information for deep knowledge tracing using attentional bi-interaction (2021)

  18. Krizhevsky, A., Sutskever, I., Hinton, G. E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84–90 (2017)

    Article  Google Scholar 

  19. Liu, Y., Hua, W., Qu, J., Xin, K., Zhou, X.: Temporal knowledge completion with context-aware embeddings. World Wide Web 24(2), 675–695 (2021)

    Article  Google Scholar 

  20. Liu, Q., Huang, Z., Yin, Y., Chen, E., Xiong, H., Su, Y., Hu, G.: Ekt: Exercise-aware knowledge tracing for student performance prediction. IEEE Trans. Knowl. Data Eng. 33(1), 100–115 (2019)

    Article  Google Scholar 

  21. Liu, K., Liu, W., Ma, H., Huang, W., Dong, X.: Generalized zero-shot learning for action recognition with web-scale video data. World Wide Web 22(2), 807–824 (2019)

    Article  Google Scholar 

  22. Liu, T., Pan, X., Wang, X., Feenstra, K.A., Heringa, J., Huang, Z.: Predicting the relationships between gut microbiota and mental disorders with knowledge graphs. Health Inf. Sci. Syst. 9(1), 1–9 (2021)

    Article  Google Scholar 

  23. Liu, T., Pan, X., Wang, X., Feenstra, K.A., Huang, Z.: Exploring the Microbiota-Gut-Brain axis for mental disorders with knowledge graphs. J. Artif. Intell. Med. Sci. (2020)

  24. Nagatani, K., Zhang, Q., Sato, M., Chen, Y.Y., Chen, F., Ohkuma, T.: Augmenting knowledge tracing by considering forgetting behavior. In: The World Wide Web Conference, pp. 3101–3107 (2019)

  25. Niu, L., Fu, C., Yang, Q., Li, Z., Chen, Z., Liu, Q., Zheng, K.: Open-world knowledge graph completion with multiple interaction attention. World Wide Web 24(1), 419–439 (2021)

    Article  Google Scholar 

  26. Pandey, S., Karypis, G.: A self-attentive model for knowledge tracing. In: Proceedings of the 12th International Conference on Educational Data Mining, pp 384–389 (2019)

  27. Pandey, S., Srivastava, J.: RKT: Relation-aware self-attention for knowledge tracing. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pp 1205–1214 (2020)

  28. Pavlik, P.I. Jr, Cen, H., Koedinger, K.R: Performance factors analysis–A new alternative to knowledge tracing. Online submission (2009)

  29. Piech, C., Bassen, J., Huang, J., Ganguli, S., Sahami, M., Guibas, L.J., Sohl-Dickstein, J.: Long short-term memory. Neural Comput. 8(9), 1735–1780 (1997)

    Google Scholar 

  30. Piech, C., Bassen, J., Huang, J., Ganguli, S., Sahami, M., Guibas, L. J., Sohl-Dickstein, J.: Deep knowledge tracing. Adv. Neural Inf. Process. Syst., 505–513 (2015)

  31. Rollinson, J., Emma, B.: From Predictive models to instructional policies. Int. Educ. Data Mining Soc. (2015)

  32. Wang, Z., Li, L., Zeng, D.: Knowledge-enhanced natural language inference based on knowledge graphs. In: Proceedings of the 28th International Conference on Computational Linguistics (2020)

  33. Wilson, K. H., Karklin, Y., Han, B., Ekanadham, C.: Back to the basics: Bayesian extensions of IRT outperform neural networks for proficiency estimation. arXiv:1604.02336 (2016)

  34. Xie, R., Liu, Z., Jia, J., Luan, H., Sun, M.: Representation learning of knowledge graphs with entity descriptions. In: Proceedings of the AAAI Conference on Artificial Intelligence, 30(1) (2016)

  35. Xie, R., Liu, Z., Jia, J., Luan, H., Sun, M.: Deep hierarchical knowledge tracing. In: Proceedings of the 12th International Conference on Educational Data Mining (2019)

  36. Yeung, C.K.: Deep-IRT: Make deep learning based knowledge tracing explainable using item response theory. arXiv:1904.11738 (2019)

  37. Yeung, C.K., Yeung, D.Y.: Addressing two problems in deep knowledge tracing via Prediction-Consistent regularization. In: Proceedings of the Fifth Annual ACM Conference on Learning at Scale (2018)

  38. Zhang, J., Shi, X., King, I., Yeung, D.Y.: Dynamic key-value memory networks for knowledge tracing. In: Proceedings of the 26th International Conference on World Wide Web, pp 765–774 (2017)

  39. Zhang, L., Xiong, X., Zhao, S., Botelho, A., Heffernan, N.T.: Incorporating rich features into deep knowledge tracing. In: Proceedings of the Fourth ACM Conference on Learning@scale, pp 169–172 (2017)

  40. Zhang, M., Zhu, J., Wang, Z., Chen, Y.: Providing personalized learning guidance in MOOCs by multi-source data analysis. World Wide Web 22 (3), 1189–1219 (2019)

    Article  Google Scholar 

Download references

Acknowledgements

We would like to thank the anonymous reviewers for their helpful comments. The research is supported by the National Natural Science Foundation of China (61702532, 61690203).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ting Wang.

Ethics declarations

Conflict of Interests

The authors declare that they have no conflict of interest.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

He, L., Tang, J., Li, X. et al. Multi-type factors representation learning for deep learning-based knowledge tracing. World Wide Web 25, 1343–1372 (2022). https://doi.org/10.1007/s11280-022-01041-2

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11280-022-01041-2

Keywords

Navigation