Abstract
In personalized adaptive systems, the learner’s progress toward clearly defined goals is continually assessed, the assessment occurs when a student is ready to demonstrate competency, and supporting materials are tailored to the needs of each learner. Despite the promise of adaptive personalized learning, there is a lack of evidence-based instructional design, transparency in many of the models and algorithms used to provide adaptive technology or a framework for rapid experimentation with different models. ALOSI (Adaptive Learning Open Source Initiative) provides open source adaptive learning technology and a common framework to measure learning gains and learner behavior. This paper provides an overview of adaptive learning functionality developed by Harvard and Microsoft in collaboration with edX and other partners, and shared results the recent deployment in Microsoft MOOC on edX. The study explored the effects of two different strategies for adaptive problems (i.e., assessment items) on knowledge and skills development. We found that the implemented adaptivity in assessment, with emphasis on remediation is associated with a substantial increase in learning gains, while producing no big effect on the drop-out. Further research is needed to confirm these findings and explore additional possible effects and implications to course design.
R. Rubin—Independent.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Everywhere in this paper, by p value we mean the p-value from the two-tailed t-test, and by the effect size (ES) we mean Cohen’s d.
References
Hawkins, W.J., Heffernan, N.T., Baker, R.S.J.D.: Learning Bayesian knowledge tracing parameters with a knowledge heuristic and empirical probabilities. In: Trausan-Matu, S., Boyer, K.E., Crosby, M., Panourgia, K. (eds.) ITS 2014. LNCS, vol. 8474, pp. 150–155. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-07221-0_18
Koedinger, K., Stamper, J.: A data driven approach to the discovery of better cognitive models. In: Baker, R.S.J.d., Merceron, A., Pavlik Jr., P.I. (eds.) Proceedings of the 3rd International Conference on Educational Data Mining (EDM 2010), Pittsburgh, pp. 325–326 (2010)
Pardos, Z., Tang, S., Davis, D., Vu Le, C.: Enabling real-time adaptivity in MOOCs with a personalized next-step recommendation framework. In: Proceedings of the Fourth ACM Conference on Learning @ Scale (2017)
Rosen, Y., Rushkin, I., Ang, A., Federicks, C., Tingley, D., Blink, M.-J.: Designing adaptive assessments in MOOCs. In: Proceedings of the Fourth ACM Conference on Learning @ Scale (2017)
Rushkin, I., Rosen, Y., Ang, A., Fredericks, C., Tingley, D., Blink, M.J., Lopez, G.: Adaptive assessment experiment in a HarvardX MOOC. In: Proceedings of the 10th International Conference on Educational Data Mining (2017)
Stamper, J.C., Eagle, M., Barnes, T., Croy, M.: Experimental evaluation of automatic hint generation for a logic tutor. In: Biswas, G., Bull, S., Kay, J., Mitrovic, A. (eds.) AIED 2011. LNCS (LNAI), vol. 6738, pp. 345–352. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21869-9_45
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Rosen, Y. et al. (2018). Adaptive Learning Open Source Initiative for MOOC Experimentation. In: Penstein Rosé, C., et al. Artificial Intelligence in Education. AIED 2018. Lecture Notes in Computer Science(), vol 10948. Springer, Cham. https://doi.org/10.1007/978-3-319-93846-2_57
Download citation
DOI: https://doi.org/10.1007/978-3-319-93846-2_57
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-93845-5
Online ISBN: 978-3-319-93846-2
eBook Packages: Computer ScienceComputer Science (R0)