ABSTRACT
In today's data-driven educational technologies, algorithms have a pivotal impact on student experiences and outcomes. Therefore, it is critical to take steps to minimize biases, to avoid perpetuating or exacerbating inequalities. In this paper, we investigate the degree to which algorithmic biases are present in two learning analytics models: knowledge estimates based on Bayesian Knowledge Tracing (BKT) and carelessness detectors. Using data from a learning platform used across the United States at scale, we explore algorithmic bias following three different approaches: 1) analyzing the performance of the models on every demographic group in the sample, 2) comparing performance across intersectional groups of these demographics, and 3) investigating whether the models trained using specific groups can be transferred to demographics that were not observed during the training process. Our experimental results show that the performance of these models is close to equal across all the demographic and intersectional groups. These findings establish the feasibility of validating educational algorithms for intersectional groups and indicate that these algorithms can be fairly used for diverse students at scale.
- Ma. Victoria Almeda and Ryan S. Baker. 2020. Predicting Student Participation in STEM Careers: The Role of Affect and Engagement during Middle School. Journal of Educational Data Mining 12, 2 (2020), 33–47.Google Scholar
- Husni Almoubayyed, Stephen E. Fancsali, and Steve Ritter. 2023. Generalizing Predictive Models of Reading Ability in Adaptive Mathematics Software. In Proceedings of the 16th International Conference on Educational Data Mining.Google Scholar
- Ryan S. Baker. 2023. Big Data and Education. 7th Edition. Philadelphia, PA: University of Pennsylvania.Google Scholar
- Ryan S. J. d.Baker, Albert. T. Corbett, and Vincent Aleven. 2008. More accurate student modeling through contextual estimation of slip and guess probabilities in bayesian knowledge tracing. In Intelligent Tutoring Systems: 9th International Conference, ITS 2008, Montreal, Canada, June 23-27, 2008, Proceedings 9, 406-415. Springer Berlin Heidelberg.Google ScholarDigital Library
- Ryan S. J. d. Baker, Albert T. Corbett, Sujith M Gowda, Angela Z Wagner, Benjamin A MacLaren, Linda R Kauffman, Aaron P Mitchell, and Stephen Giguere. 2010. Contextual slip and prediction of student performance after use of an intelligent tutor. In User Modeling, Adaptation, and Personalization: 18th International Conference, UMAP 2010, Big Island, HI, USA, June 20-24, 2010. Proceedings 18, Springer, 52–63.Google Scholar
- Ryan S. Baker and Aaron Hawn. 2022. Algorithmic bias in education. International Journal of Artificial Intelligence in Education (2021), 1–41.Google Scholar
- Leo Breiman. 2017. Classification and regression trees. Routledge.Google Scholar
- Brent Bridgeman, Catherine Trapani, and Yigal Attali. 2012. Comparison of human and machine scoring of essays: Differences by gender, ethnicity, and country. Applied Measurement in Education 25, 1 (2012), 27–40.Google ScholarCross Ref
- Ángel Alexander Cabrera, Will Epperson, Fred Hohman, Minsuk Kahng, Jamie Morgenstern, and Duen Horng Chau. 2019. FairVis: Visual analytics for discovering intersectional bias in machine learning. In 2019 IEEE Conference on Visual Analytics Science and Technology (VAST), IEEE, 46–56.Google ScholarCross Ref
- Nitesh V. Chawla, Kevin W. Bowyer, Lawrence O. Hall, and W. Philip Kegelmeyer. 2002. SMOTE: synthetic minority over-sampling technique. Journal of artificial intelligence research 16, (2002), 321–357.Google Scholar
- John Clement. 1982. Students’ preconceptions in introductory mechanics. American Journal of physics 50, 1, 66–71.Google Scholar
- Kimberle Crenshaw. 1991. Race, gender, and sexual harassment. Sothern California Law Review 65.Google Scholar
- Albert T. Corbett and John R. Anderson. 1995. Knowledge tracing: Modeling the acquisition of procedural knowledge. User Modeling and User-Adapted Interaction 4, 4, 253-278.Google ScholarCross Ref
- Oscar Blessed Deho, Chen Zhan, Jiuyong Li, Jixue Liu, Lin Liu, and Thuc Duy Le. 2022. How do the existing fairness metrics and unfairness mitigation algorithms contribute to ethical learning analytics? British Journal of Educational Technology 53, 4 (2022), 822–843.Google ScholarCross Ref
- Merrill T. Eaton, Louis A. D'Amico, and Beeman N. Phillips. 1956. Problem behavior in school. Journal of Educational Psychology 47, 6, 350.Google ScholarCross Ref
- Stephen Fancsali. 2015. Confounding Carelessness? Exploring Causal Relationships Between Carelessness, Affect, Behavior, and Learning in Cognitive Tutor Algebra Using Graphical Causal Models. In Educational Data Mining, 508–511.Google Scholar
- Josh Gardner, Christopher Brooks, and Ryan Baker. 2019. Evaluating the fairness of predictive student models through slicing analysis. In Proceedings of the 9th international conference on learning analytics & knowledge, 225–234.Google ScholarDigital Library
- Theophile Gervet, Ken Koedinger, Jeff Schneider, Tom Mitchell, and others. 2020. When is deep learning the best approach to knowledge tracing? Journal of Educational Data Mining 12, 3 (2020), 31–54.Google Scholar
- Avijit Ghosh, Lea Genuit, and Mary Reagan. 2021. Characterizing intersectional group fairness with worst-case comparisons. In Artificial Intelligence Diversity, Belonging, Equity, and Inclusion, Proceedings of Machine Learning Research 14, 22–34.Google Scholar
- Qian Hu and Huzefa Rangwala. 2020. Towards Fair Educational Data Mining: A Case Study on Detecting At-Risk Students. International Educational Data Mining Society (2020).Google Scholar
- Yun Huang, Vincent Aleven, Elizabeth McLaughlin, and Kenneth Koedinger. 2020. A general multi-method approach to design-loop adaptivity in intelligent tutoring systems. In Artificial Intelligence in Education: 21st International Conference, AIED 2020, Ifrane, Morocco, July 6–10, 2020, Proceedings, Part II 21, Springer, 124–129.Google ScholarDigital Library
- Mohammad Khalil, Paul Prinsloo, and Sharon Slade. 2023. Fairness, Trust, Transparency, Equity, and Responsibility in Learning Analytics. Journal of Learning Analytics 10, 1 (2023), 1–7.Google ScholarCross Ref
- René F. Kizilcec and Hansol Lee. 2022. Algorithmic fairness in education. In The ethics of artificial intelligence in education. Routledge, 174–202.Google Scholar
- Matt J. Kusner, Joshua Loftus, Chris Russell, and Ricardo Silva. 2017. Counterfactual fairness. Advances in neural information processing systems 30.Google Scholar
- Lin Li, Lele Sha, Yuheng Li, Mladen Raković, Jia Rong, Srecko Joksimovic, Neil Selwyn, Dragan Gašević, and Guanliang Chen. 2023. Moral Machines or Tyranny of the Majority? A Systematic Review on Predictive Bias in Education. In LAK23: 13th International Learning Analytics and Knowledge Conference, 499–508.Google ScholarDigital Library
- Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, and others. 2011. Scikit-learn: Machine learning in Python. Journal of machine Learning research 12, (2011), 2825–2830.Google ScholarDigital Library
- Jaclyn Ocumpaugh, Ryan Baker, Sujith Gowda, Neil Heffernan, and Cristina Heffernan. 2014. Population validity for educational data mining models: A case study in affect detection. British Journal of Educational Technology 45, 3, 487–501.Google ScholarCross Ref
- Amy Ogan, Erin Walker, Ryan Baker, Ma. Mercedes T. Rodrigo, Jose Carlo Soriano, and Maynor Jimenez Castro. 2015. Towards understanding how to assess help-seeking behavior across cultures. International Journal of Artificial Intelligence in Education 25, 229–248.Google ScholarCross Ref
- Luc Paquette, Jaclyn Ocumpaugh, Ziyue Li, Alexandra Andres, and Ryan Baker. 2020. Who's Learning? Using Demographics in EDM Research. Journal of Educational Data Mining 12, 3 (2020), 1–30.Google Scholar
- Radek Pelánek. 2017. Bayesian knowledge tracing, logistic models, and beyond: an overview of learner modeling techniques. User Modeling and User-Adapted Interaction 27, 313–350.Google ScholarDigital Library
- Steven Ritter, Jhon R. Anderson, Kenneth R. Koedinger, and Albert Corbett. 2007. Cognitive Tutor: Applied research in mathematics education. Psychonomic bulletin & review 14, 249-255.Google Scholar
- Nathalie Rzepka, Linda Fernsel, Hans-Georg Müller, Katrarina Simbeck, and Niels Pinkwart. 2023. Unbias me! Mitigating Algorithmic Bias for Less-studied Demographic Groups in the Context of Language Learning Technology. Computer-Based Learning in Context 6, 1, 1-23.Google Scholar
- Nathalie Rzepka, Katharina Simbeck, Hans-Georg Müller, and Niels Pinkwart. 2022. Fairness of In-session Dropout Prediction. In CSEDU (2), 316–326.Google Scholar
- Maria Ofelia San Pedro, Ryan Baker, Alex Bowers, and Neil Heffernan. 2013. Predicting college enrollment from student interaction with an intelligent tutoring system in middle school. In Educational Data Mining 2013.Google Scholar
- Maria Ofelia Z. San Pedro, Ryan S. J. d. Baker, and Ma. Mercedes T Rodrigo. 2014. Carelessness and affect in an intelligent tutoring system for mathematics. International Journal of Artificial Intelligence in Education 24, 189–210.Google ScholarCross Ref
- Maria Ofelia San Pedro, Jaclyn Ocumpaugh, Ryan S Baker, and Neil T Heffernan. 2014. Predicting STEM and Non-STEM College Major Enrollment from Middle School Interaction with Mathematics Educational Software. In Educational Data Mining, 276–279.Google Scholar
- Richard Scruggs, Ryan S. Baker, and Bruce M. McLaren. 2020. Extending Deep Knowledge Tracing: Inferring Interpretable Knowledge and Predicting Post System Performance. In Proceedings of the 28th International Conference on Computers in Education.Google Scholar
- David Sculley, Jasper Snoek, Alex Wiltschko, and Ali Rahimi. 2018. Winner's curse? On pace, progress, and empirical rigor.Google Scholar
- Lele Sha, Mladen Raković, Angel Das, Dragan Gašević, and Guanliang Chen. 2022. Leveraging class balancing techniques to alleviate algorithmic bias for predictive tasks in education. IEEE Transactions on Learning Technologies 15, 4 (2022), 481–492.Google ScholarDigital Library
- Jonathan Vasquez Verdugo, Xavier Gitiaux, Cesar Ortega, and Huzefa Rangwala. 2022. Faired: A systematic fairness analysis approach applied in a higher educational context. In LAK22: 12th International Learning Analytics and Knowledge Conference, 271–281.Google ScholarDigital Library
- Angelina Wang, Vikram V. Ramaswamy, and Olga Russakovsky. 2022. Towards intersectionality in machine learning: Including more identities, handling underrepresentation, and performing evaluation. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, 336–349.Google ScholarDigital Library
- Renzhe Yu, Hansol Lee, and René F Kizilcec. 2021. Should college dropout prediction models include protected attributes? In Proceedings of the eighth ACM conference on learning@ scale, 91–100.Google ScholarDigital Library
- Michael Yudelson, Steve Fancsali, Steve Ritter, Susan Berman, Tristan Nixon, and Ambarish Joshi. 2014. Better data beats big data. In Educational data mining 2014.Google Scholar
- Jiayi Zhang, Rohini Das, Ryan Baker, and Richard Scruggs. 2021. Knowledge tracing models’ predictive performance when a student starts a skill. In Proceedings of the 14th International Conference on Educational Data Mining. EDM, Paris, France, 625–629.Google Scholar
- Jiayi Zhang, Juliana Ma. Alexandra L. Andres, Stephen Hutt, Ryan S. Baker, Jaclyn Ocumpaugh, Nidhi Nasiar, Caitlin Mills, Jamiella Brooks, Sheela Sethuaman, Tyron Young, and others. 2022. Using Machine Learning to Detect SMART Model Cognitive Operations in Mathematical Problem-Solving Process. Journal of Educational Data Mining 14, 3 (2022), 76–108.Google Scholar
Index Terms
- Investigating Algorithmic Bias on Bayesian Knowledge Tracing and Carelessness Detectors
Recommendations
Investigating User Perception of Gender Bias in Image Search: The Role of Sexism
SIGIR '18: The 41st International ACM SIGIR Conference on Research & Development in Information RetrievalThere is growing evidence that search engines produce results that are socially biased, reinforcing a view of the world that aligns with prevalent social stereotypes. One means to promote greater transparency of search algorithms - which are typically ...
Bias in knowledge graph embeddings
ASONAM '20: Proceedings of the 12th IEEE/ACM International Conference on Advances in Social Networks Analysis and MiningIn this paper, we study bias in knowledge graph embeddings. We focus on gender bias in occupations, but our approach is applicable to other types of bias. We start by proposing measures for identifying bias in the dataset (i.e., in the KG) and then ...
Bias against research on gender bias
The bias against women in academia is a documented phenomenon that has had detrimental consequences, not only for women, but also for the quality of science. First, gender bias in academia affects female scientists, resulting in their ...
Comments