ABSTRACT
In this paper, we explore the trade-off between fairness and accuracy when data is biased and unbiased. We introduce two versions of a modified loss function: Group Equity and Group Equality Binary Cross Entropy (BCE). Both functions add the maximum group loss to the standard BCE loss and use a hyperparameter to adjust its significance. Group Equity BCE is more suited for unbiased data, while Group Equality BCE is useful for heavily bias data. We test both by training logistic regression models on Home Mortgage Disclosure Act (HMDA) data. We evaluate our data and models using multiple fairness metrics, including p-value and statistical parity difference. We observe race and race-sex group bias. Our results show that Group Equity BCE maintains high accuracy while slightly improving group fairness. Group Equality BCE in- creases group fairness substantially. However, there is usually a steep decline in accuracy.
- Alekh Agarwal, Miroslav Dudik, and Zhiwei Steven Wu. 2019. Fair Regression: Quantitative Definitions and Reduction-Based Algorithms. In Proceedings of the 36th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, Long Beach, USA, 120--129. https://proceedings.mlr.press/v97/agarwal19d.htmlGoogle Scholar
- Marco Almada. 2019. Human Intervention in Automated Decision-Making: Toward the Construction of Contestable Systems. In Proceedings of the Seventeenth International Conference on Artificial Intelligence and Law (ICAIL '19). Association for Computing Machinery, Montreal, Canada, 2--11. https://doi.org/10.1145/3322640.3326699Google ScholarDigital Library
- McKane Andrus and Sarah Villeneuve. 2022. Demographic-Reliant Algorithmic Fairness: Characterizing the Risks of Demographic Data Collection in the Pursuit of Fairness. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (FAccT '22). Association for Computing Machinery, Seoul, Republic of Korea, 1709--1721. https://doi.org/10.1145/3531146.3533226Google ScholarDigital Library
- Mustafa Atay, Hailey Gipson, Tony Gwyn, and Kaushik Roy. 2021. Evaluation of Gender Bias in Facial Recognition with Traditional Machine Learning Algorithms. In 2021 IEEE Symposium Series on Computational Intelligence (SSCI). Orlando, USA, 1--7. https://doi.org/10.1109/SSCI50451.2021.9660186Google ScholarCross Ref
- Ricardo Baeza-Yates. 2018. Bias on the Web. Commun. ACM 61, 6 (May 2018), 54--61. https://doi.org/10.1145/3209581Google ScholarDigital Library
- Ryan S. Baker and Aaron Hawn. 2022. Algorithmic Bias in Education. International Journal of Artificial Intelligence in Education 32, 4 (Dec 2022), 1052--1092.Google ScholarCross Ref
- Rachel K. E. Bellamy, Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Kalapriya Kannan, Pranay Lohia, Jacquelyn Martino, Sameep Mehta, Aleksandra Mojsilović, Seema Nagar, Karthikeyan Natesan Ramamurthy, John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh, Kush R. Varshney, and Yunfeng Zhang. 2019. AI Fairness 360: An Extensible Toolkit for Detecting and Mitigating Algorithmic Bias. IBM Journal of Research and Development 63, 4/5 (2019), 4:1-4:15. https://doi.org/10.1147/JRD.2019.2942287Google ScholarCross Ref
- Yuriy Brun and Alexandra Meliou. 2018. Software Fairness. In Proceedings of the 2018 26th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE 2018). Association for Computing Machinery, Lake Buena Vista, USA, 754--759. https://doi.org/10.1145/3236024.3264838Google ScholarDigital Library
- Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency (Proceedings of Machine Learning Research, Vol. 81), Sorelle A. Friedler and Christo Wilson (Eds.). PMLR, Atlanta, USA, 77--91. https://proceedings.mlr.press/v81/buolamwini18a.htmlGoogle Scholar
- William Cai, Ro Encarnacion, Bobbie Chern, Sam Corbett-Davies, Miranda Bogen, Stevie Bergman, and Sharad Goel. 2022. Adaptive Sampling Strategies to Construct Equitable Training Datasets. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (FAccT '22). Association for Computing Machinery, Seoul, Republic of Korea, 1467--1478. https://doi.org/10.1145/3531146.3533203Google ScholarDigital Library
- Richard J. Chen, Judy J. Wang, Drew F. K. Williamson, Tiffany Y. Chen, Jana Lipkova, Ming Y. Lu, Sharifa Sahai, and Faisal Mahmood. 2023. Algorithmic Fairness in Artificial Intelligence for Medicine and Healthcare. Nat Biomed Eng 7, 6 (June 2023), 719--742.Google ScholarCross Ref
- Valeriia Cherepanova, Steven Reich, Samuel Dooley, Hossein Souri, John Dickerson, Micah Goldblum, and Tom Goldstein. 2023. A Deep Dive into Dataset Imbalance and Bias in Face Identification. In Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society (AIES '23). Association for Computing Machinery, Montreal, Canada, 229--247. https://doi.org/10.1145/3600211.3604691Google ScholarDigital Library
- Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic Decision Making and the Cost of Fairness. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD '17). Association for Computing Machinery, Halifax, Canada, 797--806. https://doi.org/10.1145/3097983.3098095Google ScholarDigital Library
- Bo Cowgill, Fabrizio Dell'Acqua, Samuel Deng, Daniel Hsu, Nakul Verma, and Augustin Chaintreau. 2020. Biased Programmers? Or Biased Data? A Field Experiment in Operationalizing AI Ethics. In Proceedings of the 21st ACM Conference on Economics and Computation (EC '20). Association for Computing Machinery, Virtual Event, Hungary, 679--681. https://doi.org/10.1145/3391403.3399545Google ScholarDigital Library
- Alexander D'Amour, Hansa Srinivasan, James Atwood, Pallavi Baljekar, David Sculley, and Yoni Halpern. 2020. Fairness is Not Static: Deeper Understanding of Long Term Fairness via Simulation Studies. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT* '20). Association for Computing Machinery, Barcelona, Spain, 525--534. https://doi.org/10.1145/3351095.3372878Google ScholarDigital Library
- Jonathan Dodge, Q. Vera Liao, Yunfeng Zhang, Rachel K. E. Bellamy, and Casey Dugan. 2019. Explaining Models: An Empirical Study of How Explanations Impact Fairness Judgment. In Proceedings of the 24th International Conference on Intelligent User Interfaces (IUI '19). Association for Computing Machinery, Marina del Ray, USA, 275--285. https://doi.org/10.1145/3301275.3302310Google ScholarDigital Library
- Mengnan Du, Fan Yang, Na Zou, and Xia Hu. 2021. Fairness in Deep Learning: A Computational Perspective. IEEE Intelligent Systems 36, 4 (2021), 25--34. https://doi.org/10.1109/MIS.2020.3000681Google ScholarDigital Library
- Hadi Elzayn, Shahin Jabbari, Christopher Jung, Michael Kearns, Seth Neel, Aaron Roth, and Zachary Schutzman. 2019. Fair Algorithms for Learning in Allocation Problems. In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* '19). Association for Computing Machinery, Atlanta, USA, 170--179. https://doi.org/10.1145/3287560.3287571Google ScholarDigital Library
- Alessandro Fabris, Stefano Messina, Gianmaria Silvello, and Gian Antonio Susto. 2022. Algorithmic Fairness Datasets: The Story so Far. Data Mining and Knowledge Discovery 36, 6 (Nov 2022), 2074--2152. https://doi.org/10.1007/s10618-022-00854-zGoogle ScholarDigital Library
- Sina Fazelpour and Zachary C. Lipton. 2020. Algorithmic Fairness from a Non-Ideal Perspective. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES '20). Association for Computing Machinery, New York, USA, 57--63. https://doi.org/10.1145/3375627.3375828Google ScholarDigital Library
- Sainyam Galhotra, Karthikeyan Shanmugam, Prasanna Sattigeri, and Kush R. Varshney. 2022. Causal Feature Selection for Algorithmic Fairness. In Proceedings of the 2022 International Conference on Management of Data (SIGMOD '22). Association for Computing Machinery, Philadelphia, USA, 276--285. https://doi.org/10.1145/3514221.3517909Google ScholarDigital Library
- Ben Green and Yiling Chen. 2019. Disparate Interactions: An Algorithm-in-the-Loop Analysis of Fairness in Risk Assessments. In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* 19). Association for Computing Machinery, Atlanta, USA, 90--99. https://doi.org/10.1145/3287560.3287563Google ScholarDigital Library
- Matan Halevy, Camille Harris, Amy Bruckman, Diyi Yang, and Ayanna Howard. 2021. Mitigating Racial Biases in Toxic Language Detection with an Equity-Based Ensemble Framework. In Proceedings of the 1st ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization (EAAMO '21). Association for Computing Machinery, New York, USA, Article 7, 11 pages. https://doi.org/10.1145/3465416.3483299Google ScholarDigital Library
- Hoda Heidari, Michele Loi, Krishna P. Gummadi, and Andreas Krause. 2019. A Moral Framework for Understanding Fair ML through Economic Models of Equality of Opportunity. In Proceedings of the Conference on Fairness, Accountability, and Transparency (FAT* '19). Association for Computing Machinery, Atlanta, USA, 181--190. https://doi.org/10.1145/3287560.3287584Google ScholarDigital Library
- Kenneth Holstein, Jennifer Wortman Vaughan, Hal Daumé, Miro Dudik, and Hanna Wallach. 2019. Improving Fairness in Machine Learning Systems: What Do Industry Practitioners Need?. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI '19). Association for Computing Machinery, Glasgow, Scotland, 1--16. https://doi.org/10.1145/3290605.3300830Google ScholarDigital Library
- Gabbrielle M. Johnson. 2021. Algorithmic Bias: On the Implicit Biases of Social Technology. Synthese 198, 10 (Oct 2021), 9941--9961.Google ScholarCross Ref
- Maria Kasinidou, Styliani Kleanthous, Pinar Barlas, and Jahna Otterbacher. 2021. I Agree with the Decision, but They Didn't Deserve This: Future Developers' Perception of Fairness in Algorithmic Decisions. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (FAccT '21). Association for Computing Machinery, Virtual Event, Canada, 690--700. https://doi.org/10.1145/3442188.3445931Google ScholarDigital Library
- Atoosa Kasirzadeh and Andrew Smart. 2021. The Use and Misuse of Counterfactuals in Ethical Machine Learning. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (FAccT '21). Association for Computing Machinery, Virtual Event, Canada, 228--236. https://doi.org/10.1145/3442188.3445886Google ScholarDigital Library
- Michael Katell, Meg Young, Dharma Dailey, Bernease Herman, Vivian Guetler, Aaron Tam, Corinne Bintz, Daniella Raz, and P. M. Krafft. 2020. Toward Situated Interventions for Algorithmic Equity: Lessons from the Field. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT* '20). Association for Computing Machinery, Barcelona, Spain, 45--55. https://doi.org/10.1145/3351095.3372874Google ScholarDigital Library
- Jon Kleinberg. 2018. Inherent Trade-Offs in Algorithmic Fairness. In Abstracts of the 2018 ACM International Conference on Measurement and Modeling of Computer Systems (SIGMETRICS '18). Association for Computing Machinery, Irvine, USA, 40. https://doi.org/10.1145/3219617.3219634Google ScholarDigital Library
- Anoop Krishnan, Ali Almadan, and Ajita Rattani. 2020. Understanding Fairness of Gender Classification Algorithms Across Gender-Race Groups. In 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA). Miami, USA, 1028--1035. https://doi.org/10.1109/ICMLA51294.2020.00167Google ScholarCross Ref
- Kimmo Kärkkäinen and Jungseock Joo. 2021. FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age for Bias Measurement and Mitigation. In 2021 IEEE Winter Conference on Applications of Computer Vision (WACV). Virtual Event, 1547--1557. https://doi.org/10.1109/WACV48630.2021.00159Google ScholarCross Ref
- Min Kyung Lee, Daniel Kusbit, Anson Kahng, Ji Tae Kim, Xinran Yuan, Allissa Chan, Daniel See, Ritesh Noothigattu, Siheon Lee, Alexandros Psomas, and Ariel D. Procaccia. 2019. WeBuildAI: Participatory Framework for Algorithmic Governance. Proc. ACM Hum.-Comput. Interact. 3, CSCW, Article 181 (Nov 2019), 35 pages. https://doi.org/10.1145/3359283Google ScholarDigital Library
- Michelle Seng Ah Lee, Luciano Floridi, and Jatinder Singh. 2021. Formalising Trade-Offs Beyond Algorithmic Fairness: Lessons from Ethical Philosophy and Welfare Economics. AI and Ethics 1, 4 (Nov 2021), 529--544.Google Scholar
- Jinyang Li, Yuval Moskovitch, and H. V. Jagadish. 2021. DENOUNCER: Detection of Unfairness in Classifiers. Proc. VLDB Endow. 14, 12 (July 2021), 2719--2722. https://doi.org/10.14778/3476311.3476328Google ScholarDigital Library
- Michael A. Madaio, Luke Stark, Jennifer Wortman Vaughan, and Hanna Wallach. 2020. Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in AI. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI '20). Association for Computing Machinery, Honolulu, USA, 1--14. https://doi.org/10.1145/3313831.3376445Google ScholarDigital Library
- Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, and Aram Galstyan. 2021. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 54, 6, Article 115 (July 2021), 35 pages. https://doi.org/10.1145/3457607Google ScholarDigital Library
- Vishwali Mhasawade, Yuan Zhao, and Rumi Chunara. 2021. Machine Learning and Algorithmic Fairness in Public and Population Health. Nature Machine Intelligence 3, 8 (Aug 2021), 659--666. https://doi.org/10.1038/s42256-021-00373-4Google ScholarCross Ref
- Joshua R. Minot, Nicholas Cheney, Marc Maier, Danne C. Elbers, Christopher M. Danforth, and Peter Sheridan Dodds. 2022. Interpretable Bias Mitigation for Textual Data: Reducing Genderization in Patient Notes While Maintaining Classification Performance. ACM Trans. Comput. Healthcare 3, 4, Article 39 (Nov 2022), 41 pages. https://doi.org/10.1145/3524887Google ScholarDigital Library
- Shira Mitchell, Eric Potash, Solon Barocas, Alexander D'Amour, and Kristian Lum. 2021. Algorithmic Fairness: Choices, Assumptions, and Definitions. Annual Review of Statistics and Its Application 8, 1 (2021), 141--163. https://doi.org/10.1146/annurev-statistics-042720-125902 arXiv:https://doi.org/10.1146/annurev-statistics-042720-125902Google ScholarCross Ref
- Deirdre K. Mulligan, Joshua A. Kroll, Nitin Kohli, and Richmond Y. Wong. 2019. This Thing Called Fairness: Disciplinary Confusion Realizing a Value in Technology. Proc. ACM Hum.-Comput. Interact. 3, CSCW, Article 119 (Nov 2019), 36 pages. https://doi.org/10.1145/3359221Google ScholarDigital Library
- Mohammadmehdi Naghiaei, Hossein A. Rahmani, and Yashar Deldjoo. 2022. CPFair: Personalized Consumer and Producer Fairness Re-Ranking for Recommender Systems. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR '22). Association for Computing Machinery, Madrid, Spain, 770--779. https://doi.org/10.1145/3477495.3531959Google ScholarDigital Library
- Mustafa Safa Ozdayi and Murat Kantarcioglu. 2021. The Impact of Data Distribution on Fairness and Robustness in Federated Learning. In 2021 Third IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA). Atlanta, USA, 191--196. https://doi.org/10.1109/TPSISA52974.2021.00022Google ScholarCross Ref
- Trishan Panch, Heather Mattie, and Rifat Atun. 2019. Artificial Intelligence and Algorithmic Bias: Implications for Health Systems. J Glob Health 9, 2 (Dec 2019), 010318.Google ScholarCross Ref
- Eliana Pastor, Andrew Gavgavian, Elena Baralis, and Luca de Alfaro. 2021. How Divergent is Your Data? Proc. VLDB Endow. 14, 12 (July 2021), 2835--2838. https://doi.org/10.14778/3476311.3476357Google ScholarDigital Library
- Romila Pradhan, Jiongli Zhu, Boris Glavic, and Babak Salimi. 2022. Interpretable Data-Based Explanations for Fairness Debugging. In Proceedings of the 2022 International Conference on Management of Data (SIGMOD '22). Association for Computing Machinery, Philadelphia, USA, 247--261. https://doi.org/10.1145/3514221.3517886Google ScholarDigital Library
- Public HMDA. 2018. Public HMDA - LAR Data Fields. https://ffiec.cfpb.gov/documentation/publications/loan-level-datasets/lar-data-fieldsGoogle Scholar
- Manish Raghavan, Solon Barocas, Jon Kleinberg, and Karen Levy. 2020. Mitigating Bias in Algorithmic Hiring: Evaluating Claims and Practices. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (FAT* '20). Association for Computing Machinery, Barcelona, Spain, 469--481. https://doi.org/10.1145/3351095.3372828Google ScholarDigital Library
- Inioluwa Deborah Raji and Joy Buolamwini. 2019. Actionable Auditing: Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (AIES '19). Association for Computing Machinery, Honolulu, USA, 429--435. https://doi.org/10.1145/3306618.3314244Google ScholarDigital Library
- Ricardo Salazar, Felix Neutatz, and Ziawasch Abedjan. 2021. Automated Feature Engineering for Algorithmic Fairness. Proc. VLDB Endow. 14, 9 (May 2021), 1694--1702. https://doi.org/10.14778/3461535.3463474Google ScholarDigital Library
- Nripsuta Ani Saxena, Karen Huang, Evan DeFilippis, Goran Radanovic, David C. Parkes, and Yang Liu. 2019. How Do Fairness Definitions Fare? Examining Public Attitudes Towards Algorithmic Definitions of Fairness. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society (AIES '19). Association for Computing Machinery, Honolulu, USA, 99--106. https://doi.org/10.1145/3306618.3314248Google ScholarDigital Library
- Megha Srivastava, Hoda Heidari, and Andreas Krause. 2019. Mathematical Notions vs. Human Perception of Fairness: A Descriptive Approach to Fairness for Machine Learning. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD '19). Association for Computing Machinery, Anchorage, USA, 2459--2468. https://doi.org/10.1145/3292500.3330664Google ScholarDigital Library
- Harini Suresh and John Guttag. 2021. A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle. In Proceedings of the 1st ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization (-, NY, USA) (EAAMO '21). Association for Computing Machinery, New York, USA, Article 17, 9 pages. https://doi.org/10.1145/3465416.3483305Google ScholarDigital Library
- Sakshi Udeshi, Pryanshu Arora, and Sudipta Chattopadhyay. 2018. Automated Directed Fairness Testing. In Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering (ASE '18). Association for Computing Machinery, Montpellier, France, 98--108. https://doi.org/10.1145/3238147.3238165Google ScholarDigital Library
- Michael Veale, Max Van Kleek, and Reuben Binns. 2018. Fairness and Accountability Design Needs for Algorithmic Support in High-Stakes Public Sector Decision-Making. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI '18). Association for Computing Machinery, Montreal, Canada, 1--14. https://doi.org/10.1145/3173574.3174014Google ScholarDigital Library
- Ruotong Wang, F. Maxwell Harper, and Haiyi Zhu. 2020. Factors Influencing Perceived Fairness in Algorithmic Decision-Making: Algorithm Outcomes, Development Procedures, and Individual Differences. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI '20). Association for Computing Machinery, Honolulu, USA, 1--14. https://doi.org/10.1145/3313831.3376813Google ScholarDigital Library
- Ke Yang and Julia Stoyanovich. 2017. Measuring Fairness in Ranked Outputs. In Proceedings of the 29th International Conference on Scientific and Statistical Database Management (SSDBM '17). Association for Computing Machinery, Chicago, USA, Article 22, 6 pages. https://doi.org/10.1145/3085504.3085526Google ScholarDigital Library
- Mengdi Zhang, Jun Sun, Jingyi Wang, and Bing Sun. 2023. TestSGD: Interpretable Testing of Neural Networks against Subtle Group Discrimination. ACM Trans. Softw. Eng. Methodol. 32, 6, Article 137 (Sep 2023), 24 pages. https://doi.org/10.1145/3591869Google ScholarDigital Library
- Yunfeng Zhang, Rachel K. E. Bellamy, and Kush R. Varshney. 2020. Joint Optimization of AI Fairness and Utility: A Human-Centered Approach. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (New York, NY, USA) (AIES '20). Association for Computing Machinery, New York, USA, 400--406. https://doi.org/10.1145/3375627.3375862Google ScholarDigital Library
- Yunhan Zheng, Qingyi Wang, Dingyi Zhuang, Shenhao Wang, and Jinhua Zhao. 2023. Fairness-Enhancing Deep Learning for Ride-Hailing Demand Prediction. IEEE Open Journal of Intelligent Transportation Systems 4 (2023), 551--569. https://doi.org/10.1109/OJITS.2023.3297517Google ScholarCross Ref
Index Terms
- Exploring Fairness-Accuracy Trade-Offs in Binary Classification: A Comparative Analysis Using Modified Loss Functions
Recommendations
Algorithmic bias in data-driven innovation in the age of AI
AbstractData-driven innovation (DDI) gains its prominence due to its potential to transform innovation in the age of AI. Digital giants Amazon, Alibaba, Google, Apple, and Facebook, enjoy sustainable competitive advantages from DDI. However, ...
Highlights- This study identifies the sources of algorithmic bias in data-driven innovations.
Data and algorithmic bias in the web
WebSci '16: Proceedings of the 8th ACM Conference on Web ScienceThe Web is the largest public big data repository that humankind has created. In this overwhelming data ocean we need to be aware of the quality of data extracted from it. One important quality issue is data bias, which appears in different forms. These ...
Lessons from debiasing data for fair and accurate predictive modeling in education
AbstractThe past few years have witnessed an explosion of attention given to the bias displayed by Machine Learning (ML) techniques towards different groups of people (e.g., female vs. male). Although ML techniques have been widely adopted in education, ...
Comments