Abstract
There has been a significant recent interest in algorithmic fairness within data-driven systems. In this paper, we consider group fairness within Case-based Reasoning. Group fairness targets to ensure parity of outcomes across pre-specified sensitive groups, defined on the basis of extant entrenched discrimination. Addressing the context of binary decision choice scenarios over binary sensitive attributes, we develop three separate fairness interventions that operate at different stages of the CBR process. These techniques, called Label Flipping (LF), Case Weighting (CW) and Weighted Adaptation (WA), use distinct strategies to enhance group fairness in CBR decision making. Through an extensive empirical evaluation over several popular datasets and against natural baseline methods, we show that our methods are able to achieve significant enhancements in fairness at low detriment to accuracy, thus illustrating effectiveness of our methods at advancing fairness.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
An experiment with the edited nearest-neighbor rule. IEEE Trans. Syst. Man Cybern. SMC-6(6), 448–452 (1976). https://doi.org/10.1109/TSMC.1976.4309523
Abraham, S.S.: FairLOF: fairness in outlier detection. Data Sci. Eng. 6, 485–499 (2021)
Abraham, S.S., Sundaram, S.S., et al.: Fairness in clustering with multiple sensitive attributes. EDBT (2020)
Barocas, S., Hardt, M., Narayanan, A.: Fairness and Machine Learning: Limitations and Opportunities (2019). http://www.fairmlbook.org
Biega, A.J., Gummadi, K.P., Weikum, G.: Equity of attention: amortizing individual fairness in rankings. In: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, pp. 405–414 (2018)
Binns, R.: On the apparent conflict between individual and group fairness. In: Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, pp. 514–524 (2020)
Blanzeisky, W., Cunningham, P., Kennedy, K.: Introducing a family of synthetic datasets for research on bias in machine learning (2021)
Blanzeisky, W., Smyth, B., Cunningham, P.: Algorithmic bias and fairness in case-based reasoning. In: Keane, M.T., Wiratunga, N. (eds.) ICCBR 2022. LNAI, vol. 13405, pp. 48–62. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-14923-8_4
Breunig, M.M., Kriegel, H.P., Ng, R.T., Sander, J.: LOF: identifying density-based local outliers. In: Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, pp. 93–104 (2000)
Chang, K.W., Prabhakaran, V., Ordonez, V.: Bias and fairness in natural language processing. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): Tutorial Abstracts (2019)
Chouldechova, A., Roth, A.: A snapshot of the frontiers of fairness in machine learning. Commun. ACM 63(5), 82–89 (2020)
Deepak, P., Abraham, S.S.: Fair outlier detection. In: 21th International Conference on Web Information Systems Engineering: WISE 2020, pp. 447–462 (2020)
Dressel, J., Farid, H.: The accuracy, fairness, and limits of predicting recidivism. Sci. Adv. 4(1), eaao5580 (2018). https://doi.org/10.1126/sciadv.aao5580
Dwork, C., Hardt, M., Pitassi, T., Reingold, O., Zemel, R.: Fairness through awareness. In: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp. 214–226 (2012)
Ekstrand, M.D., Burke, R., Diaz, F.: Fairness and discrimination in retrieval and recommendation. In: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 1403–1404 (2019)
Hertweck, C., Heitz, C., Loi, M.: On the moral justification of statistical parity. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 747–757 (2021)
Kohavi, R.: Scaling up the accuracy of Naive-Bayes classifiers: a decision-tree hybrid. In: KDD (1997)
Kriegel, H.P., Kröger, P., Schubert, E., Zimek, A.: Loop: local outlier probabilities. In: Proceedings of the 18th ACM Conference on Information and Knowledge Management, pp. 1649–1652 (2009)
Kulshrestha, J., et al.: Search bias quantification: investigating political bias in social media and web search. Inf. Retrieval J. 22, 188–227 (2019)
Kunaver, M., Porl, T.: Diversity in recommender systems a survey. Know.-Based Syst. 123(C), 154–162 (2017). https://doi.org/10.1016/j.knosys.2017.02.009
Le Quy, T., Roy, A., Iosifidis, V., Zhang, W., Ntoutsi, E.: A survey on datasets for fairness-aware machine learning. Wiley Interdisc. Rev.: Data Min. Knowl. Discovery 12(3), e1452 (2022)
Narayanan, A.: Translation tutorial: 21 fairness definitions and their politics. In: Proceedings of the Conference on Fairness Accountability and Transparency, New York, USA, vol. 1170, p. 3 (2018)
Patro, G.K., Biswas, A., Ganguly, N., Gummadi, K.P., Chakraborty, A.: FairRec: two-sided fairness for personalized recommendations in two-sided platforms. In: Proceedings of the Web Conference 2020, pp. 1194–1204 (2020)
Pessach, D., Shmueli, E.: Algorithmic fairness. arXiv preprint arXiv:2001.09784 (2020)
Pessach, D., Shmueli, E.: A review on fairness in machine learning. ACM Comput. Surv. (CSUR) 55(3), 1–44 (2022)
Quijano-Sanchez, L., Recio-Garcia, J.A., Diaz-Agudo, B., Jimenez-Diaz, G.: Social factors in group recommender systems. ACM Trans. Intell. Syst. Technol. 4(1) (2013). https://doi.org/10.1145/2414425.2414433
Yang, K., Gkatzelis, V., Stoyanovich, J.: Balanced ranking with diversity constraints. arXiv preprint arXiv:1906.01747 (2019)
Yeh, I.C., Lien, C.: The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients. Expert Syst. Appl. 36(2, Part 1), 2473–2480 (2009). https://doi.org/10.1016/j.eswa.2007.12.020
Zehlike, M., Bonchi, F., Castillo, C., Hajian, S., Megahed, M., Baeza-Yates, R.: Fa* ir: a fair top-k ranking algorithm. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pp. 1569–1578 (2017)
Zehlike, M., Yang, K., Stoyanovich, J.: Fairness in ranking: a survey. arXiv preprint arXiv:2103.14000 (2021)
Zhang, Y., et al.: Causal intervention for leveraging popularity bias in recommendation. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 11–20 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Ethics declarations
Future Work
While our separate methods were designed for fairness interventions at different stages of the CBR process, it would be interesting to understand the complementarity between these methods and exploit them for application in scenarios where the user has control over all stages of the CBR process. We are considering extending this to cover multi-choice and structured decision scenarios and multi-valued sensitive attributes.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Mitra, S., Mathew, D., P., D., Chakraborti, S. (2023). Group Fairness in Case-Based Reasoning. In: Massie, S., Chakraborti, S. (eds) Case-Based Reasoning Research and Development. ICCBR 2023. Lecture Notes in Computer Science(), vol 14141. Springer, Cham. https://doi.org/10.1007/978-3-031-40177-0_14
Download citation
DOI: https://doi.org/10.1007/978-3-031-40177-0_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-40176-3
Online ISBN: 978-3-031-40177-0
eBook Packages: Computer ScienceComputer Science (R0)