skip to main content
research-article

The Many Facets of Data Equity

Published: 07 February 2023 Publication History

Abstract

Data-driven systems can induce, operationalize, and amplify systemic discrimination in a variety of ways. As data scientists, we tend to prefer to isolate and formalize equity problems to make them amenable to narrow technical solutions. However, this reductionist approach is inadequate in practice. In this article, we attempt to address data equity broadly, identify different ways in which it is manifest in data-driven systems, and propose a research agenda.

References

[1]
Chirag Agarwal and Sara Hooker. 2020. Estimating example difficulty using variance of gradients. CoRR abs/2008.11600 (2020). arXiv:2008.11600. https://arxiv.org/abs/2008.11600.
[2]
Elizabeth S. Anderson. 1999. What is the point of equality?Ethics 109, 2 (1999), 287–337. http://www.jstor.org/stable/10.1086/233897.
[3]
Abolfazl Asudeh, H. V. Jagadish, You Wu, and Cong Yu. 2020. On detecting cherry-picked trendlines. Proceedings of the VLDB Endowment 13, 6 (2020), 939–952.
[4]
Abolfazl Asudeh, Zhongjun Jin, and H. V. Jagadish. 2019. Assessing and remedying coverage for a given dataset. In IEEE International Conference on Data Engineering. IEEE, Macau, 554–565.
[5]
Abolfazl Asudeh, Nima Shahbazi, Zhongjun Jin, and H. V. Jagadish. 2021. Identifying insufficient data coverage for ordinal continuous-valued attributes. In SIGMOD’21: International Conference on Management of Data, Virtual Event, China, June 20–25, 2021, Guoliang Li, Zhanhuai Li, Stratos Idreos, and Divesh Srivastava (Eds.). ACM, 129–141. DOI:
[6]
Ricardo Baeza-Yates. 2018. Bias on the web. Communications of the ACM 61, 6 (2018), 54–61. DOI:
[7]
Andrés F. Barrientos, Alexander Bolton, Tom Balmat, Jerome P. Reiter, John M. De Figueiredo, Ashwin Machanavajjhala, Yan Chen, Charles Kneifel, and Mark DeLong. 2017. A framework for sharing confidential research data, applied to investigating differential pay by race in the US Government. Technical Report. National Bureau of Economic Research.
[8]
Steffen Bickel, Michael Brückner, and Tobias Scheffer. 2009. Discriminative learning under covariate shift. Journal of Machine Learning Research 10 (2009), 2137–2155. https://dl.acm.org/citation.cfm?id=1755858.
[9]
Felix Biessmann, David Salinas, Sebastian Schelter, Philipp Schmidt, and Dustin Lange. 2018. Deep learning for missing value imputation in tables with non-numerical data. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management. ACM, Turin, 2017–2025.
[10]
Miranda Bogen and Aaron Rieke. 2018. Help wanted: An examination of hiring algorithms, equity, and bias. Upturn (2018). Available from https://www.upturn.org/work/help-wanted/.
[11]
Geoffery C. Bowker and Susan Leigh Star. 2000. Sorting Things out: Classification and Its Consequences. MIT Press, Cambridge, MA.
[12]
Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on Fairness, Accountability and Transparency (FAT’18), February 23–24, 2018, New York, NY. 77–91. http://proceedings.mlr.press/v81/buolamwini18a.html.
[13]
Lucius Bynum, Joshua R. Loftus, and Julia Stoyanovich. 2021. Disaggregated interventions to reduce inequality. In ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization (EAAMO’21), Virtual Event, USA, October 5–9, 2021. ACM, 2:1–2:13. DOI:
[14]
Toon Calders, Faisal Kamiran, and Mykola Pechenizkiy. 2009. Building classifiers with independency constraints. In IEEE International Conference on Data Mining Workshops (ICDMW’09). IEEE, Miami, 13–18.
[15]
Irene Chen, Fredrik D. Johansson, and David Sontag. 2018. Why is my classifier discriminatory?. In Advances in Neural Information Processing Systems. 3539–3550.
[16]
Silvia Chiappa. 2019. Path-specific counterfactual fairness. In 33rd AAAI Conference on Artificial Intelligence (AAAI’19, 31st Innovative Applications of Artificial Intelligence Conference (IAAI’19), 9th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI’19), Honolulu, Hawaii, January 27–February 1, 2019. AAAI Press, 7801–7808. DOI:
[17]
Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data 5, 2 (2017), 153–163.
[18]
P. H. Collins. 2000. Black Feminist Thought: Knowledge, Consciousness, and the Politics of Empowerment. Routledge, New York, NY.
[19]
Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic decision making and the cost of fairness. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 797–806.
[20]
Council of Europe. 2020. Ad Hoc Committee on Artificial Intelligence (CAHAI), Feasibility Study. Retrieved from https://rm.coe.int/cahai-2020-23-final-eng-feasibility-study-/1680a0c6da.
[21]
Kimberle Crenshaw. 1989. Demarginalizing the intersection of race and sex: A black feminist critique of antidiscrimination doctrine, feminist theory and antiracist politics. University of Chicago Legal Forum1 (1989), 139–167.
[22]
Kimberle Crenshaw. 1991. Mapping the margins: Intersectionality, identity politics, and violence against women of color. Stanford Law Review 43, 6 (1991), 1241–1299. http://www.jstor.org/stable/1229039.
[23]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference. ACM, 214–226.
[24]
Ronald Dworkin. 1981. What is equality? Part 1: Equality of welfare. Philosophy and Public Affairs 10, 4 (1981), 185–246.
[25]
Ronald Dworkin. 1981. What is equality? Part 2: Equality of resources. Philosophy and Public Affairs 10, 4 (1981), 283–345.
[26]
Motahhare Eslami. 2017. Understanding and designing around users’ interaction with hidden algorithms in sociotechnical systems. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing (CSCW’17), Portland, OR, February 25–March 1, 2017, Companion Volume, Charlotte P. Lee, Steven E. Poltrock, Louise Barkhuus, Marcos Borges, and Wendy A. Kellogg (Eds.). ACM, 57–60. http://dl.acm.org/citation.cfm?id=3024947.
[27]
Claude S. Fischer, Michael Hout, Martín Sánchez Jankowski, Samuel R. Lucas, Ann Swidler, and Kim Voss. 2021. Inequality by Design: Cracking the Bell Curve Myth. Princeton University Press. DOI:
[28]
Batya Friedman and Helen Nissenbaum. 1996. Bias in computer systems. ACM Transactions on Information Systems 14, 3 (1996), 330–347. DOI:
[29]
Dominique DuBois Gilliard. 2018. Rethinking incarceration: Advocating for justice that restores. InterVarsity Press.
[30]
Stefan Grafberger, Julia Stoyanovich, and Sebastian Schelter. 2021. Lightweight inspection of data preprocessing in native machine learning pipelines. In 11th Conference on Innovative Data Systems Research (CIDR’21), Online Proceedings. www.cidrdb.org.
[31]
Maxim Grechkin, Hoifung Poon, and Bill Howe. 2018. EZLearn: Exploiting organic supervision in automated data annotation. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI’18), July 13–19, 2018, Stockholm, Sweden, Jérôme Lang (Ed.). ijcai.org, 4085–4091. DOI:
[32]
Yifan Guan, Abolfazl Asudeh, Pranav Mayuram, H. V. Jagadish, Julia Stoyanovich, Gerome Miklau, and Gautam Das. 2019. MithraRanking: A system for responsible ranking design. In Proceedings of the 2019 International Conference on the Management of Data, SIGMOD. ACM, 1913–1916. DOI:
[33]
Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. 2019. A survey of methods for explaining black box models. ACM Computing Surveys 51, 5 (2019), 93:1–93:42. DOI:
[34]
P. L. Hammack. 2018. The Oxford Handbook of Social Psychology and Social Justice. Oxford University Press. 2017034131https://books.google.com/books?id=ZY9HDwAAQBAJ.
[35]
Moritz Hardt, Eric Price, Nati Srebro, et al. 2016. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems. 3315–3323.
[36]
Melanie Herschel, Ralf Diestelkämper, and Houssem Ben Lahmar. 2017. A survey on provenance: What for? What form? What from?VLDB Journal 26, 6 (2017), 881–906. DOI:
[37]
Anna Lauren Hoffmann. 2021. Terms of inclusion: Data, discourse, violence. New Media & Society 23, 12 (2021), 3539–3556. DOI:
[38]
Sara Hooker, Nyalleng Moorosi, Gregory Clark, Samy Bengio, and Emily Denton. 2020. Characterising bias in compressed models. (2020). arXiv:cs.LG/2010.03058
[39]
J. House and D. R. Williams. 2000. Understanding and Reducing Socioeconomic and Racial/Ethnic Disparities in Health. National Academy Press, Washington, DC., 81–124. Reprinted in Richard Hofrichter (ed.) Health and Social Justice: Politics, Ideology, and Inequity in the Distribution of Disease. Jossey-Bass, pp. 89–131, 2003.
[40]
Jiayuan Huang, Alexander J. Smola, Arthur Gretton, Karsten M. Borgwardt, and Bernhard Schölkopf. 2006. Correcting sample selection bias by unlabeled data. In Advances in Neural Information Processing Systems 19, Proceedings of the 20th Annual Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, December 4–7, 2006, Bernhard Schölkopf, John C. Platt, and Thomas Hofmann (Eds.). MIT Press, 601–608. http://papers.nips.cc/paper/3075-correcting-sample-selection-bias-by-unlabeled-data.
[41]
Stephen J. Mooney, Kate Hosford, Bill Howe, An Yan, Meghan Winters, Alon Bassok, and Jana Hirsch. 2019. Freedom from the station: Spatial equity in access to dockless bike share. Journal of Transport Geography 74 (12019), 91–96. DOI:
[42]
H. V. Jagadish, Johannes Gehrke, Alexandros Labrinidis, Yannis Papakonstantinou, Jignesh M. Patel, Raghu Ramakrishnan, and Cyrus Shahabi. 2014. Big data and its technical challenges. Communications of the ACM 57, 7 (2014), 86–94.
[43]
H. V. Jagadish, Julia Stoyanovich, and Bill Howe. 2021. COVID-19 brings data equity challenges to the fore. ACM Digital Government: Research and Practice 2, 2 (2021).
[44]
Zhongjun Jin, Mengjing Xu, Chenkai Sun, Abolfazl Asudeh, and H. V. Jagadish. 2020. MithraCoverage: A system for investigating population bias for intersectional fairness. In Proceedings of the ACM SIGMOD International Conference on Management of Data. ACM, 2721–2724.
[45]
Joost Kappelhof. 2017. Survey research and the quality of survey data among ethnic minorities. Total Survey Error in Practice (2017), 235–252.
[46]
Falaah Arif Khan, Eleni Manis, and Julia Stoyanovich. 2021. Fairness as equality of opportunity: Normative guidance from political philosophy. CoRR abs/2106.08259 (2021). arXiv:2106.08259. https://arxiv.org/abs/2106.08259.
[47]
Niki Kilbertus, Mateo Rojas-Carulla, Giambattista Parascandolo, Moritz Hardt, Dominik Janzing, and Bernhard Schölkopf. 2017. Avoiding discrimination through causal reasoning. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4–9, 2017, Long Beach, CA, Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (Eds.). 656–666. https://proceedings.neurips.cc/paper/2017/hash/f5f8590cd58a54e94377e6ae2eded4d9-Abstract.html.
[48]
Daniel Kluttz, Nitin Kohli, and Dierdre Mulligan. 2020. Shaping our tools: Contestability as a means to promote responsible algorithmic decision making in the professions. In After the Digital Tornado: Networks, Algorithms, Humanity, Kevin Werback (Ed.). Cambridge University Press, 137–152.
[49]
Arun Kumar, Robert McCann, Jeffrey F. Naughton, and Jignesh M. Patel. 2015. Model selection management systems: The next frontier of advanced analytics. SIGMOD Record 44, 4 (2015), 17–22. DOI:
[50]
Matt J. Kusner, Joshua R. Loftus, Chris Russell, and Ricardo Silva. 2017. Counterfactual fairness. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4–9, 2017, Long Beach, CA, Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (Eds.). 4066–4076. https://proceedings.neurips.cc/paper/2017/hash/a486cd07e4ac3d270571622f4f316ec5-Abstract.html.
[51]
Yin Lin, Yifan Guan, Abolfazl Asudeh, and H. V. Jagadish. 2020. Identifying insufficient data coverage in databases with multiple relations. Proceedings of the VLDB Endowment 13, 12 (2020), 2229–2242.
[52]
Zachary C. Lipton, Yu-Xiang Wang, and Alexander J. Smola. 2018. Detecting and correcting for label shift with black box predictors. In Proceedings of the 35th International Conference on Machine Learning (ICML’18), Stockholmsmässan, Stockholm, Sweden, July 10–15, 2018 (Proceedings of Machine Learning Research), Jennifer G. Dy and Andreas Krause (Eds.), Vol. 80. PMLR, 3128–3136. http://proceedings.mlr.press/v80/lipton18a.html.
[53]
Joshua R. Loftus, Chris Russell, Matt J. Kusner, and Ricardo Silva. 2018. Causal reasoning for algorithmic fairness. arXiv preprint arXiv:1805.05859.
[54]
Kristian Lum, Samarth Swarup, Stephen Eubank, and James Hawdon. 2014. The contagious nature of imprisonment: An agent-based model to explain racial disparities in incarceration rates. Interface, Journal of the Royal Society 11, 98 (2014).
[55]
David Madras, Elliot Creager, Toniann Pitassi, and Richard Zemel. 2018. Learning adversarially fair and transferable representations. arXiv preprint arXiv:1802.06309.
[56]
Charles W. Mills. 2014. The Racial Contract. Cornell University Press.
[57]
Yuval Moskovitch and H. V. Jagadish. 2020. COUNTATA: Dataset labeling using pattern counts. Proc. VLDB Endow. 13, 12 (2020), 2829–2832. DOI:
[58]
Yuval Moskovitch and H. V. Jagadish. 2021. Patterns count-based labels for datasets. In 37th IEEE International Conference on Data Engineering, ICDE 2021, Chania, Greece, April 19–22, 2021. IEEE, 1961–1966. DOI:
[59]
Deirdre K. Mulligan and Kenneth A. Bamberger. 2019. Procurement as policy: Administrative process for machine learning. Berkeley Technology Law Journal 34, 3 (2019), 773–852.
[60]
Razieh Nabi and Ilya Shpitser. 2018. Fair inference on outcomes. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence (AAAI’18), the 30th Innovative Applications of Artificial Intelligence (IAAI’18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI’18), New Orleans, LA, February 2–7, 2018, Sheila A. McIlraith and Kilian Q. Weinberger (Eds.). AAAI Press, 1931–1940. https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16683.
[61]
Fatemeh Nargesian, Abolfazl Asudeh, and H. V. Jagadish. 2021. Tailoring data source distributions for fairness-aware data integration. Proceedings of the VLDB Endowment 14, 11 (2021).
[62]
Haoyue Ping, Julia Stoyanovich, and Bill Howe. 2017. DataSynthesizer: Privacy-preserving synthetic datasets. In Proceedings of the 29th International Conference on Scientific and Statistical Database Management (SSDBM’17). ACM, New York, NY, Article 42, 5 pages. DOI:
[63]
Neoklis Polyzotis, Sudip Roy, Steven Euijong Whang, and Martin Zinkevich. 2018. Data lifecycle challenges in production machine learning: A survey. SIGMOD Records 47, 2 (2018), 17–28. DOI:
[64]
John Rawls. 1971. A Theory of Justice. Harvard University Press. http://www.jstor.org/stable/j.ctvjf9z6v.
[65]
Richard V. Reeves and Dimitrios Halikias. 2017. Race gaps in SAT scores highlight inequality and hinder upward mobility. Retrieved August 14, 2019 from https://www.brookings.edu/research/race-gaps-in-sat-scores-highlight-inequality-and-hinder-upward-mobility/.
[66]
John E. Roemer and Alain Trannoy. 2015. Equality of opportunity. In Handbook of Income Distribution, A. B. Atkinson and F. Bourguignon (Eds.). Elsevier, Vol. 2, 217–300.
[67]
Lucas Rosenblatt, Xiaoyan Liu, Samira Pouyanfar, Eduardo de Leon, Anuj Desai, and Joshua Allen. 2020. Differentially private synthetic data: Applied evaluations and enhancements. CoRR abs/2011.05537 (2020). arXiv:2011.05537. https://arxiv.org/abs/2011.05537.
[68]
Babak Salimi, Luke Rodriguez, Bill Howe, and Dan Suciu. 2019. Capuchin: Causal database repair for algorithmic fairness. CoRR abs/1902.08283 (2019). arXiv:1902.08283. http://arxiv.org/abs/1902.08283.
[69]
Babak Salimi, Luke Rodriguez, Bill Howe, and Dan Suciu. 2019. Interventional fairness: Causal database repair for algorithmic fairness. In Proceedings of the 2019 International Conference on Management of Data (SIGMOD’19). ACM, New York, NY, 793–810. DOI:
[70]
Sebastian Schelter, Felix Biessmann, Tim Januschowski, David Salinas, Stephan Seufert, Gyuri Szarvas, Manasi Vartak, Samuel Madden, Hui Miao, Amol Deshpande, et al. 2018. On challenges in machine learning model management.IEEE Data Engineering Bulletin 41, 4 (2018), 5–15.
[71]
Sebastian Schelter, Yuxuan He, Jatin Khilnani, and Julia Stoyanovich. 2020. FairPrep: Promoting data to a first-class citizen in studies on fairness-enhancing interventions. In EDBT, Angela Bonifati, Yongluan Zhou, Marcos Antonio Vaz Salles, Alexander Böhm, Dan Olteanu, George H. L. Fletcher, Arijit Khan, and Bin Yang (Eds.). OpenProceedings.org, 395–398. DOI:
[72]
Sebastian Schelter, Tammo Rukat, and Felix Bießmann. 2020. Learning to validate the predictions of black box classifiers on unseen data. In Proceedings of the 2020 International Conference on Management of Data, SIGMOD Conference 2020, online conference [Portland, OR], June 14–19, 2020, David Maier, Rachel Pottinger, AnHai Doan, Wang-Chiew Tan, Abdussalam Alawini, and Hung Q. Ngo (Eds.). ACM, 1289–1299. DOI:
[73]
David Sculley, Gary Holt, Daniel Golovin, Eugene Davydov, Todd Phillips, Dietmar Ebner, Vinay Chaudhary, Michael Young, Jean-Francois Crespo, and Dan Dennison. 2015. Hidden technical debt in machine learning systems. In Advances in Neural Information Processing Systems. 2503–2511.
[74]
Camelia Simoiu, Sam Corbett-Davies, Sharad Goel, et al. 2017. The problem of infra-marginality in outcome tests for discrimination. Annals of Applied Statistics 11, 3 (2017), 1193–1216.
[75]
Judith Simon. 2015. Distributed epistemic responsibility in a hyperconnected era. In The Online Manifesto: Being Human in a Hyperconnected Era. 145–159. DOI:
[76]
Dean Spade. 2015. Normal Life: Administrative Violence, Critical Trans Politics, and the Limits of Law. Duke University Press. http://www.jstor.org/stable/j.ctv123x7qx.
[77]
Dean Spade. 2015. Normal life: Administrative violence, critical trans politics, and the limits of law. Duke University Press.
[78]
Julia Stoyanovich. 2020. Testimony of Julia Stoyanovich before New York City Council Committee on Technology regarding Int 1894–2020, Sale of automated employment decision tools. Retrieved from https://dataresponsibly.github.io/documents/Stoyanovich_Int1894Testimony.pdf.
[79]
Julia Stoyanovich and Ellen P. Goodman. Revealing algorithmic rankers. Retrieved February 17, 2019 from http://freedom-to-tinker.com/2016/08/05/revealing-algorithmic-rankers/.
[80]
Julia Stoyanovich and Ellen P. Goodman. 2016. Revealing algorithmic rankers. Freedom to Tinker, Center for Information Technology Policy, Princeton University. Retrieved from http://freedom-to-tinker.com/2016/08/05/revealing-algorithmic-rankers.
[81]
Julia Stoyanovich and Bill Howe. 2018. Refining the concept of a nutritional label for data and models. Freedom to Tinker, Center for Information Technology Policy, Princeton University.Retrieved from https://freedom-to-tinker.com/2018/05/03/refining-the-concept-of-a-nutritional-label-for-data-and-models.
[82]
Julia Stoyanovich and Bill Howe. 2019. Nutritional labels for data and models. IEEE Data Engineering Bulletin 42, 3 (2019), 13–23. http://sites.computer.org/debull/A19sept/p13.pdf.
[83]
Julia Stoyanovich, Bill Howe, and H. V. Jagadish. 2020. Responsible data management. PVLDB 13, 12 (2020), 3474–3489. DOI:
[84]
Julia Stoyanovich, Jay J. Van Bavel, and Tessa V. West. 2020. The imperative of interpretable machines. Nature Machine Intelligence 2 (2020), 197–199. DOI:
[85]
Chenkai Sun, Abolfazl Asudeh, H. V. Jagadish, Bill Howe, and Julia Stoyanovich. 2019. Mithralabel: Flexible dataset nutritional labels for responsible data science. In Proceedings of the ACM International Conference on Information and Knowledge Management. ACM, 2893–2896.
[86]
Supreme Court of the United States. 1995. ADARAND CONSTRUCTORS, Inc. v. PEÑA, 515 U.S. 200 (1995), No. 93-1841. Retrieved from https://supreme.justia.com/cases/federal/us/515/200/#tab-opinion-1959723.
[87]
Supreme Court of the United States. 1996. UNITED STATES v. VIRGINIA, 518 U.S. 515 (1996), No. 94-1941. Retrieved from https://supreme.justia.com/cases/federal/us/515/200/#tab-opinion-1959723.
[88]
Supreme Court of the United States. 2009. Ricci v. DeStefano (Nos. 7-1428 and 8-328), 530 F. 3d 87, reversed and remanded. Retrieved from https://www.law.cornell.edu/supct/html/07-1428.ZO.html.
[89]
Latanya Sweeney. 2002. k-Anonymity: A model for protecting privacy. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 10, 5 (2002), 557–570. DOI:
[90]
Ian Kennedy, Timothy A. Thomas, Ott Toomet, and Alex Ramiller. The State of Evictions: Results from the University of Washington Evictions Project. https://evictionresearch.net/washington.
[91]
Reihaneh Torkzadehmahani, Peter Kairouz, and Benedict Paten. 2019. DP-CGAN: Differentially private synthetic data and label generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 0–0.
[92]
U.S. Equal Employment Opportunity Commission (EEOC). 1979. Uniform guidelines on employee selection procedures. (March1979).
[93]
Chugui Xu, Ju Ren, Yaoxue Zhang, Zhan Qin, and Kui Ren. 2017. DPPro: Differentially private high-dimensional data release via random projection. IEEE Transactions on Information Forensics and Security 12, 12 (2017), 3081–3093.
[94]
An Yan and Bill Howe. 2019. FairST: Equitable spatial and temporal demand prediction for new mobility systems. In Proceedings of the 27th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems. 552–555.
[95]
An Yan and Bill Howe. 2021. EquiTensors: Learning fair integrations of heterogeneous urban data. In ACM SIGMOD.
[96]
Ke Yang, Joshua R. Loftus, and Julia Stoyanovich. 2021. Causal intersectionality and fair ranking. In 2nd Symposium on Foundations of Responsible Computing, FORC 2021, June 9–11, 2021, Virtual Conference (LIPIcs), Katrina Ligett and Swati Gupta (Eds.), Vol. 192. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 7:1–7:20. DOI:
[97]
Ke Yang, Julia Stoyanovich, Abolfazl Asudeh, Bill Howe, H. V. Jagadish, and Gerome Miklau. 2018. A nutritional label for rankings. In Proceedings of the 2018 International Conference on Management of Data, SIGMOD Conference 2018, Houston, TX, June 10–15, 2018. 1773–1776. DOI:
[98]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P. Gummadi. 2017. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, 1171–1180.
[99]
Meike Zehlike, Ke Yang, and Julia Stoyanovich. 2021. Fairness in ranking: A survey. CoRR abs/2103.14000 (2021). arXiv:2103.14000. https://arxiv.org/abs/2103.14000.
[100]
Junzhe Zhang and Elias Bareinboim. 2018. Fairness in decision-making–the causal explanation formula. In 32nd AAAI Conference on Artificial Intelligence.
[101]
Jun Zhang, Graham Cormode, Cecilia M. Procopiuc, Divesh Srivastava, and Xiaokui Xiao. 2017. PrivBayes: Private data release via Bayesian networks. ACM Transactions on Database Systems 42, 4 (2017), 25:1–25:41. DOI:

Cited By

View all
  • (2025)Enhancing Business Process Models with Ethical ConsiderationsEnterprise Design, Operations, and Computing. EDOC 2024 Workshops10.1007/978-3-031-79059-1_1(3-17)Online publication date: 9-Feb-2025
  • (2024)Towards Detecting Unintended Behaviors in Machine Learning Algorithms2024 AIAA DATC/IEEE 43rd Digital Avionics Systems Conference (DASC)10.1109/DASC62030.2024.10749477(1-10)Online publication date: 29-Sep-2024
  • (2022)Fairness & friends in the data science eraAI & SOCIETY10.1007/s00146-022-01472-5Online publication date: 9-Jun-2022

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Journal of Data and Information Quality
Journal of Data and Information Quality  Volume 14, Issue 4
December 2022
173 pages
ISSN:1936-1955
EISSN:1936-1963
DOI:10.1145/3563905
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 February 2023
Online AM: 08 August 2022
Accepted: 24 April 2022
Revised: 12 March 2022
Received: 31 July 2021
Published in JDIQ Volume 14, Issue 4

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Data equity
  2. ethics
  3. responsible data science
  4. Fairness in AI

Qualifiers

  • Research-article
  • Refereed

Funding Sources

  • US National Science Foundation

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)162
  • Downloads (Last 6 weeks)18
Reflects downloads up to 25 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Enhancing Business Process Models with Ethical ConsiderationsEnterprise Design, Operations, and Computing. EDOC 2024 Workshops10.1007/978-3-031-79059-1_1(3-17)Online publication date: 9-Feb-2025
  • (2024)Towards Detecting Unintended Behaviors in Machine Learning Algorithms2024 AIAA DATC/IEEE 43rd Digital Avionics Systems Conference (DASC)10.1109/DASC62030.2024.10749477(1-10)Online publication date: 29-Sep-2024
  • (2022)Fairness & friends in the data science eraAI & SOCIETY10.1007/s00146-022-01472-5Online publication date: 9-Jun-2022

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

HTML Format

View this article in HTML Format.

HTML Format

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media