skip to main content
10.1145/3378393.3402507acmconferencesArticle/Chapter ViewAbstractPublication PagescompassConference Proceedingsconference-collections
research-article

The Role of In-Group Bias and Balanced Data: A Comparison of Human and Machine Recidivism Risk Predictions

Published: 01 July 2020 Publication History

Abstract

Fairness and bias in automated decision-making gain importance as the prevalence of algorithms increases in different areas of social life. This paper contributes to the discussion of algorithmic fairness with a crowdsourced vignette survey on recidivism risk assessment, which we compare to previous studies on this topic and to predictions of an automated recidivism risk tool. We use the case of the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) and the Broward County dataset of pre-trial defendants as a data source and for purposes of comparability with the earlier analysis. In our survey, each respondent assessed recidivism risk for a set of vignettes describing real defendants, where each set was balanced with regard to the defendants' race and re-offender status. The survey ensured a 50: 50 ratio of black and white respondents. We found that predictions in our survey---while less accurate---were considerably more fair in terms of equalized odds than previous surveys. We attribute it to the differences in survey design: using the balanced set of vignettes and not providing feedback after responding to each vignette. We also analyzed the performance and fairness of predictions by race of respondent and defendant. We found that both white and black respondents tend to favor defendants of their own race, but the magnitude of the effect is relatively small. In addition to the survey, we train two statistical models, one trained with balanced data and other with unbalanced data. We observe that the model trained on balanced data is substantially more fair and possess less in-group bias.

References

[1]
Michelle Alexander. 2010. The new Jim Crow: Mass incarceration in the age of colorblindness. The New Press, New York, NY.
[2]
Solon Barocas and Andrew Selbst. 2016. Big Data's Disparate Impact. California Law Review 104, 1 (2016), 671--732. https://doi.org/10.15779/Z38BG31
[3]
Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. 2018. Fairness in Criminal Justice Risk Assessments: The State of the Art. Sociological Methods and Research (2018), 1--42. https://doi.org/10.1177/0049124118782533
[4]
Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classificatio. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency, Sorelle A. Friedler and Christo Wilson (Eds.), Vol. 81. New York, 77--91. https://doi.org/10.2147/OTT.S126905
[5]
Toon Calders, Faisal Kamiran, and Mykola Pechenizkiy. 2009. Building classifiers with independency constraints. In Data mining workshops, 2009. ICDMW'09. IEEE international conference on. IEEE, 13--18.
[6]
Alexandra Chouldechova, Emily Putnam-Hornstein, Suzanne Dworak-Peck, Diana Benavides-Prado, Oleksandr Fialko, Rhema Vaithianathan, Sorelle A Friedler, and Christo Wilson. 2018. A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. Proceedings of Machine Learning Research 81 (2018), 1--15. http://proceedings.mlr.press/v81/chouldechova18a/chouldechova18a.pdf
[7]
Alexandra Chouldechova and Aaron Roth. 2018. The frontiers of fairness in machine learning. arXiv preprint arXiv:1810.08810 (2018).
[8]
Sam Corbett-Davies and Sharad Goel. 2018. The measure and mismeasure of fairness: A critical review of fair machine learning. arXiv preprint arXiv:1808.00023 (2018).
[9]
Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic decision making and the cost of fairness. Proceedings of KDD '17, Halifax, NS, Canada (2017). https://doi.org/10.1145/3097983.309809 arXiv:1701.08230
[10]
Kate Crawford. 2017. The Trouble with Bias. Keynote at the Neural Information Processing Systems (NIPS) Conference (2017). https://www.youtube.com/watch?v=fMym_BKWQzk
[11]
Kimberle Crenshaw. 1989. Demarginalizing the Intersection of Race and Sex: A Black Feminist Critique of Antidiscrimination Doctrine, Feminist Theory and Antiracist Politics. University of Chicago Legal Forum 1989, 1 (1989), 139--167. https://chicagounbound.uchicago.edu/cgi/viewcontent.cgi?article=1052&context=uclf
[12]
Julia Dressel and Hany Farid. 2018. The accuracy, fairness, and limits of predicting recidivism. Science Advances 4, 1 (2018), 1--6. https://doi.org/10.1126/sciadv.aao5580
[13]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. ACM, 214--226.
[14]
Michael Feldman, Sorelle Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM (2015), 259--268. arXiv:1412.3756 http://arxiv.org/abs/1412.3756
[15]
Jean Garcia-Gathright, Aaron Springer, and Henriette Cramer. 2018. Assessing and Addressing Algorithmic Bias - But Before We Get There. 2018 AAAI Spring Symposium Series Assessing (2018). arXiv:1809.03332 http://arxiv.org/abs/1809.03332
[16]
Moritz Hardt, Eric Price, and Nathan Srebro. 2016. Equality of Opportunity in Supervised Learning. Advances in Neural Information Processing Systems (2016), 3315--3323. arXiv:1610.02413 http://arxiv.org/abs/1610.02413
[17]
Anna Lauren Hoffmann. 2019. Where fairness fails: data, algorithms, and the limits of antidiscrimination discourse. Information, Communication & Society 22, 7 (jun 2019), 900--915. https://doi.org/10.1080/1369118X.2019.1573912
[18]
Kenneth Holstein, Jennifer Wortman Vaughan, Hal Daumé, Miro Dudík, and Hanna Wallach. 2018. Improving fairness in machine learning systems: What do industry practitioners need? (2018). https://doi.org/10.1145/3290605.3300830 arXiv:1812.05239
[19]
D Bradford Hunt. 2005. Redlining. Encyclopedia of Chicago (2005).
[20]
Carter Jernigan and Behram FT Mistree. 2009. Gaydar: Facebook friendships expose sexual orientation. First Monday 14, 10 (2009).
[21]
Faisal Kamiran and Toon Calders. 2012. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems 33, 1 (2012), 1--33.
[22]
Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. 2012. Fairness-aware classifier with prejudice remover regularizer. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Springer, 35--50.
[23]
Matthew Kay, Cynthia Matuszek, and Sean A. Munson. 2015. Unequal Representation and Gender Stereotypes in Image Search Results for Occupations. (2015), 3819--3828. https://doi.org/10.1145/2702123.2702520
[24]
Danielle Kehl, Priscilla Guo, and Samuel Kessler. 2017. Algorithms in the Criminal Justice System: Assessing the Use of Risk Assessments in Sentencing. http://nrs.harvard.edu/urn-3:HUL.InstRepos:33746041
[25]
Jon Kleinberg, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2017. Human Decisions and Machine Predictions. The Quarterly Journal of Economics January (aug 2017), 237--293. https://doi.org/10.1093/qje/qjx032
[26]
Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2017. Inherent trade-offs in the fair determination of risk scores. Innovations in Theoretical Computer Science (2017).
[27]
Max Kuhn. 2008. Building Predictive Models in R Using the caret Package. Journal of Statistical Software, Articles 28, 5 (2008), 1--26. https://doi.org/10.18637/jss.v028.i05
[28]
Je Larson, Surya Mattu, Lauren Kirchner, and Julia Angwin. 2016. How We Analyzed the COMPAS Recidivism Algorithm. Technical Report. ProPublica. https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm
[29]
Leib Litman, Jonathan Robinson, and Tzvi Abberbock. 2017. TurkPrime. com: A versatile crowdsourcing data acquisition platform for the behavioral sciences. Behavior research methods 49, 2 (2017), 433--442.
[30]
Douglass S. Massey. 2007. Categorically Unequal. The American Stratification System. Russell Sage Foundation, New York.
[31]
Shira Mitchell, Eric Potash, and Solon Barocas. 2018. Prediction-based decisions and fairness: A catalogue of choices, assumptions, and definitions. arXiv preprint arXiv:1811.07867 (2018).
[32]
Sofia Olhede and Russell Rodrigues. 2017. Fairness and transparency in the age of the algorithm. Significance 14, 2 (2017), 8--9. https://doi.org/10.1111/j.1740-9713.2017.01012.x arXiv:1609.05807
[33]
Cathy O'Neill. 2016. Weapons of Math Destruction. How Big Data Increases Inequality and Threatens Democracy. Crown/Archetype, New York.
[34]
Becky Pettit and Bruce Western. 2004. Mass Imprisonment and the Life Course: Race and Class Inequality in U.S. Incarceration. American Sociological Review 69, 2 (apr 2004), 151--169. https://doi.org/10.1177/000312240406900201
[35]
Latanya Sweeney. 2013. Discrimination in Online Ad Delivery. Queue 11, 3 (mar 2013), 10. https://doi.org/10.1145/2460276.2460278
[36]
Charles Tilly. 1999. Durable Inequality. University of California Press.
[37]
Michael Veale, Max Van Kleek, and Reuben Binns. 2018. Fairness and Accountability Design Needs for Algorithmic Support in High-Stakes Public Sector Decision-Making. (2018). https://doi.org/10.1145/3173574.3174014 arXiv:1802.01029
[38]
Blake Woodworth, Suriya Gunasekar, Mesrob I Ohannessian, and Nathan Srebro. 2017. Learning non-discriminatory predictors. Preprint arXiv:1702.06081 (2017).
[39]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2017. Fairness constraints: Mechanisms for fair classification. In Artificial Intelligence and Statistics. 962--970.
[40]
Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In Proceedings of the 30th International Conference on Machine Learning (ICML-13). 325--333.
[41]
Zhe Zhang and Daniel B Neill. 2016. Identifying significant predictive bias in classifiers. arXiv preprint arXiv:1611.08292 (2016).

Cited By

View all
  • (2024)Algorithms and Recidivism: A Multi-Disciplinary Systematic ReviewProceedings of the 2024 AAAI/ACM Conference on AI, Ethics, and Society10.5555/3716662.3716775(1292-1305)Online publication date: 21-Oct-2024
  • (2024)Fairness for Deep Learning Predictions Using Bias Parity Score Based Loss Function RegularizationInternational Journal on Artificial Intelligence Tools10.1142/S021821302460003033:03Online publication date: 25-Apr-2024
  • (2024)Safeguarding the Future of Artificial Intelligence: An AI BlueprintArtificial Intelligence for Security10.1007/978-3-031-57452-8_1(3-22)Online publication date: 17-Apr-2024
  • Show More Cited By

Index Terms

  1. The Role of In-Group Bias and Balanced Data: A Comparison of Human and Machine Recidivism Risk Predictions

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      COMPASS '20: Proceedings of the 3rd ACM SIGCAS Conference on Computing and Sustainable Societies
      June 2020
      359 pages
      ISBN:9781450371292
      DOI:10.1145/3378393
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 01 July 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Bias
      2. Experimentation
      3. Racial discrimination
      4. Statistical Analysis
      5. Survey Analysis

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Conference

      COMPASS '20
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 25 of 50 submissions, 50%

      Upcoming Conference

      COMPASS '25

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)51
      • Downloads (Last 6 weeks)9
      Reflects downloads up to 03 Mar 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Algorithms and Recidivism: A Multi-Disciplinary Systematic ReviewProceedings of the 2024 AAAI/ACM Conference on AI, Ethics, and Society10.5555/3716662.3716775(1292-1305)Online publication date: 21-Oct-2024
      • (2024)Fairness for Deep Learning Predictions Using Bias Parity Score Based Loss Function RegularizationInternational Journal on Artificial Intelligence Tools10.1142/S021821302460003033:03Online publication date: 25-Apr-2024
      • (2024)Safeguarding the Future of Artificial Intelligence: An AI BlueprintArtificial Intelligence for Security10.1007/978-3-031-57452-8_1(3-22)Online publication date: 17-Apr-2024
      • (2023)Mitigating Voter Attribute Bias for Fair Opinion AggregationProceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3600211.3604660(170-180)Online publication date: 8-Aug-2023
      • (2022)Data-Centric Factors in Algorithmic FairnessProceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3514094.3534147(396-410)Online publication date: 26-Jul-2022
      • (2022)The Influences of Task Design on Crowdsourced Judgement: A Case Study of Recidivism Risk EvaluationProceedings of the ACM Web Conference 202210.1145/3485447.3512239(1685-1696)Online publication date: 25-Apr-2022
      • (2022)Mitigating Observation Biases in Crowdsourced Label Aggregation2022 26th International Conference on Pattern Recognition (ICPR)10.1109/ICPR56361.2022.9956439(1171-1177)Online publication date: 21-Aug-2022
      • (2021)The Impact of Algorithmic Risk Assessments on Human Predictions and its Analysis via Crowdsourcing StudiesProceedings of the ACM on Human-Computer Interaction10.1145/34795725:CSCW2(1-24)Online publication date: 18-Oct-2021
      • (2021)A sociotechnical view of algorithmic fairnessInformation Systems Journal10.1111/isj.1237032:4(754-818)Online publication date: 7-Oct-2021
      • (2020)Using Bias Parity Score to Find Feature-Rich Models with Least Relative BiasTechnologies10.3390/technologies80400688:4(68)Online publication date: 14-Nov-2020
      • Show More Cited By

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media