skip to main content
10.1145/3287560.3287586acmconferencesArticle/Chapter ViewAbstractPublication PagesfacctConference Proceedingsconference-collections
research-article

Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees

Published: 29 January 2019 Publication History

Abstract

Developing classification algorithms that are fair with respect to sensitive attributes of the data is an important problem due to the increased deployment of classification algorithms in societal contexts. Several recent works have focused on studying classification with respect to specific fairness metrics, modeled the corresponding fair classification problem as constrained optimization problems, and developed tailored algorithms to solve them. Despite this, there still remain important metrics for which there are no fair classifiers with theoretical guarantees; primarily because the resulting optimization problem is non-convex. The main contribution of this paper is a meta-algorithm for classification that can take as input a general class of fairness constraints with respect to multiple non-disjoint and multi-valued sensitive attributes, and which comes with provable guarantees. In particular, our algorithm can handle non-convex "linear fractional" constraints (which includes fairness constraints such as predictive parity) for which no prior algorithm was known. Key to our results is an algorithm for a family of classification problems with convex constraints along with a reduction from classification problems with linear fractional constraints to this family. Empirically, we observe that our algorithm is fast, can achieve near-perfect fairness with respect to various fairness metrics, and the loss in accuracy due to the imposed fairness constraints is often small.

References

[1]
ACM. 2017. Statement on Algorithmic Transparency and Accountability. https:/www.acm.org/binaries/content/assets/public-policy/2017_usacm_statement_algorithms.pdf.
[2]
An Act. 1964. Civil Rights Act. Title VII, Equal Employment Opportunities (1964).
[3]
Alekh Agarwal, Alina Beygelzimer, Miroslav Dudik, John Langford, and Hanna M. Wallach. 2018. A Reductions Approach to Fair Classification. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018. 60--69.
[4]
Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. https://gir.hub.com/propublica/compas-analysis.
[5]
Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks. ProPublica, May (2016).
[6]
Solon Barocas and Andrew D Selbst. 2016. Big data's disparate impact. California Law Review (2016).
[7]
Richard Berk. 2009. The role of race in forecasts of violent crime. Race and social problems (2009).
[8]
Stephen Boyd and Almir Mutapcic. 2008. Stochastic subgradient methods. Lecture Notes for EE364b, Stanford University (2008).
[9]
Toon Calders and Sicco Verwer. 2010. Three naive Bayes approaches for discrimination-free classification. Data Min. Knowl. Discov. 21, 2 (2010), 277--292.
[10]
L Elisa Celis, Lingxiao Huang, and Nisheeth K Vishnoi. 2018. Multiwinner voting with fairness constraints. In Proceedings of the Twenty-seventh International Joint Conference on Artificial Intelligence and the Twenty-third European Conference on Artificial Intelligence, IJCAI-ECAI.
[11]
L. Elisa Celis, Vijay Keswani, Amit Deshpande, Tarun Kathuria, Damian Straszak, and Nisheeth K. Vishnoi. 2018. Fair and Diverse DPP-based Data Summarization. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018.
[12]
L. Elisa Celis, Damian Straszak, and Nisheeth K. Vishnoi. 2018. Ranking with Fairness Constraints. In Proceedings of the fourty-fifth International Colloquium on Automata, Languages, and Programming ICALP.
[13]
L. Elisa Celis and Nisheeth K Vishnoi. 2017. Fair Personalization. In Fairness, Accountability, and Transparency in Machine Learning.
[14]
Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153--163.
[15]
Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic Decision Making and the Cost of Fairness. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2017.797--806.
[16]
Amit Datta, Michael Carl Tschantz, and Anupam Datta. 2015. Automated Experiments on Ad Privacy Settings. Proceedings on Privacy Enhancing Technologies (2015).
[17]
Bill Dedman et al. 1988. The Color of Money. Atlanta Journal-Constitution (1988).
[18]
Dua Dheeru and Efi Karra Taniskidou. 2017. UCI Machine Learning Repository. http://archive.ics.uci.edu/ml.
[19]
Ilias Diakonikolas. 2016. Learning Structured Distributions. Handbook of Big Data 267 (2016).
[20]
William Dieterich, Christina Mendoza, and Tim Brennan. 2016. COMPAS risk scales: Demonstrating accuracy equity and predictive parity. Northpoint Inc (2016).
[21]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Innovations in Theoretical Computer Science 2012. ACM, 214--226.
[22]
Cynthia Dwork, Nicole Immorlica, Adam Tauman Kalai, and Mark D. M. Leiserson. 2018. Decoupled Classifiers for Group-Fair and Efficient Machine Learning. In Fairness, Accountability, and Transparency in Machine Learning. 119--133.
[23]
ENTHOUGHT. 2018. SciPy. https:/www.scipy.org/.
[24]
Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2015. ACM, 259--268.
[25]
Benjamin Fish, Jeremy Kun, and Ádám D Lelkes. 2016. A confidence-based approach for balancing fairness and accuracy. In Proceedings of the 2016 SIAM International Conference on Data Mining. SIAM, 144--152.
[26]
Anthony W Flores, Kristin Bechtel, and Christopher T Lowenkamp. 2016. False Positives, False Negatives, and False Analyses: A Rejoinder to Machine Bias: There's Software Used across the Country to Predict Future Criminals. And It's Biased against Blacks. Fed. Probation 80 (2016), 38.
[27]
Sorelle A Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. 2016. On the (Im) Possibility of Fairness. arXiv preprint arXiv:1609.07236 (2016).
[28]
Sorelle A Friedler, Carlos Scheidegger, Suresh Venkatasubramanian, Sonam Choudhary, Evan P Hamilton, and Derek Roth. 2018. A Comparative Study of Fairness-Enhancing Interventions in Machine Learning. arXiv preprint arXiv: 1802.04422 (2018).
[29]
Naman Goel, Mohammad Yaghini, and Boi Faltings. 2018. Non-Discriminatory Machine Learning through Convex Fairness Criteria. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018.
[30]
Sharad Goel, Justin M Rao, Ravi Shroff, et al. 2016. Precinct or prejudice? Understanding racial disparities in New York City's stop-and-frisk policy. The Annals of Applied Statistics 10, 1 (2016), 365--394.
[31]
Gabriel Goh, Andrew Cotter, Maya R. Gupta, and Michael P. Friedlander. 2016. Satisfying Real-world Goals with Dataset Constraints. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016. 2415--2423.
[32]
Nina Grgić-Hlača, Elissa M Redmiles, Krishna P Gummadi, and Adrian Weller. 2018. Human Perceptions of Fairness in Algorithmic Decision Making: A Case Study of Criminal Risk Prediction. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, WWW 2018. 903--912.
[33]
Nina Grgić-Hlača, Muhammad Bilal Zafar, Krishna P Gummadi, and Adrian Weller. 2018. Beyond Distributive Fairness in Algorithmic Decision Making: Feature Selection for Procedurally Fair Learning. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, 2018.
[34]
Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of Opportunity in Supervised Learning. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016. 3315--3323.
[35]
Mara Hvistendahl. 2016. Can "predictive policing" prevent crime before it happens. Science Magazine 28 (2016).
[36]
Matthew Joseph, Michael Kearns, Jamie H Morgenstern, and Aaron Roth. 2016. Fairness in learning: Classic and contextual bandits. In Advances in Neural Information Processing Systems. 325--333.
[37]
Faisal Kamiran and Toon Calders. 2009. Classifying without discriminating. In Computer, Control and Communication, 2009. IC4 2009. 2nd International Conference on. IEEE, 1--6.
[38]
Faisal Kamiran and Toon Calders. 2012. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems 33, 1 (2012), 1--33.
[39]
Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. 2012. Fairness-Aware Classifier with Prejudice Remover Regularizer. In In Proceedings of Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2012. 35--50.
[40]
Michael Kearns, Aaron Roth, and Zhiwei Steven Wu. 2017. Meritocratic fairness for cross-population selection. In International Conference on Machine Learning.
[41]
Jon M. Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2017. Inherent Trade-Offs in the Fair Determination of Risk Scores. In 8th Innovations in Theoretical Computer Science Conference, ITCS, 2017. 43:1--43:23.
[42]
Emmanouil Krasanakis, Eleftherios Spyromitros-Xioufis, Symeon Papadopoulos, and Yiannis Kompatsiaris. 2018. Adaptive Sensitive Reweighting to Mitigate Bias in Fairness-aware Classification. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, WWW 2018. International World Wide Web Conferences Steering Committee.
[43]
Jeff Larson, Surya Mattu, Lauren Kirchner, and Julia Angwin. 2016. How we analyzed the COMPAS recidivism algorithm. ProPublica (5 2016) 9 (2016).
[44]
Binh Thanh Luong, Salvatore Ruggieri, and Franco Turini. 2011. k-NN as an implementation of situation testing for discrimination discovery and prevention. In Proceedings of the 17th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2011. ACM, 502--510.
[45]
Susan Magarey. 2004. The sex discrimination act 1984. Australian Feminist Law Journal (2004).
[46]
Aditya Krishna Menon and Robert C. Williamson. 2018. The cost of fairness in binary classification. In Conference on Fairness, Accountability and Transparency, FAT 2018. 107--118.
[47]
Claire Cain Miller. 2015. Can an algorithm hire better than a human. The New York Times 25 (2015).
[48]
Ankur Moitra and Gregory Valiant. 2010. Settling the polynomial learnability of mixtures of gaussians. In Foundations of Computer Science (FOCS), 2010 51st Annual IEEE Symposium on. IEEE, 93--102.
[49]
Harikrishna Narasimhan, Rohit Vaish, and Shivani Agarwal. 2014. On the Statistical Consistency of Plug-in Classifiers for Non-decomposable Performance Measures. In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014. 1493--1501.
[50]
Arvind Narayanan. 2018. Tutorial: 21 Fairness Definitions and Their Politics. https:/www.youtube.com/watch?v=jIXIuYdnyyk.
[51]
Northpointe. 2012. Compas risk and need assessment systems. http://www.northpointeinc.com/files/downloads/FAQ_Document.pdf.
[52]
United States. Executive Office of the President and John Podesta. 2014. Big data: Seizing opportunities, preserving values. White House, Executive Office of the President.
[53]
Dino Pedreshi, Salvatore Ruggieri, and Franco Turini. 2008. Discrimination-aware data mining. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2008. ACM, 560--568.
[54]
Geoff Pleiss, Manish Raghavan, Felix Wu, Jon M. Kleinberg, and Kilian Q. Weinberger. 2017. On Fairness and Calibration. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017 5684--5693.
[55]
Novi Quadrianto and Viktoriia Sharmanska. 2017. Recycling Privileged Learning and Distribution Matching for Fairness. In Advances in Neural Information Processing Systems. 677--688.
[56]
WhiteHouse. 2016. Big data: A report on algorithmic systems, opportunity, and civil rights. Executive Office of the President.
[57]
Blake E. Woodworth, Suriya Gunasekar, Mesrob I. Ohannessian, and Nathan Srebro. 2017. Learning Non-Discriminatory Predictors. In Proceedings of the 30th Conference on Learning Theory, COLT 2017. 1920--1953.
[58]
Ke Yang and Julia Stoyanovich. 2017. Measuring Fairness in Ranked Outputs. In SSDBM.
[59]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, and Krishna P. Gummadi. 2017. Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment. In Proceedings of the 26th International Conference on World Wide Web, WWW 2017 1171--1180.
[60]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, and Krishna P. Gummadi. 2017. Fairness Constraints: Mechanisms for Fair Classification. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017 962--970.
[61]
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, Krishna P. Gummadi, and Adrian Weller. 2017. From Parity to Preference-based Notions of Fairness in Classification. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017. 228--238.
[62]
Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In Proceedings of the 30th International Conference on Machine Learning, ICML 2013. 325--333.
[63]
Indre Zliobaite. 2017. Measuring discrimination in algorithmic decision making. Data Min. Knowl. Discov. (2017).

Cited By

View all
  • (2024)Algorithmic Bias in Recommendation Systems and Its Social Impact on User BehaviorInternational Theory and Practice in Humanities and Social Sciences10.70693/itphss.v1i1.2041:1(290-303)Online publication date: 17-Dec-2024
  • (2024)FRAPPÉProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3694044(48321-48343)Online publication date: 21-Jul-2024
  • (2024)Metric-Independent Mitigation of Unpredefined Bias in Machine ClassificationIntelligent Computing10.34133/icomputing.00833Online publication date: 8-Apr-2024
  • Show More Cited By

Index Terms

  1. Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    FAT* '19: Proceedings of the Conference on Fairness, Accountability, and Transparency
    January 2019
    388 pages
    ISBN:9781450361255
    DOI:10.1145/3287560
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 29 January 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Algorithmic Fairness
    2. Classification

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    FAT* '19
    Sponsor:

    Upcoming Conference

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)244
    • Downloads (Last 6 weeks)20
    Reflects downloads up to 22 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Algorithmic Bias in Recommendation Systems and Its Social Impact on User BehaviorInternational Theory and Practice in Humanities and Social Sciences10.70693/itphss.v1i1.2041:1(290-303)Online publication date: 17-Dec-2024
    • (2024)FRAPPÉProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3694044(48321-48343)Online publication date: 21-Jul-2024
    • (2024)Metric-Independent Mitigation of Unpredefined Bias in Machine ClassificationIntelligent Computing10.34133/icomputing.00833Online publication date: 8-Apr-2024
    • (2024)Investigating and Mitigating the Performance–Fairness Tradeoff via Protected-Category SamplingElectronics10.3390/electronics1315302413:15(3024)Online publication date: 31-Jul-2024
    • (2024)Strategies to improve fairness in artificial intelligence:A systematic literature reviewEducation for Information10.3233/EFI-240045(1-24)Online publication date: 25-Jul-2024
    • (2024)Inside the Black Box: Detecting and Mitigating Algorithmic Bias Across Racialized Groups in College Student-Success PredictionAERA Open10.1177/2332858424125874110Online publication date: 10-Jul-2024
    • (2024)MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual PredictionsProceedings of the ACM on Software Engineering10.1145/36608011:FSE(2121-2143)Online publication date: 12-Jul-2024
    • (2024)FairHash: A Fair and Memory/Time-efficient HashmapProceedings of the ACM on Management of Data10.1145/36549392:3(1-29)Online publication date: 30-May-2024
    • (2024)Algorithmic Fairness in Performative Policy Learning: Escaping the Impossibility of Group FairnessProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658929(616-630)Online publication date: 3-Jun-2024
    • (2024)Fairness in Machine Learning: A SurveyACM Computing Surveys10.1145/361686556:7(1-38)Online publication date: 9-Apr-2024
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media