skip to main content
10.1145/3531146.3533217acmotherconferencesArticle/Chapter ViewAbstractPublication PagesfacctConference Proceedingsconference-collections
research-article
Open access

Surfacing Racial Stereotypes through Identity Portrayal

Published: 20 June 2022 Publication History

Abstract

Content warning: this paper discusses and contains content that may be offensive or upsetting.
People express racial stereotypes through conversations with others, increasingly in a digital format; as a result, the ability to computationally identify racial stereotypes could be beneficial to help mitigate some of the harmful effects of stereotyping. In this work, we seek to better understand how we can computationally surface racial stereotypes in text by identifying linguistic features associated with differences in racial identity portrayal, focused on two races (Black and White). We collect novel data of individuals’ self-presentation via crowdsourcing, where each crowdworker answers a set of prompts from their own perspective (real identity), and from the perspective of another racial identity (portrayed identity), keeping the gender constant. We use these responses as a dataset to identify stereotypes. Through a series of experiments based on classifications between real and portrayed identities, we show that generalizations and stereotypes appear to be more prevalent amongst white participants than black participants. Through analyses of predictive words and word usage patterns, we find that some of the most predictive features of an author portraying a different racial identity are known stereotypes, and reveal how people of different identities see themselves and others.

References

[1]
Omar Ali, Nancy Scheidt, Alexander Gegov, Ella Haig, Mo Adda, and Benjamin Aziz. 2020. Automated Detection of Racial Microaggressions using Machine Learning. In 2020 IEEE Symposium Series on Computational Intelligence (SSCI). 2477–2484. https://doi.org/10.1109/SSCI47803.2020.9308569
[2]
Ketra L. Armstrong. 1999. Nike’s Communication with Black Audiences: A Sociological Analysis of Advertising Effectiveness via Symbolic Interactionism. Journal of Sport and Social Issues 23, 3 (Aug. 1999), 266–286. https://doi.org/10.1177/0193723599233003 Publisher: SAGE Publications Inc.
[3]
Yvette Assilaméhou, Nadia Lepastourel, and Benoit Testé. 2013. How the Linguistic Intergroup Bias Affects Group Perception: Effects of Language Abstraction on Generalization to the Group. The Journal of Social Psychology 153, 1 (Jan. 2013), 98–108. https://doi.org/10.1080/00224545.2012.711380
[4]
Zinzi D. Bailey, Nancy Krieger, Madina Agénor, Jasmine Graves, Natalia Linos, and Mary T. Bassett. 2017. Structural racism and health inequities in the USA: evidence and interventions. The Lancet 389, 10077 (2017), 1453–1463. https://doi.org/10.1016/S0140-6736(17)30569-X
[5]
Hilary B. Bergsieker, J. Nicole Shelton, and Jennifer A. Richeson. 2010. To be liked versus respected: Divergent goals in interracial interactions. Journal of Personality and Social Psychology 99, 2 (Aug. 2010), 248–264. https://doi.org/10.1037/a0018474
[6]
Tolga Bolukbasi, Kai-Wei Chang, James Zou, Venkatesh Saligrama, and Adam Kalai. 2016. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. In Proceedings of the 30th International Conference on Neural Information Processing Systems (Barcelona, Spain) (NIPS’16). Curran Associates Inc., Red Hook, NY, USA, 4356–4364.
[7]
James J. Bradac and Randall Wisegarver. 1984. Ascribed Status, Lexical Diversity, and Accent: Determinants of Perceived Status, Solidarity, and Control of Speech Style. Journal of Language and Social Psychology 3, 4 (Dec. 1984), 239–255. https://doi.org/10.1177/0261927X8400300401 Publisher: SAGE Publications Inc.
[8]
Jennifer E. Bruening. 2005. Gender and Racial Analysis in Sport: Are All the Women White and All the Blacks Men?Quest 57, 3 (Aug. 2005), 330–349. https://doi.org/10.1080/00336297.2005.10491861
[9]
Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (14 April 2017), 183–186. https://doi.org/10.1126/science.aal4230
[10]
Miguel Ángel Cano, Seth J. Schwartz, David P. MacKinnon, Brian T. H. Keum, Guillermo Prado, Flavio F. Marsiglia, Christopher P. Salas-Wright, Cory L. Cobb, Luz M. Garcini, Mario De La Rosa, Mariana Sánchez, Abir Rahman, Laura M. Acosta, Angelica M. Roncancio, and Marcel A. de Dios. 2021. Exposure to ethnic discrimination in social media and symptoms of anxiety and depression among Hispanic emerging adults: Examining the moderating role of gender. Journal of Clinical Psychology 77, 3 (2021), 571–586. https://doi.org/10.1002/jclp.23050 _eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/jclp.23050.
[11]
Jordan Carpenter, Daniel Preotiuc-Pietro, Lucie Flekova, Salvatore Giorgi, Courtney Hagan, Margaret L. Kern, Anneke E. K. Buffone, Lyle Ungar, and Martin E. P. Seligman. 2017. Real Men Don’t Say “Cute”: Using Automatic Language Analysis to Isolate Inaccurate Aspects of Stereotypes. Social Psychological and Personality Science 8, 3 (April 2017), 310–322. https://doi.org/10.1177/1948550616671998
[12]
Cindy Chung and James Pennebaker. 2007. The Psychological Functions of Function Words.In Social communication.Psychology Press, New York, NY, US, 343–359.
[13]
D Anthony Clark, Lisa Spanierman, Tamilia Reed, Jason Soble, and Sharon Cabana. 2011. Documenting Web log Expressions of Racial Microaggressions That Target American Indians. Journal of Diversity in Higher Education 4 (March 2011), 39–50. https://doi.org/10.1037/a0021762
[14]
Kimberle Crenshaw. [n.d.]. Demarginalizing the Intersection of Race and Sex: A Black Feminist Critique of Antidiscrimination Doctrine, Feminist Theory and Antiracist Politics. ([n. d.]), 31.
[15]
Jenna Cryan, Shiliang Tang, Xinyi Zhang, Miriam Metzger, Haitao Zheng, and Ben Y. Zhao. 2020. Detecting Gender Stereotypes: Lexicon vs. Supervised Learning Methods. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. ACM, Honolulu HI USA, 1–11. https://doi.org/10.1145/3313831.3376488
[16]
Jessie Daniels. 2013. Race and racism in Internet Studies: A review and critique. New Media & Society 15, 5 (Aug. 2013), 695–719. https://doi.org/10.1177/1461444812462849
[17]
Thomas Davidson, Debasmita Bhattacharya, and Ingmar Weber. 2019. Racial Bias in Hate Speech and Abusive Language Detection Datasets. In Proceedings of the Third Workshop on Abusive Language Online. Association for Computational Linguistics, Florence, Italy, 25–35. https://doi.org/10.18653/v1/W19-3504
[18]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, 4171–4186. https://doi.org/10.18653/v1/N19-1423
[19]
Ashley Fetters. 2020. The Many Faces of the ‘Wine Mom’. https://www.theatlantic.com/family/archive/2020/05/wine-moms-explained/612001/
[20]
Anjalie Field, Su Lin Blodgett, Zeerak Waseem, and Yulia Tsvetkov. 2021. A Survey of Race, Racism, and Anti-Racism in NLP. arXiv:2106.11410 [cs] (July 2021). http://arxiv.org/abs/2106.11410 arXiv:2106.11410.
[21]
Gilbert C. Gee and Chandra L. Ford. 2011. STRUCTURAL RACISM AND HEALTH INEQUITIES: Old Issues, New Directions. Du Bois Review: Social Science Research on Race 8, 1 (2011), 115–132. https://doi.org/10.1017/S1742058X11000130
[22]
Joelle Sano Gilmore and Amy Jordan. 2012. Burgers and basketball: Race and stereotypes in food and beverage advertising aimed at children in the US. Journal of Children and Media 6, 3 (Aug. 2012), 317–332. https://doi.org/10.1080/17482798.2012.673498
[23]
Samuel R Gross, Maurice Possley, and Klara Stephens. [n.d.]. Race and Wrongful Convictions in the United States. ([n. d.]).
[24]
McKenzie Himelein-Wachowiak, Salvatore Giorgi, Amanda Devoto, Muhammad Rahman, Lyle Ungar, H Andrew Schwartz, David H Epstein, Lorenzo Leggio, Brenda Curtis, 2021. Bots and misinformation spread on social media: Implications for COVID-19. Journal of Medical Internet Research 23, 5 (2021), e26933.
[25]
Juliana Menasce Horowitz, Anna Brown, and Kiana Cox. 2019. The role of race and ethnicity in Americans’ lives. https://www.pewresearch.org/social-trends/2019/04/09/the-role-of-race-and-ethnicity-in-americans-personal-lives/
[26]
Brandon A. Jackson and Mary Margaret Hui. 2017. Looking for Brothers: Black Male Bonding at a Predominantly White Institution. Journal of Negro Education 86, 4 (2017), 463–478. https://muse.jhu.edu/article/802718 Publisher: Journal of Negro Education.
[27]
May Jiang and Christiane Fellbaum. 2020. Interdependencies of Gender and Race in Contextualized Word Embeddings. In Proceedings of the Second Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics, Barcelona, Spain (Online), 17–25. https://aclanthology.org/2020.gebnlp-1.2
[28]
Kenneth Joseph and Jonathan Morgan. 2020. When do Word Embeddings Accurately Reflect Surveys on our Beliefs About People?. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 4392–4415. https://doi.org/10.18653/v1/2020.acl-main.405
[29]
Anastacia Kurylo. 2012. What Are They Like? Non-Expert Definitions of Stereotypes and Their Implications for Stereotype Maintenance. Qualitative Research in Psychology 9, 4 (Oct. 2012), 337–350. https://doi.org/10.1080/14780887.2010.500517
[30]
Ryan Lavalley and Khalilah Robinson Johnson. 2020. Occupation, injustice, and anti-Black racism in the United States of America. Journal of Occupational Science (Sept. 2020), 1–13. https://doi.org/10.1080/14427591.2020.1810111
[31]
Michael Lepori. 2020. Unequal Representations: Analyzing Intersectional Biases in Word Embeddings Using Representational Similarity Analysis. In Proceedings of the 28th International Conference on Computational Linguistics. International Committee on Computational Linguistics, Barcelona, Spain (Online), 1720–1728. https://doi.org/10.18653/v1/2020.coling-main.151
[32]
Xiao Ma, Jeff Hancock, and Mor Naaman. 2016. Anonymity, Intimacy and Self-Disclosure in Social Media. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. ACM, San Jose California USA, 3857–3869. https://doi.org/10.1145/2858036.2858414
[33]
C. Neil Macrae, Charles Stangor, Miles Hewstone, and Leslie A Zebrowitz. 1996. Physical appearance as a basis of stereotyping. Guilford Press, 79–120.
[34]
Thomas Manzini, Lim Yao Chong, Alan W Black, and Yulia Tsvetkov. 2019. Black is to Criminal as Caucasian is to Police: Detecting and Removing Multiclass Bias in Word Embeddings. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, 615–621. https://doi.org/10.18653/v1/N19-1062
[35]
Henrika McCoy. 2020. Black Lives Matter, and Yes, You are Racist: The Parallelism of the Twentieth and Twenty-First Centuries. Child and Adolescent Social Work Journal 37, 5 (Oct. 2020), 463–475. https://doi.org/10.1007/s10560-020-00690-4
[36]
Jack Merullo, Luke Yeh, Abram Handler, Alvin Grissom II, Brendan O’Connor, and Mohit Iyyer. 2019. Investigating Sports Commentator Bias within a Large Corpus of American Football Broadcasts. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 6355–6361. https://doi.org/10.18653/v1/D19-1666
[37]
James Moody. 2001. Race, School Integration, and Friendship Segregation in America. Amer. J. Sociology 107, 3 (Nov. 2001), 679–716. https://doi.org/10.1086/338954 Publisher: The University of Chicago Press.
[38]
Samuel D. Museus and Kimberly A. Truong. 2013. Racism and Sexism in Cyberspace: Engaging Stereotypes of Asian American Women and Men to Facilitate Student Learning and Development. About Campus: Enriching the Student Learning Experience 18, 4 (Sept. 2013), 14–21. https://doi.org/10.1002/abc.21126
[39]
Anna P. Goddu, Katie J. O’Conor, Sophie Lanzkron, Mustapha O. Saheed, Somnath Saha, Monica E. Peek, Carlton Haywood, and Mary Catherine Beach. 2018. Do Words Matter? Stigmatizing Language and the Transmission of Bias in the Medical Record. Journal of General Internal Medicine 33, 5 (May 2018), 685–691. https://doi.org/10.1007/s11606-017-4289-2
[40]
Orestis Papakyriakopoulos, Simon Hegelich, Juan Carlos Medina Serrano, and Fabienne Marco. 2020. Bias in Word Embeddings. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (Barcelona, Spain) (FAT* ’20). Association for Computing Machinery, New York, NY, USA, 446–457. https://doi.org/10.1145/3351095.3372843
[41]
J. Pennebaker, Cindy K. Chung, Molly Ireland, A. Gonzales, and R. Booth. 2011. The Development and Psychometric Properties of LIWC2007.
[42]
Rosentene B. Purnell. 1982. Teaching Them to Curse: Racial Bias in Language, Pedagogy and Practices. Phylon (1960-) 43, 3 (1982), 231–241. https://doi.org/10.2307/274820 Publisher: Clark Atlanta University.
[43]
Verónica Pérez-Rosas, Quincy Davenport, Anna Mengdan Dai, Mohamed Abouelenien, and Rada Mihalcea. 2017. Identity Deception Detection. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Asian Federation of Natural Language Processing, Taipei, Taiwan, 885–894. https://www.aclweb.org/anthology/I17-1089
[44]
Douglas Rice, Jesse H Rhodes, and Tatishe Nteta. 2019. Racial bias in legal language. Research & Politics 6, 2 (2019), 2053168019848930.
[45]
Douglas Rice, Jesse H. Rhodes, and Tatishe Nteta. 2019. Racial bias in legal language. Research & Politics 6, 2 (April 2019), 2053168019848930. https://doi.org/10.1177/2053168019848930 Publisher: SAGE Publications Ltd.
[46]
Cynthia L. Robinson. 2011. Hair as Race: Why “Good Hair” May Be Bad for Black Females. Howard Journal of Communications 22, 4 (Oct. 2011), 358–376. https://doi.org/10.1080/10646175.2011.617212
[47]
Zick Rubin. 1975. Disclosing oneself to a stranger: Reciprocity and its limits. Journal of Experimental Social Psychology 11, 3 (1975), 233–260. https://doi.org/10.1016/S0022-1031(75)80025-4
[48]
Maarten Sap, Dallas Card, Saadia Gabriel, Yejin Choi, and Noah A. Smith. 2019. The Risk of Racial Bias in Hate Speech Detection. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 1668–1678. https://doi.org/10.18653/v1/P19-1163
[49]
Laura West Steck, Druann Maria Heckert, and D. Alex Heckert. 2003. The salience of racial identity among African-American and white students. Race and Society 6, 1 (Jan. 2003), 57–73. https://doi.org/10.1016/j.racsoc.2004.09.005
[50]
Amelia Tait. 2018. Karen, Sharon, Becky, and Chad: How it feels when your name becomes a meme. https://www.newstatesman.com/science-tech/2018/01/karen-sharon-becky-and-chad-how-it-feels-when-your-name-becomes-meme
[51]
Rob Voigt, Nicholas P. Camp, Vinodkumar Prabhakaran, William L. Hamilton, Rebecca C. Hetey, Camilla M. Griffiths, David Jurgens, Dan Jurafsky, and Jennifer L. Eberhardt. 2017. Language from police body camera footage shows racial disparities in officer respect. Proceedings of the National Academy of Sciences 114, 25 (June 2017), 6521–6526. https://doi.org/10.1073/pnas.1702413114
[52]
Zeerak Waseem. 2016. Are You a Racist or Am I Seeing Things? Annotator Influence on Hate Speech Detection on Twitter. In Proceedings of the First Workshop on NLP and Computational Social Science. Association for Computational Linguistics, Austin, Texas, 138–142. https://doi.org/10.18653/v1/W16-5618
[53]
Amanda Williams, Clio Oliver, Katherine Aumer, and Chanel Meyers. 2016. Racial microaggressions and perceptions of Internet memes. Computers in Human Behavior 63 (Oct. 2016), 424–432. https://doi.org/10.1016/j.chb.2016.05.067
[54]
Guanhua Zhang, Bing Bai, Junqi Zhang, Kun Bai, Conghui Zhu, and Tiejun Zhao. 2020. Demographics Should Not Be the Reason of Toxicity: Mitigating Discrimination in Text Classifications with Instance Weighting. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 4134–4145. https://doi.org/10.18653/v1/2020.acl-main.380

Cited By

View all
  • (2025)Large language models that replace human participants can harmfully misportray and flatten identity groupsNature Machine Intelligence10.1038/s42256-025-00986-zOnline publication date: 17-Feb-2025
  • (2024)Ethnic Classifications in Algorithmic Fairness: Concepts, Measures and Implications in PracticeProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658902(237-253)Online publication date: 3-Jun-2024
  • (2024)Towards robust neural networks: Exploring counterfactual causality-based repairExpert Systems with Applications10.1016/j.eswa.2024.125082257(125082)Online publication date: Dec-2024
  • Show More Cited By

Index Terms

  1. Surfacing Racial Stereotypes through Identity Portrayal
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Other conferences
      FAccT '22: Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency
      June 2022
      2351 pages
      ISBN:9781450393522
      DOI:10.1145/3531146
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 20 June 2022

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. datasets
      2. racial bias
      3. stereotypes

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Funding Sources

      Conference

      FAccT '22
      Sponsor:

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)521
      • Downloads (Last 6 weeks)81
      Reflects downloads up to 08 Mar 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2025)Large language models that replace human participants can harmfully misportray and flatten identity groupsNature Machine Intelligence10.1038/s42256-025-00986-zOnline publication date: 17-Feb-2025
      • (2024)Ethnic Classifications in Algorithmic Fairness: Concepts, Measures and Implications in PracticeProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658902(237-253)Online publication date: 3-Jun-2024
      • (2024)Towards robust neural networks: Exploring counterfactual causality-based repairExpert Systems with Applications10.1016/j.eswa.2024.125082257(125082)Online publication date: Dec-2024
      • (2023)Whose opinions do language models reflect?Proceedings of the 40th International Conference on Machine Learning10.5555/3618408.3619652(29971-30004)Online publication date: 23-Jul-2023
      • (2023)Machine Ethics Research: Promises and Potential PitfallsIEEE Intelligent Systems10.1109/MIS.2023.328316938:4(62-68)Online publication date: 1-Jul-2023
      • (2023)Author Gender Identification Considering Gender BiasArtificial Intelligence and Cognitive Science10.1007/978-3-031-26438-2_17(214-225)Online publication date: 23-Feb-2023
      • (2022)Detection of Racist Language in French TweetsInformation10.3390/info1307031813:7(318)Online publication date: 29-Jun-2022
      • (2022)COVID-19 bailout nationalism: a predicament in saving small immigrant-owned businesses in South AfricaInternational Journal of Research in Business and Social Science (2147- 4478)10.20525/ijrbs.v11i6.196111:6(33-42)Online publication date: 12-Sep-2022

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format.

      HTML Format

      Login options

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media