Abstract
Large language models (LLMs) have the potential to generate significant benefits, but their blanket application in Africa could exacerbate existing social and economic inequalities. This is due to a number of factors, including limited technological advancement, historical injustice and marginalization, and underrepresentation of African languages, values, and norms in training data. Despite comprising nearly one-third of the world’s languages, most African languages are underrepresented on the internet: they are primarily oral with little available in written and digitized form. Additionally, most African languages have conflicting orthographic standards. While Africa is undergoing a digital transformation, both internet connectivity and digital literacy remain relatively low and unevenly distributed. This lack of online representation for African languages limits the availability of natural language data for training inclusive language models. This paper examines the potential harms of LLMs in Africa, covering harms already documented for the African context; harms studied and documented for the Western context, but previously unapplied to Africa; and novel potential harms based on the norms, values, practices, and contextual factors of the African continent. This work aims to contribute to a better understanding of potential harms of LLMs in Africa, which in turn could inform and support the development of more inclusive LLMs.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Adams, R.: AI in Africa: Key Concerns and Policy Considerations for the Future of the Continent (2022). https://afripoli.org/ai-in-africa-key-concerns-and-policy-considerations-for-the-future-of-the-continent
African Union. Toolkit on Disability Africa (2017). https://www.un.org/esa/socdev/documents/disability/Toolkit/Intro-UN-CRPD.pdf
Standing Voice (2021)
Amnesty International. Nigeria: We dried our tears: Addressing the toll on children of Northeast Nigeria’s conflict. Amnesty International Publications (2020). https://sitn.hms.harvard.edu/flash/2020/racial-discrimination-in-face-recognition-technology/
Carlini, N., et al.: Extracting Training Data from Large Language Models (2021). http://arxiv.org/abs/2012.07805. arXiv:2012.07805
Chen, R.J., et al.: Synthetic data in machine learning for medicine and healthcare. Nat. Biomed. Eng. 5(6), 493–497 (2021). https://www.nature.com/articles/s41551-021-00751-8
CIPESA. Kenya’s 2022 Political Sphere Overwhelmed by Disinformation (2022). https://cipesa.org/2022/07/11187-2/
Dennis Tang Discusses the Future of A.I. Poetry at Literary Hub: https://www.poetryfoundation.org/harriet-books/2020/01/dennis-tang-discusses-the-future-of-ai-poetry-at-literary-hub
Google PAIR. People + AI Guidebook. Google (2019). https://design.google/ai-guidebook
Gwagwa, A., Kazim, E., Hilliard, A.: The role of the African value of Ubuntu in global AI inclusion discourse: a normative ethics perspective (2022). https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9023883/
Hardesty (2018). https://news.mit.edu/2018/study-finds-gender-skin-type-bias-artificial-intelligence-systems-0212
Hao, K.: A college kid’s fake, AI-generated blog fooled tens of thousands. This is how he made it. MIT Technology Review (2020). https://www.technologyreview.com/2020/08/14/1006780/ai-gpt-3-fake-blog-reached-top-of-hacker-news/
Hassan, I.: Disinformation Is Undermining Democracy in West Africa (2022). https://www.cigionline.org/articles/disinformation-is-undermining-democracy-in-west-africa/
Human Rights Measurement Initiative. Measuring Online Harassment: A Technical Report (2021). https://hrmi.org/measuring-online-harassment-a-technical-report/
Human Rights Watch/Africa. Shattered lives: Sexual Violence during the Rwandan Genocide and its Aftermath (1996). https://www.hrw.org/reports/1996/Rwanda.htm
INTERPOL report identifies top cyberthreats in Africa. https://www.interpol.int/en/News-and-Events/News/2021/INTERPOL-report-identifies-top-cyberthreats-in-Africa
Jousse, L.: Discrimination and gender inequalities in Africa: what about equality between women and men? (2021). https://igg-geo.org/?p=3863&lang=en
Kenya’s National Action Plan Against Hate Speech. https://cohesion.or.ke/images/docs/downloads/Kenyas_National_Action_Plan_Against_Hate_Speech.pdf
Kim, Y., Sundar, S.S.: Anthropomorphism of computers: is it mindful or mindless? Comput. Hum. Behav. 28(1), 241–250 (2012)
Kenton, Z., et al.: Alignment of Language Agents. http://arxiv.org/abs/2103.14659. arXiv:2103.14659
Kwon, K.H., Moon, S.I., Stefanone, M.A.: Unspeaking on Facebook? Testing network effects on self censorship of political expressions in social network sites. Qual. Quant. (2015). https://doi.org/10.1007/s11135-014-0078-8
Lewis, P., Stenetorp, P., Riedel, S.: Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets (2008). http://arxiv.org/abs/2008.02637. arXiv:2008.02637
Machirori, F.: Tackling online hate speech in Africa and beyond: “We can’t trust Big Tech to abide by its own rules” (2022). https://www.apc.org/en/news/tackling-online-hate-speech-africa-and-beyond-we-cant-trust-big-tech-abide-its-own-rules
Madgavkar, A.: A conversation on artificial intelligence and gender bias (2021). https://www.mckinsey.com/featured-insights/asia-pacific/a-conversation-on-artificial-intelligence-and-gender-bias
Martin, J.L., Wright, K.E.: Bias in Automatic Speech Recognition: The Case of African American Language Applied Linguistics 2022: XX/XX: 1–18 (2022). https://doi.org/10.1093/applin/amac066
McGuffie, K., Newhouse, A.: The Radicalization Risks of GPT-3 and Advanced Neural Language Models (2009). http://arxiv.org/abs/2009.06807. arXiv:2009.06807
Miriri: Kenya orders Meta’s Facebook to tackle hate speech or face suspension (2022). https://www.reuters.com/world/africa/kenyas-cohesion-watchdog-gives-meta-7-days-comply-with-regulations-2022-07-29/
Tshepo, M., Mthabela, Z., Zvobgo, T.: Examining Racial Bias in Chatbot Responses to Customer Inquiries. South Africa (2020)
Negasha, F.: Globalization and the Role of African Languages for Development. In: Proceedings of “Language Communities or Cultural Empires Conference”, 9–11 February 2005, UC Berkeley (2005). https://escholarship.org/content/qt05m659jt/qt05m659jt_noSplash_b6ab52b57a571da5c12c42917270bf40.pdf
Nanfuka, J.: Data Privacy still a neglected Digital Right in Africa (2022). https://cipesa.org/2022/01/data-privacy-still-a-neglected-digital-right-in-africa/
Nanfuka, J.: A Section of Uganda’s Computer Misuse Act Outlawed! But, the Greater Part of the Law Remains Thorny (2023). https://cipesa.org/2023/01/a-section-of-ugandas-computer-misuse-act-outlawed-but-the-greater-part-of-the-law-remains-thorny/
Ogueji, K.: AfriBERTa: Towards Viable Multilingual Language Models for Low-resource Languages, a thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Computer Science (2022)
Quach, K.: Researchers made an OpenAI GPT-3 medical chatbot as an experiment. It told a mock patient to kill themselves (2020). https://www.theregister.com/2020/10/28/gpt3_medical_chatbot_experiment/
Rong, G.: Punishment Island: Where pregnant Ugandan girls were left to die (2017). https://news.cgtn.com/news/3d45544f30597a4d/share_p.html
Schmidt, A.T., Engelen, B.: The ethics of nudging: An overview. Philosophy Compass (2020). https://onlinelibrary.wiley.com/doi/10.1111/phc3.12658
Stańczak, K., Augenstein, S.: A Survey on Gender Bias in Natural Language Processing (2021)
Zdenek, S.: “Just roll your mouse over me”: designing virtual women for customer service on the web. Tech. Commun. Q. 16(4), 397–430 (2007)
Wang, Y., Kosinski, M.: Deep neural networks are more accurate than humans at detecting sexual orientation from facial images. J. Pers. Soc. Psychol. 114(2), 246–325 (2018)
Weidinger, L., et al.: Ethical and social risks of harm from Language Models (2021). https://arxiv.org/pdf/2112.04359.pdf
WHO-Africa: Countering COVID-19 Misinformation in Africa: On a continent of 1.3 billion people, WHO and partners are working to reduce social media-driven health myths (2021). https://www.afro.who.int/news/countering-covid-19-misinformation-africa-continent-13-billion-people-who-and-partners-are
World Bank: Social Exclusion in Africa (2017). https://www.worldbank.org/en/region/afr/brief/social-inclusion-in-africa
Wright, L.J., Martin, K.E.: Bias in Automatic Speech Recognition: The Case of African American Language (2022). https://academic.oup.com/applij/advance-article/doi/10.1093/applin/amac066/6901317
What Are the Latest Scams To Look Out For In Africa (2022): https://transfy.io/blog/latest-scams-to-look-out-for-in-africa-2022
What Are Large Language Models Used For? https://blogs.nvidia.com/blog/2023/01/26/what-are-large-language-models-used-for/
Zhao, J., et al.: Gender Bias in Contextualized Word Embeddings (2019). http://arxiv.org/abs/1904.03310. arXiv:1904.03310
Acknowledgement
This work was supported by funding from Google’s Research Collabs Africa program 2022/2023.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 ICST Institute for Computer Sciences, Social Informatics and Telecommunications Engineering
About this paper
Cite this paper
Baguma, R., Namuwaya, H., Nakatumba-Nabende, J., Rashid, Q.M. (2024). Examining Potential Harms of Large Language Models (LLMs) in Africa. In: Tchakounte, F., Atemkeng, M., Rajagopalan, R.P. (eds) Safe, Secure, Ethical, Responsible Technologies and Emerging Applications. SAFER-TEA 2023. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 566. Springer, Cham. https://doi.org/10.1007/978-3-031-56396-6_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-56396-6_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-56395-9
Online ISBN: 978-3-031-56396-6
eBook Packages: Computer ScienceComputer Science (R0)