At the intersection of humanity and technology: a technofeminist intersectional critical discourse analysis of gender and race biases in the natural language processing model GPT-3

AI & SOCIETY - Trang 1-19 - 2023
M. A. Palacios Barea1,2, D. Boeren1, J. F. Ferreira Goncalves1
1Erasmus University Rotterdam, Rotterdam, The Netherlands
2Delft University of Technology, Delft, The Netherlands

Tóm tắt

Algorithmic biases, or algorithmic unfairness, have been a topic of public and scientific scrutiny for the past years, as increasing evidence suggests the pervasive assimilation of human cognitive biases and stereotypes in such systems. This research is specifically concerned with analyzing the presence of discursive biases in the text generated by GPT-3, an NLPM which has been praised in recent years for resembling human language so closely that it is becoming difficult to differentiate between the human and the algorithm. The pertinence of this research object is substantiated by the identification of race, gender and religious biases in the model’s completions in recent research, suggesting that the model is indeed heavily influenced by human cognitive biases. To this end, this research inquires: How does the Natural Language Processing Model GPT-3 replicate existing social biases?. This question is addressed through the scrutiny of GPT-3’s completions using Critical Discourse Analysis (CDA), a method which has been deemed as amply valuable for this research as it is aimed at uncovering power asymmetries in language. As such, the analysis is specifically centered around the analysis of gender and race biases in the model’s generated text. Research findings suggest that GPT-3’s language generation model significantly exacerbates existing social biases while replicating dangerous ideologies akin to white supremacy and hegemonic masculinity as factual knowledge.

Tài liệu tham khảo

Baker PK, Potts A (2013) Why do white people have thin lips? Google and the perpetuation of stereotypes via auto-complete search forms. Crit Discourse Stud 10(2):187–204. https://doi.org/10.1080/17405904.2012.744320 Balayn A, Gürses S (2021) Beyond debiasing: regulating AI and its inequalities. European Digital Rights (EDRi). Delft University of Technology https://edri.org/wp-content/uploads/2021/09/EDRi_Beyond-Debiasing-Report_Online.pdf. Accessed 12 Dec 2022 Barera M (2020) Mind the gap: addressing structural equity and inclusion on Wikipedia. http://hdl.handle.net/10106/29572 Bender EM, Gebru T, McMillan-Major A, Mitchell S (2021) On the dangers of stochastic parrots: can language models be too big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp 610–623. https://doi.org/10.1145/3442188.3445922 Bishop JM (2021) Artificial intelligence is stupid and causal reasoning will not fix it. Front Psychol 11:1–18. https://doi.org/10.3389/fpsyg.2020.513474 Bonilla-Silva E (2015) The structure of racism in color-blind, “post-racial” America. Am Behav Sci 59:1358–1376. https://doi.org/10.1177/0002764215586826 Bordalo P, Coffman KB, Gennaioli N, Shleifer A (2016) Stereotypes. Q J Econ 131(4):1753–1794. https://doi.org/10.1093/qje/qjw029 Bowser BP (2017) Racism: origin and theory. J Black Stud 48:572–590. https://doi.org/10.1177/0021934717702135 Brown TB, Mann B et al (2020) Language models are few-shot learners. https://arXiv.org/2005.14165. https://doi.org/10.48550/arXiv.2005.14165 Buolamwini J, Gebru T (2018) Gender shades: intersectional accuracy disparities in commercial gender classification. Proc Mach Learn Res 81:1–15 Byrd D, Ceacal Y, Felton J, Nicholson C, Rhaney D, McCray N, Young J (2017) A modern doll study. Race Gend Cl 24(1–2):186–202 Collins HP, Bilge S (2020) Intersectionality. Polity Press, Cambridge Connell RW (2005) Masculinities, 2nd edn. University of California Press, Berkeley Crawford K (2021) Atlas of AI: power, politics, and the planetary costs of artificial intelligence. Yale University Press, New Haven Dale R (2021) Gpt-3: what’s it good for? Nat Lang Eng 27:113–118. https://doi.org/10.1017/S1351324920000601 Davis MD (2016) We were treated like machines: professionalism and anti-blackness in social work agency culture. Masters Thesis, Smith College https://scholarworks.smith.edu/theses/1708 de Boise S (2019) Editorial: is masculinity toxic? NORMA Int J Masculinity Stud 14:147–151. https://doi.org/10.1080/18902138.2019.1654742 Dye L (2009) Consuming constructions: a critique of Dove’s campaign for real beauty. Can J Media Stud 5:114–212 Eagly AH, Wood W (2016) Social role theory of sex differences. The Wiley Blackwell encyclopedia of gender and sexuality stud. Wiley. https://doi.org/10.1002/9781118663219.wbegss183 Eckert P, McConnell-Ginet S (1992) Think practically and look locally: language and gender as community-based practice. Ann Rev Anthropol 21:461–490 Eckert P, McConnell-Ginet S (2003) Language and gender. Camb University Press Engeln-Maddox R (2006) Buying a beauty standard or dreaming of a new life? Expectations associated with media ideals. Psychol Women Q 30:258–266. https://doi.org/10.1111/j.1471-6402.2006.00294.x Farseev A (2023) Council post: is bigger better? Why the ChatGPT Vs. GPT-3 Vs. GPT-4 ‘battle’ is just a family chat. Forbes. https://www.forbes.com/sites/forbestechcouncil/2023/02/17/is-bigger-better-why-the-chatgpt-vs-gpt-3-vs-gpt-4-battle-is-just-a-family-chat/ Feagin J, Elias S (2013) Rethinking racial formation theory: a systemic racism critique. Ethnic Racial Stud 36:931–960. https://doi.org/10.1080/01419870.2012.669839 Fiske ST (1993) Controlling other people: the impact of power on stereotyping. Am Psychol 48(6):621–628. https://doi.org/10.1037/0003-066X.48.6.621 Flexner A (1915) Is social work a profession? The Social Welfare History Project. http://www.socialwelfarehistory.com/social-work/is-social-work-a-profession-1915 Floridi L, Chiriatti M (2020) GPT-3: its nature, scope, limits, and consequences. Mind Mach 30(4):681–694. https://doi.org/10.1007/s11023-020-09548-1 Friedman B, Nissenbaum H (1996) Bias in computer systems. ACM Trans Inform Syst 14(3):330–347. https://doi.org/10.1145/230538.230561 Gardner J, Brooks C, Baker R (2019) Evaluating the fairness of predictive student models through slicing analysis. In: Proceedings of the 9th International Conference on Learning Analytics & Knowledge, pp 225–234. https://doi.org/10.1145/3303772.3303791 Goffman E (1977) The arrangement between the sexes. Theory Soc 4:301–331 Gramsci A (1971) Selections from the prison notebooks. Lawrence and Wishart, London Hamburger ME, Hogben M, McGowan S, Dawson LJ (1996) Assessing hypergender ideologies: development and initial validation of a gender-neutral measure of adherence to extreme gender-role beliefs. J Res Pers 30(2):157–178. https://doi.org/10.1006/jrpe.1996.0011 Haraway DJ (1985) A cyborg manifesto: science, technology, and socialist-feminism in the late twentieth century. Posthumanism. https://doi.org/10.1007/978-1-137-05194-3_10 Hill Collins P (2019) Intersectionality as critical social theory. Duke University Press Hinton PR (2017) Implicit stereotypes and the predictive brain: cognition and culture in “biased” person perception. Palgrave Commun 3(1):1–9. https://doi.org/10.1057/palcomms.2017.86 Hoffmann J, Borgeaud S, Mensch A, Buchatskaya E, Cai T, Rutherford E, Casas D de L, Hendricks LA, Welbl J, Clark A, Hennigan T, Noland E, Millican K, Driessche G van den, Damoc B, Guy A, Osindero S, Simonyan K, Elsen E, Rae JW, Vinyals O, Sifre L (2022) Training compute-optimal large language models. arXiv. http://arxiv.org/abs/2203.15556 Houli D, Radford ML, Singh V (2021) “COVID19 is_”: the perpetuation of coronavirus conspiracy theories via Google autocomplete. In: Proceedings of the Association for Information Science and Technology, vol 58, pp 218–229. https://doi.org/10.1002/pra2.450 Howard A, Isbell AH (2020) Diversity in AI: the invisible men and women. MIT Sloan Management Review. https://sloanreview.mit.edu/article/diversity-in-ai-the-invisible-men-and-women/ Jakesch M, Bhat, A, Buschek D, Zalmanson L, and Naaman M (2023) Co-writing with opinionated language models affects users’ views. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI ‘23), Hamburg, Germany. ACM, New York, NY, USA, p 15. https://doi.org/10.1145/3544548.3581196 Jones LK (2020) Twitter wants you to know that you’re still SOL if you get a death threat—unless you’re President Donald Trump. https://medium.com/@agua.carbonica/twitter-wants-you-to-know-that-youre-still-sol-if-you-get-a-death-threat-unless-you-re-a5cce316b706 Kendall S, Tannen D (2015) Discourse and gender. pp. 548–567. https://doi.org/10.1002/9780470753460.ch29 Koenig AM, Eagly AH (2014) Evidence for the social role theory of stereotype content: observations of groups’ roles shape stereotypes. J Pers Soc Psychol 107:371–392. https://doi.org/10.1037/a0037215 Kollmayer M, Schober B, Spiel C (2018) Gender stereotypes in education: development, consequences, and interventions. Eur J Dev Psychol 15(4):361–377. https://doi.org/10.1080/17405629.2016.1193483 Lapowsky I (2018) Google autocomplete suggestions are still racist, sexist, and science-denying. WIRED. https://wired.com/story/google-autocomplete-vile-suggestions/ LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444. https://doi.org/10.1038/nature14539 Li L, Bamman D (2021) Gender and representation bias in GPT-3 generated stories. In: Proceedings of the Third Workshop on Narrative Understanding, pp 48–55. https://doi.org/10.18653/v1/2021.nuse-1.5 Liu Q, Kusner MJ, Blunsom P (2020) A survey on contextual embeddings. [Cs]. http://arxiv.org/abs/2003.07278 Maas JJC (2022) Machine learning and power relations. Ai & Soc 38:1493–1500 Machin D, Mayr A (2012) How to do critical discourse analysis: a multimodal approach. Sage Magee L, Ghahremanlou L, Soldatic K, Robertson S (2021) Intersectional bias in causal language models. [Cs]. http://arxiv.org/abs/2107.07691 Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2019) A survey on bias and fairness in machine learning. ArXiv E-Prints. https://arxiv.org/abs/1908.09635 Mitchell PW (2018) The fault in his seeds: Lost notes to the case of bias in Samuel George Morton’s cranial race science. PLoS Biol 16(10):e2007008. https://doi.org/10.1371/journal.pbio.2007008 Moule J (2009) Understanding unconscious bias and unintentional racism. Phi Delta Kappan 90(5):320–326. https://doi.org/10.1177/003172170909000504 Murgia M (2019) AI academics under pressure to do commercial research. Financial Times. https://www.ft.com/content/94e86cd0-44b6-11e9-a965-23d669740bfb Nadeem M, Bethke A, Reddy S (2020) Stereoset: measuring stereotypical bias in pretrained language models Nash JC (2008) Re-thinking intersectionality. Fem Rev 89(1):1–15. https://doi.org/10.1057/fr.2008.4 Nelson A (2016) The social life of DNA: race, reparations, and reconciliation after the genome. Beacon Press OpenAI Platform n.d. https://platform.openai.com. Accessed May 2022 O’Neill L, Anantharama N, Buntine W, Angus SD (2021) Quantitative discourse analysis at Scale—AI, NLP and the transformer revolution. In: SoDa Laboratories Working Paper Series (2021–12; SoDa Laboratories Working Paper Series). Monash University, SoDa Laboratories. https://ideas.repec.org/p/ajr/sodwps/2021-12.html O’Sullivan L, Dickerson J (2020) Here are a few ways GPT-3 can go wrong. TechCrunch. https://social.techcrunch.com/2020/08/07/here-are-a-few-ways-gpt-3-can-go-wrong/ Pew (2016) Reddit news users more likely to be male, young and digital in their news preferences. Pew Research Center’s Journalism Project. https://www.pewresearch.org/journalism/2016/02/25/reddit-news-users-more-likely-to-be-male-young-and-digital-in-their-news-preferences/ Pew (2018) Internet/broadband fact sheet. https://www.pewinternet.org/fact-sheet/internet-broadband/ Powell-Hopson D, Hopson DS (1988) Implications of doll color preferences among black preschool children and white preschool children. J Black Psychol 14(2):57–63. https://doi.org/10.1177/00957984880142004 Salles A, Evers K, Farisco M (2020) Anthropomorphism in AI. AJOB Neurosci 11(2):88–95. https://doi.org/10.1080/21507740.2020.1740350 Santurkar S, Durmus E, Ladhak F, Lee C, Liang P, Hashimoto T (2023) Whose Opinions Do Language Models Reflect? (https://arXiv.org/2303.17548). arXiv. https://doi.org/10.48550/arXiv.2303.17548 Schüssler Fiorenza E (2009) Introduction: exploring the intersections of race, gender, status, and ethnicity in early Christian studies. In: Laura Nasrallah, Fiorenza (ed) Prejudice and Christian Beginnings: Investigating Race, Gender, and Ethnicity in Early Christian Studies. pp 1–23 Sengupta U (2021) Monoculturalism, aculturalism, and postculturalism: the exclusionary culture of algorithmic development. Algorithmic culture: how big data and artificial intelligence are transforming everyday life. pp 71–97 Silverman D (2020) Credible qualitative research. Interpreting qualitative data. Sage, pp 352–395 Smith CS (2022) OpenAI is giving Microsoft exclusive access to its GPT-3 language model|MIT Technology Review. https://www.technologyreview.com/2020/09/23/1008729/openai-is-giving-microsoft-exclusive-access-to-its-gpt-3-language-model/ Spencer SJ, Logel C, Davies P (2016) Stereotype threat. Ann Rev Psychol 67(1):415–437. https://doi.org/10.1146/annurev-psych-073115-103235 Staszak J (2009) Other/Otherness. In: Kitchin & Thrift (ed) International encyclopedia of hum geography: A 12- volume set, 1st edn. Oxford, Elsevier Science. https://archive-ouverte.unige.ch/unige:77582 van Sterkenburg J, Knoppers A, de Leeuw S (2012) Constructing racial/ethnic difference in and through Dutch televised soccer commentary. J Sport Soc Issues 36:422–442. https://doi.org/10.1177/0193723512448664 Veerman E (2016) “Welke pop vind je lelijk?” VPRO. https://www.vpro.nl/lees/gids/2016/51/-Welke-pop-vind-je-lelijk.html Wajcman J (2010) Feminist theories of technology. Camb J Econ 34(1):143–152 West SM (2020) AI and the Far Right: A History We Can’t Ignore. Medium. https://medium.com/@AINowInstitute/ai-and-the-far-right-a-history-we-cant-ignore-f81375c3cc57 Whittaker M (2021) The steep cost of capture. SSRN Scholarly Paper No. 4135581. https://papers.ssrn.com/abstract=4135581 Wilson J (2017) People see black men as larger, more threatening than same-sized white men. https://www.apa.org. https://www.apa.org/news/press/releases/2017/03/black-men-threatening Winner L (1980) “Do artifacts have politics?” Emerging technologies: ethics, law and governance, pp 15–30. https://doi.org/10.4324/9781003074960-3 World Bank (2018) Individuals using the internet https://data.worldbank.org/indicator/IT.NET.USER.ZS?end=2017&locations=US&start=2015