At the intersection of humanity and technology: a technofeminist intersectional critical discourse analysis of gender and race biases in the natural language processing model GPT-3
AI & SOCIETY - Trang 1-19 - 2023
Tóm tắt
Algorithmic biases, or algorithmic unfairness, have been a topic of public and scientific scrutiny for the past years, as increasing evidence suggests the pervasive assimilation of human cognitive biases and stereotypes in such systems. This research is specifically concerned with analyzing the presence of discursive biases in the text generated by GPT-3, an NLPM which has been praised in recent years for resembling human language so closely that it is becoming difficult to differentiate between the human and the algorithm. The pertinence of this research object is substantiated by the identification of race, gender and religious biases in the model’s completions in recent research, suggesting that the model is indeed heavily influenced by human cognitive biases. To this end, this research inquires: How does the Natural Language Processing Model GPT-3 replicate existing social biases?. This question is addressed through the scrutiny of GPT-3’s completions using Critical Discourse Analysis (CDA), a method which has been deemed as amply valuable for this research as it is aimed at uncovering power asymmetries in language. As such, the analysis is specifically centered around the analysis of gender and race biases in the model’s generated text. Research findings suggest that GPT-3’s language generation model significantly exacerbates existing social biases while replicating dangerous ideologies akin to white supremacy and hegemonic masculinity as factual knowledge.
Tài liệu tham khảo
Baker PK, Potts A (2013) Why do white people have thin lips? Google and the perpetuation of stereotypes via auto-complete search forms. Crit Discourse Stud 10(2):187–204. https://doi.org/10.1080/17405904.2012.744320
Balayn A, Gürses S (2021) Beyond debiasing: regulating AI and its inequalities. European Digital Rights (EDRi). Delft University of Technology https://edri.org/wp-content/uploads/2021/09/EDRi_Beyond-Debiasing-Report_Online.pdf. Accessed 12 Dec 2022
Barera M (2020) Mind the gap: addressing structural equity and inclusion on Wikipedia. http://hdl.handle.net/10106/29572
Bender EM, Gebru T, McMillan-Major A, Mitchell S (2021) On the dangers of stochastic parrots: can language models be too big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp 610–623. https://doi.org/10.1145/3442188.3445922
Bishop JM (2021) Artificial intelligence is stupid and causal reasoning will not fix it. Front Psychol 11:1–18. https://doi.org/10.3389/fpsyg.2020.513474
Bonilla-Silva E (2015) The structure of racism in color-blind, “post-racial” America. Am Behav Sci 59:1358–1376. https://doi.org/10.1177/0002764215586826
Bordalo P, Coffman KB, Gennaioli N, Shleifer A (2016) Stereotypes. Q J Econ 131(4):1753–1794. https://doi.org/10.1093/qje/qjw029
Bowser BP (2017) Racism: origin and theory. J Black Stud 48:572–590. https://doi.org/10.1177/0021934717702135
Brown TB, Mann B et al (2020) Language models are few-shot learners. https://arXiv.org/2005.14165. https://doi.org/10.48550/arXiv.2005.14165
Buolamwini J, Gebru T (2018) Gender shades: intersectional accuracy disparities in commercial gender classification. Proc Mach Learn Res 81:1–15
Byrd D, Ceacal Y, Felton J, Nicholson C, Rhaney D, McCray N, Young J (2017) A modern doll study. Race Gend Cl 24(1–2):186–202
Collins HP, Bilge S (2020) Intersectionality. Polity Press, Cambridge
Connell RW (2005) Masculinities, 2nd edn. University of California Press, Berkeley
Crawford K (2021) Atlas of AI: power, politics, and the planetary costs of artificial intelligence. Yale University Press, New Haven
Dale R (2021) Gpt-3: what’s it good for? Nat Lang Eng 27:113–118. https://doi.org/10.1017/S1351324920000601
Davis MD (2016) We were treated like machines: professionalism and anti-blackness in social work agency culture. Masters Thesis, Smith College https://scholarworks.smith.edu/theses/1708
de Boise S (2019) Editorial: is masculinity toxic? NORMA Int J Masculinity Stud 14:147–151. https://doi.org/10.1080/18902138.2019.1654742
Dye L (2009) Consuming constructions: a critique of Dove’s campaign for real beauty. Can J Media Stud 5:114–212
Eagly AH, Wood W (2016) Social role theory of sex differences. The Wiley Blackwell encyclopedia of gender and sexuality stud. Wiley. https://doi.org/10.1002/9781118663219.wbegss183
Eckert P, McConnell-Ginet S (1992) Think practically and look locally: language and gender as community-based practice. Ann Rev Anthropol 21:461–490
Eckert P, McConnell-Ginet S (2003) Language and gender. Camb University Press
Engeln-Maddox R (2006) Buying a beauty standard or dreaming of a new life? Expectations associated with media ideals. Psychol Women Q 30:258–266. https://doi.org/10.1111/j.1471-6402.2006.00294.x
Farseev A (2023) Council post: is bigger better? Why the ChatGPT Vs. GPT-3 Vs. GPT-4 ‘battle’ is just a family chat. Forbes. https://www.forbes.com/sites/forbestechcouncil/2023/02/17/is-bigger-better-why-the-chatgpt-vs-gpt-3-vs-gpt-4-battle-is-just-a-family-chat/
Feagin J, Elias S (2013) Rethinking racial formation theory: a systemic racism critique. Ethnic Racial Stud 36:931–960. https://doi.org/10.1080/01419870.2012.669839
Fiske ST (1993) Controlling other people: the impact of power on stereotyping. Am Psychol 48(6):621–628. https://doi.org/10.1037/0003-066X.48.6.621
Flexner A (1915) Is social work a profession? The Social Welfare History Project. http://www.socialwelfarehistory.com/social-work/is-social-work-a-profession-1915
Floridi L, Chiriatti M (2020) GPT-3: its nature, scope, limits, and consequences. Mind Mach 30(4):681–694. https://doi.org/10.1007/s11023-020-09548-1
Friedman B, Nissenbaum H (1996) Bias in computer systems. ACM Trans Inform Syst 14(3):330–347. https://doi.org/10.1145/230538.230561
Gardner J, Brooks C, Baker R (2019) Evaluating the fairness of predictive student models through slicing analysis. In: Proceedings of the 9th International Conference on Learning Analytics & Knowledge, pp 225–234. https://doi.org/10.1145/3303772.3303791
Goffman E (1977) The arrangement between the sexes. Theory Soc 4:301–331
Gramsci A (1971) Selections from the prison notebooks. Lawrence and Wishart, London
Hamburger ME, Hogben M, McGowan S, Dawson LJ (1996) Assessing hypergender ideologies: development and initial validation of a gender-neutral measure of adherence to extreme gender-role beliefs. J Res Pers 30(2):157–178. https://doi.org/10.1006/jrpe.1996.0011
Haraway DJ (1985) A cyborg manifesto: science, technology, and socialist-feminism in the late twentieth century. Posthumanism. https://doi.org/10.1007/978-1-137-05194-3_10
Hill Collins P (2019) Intersectionality as critical social theory. Duke University Press
Hinton PR (2017) Implicit stereotypes and the predictive brain: cognition and culture in “biased” person perception. Palgrave Commun 3(1):1–9. https://doi.org/10.1057/palcomms.2017.86
Hoffmann J, Borgeaud S, Mensch A, Buchatskaya E, Cai T, Rutherford E, Casas D de L, Hendricks LA, Welbl J, Clark A, Hennigan T, Noland E, Millican K, Driessche G van den, Damoc B, Guy A, Osindero S, Simonyan K, Elsen E, Rae JW, Vinyals O, Sifre L (2022) Training compute-optimal large language models. arXiv. http://arxiv.org/abs/2203.15556
Houli D, Radford ML, Singh V (2021) “COVID19 is_”: the perpetuation of coronavirus conspiracy theories via Google autocomplete. In: Proceedings of the Association for Information Science and Technology, vol 58, pp 218–229. https://doi.org/10.1002/pra2.450
Howard A, Isbell AH (2020) Diversity in AI: the invisible men and women. MIT Sloan Management Review. https://sloanreview.mit.edu/article/diversity-in-ai-the-invisible-men-and-women/
Jakesch M, Bhat, A, Buschek D, Zalmanson L, and Naaman M (2023) Co-writing with opinionated language models affects users’ views. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI ‘23), Hamburg, Germany. ACM, New York, NY, USA, p 15. https://doi.org/10.1145/3544548.3581196
Jones LK (2020) Twitter wants you to know that you’re still SOL if you get a death threat—unless you’re President Donald Trump. https://medium.com/@agua.carbonica/twitter-wants-you-to-know-that-youre-still-sol-if-you-get-a-death-threat-unless-you-re-a5cce316b706
Kendall S, Tannen D (2015) Discourse and gender. pp. 548–567. https://doi.org/10.1002/9780470753460.ch29
Koenig AM, Eagly AH (2014) Evidence for the social role theory of stereotype content: observations of groups’ roles shape stereotypes. J Pers Soc Psychol 107:371–392. https://doi.org/10.1037/a0037215
Kollmayer M, Schober B, Spiel C (2018) Gender stereotypes in education: development, consequences, and interventions. Eur J Dev Psychol 15(4):361–377. https://doi.org/10.1080/17405629.2016.1193483
Lapowsky I (2018) Google autocomplete suggestions are still racist, sexist, and science-denying. WIRED. https://wired.com/story/google-autocomplete-vile-suggestions/
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444. https://doi.org/10.1038/nature14539
Li L, Bamman D (2021) Gender and representation bias in GPT-3 generated stories. In: Proceedings of the Third Workshop on Narrative Understanding, pp 48–55. https://doi.org/10.18653/v1/2021.nuse-1.5
Liu Q, Kusner MJ, Blunsom P (2020) A survey on contextual embeddings. [Cs]. http://arxiv.org/abs/2003.07278
Maas JJC (2022) Machine learning and power relations. Ai & Soc 38:1493–1500
Machin D, Mayr A (2012) How to do critical discourse analysis: a multimodal approach. Sage
Magee L, Ghahremanlou L, Soldatic K, Robertson S (2021) Intersectional bias in causal language models. [Cs]. http://arxiv.org/abs/2107.07691
Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2019) A survey on bias and fairness in machine learning. ArXiv E-Prints. https://arxiv.org/abs/1908.09635
Mitchell PW (2018) The fault in his seeds: Lost notes to the case of bias in Samuel George Morton’s cranial race science. PLoS Biol 16(10):e2007008. https://doi.org/10.1371/journal.pbio.2007008
Moule J (2009) Understanding unconscious bias and unintentional racism. Phi Delta Kappan 90(5):320–326. https://doi.org/10.1177/003172170909000504
Murgia M (2019) AI academics under pressure to do commercial research. Financial Times. https://www.ft.com/content/94e86cd0-44b6-11e9-a965-23d669740bfb
Nadeem M, Bethke A, Reddy S (2020) Stereoset: measuring stereotypical bias in pretrained language models
Nash JC (2008) Re-thinking intersectionality. Fem Rev 89(1):1–15. https://doi.org/10.1057/fr.2008.4
Nelson A (2016) The social life of DNA: race, reparations, and reconciliation after the genome. Beacon Press
OpenAI Platform n.d. https://platform.openai.com. Accessed May 2022
O’Neill L, Anantharama N, Buntine W, Angus SD (2021) Quantitative discourse analysis at Scale—AI, NLP and the transformer revolution. In: SoDa Laboratories Working Paper Series (2021–12; SoDa Laboratories Working Paper Series). Monash University, SoDa Laboratories. https://ideas.repec.org/p/ajr/sodwps/2021-12.html
O’Sullivan L, Dickerson J (2020) Here are a few ways GPT-3 can go wrong. TechCrunch. https://social.techcrunch.com/2020/08/07/here-are-a-few-ways-gpt-3-can-go-wrong/
Pew (2016) Reddit news users more likely to be male, young and digital in their news preferences. Pew Research Center’s Journalism Project. https://www.pewresearch.org/journalism/2016/02/25/reddit-news-users-more-likely-to-be-male-young-and-digital-in-their-news-preferences/
Pew (2018) Internet/broadband fact sheet. https://www.pewinternet.org/fact-sheet/internet-broadband/
Powell-Hopson D, Hopson DS (1988) Implications of doll color preferences among black preschool children and white preschool children. J Black Psychol 14(2):57–63. https://doi.org/10.1177/00957984880142004
Salles A, Evers K, Farisco M (2020) Anthropomorphism in AI. AJOB Neurosci 11(2):88–95. https://doi.org/10.1080/21507740.2020.1740350
Santurkar S, Durmus E, Ladhak F, Lee C, Liang P, Hashimoto T (2023) Whose Opinions Do Language Models Reflect? (https://arXiv.org/2303.17548). arXiv. https://doi.org/10.48550/arXiv.2303.17548
Schüssler Fiorenza E (2009) Introduction: exploring the intersections of race, gender, status, and ethnicity in early Christian studies. In: Laura Nasrallah, Fiorenza (ed) Prejudice and Christian Beginnings: Investigating Race, Gender, and Ethnicity in Early Christian Studies. pp 1–23
Sengupta U (2021) Monoculturalism, aculturalism, and postculturalism: the exclusionary culture of algorithmic development. Algorithmic culture: how big data and artificial intelligence are transforming everyday life. pp 71–97
Silverman D (2020) Credible qualitative research. Interpreting qualitative data. Sage, pp 352–395
Smith CS (2022) OpenAI is giving Microsoft exclusive access to its GPT-3 language model|MIT Technology Review. https://www.technologyreview.com/2020/09/23/1008729/openai-is-giving-microsoft-exclusive-access-to-its-gpt-3-language-model/
Spencer SJ, Logel C, Davies P (2016) Stereotype threat. Ann Rev Psychol 67(1):415–437. https://doi.org/10.1146/annurev-psych-073115-103235
Staszak J (2009) Other/Otherness. In: Kitchin & Thrift (ed) International encyclopedia of hum geography: A 12- volume set, 1st edn. Oxford, Elsevier Science. https://archive-ouverte.unige.ch/unige:77582
van Sterkenburg J, Knoppers A, de Leeuw S (2012) Constructing racial/ethnic difference in and through Dutch televised soccer commentary. J Sport Soc Issues 36:422–442. https://doi.org/10.1177/0193723512448664
Veerman E (2016) “Welke pop vind je lelijk?” VPRO. https://www.vpro.nl/lees/gids/2016/51/-Welke-pop-vind-je-lelijk.html
Wajcman J (2010) Feminist theories of technology. Camb J Econ 34(1):143–152
West SM (2020) AI and the Far Right: A History We Can’t Ignore. Medium. https://medium.com/@AINowInstitute/ai-and-the-far-right-a-history-we-cant-ignore-f81375c3cc57
Whittaker M (2021) The steep cost of capture. SSRN Scholarly Paper No. 4135581. https://papers.ssrn.com/abstract=4135581
Wilson J (2017) People see black men as larger, more threatening than same-sized white men. https://www.apa.org. https://www.apa.org/news/press/releases/2017/03/black-men-threatening
Winner L (1980) “Do artifacts have politics?” Emerging technologies: ethics, law and governance, pp 15–30. https://doi.org/10.4324/9781003074960-3
World Bank (2018) Individuals using the internet https://data.worldbank.org/indicator/IT.NET.USER.ZS?end=2017&locations=US&start=2015