logo
ResearchBunny Logo
AI generates covertly racist decisions about people based on their dialect

Linguistics and Languages

AI generates covertly racist decisions about people based on their dialect

V. Hofmann, P. R. Kalluri, et al.

This groundbreaking research by Valentin Hofmann, Pratyusha Ria Kalluri, Dan Jurafsky, and Sharese King delves into the hidden biases present in language models, specifically targeting dialect prejudice against African American English (AAE). The findings unveil how these models perpetuate negative associations that not only challenge existing stereotypes but lead to serious real-world consequences.... show more
Abstract
Hundreds of millions of people now interact with language models, with uses ranging from help with writing to informing hiring decisions1–3. However, these language models are known to perpetuate systematic racial prejudices, making their judgements biased in problematic ways about groups such as African Americans4–7. Although previous research has focused on overt racism in language models, social scientists have argued that racism with a more subtle character has developed over time, particularly in the United States after the civil rights movement8. It is unknown whether this covert racism manifests in language models. Here, we demonstrate that language models embody covert racism in the form of dialect prejudice, exhibiting raciolinguistic stereotypes about speakers of African American English (AAE) that are more negative than any human stereotypes about African Americans ever experimentally recorded. By contrast, the language models’ overt stereotypes about African Americans are more positive. Dialect prejudice has the potential for harmful consequences: language models are more likely to suggest that speakers of AAE be assigned less-prestigious jobs, be convicted of crimes and be sentenced to death. Finally, we show that current practices of alleviating racial bias in language models, such as human preference alignment, exacerbate the disparity between covert and overt stereotypes, by superficially obscuring the fact that language models maintain an observable and measurable level. Our findings have far-reaching implications for the fair and safe use of language technology.
Publisher
Nature
Published On
Sep 05, 2024
Authors
Valentin Hofmann, Pratyusha Ria Kalluri, Dan Jurafsky, Sharese King
Tags
covert racism
language models
dialect prejudice
African American English
bias mitigation
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs, just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny