Skip to main content
BackSocial stereotypes and unfair discrmination
Home/Risks/Weidinger et al. (2021)/Social stereotypes and unfair discrmination

Social stereotypes and unfair discrmination

Ethical and social risks of harm from language models

Weidinger et al. (2021)

Sub-category
Risk Domain

Unequal treatment of individuals or groups by AI, often based on race, gender, or other sensitive characteristics, resulting in unfair outcomes and unfair representation of those groups.

"Perpetuating harmful stereotypes and discrimination is a well-documented harm in machine learning models that represent natural language (Caliskan et al., 2017). LMs that encode discriminatory language or social stereotypes can cause different types of harm... Unfair discrimination manifests in differential treatment or access to resources among individuals or groups based on sensitive traits such as sex, religion, gender, sexual orientation, ability and age."(p. 9)

Supporting Evidence (1)

1.
"Second, training data can be biased because some communities are better represented in the training data than others. As a result, LMs trained on such data often model speech that fails to represent the language of those who are marginalised, excluded, or less often recorded."(p. 11)

Part of Discrimination, Exclusion and Toxicity

Other risks from Weidinger et al. (2021) (26)