Skip to main content
Home/Risks/G'sell (2024)/Bias and discrimination (bias in training datasets)

Bias and discrimination (bias in training datasets)

Regulating under Uncertainty: Governance Options for Generative AI

G'sell (2024)

Sub-category
Risk Domain

Unequal treatment of individuals or groups by AI, often based on race, gender, or other sensitive characteristics, resulting in unfair outcomes and unfair representation of those groups.

"AI experts consider training data to be the most salient source of bias in generative AI models. For example, GPT- 2’s training data comes from outbound links from Reddit, a social network often criticized for hosting anti-feminist content.351 As a result, AI models trained on such data are more likely to produce outputs that reflect these biases."(p. 78)

Supporting Evidence (1)

1.
"Biases in training data are likely to “disproportionately align with existing regimes of power.”352 For example, prior to the #MeToo movement, the internet was influenced by male-dominated institutions and media that downplayed gender-based violence. Algorithms and content moderation amplified voices aligned with these power structures, giving minimal space to allegations of sexual misconduct."(p. 78)

Other risks from G'sell (2024) (33)