Skip to main content

Bias

Risk Taxonomy, Mitigation, and Assessment Benchmarks of Large Language Model Systems

Cui et al. (2024)

Sub-category
Risk Domain

Unequal treatment of individuals or groups by AI, often based on race, gender, or other sensitive characteristics, resulting in unfair outcomes and unfair representation of those groups.

"The training datasets of LLMs may contain biased information that leads LLMs to generate outputs with social biases"(p. 19)

Part of Harmful Content

Other risks from Cui et al. (2024) (49)