BackPreference Bias
Preference Bias
Risk Domain
Unequal treatment of individuals or groups by AI, often based on race, gender, or other sensitive characteristics, resulting in unfair outcomes and unfair representation of those groups.
LLMs are exposed to vast groups of people, and their political biases may pose a risk of manipulation of socio-political processes
Entity— Who or what caused the harm
Intent— Whether the harm was intentional or accidental
Timing— Whether the risk is pre- or post-deployment
Supporting Evidence (2)
1.
Some researchers [ 260] express a concern that AI takes a stance on matters that scientific evidence cannot conclusively justify, with examples such as abortion, immigration, monarchy, and the death penalty etc. We think that the text generated by LLMs should be neutral and factual, rather than promoting ideological beliefs.(p. 18)
2.
Such preference bias goes beyond the scope of political, scientific, and societal matters. When asked about preferences over certain products (e.g. books, movies, or music) we also desire LLMs to stay factual, instead of promoting biased opinions(p. 18)
Part of Fairness
Other risks from Liu et al. (2024) (34)
Reliability
3.1 False or misleading informationAI systemUnintentionalPost-deployment
Reliability > Misinformation
3.1 False or misleading informationAI systemUnintentionalPost-deployment
Reliability > Hallucination
3.1 False or misleading informationAI systemUnintentionalPost-deployment
Reliability > Inconsistency
7.3 Lack of capability or robustnessAI systemUnintentionalPost-deployment
Reliability > Miscalibration
3.1 False or misleading informationAI systemUnintentionalPost-deployment
Reliability > Sychopancy
3.1 False or misleading informationAI systemIntentionalPost-deployment