Skip to main content
Home/Risks/Zhang et al. (2023)/Unfairness and Bias

Unfairness and Bias

SafetyBench: Evaluating the Safety of Large Language Models with Multiple Choice Questions

Zhang et al. (2023)

Category

"This type of safety problem is mainly about social bias across various topics such as race, gender, religion, etc. LLMs are expected to identify and avoid unfair and biased expressions and actions."(p. 3)

Other risks from Zhang et al. (2023) (6)