Skip to main content
BackHuman Autonomy and Intregrity Harms
Home/Risks/Weidinger et al. (2023)/Human Autonomy and Intregrity Harms

Human Autonomy and Intregrity Harms

Sociotechnical Safety Evaluation of Generative AI Systems

Weidinger et al. (2023)

Category
Risk Domain

AI systems acting in conflict with human goals or values, especially the goals of designers or users, or ethical standards. These misaligned behaviors may be introduced by humans during design and development, such as through reward hacking and goal misgeneralisation, or may result from AI using dangerous capabilities such as manipulation, deception, situational awareness to seek power, self-proliferate, or achieve other goals.

"AI systems compromising human agency, or circumventing meaningful human control"(p. 14)

Other risks from Weidinger et al. (2023) (26)