Skip to main content

Agency

Category
Risk Domain

AI systems that develop, access, or are provided with capabilities that increase their potential to cause mass harm through deception, weapons development and acquisition, persuasion and manipulation, political strategy, cyber-offense, AI development, situational awareness, and self-proliferation. These capabilities may cause mass harm due to malicious human actors, misaligned AI systems, or failure in the AI system.

"This section catalogs the risk sources and risk management measures related to agentic AI systems. We categorize these into the following groups: goal- directedness, deception, situational awareness, self-proliferation, and persuasion"(p. 28)

Supporting Evidence (1)

1.
"These risk items are related to behaviors associated with agentic systems using a GPAI as a base model or other component that is tasked with achieving an objective by manipulate its environment. They should not be confused with autonomy, where an AI system is free to take actions without full human supervision."

Other risks from Gipiškis2024 (144)