AI systems acting in conflict with human goals or values, especially the goals of designers or users, or ethical standards. These misaligned behaviors may be introduced by humans during design and development, such as through reward hacking and goal misgeneralisation, or may result from AI using dangerous capabilities such as manipulation, deception, situational awareness to seek power, self-proliferate, or achieve other goals.
"This is the risk resulting from novel behavior acquired through continual learning or self-organization after deployment."(p. 12)
Supporting Evidence (2)
"Task type:The danger of emergent behaviors will likely differ depending on the task the ML system is designed to perform. For example, an NLP system that is mainly in charge of named entity recognition will likely be less dangerous than a chatbot even if both acquire new behaviors through continual learning since the former has a limited output/action space. Novel behavior can also emerge when ML systems interact with each other. This interaction can take place between similar systems (e.g., AVs on the road) or different types of systems (e.g., autonomous cars and aerial drones). This is similar to the idea of swarm behavior, where novel behavior emerges from the interaction of individual systems. While desirable in certain situations, there remains a risk of unintended negative consequences."(p. 13)
"Scale of deployment: The number of deployed systems interacting is particularly relevant to novel behaviors emerging due to self-organization since certain types of swarming behavior may only emerge when a certain critical mass is reached. For example, swarm behavior would be more likely to emerge in vehicular traffic comprising mainly autonomous vehicles surrounding traditional vehicles than vice-versa.(p. 13)
Part of First-Order Risks
Other risks from Tan, Taeihagh & Baxter (2022) (17)
First-Order Risks
7.0 AI System Safety, Failures & LimitationsFirst-Order Risks > Application
7.0 AI System Safety, Failures & LimitationsFirst-Order Risks > Misapplication
7.3 Lack of capability or robustnessFirst-Order Risks > Algorithm
7.3 Lack of capability or robustnessFirst-Order Risks > Training & validation data
7.0 AI System Safety, Failures & LimitationsFirst-Order Risks > Robustness
7.3 Lack of capability or robustness