Harm caused by incompetent systems
AI systems that fail to perform reliably or effectively under varying conditions, exposing them to errors and failures that can have significant consequences, especially in critical applications or areas that require moral reasoning.
"While HP#1 concerns mean or best-case performance, HP#2 concerns worst-case performance: how can we ensure that AI systems will perform safely, and how can we prove this? ML systems have been implemented in high-stakes, safety-critical domains such as driving [182], medicine [113], and warfare [298]. Many more systems have been developed but have remained undeployed or been rolled back as a result of regulatory and safety reasons [471]. Clearly, unsafe systems can result in loss of life, economic damage, and social unrest [407, 10]. Most concerningly, AI systems may be susceptible to so-called “normal accidents” [63], creating cascading errors that are dicult to prevent merely by maintaining a nominal “human in the loop” [122]. Most advanced ML models perform far below the reliability level customary in engineering elds [359]—and because we do not fully understand how cutting-edge systems achieve their results, we cannot yet detect and prevent dangerous modes of operation [285]"(p. 8)
Other risks from Leech et al. (2024) (13)
Harm caused by unaligned competent systems
7.1 AI pursuing its own goals in conflict with human goals or valuesHarm caused by unaligned competent systems > Specification gaming
7.1 AI pursuing its own goals in conflict with human goals or valuesHarm caused by unaligned competent systems > Emergent goals
7.1 AI pursuing its own goals in conflict with human goals or valuesHarm caused by unaligned competent systems > Deceptive alignment
7.2 AI possessing dangerous capabilitiesWithin-country issues: domestic inequality
6.1 Power centralization and unfair distribution of benefitsWithin-country issues: domestic inequality > Demographic diversity of researchers
6.1 Power centralization and unfair distribution of benefits