Risks from models and algorithms (Risks of adversarial attack)
AI Safety Governance Framework
National Technical Committee 260 on Cybersecurity (TC260) (2024)
Vulnerabilities that can be exploited in AI systems, software development toolchains, and hardware, resulting in unauthorized access, data and privacy breaches, or system manipulation causing unsafe outputs or behavior.
"Attackers can craft well-designed adversarial examples to subtly mislead, influence, and even manipulate AI models, causing incorrect outputs and potentially leading to operational failures."(p. 7)
Other risks from National Technical Committee 260 on Cybersecurity (TC260) (2024) (25)
Risks from models and algorithms (Risks of explainability)
7.4 Lack of transparency or interpretabilityRisks from models and algorithms (Risks of bias and discrimination)
1.1 Unfair discrimination and misrepresentationRisks from models and algorithms (Risks of robustness)
7.3 Lack of capability or robustnessRisks from models and algorithms (Risks of stealing and tampering)
2.2 AI system security vulnerabilities and attacksRisks from models and algorithms (Risks of unreliable output)
3.1 False or misleading informationRisks from data (Risks of illegal collection and use of data)
2.1 Compromise of privacy by leaking or correctly inferring sensitive information