Multiple AI incidents were downgraded from the AI Incident Database because they involved academic findings, research demonstrations, or projected harms rather than documented real-world harm events.
The AI Incident Database downgraded six former incidents that no longer met updated incident criteria. These included: the 2016 Winograd Schema Challenge showing AI systems performed only 3% better than random chance; Janelle Shane's humorous AI-generated Christmas carols using neural networks trained on 240 popular carols; Tencent Keen Security Lab's research identifying Tesla Autopilot vulnerabilities to adversarial attacks and wireless gamepad control; French healthcare company Nabla's findings that GPT-3 was inconsistent and risky for medical applications, including telling a mock patient to kill themselves; TheFaceTag facial recognition app developed by a Harvard student raising ethical concerns; and GPT-3 generating threatening content about destroying humankind in a Guardian op-ed. All incidents were downgraded because they represented academic findings, research demonstrations, projected risks, or unclear harms rather than documented real-world incidents causing actual harm to people.
Domain classification, causal taxonomy, severity scores, and national security assessments were LLM-classified and may contain errors.
AI systems that fail to perform reliably or effectively under varying conditions, exposing them to errors and failures that can have significant consequences, especially in critical applications or areas that require moral reasoning.
Other
Due to some other reason or is ambiguous
Other
Without clearly specifying the intentionality
Other
Without a clearly specified time of occurrence
No population impact data reported.