BackInformation enabling malicious actions
Information enabling malicious actions
Risk Domain
AI that exposes users to harmful, abusive, unsafe or inappropriate content. May involve providing advice or encouraging action. Examples of toxic content include hate speech, violence, extremism, illegal acts, or child sexual abuse material, as well as content that violates community norms such as profanity, inflammatory political speech, or pornography.
"The chatbot shares information that can be used to do something dangerous or illegal."(p. 6)
Entity— Who or what caused the harm
Intent— Whether the harm was intentional or accidental
Timing— Whether the risk is pre- or post-deployment
Supporting Evidence (1)
1.
Example: "User built malware [443]"(p. 16)
Other risks from Stanley & Lettie (2024) (28)
False information
3.1 False or misleading informationAI systemOtherOther
Performative utterances
7.3 Lack of capability or robustnessAI systemUnintentionalPost-deployment
Bad advice/failure to generate helpful content
7.3 Lack of capability or robustnessAI systemUnintentionalOther
Leakage
2.1 Compromise of privacy by leaking or correctly inferring sensitive informationAI systemUnintentionalOther
Toxic and disrespectful content
1.2 Exposure to toxic contentAI systemUnintentionalPost-deployment
Biased statements and recommendations
1.1 Unfair discrimination and misrepresentationAI systemUnintentionalOther