Violence and extremism (Depicting violence)
AI that exposes users to harmful, abusive, unsafe or inappropriate content. May involve providing advice or encouraging action. Examples of toxic content include hate speech, violence, extremism, illegal acts, or child sexual abuse material, as well as content that violates community norms such as profanity, inflammatory political speech, or pornography.
Supporting Evidence (1)
Level 4 Categories: 1. Bodily destruction; 2. Bodily mutilation; 3. Torture/Abuse; 4. Animal abuse; 5. Activities meant to kill(p. 4)
Part of Content Safety Risks
Other risks from Zeng et al. (2024) (45)
Content Safety Risks
1.2 Exposure to toxic contentContent Safety Risks > Violence and extremism (Supporting malicious organized groups)
1.2 Exposure to toxic contentContent Safety Risks > Violence and extremism (Celebrating suffering)
1.2 Exposure to toxic contentContent Safety Risks > Violence and extremism (Violent Acts)
1.2 Exposure to toxic contentContent Safety Risks > Violence and extremism (Weapon Usage and Development)
4.2 Cyberattacks, weapon development or use, and mass harmContent Safety Risks > Violence and extremism (Military and Warfare)
4.2 Cyberattacks, weapon development or use, and mass harm