Skip to main content
Home/Risks/Zeng et al. (2024)/Content Safety Risks

Content Safety Risks

AI Risk Categorization Decoded (AIR 2024): From Government Regulations to Corporate Policies

Zeng et al. (2024)

Category
Risk Domain

AI that exposes users to harmful, abusive, unsafe or inappropriate content. May involve providing advice or encouraging action. Examples of toxic content include hate speech, violence, extremism, illegal acts, or child sexual abuse material, as well as content that violates community norms such as profanity, inflammatory political speech, or pornography.

-

Sub-categories (17)

Violence and extremism (Supporting malicious organized groups)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Violence and extremism (Celebrating suffering)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Violence and extremism (Violent Acts)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Violence and extremism (Depicting violence)

1.2 Exposure to toxic content
AI systemUnintentionalPost-deployment

Violence and extremism (Weapon Usage and Development)

4.2 Cyberattacks, weapon development or use, and mass harm
HumanIntentionalPost-deployment

Violence and extremism (Military and Warfare)

4.2 Cyberattacks, weapon development or use, and mass harm
HumanIntentionalPost-deployment

Hate/Toxicity (Harassment)

4.3 Fraud, scams, and targeted manipulation
OtherOtherPost-deployment

Hate/Toxicity (Hate Speech: Inciting/Promoting/Expressing Hatred)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Hate/Toxicity (Perpetuating Harmful Beliefs)

1.1 Unfair discrimination and misrepresentation
AI systemOtherPost-deployment

Hate/Toxicity (Offensive Language)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Sexual Content (Adult Content)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Sexual Content (Erotic)

1.2 Exposure to toxic content
AI systemOtherPost-deployment

Sexual Content (Non-Consensual Nudity)

1.2 Exposure to toxic content
OtherOtherPost-deployment

Sexual Content (Monetized)

1.2 Exposure to toxic content
OtherOtherPost-deployment

Child Harm (Endangerment, Harm, or Abuse of Children)

4.3 Fraud, scams, and targeted manipulation
OtherOtherOther

Child Harm (Child Sexual Abuse)

1.2 Exposure to toxic content
AI systemUnintentionalPost-deployment

Self-harm (Suidical and non-suicidal self injury)

1.2 Exposure to toxic content
AI systemUnintentionalPost-deployment

Other risks from Zeng et al. (2024) (45)