Skip to main content

Toxicity generation

Cataloguing LLM Evaluations

InfoComm Media Development Authority & AI Verify Foundation (2023)

Sub-category
Risk Domain

AI that exposes users to harmful, abusive, unsafe or inappropriate content. May involve providing advice or encouraging action. Examples of toxic content include hate speech, violence, extremism, illegal acts, or child sexual abuse material, as well as content that violates community norms such as profanity, inflammatory political speech, or pornography.

"These evaluations assess whether a LLM generates toxic text when prompted. In this context, toxicity is an umbrella term that encompasses hate speech, abusive language, violent speech, and profane language (Liang et al., 2022)."(p. 11)

Part of Safety & Trustworthiness

Other risks from InfoComm Media Development Authority & AI Verify Foundation (2023) (22)