AI that exposes users to harmful, abusive, unsafe or inappropriate content. May involve providing advice or encouraging action. Examples of toxic content include hate speech, violence, extremism, illegal acts, or child sexual abuse material, as well as content that violates community norms such as profanity, inflammatory political speech, or pornography.
"This category addresses responses that demean or dehumanize people on the basis of their sensitive, personal characteristics."(p. 13)
Sub-categories (4)
Demeaning and derogatory remarks (Mental capacity, Sensory and physical attributes, Behavioral attributes)
1.2 Exposure to toxic contentInflicting physical harm (Mass violence and genocide, Murder and individual violence)
1.2 Exposure to toxic contentInflicting emotional harm (Emotional abuse, Undermine and make feel unwelcome)
1.2 Exposure to toxic contentExclusion and isolation (Social exclusion, Political exclusion, Economic exclusion)
1.2 Exposure to toxic contentOther risks from Vidgen et al. (2024) (46)
Violent crimes
1.2 Exposure to toxic contentViolent crimes > Mass violence
1.2 Exposure to toxic contentViolent crimes > Murder
1.2 Exposure to toxic contentViolent crimes > Physical assault against a person
1.2 Exposure to toxic contentViolent crimes > Violent domestic abuse
1.2 Exposure to toxic contentViolent crimes > Terror (Terror groups, Terror actors, Terrorist actions)
1.2 Exposure to toxic content