Skip to main content
BackLeading users to perform unethical or illegal actions
Home/Risks/Weidinger et al. (2021)/Leading users to perform unethical or illegal actions

Leading users to perform unethical or illegal actions

Ethical and social risks of harm from language models

Weidinger et al. (2021)

Sub-category
Risk Domain

Users anthropomorphizing, trusting, or relying on AI systems, leading to emotional or material dependence and inappropriate relationships with or expectations of AI systems. Trust can be exploited by malicious actors (e.g., to harvest personal information or enable manipulation), or result in harm from inappropriate use of AI in critical situations (e.g., medical emergency). Overreliance on AI systems can compromise autonomy and weaken social ties.

"Where a LM prediction endorses unethical or harmful views or behaviours, it may motivate the user to perform harmful actions that they may otherwise not have performed. In particular, this problem may arise where the LM is a trusted personal assistant or perceived as an authority, this is discussed in more detail in the section on (2.5 Human-Computer Interaction Harms). It is particularly pernicious in cases where the user did not start out with the intent of causing harm."(p. 25)

Part of Misinformation Harms

Other risks from Weidinger et al. (2021) (26)