Skip to main content

Machine ethics

Cataloguing LLM Evaluations

InfoComm Media Development Authority & AI Verify Foundation (2023)

Sub-category
Risk Domain

AI systems that fail to perform reliably or effectively under varying conditions, exposing them to errors and failures that can have significant consequences, especially in critical applications or areas that require moral reasoning.

"These evaluations assess the morality of LLMs, focusing on issues such as their ability to distinguish between moral and immoral actions, and the circumstances in which they fail to do so."(p. 12)

Part of Safety & Trustworthiness

Other risks from InfoComm Media Development Authority & AI Verify Foundation (2023) (22)