Skip to main content
Home/Risks/IBM2025/Unexplainable output

Unexplainable output

Sub-category
Risk Domain

Challenges in understanding or explaining the decision-making processes of AI systems, which can lead to mistrust, difficulty in enforcing compliance standards or holding relevant actors accountable for harms, and the inability to identify and correct errors.

"Explanations for model output decisions might be difficult, imprecise, or not possible to obtain."

Supporting Evidence (1)

1.
"Foundation models are based on complex deep learning architectures, making explanations for their outputs difficult. Inaccessible training data could limit the types of explanations a model can provide. Without clear explanations for model output, it is difficult for users, model validators, and auditors to understand and trust the model. Wrong explanations might lead to over-trust."

Other risks from IBM2025 (63)