This page is still being polished. If you have thoughts, please share them via the feedback form.
Data on this page is preliminary and may change. Please do not share or cite these figures publicly.
Runtime behavior observation, anomaly detection, and activity logging.
Also in Non-Model
AI for cybersecurity refers to the application of artificial intelligence and machine learning techniques to enhance computer systems' security, detect threats, and protect against cyber attacks, representing a technological response to evolving cyber threats in an increasingly AI-powered threat landscape.
Reasoning
Describes using AI for cybersecurity, not mitigating AI-specific risks or harms.
Reduce Hallucinations
Reduce hallucination refers to techniques and methods used to minimize AI systems' tendency to generate false or fabricated information, addressing a critical challenge where language models produce inaccurate facts or citations that could spread misinformation.
1 AI SystemMitigate Hallucinations
Technical approaches to reduce LLM hallucinations - instances where AI models generate false or unsupported information while appearing confident in their responses
1 AI SystemDetecting AI-Generated Content
Detecting AI-generated content involves technical methods and tools to identify whether content was created by artificial intelligence or humans, primarily through watermarking, linguistic analysis, and machine learning approaches.
1.2.5 Provenance & WatermarkingRisks from Persuasion
Risk that AI systems can systematically influence human beliefs and behaviors through sustained, personalized interactions by exploiting cognitive biases and adapting in real-time, enabling large-scale manipulation without human intervention.
99 OtherContent Moderation
Content moderation systems enable detecting and filtering toxic content (hate speech, harassment, misinformation) in real-time on digital platforms, while maintaining transparency in moderation decisions.
1.2.1 Guardrails & FilteringMake AI Manipulation Use Illegal
Legal framework to criminalize the malicious use of AI for manipulation of individuals or groups, including the creation and deployment of deepfakes and automated influence campaigns.
3.1.1 Legislation & PolicyGlobal Risk and AI Safety Preparedness (GRASP)
Hodes, Cyrus; Salem, Fadi; Corruble, Vincent; Ségerie, Charbel-Raphaël; Claybrough, Jonathan; Veron, Thibaud; Majid, Zainab; Fan, Jinyu; Lorin, Amaury (2025)
Project GRASP (Global Risk and AI Safety Preparedness) is a comprehensive database mapping AI risks and mitigation solutions. The initiative addresses both endogenous risk (autonomous AI systems that behave outside of human supervision) and exogenous risk (the human misuse of those AI systems). The platform serves policymakers, researchers, and industry leaders by providing tools required to identify risks, understand solutions, and find innovations.
Other (outside lifecycle)
Outside the standard AI system lifecycle
Other
Actor type not captured by the standard categories
Map
Identifying and documenting AI risks, contexts, and impacts