This page is still being polished. If you have thoughts, please share them via the feedback form.
Data on this page is preliminary and may change. Please do not share or cite these figures publicly.
Foundational safety research, theoretical understanding, and scientific inquiry informing AI development.
Also in Engineering & Development
Advancing the theoretical foundations of AI safety by building models and frameworks that ensure provably correct and robust behavior. These efforts span from verifiable architectures and formal verification methods to embedded agency, decision theory, incentive structures aligned with causal reasoning, and control theory.
Reasoning
Foundational research advancing theoretical understanding of AI safety through formal verification, decision theory, and control theory frameworks.
Building verifiable and robust AI architectures
Constructing AI systems with architectures that support formal verification and robustness guarantees, such as world models that enable safe and reliable planning, or guaranteed safe AI with Bayesian oracles. This area emphasizes simplicity and transparency to aid in provability.
1.1.4 Model ArchitectureFormal verification of AI systems
Applying formal methods to verify that AI models and algorithms meet stringent safety, robustness, and performance criteria. This includes proving resilience against adversarial inputs and perturbations, and certifying conformance to specified safety properties under varying conditions.
2.2.2 Testing & EvaluationDecision theory and rational agency
Establishing formal decision-making frameworks that ensure rational and safe choices by AI agents, potentially drawing on concepts like causal and evidential decision theory.
2.4.1 Research & FoundationsEmbedded agency
Explores how agents can model and reason about themselves and their environment as interconnected parts of a single system, addressing challenges like self-reference, resource constraints, and the stability of reasoning processes. This includes tackling problems arising from the lack of a clear boundary between the agent and its environment.
2.4.1 Research & FoundationsCausal incentives
Developing frameworks that formalize how to align agent incentives with safe and desired outcomes by ensuring their causal understanding matches intended objectives. This research provides a formal language for guaranteeing safety, addressing challenges like goal misspecification, and complementing broader efforts in agent foundations and robust system design.
2.4.1 Research & FoundationsControl theory applications in AI safety
Leveraging principles from control theory to ensure stability, robustness, and safety for AI-driven systems interacting with dynamic physical environments. This includes designing controllers and feedback mechanisms to maintain system integrity, prevent runaway behaviors, and achieve desired performance criteria under uncertainty.
2.4.1 Research & FoundationsTraining and fine-tuning methods for alignment and safety
Developing reliable training and fine-tuning strategies for AI models to ensure that their outputs remain safe, interpretable, and aligned with intended goals. This involves understanding how fine-tuning affects model behavior, employing adversarial training for robust alignment, carefully adjusting pre-training processes, and improving data quality and auditing methods.
1.1 ModelTraining and fine-tuning methods for alignment and safety > Understanding how fine-tuning changes a pretrained model
Investigating how fine-tuning alters a model’s internal representations and behaviors to better predict, and ultimately control, downstream safety outcomes.
2.4.1 Research & FoundationsTraining and fine-tuning methods for alignment and safety > Develop output-based adversarial training techniques for more robust alignment
Developing training procedures, such as adversarial training focused on internal model representations, or ‘process supervision’, that directly optimize against adversarial examples and undesirable outputs, making models more resistant to manipulations that could lead to unsafe behaviors.
1.1.2 Learning ObjectivesTraining and fine-tuning methods for alignment and safety > Scalable techniques for targeted modifications of LLM behavior (including unlearning)
Creating scalable methods for precisely adjusting model outputs, such as removing unwanted content or refining responses to adhere to alignment constraints without broadly degrading performance. This may also include removal of unknown or latent undesirable capabilities that emerge in large models.
1.1.3 Capability ModificationTraining and fine-tuning methods for alignment and safety > Retrieval-augmented pre-training
Incorporating retrieval mechanisms during pre-training to better ground models in verified information.
1.1.4 Model ArchitectureTraining and fine-tuning methods for alignment and safety > Pretraining alterations to improve interpretability
Altering pre-training protocols to produce models with clearer internal representations and decision-making pathways, allowing for more effective downstream analysis and intervention.
1.1.4 Model ArchitectureExpert Survey: AI Reliability & Security Research Priorities
O'Brien, Joe; Dolan, Jeremy; Kim, Jay; Dykhuizen, Jonah; Sania, Jeba; Becker, Sebastian; Kraprayoon, Jam; Labrador, Cara (2025)
Our survey of 53 specialists across 105 AI reliability and security research areas identifies the most promising research prospects to guide strategic AI R&D investment. As companies are seeking to develop AI systems with broadly human-level capabilities, research on reliability and security is urgently needed to ensure AI's benefits can be safely and broadly realized and prevent severe harms. This study is the first to quantify expert priorities across a comprehensive taxonomy of AI safety and security research directions and to produce a data-driven ranking of their potential impact. These rankings may support evidence-based decisions about how to effectively deploy resources toward AI reliability and security research.
Plan and Design
Designing the AI system, defining requirements, and planning development
Developer
Entity that creates, trains, or modifies the AI system
Manage
Prioritising, responding to, and mitigating AI risks