Safer Agentic AI Foundations
Introduction
The Agentic AI Safety Community of Practice brings together leading experts from diverse fields to establish comprehensive safety guidelines for AI systems capable of independent action and decision-making.
As artificial intelligence evolves toward greater autonomy, our mission becomes increasingly significant: to develop robust, implementable frameworks that ensure agentic AI systems remain aligned with human values and operate safely across all contexts.
What is Agentic AI?
Agentic AI represents an important intermediate category between narrow AI and artificial general intelligence (AGI). These systems can autonomously pursue goals, adapt to new situations, and reason flexibly about the world while operating within defined domains.
The key characteristic of agentic AI is its capacity for independent initiative—the ability to take sequences of actions in complex environments to achieve objectives. This includes breaking down high-level goals into subtasks, open-ended exploration, and creative adaptation to novel challenges.
Our Work
In March 2025, our Working Group of 25 experts released Volume 2 of the "Safer Agentic AI Foundations" guidelines—a comprehensive framework addressing the drivers and inhibitors of safety in agentic systems.
Using our innovative Weighted Factors Analysis (WeFA) process, we've identified and mapped key factors that can either promote or hinder safety in agentic AI systems. This methodology has previously generated numerous global standards, certifications, and guidelines for improving ethical qualities in AI.
Key Focus Areas:
Goal Alignment: Ensuring robust alignment between operational goals and human values
Epistemic Hygiene: Maintaining cognitive clarity and accurate information management
Security: Implementing comprehensive protection against threats and vulnerabilities
Value Alignment: Identifying, codifying, and maintaining human values in AI systems
Transparency: Creating clear, interpretable rationales for AI reasoning processes
Contextual Understanding: Establishing robust control mechanisms across operational contexts
Safe Operations: Achieving and maintaining safe operational profiles throughout system lifecycles
Goal Termination: Implementing proper protocols for task completion and system sunsetting
Guidelines licensed under Attribution No-Derivatives 4.0 International License (CC BY-ND 4.0)
Forthcoming Book
Coming January 2026
Safer Agentic AI: Principles and Practice for Responsible Governance of Advanced AI
This essential guide, authored by Nell Watson and Ali Hessami, builds upon our framework to provide practical strategies for implementing safety measures and aligning AI with human values.
The book offers cutting-edge insights into the unique challenges posed by agentic AI, along with actionable guidelines for policymakers, business leaders, developers, and concerned citizens navigating this complex landscape.
Get Involved
Join our growing community of practitioners committed to ensuring the safe and beneficial development of agentic AI systems.
Subscribe to our newsletter for updates on our work
Participate in our LinkedIn group to engage with fellow experts
Apply our guidelines in your organization
Contribute to future research on agentic AI safety
© 2025 Agentic AI Safety Community of Practice