Stanford Center for AI Safety

Advancing Safe & Trustworthy AI

Developing rigorous techniques for building AI systems that are safe, robust, and beneficial for society.






OUR MISSION

The Stanford Center for AI Safety leads research, education, and policy to ensure AI systems are safe, trustworthy, and beneficial for humanity.

OUR VISION

We envision a future where AI technologies are developed and deployed responsibly, with robust safety guarantees, transparent decision-making, and alignment with human values. Through interdisciplinary collaboration spanning computer science, engineering, law, and social sciences, we aim to shape the policies and practices that govern AI development worldwide.

RESEARCH AREAS

Formal Methods

Using precise mathematical modeling to ensure the safety, security, and robustness of conventional software and hardware systems.

Learning & Control

Designing systems that intelligently balance learning under uncertainty and acting safely in dynamic environments.

Transparency

Understanding safety in the context of fairness, accountability, and explainability for autonomous and intelligent systems.

AI Governance & Policy

Developing frameworks for responsible AI governance, working with policymakers and industry to create standards and best practices.

Human-AI Interaction

Studying how humans and AI systems can work together effectively and safely through trust calibration and human oversight mechanisms.

EXPLORE

Stay Connected

Subscribe to our newsletter for the latest updates on AI safety research, events, and opportunities.

Join Our Mailing List