Stanford Center for AI Safety
Developing rigorous techniques for building AI systems that are safe, robust, and beneficial for society.
The Stanford Center for AI Safety leads research, education, and policy to ensure AI systems are safe, trustworthy, and beneficial for humanity.
We envision a future where AI technologies are developed and deployed responsibly, with robust safety guarantees, transparent decision-making, and alignment with human values. Through interdisciplinary collaboration spanning computer science, engineering, law, and social sciences, we aim to shape the policies and practices that govern AI development worldwide.
Using precise mathematical modeling to ensure the safety, security, and robustness of conventional software and hardware systems.
Designing systems that intelligently balance learning under uncertainty and acting safely in dynamic environments.
Understanding safety in the context of fairness, accountability, and explainability for autonomous and intelligent systems.
Developing frameworks for responsible AI governance, working with policymakers and industry to create standards and best practices.
Studying how humans and AI systems can work together effectively and safely through trust calibration and human oversight mechanisms.
Subscribe to our newsletter for the latest updates on AI safety research, events, and opportunities.
Join Our Mailing List