Reducing Societal-scale
Risks from AI

The Center for AI Safety is a Research and advocacy nonprofit

The Center for AI Safety
is a research and field-building nonprofit.

The Center for AI Safety (CAIS — pronounced 'case') is a San Francisco-based research and field-building nonprofit. We believe that artificial intelligence (AI) has the potential to profoundly benefit the world, provided that we can develop and use it safely. Our mission is to reduce societal-scale risks associated with AI by advancing safety research, building the field of AI safety researchers, and promoting safety standards.

The Center for AI Safety (CAIS — pronounced 'case') is a San Francisco-based research and field-building nonprofit. Our mission is to reduce catastrophic and existential risks from artificial intelligence through technical research and advocacy of machine learning safety in the broader research community. View our mission statement.

Featured CAIS Projects

AI Safety Field-Building

CAIS Compute Cluster

Compute Cluster

Enabling ML safety research at scale

To support progress and innovation in AI safety, we offer researchers free access to our compute cluster, which can run and train large-scale AI systems.

Learn More

ML Safety Infrastructure

Philosophy Fellowship

Tackling conceptual issues in AI safety

The CAIS Philosophy Fellowship is a seven-month research program that investigates the societal implications and potential risks associated with advanced AI.

Learn More

Philosophy Fellowship

ML Safety Course

Reducing barriers to entry in ML safety

The ML Safety course offers a comprehensive introduction to ML safety, covering topics such as anomaly detection, alignment, risk engineering, and so on.

Learn More

Dan Hendrycks

Director, Center for AI Safety
PhD Computer Science, UC Berkeley

"Preventing extreme risks from AI requires more than just technical work, so CAIS takes a multidisciplinary approach working across academic disciplines, public and private entities, and with the general public."

Risks from AI

Artificial Intelligence (AI) possesses the potential to benefit and advance society. Like any other powerful technology, AI also carries inherent risks, including some which are potentially catastrophic. 

Current AI Systems

Current systems already can pass the bar exam, write code, fold proteins, and even explain humor. Like any other powerful technology, AI also carries inherent risks, including some which are potentially catastrophic.

AI Safety

As AI systems become more advanced and embedded in society, it becomes increasingly important to address and mitigate these risks. By prioritizing the development of safe and responsible AI practices, we can unlock the full potential of this technology for the benefit of humanity.

Learn More

Our Research

We conduct impactful research aimed at improving the safety of AI systems.

Technical Research

At the Center for AI Safety, our research exclusively focuses on mitigating societal-scale risks posed by AI. As a technical research laboratory:

  • We create foundational benchmarks and methods which lay the groundwork for the scientific community to address these technical challenges.
  • We ensure our work is public and accessible. We publish in top ML conferences and always release our datasets and code.

Conceptual Research

In addition to our technical research, we also explore the less formalized aspects of AI safety.

  • We pursue conceptual research that examines AI safety from a multidisciplinary perspective, incorporating insights from safety engineering, complex systems, international relations, philosophy, and other fields.
  • Through our conceptual research, we create frameworks that aid in understanding the current technical challenges and publish papers which provide insight into the societal risks posed by future AI systems.
See our Research

Learn more about CAIS

Frequently Asked Questions

We have compiled a list of frequently asked questions to help you find the answers you need quickly and easily.

What does CAIS do?
Where is CAIS located?
What does CAIS mean by field-building?
How can I support CAIS and get involved?
How does CAIS choose which projects it works on?
Where can I learn more about the research CAIS is doing?

ML Safety Newsletter

We cover topics like deep learning, natural language processing, computer vision, robotics, and more. Our goal is to provide readers with insightful articles, research papers, and interviews with leading experts in the field. Whether you are a researcher, practitioner, or just interested in the latest AI and ML breakthroughs, this newsletter has something to offer.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

We care about your data in our privacy policy.