Organizations
Research labs, nonprofits, and companies working on AI alignment.
Anthropic
AI safety company focused on building reliable, interpretable, and steerable AI systems.
OpenAI
AI research company developing advanced AI systems including GPT and ChatGPT.
DeepMind
Google-owned AI research laboratory working on general-purpose AI.
MIRI
Machine Intelligence Research Institute, focused on foundational AI alignment research.
ARC
Alignment Research Center, focused on theoretical and empirical alignment research.
Center for AI Safety
Nonprofit focused on reducing societal-scale risks from AI through research and advocacy.
Redwood Research
AI safety research lab focused on interpretability and alignment techniques.
METR
Model Evaluation and Threat Research, focused on evaluating dangerous AI capabilities.
Conjecture
AI safety company developing alignment techniques and safe AI systems.
Future of Humanity Institute
Oxford-based research center focused on existential risks including AI.
CHAI
Center for Human-Compatible AI at UC Berkeley, led by Stuart Russell.
EleutherAI
Grassroots collective focused on open-source AI research and safety.