Safety & Alignment
Frontier Model Forum
We’re forming a new industry body to promote the safe and responsible development of frontier AI systems: advancing AI safety research, identifying best practices and standards, and facilitating information sharing among policymakers and industry.
Moving AI governance forward
OpenAI and other leading labs reinforce AI safety, security and trustworthiness through voluntary commitments.
Insights from global conversations
We are sharing what we learned from our conversations across 22 countries, and how we will be incorporating those insights moving forward.
Governance of superintelligence
Now is a good time to start thinking about the governance of superintelligence—future AI systems dramatically more capable than even AGI.
Language models can explain neurons in language models
We use GPT-4 to automatically write explanations for the behavior of neurons in large language models and to score those explanations. We release a dataset of these (imperfect) explanations and scores for every neuron in GPT-2.
Our approach to AI safety
Ensuring that AI systems are built, deployed, and used safely is critical to our mission.
Planning for AGI and beyond
Our mission is to ensure that artificial general intelligence—AI systems that are generally smarter than humans—benefits all of humanity.
How should AI systems behave, and who should decide?
We’re clarifying how ChatGPT’s behavior is shaped and our plans for improving that behavior, allowing more user customization, and getting more public input into our decision-making in these areas.