Summary of our gpt-oss methodology review
Details on external recommendations from METR for gpt-oss Preparedness experiments and follow-up from OpenAI.
Topic feed
Safety evaluations, red teaming, preparedness, and model risk testing.
Details on external recommendations from METR for gpt-oss Preparedness experiments and follow-up from OpenAI.
ChatGPT agent System Card: OpenAI’s agentic model unites research, browser automation, and code tools with safeguards under the Preparedness Framework.
Current views on information relevant for visibility into frontier AI risk.
Sharing our updated framework for measuring and protecting against severe harm from frontier AI capabilities.
We’re exploring the frontiers of AGI, prioritizing technical safety, proactive risk assessment, and collaboration with the AI community.
This report outlines the safety work carried out prior to releasing deep research including external red teaming, frontier risk evaluations according to our Preparedness Framework, and an overview of the mitigations we built in to address key risk areas.
This report outlines the safety work carried out for the OpenAI o3-mini model, including safety evaluations, external red teaming, and Preparedness Framework evaluations.
Why pre-deployment testing is not an adequate framework for AI risk management
This report outlines the safety work carried out prior to releasing OpenAI o1 and o1-mini, including external red teaming and frontier risk evaluations according to our Preparedness Framework.
Suggestions for expanded guidance on capability elicitation and robust model safeguards in the U.S. AI Safety Institute’s draft document “Managing Misuse Risk for Dual-Use Foundation Models” (NIST AI 800-1).
Comments on NIST’s draft document “AI Risk Management Framework: Generative AI Profile.”