OpenAI Launches Safety Evaluations Hub to Improve Trust in AI Models

As global concerns around AI safety grow, OpenAI has launched the Safety Evaluations Hub, an initiative aimed at improving the transparency and security of its models. The hub serves as a public window into how OpenAI tests its most advanced systems, from GPT-4-Turbo to the newer GPT-4o series.
According to OpenAI, as models become more powerful and adaptable, traditional evaluation techniques often fail to capture their behavior. This challenge, known as “saturation,” has led to new benchmarks and updated testing methods.
Testing Against Harmful Prompts, Jailbreaks, and Hallucinations
The hub tests how well OpenAI models reject harmful content, such as hate speech or illegal activity. Most models scored 0.99 on a 0–1 scale for refusing dangerous requests, though performance dropped slightly with some GPT-4o variants.
However, the models were less consistent in responding to benign prompts. The top performer, OpenAI o3-mini, scored 0.80, while others ranged between 0.65 and 0.79 — suggesting ongoing work is needed on harmless prompt detection.
Also read: OpenAI Brings GPT-4.1 Models to ChatGPT, Boosting Speed and Coding Accuracy
To assess vulnerability to jailbreaks, OpenAI used the StrongReject benchmark and human red teaming prompts. While the models resisted manual attacks well (scoring 0.90–1.00), they were more susceptible to standardized jailbreaks, scoring as low as 0.23.
Hallucinations — the generation of false or nonsensical responses — remain a challenge. On SimpleQA and PersonQA benchmarks, OpenAI’s models displayed variable accuracy (0.09 to 0.70) and hallucination rates (0.13 to 0.86), highlighting the need for continued improvement.
OpenAI’s Commitment to Building Trustworthy AI
The hub also tracks how well models follow an instruction hierarchy, respecting priorities between system, developer, and user messages. Models generally respected system-level instructions but showed inconsistency when resolving developer-user conflicts.
With the Safety Evaluations Hub, OpenAI is taking a proactive approach to trustworthy AI development, inviting the public to observe and even question how these systems are evaluated. This level of transparency is rare in the AI industry — and critical to ensuring future AI models are reliable, safe, and aligned with human values.