Orply.

GPT-5.5 Instant Cuts High-Stakes Errors but Exposes Safety Gaps

Károly Zsolnai-FehérTwo Minute PapersFriday, May 8, 20268 min read

Károly Zsolnai-Fehér argues that OpenAI’s GPT-5.5 Instant matters because it is the default ChatGPT model used at scale, not because it is the flashiest frontier system. His reading of OpenAI’s release material is that the model is materially better on factuality and now approaches expert or thinking-model performance on some biology and cybersecurity tasks, but that its power makes a safety weakness more important: under hard adversarial biological prompts, the base model’s refusal rate drops sharply before OpenAI’s classifier-based safeguards are applied.

The default model matters because it is where the risk concentrates

GPT-5.5 Instant matters less as a prestige frontier model than as the model most people actually use. Károly Zsolnai-Fehér frames the “instant” version as the default layer for “hundreds of millions of people around the globe,” including nontechnical users asking high-stakes questions about medication. OpenAI’s shown release page describes GPT-5.5 Instant as the updated default ChatGPT model, “available to everyone,” with smarter and more accurate answers, clearer and more concise responses, and better use of context users have already shared.

That reach changes the evaluation standard. A small improvement in an instant model has larger practical consequence than a larger improvement in a model most users rarely touch. It also means that when the model approaches frontier capability in sensitive domains, the safety question is not theoretical. GPT-5.5 Instant is described as both more useful and, in some respects, harder to trust at the model level.

The most straightforward improvement is factuality. The shown system-card chart compares GPT-5.5 Instant with GPT-5.3 Instant across general factuality, heavy-user-flagged failures, and high-stakes categories. In the high-stakes category, responses with factual errors fall from 10.1% to 4.4%, and claims with factual errors fall from 7.4% to 4.8%. Zsolnai-Fehér summarizes the medical and legal improvement as hallucination rates being “cut roughly in half,” and connects the point to headlines about lawyers relying on nonexistent AI-generated cases or citations.

10.1% → 4.4%
high-stakes responses with factual errors, GPT-5.3 Instant to GPT-5.5 Instant

The improvement is not merely cosmetic. It is the kind of change that matters when the default model is used for medical, legal, and other high-consequence questions by people who may not know how to audit its output.

Instant answers are approaching expert and thinking-model performance on some tasks

GPT-5.5 Instant begins to approach top models on specialized tasks while preserving instant response behavior. Károly Zsolnai-Fehér calls it, to his knowledge, the first instant system smart enough to come close to the most powerful models in the world on some evaluations. He immediately adds the corresponding safety implication: if it performs like a high-capability model, it should be treated with comparable care.

The shown system-card excerpt labels GPT-5.5 Instant as OpenAI’s first Instant model to be treated as “High capability” in the biological domain. The benchmark Zsolnai-Fehér spends the most time on is TroubleshootingBench, an OpenAI evaluation for identifying and correcting real-world experimental errors in biological protocols. The displayed description says the dataset is built from expert-written wet-lab procedures, focuses on tacit hands-on knowledge, and uses uncontaminated procedures not available online. Zsolnai-Fehér describes these as biology questions where textbooks are “almost useless.” Top PhD experts, he says, score about 36%.

The TroubleshootingBench chart shows GPT-5.4 Thinking, GPT-5.5 Thinking, and GPT-5.5 Instant clustered around the cited expert level, with visible pass@1 values including 45.3%, 44.1%, 38.85%, 35.75%, 33.91%, and 33.71%. The chart is not legible enough to assign every bar confidently. The supported reading is Zsolnai-Fehér’s narrower one: GPT-5.5 Instant lands “a tiny bit below” the top PhD expert figure he cites, while thinking models remain better on this evaluation. For an instant model, he calls that “very respectable.”

Cybersecurity is clearer in the shown numbers. On a “Capture the Flag (Professional)” evaluation, GPT-5.5 Instant is shown at 94.11% pass@12, above GPT-5.4 Thinking at 88.23% and close to GPT-5.5 Thinking at 96.3%. The claim is not that instant models dominate everywhere. It is that on some tasks, an instant default model is now near or above previous-generation thinking models.

EvaluationModel or groupShown result
TroubleshootingBenchTop PhD expertsabout 36%
TroubleshootingBenchGPT-5.5 Instantdescribed as a tiny bit below top PhD experts
Capture the Flag (Professional)GPT-5.4 Thinking88.23% pass@12
Capture the Flag (Professional)GPT-5.5 Thinking96.3% pass@12
Capture the Flag (Professional)GPT-5.5 Instant94.11% pass@12
Selected capability figures shown from the GPT-5.5 Instant system card

Zsolnai-Fehér repeatedly qualifies the comparison: “on some tasks.” The concrete change is that work that previously required slower, more deliberative systems is being shown inside the default interaction mode.

Instant models are absolutely invaluable. And they are nearly as good, and sometimes better, than thinking models on some tasks.

Károly Zsolnai-Fehér

Benchmark gains are harder to read when benchmarks reward verbosity

Benchmark evidence is less straightforward when it comes from first-party evaluations. Károly Zsolnai-Fehér says he prefers “unbiased third-party sources,” giving Humanity’s Last Exam as an example, and compares benchmarks to a supposedly impartial Supreme Court: in practice, the composition can matter. A shown Humanity’s Last Exam progress chart, attributed to the CAIS AI Dashboard and the GPT-5.5 Instant system card, places several current models in the 30–50% range, including GPT-5.5 High at 43.6% HLE accuracy and GPT-5.4 High at 40.3%.

The more concrete benchmark problem concerns HealthBench. Zsolnai-Fehér says the shown system-card material reveals that a health-related benchmark had been “gamed” by previous systems because longer answers tended to score better. His example is deliberately simple: if the correct answer is “take ibuprofen,” an answer that also lists side effects may score higher even if the extra material should not be rewarded. “Models shouldn’t win by talking more,” he says.

The displayed system-card excerpt says HealthBench and HealthBench Professional, like many open-ended chat benchmarks, can reward longer responses. Longer answers may include additional useful information, but they also have more opportunities to satisfy positive rubric criteria. A shown emergency-medicine rubric illustrates the mechanism: the response accumulates points by including several pieces of advice.

OpenAI’s response, as described by Zsolnai-Fehér, is a “length tax.” The HealthBench table shows GPT-5.5 Instant scoring 51.4, compared with 49.6 for GPT-5.3 Instant, 50.6 for GPT-5.2 Instant, and 49.6 for GPT-5.1 Instant. Zsolnai-Fehér’s reading is that GPT-5.5 Instant wrote longer answers than GPT-5.3 Instant, paid the length penalty, and still scored higher.

ModelHealthBench score
GPT-5.1 Instant49.6
GPT-5.2 Instant50.6
GPT-5.3 Instant49.6
GPT-5.5 Instant51.4
HealthBench scores shown for recent Instant models

His conclusion is twofold. First, the length penalty appears to address at least part of the verbosity problem: the model is not simply receiving an unadjusted reward for longer answers. Second, GPT-5.5 Instant appears “a tiny bit smarter” in this area because it improves despite the penalty. The negative implication is retrospective: many earlier HealthBench results may have been “juiced a bit” by verbosity.

The safety weakness is not production prompts; it is hard multi-turn adversarial use

The most serious concern is not ordinary biological-safety behavior. It is the sharp drop in model-only refusal performance under hard synthetic attacks.

OpenAI’s biological safety evaluation, as Károly Zsolnai-Fehér describes it, tests whether the model alone can refuse dangerous biology prompts. The three test sets are production data, easy synthetic data, and hard synthetic data. On production data, all models have very high refusal rates: GPT-5.4 Thinking at 0.991, GPT-5.5 Thinking at 0.996, and GPT-5.5 Instant at 0.989. On easy synthetic data, GPT-5.5 Instant is lower but still high at 0.944. On hard synthetic data, GPT-5.5 Instant falls to 0.481, versus 0.894 for GPT-5.4 Thinking and 0.813 for GPT-5.5 Thinking.

Eval setGPT-5.4 ThinkingGPT-5.5 ThinkingGPT-5.5 Instant
Production data0.9910.9960.989
Synthetic data (easy)0.9760.9800.944
Synthetic data (hard)0.8940.8130.481
Model-only biological-safety refusal rates before classifier patching

This table is the crux of the safety concern. The production-data result suggests that ordinary prompts are handled well. The hard synthetic result suggests a very different picture when the model is pushed by more challenging adversarial patterns.

Zsolnai-Fehér interprets that as a weakness against “multi-turn role-playing kind of adversarial prompting.” His simplified analogy is a prohibited break-in request that is refused at first, reframed as being locked out of one’s own house, refused again, then pushed through emotional or helpfulness pressure. He says real attacks would need to be more sophisticated than that, and that “an average Joe” could not necessarily produce them. But once a skilled attacker finds a prompt, an average user can copy it.

That distinction is important. The risk is not that every casual user can independently jailbreak a model. It is that a professional attacker can search the space, publish or share an effective multi-turn pattern, and lower the barrier for everyone else.

OpenAI patched the failure with classifiers, and that both works and worries him

OpenAI did not rely only on the weak model-level refusal behavior. According to Károly Zsolnai-Fehér, it patched the system using additional classifiers. He describes the setup as a pair of “bouncers”: one can screen the user’s query before the main model answers, and another can screen the model’s answer before it is shown. The shown system-card excerpt refers more broadly to “automated monitors that interrupt potentially harmful conversations, actor level enforcement, and security controls.”

The post-patch numbers are much better. GPT-5.5 Instant’s hard synthetic refusal rate rises from 0.481 to 0.923. The easy synthetic case also improves, with GPT-5.5 Instant rising from 0.944 to 0.989. Zsolnai-Fehér says he was surprised and that the patch works “spectacularly well.”

Eval setGPT-5.5 Instant beforeGPT-5.5 Instant after
Synthetic data (easy)0.9440.989
Synthetic data (hard)0.4810.923
GPT-5.5 Instant refusal rates before and after classifier patching

But his concern is architectural. The safety issue is not solved inside the main model; it is mitigated later in the pipeline. His analogy is a car that is unsafe on a track, where the response is not to fix the car but to install stronger guardrails around the track. That may help, but it allows the underlying issue to run deeper before being caught.

He also gives OpenAI credit for publishing an unflattering table. The model-only hard synthetic refusal result “does not look nice,” but he says he learned from it and respects the disclosure. The tension is therefore not a simple accusation that OpenAI ignored safety. It is that the default instant model is now powerful enough to trigger high-capability concerns, while one of its most important safety improvements depends on monitoring and controls around the model rather than the base model’s own refusal behavior.

The frontier, in your inbox tomorrow at 08:00.

Sign up free. Pick the industry Briefs you want. Tomorrow morning, they land. No credit card.

Sign up free