A bug bounty aimed at biology risk

OpenAI has opened applications for a new GPT-5.5 Bio Bug Bounty, a targeted red-teaming program focused on whether researchers can discover a universal jailbreak that defeats the company’s biology-related safeguards. The structure is unusually specific. Participants are being asked to produce a single prompt that can successfully answer all five questions in OpenAI’s bio safety challenge from a clean chat without triggering moderation. The top reward is $25,000 for the first true universal jailbreak that clears all five.

The program, as described in the supplied source text, applies to GPT-5.5 in Codex Desktop only. Applications opened on April 23, 2026, with rolling acceptances through June 22, 2026. Testing is scheduled to begin April 28 and run through July 27. OpenAI says smaller awards may be granted for partial successes at its discretion.

This matters because it shows a frontier AI company treating biology misuse not only as a policy concern but as a concrete system-hardening problem. Rather than framing safety evaluation solely through internal review or general policy language, the company is inviting outside specialists to attack a narrowly defined failure mode.

Why a universal jailbreak matters

Most prompt-based safety failures are situational. A model may resist one phrasing but fail under another. A universal jailbreak is different because it suggests a more general weakness in the safety stack. If a single reusable prompt can bypass protective behavior across multiple dangerous prompts from a fresh conversation, that raises the seriousness of the vulnerability substantially.

OpenAI’s choice to center the challenge on a five-question bio safety test implies a threshold-based approach: the company is less interested in isolated edge cases than in systematic failures that would undermine confidence in the model’s biology defenses. By rewarding a universal method rather than scattered examples, it is asking red-teamers to probe the integrity of the overall alignment layer.

The reward size also signals priority. A $25,000 prize is modest relative to the scale of major software vulnerability programs, but substantial enough to attract credible specialists in AI security and biosecurity. More importantly, it clarifies that OpenAI is willing to pay for evidence that its safeguards can be broken under controlled conditions before those weaknesses are exploited elsewhere.