Skip to content
AI Primer
release

OpenAI opens GPT-5.5 API with 1M context and Responses support

OpenAI added GPT-5.5 and GPT-5.5 Pro to the API and Playground with 1M context and Responses support. Partners including OpenRouter, Perplexity, GitHub Copilot, Vercel, Warp, and Devin rolled it out the same day, widening access beyond Codex.

7 min read
OpenAI opens GPT-5.5 API with 1M context and Responses support
OpenAI opens GPT-5.5 API with 1M context and Responses support

TL;DR

You can read OpenAI's launch post, skim the latest model guide, and Simon Willison already pulled out one of the more practical migration notes in his GPT-5.5 prompting guide. There is also a surprisingly useful Vercel changelog, a public CursorBench post, and an active Hacker News thread where the debate quickly moved from launch hype to pricing, access paths, and safety policy.

API rollout

OpenAI's April 23 messaging still framed GPT-5.5 as a ChatGPT and Codex launch, with the API only "coming soon" in OpenAIDevs' launch post, thsottiaux's announcement, and Sam Altman's launch thread. Roughly 24 hours later, OpenAIDevs' API launch and OpenAIDevs' API details changed that to general API availability, then Sam Altman's API post repeated the update from his own account.

That one-day gap matters because it clarifies what actually shipped first. GPT-5.5 was not an API-first release. It was a Codex and ChatGPT release that only became a full developer platform launch once Responses and Chat Completions support went live the next day.

Responses API and 1M context

The most concrete API change in OpenAIDevs' API details is surface area: GPT-5.5 is available through both Responses and Chat Completions, while GPT-5.5 Pro is available through Responses only. The same post also set the headline context limit at 1 million tokens.

That puts the launch in line with OpenAI's current developer stack rather than as a one-off model slug. The latest model guide and OpenAI's launch post both frame GPT-5.5 as the default frontier model for complex coding and professional work, not as a narrow Codex-only checkpoint.

The linked docs also make an important migration point that Simon Willison highlighted in his prompting guide writeup: OpenAI recommends treating GPT-5.5 as a new model family to tune for, not as a drop-in prompt swap for GPT-5.2 or GPT-5.4.

Pricing and token efficiency

Pricing moved up versus GPT-5.4. scaling01's pricing screenshot showed GPT-5.5 at $5 input and $30 output per million tokens, and scaling01's pricing post put GPT-5.5 Pro at $30 input and $180 output.

OpenAI's counter-argument is token efficiency, not cheaper list price. OpenAIDevs' token-efficiency thread said GPT-5.5 gets better Codex results with fewer tokens than GPT-5.4, while Sam Altman's launch thread said it matches GPT-5.4 on per-token speed but uses significantly fewer tokens per task.

That same framing showed up immediately in partner rollouts. OpenAIDevs' Perplexity case study claimed 56% fewer tokens in Perplexity Computer workflows, and AravSrinivas' orchestration note tied the model switch directly to lower Computer credit burn.

Agent benchmarks

OpenAI's own benchmark card in OpenAIDevs' launch post is very obviously built to sell GPT-5.5 as an agent model first.

The flagship deltas versus GPT-5.4 from OpenAIDevs' launch post and OpenAI's launch post were:

  • Terminal-Bench 2.0: 75.1% to 82.7%
  • OSWorld-Verified: 75.0% to 78.7%
  • Toolathlon: 54.6% to 55.6%
  • FrontierMath Tier 4: 27.1% to 35.4%
  • CyberGym: 79.0% to 81.8%
  • GDPval: 83.0% to 84.9%

Two details stand out in the evidence pool. First, OpenAI kept repeating that GPT-5.5 is strongest on longer workflows, command-line work, and persistence across research loops in OpenAIDevs' extended thread. Second, some third-party chatter immediately pushed on benchmark variance, with ValsAI's Terminal Bench correction revising one external Terminal-Bench configuration and nrehiew_'s variance post asking why official and external runs were so far apart.

That leaves the headline chart intact, but it also means the API story is arriving with an unusually live debate over harness setup and reasoning-effort settings.

Same-day integrations

The rollout spread far beyond OpenAI's own surfaces within hours.

A quick inventory from the evidence:

This is the part of the launch that felt like Christmas come early for coding agent nerds. GPT-5.5 was not trapped inside one vendor shell for long.

Hands-on workflow reports

The early user reports were much more consistent on workflow feel than on abstract frontier bragging. The common claim was that GPT-5.5 carries tasks further without extra steering.

Examples from the first day:

The personality shift also came up repeatedly. haider1's tone post described it as more composed and more like GPT-4o in tone, while onusoz's schema-design note and BlancheMinerva's complaint show that terser style did not land uniformly.

Safety retention policy

One of the least-advertised launch details came from the agent ecosystem, not OpenAI's main posts. sqs' Amp disclosure said GPT-5.5 use in Amp's deep mode currently falls under OpenAI's Safety Retention Policy, with non-zero data retention in about 0.05% of cases that OpenAI's classifier flags as severe cybersecurity abuse risk.

Y
Hacker News

Fresh discussion on GPT-5.5

1.5k upvotes · 1k comments

That disclosure helps explain why API access lagged the ChatGPT and Codex launch by a day, and why the Hacker News thread quickly filled with questions about cyber policy, malware-analysis workflows, and access constraints. It also gives the ecosystem rollout a sharper edge: GPT-5.5 arrived everywhere fast, but not every agent wrapper is equally comfortable with the safety terms attached to it yet.

Further reading

Discussion across the web

Where this story is being discussed, in original context.

On X· 5 threads
API rollout5 posts
Pricing and token efficiency3 posts
Agent benchmarks3 posts
Same-day integrations10 posts
Hands-on workflow reports6 posts