Anthropic added a beta advisor tool to the Messages API so Sonnet or Haiku can call Opus mid-run inside one request. Anthropic says Sonnet plus Opus scored 2.7 points higher on SWE-bench Multilingual while cutting per-task cost 11.9%.

anthropic-beta: advisor-tool-2026-03-01 header in Anthropic's API release notes, with advisor usage billed at Opus rates and executor usage billed at Sonnet or Haiku rates in the pricing docs.You can read the full launch post, check the API release note, and the interesting bit is how little machinery Anthropic is asking for. The product thread frames it as one tool added to a Messages call, TestingCatalog's diagram shows the shared-context handoff, and scaling01's benchmark screenshots surface the weirdest number in the launch, Haiku more than doubling its BrowseComp score when Opus gets pulled in selectively.
Anthropic is packaging the advisor pattern as a server-side tool, not as a separate orchestration product. In the launch post, Sonnet or Haiku stays in the driver's seat as the executor, handling tools, reading results, and iterating until it hits a decision point it cannot clear cheaply.
At that point, according to Anthropic's API description, the executor consults Opus and continues inside the same /v1/messages request. The blog's example request adds advisor_20260301 to the tools array with an Opus model name and a max_uses cap, which makes this look more like a targeted escalation hook than a new agent framework.
Anthropic also leans hard on the contrast with the usual sub-agent pattern in the official post: instead of a larger model decomposing work and delegating downward, a cheaper executor runs end to end and only escalates upward when it gets stuck.
The diagram circulating with the launch captures the important implementation detail. As TestingCatalog's screenshot shows, the executor and advisor read the same conversation, tool state, and history, so Anthropic is not asking developers to serialize state into a second request or manage a side channel for planning.
That matches the official wording in the blog post: Opus reviews curated shared context, returns a plan, correction, or stop signal, and then the executor resumes. Opus does not call tools itself and does not produce the user-facing answer.
That design explains why the cost story can work at all. As Lance Martin's thread puts it, the expensive model gets used for the hard branch points instead of grinding through the entire trajectory.
Anthropic led with the SWE-bench Multilingual number, but the launch material actually shows two different stories depending on executor.
For Sonnet 4.6 plus Opus advisor, Anthropic reports:
For Haiku 4.5 plus Opus advisor, the gains are larger in absolute score, but so is the added spend versus Haiku solo. scaling01's screenshots show:
The official post adds one more useful comparison: Haiku plus Opus still trails Sonnet solo in score, but Anthropic says it costs 85% less per task, which gives the launch a second persona beyond coding agents, cheap high-volume workloads that occasionally need a smarter plan.
The API release note is more specific than the tweet thread about rollout. In Anthropic's release notes, the advisor tool is listed as a public beta feature launched on April 9, with the required header anthropic-beta: advisor-tool-2026-03-01.
The launch post says advisor tokens are reported separately in the usage block and that developers can cap invocations with max_uses. Pricing follows the underlying models rather than a separate tool fee: Anthropic's pricing page lists Opus 4.6 at $5 per million input tokens and $25 per million output tokens, Sonnet 4.6 at $3 and $15, and Haiku 4.5 at $1 and $5.
One small but revealing line in Lance Martin's follow-up is that Anthropic is selling the advisor pattern as token efficiency, not just model mixing. The bet is that a more capable model can be cheaper when it only gets called for the forks where a weaker model would otherwise waste turns.