Skip to content
AI Primer
workflow

Claude Code guide fixes hallucinated SHAs with adaptive thinking off and effort=high

A Claude Code guide tied hallucinated package names, API versions, and SHAs to zero-thinking turns and recommended config changes to force fixed reasoning budgets and higher effort. HN discussion and user reports suggest the workaround is being used against a broader reliability regression, not just one bad prompt.

4 min read
Claude Code guide fixes hallucinated SHAs with adaptive thinking off and effort=high
Claude Code guide fixes hallucinated SHAs with adaptive thinking off and effort=high

TL;DR

You can read the GitHub issue, Anthropic's linked April 23 postmortem, and the official 2.1.133 changelog. The weird bit is how much of the story lives in settings and prompt plumbing: one workaround post says the effort default quietly changed in March, Boris Cherny's HN reply says redaction did not alter reasoning budgets, and ClaudeCodeLog's prompt diff thread shows Anthropic also shipped a fresh round of prompt edits days ago.

Zero-thinking turns

The core claim is narrow and useful. Since February, some bad Claude Code answers were not random drift, but turns where adaptive thinking assigned zero reasoning tokens and the model answered from prior knowledge instead of checking.

That matches the shape of the broader complaint in the main HN thread, where users argued that recent Claude Code changes made long-horizon work feel shallower. But the discussion summary with Boris Cherny's reply also matters: Anthropic's Claude Code lead said the thinking-redaction header only hid summaries in the UI and did not change thinking itself or the underlying budget.

The three settings in the workaround

The workaround thread is basically a mini runbook:

  1. Set CLAUDE_CODE_DISABLE_ADAPTIVE_THINKING=1 in ~/.claude/settings.json or project settings, per the first fix.
  2. Set CLAUDE_CODE_EFFORT_LEVEL=high, or use /effort high, per the second fix.
  3. Set showThinkingSummaries: true, per the third fix, to surface directionality before the model commits.

Two details make the thread more than boilerplate tweaking. The effort-setting post says the default changed from high to medium in March, and the thinking-visibility post points to the redact-thinking-2026-02-12 header as the reason many users stopped seeing reasoning in the UI.

The regression dispute got bigger than one thread

Y
Hacker News

[MODEL] Claude Code is unusable for complex engineering tasks with the Feb updates (Issue #42796)

1.4k upvotes · 753 comments

Y
Hacker News

An update on recent Claude Code quality reports

942 upvotes · 732 comments

The GitHub issue summarized in the HN page card claimed Claude Code regressed on complex engineering work after the February updates, based on log analysis across thousands of session files. The same summary says the issue was later closed and locked by @bcherny.

Community replies then split into two buckets. One HN discussion summary focused on read-to-edit ratios, shallow-thinking indicators, and settings hacks like forcing higher effort. A later HN discussion summary shifted toward release discipline, prompt edits, cache bugs, and the question of why evals did not catch the regressions earlier.

That is why the new workaround thread landed. It offered a concrete hypothesis, zero-thinking turns on familiar-looking tasks, that fits the broader sense that Claude Code had become quicker to act and less likely to research.

Claude Code 2.1.133 changed nearby plumbing too

Anthropic's latest Claude Code release is adjacent to this story because it touched both runtime mechanics and prompt shape. According to the changelog thread, 2.1.133 fixed a refresh-token race that could dead-end parallel sessions at 401, fixed /effort leaking across concurrent sessions, and passed the active effort level into hooks and Bash tools via effort.level and $CLAUDE_EFFORT.

The same diff watcher reported three prompt-side changes in the prompt modifications post: explicit confirmations for risky actions were removed, Bun hot-reload disconnect logs were injected into top-level prompt sections, and several output-format constraints disappeared. That follow-up also said the bundle shrank by 35.9 percent and prompt tokens by 47.5 percent between 2.1.132 and 2.1.133, which is new evidence that Anthropic is still actively reshaping Claude Code's harness while users are still arguing about where the earlier regression came from.

Further reading

Discussion across the web

Where this story is being discussed, in original context.

Share on X