The February regression: how a single update broke enterprise trust
The most technically significant thread in this crisis traces back to a specific date: February 12, 2026, when Anthropic rolled out thinking content redaction in Claude Code. According to AMD's Senior Director of AI Stella Laurenzo, who analyzed 17,871 thinking blocks and 234,760 tool calls across 6,852 coding sessions, the quality degradation correlated precisely with this update. The model began stopping reading code before making changes and defaulting to shallow reasoning -- what Laurenzo described as defaulting to "the cheapest action available."
This is not a vague complaint about vibes. The analysis documented 60-80% failure rates for multi-step tasks per SWE-EVO 2025 benchmarks, and the regression was measurable across enterprise-scale engineering workflows. The implication is deeply uncomfortable for Anthropic: by redacting thinking content -- likely to reduce compute costs or protect intellectual property in the reasoning chain -- the company may have inadvertently crippled the deep reasoning capabilities that made Claude Code attractive for complex engineering in the first place. Chandrika Dutt of Avasant framed it as fundamentally "a capacity and cost issue," suggesting Anthropic is making implicit tradeoffs between serving more users and maintaining reasoning depth. For enterprise customers who adopted Claude Code specifically for its superior reasoning on complex tasks, this tradeoff is unacceptable.



