The week's story is competitive pressure showing up in two places at once.
OpenAI shipped GPT-5.5 on Wednesday with the same agentic-coding pitch as Claude. The API followed a day later: $5 per million input tokens, $30 per million output tokens, 1M-token context window. Anthropic spent the same week reversing its own Claude Code regression stack - a caching bug that dropped thinking history, a verbosity prompt change that hurt coding quality, a quietly-lowered default reasoning effort - all rolled back, xhigh restored as the coding default, usage limits credited for affected subscribers. Then on Friday night they shipped 2.1.120, which crashed claude --resume and claude --continue with a JS runtime error and reintroduced a macOS sandbox failure on resumed sessions. The npm latest tag stayed pinned to 2.1.119, no GitHub Release was cut, and ~25 bug reports landed within 48 hours. The model race is one story; the harness around the model is a different and increasingly louder one.
From us: we shipped Yaw Mode for Claude Code - an overlay that bundles rules and agents while preserving conversation history across sessions. Plus a public methodology for grading any MCP server against the spec in 30 seconds (88 criteria, A–F), a Tailscale audit prompt that resolves five common scenarios in one agent run, and a piece on the local-vs-remote MCP deployment tradeoff.
From the Yaw blog
- Claude Code Yaw Mode - overlay bundle preserving conversation history across sessions
- Grading MCP Servers A to F: 88 Tests Against the Spec - open methodology, 30-second runtime, any MCP server
- Five Tailscale Audits I Now Answer in One Prompt - composing API calls into one agent run
- Local vs Remote MCP Servers: When Each One Is Right - deployment tradeoffs for agents and teams
Claude Code this week
- The Apr 23 post-mortem - Anthropic publicly identifies the three changes behind the month-long quality dip (Mar 4 reasoning-effort downgrade, Mar 26 caching bug, Apr 16 verbosity prompt), confirms all reverted/fixed, resets subscriber usage limits
- 2.1.120 (Apr 24) -
--resume/--continuecrash withUKH/FKH/g9H is not a function; macOS sandbox regression on resumed sessions. Anthropic posted a status-page advisory Apr 25 01:45 UTC: auto-update rolls clients back to 2.1.119 on next check; npmlatesttag held at 2.1.119; no GitHub Release was cut for 2.1.120 - 2.1.119 (Apr 23) - persistent
/configwrites to~/.claude/settings.json; PowerShell commands auto-approvable; MCP servers connect in parallel - 2.1.118 (Apr 23) -
/costand/statsmerged into/usage; vim visual/visual-line modes; hooks can invoke MCP tools directly - 2.1.117 (Apr 22) -
/modelselection persists across restarts;CLAUDE_CODE_FORK_SUBAGENT=1enables forked subagents - 2.1.116 (Apr 20) -
/resume67% faster on large sessions; inline progress in the thinking spinner
The broader week
- OpenAI introducing GPT-5.5 - Wed Apr 23, agentic coding, computer use, ChatGPT and Codex rollout
- GPT-5.5 and GPT-5.5 Pro now in the API (Apr 24) - $5/$30 and $30/$180 per 1M tokens, 1M context window
- DeepSeek V4 preview (Apr 24) - open-source, beats other open models on math and coding, only trails Gemini 3.1-Pro on world knowledge
- Anthropic and Amazon: up to 5 GW of new compute - Apr 20