Server-Side Token Injection (v2.1.100+)

Description

Starting in v2.1.100, an undocumented server-side mechanism causes approximately 20,000 extra cache_creation_input_tokens to be billed per request despite slightly smaller client payloads. Proxy-verified across 40+ sessions with clean bimodal distribution: ~50K tokens pre-v2.1.100 cluster vs. ~71K v2.1.100+ cluster. Tokens enter the model's context window, competing with user instructions. Cause is unconfirmed by Anthropic; community speculation includes expanded session memory injection, expanded safety classifier context, or a User-Agent-version-keyed server routing change. GitHub #46917 remains open.

Key claims

Relations

Sources

src-20260423-542f02260352