Claude Code Faces Criticism Over Excessive Token Usage in Code Generation
Developers are raising alarms about excessive token consumption in Anthropic's Claude Code, leading to performance bottlenecks and increased costs. The issue, documented on GitHub and discussed on Hacker News, highlights broader concerns about AI efficiency in developer tooling.

Claude Code Faces Criticism Over Excessive Token Usage in Code Generation
Developers using Anthropic’s Claude Code AI assistant are reporting significant inefficiencies tied to excessive token usage during code generation and editing tasks. The issue, detailed in a GitHub issue (#16856), has sparked a wave of concern across developer communities, with users noting that the model often consumes far more computational resources than necessary—leading to slower response times, higher API costs, and diminished usability in production environments.
The term "excessive," as defined by Dictionary.com, refers to "going beyond what is reasonable, necessary, or appropriate." In the context of AI-powered coding tools, this definition resonates strongly: when a model generates 5,000 tokens to complete a task that could be accomplished in 500, the behavior is not merely inefficient—it is operationally unsustainable at scale. According to GitHub user reports, Claude Code frequently repeats code snippets, expands trivial comments into paragraphs, and generates verbose boilerplate code even when minimal output is requested.
On Hacker News, where the issue garnered 52 points and 13 comments, developers shared anecdotal evidence of token bloat. One user noted that a simple function refactor triggered a 3,200-token response, while another reported that Claude Code would "re-explain the entire project architecture" after each minor edit. These behaviors contrast sharply with competing models like GitHub Copilot, which are optimized for precision and brevity. The pattern suggests a design philosophy prioritizing comprehensiveness over concision—a trade-off that may appeal to beginners but alienates experienced engineers working under tight resource constraints.
While Oxford Learner’s Dictionary defines "excessive" as "more than is needed, allowed, or reasonable," the implications in this case extend beyond semantics. In enterprise settings, token usage directly correlates with API billing. Anthropic’s pricing model, which charges per token, means that inefficient generation can lead to unexpectedly high operational costs. A single developer using Claude Code for several hours a day could incur hundreds of dollars in monthly fees due to inflated token consumption—a burden that becomes prohibitive for startups and open-source maintainers.
Notably, the GitHub issue does not accuse Anthropic of malfeasance but rather calls for configurability. Users are requesting options to enforce "concise mode," limit output length, or disable verbose explanations. Some have suggested implementing a token budget system, similar to those found in enterprise LLM platforms, where users can cap responses to a predefined threshold. These are not fringe requests—they reflect a growing demand for transparency and control in AI-assisted development tools.
Interestingly, while Merriam-Webster includes definitions tied to substance abuse and behavioral excess, the core linguistic thread remains consistent: excessive behavior, whether in addiction or in AI output, is characterized by a lack of restraint and proportionality. In the case of Claude Code, the lack of restraint is not biological but algorithmic—a consequence of training objectives that reward verbosity over precision.
Anthropic has yet to issue a public response to the issue. However, given the visibility of the GitHub thread and the growing chorus of developer feedback, pressure is mounting for a technical resolution. The broader industry is watching: as AI coding assistants become central to software workflows, efficiency will be as critical as accuracy. The challenge for Anthropic is not just to generate code—but to generate the right amount of code.
For now, developers are advised to monitor token usage closely and consider supplementing Claude Code with other tools until optimizations are implemented. The future of AI-assisted programming depends not just on intelligence—but on discipline.


