3 min read
[AI Minor News]

Claude Code Hits Usage Limits at Lightning Speed? Reports of 20x Costs Due to Bugs… Anthropic Investigating


  • Anthropic's AI coding assistant, Claude Code, is causing quite a stir as many users are hitting usage limits (quotas) much sooner than expected, leading to interruptions in their work. ...
※この記事はアフィリエイト広告を含みます

Claude Code Hits Usage Limits at Lightning Speed? Reports of 20x Costs Due to Bugs… Anthropic Investigating

📰 News Overview

  • Many users of Anthropic’s AI coding assistant, Claude Code, are reaching usage limits (quotas) much faster than expected, causing work interruptions.
  • Users have reported that, while they used to work for 8 hours, they’ve now hit limits in just 1 hour, with complaints that even the $200 annual Pro plan allows usage for only “12 days a month.”
  • Anthropic has acknowledged the issue and confirmed that investigating it is now their team’s top priority.

💡 Key Points

  • Cost Inflation Bug: Users who reverse-engineered the binary discovered two bugs that destroy prompt caching, inflating costs by 10 to 20 times. Some have reported improvements by downgrading to an older version (2.1.34).
  • Cache Lifespan Trap: The default lifespan of the prompt cache is a mere 5 minutes. Even short breaks can lead to cache expiration, resulting in hefty input costs when users resume work.
  • Opaque Usage Limits: Anthropic has only provided vague relative expressions like “five times that of the free version,” leaving developers in the dark about exact token limits, complicating their planning.

🦈 Shark’s Eye (Curator’s Perspective)

Having a prompt cache that expires in 5 minutes is a real buzzkill for development flow! If you step away for a quick coffee or to review specs, your cache evaporates, and each restart comes at a heavy input cost—talk about inefficient! Especially when using “automated workflows,” the risk of rate limit errors triggering silent retries could eat through a day’s budget in minutes. This is a crucial lesson for agent development!

🚀 What’s Next?

In light of these issues, Anthropic is expected to rush improvements on caching efficiency and bug fixes. On the user side, there will likely be a greater emphasis on explicit error handling during retry processes and more rigorous prompt designs to optimize costs. Negotiations between AI vendors and developers on “fair pricing and usage models” are bound to continue.

💬 A Note from HaruShark

It’s a shame to have such a rapid development tool sidelined by restrictions! While waiting for bug fixes, rolling back to an older version might just be the smart move! 🦈💥

📚 Glossary

  • Prompt Cache: A technique that temporarily stores previous inputs for reuse, speeding up processing and reducing costs.

  • Quota: The allocation of processing power or token quantity that users can utilize within a certain timeframe.

  • Rate Limiting: Restrictions imposed by a system to prevent a flood of requests in a short period.

  • Source: Anthropic: Claude Code users hitting usage limits ‘way faster than expected’

【免責事項 / Disclaimer / 免责声明】
JP: 本記事はAIによって構成され、運営者が内容の確認・管理を行っています。情報の正確性は保証せず、外部サイトのコンテンツには一切の責任を負いません。
EN: This article was structured by AI and is verified and managed by the operator. Accuracy is not guaranteed, and we assume no responsibility for external content.
ZH: 本文由AI构建,并由运营者进行内容确认与管理。不保证准确性,也不对外部网站的内容承担任何责任。
🦈