不,Anthropic 的每个 Claude Code 用户不会花费 5000 美元。
No, it doesn't cost Anthropic $5k per Claude Code user

原始链接: https://martinalderson.com/posts/no-it-doesnt-cost-anthropic-5k-per-claude-code-user/

最近关于Anthropic的每月200美元Claude Code Max计划每用户亏损4800美元的报道具有误导性。5000美元的数字源于Cursor的内部成本——他们以接近零售价的价格*使用*该模型,而非Anthropic的实际计算费用。 虽然重度用户*可能*会根据Anthropic的定价(每百万token 5美元/25美元)达到5000美元的API等效使用量,但这并不反映Anthropic的潜在成本。与OpenRouter等平台上的类似开源模型(Qwen, Kimi)相比,推理成本大约低10倍——对于重度用户而言,可能在500美元左右。 Anthropic可能在大多数订阅者身上实现盈亏平衡或盈利,因为只有不到5%的用户达到使用上限。核心问题并非推理成本,而是模型训练、研究人员薪资和计算基础设施的巨大开支。 “昂贵的AI推理”的说法有利于前沿实验室,为其高API溢价和抑制竞争辩护。基于OpenRouter的定价,现实的观点表明推理的盈利能力远高于通常认为的。

一篇近期文章在Hacker News上引发了关于使用Anthropic的Claude Code成本的讨论。最初的说法,在福布斯的一篇文章之后在LinkedIn和Twitter上流传,认为每月200美元的Claude Code Max计划可能导致用户高达5000美元的计算成本。 然而,讨论中链接的一篇文章(martinalderson.com)认为,这些利润空间与Anthropic的Dario Amodei的估计相比被严重高估。用户们争论着驱动Claude的Opus模型的大小,猜测范围从1万亿到2万亿参数。 对话还涉及了Cursor用户(特别是“power-Cursor”用户)的成本,仍然是5000美元,以及关于重复编码任务可能导致RSI(重复性劳损)的幽默观察。最终,讨论的核心在于澄清使用Claude Code相关的实际计算成本。
相关文章

原文

My LinkedIn and Twitter feeds are full of screenshots from the recent Forbes article on Cursor claiming that Anthropic's $200/month Claude Code Max plan can consume $5,000 in compute. The relevant quote:

Today, that subsidization appears to be even more aggressive, with that $200 plan able to consume about $5,000 in compute, according to a different person who has seen analyses on the company's compute spend patterns.

This is being shared as proof that Anthropic is haemorrhaging money on inference. It doesn't survive basic scrutiny.

What the $5,000 figure actually is

I'm fairly confident the Forbes sources are confusing retail API prices with actual compute costs. These are very different things.

Anthropic's current API pricing for Opus 4.6 is $5 per million input tokens and $25 per million output tokens. At those prices, yes - a heavy Claude Code Max 20 user could rack up $5,000/month in API-equivalent usage. That maths checks out.

But API pricing is not what it costs Anthropic to serve those tokens.

The OpenRouter reality check

The best way to estimate what inference actually costs is to look at what open-weight models of similar size are priced at on OpenRouter - where multiple providers compete on price.

Qwen 3.5 397B-A17B is a good comparison point. It's a large MoE model, broadly comparable in architecture size to what Opus 4.6 is likely to be. Equally, so is Kimi K2.5 1T params with 32B active, which is probably approaching the upper limit of what you can efficiently serve.

Here's what the pricing looks like:

OpenRouter pricing showing Qwen 3.5 397B and Kimi K2.5 at roughly 10% of Claude Opus 4.6 API pricing per token The Qwen 3.5 397B model on OpenRouter (via Alibaba Cloud) costs _$0.39_ per million input tokens and _$2.34_ per million output tokens. Compare that to Opus 4.6's API pricing of $5/$25. Kimi K2.5 is even cheaper at $0.45 per million input tokens and $2.25 output.

That's roughly 10x cheaper.

And this ratio holds for cached tokens too - DeepInfra charges $0.07/MTok for cache reads on Kimi K2.5 vs Anthropic's $0.50/MTok.

These OpenRouter providers are running a business. They have to cover their compute costs, pay for GPUs, and make a margin. They're not charities. If so many can serve a model of comparable size at ~10% of Anthropic's API price and remain in business, it is hard for me to believe that they are all taking enormous losses (at ~the exact same rate range).

If a heavy Claude Code Max user consumes $5,000 worth of tokens at Anthropic's retail API prices, and the actual compute cost is roughly 10% of that, Anthropic is looking at approximately $500 in real compute cost for the heaviest users.

That's a loss of $300/month on the most extreme power users - not $4,800.

However, most users don't come anywhere near the limit. Anthropic themselves said when they introduced weekly caps that fewer than 5% of subscribers would be affected. I personally use the Max 20x plan and probably consume around 50% of my weekly token budget and it's hard to use that many tokens without getting serious RSI. At that level of usage, the maths works out to roughly break-even or profitable for Anthropic.

So who is actually losing $5,000?

The real story is actually in the article. The $5,000 figure comes from Cursor's internal analysis. And for Cursor, the number probably is roughly correct - because Cursor has to pay Anthropic's retail API prices (or close to it) for access to Opus 4.6.

So to provide a Claude Code-equivalent experience using Opus 4.6, it would cost Cursor ~$5,000 per power user per month. But it would cost Anthropic perhaps $500 max.

And the real issue for Cursor is that developers want to use the Anthropic models, even in Cursor itself. They have real "brand awareness", and they are genuinely better than the cheaper open weights models - for now at least. It's a real conundrum for them.

Anthropic is not a profitable company. But inference isn't why.

Obviously Anthropic isn't printing free cashflow. The costs of training frontier models, the enormous salaries required to hire top AI researchers, the multi-billion dollar compute commitments - these are genuinely massive expenses that dwarf inference costs.

But on a per-user, per-token basis for inference? I believe Anthropic is very likely profitable - potentially very profitable - on the average Claude Code subscriber.

The "AI inference is a money pit" narrative is misinformation that actually plays into the hands of the frontier labs. If everyone believes that serving tokens is wildly expensive, nobody questions the 10x+ markups on API pricing. It discourages competition and makes the moat look deeper than it is.

If you want to understand the real economics of AI inference, don't take API prices at face value. Look at what competitive open-weight model providers charge on OpenRouter. That's a much closer proxy for what it actually costs to run these models - and it's a fraction of what the frontier labs charge.

联系我们 contact @ memedata.com