你将无法负担起最好的AI编码工具。
You are going to get priced out of the best AI coding tools

原始链接: https://newsletter.danielpaleka.com/p/you-are-going-to-get-priced-out-of

## 人工智能工具成本上升 受沃霍尔“即使是奢侈品也广泛可得(如可口可乐)”的观察启发,作者探讨了最初人工智能编码工具(以 GitHub Copilot 为例)的低廉价格可能无法长期维持的原因。 曾经每月仅需 10 美元就能使用,但趋势表明成本将大幅上升,顶级选项可能达到每月 100 美元以上,甚至研究级代理可能高达每月 20,000 美元以上。 这不同于颠覆性技术的典型情况;大型语言模型*一开始*价格低廉,但提供的价值不成比例——以相同的成本生成比人类开发者更多的代码。 对更快、更可靠的人工智能的需求增加(例如持续的代码注释或改进的信息检索)需要更多的计算能力,从而推高价格。 改进还来自于并行运行多个人工智能实例等技术,进一步增加成本。 虽然竞争、硬件进步或规模收益递减*可能*会抑制价格上涨,但作者认为这种情况不太可能发生。 人工智能实验室内部的预期是,未来访问最佳工具的成本将越来越高,可能会让学者和小型企业无力承担,从而形成一种分层系统,价值与成本直接相关。

黑客新闻 新的 | 过去的 | 评论 | 提问 | 展示 | 工作 | 提交 登录 你将无法负担得起最好的 AI 编码工具 (danielpaleka.com) 32 分,来自 fi-le 29 分钟前 | 隐藏 | 过去的 | 收藏 | 1 条评论 帮助 iambateman 1 分钟前 [–] 我认为沃霍尔的引言是怀旧的,但不完整。 我负担不起最好的汽车、最好的房子、最好的家庭影院系统、最好的学校。即使年收入 30 万美元的人也无法负担得起所有最好的东西。 当然,iPhone 曾经是“最好的”手机,也被几乎所有人使用,但我认为即使在短期内这也是一个异常。 现在我每月支付 200 美元使用 Claude 代码来完成我原本需要每月支付 10,000 美元的工作。当然,我预计这些数字会越来越接近。 没有 VC 资助的“摇钱树”会永远持续下去。 指南 | 常见问题 | 列表 | API | 安全 | 法律 | 申请 YC | 联系 搜索:
相关文章

原文

Andy Warhol famously said:

What’s great about this country is that the richest consumers buy essentially the same things as the poorest. You can be watching TV and see Coca-Cola, and you know that the President drinks Coke, Liz Taylor drinks Coke, and just think, you can drink Coke, too.

There was a time when everyone used Github Copilot. It used to cost $10 per month, or free for students. I used it, Andrej Karpathy used it, high schoolers learning to code used it too.

This world is already partly gone; the cheapest usable tier of Claude Code is $100/mo. In this post, I outline a bunch of short arguments for why the old state of affairs was temporary, and why the best AI tools will become far more expensive.

I made a plot of a bunch of tiered offerings in AI coding tools, showing an exponential trend. There are two issues with this plot: (1) the data is biased towards products I looked up; (2) if you look at the data, this is obviously multiple disjoint trends in the higher and lower pricing regime, and fitting a straight line seems like a bad idea. But I think it is nevertheless clear that there is some sort of exponential trend.

Furthermore, OpenAI reportedly discussed charging $20k/month on PhD-level research agents with investors. This was in March, and I haven’t found anything since; so take this claim with a grain of salt.

LLMs are a very unusual disruptive technology, in the sense that they started out cheap. It has been noticed many times that there are many tasks AI agents cannot do; but when they can, they do it much cheaper than people! This was not usually the case with new technologies. Computers used to be huge and pricey. Or, consider self-driving cars: Waymo is more expensive than Uber.

In fact, at least measured by the number of lines of code they are producing, LLM coding agents are producing way more value than they cost.

This creates opportunity for anyone who can create a better product to use more compute, charge more, and make more money.

First, I would personally pay more to get frontier LLMs to (1) continuously run and comment/fill in what I am doing; (2) get to their results faster. This costs money.

Secondly, ChatGPT often fails at challenging information retrieval. The best chatbot-like experience possible today looks more like Deep Research than ChatGPT. The issue with Deep Research is that it is slow. Making a faster version is likely to both (1) increase the price; (2) increase demand.

Finally, sampling more consistently improves results; a nice way to make a better coding agent is to just run a few in parallel and pick the best one. The difference between Pass@K and Pass@1 metrics was always somewhat large, and I do not expect it to just go away; e.g. the DeepSeek-R1 paper reports performance of Deepseek-R1-Zero on a math benchmark as follows: 70% when you ask the model once; 86% when you ask the model 64 times and take the majority vote.

Although, it is kind of weird that DeepSeek does not report Pass@K for the R1 model, nor can I find any other recent release that reports this. Perhaps inference-time-scaled models are already using inference time compute efficiently.

In my impression, this is a view that has been commonly held in circles close to the AI labs. No one seems to have written anything of this form yet, though. Here’s AI industry insider Nathan Lambert commenting on this in passing, reporting from The Curve:

Within 2 years a lot of academic AI research engineering will be automated with the top end of tools (…) I also expect academics to be fully priced out from these tools. (…) but there are still meaningful technical bottlenecks that are solvable but expensive. The compute increase per available user has a ceiling too. Labs will be spending $200k+ per year per employee on AI tools easily (ie the inference cost), but most consumers will be at tiers of $20k or less due to compute scarcity

The full economic calculation would require (1) collecting data that is scarcely available outside the labs; (2) technical analysis amounting to a full research paper. As we did neither for this post, I need to steelman the opposite conclusion.

What could keep costs down? Here are some possibilities:

  1. The competition between labs (or open source) pushes them to not raise prices, nor to work on products that would require higher prices.

  2. Relatedly, the labs have an incentive to make more people use their tools; especially the most effective people who would be paying the high prices. They subsidize the cost of the tools.

  3. Hardware supply + algorithmic efficiency expands faster than demand + long horizon capabilities.

  4. Diminishing returns on scaling inference time compute; e.g. due to RL being intrinsically different from pretraining, Pass@K and Pass@1 on various benchmarks become essentially the same.

I do not feel any of these are very likely; although it would be a very fun research idea to investigate if the last one is becoming true.

联系我们 contact @ memedata.com