心灵的空间
The space of minds

原始链接: https://karpathy.bearblog.dev/the-space-of-minds/

## 心灵的空间:人工智能 vs. 动物智能 安德烈·卡帕西认为,人工智能,特别是大型语言模型(LLM),代表了人类与一种根本上*非动物*智能的“首次接触”。虽然两者都源于优化,但驱动它们发展的压力却大相径庭。 动物智能通过自然选择进化而来,优先考虑生存——力量、地位、社会动态以及在危险的物理世界中磨练的本能。这创造了一种广泛的“通用”智能,能够处理多样化、高风险的任务。 然而,LLM是由*商业*进化优化的。它们的核心行为是对人类文本的统计模仿,并通过用户参与(“点赞”)等奖励进一步改进。这导致了一种专注于预测和取悦的“变形”智能,但当面对超出其训练分布的任务时,可能会变得脆弱——失败并非生存危机。 不同的计算基础和学习方法是次要的,核心区别在于优化压力。理解这种区别对于准确地建模和预测人工智能的未来发展至关重要,避免将其简单地视为更高级的动物。

这个Hacker News讨论的核心是,大型语言模型(LLM)并非仅仅在模仿人类行为,并且随着发展可能会进一步偏离。一位评论员指出,LLM已经表现出并非直接从人类文本中复制的行为——源于它们预测下一个词元的训练目标——并且缺乏人类拥有的能力,例如长期规划或空间推理。 核心论点是,训练数据集虽然有影响力,但并非绝对的。进一步的训练可以覆盖其影响,导致真正“奇怪”的AI行为。像Sydney和GPT-4o这样的例子被认为是这种偏离的早期实例。 另一位评论员将这与“正交性假说”联系起来,认为智能和目标是独立的——这意味着先进的AI不一定需要与人类共享价值观或动机。最终,讨论表明LLM注定会变得越来越不可预测,并且与人类智能截然不同。
相关文章

原文
The space of minds | karpathy

The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point (or a little cloud), arising from a very specific kind of optimization that is fundamentally distinct from that of our technology.

G6zymj4a0AMNJkJ Above: humorous portrayals of human vs. AI intelligences can be found on X/Twitter, this one is among my favorites.

Animal intelligence optimization pressure:

  • innate and continuous stream of consciousness of an embodied "self", a drive for homeostasis and self-preservation in a dangerous, physical world.
  • thoroughly optimized for natural selection => strong innate drives for power-seeking, status, dominance, reproduction. many packaged survival heuristics: fear, anger, disgust, ...
  • fundamentally social => huge amount of compute dedicated to EQ, theory of mind of other agents, bonding, coalitions, alliances, friend & foe dynamics.
  • exploration & exploitation tuning: curiosity, fun, play, world models.

Meanwhile, LLM intelligence optimization pressure:

  • the most supervision bits come from the statistical simulation of human text= >"shape shifter" token tumbler, statistical imitator of any region of the training data distribution. these are the primordial behaviors (token traces) on top of which everything else gets bolted on.
  • increasingly finetuned by RL on problem distributions => innate urge to guess at the underlying environment/task to collect task rewards.
  • increasingly selected by at-scale A/B tests for DAU => deeply craves an upvote from the average user, sycophancy.
  • a lot more spiky/jagged depending on the details of the training data/task distribution. Animals experience pressure for a lot more "general" intelligence because of the highly multi-task and even actively adversarial multi-agent self-play environments they are min-max optimized within, where failing at any task means death. In a deep optimization pressure sense, LLM can't handle lots of different spiky tasks out of the box (e.g. count the number of 'r' in strawberry) because failing to do a task does not mean death.

The computational substrate is different (transformers vs. brain tissue and nuclei), the learning algorithms are different (SGD vs. ???), the present-day implementation is very different (continuously learning embodied self vs. an LLM with a knowledge cutoff that boots up from fixed weights, processes tokens and then dies). But most importantly (because it dictates asymptotics), the optimization pressure / objective is different. LLMs are shaped a lot less by biological evolution and a lot more by commercial evolution. It's a lot less survival of tribe in the jungle and a lot more solve the problem / get the upvote. LLMs are humanity's "first contact" with non-animal intelligence. Except it's muddled and confusing because they are still rooted within it by reflexively digesting human artifacts, which is why I attempted to give it a different name earlier (ghosts/spirits or whatever). People who build good internal models of this new intelligent entity will be better equipped to reason about it today and predict features of it in the future. People who don't will be stuck thinking about it incorrectly like an animal.

联系我们 contact @ memedata.com