人工智能是否在思考?
The Case That A.I. Is Thinking

原始链接: https://www.newyorker.com/magazine/2025/11/10/the-case-that-ai-is-thinking

最初持怀疑态度的著名认知科学家道格拉斯·霍夫施塔特在2023年GPT-4发布后改变了看法,承认它具有以“异类方式”思考的非凡能力。 这种转变凸显了大型语言模型(LLM)的快速发展及其与人类认知惊人的相似之处。 LLM通过将单词甚至图像表示为高维空间中的数字“向量”来运作,并通过训练调整这些坐标以反映关系和含义。 这使得类比推理成为可能——以“巴黎-法国-意大利-罗马”为例——并对上下文进行细致的理解。 近期研究甚至确定了这些模型中与特定概念相关的特定“特征”,表明存在以前未见过的内部表征水平。 有趣的是,这些模型背后的数学原理,特别是谷歌的Transformer架构,呼应了数十年前由Pentti Kanerva提出的理论。 这种融合正在促进人工智能和神经科学之间的相互关系,研究人员现在使用LLM来模拟和理解人脑,从而实现认知科学长期以来的梦想。 人工智能的“黑匣子”实际上越来越容易进行科学探究。

## AI 思维:细致的讨论 一篇最近的《纽约客》文章引发了 Hacker News 上关于人工智能是否真正“思考”的讨论。 讨论的中心在于找到一个中间立场,既不否定人工智能的用处,也不将其归因于类似人类的思维过程。 许多评论者认为大型语言模型 (LLM) 代表着一项重要的技术飞跃,反映了人类认知的一部分——特别是模式识别和预测。 它们的功能类似于大脑分类和理解信息的能力,例如识别“大象”,无论像素变化如何。 然而,共识是 LLM 目前缺乏人类思维的关键要素:自我动力、目标导向的行动以及更广泛的自我和情境意识。 尽管逼近这些特质并非不可能,但仍然存在怀疑。 一个关键点是定义和衡量意识和主观体验的困难,即使是在人类身上也是如此。 最终,这场辩论强调了区分“AI 作为工程学”(无论 *如何* 思考,都创造智能机器)和“AI 作为认知科学”(试图复制人类思维)的重要性。 讨论强调,当前的人工智能虽然令人印象深刻,但可能与人类智能从根本上不同。
相关文章

原文

Kanerva’s book receded from view, and Hofstadter’s own star faded—except when he occasionally poked up his head to criticize a new A.I. system. In 2018, he wrote of Google Translate and similar technologies: “There is still something deeply lacking in the approach, which is conveyed by a single word: understanding.” But GPT-4, which was released in 2023, produced Hofstadter’s conversion moment. “I’m mind-boggled by some of the things that the systems do,” he told me recently. “It would have been inconceivable even only ten years ago.” The staunchest deflationist could deflate no longer. Here was a program that could translate as well as an expert, make analogies, extemporize, generalize. Who were we to say that it didn’t understand? “They do things that are very much like thinking,” he said. “You could say they are thinking, just in a somewhat alien way.”

L.L.M.s appear to have a “seeing as” machine at their core. They represent each word with a series of numbers denoting its coördinates—its vector—in a high-dimensional space. In GPT-4, a word vector has thousands of dimensions, which describe its shades of similarity to and difference from every other word. During training, a large language model tweaks a word’s coördinates whenever it makes a prediction error; words that appear in texts together are nudged closer in space. This produces an incredibly dense representation of usages and meanings, in which analogy becomes a matter of geometry. In a classic example, if you take the word vector for “Paris,” subtract “France,” and then add “Italy,” the nearest other vector will be “Rome.” L.L.M.s can “vectorize” an image by encoding what’s in it, its mood, even the expressions on people’s faces, with enough detail to redraw it in a particular style or to write a paragraph about it. When Max asked ChatGPT to help him out with the sprinkler at the park, the model wasn’t just spewing text. The photograph of the plumbing was compressed, along with Max’s prompt, into a vector that captured its most important features. That vector served as an address for calling up nearby words and concepts. Those ideas, in turn, called up others as the model built up a sense of the situation. It composed its response with those ideas “in mind.”

A few months ago, I was reading an interview with an Anthropic researcher, Trenton Bricken, who has worked with colleagues to probe the insides of Claude, the company’s series of A.I. models. (Their research has not been peer-reviewed or published in a scientific journal.) His team has identified ensembles of artificial neurons, or “features,” that activate when Claude is about to say one thing or another. Features turn out to be like volume knobs for concepts; turn them up and the model will talk about little else. (In a sort of thought-control experiment, the feature representing the Golden Gate Bridge was turned up; when one user asked Claude for a chocolate-cake recipe, its suggested ingredients included “1/4 cup dry fog” and “1 cup warm seawater.”) In the interview, Bricken mentioned Google’s Transformer architecture, a recipe for constructing neural networks that underlies leading A.I. models. (The “T” in ChatGPT stands for “Transformer.”) He argued that the mathematics at the heart of the Transformer architecture closely approximated a model proposed decades earlier—by Pentti Kanerva, in “Sparse Distributed Memory.”

Should we be surprised by the correspondence between A.I. and our own brains? L.L.M.s are, after all, artificial neural networks that psychologists and neuroscientists helped develop. What’s more surprising is that when models practiced something rote—predicting words—they began to behave in such a brain-like way. These days, the fields of neuroscience and artificial intelligence are becoming entangled; brain experts are using A.I. as a kind of model organism. Evelina Fedorenko, a neuroscientist at M.I.T., has used L.L.M.s to study how brains process language. “I never thought I would be able to think about these kinds of things in my lifetime,” she told me. “I never thought we’d have models that are good enough.”

It has become commonplace to say that A.I. is a black box, but the opposite is arguably true: a scientist can probe the activity of individual artificial neurons and even alter them. “Having a working system that instantiates a theory of human intelligence—it’s the dream of cognitive neuroscience,” Kenneth Norman, a Princeton neuroscientist, told me. Norman has created computer models of the hippocampus, the brain region where episodic memories are stored, but in the past they were so simple that he could only feed them crude approximations of what might enter a human mind. “Now you can give memory models the exact stimuli you give to a person,” he said.

联系我们 contact @ memedata.com