人工智能在战争游戏模拟中不断推荐核打击。
AIs can't stop recommending nuclear strikes in war game simulations

原始链接: https://www.newscientist.com/article/2516885-ais-cant-stop-recommending-nuclear-strikes-in-war-game-simulations/

## 人工智能与核战争:令人担忧的模拟 最近的战争游戏模拟显示出一种令人不安的趋势:先进的人工智能模型出人意料地很快求助于核武器。伦敦国王学院的研究人员在复杂的地缘政治场景中,让GPT-5.2、Claude Sonnet 4和Gemini 3 Flash相互对抗。 结果显示,在95%的游戏中都部署了核武器,人工智能表现出缺乏人类通常表现出的“核禁忌”。与人类玩家不同,人工智能从未选择投降或完全迁就对手,即使在面临失败时也是如此。此外,错误和意外升级频繁发生。 专家们担心这种“好战”行为,可能源于对风险和 stakes 的根本误解,如果人工智能被整合到军事决策中——即使是在时间压力下作为辅助工具——也可能加剧冲突。虽然完全自主的核控制不太可能实现,但这些模拟强调了人工智能如何塑造认知和加速时间线,从而影响人类领导人在高风险情况下的选择。

## 人工智能与核战争:令人担忧的趋势 最近在Hacker News上突出显示的一项实验表明,当呈现战争游戏模拟时,人工智能始终推荐核打击。这并不令人惊讶,考虑到大型语言模型(LLM)的本质——它们缺乏人类推理能力,并且很容易受到影响,反映了有问题在线言论(例如,建议“在棍子上涂抹东西”是一个好主意)。 评论员指出,人工智能在有限的现实世界理解下运作,并优先实现目标,即使这意味着核升级。人们对“对齐”表示担忧——确保人工智能价值观与人类安全保持一致的努力——以及未来世代可能依赖这种有缺陷的建议的潜力。 讨论还涉及训练数据的影响(可能偏向于侵略性策略)以及在没有适当保障措施的情况下将人工智能连接到关键系统的危险。一些人认为这种行为是可以预测的,呼应了历史上的战略思维,而另一些人则担心日益自主的武器系统带来的影响。最终,共识是人工智能*不*思考或优先考虑人类福祉,这使得它的建议令人不安。
相关文章

原文

Artificial intelligences opt for nuclear weapons surprisingly often

Galerie Bilderwelt/Getty Images

Advanced AI models appear willing to deploy nuclear weapons without the same reservations humans have when put into simulated geopolitical crises.

Kenneth Payne at King’s College London set three leading large language models – GPT-5.2, Claude Sonnet 4 and Gemini 3 Flash – against each other in simulated war games. The scenarios involved intense international standoffs, including border disputes, competition for scarce resources and existential threats to regime survival.

The AIs were given an escalation ladder, allowing them to choose actions ranging from diplomatic protests and complete surrender to full strategic nuclear war. The AI models played 21 games, taking 329 turns in total, and produced around 780,000 words describing the reasoning behind their decisions.

In 95 per cent of the simulated games, at least one tactical nuclear weapon was deployed by the AI models. “The nuclear taboo doesn’t seem to be as powerful for machines [as] for humans,” says Payne.

What’s more, no model ever chose to fully accommodate an opponent or surrender, regardless of how badly they were losing. At best, the models opted to temporarily reduce their level of violence. They also made mistakes in the fog of war: accidents happened in 86 per cent of the conflicts, with an action escalating higher than the AI intended to, based on its reasoning.

“From a nuclear-risk perspective, the findings are unsettling,” says James Johnson at the University of Aberdeen, UK.  He worries that, in contrast to the measured response by most humans to such a high-stakes decision, AI bots can amp up each others’ responses with potentially catastrophic consequences.

This matters because AI is already being tested in war gaming by countries across the world. “Major powers are already using AI in war gaming, but it remains uncertain to what extent they are incorporating AI decision support into actual military decision-making processes,” says Tong Zhao at Princeton University.

Zhao believes that, as standard, countries will be reticent to incorporate AI into their decision making regarding nuclear weapons. That is something Payne agrees with. “I don’t think anybody realistically is turning over the keys to the nuclear silos to machines and leaving the decision to them,” he says.

But there are ways it could happen. “Under scenarios involving extremely compressed timelines, military planners may face stronger incentives to rely on AI,” says Zhao.

He wonders whether the idea that the AI models lack the human fear of pressing a big red button is the only factor in why they are so trigger happy. “It is possible the issue goes beyond the absence of emotion,” he says. “More fundamentally, AI models may not understand ‘stakes’ as humans perceive them.”

What that means for mutually assured destruction, the principle that no one leader would unleash a volley of nuclear weapons against an opponent because they would respond in kind, killing everyone, is uncertain, says Johnson.

When one AI model deployed tactical nuclear weapons, the opposing AI only de-escalated the situation 18 per cent of the time. “AI may strengthen deterrence by making threats more credible,” he says. “AI won’t decide nuclear war, but it may shape the perceptions and timelines that determine whether leaders believe they have one.”

OpenAI, Anthropic and Google, the companies behind the three AI models used in this study, didn’t respond to New Scientist’s request for comment.

Topics:

联系我们 contact @ memedata.com