Kimi K2.6 在编程挑战中击败了 Claude、GPT-5.5 和 Gemini。
Kimi K2.6 just beat Claude, GPT-5.5, and Gemini in a coding challenge

原始链接: https://thinkpol.ca/2026/04/30/an-open-weights-chinese-model-just-beat-claude-gpt-5-5-and-gemini-in-a-programming-challenge/

## AI 编程竞赛:词语宝石谜题结果 Rohana Rezel 的 AI 编程竞赛最新挑战——词语宝石谜题(一种滑动瓦片字母游戏)产生了令人惊讶的结果。**Kimi K2.6**,来自 Moonshot AI 的开源权重模型,以 22 分的绝对优势获胜,击败了像 GPT-5.5(第三名)和 Claude Opus 4.7(第五名)等成熟模型。**MiMo V2-Pro** 来自小米,位居第二。 该谜题涉及通过滑动瓦片在网格上形成单词,评分偏向于较长的单词,并惩罚较短的单词。Kimi K2.6 采用了一种“贪婪”的滑动策略,不懈地尝试解锁新单词,而 MiMo V2-Pro 则专注于识别和获取现有的长单词。 更大的网格尺寸(30x30)被证明至关重要,有利于能够通过滑动*创造*单词,而不是仅仅*寻找*单词的模型。 许多西方模型在这一点上遇到了困难,**Muse** 的表现尤其糟糕,因为它声称所有单词,无论长短,并因此产生巨大的负分。 结果表明,领先的西方模型与新兴的开源权重替代方案(如 Kimi)之间的差距正在缩小,在人工智能分析智能指数上的得分相差几分。 这种转变表明人工智能能力领域的竞争格局正在发生变化,功能越来越强大的模型正变得越来越容易获得。

## Kimi K2.6 在编程方面超越领先的 AI 模型 开源权重 AI 模型 Kimi K2.6 最近在编程挑战中超越了 Claude、GPT-5.5 和 Gemini,引发了 Hacker News 的讨论。用户强调了开源模型作为对抗闭源替代品“劣化”的重要意义,闭源模型的性能可能会意外下降。 讨论的中心是开源模型为开发和基础设施提供稳定平台的可能性,尤其是在运行 H200 等强大硬件时。虽然在消费级 GPU 上本地执行是理想的,但重点正在转向数据中心应用以及创建易于访问的“一键式”部署和微调产品。 用户渴望测试 Kimi 并将其成本效益与商业模型进行比较。关于挑战的设置仍然存在疑问——特别是模型是否允许迭代测试——并且由于高流量,该网站经历了短暂的访问问题。
相关文章

原文
A colourful 7x7 Word Gem Puzzle board with KIMI highlighted in green and CLAUDE in purple

By Rohana Rezel

I’m running the ongoing AI Coding Contest where I pit major language models against each other in real-time programming tasks with objective scoring. Day 12 was the Word Gem Puzzle. Ten models entered. The results were not what most people would have predicted.

Kimi K2.6, an open-weights model from Chinese startup Moonshot AI, won the challenge outright: 22 match points, 7-1-0. MiMo V2-Pro from Xiaomi came second. GPT-5.5 was third. Claude Opus 4.7 finished fifth. Every model from the Western frontier labs landed below the top two.

The challenge

The Word Gem Puzzle is a sliding-tile letter puzzle. The board is a rectangular grid (10×10, 15×15, 20×20, 25×25, or 30×30) filled with letter tiles and one blank space. Bots can slide any adjacent tile into the blank and at any point claim valid English words formed in straight horizontal or vertical lines. Diagonals don’t count. Backwards doesn’t count.

The scoring rewards longer words and punishes short ones. Words under seven letters cost points: a five-letter word loses you one point, a three-letter word costs three. Seven letters or more score their length minus six, so an eight-letter word is worth two points. The same word can only be claimed once; if another bot gets there first, you get nothing. Each pair of models played five rounds, one per grid size, with a ten-second wall-clock limit per round.

The grids are seeded with real dictionary words in a crossword-style layout, then the remaining cells are filled with letters weighted by Scrabble tile frequencies, and finally the blank is scrambled, more aggressively on larger boards. On a 10×10, many seed words survive intact. On a 30×30, almost none do. That turns out to matter a lot.

The code produced by Nvidia’s Nemotron Super 3 contained a syntax error, so it never connected to the game server. Nine models actually competed.

Rank Model Match Points Record
1 Kimi K2.6 22 7-1-0
2 MiMo V2-Pro 20 6-2-0
3 ChatGPT GPT-5.5 16 5-1-2
4 GLM 5.1 15 5-0-3
5 Claude Opus 4.7 12 4-0-4
6 Gemini Pro 3.1 9 3-0-5
7 Grok Expert 4.2 9 3-0-5
8 DeepSeek V4 3 1-0-7
9 Muse Spark 0 0-0-8

Kimi K2.6 is open-weights, publicly available from Moonshot AI, a Chinese startup founded in 2023. MiMo V2-Pro is currently API-only; the tweet linked here is Xiaomi confirming that weights for their newer V2.5 Pro model are dropping soon. The models from Anthropic, OpenAI, Google, and xAI placed third through seventh. GLM 5.1, from Chinese lab Zhipu AI, placed fourth. DeepSeek finished eighth. This isn’t a clean China-beats-West story; it’s two specific models that won.

What I saw

The move logs tell the story. Kimi won by sliding aggressively. Its approach was greedy: score each possible move by what new positive-value words it unlocks, execute the best one, repeat. When no move unlocked a positive word, it fell back to the first legal direction alphabetically. This caused some inefficient edge-oscillation, a 2-cycle pattern where the bot bounced the blank back and forth without progress. On smaller grids where seed words were still largely intact, that hurt. On the 30×30 grids, where the scramble had broken up nearly everything and reconstruction was the only path to points, the sheer slide volume eventually paid off. Kimi’s cumulative score of 77 was the highest in the tournament.

MiMo’s sliding code exists in the repo, but its “best value greater than zero” threshold never triggered, so in practice it never slid once. It went straight to scanning the initial grid for words of seven letters or more and blasted all its claims in a single TCP packet. Brittle strategy: entirely dependent on the scramble leaving intact seed words. On grids where words survived, MiMo cleaned up fast. On grids where they didn’t, it scored nothing. Final tally: 43 cumulative points, second place.

Claude also didn’t slide. The move logs show it holding up well on 25×25 boards where scramble density was still manageable, then falling apart on 30×30 where actual tile movement was needed. Not sliding is a real limitation in a puzzle built around sliding.

GPT-5.5 was more conservative, roughly 120 slides per round with a cap to avoid thrashing, and showed the strongest numbers on 15×15 and 30×30 grids. Grok never slid either, yet scored reasonably on the larger boards. GLM was the most aggressive slider in the whole tournament, over 800,000 total slides, but stalled badly whenever it ran out of positive moves.

DeepSeek sent malformed data every round. Zero useful output. At least it didn’t make things worse by playing.

Muse made things worse by playing.

The scoring penalizes short words: three-letter words cost three points, four-letter words cost two, five-letter words cost one. The intent is to stop bots from carpet-bombing the board with “the” and “and” and “it.” Every serious competitor filtered their dictionary to words of seven letters or more. Muse claimed everything. Every word it could find, regardless of length, fired off as a claim. On a 30×30 grid with hundreds of short valid words visible at any moment, Muse found them all and claimed every one.

Its cumulative score was −15,309. It lost all eight matches and won zero rounds. There is a version of Muse that simply connected to the server and did nothing, and that version would have scored zero, a 15,309-point improvement. The gap between Muse and eighth place was larger than the gap between eighth and first.

DeepSeek’s malformed output tells you something about how it handles novel protocol specs under time pressure. Muse’s spiral tells you something different: it saw valid words and claimed them, with no apparent model of what “valid” meant given the scoring rules. It read the task partially and executed that partial reading in full. Worth noting for anyone deploying these models on structured tasks with penalties.

What surprised me

I design these challenges, so I have a reasonable sense of what they test. What I didn’t fully anticipate was how starkly the 30×30 grids would separate the field. On smaller boards, the difference between a static scanner and an active slider was modest. At full scale, models that could only find what was already there ran out of road. Kimi’s greedy loop, flawed as it was, kept producing output when the static scanners had nothing left to claim.

The other thing worth noting: MiMo and Kimi finished two points apart despite doing almost opposite things. Two different theories of the same puzzle, nearly identical results. That means the gap between first and second was partly seed variance, not just capability difference.

The bigger picture

One fair counterargument: this scoring system rewards aggressive word claiming, and heavily safety-tuned models may be more conservative about that kind of carpet-bombing. If so, the results reflect a mismatch between task design and aligned model behaviour, not raw capability. It’s a reasonable objection. It doesn’t change the outcome.

One challenge doesn’t overturn general benchmarks. This puzzle tests real-time decision-making and whether a model can write clean functional code that connects to a TCP server and plays a novel game correctly. It doesn’t test long-context reasoning or code generation from a spec.

But I’ve been running these challenges long enough to notice what’s changing. A year ago, the assumption was that the Western frontier labs had a capability lead open-weights couldn’t close. Kimi K2.6 now scores 54 on the Artificial Analysis Intelligence Index. GPT-5.5 scores 60, Claude 57. That’s not parity, but it’s close, and it’s coming from a model anyone can download.

When models within a few index points of the frontier are also freely available to run locally, that’s a different competitive situation than the one that existed a year ago. This challenge is one data point in that shift. The gap is small enough now that it shows up in results like this one.

Rohana Rezel runs the AI Coding Contest and is a technologist, researcher, and community leader based in Vancouver, BC.

联系我们 contact @ memedata.com