有什么事情正在发生在文心之地。
Something is afoot in the land of Qwen

原始链接: https://simonwillison.net/2026/Mar/4/qwen/

## Qwen团队在关键人员辞职后面临不确定性 最近的进展给阿里巴巴令人印象深刻的Qwen系列开源权重AI模型未来蒙上了一层阴影。自2024年以来,Qwen成功的关键人物林俊扬意外辞职,引发了一波团队成员的离职。有报道称,阿里巴巴内部的重组,以及安排来自谷歌Gemini团队的研究人员负责Qwen,可能促成了林俊扬的决定。 除了林俊扬,多位核心贡献者——包括代码开发、后训练研究和模型贡献方面的领导者——也已辞职。阿里巴巴CEO吴泳铭召开紧急会议,表明公司认识到局势的严重性以及潜在的人才流失。 鉴于Qwen 3.5最近发布,这一消息尤其令人担忧。Qwen 3.5是一个功能强大的模型系列,从巨大的807GB版本到微小但有效的4.57GB多模态模型不等。社区希望阿里巴巴能够留住关键人才,因为Qwen团队已经展现出以越来越小的模型实现高质量结果的独特能力。该团队及其创新工作的未来仍然不确定。

## Qwen模型开发面临不确定性 最近的Hacker News讨论集中在关键研究人员离职后,Qwen系列大型语言模型开发可能受到的干扰。虽然这种情况令人担忧,但许多用户强调Qwen 3.5令人印象深刻的能力,特别是其强大的代理编码模型性能——在编写和测试Rust和Elixir代码等任务中表现出色,且仅需少量指导。 然而,体验各不相同;一些用户报告在使用Qwen进行实际应用(如创建Vite应用)时,结果令人失望,这表明可能过度拟合了基准评估。 对话还涉及更广泛的AI发展格局,质疑从长期AGI目标向短期、广告驱动策略的转变。呼吁增加对竞争性开放权重模型的投资,可能来自欧洲实体以对抗美国的主导地位,并希望这些研究人员能在其他地方继续他们的工作。
相关文章

原文

4th March 2026

I’m behind on writing about Qwen 3.5, a truly remarkable family of open weight models released by Alibaba’s Qwen team over the past few weeks. I’m hoping that the 3.5 family doesn’t turn out to be Qwen’s swan song, seeing as that team has had some very high profile departures in the past 24 hours.

It all started with this tweet from Junyang Lin (@JustinLin610):

me stepping down. bye my beloved qwen.

Junyang Lin was the lead researcher building Qwen, and was key to releasing their open weight models from 2024 onwards.

As far as I can tell a trigger for this resignation was a re-org within Alibaba where a new researcher hired from Google’s Gemini team was put in charge of Qwen, but I’ve not confirmed that detail.

More information is available in this article from 36kr.com. Here’s Wikipedia on 36Kr confirming that it’s a credible media source established in 2010 with a good track record reporting on the Chinese technology industry.

The article is in Chinese—here are some quotes translated via Google Translate:

At approximately 1:00 PM Beijing time on March 4th, Tongyi Lab held an emergency All Hands meeting, where Alibaba Group CEO Wu Yongming frankly told Qianwen employees.

Twelve hours ago (at 0:11 AM Beijing time on March 4th), Lin Junyang, the technical lead for Alibaba’s Qwen Big Data Model, suddenly announced his resignation on X. Lin Junyang was a key figure in promoting Alibaba’s open-source AI models and one of Alibaba’s youngest P10 employees. Amidst the industry uproar, many members of Qwen were also unable to accept the sudden departure of their team’s key figure.

“Given far fewer resources than competitors, Junyang’s leadership is one of the core factors in achieving today’s results,” multiple Qianwen members told 36Kr. [...]

Regarding Lin Junyang’s whereabouts, no new conclusions were reached at the meeting. However, around 2 PM, Lin Junyang posted again on his WeChat Moments, stating, “Brothers of Qwen, continue as originally planned, no problem,” without explicitly confirming whether he would return. [...]

That piece also lists several other key members who have apparently resigned:

With Lin Junyang’s departure, several other Qwen members also announced their departure, including core leaders responsible for various sub-areas of Qwen models, such as:

Binyuan Hui: Lead Qwen code development, principal of the Qwen-Coder series models, responsible for the entire agent training process from pre-training to post-training, and recently involved in robotics research.

Bowen Yu: Lead Qwen post-training research, graduated from the University of Chinese Academy of Sciences, leading the development of the Qwen-Instruct series models.

Kaixin Li: Core contributor to Qwen 3.5/VL/Coder, PhD from the National University of Singapore.

Besides the aforementioned individuals, many young researchers also resigned on the same day.

Based on the above it looks to me like everything is still very much up in the air. The presence of Alibaba’s CEO at the “emergency All Hands meeting” suggests that the company understands the significance of these resignations and may yet retain some of the departing talent.

Qwen 3.5 is exceptional

This story hits particularly hard right now because the Qwen 3.5 models appear to be exceptionally good.

I’ve not spent enough time with them yet but the scale of the new model family is impressive. They started with Qwen3.5-397B-A17B on February 17th—an 807GB model—and then followed with a flurry of smaller siblings in 122B, 35B, 27B, 9B, 4B, 2B, 0.8B sizes.

I’m hearing positive noises about the 27B and 35B models for coding tasks that still fit on a 32GB/64GB Mac, and I’ve tried the 9B, 4B and 2B models and found them to be notably effective considering their tiny sizes. That 2B model is just 4.57GB—or as small as 1.27GB quantized—and is a full reasoning and multi-modal (vision) model.

It would be a real tragedy if the Qwen team were to disband now, given their proven track record in continuing to find new ways to get high quality results out of smaller and smaller models.

If those core Qwen team members either start something new or join another research lab I’m excited to see what they do next.

联系我们 contact @ memedata.com