深度搜索推理引擎开源之路
The path to open-sourcing the DeepSeek inference engine

原始链接: https://github.com/deepseek-ai/open-infra-index/tree/main/OpenSourcing_DeepSeek_Inference_Engine

DeepSeek 致力于回馈开源社区,认识到开源社区对其 AGI 进展的关键作用,特别是在 PyTorch 和 vLLM 等工具方面。继最近开源库发布获得积极反响后,DeepSeek 计划开源其内部推理引擎,该引擎对于部署 DeepSeek-V3 和 DeepSeek-R1 等模型至关重要。然而,由于代码库与早期 vLLM 分支存在差异、基础设施依赖以及有限的维护带宽,完整的开源发布不可行。因此,DeepSeek 将通过提取和贡献模块化、可重用的组件以及共享设计优化,积极与现有的开源项目合作。DeepSeek 的目标是培育一个同步的生态系统,通过在新的模型发布之前同步推理相关的工程工作,使得尖端的 AI 能力能够在各种硬件平台上无缝实施。这种方法确保了对开源生态系统的可持续贡献,并促进了其进步的广泛普及。

Hacker News 最新 | 往期 | 评论 | 提问 | 展示 | 招聘 | 提交 登录 开源 DeepSeek 推理引擎之路 (github.com/deepseek-ai) 14 分 by Palmik 27 分钟前 | 隐藏 | 往期 | 收藏 | 讨论 加入我们,参加 6 月 16-17 日在旧金山举办的 AI 初创公司学校! 指南 | 常见问题 | 列表 | API | 安全 | 法律 | 申请 YC | 联系我们 搜索:

原文

A few weeks ago, during Open Source Week, we open-sourced several libraries. The response from the community has been incredibly positive - sparking inspiring collaborations, productive discussions, and valuable bug fixes. Encouraged by this, we’ve decided to take another step forward: contributing our internal inference engine back to the open-source community.

We are deeply grateful for the open-source ecosystem, without which our progress toward AGI would not be possible. Our training framework relies on PyTorch, and our inference engine is built upon vLLM, both of which have been instrumental in accelerating the training and deployment of DeepSeek models.

Given the growing demand for deploying models like DeepSeek-V3 and DeepSeek-R1, we want to give back to the community as much as we can. While we initially considered open-sourcing our full internal inference engine, we identified several challenges:

  • Codebase Divergence: Our engine is based on an early fork of vLLM from over a year ago. Although structurally similar, we’ve heavily customized it for DeepSeek models, making it difficult to extend for broader use cases.
  • Infrastructure Dependencies: The engine is tightly coupled with our internal infrastructure, including cluster management tools, making it impractical for public deployment without significant modifications.
  • Limited Maintenance Bandwidth: As a small research team focused on developing better models, we lack bandwidth to maintain a large open-source project.

Considering these challenges, we’ve decided to collaborate with existing open-source projects as more sustainable alternatives.

Moving forward, we will work closely with existing open-source projects to:

  • Extract Standalone Features: Modularize and contribute reusable components as independent libraries.
  • Share Optimizations: Contribute design improvements and implementation details directly.

We are profoundly grateful for the open-source movement - from operating systems and programming languages to machine learning frameworks and inference engines. It’s an honor to contribute to this thriving ecosystem and to see our models and code embraced by the community. Together, let’s push the boundaries of AGI and ensure its benefits serve all of humanity.

Note

To clarify, this article outlines our approach to open-sourcing of our DeepSeek-Inference-Engine codebase only. Regarding future model releases, we maintain an open and collaborative stance towards both the open-source community and hardware partners. We commit to proactively synchronizing inference-related engineering efforts prior to new model launches, with the goal of enabling the community to achieve state-of-the-art (SOTA) support from Day-0. Our ultimate aim is to foster a synchronized ecosystem where cutting-edge AI capabilities can be seamlessly implemented across diverse hardware platforms upon official model releases.

联系我们 contact @ memedata.com