mirror of
https://github.com/deepseek-ai/open-infra-index.git
synced 2025-05-13 23:13:16 -04:00
blog: 202504 The Path to Open-Sourcing the DeepSeek Inference Engine (DeepSeek-vLLM))
This commit is contained in:
parent
b34890e010
commit
72a0be9034
2 changed files with 53 additions and 0 deletions
51
OpenSourcing_DeepSeek-vLLM/README.md
Normal file
51
OpenSourcing_DeepSeek-vLLM/README.md
Normal file
|
@ -0,0 +1,51 @@
|
||||||
|
# The Path to Open-Sourcing the DeepSeek Inference Engine (DeepSeek-vLLM)
|
||||||
|
|
||||||
|
A few weeks ago,
|
||||||
|
during [Open Source Week](https://github.com/deepseek-ai/open-infra-index?tab=readme-ov-file#202502-open-source-week),
|
||||||
|
we open-sourced several libraries.
|
||||||
|
The response from the community has been incredibly positive - sparking inspiring collaborations, productive
|
||||||
|
discussions, and valuable bug fixes.
|
||||||
|
Encouraged by this, we’ve decided to take another step forward: contributing our internal inference engine back to the
|
||||||
|
open-source community.
|
||||||
|
|
||||||
|
We are deeply grateful for the open-source ecosystem, without which our progress toward AGI would not be possible.
|
||||||
|
Our training framework relies on [PyTorch](https://github.com/pytorch/pytorch), and our inference engine is built
|
||||||
|
upon [vLLM](https://github.com/pytorch/pytorch),
|
||||||
|
both of which have ben instrumental in accelerating the training and deployment of DeepSeek models.
|
||||||
|
|
||||||
|
Given the growing demand for deploying models like [DeepSeek-V3](https://github.com/deepseek-ai/DeepSeek-V3)
|
||||||
|
and [DeepSeek-R1](https://github.com/deepseek-ai/DeepSeek-R1), we want to give back to the community as much as we can.
|
||||||
|
While we initially considered open-sourcing our full internal inference engine, we identified several challenges:
|
||||||
|
|
||||||
|
- **Codebase Divergence**: Our engine is based on an early fork of vLLM from over a year ago. Although structurally
|
||||||
|
similar, we’ve heavily customized it for DeepSeek models, making it difficult to extend for broader use cases.
|
||||||
|
- **Infrastructure Dependencies**: The engine is tightly coupled with our internal infrastructure, including cluster
|
||||||
|
management tools, making it impractical for public deployment without significant modifications.
|
||||||
|
- **Limited Maintenance Bandwidth**: As a small research team focused on developing better models, we lack bandwidth to
|
||||||
|
maintain a large open-source project.
|
||||||
|
|
||||||
|
Considering these challenges, we’ve decided to collaborate with the vLLM project as a more sustainable alternative.
|
||||||
|
vLLM stands out for its broad hardware support, extensive model compatibility, and state-of-the-art performance.
|
||||||
|
Its modular architecture gives us confidence that we can adapt and contribute key components from our internal engine to
|
||||||
|
benefit the wider community.
|
||||||
|
|
||||||
|
Moving forward, we will work closely with the vLLM team to:
|
||||||
|
|
||||||
|
- **Extract Standalone Features**: Modularize and contribute reusable components as independent libraries.
|
||||||
|
- **Share Optimizations**: Contribute design improvements and implementation details directly to vLLM.
|
||||||
|
|
||||||
|
We are profoundly grateful for the open-source movement - from operating systems and programming languages to machine
|
||||||
|
learning frameworks and inference engines. It’s an honor to contribute to this thriving ecosystem and to see our models
|
||||||
|
and code embraced by the community. Together, let’s push the boundaries of AGI and ensure its benefits serve all of
|
||||||
|
humanity.
|
||||||
|
|
||||||
|
This page will serve as a central hub for tracking our related open-source efforts on the inference engine.
|
||||||
|
|
||||||
|
> [!NOTE]
|
||||||
|
> **To clarify, this article outlines our approach to open-sourcing of our DeepSeek-vLLM codebase only.
|
||||||
|
> Regarding future model releases, we maintain an open and collaborative stance towards both the open-source community
|
||||||
|
> and hardware partners.
|
||||||
|
> We commit to proactively synchronizing inference-related engineering efforts prior to new model launches, with the
|
||||||
|
> goal of enabling the community to achieve state-of-the-art (SOTA) support from Day-0. Our ultimate aim is to foster a
|
||||||
|
> synchronized ecosystem where cutting-edge AI capabilities can be seamlessly implemented across diverse hardware
|
||||||
|
> platforms upon official model releases.**
|
|
@ -9,6 +9,8 @@
|
||||||
|
|
||||||
# Hello, DeepSeek Open Infra!
|
# Hello, DeepSeek Open Infra!
|
||||||
|
|
||||||
|
## 202504 [The Path to Open-Sourcing the DeepSeek Inference Engine (DeepSeek-vLLM))](OpenSourcing_DeepSeek-vLLM/README.md)
|
||||||
|
|
||||||
## 202502 Open-Source Week
|
## 202502 Open-Source Week
|
||||||
We're a tiny team @deepseek-ai pushing our limits in AGI exploration.
|
We're a tiny team @deepseek-ai pushing our limits in AGI exploration.
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue