From b6fc01a34b3e574ef12886eff2a52b9864081202 Mon Sep 17 00:00:00 2001 From: wangxiyuan Date: Tue, 30 Sep 2025 17:53:55 +0800 Subject: [PATCH] Refresh ascend link Signed-off-by: wangxiyuan --- _posts/2025-09-29-deepseek-v3-2.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/_posts/2025-09-29-deepseek-v3-2.md b/_posts/2025-09-29-deepseek-v3-2.md index 9ef9fee..4702c3b 100644 --- a/_posts/2025-09-29-deepseek-v3-2.md +++ b/_posts/2025-09-29-deepseek-v3-2.md @@ -102,7 +102,7 @@ We are also excited to announce out-of-the-box Blackwell support for this model. We are barely touching the surface of the optimization for DSA and related sparse attention in vLLM. In the coming weeks: * We plan to expand the architectures supported beyond Hopper and Blackwell. -* We will expand the support to other hardwares such as AMD and TPU. With vLLM's extensible systems, developers can add support for models directly. For example, [vllm-ascend](https://github.com/vllm-project/vllm-ascend/tree/v0.9.1-dev/examples/deepseek.md) and [vllm-mlu](https://github.com/Cambricon/vllm-mlu) already support DeepSeek V3.2! +* We will expand the support to other hardwares such as AMD and TPU. With vLLM's extensible systems, developers can add support for models directly. For example, [vllm-ascend](https://github.com/vllm-project/vllm-ascend/releases/tag/v0.11.0rc0) and [vllm-mlu](https://github.com/Cambricon/vllm-mlu) already support DeepSeek V3.2! * We continuously test large-scale wide EP serving and disaggregation. * You will soon be able to run an end-to-end RL loop with this model. * We will explore the "masked MHA mode for short sequence prefilling" from DeepSeek.