Skip to content

Conversation

mgoin
Copy link
Member

@mgoin mgoin commented Oct 1, 2025

Purpose

Address #26042 for now to unblock the release. We will take a performance regression on B200 due to falling back to FA2 prefill, but we need correctness first.

Test Plan

Test Result

Base commands:
vllm serve deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct
python tests/evals/gsm8k/gsm8k_eval.py

# Hopper

vllm serve deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct
Accuracy: 0.794

# Blackwell

## Default backends (FlashInfer prefill and CUTLASS MLA decode)
vllm serve deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct
Accuracy: 0.216

## Using FA2 prefill and CUTLASS MLA decode
VLLM_DISABLE_FLASHINFER_PREFILL=1 vllm serve deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct
Accuracy: 0.785

Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates the default value for VLLM_DISABLE_FLASHINFER_PREFILL to True, effectively disabling FlashInfer MLA prefill by default to address issues with chunked prefill. The change is applied consistently to both the type-hinted default and the environment variable parsing logic. While the change itself is correct, I've identified a related potential issue: this environment variable, which likely influences the computation graph, is not included in the cache key computation. This could lead to incorrect cache hits if the flag is toggled.

@pavanimajety
Copy link
Contributor

@simon-mo simon-mo added this to the v0.11.0 Cherry Picks milestone Oct 1, 2025
@mgoin mgoin added bug Something isn't working ready ONLY add when PR is ready to merge/full CI is needed deepseek Related to DeepSeek models labels Oct 1, 2025
@mgoin
Copy link
Member Author

mgoin commented Oct 1, 2025

Unfortunately the FA2 backend in flashinfer still has the same issue, so it is likely vLLM preparing inputs improperly for flashinfer

Copy link
Collaborator

@LucasWilkinson LucasWilkinson left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

use: #26063 instead

@LucasWilkinson LucasWilkinson removed this from the v0.11.0 Cherry Picks milestone Oct 2, 2025
@mgoin mgoin closed this Oct 2, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug Something isn't working deepseek Related to DeepSeek models ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants