-
Notifications
You must be signed in to change notification settings - Fork 299
feat: add LLM-D profile for E2E testing framework #705
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
✅ Deploy Preview for vllm-semantic-router ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
👥 vLLM Semantic Team NotificationThe following members have been identified for the changed files in this PR and have been automatically assigned: 📁
|
5f82dc9 to
f3998d5
Compare
|
@Xunzhuo It's ready for review. |
|
Cool, a quick question, any test cases for features combine VSR features and llm-d ? |
|
@Xunzhuo Yes
|
|
/gemini summary |
|
My question is looking at the testcases they are around the functionality of llm-d, that is good but the integration test should add more cases around vsr + llm-d like verifying model auto selection + llm-d, ideally should reuse the shared test cases |
|
@Xunzhuo Thanks for the review. I will add VSR + LLM-D integration tests, reuse existing VSR auto-routing cases, verify selected-model consistency with the LLM-D backend, and include the failover path. |
|
update: all use shared testcase. ================================================================================
TEST RESULTS
================================================================================
✅ PASSED - chat-completions-request (2.143574634s)
✅ PASSED - chat-completions-stress-request (20.746133879s)
✅ PASSED - chat-completions-progressive-stress (22.525980752s)
✅ PASSED - domain-classify (1m15.987869552s)
================================================================================
Total: 4 | Passed: 4 | Failed: 0
================================================================================
[Runner] 📝 Collecting semantic-router logs...
[Runner] ✅ Semantic router logs saved to: semantic-router-logs.txt
[Runner] ✅ All tests passed! |
Signed-off-by: samzong <[email protected]>
|
@Xunzhuo Now it has been processed, llmd_* testcases have been removed. all cases have been reused, and has test passed. |
…stcases Signed-off-by: samzong <[email protected]>
|
very cool. Thanks |
Signed-off-by: samzong <[email protected]>

FIX #658 (link existing issues this PR will resolve)
LLM-D profile e2e
Summary
Add an LLM‑D e2e profile to validate LLM-based decision routing in a standard K8s/Kind environment, with focus on correctness, resilience, and basic observability.
What's changed
LLM-Dprofile undere2e/profiles/llm-d/(install, config, health checks, cleanup), selectable viaE2E_PROFILE=llm-d.LLM-Dprofile ine2e/cmd/e2e/main.goso it is managed through the unified e2e entrypoint.llm-ddeployment..github/workflows/integration-test-k8s.ymlto includeE2E_PROFILE=llm-din the k8s integration test matrix.tools/make/e2e.mkande2e/README.mdwith LLM-D usage, examples, and basic troubleshooting.How to run
make e2e-setup E2E_PROFILE=llm-dmake e2e-test E2E_PROFILE=llm-dmake e2e-test E2E_PROFILE=llm-d E2E_VERBOSE=1