Skip to content

Conversation

@junya-takayama
Copy link
Collaborator

OpenAIChatAPI uses longer retry backoff durations to suit commercial services.
However, VLLMServeLM runs on a self-hosted server, so shorter wait times are sufficient.
This PR reduces the retry wait time for VLLMServeLM.

@junya-takayama junya-takayama force-pushed the eliminate_wait_times_on_VLLMServeLM branch from a05cb1b to ddf90c1 Compare December 25, 2025 09:09
@junya-takayama junya-takayama requested a review from a team December 25, 2025 09:24
@junya-takayama junya-takayama merged commit 3b956be into main Dec 26, 2025
7 of 8 checks passed
@junya-takayama junya-takayama deleted the eliminate_wait_times_on_VLLMServeLM branch December 26, 2025 03:16
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants