How to pass vLLM inference server error messages to GUI / open-webui users? #10857
Unanswered
c-hoffmann
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I run vLLM and open-webui, both installed via pip instead of docker.
My issue is that vLLM error messages don't get passed through to the open-webui user as expected. An easy way to show this is to have a conversation that exceeds the max_model_len / context length. Here's my vLLM output when this occurs:
And here's what open-webui shows:
How can I pass the issue to the GUI more appropriately? Perhaps in this case telling the user to start a new conversation?
Beta Was this translation helpful? Give feedback.
All reactions