Autocomplete not working with VLLM Backend #8617
Unanswered
tg-arraylabs
asked this question in
Help
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I'm self hosting VLLM using Qwen/Qwen3-Coder-30B-A3B-Instruct as my model. I'm able to use this model without issue for the chat role, but autocomplete doesn't work. I can see the autocomplete requests getting POSTed to VLLM, but nothing comes back in my IDE.
Any advice would be helpful.
Here is my continue.dev configuration:
VLLM configuration:
Beta Was this translation helpful? Give feedback.
All reactions