[Feat]: Adding vllm support. #115
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
sets up a vllm proxy to a locally running ollama server. configures a connection and runs tests against it.
removes opentelemetry dumps to console
removes progress pushing state to the console both of which were polluting the logs
Added "HOSTED_VLLM_API_BASE" in UI key for vllm support.
https://docs.litellm.ai/docs/providers/vllm#embeddings
TODO
Steps to run the vllm proxy to ollama locally
cd examples/vllm_server python vllm_ollama_server.pyTested by running the command:
or
or
Screenshots/Results
Reference Links
#110
Checklist
This PR includes the following (tick all that apply):