You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be nice to have a simple example of this proxying to ollama instead of openai. Ollama itself could be a pod or an external service (e.g. in k3s back to the docker host). If serving in the example, we could use a small model like qwen2.5:0.5b to keep the inference lag down.
The text was updated successfully, but these errors were encountered:
It would be nice to have a simple example of this proxying to ollama instead of openai. Ollama itself could be a pod or an external service (e.g. in k3s back to the docker host). If serving in the example, we could use a small model like qwen2.5:0.5b to keep the inference lag down.
The text was updated successfully, but these errors were encountered: