Separate profile for Autocomplete #5109
isupovs
started this conversation in
1. Feature requests
Replies: 2 comments 1 reply
-
|
@isupovs you should be able to do this- here's an instructional video: https://www.youtube.com/watch?v=iNioZeF4-yY |
Beta Was this translation helpful? Give feedback.
1 reply
-
|
Yes, that's fair |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have two computers running local LLMs. On the first machine, I run a large model that occupies all available VRAM.
As I understand it, the Autocomplete feature currently supports only Mistral-based models, which is fine. However, I would like to run Autocomplete on a second machine, while keeping the main assistant on the first one.
At the moment, there is no way to configure Autocomplete to use a separate profile or a different endpoint/host. This makes it impossible to split workloads between machines.
Beta Was this translation helpful? Give feedback.
All reactions