Skip to content

Commit 491a993

Browse files
committed
Tell Ollama to not unload models automatically
1 parent 8446f03 commit 491a993

File tree

1 file changed

+6
-4
lines changed

1 file changed

+6
-4
lines changed

backend/internal/engine/ollama/engine_ollama.go

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -40,9 +40,10 @@ type chatMessage struct {
4040
}
4141

4242
type chatPayload struct {
43-
Model string `json:"model"`
44-
Messages []chatMessage `json:"messages"`
45-
Options optionsPayload `json:"options"`
43+
Model string `json:"model"`
44+
Messages []chatMessage `json:"messages"`
45+
Options optionsPayload `json:"options"`
46+
KeepAlive int `json:"keep_alive"`
4647
}
4748

4849
type chatResponse struct {
@@ -74,7 +75,8 @@ func (this *OllamaEngineBackend) Process(work *types.Request, model *data.Model,
7475
log.Printf("OllamaEngineBackend Process(): Temperature: %f, TopP: %f\n", preset.Temperature, preset.TopP)
7576
previousMessages := work.Messages[0 : len(work.Messages)-1]
7677
payload := &chatPayload{
77-
Model: model.InternalModelID,
78+
Model: model.InternalModelID,
79+
KeepAlive: -1,
7880
Messages: slices.Map(previousMessages, func(m data.Message) chatMessage {
7981
mRole := "user"
8082
if m.Role == role.Assistant {

0 commit comments

Comments
 (0)