From d4ebdadbe79a925e25ad63f980d18dafc2434626 Mon Sep 17 00:00:00 2001 From: Jeffrey Morgan Date: Wed, 27 Dec 2023 14:23:42 -0500 Subject: [PATCH] enable `cache_prompt` by default --- api/types.go | 1 - llm/ext_server.go | 4 ++-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/api/types.go b/api/types.go index 7662599a..47f0e124 100644 --- a/api/types.go +++ b/api/types.go @@ -103,7 +103,6 @@ type Options struct { MirostatEta float32 `json:"mirostat_eta,omitempty"` PenalizeNewline bool `json:"penalize_newline,omitempty"` Stop []string `json:"stop,omitempty"` - Cache bool `json:"cache,omitempty"` } // Runner options which must be set when the model is loaded into memory diff --git a/llm/ext_server.go b/llm/ext_server.go index 5a760e3e..0f7d441a 100644 --- a/llm/ext_server.go +++ b/llm/ext_server.go @@ -234,8 +234,8 @@ func predict(llm extServer, opts api.Options, ctx context.Context, predict Predi "penalize_nl": opts.PenalizeNewline, "seed": opts.Seed, "stop": opts.Stop, - "image_data": imageData, - "cache_prompt": opts.Cache, + "image_data": imageData, + "cache_prompt": true, } if predict.Format == "json" {