From 33a65e3ba3ad5666d6ba8430efbccfa6d642d1de Mon Sep 17 00:00:00 2001 From: Josh Yan Date: Mon, 1 Jul 2024 16:04:13 -0700 Subject: [PATCH] error --- llm/server.go | 3 +++ llm/status.go | 1 + 2 files changed, 4 insertions(+) diff --git a/llm/server.go b/llm/server.go index 61346069..8b63cfbd 100644 --- a/llm/server.go +++ b/llm/server.go @@ -560,6 +560,9 @@ func (s *llmServer) WaitUntilRunning(ctx context.Context) error { if s.status != nil && s.status.LastErrMsg != "" { msg = s.status.LastErrMsg } + if strings.Contains(msg, "unknown model") { + return fmt.Errorf("this model is not supported by your version of Ollama. You may need to upgrade") + } return fmt.Errorf("llama runner process has terminated: %v %s", err, msg) default: } diff --git a/llm/status.go b/llm/status.go index 8a49bd55..0f56b7f9 100644 --- a/llm/status.go +++ b/llm/status.go @@ -25,6 +25,7 @@ var errorPrefixes = []string{ "CUDA error", "cudaMalloc failed", "\"ERR\"", + "architecture", } func (w *StatusWriter) Write(b []byte) (int, error) {