From 09d8bf6730b5a3cd5449166b4299168c6bd4cb98 Mon Sep 17 00:00:00 2001 From: Bruce MacDonald Date: Wed, 9 Aug 2023 10:45:57 -0400 Subject: [PATCH] fix build errors --- llama/llama.go | 2 +- server/routes.go | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/llama/llama.go b/llama/llama.go index 2c7df1da..f8c897d4 100644 --- a/llama/llama.go +++ b/llama/llama.go @@ -416,7 +416,7 @@ func (llm *LLM) Embedding(input string) ([]float64, error) { return nil, errors.New("llama: embedding not enabled") } - tokens := llm.tokenize(input) + tokens := llm.Encode(input) if tokens == nil { return nil, errors.New("llama: tokenize embedding") } diff --git a/server/routes.go b/server/routes.go index e5903a86..9182de3b 100644 --- a/server/routes.go +++ b/server/routes.go @@ -87,13 +87,13 @@ func GenerateHandler(c *gin.Context) { } if opts.NumKeep < 0 { - promptWithSystem, err := model.Prompt(api.GenerateRequest{}) + promptWithSystem, err := model.Prompt(api.GenerateRequest{}, "") if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } - promptNoSystem, err := model.Prompt(api.GenerateRequest{Context: []int{0}}) + promptNoSystem, err := model.Prompt(api.GenerateRequest{Context: []int{0}}, "") if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return