Merge pull request #270 from jmorganca/update-llama-cpp

update llama.cpp
This commit is contained in:
Michael Yang 2023-08-03 12:09:00 -07:00 committed by GitHub
commit 85aeb42869
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
19 changed files with 623 additions and 298 deletions

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

File diff suppressed because it is too large Load diff

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build darwin
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build darwin
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build darwin
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build mpi
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build mpi
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build opencl
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,7 +1,7 @@
//go:build opencl
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*

View file

@ -128,11 +128,6 @@ func New(model string, opts api.Options) (*LLM, error) {
C.llama_backend_init(C.bool(llm.UseNUMA))
// TODO: GQA == 8 suggests 70B model which doesn't support metal
if llm.NumGQA == 8 {
llm.NumGPU = 0
}
params := C.llama_context_default_params()
params.seed = C.uint(llm.Seed)
params.n_ctx = C.int(llm.NumCtx)

View file

@ -1,5 +1,5 @@
/**
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
* llama.cpp - git 8183159cf3def112f6d1fe94815fce70e1bffa12
*
* MIT License
*