ollama/llm
Daniel Hiltgen 39928a42e8 Always dynamically load the llm server library
This switches darwin to dynamic loading, and refactors the code now that no
static linking of the library is used on any platform
2024-01-11 08:42:47 -08:00
..
ext_server Support multiple variants for a given llm lib type 2024-01-10 17:27:51 -08:00
generate Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
llama.cpp@328b83de23 revert submodule back to 328b83de23b33240e28f4e74900d1d06726f5eb1 2024-01-10 18:42:39 -05:00
dyn_ext_server.c Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
dyn_ext_server.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
dyn_ext_server.h Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
ggml.go Offload layers to GPU based on new model size estimates (#1850) 2024-01-08 16:42:00 -05:00
gguf.go Offload layers to GPU based on new model size estimates (#1850) 2024-01-08 16:42:00 -05:00
llama.go Offload layers to GPU based on new model size estimates (#1850) 2024-01-08 16:42:00 -05:00
llm.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
payload_common.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
payload_darwin.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
payload_linux.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
payload_test.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
payload_windows.go Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00