Commit graph

161 commits

Author SHA1 Message Date
Bruce MacDonald 7cf5905063 display pull progress 2023-07-06 16:34:44 -04:00
Michael Yang 580fe8951c free llama model 2023-07-06 16:34:44 -04:00
Michael Yang 68e6b4550c use prompt templates 2023-07-06 16:34:44 -04:00
Bruce MacDonald a6494f8211 pull models 2023-07-06 16:34:44 -04:00
Michael Yang 1b7183c5a1 enable metal gpu acceleration
ggml-metal.metal must be in the same directory as the ollama binary
otherwise llama.cpp will not be able to find it and load it.

1. go generate llama/llama_metal.go
2. go build .
3. ./ollama serve
2023-07-06 16:34:44 -04:00
Jeffrey Morgan 0998d4f0a4 remove debug print statements 2023-07-06 16:34:44 -04:00
Bruce MacDonald 8ea5e5e147 separate routes 2023-07-06 16:34:44 -04:00
Jeffrey Morgan fd962a36e5 client updates 2023-07-06 16:34:44 -04:00
Jeffrey Morgan 9164981d72 move prompt templates out of python bindings 2023-07-06 16:34:44 -04:00
Jeffrey Morgan 6093a88c1a add llama.cpp go bindings 2023-07-06 16:34:44 -04:00
Jeffrey Morgan 76cb60d496 wip go engine
Co-authored-by: Patrick Devine <pdevine@sonic.net>
2023-07-06 16:34:44 -04:00