docs(README): replace llama.cpp link with python bindings
This commit is contained in:
parent
4891468c1a
commit
8926969265
|
@ -27,7 +27,7 @@ Note that **llm-ls** does not export any data anywhere (other than setting a use
|
||||||
|
|
||||||
### Multiple backends
|
### Multiple backends
|
||||||
|
|
||||||
**llm-ls** is compatible with Hugging Face's [Inference API](https://huggingface.co/docs/api-inference/en/index), Hugging Face's [text-generation-inference](https://github.com/huggingface/text-generation-inference), [ollama](https://github.com/ollama/ollama) and OpenAI compatible APIs, like [llama.cpp](https://github.com/ggerganov/llama.cpp/tree/master/examples/server).
|
**llm-ls** is compatible with Hugging Face's [Inference API](https://huggingface.co/docs/api-inference/en/index), Hugging Face's [text-generation-inference](https://github.com/huggingface/text-generation-inference), [ollama](https://github.com/ollama/ollama) and OpenAI compatible APIs, like the [python llama.cpp server bindings](https://github.com/abetlen/llama-cpp-python?tab=readme-ov-file#openai-compatible-web-server).
|
||||||
|
|
||||||
## Compatible extensions
|
## Compatible extensions
|
||||||
|
|
||||||
|
|
|
@ -60,6 +60,10 @@ pub enum Backend {
|
||||||
#[serde(default = "hf_default_url")]
|
#[serde(default = "hf_default_url")]
|
||||||
url: String,
|
url: String,
|
||||||
},
|
},
|
||||||
|
// TODO:
|
||||||
|
// LlamaCpp {
|
||||||
|
// url: String,
|
||||||
|
// },
|
||||||
Ollama {
|
Ollama {
|
||||||
url: String,
|
url: String,
|
||||||
},
|
},
|
||||||
|
|
Loading…
Reference in a new issue