ollama/README.md
Jeffrey Morgan 76cb60d496 wip go engine
Co-authored-by: Patrick Devine <pdevine@sonic.net>
2023-07-06 16:34:44 -04:00

45 lines
750 B
Markdown

# Ollama
A fast runtime for large language models, powered by [llama.cpp](https://github.com/ggerganov/llama.cpp).
> _Note: this project is a work in progress. Certain models that can be run with `ollama` are intended for research and/or non-commercial use only._
## Install
Using `pip`:
```
pip install ollama
```
Using `docker`:
```
docker run ollama/ollama
```
## Quickstart
To run a model, use `ollama run`:
```
ollama run orca-mini-3b
```
You can also run models from hugging face:
```
ollama run huggingface.co/TheBloke/orca_mini_3B-GGML
```
Or directly via downloaded model files:
```
ollama run ~/Downloads/orca-mini-13b.ggmlv3.q4_0.bin
```
## Documentation
- [Development](docs/development.md)
- [Python SDK](docs/python.md)