2023-06-27 12:08:52 -04:00
|
|
|
|
# Ollama
|
2023-06-22 12:45:31 -04:00
|
|
|
|
|
2023-06-27 17:26:27 -04:00
|
|
|
|
The easiest way to run ai models.
|
2023-06-27 17:13:07 -04:00
|
|
|
|
|
|
|
|
|
## Download
|
|
|
|
|
|
2023-06-27 22:05:41 -04:00
|
|
|
|
- [macOS](https://ollama.ai/download/darwin_arm64) (Apple Silicon)
|
2023-06-27 17:13:07 -04:00
|
|
|
|
- macOS (Intel – Coming soon)
|
|
|
|
|
- Windows (Coming soon)
|
|
|
|
|
- Linux (Coming soon)
|
|
|
|
|
|
|
|
|
|
## Python SDK
|
2023-06-22 12:45:31 -04:00
|
|
|
|
|
|
|
|
|
```
|
2023-06-27 12:08:52 -04:00
|
|
|
|
pip install ollama
|
2023-06-22 12:45:31 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 17:13:07 -04:00
|
|
|
|
### Python SDK quickstart
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```python
|
|
|
|
|
import ollama
|
2023-06-27 12:48:50 -04:00
|
|
|
|
ollama.generate("./llama-7b-ggml.bin", "hi")
|
2023-06-25 13:08:03 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
### `ollama.generate(model, message)`
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
Generate a completion
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
|
|
|
|
```python
|
2023-06-27 12:51:36 -04:00
|
|
|
|
ollama.generate("./llama-7b-ggml.bin", "hi")
|
2023-06-25 13:08:03 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
### `ollama.load(model)`
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
Load a model for generation
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```python
|
2023-06-27 12:56:53 -04:00
|
|
|
|
ollama.load("model")
|
2023-06-25 13:08:03 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
### `ollama.models()`
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:44:12 -04:00
|
|
|
|
List available local models
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
|
|
|
|
```
|
|
|
|
|
models = ollama.models()
|
2023-06-25 13:08:03 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
### `ollama.serve()`
|
2023-06-25 13:10:15 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
Serve the ollama http server
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 17:36:02 -04:00
|
|
|
|
### `ollama.add(filepath)`
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 17:36:02 -04:00
|
|
|
|
Add a model by importing from a file
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
|
|
|
|
```python
|
2023-06-27 17:36:02 -04:00
|
|
|
|
ollama.add("./path/to/model")
|
2023-06-25 13:08:03 -04:00
|
|
|
|
```
|
|
|
|
|
|
2023-06-27 17:36:02 -04:00
|
|
|
|
## Cooming Soon
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
2023-06-27 17:36:02 -04:00
|
|
|
|
### `ollama.pull(model)`
|
|
|
|
|
|
|
|
|
|
Download a model
|
2023-06-27 12:08:52 -04:00
|
|
|
|
|
|
|
|
|
```python
|
2023-06-27 17:36:02 -04:00
|
|
|
|
ollama.pull("huggingface.co/thebloke/llama-7b-ggml")
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:51:36 -04:00
|
|
|
|
### `ollama.search("query")`
|
2023-06-25 14:29:26 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
Search for compatible models that Ollama can run
|
2023-06-25 14:29:26 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```python
|
|
|
|
|
ollama.search("llama-7b")
|
|
|
|
|
```
|
2023-06-25 13:08:03 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
## Future CLI
|
2023-06-25 14:29:26 -04:00
|
|
|
|
|
2023-06-27 13:51:20 -04:00
|
|
|
|
In the future, there will be an `ollama` CLI for running models on servers, in containers or for local development environments.
|
2023-06-27 12:44:12 -04:00
|
|
|
|
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```
|
2023-06-27 13:52:32 -04:00
|
|
|
|
ollama generate huggingface.co/thebloke/llama-7b-ggml "hi"
|
2023-06-27 12:44:12 -04:00
|
|
|
|
> Downloading [================> ] 66.67% (2/3) 30.2MB/s
|
2023-06-27 12:08:52 -04:00
|
|
|
|
```
|
2023-06-27 13:46:46 -04:00
|
|
|
|
|
|
|
|
|
## Documentation
|
|
|
|
|
|
|
|
|
|
- [Development](docs/development.md)
|