From 3f71ba406a6cbacd73b56e09389d416c5f9428c2 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?M=C3=A9lony=20QIN?= <4621560+cloudmelon@users.noreply.github.com> Date: Tue, 7 May 2024 18:53:08 +0200 Subject: [PATCH] Correct the kubernetes terminology (#3843) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * add details on kubernetes deployment and separate the testing process * Update examples/kubernetes/README.md thanks for suggesting this change, I agree with you and let's make this project better together ! Co-authored-by: JonZeolla --------- Co-authored-by: QIN Mélony Co-authored-by: JonZeolla --- examples/kubernetes/README.md | 26 ++++++++++++++------------ 1 file changed, 14 insertions(+), 12 deletions(-) diff --git a/examples/kubernetes/README.md b/examples/kubernetes/README.md index c522ba76..2e2444c7 100644 --- a/examples/kubernetes/README.md +++ b/examples/kubernetes/README.md @@ -7,12 +7,24 @@ ## Steps -1. Create the Ollama namespace, daemon set, and service +1. Create the Ollama namespace, deployment, and service ```bash kubectl apply -f cpu.yaml ``` +## (Optional) Hardware Acceleration + +Hardware acceleration in Kubernetes requires NVIDIA's [`k8s-device-plugin`](https://github.com/NVIDIA/k8s-device-plugin) which is deployed in Kubernetes in form of daemonset. Follow the link for more details. + +Once configured, create a GPU enabled Ollama deployment. + +```bash +kubectl apply -f gpu.yaml +``` + +## Test + 1. Port forward the Ollama service to connect and use it locally ```bash @@ -23,14 +35,4 @@ ```bash ollama run orca-mini:3b - ``` - -## (Optional) Hardware Acceleration - -Hardware acceleration in Kubernetes requires NVIDIA's [`k8s-device-plugin`](https://github.com/NVIDIA/k8s-device-plugin). Follow the link for more details. - -Once configured, create a GPU enabled Ollama deployment. - -```bash -kubectl apply -f gpu.yaml -``` + ``` \ No newline at end of file