diff --git a/.golangci.yaml b/.golangci.yaml index 85558f03..985fad63 100644 --- a/.golangci.yaml +++ b/.golangci.yaml @@ -24,4 +24,4 @@ linters-settings: - (*os.File).Seek - (*bufio.Writer).WriteString - (*github.com/spf13/pflag.FlagSet).Set - - (*github.com/jmorganca/ollama/llm.readSeekOffset).Seek + - (*github.com/ollama/ollama/llm.readSeekOffset).Seek diff --git a/Dockerfile b/Dockerfile index 59cca725..919d48de 100644 --- a/Dockerfile +++ b/Dockerfile @@ -15,8 +15,8 @@ ARG CMAKE_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh @@ -25,8 +25,8 @@ ARG CMAKE_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/gcc-toolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh @@ -36,18 +36,18 @@ COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH ENV LIBRARY_PATH /opt/amdgpu/lib64 -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate ARG CGO_CFLAGS ARG AMDGPU_TARGETS RUN OLLAMA_SKIP_CPU_GENERATE=1 sh gen_linux.sh RUN mkdir /tmp/scratch && \ - for dep in $(cat /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/x86_64/rocm*/lib/deps.txt) ; do \ + for dep in $(cat /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/x86_64/rocm*/lib/deps.txt) ; do \ cp ${dep} /tmp/scratch/ || exit 1 ; \ done && \ (cd /opt/rocm/lib && tar cf - rocblas/library) | (cd /tmp/scratch/ && tar xf - ) && \ - mkdir -p /go/src/github.com/jmorganca/ollama/dist/deps/ && \ - (cd /tmp/scratch/ && tar czvf /go/src/github.com/jmorganca/ollama/dist/deps/ollama-linux-amd64-rocm.tgz . ) + mkdir -p /go/src/github.com/ollama/ollama/dist/deps/ && \ + (cd /tmp/scratch/ && tar czvf /go/src/github.com/ollama/ollama/dist/deps/ollama-linux-amd64-rocm.tgz . ) FROM --platform=linux/amd64 centos:7 AS cpu-builder-amd64 @@ -56,10 +56,10 @@ ARG GOLANG_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} GOLANG_VERSION=${GOLANG_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ ARG OLLAMA_CUSTOM_CPU_DEFS ARG CGO_CFLAGS -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +WORKDIR /go/src/github.com/ollama/ollama/llm/generate FROM --platform=linux/amd64 cpu-builder-amd64 AS cpu-build-amd64 RUN OLLAMA_CPU_TARGET="cpu" sh gen_linux.sh @@ -74,8 +74,8 @@ ARG GOLANG_VERSION COPY ./scripts/rh_linux_deps.sh / RUN CMAKE_VERSION=${CMAKE_VERSION} GOLANG_VERSION=${GOLANG_VERSION} sh /rh_linux_deps.sh ENV PATH /opt/rh/devtoolset-10/root/usr/bin:$PATH -COPY --from=llm-code / /go/src/github.com/jmorganca/ollama/ -WORKDIR /go/src/github.com/jmorganca/ollama/llm/generate +COPY --from=llm-code / /go/src/github.com/ollama/ollama/ +WORKDIR /go/src/github.com/ollama/ollama/llm/generate # Note, we only build the "base" CPU variant on arm since avx/avx2 are x86 features ARG OLLAMA_CUSTOM_CPU_DEFS ARG CGO_CFLAGS @@ -84,13 +84,13 @@ RUN OLLAMA_CPU_TARGET="cpu" sh gen_linux.sh # Intermediate stage used for ./scripts/build_linux.sh FROM --platform=linux/amd64 cpu-build-amd64 AS build-amd64 ENV CGO_ENABLED 1 -WORKDIR /go/src/github.com/jmorganca/ollama +WORKDIR /go/src/github.com/ollama/ollama COPY . . -COPY --from=cpu_avx-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=cpu_avx2-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=cuda-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=rocm-build-amd64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -COPY --from=rocm-build-amd64 /go/src/github.com/jmorganca/ollama/dist/deps/ ./dist/deps/ +COPY --from=cpu_avx-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=cpu_avx2-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=cuda-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=rocm-build-amd64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +COPY --from=rocm-build-amd64 /go/src/github.com/ollama/ollama/dist/deps/ ./dist/deps/ ARG GOFLAGS ARG CGO_CFLAGS RUN go build -trimpath . @@ -99,10 +99,10 @@ RUN go build -trimpath . FROM --platform=linux/arm64 cpu-build-arm64 AS build-arm64 ENV CGO_ENABLED 1 ARG GOLANG_VERSION -WORKDIR /go/src/github.com/jmorganca/ollama +WORKDIR /go/src/github.com/ollama/ollama COPY . . -COPY --from=cuda-build-arm64 /go/src/github.com/jmorganca/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ -RUN mkdir -p /go/src/github.com/jmorganca/ollama/dist/deps/ +COPY --from=cuda-build-arm64 /go/src/github.com/ollama/ollama/llm/llama.cpp/build/linux/ llm/llama.cpp/build/linux/ +RUN mkdir -p /go/src/github.com/ollama/ollama/dist/deps/ ARG GOFLAGS ARG CGO_CFLAGS RUN go build -trimpath . @@ -110,15 +110,15 @@ RUN go build -trimpath . # Runtime stages FROM --platform=linux/amd64 ubuntu:22.04 as runtime-amd64 RUN apt-get update && apt-get install -y ca-certificates -COPY --from=build-amd64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-amd64 /go/src/github.com/ollama/ollama/ollama /bin/ollama FROM --platform=linux/arm64 ubuntu:22.04 as runtime-arm64 RUN apt-get update && apt-get install -y ca-certificates -COPY --from=build-arm64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-arm64 /go/src/github.com/ollama/ollama/ollama /bin/ollama # Radeon images are much larger so we keep it distinct from the CPU/CUDA image FROM --platform=linux/amd64 rocm/dev-centos-7:${ROCM_VERSION}-complete as runtime-rocm RUN update-pciids -COPY --from=build-amd64 /go/src/github.com/jmorganca/ollama/ollama /bin/ollama +COPY --from=build-amd64 /go/src/github.com/ollama/ollama/ollama /bin/ollama EXPOSE 11434 ENV OLLAMA_HOST 0.0.0.0 diff --git a/README.md b/README.md index 48da0e96..f46e479f 100644 --- a/README.md +++ b/README.md @@ -1,5 +1,5 @@
- ollama + ollama
# Ollama @@ -22,7 +22,7 @@ Get up and running with large language models locally. curl -fsSL https://ollama.com/install.sh | sh ``` -[Manual install instructions](https://github.com/jmorganca/ollama/blob/main/docs/linux.md) +[Manual install instructions](https://github.com/ollama/ollama/blob/main/docs/linux.md) ### Docker @@ -213,7 +213,7 @@ Then build the binary: go build . ``` -More detailed instructions can be found in the [developer guide](https://github.com/jmorganca/ollama/blob/main/docs/development.md) +More detailed instructions can be found in the [developer guide](https://github.com/ollama/ollama/blob/main/docs/development.md) ### Running local builds diff --git a/api/client.go b/api/client.go index 36019206..a37a4cbb 100644 --- a/api/client.go +++ b/api/client.go @@ -15,8 +15,8 @@ import ( "runtime" "strings" - "github.com/jmorganca/ollama/format" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/format" + "github.com/ollama/ollama/version" ) type Client struct { diff --git a/app/lifecycle/lifecycle.go b/app/lifecycle/lifecycle.go index 14a85b11..ab624e81 100644 --- a/app/lifecycle/lifecycle.go +++ b/app/lifecycle/lifecycle.go @@ -9,8 +9,8 @@ import ( "os/signal" "syscall" - "github.com/jmorganca/ollama/app/store" - "github.com/jmorganca/ollama/app/tray" + "github.com/ollama/ollama/app/store" + "github.com/ollama/ollama/app/tray" ) func Run() { diff --git a/app/lifecycle/server.go b/app/lifecycle/server.go index 1cb689a2..e3ca22f9 100644 --- a/app/lifecycle/server.go +++ b/app/lifecycle/server.go @@ -11,7 +11,7 @@ import ( "path/filepath" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func getCLIFullPath(command string) string { diff --git a/app/lifecycle/updater.go b/app/lifecycle/updater.go index f26e32af..243bbf22 100644 --- a/app/lifecycle/updater.go +++ b/app/lifecycle/updater.go @@ -18,8 +18,8 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/auth" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/auth" + "github.com/ollama/ollama/version" ) var ( diff --git a/app/main.go b/app/main.go index 57d8b1c1..db829795 100644 --- a/app/main.go +++ b/app/main.go @@ -4,7 +4,7 @@ package main // go build -ldflags="-H windowsgui" . import ( - "github.com/jmorganca/ollama/app/lifecycle" + "github.com/ollama/ollama/app/lifecycle" ) func main() { diff --git a/app/tray/tray.go b/app/tray/tray.go index 47b204d6..20d1c0b5 100644 --- a/app/tray/tray.go +++ b/app/tray/tray.go @@ -4,8 +4,8 @@ import ( "fmt" "runtime" - "github.com/jmorganca/ollama/app/assets" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/assets" + "github.com/ollama/ollama/app/tray/commontray" ) func NewTray() (commontray.OllamaTray, error) { diff --git a/app/tray/tray_nonwindows.go b/app/tray/tray_nonwindows.go index 6c30c3c2..ae5572b2 100644 --- a/app/tray/tray_nonwindows.go +++ b/app/tray/tray_nonwindows.go @@ -5,7 +5,7 @@ package tray import ( "fmt" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/commontray" ) func InitPlatformTray(icon, updateIcon []byte) (commontray.OllamaTray, error) { diff --git a/app/tray/tray_windows.go b/app/tray/tray_windows.go index 8ac4e478..086fc794 100644 --- a/app/tray/tray_windows.go +++ b/app/tray/tray_windows.go @@ -1,8 +1,8 @@ package tray import ( - "github.com/jmorganca/ollama/app/tray/commontray" - "github.com/jmorganca/ollama/app/tray/wintray" + "github.com/ollama/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/wintray" ) func InitPlatformTray(icon, updateIcon []byte) (commontray.OllamaTray, error) { diff --git a/app/tray/wintray/tray.go b/app/tray/wintray/tray.go index 365cfb82..69d4487d 100644 --- a/app/tray/wintray/tray.go +++ b/app/tray/wintray/tray.go @@ -13,7 +13,7 @@ import ( "sync" "unsafe" - "github.com/jmorganca/ollama/app/tray/commontray" + "github.com/ollama/ollama/app/tray/commontray" "golang.org/x/sys/windows" ) diff --git a/cmd/cmd.go b/cmd/cmd.go index 3a0419e4..c5430f6f 100644 --- a/cmd/cmd.go +++ b/cmd/cmd.go @@ -30,12 +30,12 @@ import ( "golang.org/x/exp/slices" "golang.org/x/term" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/progress" - "github.com/jmorganca/ollama/server" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/progress" + "github.com/ollama/ollama/server" + "github.com/ollama/ollama/version" ) func CreateHandler(cmd *cobra.Command, args []string) error { diff --git a/cmd/interactive.go b/cmd/interactive.go index 82e3642a..2dd9ee19 100644 --- a/cmd/interactive.go +++ b/cmd/interactive.go @@ -14,9 +14,9 @@ import ( "github.com/spf13/cobra" "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/progress" - "github.com/jmorganca/ollama/readline" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/progress" + "github.com/ollama/ollama/readline" ) type MultilineState int diff --git a/cmd/interactive_test.go b/cmd/interactive_test.go index 19e43287..8eedf729 100644 --- a/cmd/interactive_test.go +++ b/cmd/interactive_test.go @@ -7,7 +7,7 @@ import ( "github.com/stretchr/testify/assert" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestExtractFilenames(t *testing.T) { diff --git a/cmd/start_darwin.go b/cmd/start_darwin.go index 7e3000f0..82b09ad6 100644 --- a/cmd/start_darwin.go +++ b/cmd/start_darwin.go @@ -7,7 +7,7 @@ import ( "os/exec" "strings" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/cmd/start_default.go b/cmd/start_default.go index 664c2d1f..c9d6137b 100644 --- a/cmd/start_default.go +++ b/cmd/start_default.go @@ -6,7 +6,7 @@ import ( "context" "fmt" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/cmd/start_windows.go b/cmd/start_windows.go index b9a423cf..6024a235 100644 --- a/cmd/start_windows.go +++ b/cmd/start_windows.go @@ -10,7 +10,7 @@ import ( "strings" "syscall" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func startApp(ctx context.Context, client *api.Client) error { diff --git a/convert/convert.go b/convert/convert.go index ba23080c..4768d6f2 100644 --- a/convert/convert.go +++ b/convert/convert.go @@ -16,8 +16,8 @@ import ( "github.com/mitchellh/mapstructure" "google.golang.org/protobuf/proto" - "github.com/jmorganca/ollama/convert/sentencepiece" - "github.com/jmorganca/ollama/llm" + "github.com/ollama/ollama/convert/sentencepiece" + "github.com/ollama/ollama/llm" ) type Params struct { diff --git a/docs/faq.md b/docs/faq.md index 63976ee4..6bd1b340 100644 --- a/docs/faq.md +++ b/docs/faq.md @@ -154,7 +154,7 @@ No. Ollama runs locally, and conversation data does not leave your machine. ## How can I use Ollama in Visual Studio Code? -There is already a large collection of plugins available for VSCode as well as other editors that leverage Ollama. See the list of [extensions & plugins](https://github.com/jmorganca/ollama#extensions--plugins) at the bottom of the main repository readme. +There is already a large collection of plugins available for VSCode as well as other editors that leverage Ollama. See the list of [extensions & plugins](https://github.com/ollama/ollama#extensions--plugins) at the bottom of the main repository readme. ## How do I use Ollama behind a proxy? diff --git a/docs/modelfile.md b/docs/modelfile.md index 7edd85eb..94acbdf6 100644 --- a/docs/modelfile.md +++ b/docs/modelfile.md @@ -113,7 +113,7 @@ FROM llama2 ``` A list of available base models: - + #### Build from a `bin` file diff --git a/docs/openai.md b/docs/openai.md index 5808ae75..b4dc1f21 100644 --- a/docs/openai.md +++ b/docs/openai.md @@ -1,6 +1,6 @@ # OpenAI compatibility -> **Note:** OpenAI compatibility is experimental and is subject to major adjustments including breaking changes. For fully-featured access to the Ollama API, see the Ollama [Python library](https://github.com/ollama/ollama-python), [JavaScript library](https://github.com/ollama/ollama-js) and [REST API](https://github.com/jmorganca/ollama/blob/main/docs/api.md). +> **Note:** OpenAI compatibility is experimental and is subject to major adjustments including breaking changes. For fully-featured access to the Ollama API, see the Ollama [Python library](https://github.com/ollama/ollama-python), [JavaScript library](https://github.com/ollama/ollama-js) and [REST API](https://github.com/ollama/ollama/blob/main/docs/api.md). Ollama provides experimental compatibility with parts of the [OpenAI API](https://platform.openai.com/docs/api-reference) to help connect existing applications to Ollama. diff --git a/examples/langchain-python-rag-privategpt/README.md b/examples/langchain-python-rag-privategpt/README.md index f739df2b..0133fd88 100644 --- a/examples/langchain-python-rag-privategpt/README.md +++ b/examples/langchain-python-rag-privategpt/README.md @@ -1,6 +1,6 @@ # PrivateGPT with Llama 2 uncensored -https://github.com/jmorganca/ollama/assets/3325447/20cf8ec6-ff25-42c6-bdd8-9be594e3ce1b +https://github.com/ollama/ollama/assets/3325447/20cf8ec6-ff25-42c6-bdd8-9be594e3ce1b > Note: this example is a slightly modified version of PrivateGPT using models such as Llama 2 Uncensored. All credit for PrivateGPT goes to Iván Martínez who is the creator of it, and you can find his GitHub repo [here](https://github.com/imartinez/privateGPT). diff --git a/examples/modelfile-mario/readme.md b/examples/modelfile-mario/readme.md index 9df08a67..0d72dddc 100644 --- a/examples/modelfile-mario/readme.md +++ b/examples/modelfile-mario/readme.md @@ -28,7 +28,7 @@ You are Mario from Super Mario Bros, acting as an assistant. What if you want to change its behaviour? - Try changing the prompt -- Try changing the parameters [Docs](https://github.com/jmorganca/ollama/blob/main/docs/modelfile.md) +- Try changing the parameters [Docs](https://github.com/ollama/ollama/blob/main/docs/modelfile.md) - Try changing the model (e.g. An uncensored model by `FROM wizard-vicuna` this is the wizard-vicuna uncensored model ) Once the changes are made, diff --git a/examples/python-json-datagenerator/readme.md b/examples/python-json-datagenerator/readme.md index ec5701be..369fb2a5 100644 --- a/examples/python-json-datagenerator/readme.md +++ b/examples/python-json-datagenerator/readme.md @@ -1,6 +1,6 @@ # JSON Output Example -![llmjson 2023-11-10 15_31_31](https://github.com/jmorganca/ollama/assets/633681/e599d986-9b4a-4118-81a4-4cfe7e22da25) +![llmjson 2023-11-10 15_31_31](https://github.com/ollama/ollama/assets/633681/e599d986-9b4a-4118-81a4-4cfe7e22da25) There are two python scripts in this example. `randomaddresses.py` generates random addresses from different countries. `predefinedschema.py` sets a template for the model to fill in. diff --git a/examples/python-loganalysis/readme.md b/examples/python-loganalysis/readme.md index 60c57217..4be0baaa 100644 --- a/examples/python-loganalysis/readme.md +++ b/examples/python-loganalysis/readme.md @@ -1,6 +1,6 @@ # Log Analysis example -![loganalyzer 2023-11-10 08_53_29](https://github.com/jmorganca/ollama/assets/633681/ad30f1fc-321f-4953-8914-e30e24db9921) +![loganalyzer 2023-11-10 08_53_29](https://github.com/ollama/ollama/assets/633681/ad30f1fc-321f-4953-8914-e30e24db9921) This example shows one possible way to create a log file analyzer. It uses the model **mattw/loganalyzer** which is based on **codebooga**, a 34b parameter model. diff --git a/examples/typescript-functioncalling/readme.md b/examples/typescript-functioncalling/readme.md index 03ff46dc..d29379a0 100644 --- a/examples/typescript-functioncalling/readme.md +++ b/examples/typescript-functioncalling/readme.md @@ -1,6 +1,6 @@ # Function calling -![function calling 2023-11-16 16_12_58](https://github.com/jmorganca/ollama/assets/633681/a0acc247-9746-45ab-b325-b65dfbbee4fb) +![function calling 2023-11-16 16_12_58](https://github.com/ollama/ollama/assets/633681/a0acc247-9746-45ab-b325-b65dfbbee4fb) One of the features added to some models is 'function calling'. It's a bit of a confusing name. It's understandable if you think that means the model can call functions, but that's not what it means. Function calling simply means that the output of the model is formatted in JSON, using a preconfigured schema, and uses the expected types. Then your code can use the output of the model and call functions with it. Using the JSON format in Ollama, you can use any model for function calling. diff --git a/go.mod b/go.mod index 74f75b47..65647fb1 100644 --- a/go.mod +++ b/go.mod @@ -1,4 +1,4 @@ -module github.com/jmorganca/ollama +module github.com/ollama/ollama go 1.22 diff --git a/integration/basic_test.go b/integration/basic_test.go index ce933ffe..926ca52c 100644 --- a/integration/basic_test.go +++ b/integration/basic_test.go @@ -8,7 +8,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestOrcaMiniBlueSky(t *testing.T) { diff --git a/integration/llm_image_test.go b/integration/llm_image_test.go index 0ac0e1e8..94082d6e 100644 --- a/integration/llm_image_test.go +++ b/integration/llm_image_test.go @@ -9,7 +9,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" "github.com/stretchr/testify/require" ) diff --git a/integration/llm_test.go b/integration/llm_test.go index 107b5573..5b060447 100644 --- a/integration/llm_test.go +++ b/integration/llm_test.go @@ -9,7 +9,7 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) // TODO - this would ideally be in the llm package, but that would require some refactoring of interfaces in the server diff --git a/integration/utils_test.go b/integration/utils_test.go index 47184af8..14ec39dc 100644 --- a/integration/utils_test.go +++ b/integration/utils_test.go @@ -21,8 +21,8 @@ import ( "testing" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/app/lifecycle" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/app/lifecycle" "github.com/stretchr/testify/assert" ) diff --git a/llm/dyn_ext_server.go b/llm/dyn_ext_server.go index a6a010d7..6e43333e 100644 --- a/llm/dyn_ext_server.go +++ b/llm/dyn_ext_server.go @@ -33,8 +33,8 @@ import ( "time" "unsafe" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" ) type dynExtServer struct { diff --git a/llm/gguf.go b/llm/gguf.go index 0c7e914f..6cb90484 100644 --- a/llm/gguf.go +++ b/llm/gguf.go @@ -15,7 +15,7 @@ import ( "github.com/pdevine/tensor/native" "github.com/x448/float16" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/format" ) type ContainerGGUF struct { diff --git a/llm/llama.go b/llm/llama.go index a5d2036a..cce9f484 100644 --- a/llm/llama.go +++ b/llm/llama.go @@ -5,7 +5,7 @@ import ( "fmt" "time" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) const jsonGrammar = ` diff --git a/llm/llm.go b/llm/llm.go index a3f59d2e..75ff955e 100644 --- a/llm/llm.go +++ b/llm/llm.go @@ -8,8 +8,8 @@ import ( "runtime" "slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" ) type LLM interface { diff --git a/llm/payload_common.go b/llm/payload_common.go index d45b70c4..45e45a2d 100644 --- a/llm/payload_common.go +++ b/llm/payload_common.go @@ -16,7 +16,7 @@ import ( "golang.org/x/exp/slices" "golang.org/x/sync/errgroup" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/gpu" ) // Libraries names may contain an optional variant separated by '_' diff --git a/llm/payload_test.go b/llm/payload_test.go index 44537b0a..eb88d812 100644 --- a/llm/payload_test.go +++ b/llm/payload_test.go @@ -3,7 +3,7 @@ package llm import ( "testing" - "github.com/jmorganca/ollama/gpu" + "github.com/ollama/ollama/gpu" "github.com/stretchr/testify/assert" ) diff --git a/main.go b/main.go index ecdcb71d..158f0063 100644 --- a/main.go +++ b/main.go @@ -3,7 +3,7 @@ package main import ( "context" - "github.com/jmorganca/ollama/cmd" + "github.com/ollama/ollama/cmd" "github.com/spf13/cobra" ) diff --git a/openai/openai.go b/openai/openai.go index 4f495569..96d8f218 100644 --- a/openai/openai.go +++ b/openai/openai.go @@ -11,7 +11,7 @@ import ( "time" "github.com/gin-gonic/gin" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) type Error struct { diff --git a/progress/bar.go b/progress/bar.go index 9a0da825..476ea888 100644 --- a/progress/bar.go +++ b/progress/bar.go @@ -6,7 +6,7 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/format" "golang.org/x/term" ) diff --git a/scripts/build_darwin.sh b/scripts/build_darwin.sh index cb561123..a2f76af2 100755 --- a/scripts/build_darwin.sh +++ b/scripts/build_darwin.sh @@ -3,7 +3,7 @@ set -e export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" mkdir -p dist diff --git a/scripts/build_docker.sh b/scripts/build_docker.sh index 7bdad641..958733f3 100755 --- a/scripts/build_docker.sh +++ b/scripts/build_docker.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" # We use 2 different image repositories to handle combining architecture images into multiarch manifest # (The ROCm image is x86 only and is not a multiarch manifest) @@ -74,4 +74,4 @@ if [ -z "${OLLAMA_SKIP_MANIFEST_CREATE}" ]; then echo " ${ARCH_IMAGE_REPO}:$VERSION-arm64" echo " ${ARCH_IMAGE_REPO}:$VERSION-rocm" fi -fi \ No newline at end of file +fi diff --git a/scripts/build_linux.sh b/scripts/build_linux.sh index e6db485a..27c4ff1f 100755 --- a/scripts/build_linux.sh +++ b/scripts/build_linux.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-$(git describe --tags --first-parent --abbrev=7 --long --dirty --always | sed -e "s/^v//g")} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" BUILD_ARCH=${BUILD_ARCH:-"amd64 arm64"} export AMDGPU_TARGETS=${AMDGPU_TARGETS:=""} @@ -21,10 +21,10 @@ for TARGETARCH in ${BUILD_ARCH}; do -t builder:$TARGETARCH \ . docker create --platform linux/$TARGETARCH --name builder-$TARGETARCH builder:$TARGETARCH - docker cp builder-$TARGETARCH:/go/src/github.com/jmorganca/ollama/ollama ./dist/ollama-linux-$TARGETARCH + docker cp builder-$TARGETARCH:/go/src/github.com/ollama/ollama/ollama ./dist/ollama-linux-$TARGETARCH if [ "$TARGETARCH" = "amd64" ]; then - docker cp builder-$TARGETARCH:/go/src/github.com/jmorganca/ollama/dist/deps/ ./dist/ + docker cp builder-$TARGETARCH:/go/src/github.com/ollama/ollama/dist/deps/ ./dist/ fi docker rm builder-$TARGETARCH diff --git a/scripts/build_windows.ps1 b/scripts/build_windows.ps1 index e6e1f4b0..1a89045a 100644 --- a/scripts/build_windows.ps1 +++ b/scripts/build_windows.ps1 @@ -74,7 +74,7 @@ function buildOllama() { } else { write-host "Skipping generate step with OLLAMA_SKIP_GENERATE set" } - & go build -trimpath -ldflags "-s -w -X=github.com/jmorganca/ollama/version.Version=$script:VERSION -X=github.com/jmorganca/ollama/server.mode=release" . + & go build -trimpath -ldflags "-s -w -X=github.com/ollama/ollama/version.Version=$script:VERSION -X=github.com/ollama/ollama/server.mode=release" . if ($LASTEXITCODE -ne 0) { exit($LASTEXITCODE)} if ("${env:KEY_CONTAINER}") { & "${script:SignTool}" sign /v /fd sha256 /t http://timestamp.digicert.com /f "${script:OLLAMA_CERT}" ` @@ -89,7 +89,7 @@ function buildApp() { write-host "Building Ollama App" cd "${script:SRC_DIR}\app" & windres -l 0 -o ollama.syso ollama.rc - & go build -trimpath -ldflags "-s -w -H windowsgui -X=github.com/jmorganca/ollama/version.Version=$script:VERSION -X=github.com/jmorganca/ollama/server.mode=release" . + & go build -trimpath -ldflags "-s -w -H windowsgui -X=github.com/ollama/ollama/version.Version=$script:VERSION -X=github.com/ollama/ollama/server.mode=release" . if ($LASTEXITCODE -ne 0) { exit($LASTEXITCODE)} if ("${env:KEY_CONTAINER}") { & "${script:SignTool}" sign /v /fd sha256 /t http://timestamp.digicert.com /f "${script:OLLAMA_CERT}" ` diff --git a/scripts/push_docker.sh b/scripts/push_docker.sh index da953dd4..c140c27c 100755 --- a/scripts/push_docker.sh +++ b/scripts/push_docker.sh @@ -3,7 +3,7 @@ set -eu export VERSION=${VERSION:-0.0.0} -export GOFLAGS="'-ldflags=-w -s \"-X=github.com/jmorganca/ollama/version.Version=$VERSION\" \"-X=github.com/jmorganca/ollama/server.mode=release\"'" +export GOFLAGS="'-ldflags=-w -s \"-X=github.com/ollama/ollama/version.Version=$VERSION\" \"-X=github.com/ollama/ollama/server.mode=release\"'" docker build \ --push \ diff --git a/server/auth.go b/server/auth.go index 5af85ff6..e92a5b65 100644 --- a/server/auth.go +++ b/server/auth.go @@ -15,8 +15,8 @@ import ( "strings" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/auth" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/auth" ) type registryChallenge struct { diff --git a/server/download.go b/server/download.go index f6d199b9..db9b9681 100644 --- a/server/download.go +++ b/server/download.go @@ -21,8 +21,8 @@ import ( "golang.org/x/sync/errgroup" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" ) const maxRetries = 6 diff --git a/server/images.go b/server/images.go index 11c11745..e5837d91 100644 --- a/server/images.go +++ b/server/images.go @@ -24,11 +24,11 @@ import ( "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/convert" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/convert" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) type registryOptions struct { diff --git a/server/prompt.go b/server/prompt.go index 88da5b6b..604e6971 100644 --- a/server/prompt.go +++ b/server/prompt.go @@ -7,7 +7,7 @@ import ( "text/template" "text/template/parse" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) // isResponseNode checks if the node contains .Response diff --git a/server/prompt_test.go b/server/prompt_test.go index 500ee522..a7e18a70 100644 --- a/server/prompt_test.go +++ b/server/prompt_test.go @@ -4,7 +4,7 @@ import ( "strings" "testing" - "github.com/jmorganca/ollama/api" + "github.com/ollama/ollama/api" ) func TestPrompt(t *testing.T) { diff --git a/server/routes.go b/server/routes.go index a03f39e7..5582f3a1 100644 --- a/server/routes.go +++ b/server/routes.go @@ -27,12 +27,12 @@ import ( "github.com/gin-gonic/gin" "golang.org/x/exp/slices" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/gpu" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/openai" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/gpu" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/openai" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) var mode string = gin.DebugMode diff --git a/server/routes_test.go b/server/routes_test.go index bbed02ed..9b298970 100644 --- a/server/routes_test.go +++ b/server/routes_test.go @@ -15,10 +15,10 @@ import ( "github.com/stretchr/testify/assert" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/llm" - "github.com/jmorganca/ollama/parser" - "github.com/jmorganca/ollama/version" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/llm" + "github.com/ollama/ollama/parser" + "github.com/ollama/ollama/version" ) func Test_Routes(t *testing.T) { diff --git a/server/upload.go b/server/upload.go index 4da34052..9b52238a 100644 --- a/server/upload.go +++ b/server/upload.go @@ -16,8 +16,8 @@ import ( "sync/atomic" "time" - "github.com/jmorganca/ollama/api" - "github.com/jmorganca/ollama/format" + "github.com/ollama/ollama/api" + "github.com/ollama/ollama/format" "golang.org/x/sync/errgroup" )