diff options
author | Jakub N <jakubniemczyk97@gmail.com> | 2024-03-11 14:40:42 +0100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-03-11 14:40:42 +0100 |
commit | 828defefb66fc8a25404f5de845897145bf34061 (patch) | |
tree | 2f9671748c9776636c24e297ade8b76462251cdc /examples/server | |
parent | caa106d4e05a0ab94225c220b81f9e2cd522339b (diff) |
Update server docker image URLs (#5997)
Diffstat (limited to 'examples/server')
-rw-r--r-- | examples/server/README.md | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/examples/server/README.md b/examples/server/README.md index 37673905..8f8454af 100644 --- a/examples/server/README.md +++ b/examples/server/README.md @@ -123,10 +123,10 @@ You can consume the endpoints with Postman or NodeJS with axios library. You can ### Docker ```bash -docker run -p 8080:8080 -v /path/to/models:/models ggerganov/llama.cpp:server -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 +docker run -p 8080:8080 -v /path/to/models:/models ghcr.io/ggerganov/llama.cpp:server -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 # or, with CUDA: -docker run -p 8080:8080 -v /path/to/models:/models --gpus all ggerganov/llama.cpp:server-cuda -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 --n-gpu-layers 99 +docker run -p 8080:8080 -v /path/to/models:/models --gpus all ghcr.io/ggerganov/llama.cpp:server-cuda -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 --n-gpu-layers 99 ``` ## Testing with CURL |