summaryrefslogtreecommitdiff
path: root/examples/server
diff options
context:
space:
mode:
authorJakub N <jakubniemczyk97@gmail.com>2024-03-11 14:40:42 +0100
committerGitHub <noreply@github.com>2024-03-11 14:40:42 +0100
commit828defefb66fc8a25404f5de845897145bf34061 (patch)
tree2f9671748c9776636c24e297ade8b76462251cdc /examples/server
parentcaa106d4e05a0ab94225c220b81f9e2cd522339b (diff)
Update server docker image URLs (#5997)
Diffstat (limited to 'examples/server')
-rw-r--r--examples/server/README.md4
1 files changed, 2 insertions, 2 deletions
diff --git a/examples/server/README.md b/examples/server/README.md
index 37673905..8f8454af 100644
--- a/examples/server/README.md
+++ b/examples/server/README.md
@@ -123,10 +123,10 @@ You can consume the endpoints with Postman or NodeJS with axios library. You can
### Docker
```bash
-docker run -p 8080:8080 -v /path/to/models:/models ggerganov/llama.cpp:server -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080
+docker run -p 8080:8080 -v /path/to/models:/models ghcr.io/ggerganov/llama.cpp:server -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080
# or, with CUDA:
-docker run -p 8080:8080 -v /path/to/models:/models --gpus all ggerganov/llama.cpp:server-cuda -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 --n-gpu-layers 99
+docker run -p 8080:8080 -v /path/to/models:/models --gpus all ghcr.io/ggerganov/llama.cpp:server-cuda -m models/7B/ggml-model.gguf -c 512 --host 0.0.0.0 --port 8080 --n-gpu-layers 99
```
## Testing with CURL