summaryrefslogtreecommitdiff
path: root/examples/server/tests/features/server.feature
diff options
context:
space:
mode:
authorPierrick Hymbert <pierrick.hymbert@gmail.com>2024-03-23 18:07:00 +0100
committerGitHub <noreply@github.com>2024-03-23 18:07:00 +0100
commitf482bb2e4920e544651fb832f2e0bcb4d2ff69ab (patch)
tree9fabefd6f3b34aef6bf13a8469c7cdf363cc88cb /examples/server/tests/features/server.feature
parent1997577d5e121568ae39f538021733ccd4278c23 (diff)
common: llama_load_model_from_url split support (#6192)
* llama: llama_split_prefix fix strncpy does not include string termination common: llama_load_model_from_url: - fix header name case sensitive - support downloading additional split in parallel - hide password in url * common: EOL EOF * common: remove redundant LLAMA_CURL_MAX_PATH_LENGTH definition * common: change max url max length * common: minor comment * server: support HF URL options * llama: llama_model_loader fix log * common: use a constant for max url length * common: clean up curl if file cannot be loaded in gguf * server: tests: add split tests, and HF options params * common: move llama_download_hide_password_in_url inside llama_download_file as a lambda * server: tests: enable back Release test on PR * spacing Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * spacing Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> * spacing Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
Diffstat (limited to 'examples/server/tests/features/server.feature')
-rw-r--r--examples/server/tests/features/server.feature4
1 files changed, 2 insertions, 2 deletions
diff --git a/examples/server/tests/features/server.feature b/examples/server/tests/features/server.feature
index a2e0e5b3..646a4e49 100644
--- a/examples/server/tests/features/server.feature
+++ b/examples/server/tests/features/server.feature
@@ -4,8 +4,8 @@ Feature: llama.cpp server
Background: Server startup
Given a server listening on localhost:8080
- And a model url https://huggingface.co/ggml-org/models/resolve/main/tinyllamas/stories260K.gguf
- And a model file stories260K.gguf
+ And a model file tinyllamas/stories260K.gguf from HF repo ggml-org/models
+ And a model file test-model.gguf
And a model alias tinyllama-2
And 42 as server seed
# KV Cache corresponds to the total amount of tokens