summaryrefslogtreecommitdiff
path: root/examples/llava/llava-cli.cpp
diff options
context:
space:
mode:
authorMatthew Tejo <matthew.tejo@gmail.com>2023-11-06 23:43:59 -0800
committerGitHub <noreply@github.com>2023-11-07 10:43:59 +0300
commit54b4df8886103b436a4bb3b60f4d84824f9e8868 (patch)
tree248fb5a186235fec984173005f40c485d74c55c7 /examples/llava/llava-cli.cpp
parent46876d2a2c92e60579dc732cdb8cbd243b06f317 (diff)
Use params when loading models in llava-cli (#3976)
llava-cli was loading models with default params and ignoring settings from the cli. This switches to a generic function to load the params from the cli options.
Diffstat (limited to 'examples/llava/llava-cli.cpp')
-rw-r--r--examples/llava/llava-cli.cpp8
1 files changed, 3 insertions, 5 deletions
diff --git a/examples/llava/llava-cli.cpp b/examples/llava/llava-cli.cpp
index 19374c67..633afd1d 100644
--- a/examples/llava/llava-cli.cpp
+++ b/examples/llava/llava-cli.cpp
@@ -242,18 +242,16 @@ static struct llava_context * llava_init(gpt_params * params) {
llama_backend_init(params->numa);
- llama_model_params model_params = llama_model_default_params();
+ llama_model_params model_params = llama_model_params_from_gpt_params(*params);
+
llama_model * model = llama_load_model_from_file(params->model.c_str(), model_params);
if (model == NULL) {
fprintf(stderr , "%s: error: unable to load model\n" , __func__);
return NULL;
}
- llama_context_params ctx_params = llama_context_default_params();
-
+ llama_context_params ctx_params = llama_context_params_from_gpt_params(*params);
ctx_params.n_ctx = params->n_ctx < 2048 ? 2048 : params->n_ctx; // we need a longer context size to process image embeddings
- ctx_params.n_threads = params->n_threads;
- ctx_params.n_threads_batch = params->n_threads_batch == -1 ? params->n_threads : params->n_threads_batch;
llama_context * ctx_llama = llama_new_context_with_model(model, ctx_params);