summaryrefslogtreecommitdiff
path: root/examples/llama-bench/llama-bench.cpp
diff options
context:
space:
mode:
authorslaren <slarengh@gmail.com>2024-05-22 16:10:46 +0200
committerGitHub <noreply@github.com>2024-05-22 16:10:46 +0200
commitb18532a4efeca8796fea8e36195c81cbfd596a4a (patch)
treecc9966ee05ad20c4f554a430e2f4b03c40d75b8b /examples/llama-bench/llama-bench.cpp
parentfcda1128bc5f8eb7e1811708fe9d9867b9aec815 (diff)
phi3 : duplicate rope factors in each layer (#7447)
* phi3 : duplicate rope factors in each layer phi3 : set phi-3 model type as 14B model loader : simplify the process for duplicating model tensors llama-bench : remove default pg test * replace bool parameters in llama_model_loader with named flags
Diffstat (limited to 'examples/llama-bench/llama-bench.cpp')
-rw-r--r--examples/llama-bench/llama-bench.cpp2
1 files changed, 1 insertions, 1 deletions
diff --git a/examples/llama-bench/llama-bench.cpp b/examples/llama-bench/llama-bench.cpp
index 8b965e19..6bb1f70c 100644
--- a/examples/llama-bench/llama-bench.cpp
+++ b/examples/llama-bench/llama-bench.cpp
@@ -195,7 +195,7 @@ static const cmd_params cmd_params_defaults = {
/* model */ {"models/7B/ggml-model-q4_0.gguf"},
/* n_prompt */ {512},
/* n_gen */ {128},
- /* n_pg */ {{512, 128}},
+ /* n_pg */ {},
/* n_batch */ {2048},
/* n_ubatch */ {512},
/* type_k */ {GGML_TYPE_F16},