summaryrefslogtreecommitdiff
path: root/llama.cpp
diff options
context:
space:
mode:
authorrunfuture <runfuture@users.noreply.github.com>2024-02-08 18:36:19 +0800
committerGitHub <noreply@github.com>2024-02-08 12:36:19 +0200
commit4aa43fab569215a13495a7f1a0f8afc541b16d03 (patch)
tree56afda2368ff775a0fda01a543abfbfa89534a0c /llama.cpp
parenta6e514a85f0fda38ff78ec91782877ea3d19ed98 (diff)
llama : fix MiniCPM (#5392)
* fix bug for norm_rms_eps missing * to align with the same order as convert.py for model write * fix: undo HF models permute tensor * update for flake8 lint
Diffstat (limited to 'llama.cpp')
-rw-r--r--llama.cpp2
1 files changed, 2 insertions, 0 deletions
diff --git a/llama.cpp b/llama.cpp
index f8f5796a..552e0d02 100644
--- a/llama.cpp
+++ b/llama.cpp
@@ -2947,6 +2947,8 @@ static void llm_load_hparams(
} break;
case LLM_ARCH_MINICPM:
{
+ ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
+
switch (hparams.n_layer) {
case 40: model.type = e_model::MODEL_2B; break;
default: model.type = e_model::MODEL_UNKNOWN;