diff options
author | Brian <mofosyne@gmail.com> | 2024-01-11 01:09:53 +1100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-01-10 16:09:53 +0200 |
commit | 57d016ba2d46a6e22517a31a75cebb48f9e234b6 (patch) | |
tree | 874e83ce598f7bbfdf1ad632df8d81cb3e635c18 | |
parent | 329ff615699d32f596d4ebf8baba654c30064e0d (diff) |
llama : add additional suffixes for model params (#4834)
* llm_load_print_meta: Add additional suffixs for model params
* Update llama.cpp model param log
remove unneeded comments and convert from > to >=
-rw-r--r-- | llama.cpp | 10 |
1 files changed, 9 insertions, 1 deletions
@@ -3146,7 +3146,15 @@ static void llm_load_print_meta(llama_model_loader & ml, llama_model & model) { LLAMA_LOG_INFO("%s: rope_finetuned = %s\n", __func__, hparams.rope_finetuned ? "yes" : "unknown"); LLAMA_LOG_INFO("%s: model type = %s\n", __func__, llama_model_type_name(model.type)); LLAMA_LOG_INFO("%s: model ftype = %s\n", __func__, llama_model_ftype_name(model.ftype).c_str()); - LLAMA_LOG_INFO("%s: model params = %.2f B\n", __func__, ml.n_elements*1e-9); + if (ml.n_elements >= 1e12) { + LLAMA_LOG_INFO("%s: model params = %.2f T\n", __func__, ml.n_elements*1e-12); + } else if (ml.n_elements >= 1e9) { + LLAMA_LOG_INFO("%s: model params = %.2f B\n", __func__, ml.n_elements*1e-9); + } else if (ml.n_elements >= 1e6) { + LLAMA_LOG_INFO("%s: model params = %.2f M\n", __func__, ml.n_elements*1e-6); + } else { + LLAMA_LOG_INFO("%s: model params = %.2f K\n", __func__, ml.n_elements*1e-3); + } if (ml.n_bytes < GiB) { LLAMA_LOG_INFO("%s: model size = %.2f MiB (%.2f BPW) \n", __func__, ml.n_bytes/1024.0/1024.0, ml.n_bytes*8.0/ml.n_elements); } else { |