summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorGeorgi Gerganov <ggerganov@gmail.com>2024-01-14 13:26:53 +0200
committerGitHub <noreply@github.com>2024-01-14 13:26:53 +0200
commitbb0c1392479398f9aba86d9ec98db0b95ede6e6d (patch)
tree47e096747bf7fd3349a3415846f76d4125991c6b
parent9408cfdad6b1c090a7e1419d4434edc260b7e47e (diff)
llama : check LLAMA_TRACE env for extra logging (#4929)
* llama : minor fix indent * llama : check LLAMA_TRACE env for extra logging ggml-ci
-rw-r--r--llama.cpp32
1 files changed, 18 insertions, 14 deletions
diff --git a/llama.cpp b/llama.cpp
index 51821965..63f37ecd 100644
--- a/llama.cpp
+++ b/llama.cpp
@@ -2190,6 +2190,11 @@ struct llama_model_loader {
LLM_KV llm_kv = LLM_KV(LLM_ARCH_UNKNOWN);
llama_model_loader(const std::string & fname, bool use_mmap, const struct llama_model_kv_override * param_overrides_p) : file(fname.c_str(), "rb") {
+ int trace = 0;
+ if (getenv("LLAMA_TRACE")) {
+ trace = atoi(getenv("LLAMA_TRACE"));
+ }
+
struct gguf_init_params params = {
/*.no_alloc = */ true,
/*.ctx = */ &ctx_meta,
@@ -2242,11 +2247,10 @@ struct llama_model_loader {
type_max = type;
}
- // TODO: make runtime configurable
-#if 0
- struct ggml_tensor * meta = ggml_get_tensor(ctx_meta, gguf_get_tensor_name(ctx_gguf, i));
- LLAMA_LOG_INFO("%s: - tensor %4d: %32s %-8s [ %s ]\n", __func__, i, ggml_get_name(meta), ggml_type_name(type), llama_format_tensor_shape(meta).c_str());
-#endif
+ if (trace > 0) {
+ struct ggml_tensor * meta = ggml_get_tensor(ctx_meta, gguf_get_tensor_name(ctx_gguf, i));
+ LLAMA_LOG_INFO("%s: - tensor %4d: %32s %-8s [ %s ]\n", __func__, i, ggml_get_name(meta), ggml_type_name(type), llama_format_tensor_shape(meta).c_str());
+ }
}
switch (type_max) {
@@ -6451,15 +6455,15 @@ static uint8_t llama_token_to_byte(const llama_vocab& vocab, llama_token id) {
static llama_token llama_byte_to_token(const llama_vocab & vocab, uint8_t ch) {
static const char * hex = "0123456789ABCDEF";
switch (llama_vocab_get_type(vocab)) {
- case LLAMA_VOCAB_TYPE_SPM: {
- const char buf[7] = { '<', '0', 'x', hex[ch >> 4], hex[ch & 15], '>', 0 };
- return vocab.token_to_id.at(buf);
- }
- case LLAMA_VOCAB_TYPE_BPE: {
- return vocab.token_to_id.at(bytes_to_unicode_bpe(ch));
- }
- default:
- GGML_ASSERT(false);
+ case LLAMA_VOCAB_TYPE_SPM: {
+ const char buf[7] = { '<', '0', 'x', hex[ch >> 4], hex[ch & 15], '>', 0 };
+ return vocab.token_to_id.at(buf);
+ }
+ case LLAMA_VOCAB_TYPE_BPE: {
+ return vocab.token_to_id.at(bytes_to_unicode_bpe(ch));
+ }
+ default:
+ GGML_ASSERT(false);
}
}