summaryrefslogtreecommitdiff
path: root/llama.h
diff options
context:
space:
mode:
Diffstat (limited to 'llama.h')
-rw-r--r--llama.h13
1 files changed, 5 insertions, 8 deletions
diff --git a/llama.h b/llama.h
index deb09fe5..42c364c6 100644
--- a/llama.h
+++ b/llama.h
@@ -55,6 +55,7 @@ extern "C" {
bool f16_kv; // use fp16 for KV cache
bool logits_all; // the llama_eval() call computes all logits, not just the last one
bool vocab_only; // only load the vocabulary, no weights
+ bool use_mmap; // use mmap if possible
bool use_mlock; // force system to keep model in RAM
bool embedding; // embedding mode only
@@ -66,6 +67,9 @@ extern "C" {
LLAMA_API struct llama_context_params llama_context_default_params();
+ LLAMA_API bool llama_mmap_supported();
+ LLAMA_API bool llama_mlock_supported();
+
// Various functions for loading a ggml llama model.
// Allocate (almost) all memory needed for the model.
// Return NULL on failure
@@ -164,13 +168,6 @@ extern "C" {
#ifdef __cplusplus
}
-
-#include <string>
-#include <unordered_map>
-//
-// Internal function exposed for tests and benchmarks
-//
-std::unordered_map<std::string, struct ggml_tensor *>& llama_internal_get_tensor_map(struct llama_context * ctx);
#endif
-#endif
+#endif // LLAMA_H