From 213e90ed73f8ac3cd3026dc3f086beae0d414f96 Mon Sep 17 00:00:00 2001 From: Herman Semenov Date: Mon, 20 May 2024 07:33:21 +0000 Subject: ggml-opencl, llama: using reserve() if count already known (#7272) --- llama.cpp | 3 +++ 1 file changed, 3 insertions(+) (limited to 'llama.cpp') diff --git a/llama.cpp b/llama.cpp index ca3e9fcc..2025e455 100644 --- a/llama.cpp +++ b/llama.cpp @@ -16162,6 +16162,7 @@ static bool llama_control_vector_init(struct llama_control_vector & cvec, const } // make tensors + cvec.tensors.reserve(model.hparams.n_layer); cvec.tensors.push_back(nullptr); // there's never a tensor for layer 0 for (size_t il = 1; il < model.hparams.n_layer; il++) { struct ggml_context * ctx = ctx_map.at(model.buft_layer[il].buft); @@ -16170,6 +16171,8 @@ static bool llama_control_vector_init(struct llama_control_vector & cvec, const } // allocate tensors / buffers and zero + cvec.ctxs.reserve(ctx_map.size()); + cvec.bufs.reserve(ctx_map.size()); for (auto it : ctx_map) { ggml_backend_buffer_type_t buft = it.first; ggml_context * ctx = it.second; -- cgit v1.2.3