summaryrefslogtreecommitdiff
path: root/llama.cpp
diff options
context:
space:
mode:
authorGeorgi Gerganov <ggerganov@gmail.com>2024-06-04 21:23:20 +0300
committerGitHub <noreply@github.com>2024-06-04 21:23:20 +0300
commit554c247caffed64465f372661f2826640cb10430 (patch)
treece211403d5746a5cde175247a806223e83e17375 /llama.cpp
parent0cd6bd3483fa66124b76a8a8ac794d9ee18c70c1 (diff)
ggml : remove OpenCL (#7735)
ggml-ci
Diffstat (limited to 'llama.cpp')
-rw-r--r--llama.cpp10
1 files changed, 1 insertions, 9 deletions
diff --git a/llama.cpp b/llama.cpp
index 92c33f53..c05e2bdb 100644
--- a/llama.cpp
+++ b/llama.cpp
@@ -13,8 +13,6 @@
#ifdef GGML_USE_CUDA
# include "ggml-cuda.h"
-#elif defined(GGML_USE_CLBLAST)
-# include "ggml-opencl.h"
#elif defined(GGML_USE_VULKAN)
# include "ggml-vulkan.h"
#elif defined(GGML_USE_SYCL)
@@ -2406,8 +2404,6 @@ static ggml_backend_buffer_type_t llama_default_buffer_type_offload(const llama_
buft = ggml_backend_vk_buffer_type(gpu);
#elif defined(GGML_USE_SYCL)
buft = ggml_backend_sycl_buffer_type(gpu);
-#elif defined(GGML_USE_CLBLAST)
- buft = ggml_backend_opencl_buffer_type();
#elif defined(GGML_USE_KOMPUTE)
buft = ggml_backend_kompute_buffer_type(gpu);
if (buft == nullptr) {
@@ -2530,10 +2526,6 @@ static bool llama_kv_cache_init(
}
}
-#ifdef GGML_USE_CLBLAST
- offload = false;
-#endif
-
// count used buffer types
std::map<ggml_backend_buffer_type_t, int> buft_layer_count;
if (offload) {
@@ -15921,7 +15913,7 @@ bool llama_supports_mlock(void) {
}
bool llama_supports_gpu_offload(void) {
-#if defined(GGML_USE_CUDA) || defined(GGML_USE_CLBLAST) || defined(GGML_USE_METAL) || defined(GGML_USE_VULKAN) || \
+#if defined(GGML_USE_CUDA) || defined(GGML_USE_METAL) || defined(GGML_USE_VULKAN) || \
defined(GGML_USE_SYCL) || defined(GGML_USE_KOMPUTE) || defined(GGML_USE_RPC)
// Defined when llama.cpp is compiled with support for offloading model layers to GPU.
return true;