From 154e0d75fccf1784fe9ff6fd76a630b66563da3d Mon Sep 17 00:00:00 2001 From: Kawrakow <48489457+ikawrakow@users.noreply.github.com> Date: Sat, 27 Jul 2024 07:55:01 +0200 Subject: Merge mainline llama.cpp (#3) * Merging mainline - WIP * Merging mainline - WIP AVX2 and CUDA appear to work. CUDA performance seems slightly (~1-2%) lower as it is so often the case with llama.cpp/ggml after some "improvements" have been made. * Merging mainline - fix Metal * Remove check --------- Co-authored-by: Iwan Kawrakow --- docs/development/llama-star/idea-arch.key | Bin 0 -> 488591 bytes docs/development/llama-star/idea-arch.pdf | Bin 0 -> 42334 bytes 2 files changed, 0 insertions(+), 0 deletions(-) create mode 100755 docs/development/llama-star/idea-arch.key create mode 100644 docs/development/llama-star/idea-arch.pdf (limited to 'docs/development/llama-star') diff --git a/docs/development/llama-star/idea-arch.key b/docs/development/llama-star/idea-arch.key new file mode 100755 index 00000000..3e068e70 Binary files /dev/null and b/docs/development/llama-star/idea-arch.key differ diff --git a/docs/development/llama-star/idea-arch.pdf b/docs/development/llama-star/idea-arch.pdf new file mode 100644 index 00000000..4fa92c71 Binary files /dev/null and b/docs/development/llama-star/idea-arch.pdf differ -- cgit v1.2.3