summaryrefslogtreecommitdiff
path: root/.devops/llama-cpp-cublas.srpm.spec
diff options
context:
space:
mode:
authorslaren <slarengh@gmail.com>2024-03-26 01:16:01 +0100
committerGitHub <noreply@github.com>2024-03-26 01:16:01 +0100
commit280345968dabc00d212d43e31145f5c9961a7604 (patch)
tree4d0ada8b59a4c15cb6d4fe1a6b4740a30dcdb0f2 /.devops/llama-cpp-cublas.srpm.spec
parentb06c16ef9f81d84da520232c125d4d8a1d273736 (diff)
cuda : rename build flag to LLAMA_CUDA (#6299)
Diffstat (limited to '.devops/llama-cpp-cublas.srpm.spec')
-rw-r--r--.devops/llama-cpp-cublas.srpm.spec83
1 files changed, 0 insertions, 83 deletions
diff --git a/.devops/llama-cpp-cublas.srpm.spec b/.devops/llama-cpp-cublas.srpm.spec
deleted file mode 100644
index f847ebb1..00000000
--- a/.devops/llama-cpp-cublas.srpm.spec
+++ /dev/null
@@ -1,83 +0,0 @@
-# SRPM for building from source and packaging an RPM for RPM-based distros.
-# https://fedoraproject.org/wiki/How_to_create_an_RPM_package
-# Built and maintained by John Boero - boeroboy@gmail.com
-# In honor of Seth Vidal https://www.redhat.com/it/blog/thank-you-seth-vidal
-
-# Notes for llama.cpp:
-# 1. Tags are currently based on hash - which will not sort asciibetically.
-# We need to declare standard versioning if people want to sort latest releases.
-# 2. Builds for CUDA/OpenCL support are separate, with different depenedencies.
-# 3. NVidia's developer repo must be enabled with nvcc, cublas, clblas, etc installed.
-# Example: https://developer.download.nvidia.com/compute/cuda/repos/fedora37/x86_64/cuda-fedora37.repo
-# 4. OpenCL/CLBLAST support simply requires the ICD loader and basic opencl libraries.
-# It is up to the user to install the correct vendor-specific support.
-
-Name: llama.cpp-cublas
-Version: %( date "+%%Y%%m%%d" )
-Release: 1%{?dist}
-Summary: CPU Inference of LLaMA model in pure C/C++ (no CUDA/OpenCL)
-License: MIT
-Source0: https://github.com/ggerganov/llama.cpp/archive/refs/heads/master.tar.gz
-BuildRequires: coreutils make gcc-c++ git cuda-toolkit
-Requires: cuda-toolkit
-URL: https://github.com/ggerganov/llama.cpp
-
-%define debug_package %{nil}
-%define source_date_epoch_from_changelog 0
-
-%description
-CPU inference for Meta's Lllama2 models using default options.
-
-%prep
-%setup -n llama.cpp-master
-
-%build
-make -j LLAMA_CUBLAS=1
-
-%install
-mkdir -p %{buildroot}%{_bindir}/
-cp -p main %{buildroot}%{_bindir}/llamacppcublas
-cp -p server %{buildroot}%{_bindir}/llamacppcublasserver
-cp -p simple %{buildroot}%{_bindir}/llamacppcublassimple
-
-mkdir -p %{buildroot}/usr/lib/systemd/system
-%{__cat} <<EOF > %{buildroot}/usr/lib/systemd/system/llamacublas.service
-[Unit]
-Description=Llama.cpp server, CPU only (no GPU support in this build).
-After=syslog.target network.target local-fs.target remote-fs.target nss-lookup.target
-
-[Service]
-Type=simple
-EnvironmentFile=/etc/sysconfig/llama
-ExecStart=/usr/bin/llamacppcublasserver $LLAMA_ARGS
-ExecReload=/bin/kill -s HUP $MAINPID
-Restart=never
-
-[Install]
-WantedBy=default.target
-EOF
-
-mkdir -p %{buildroot}/etc/sysconfig
-%{__cat} <<EOF > %{buildroot}/etc/sysconfig/llama
-LLAMA_ARGS="-m /opt/llama2/ggml-model-f32.bin"
-EOF
-
-%clean
-rm -rf %{buildroot}
-rm -rf %{_builddir}/*
-
-%files
-%{_bindir}/llamacppcublas
-%{_bindir}/llamacppcublasserver
-%{_bindir}/llamacppcublassimple
-/usr/lib/systemd/system/llamacublas.service
-%config /etc/sysconfig/llama
-
-%pre
-
-%post
-
-%preun
-%postun
-
-%changelog