index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
Age
Commit message (
Expand
)
Author
2024-01-26
Add OpenCL add kernel (#5151)
0cc4m
2024-01-26
cmake : pass CPU architecture flags to nvcc (#5146)
Jared Van Bortel
2024-01-26
cuda : fix tensor size calculation for non-split buffer (#5145)
slaren
2024-01-26
ggml-alloc : add 10% margin to the buffer sizes (#5149)
slaren
2024-01-26
ggml : update softmax n_task calculation (#5126)
snadampal
2024-01-26
scripts : move run-with-preset.py from root to scripts folder
Georgi Gerganov
2024-01-26
tests : gitignore test-c.o
Georgi Gerganov
2024-01-26
server : refactored the task processing logic (#5065)
Xuan Son Nguyen
2024-01-26
ci : add model tests + script wrapper (#4586)
crasm
2024-01-26
metal : remove unused `n_buffers` and `buffers` (#5129)
Paul Tsochantaris
2024-01-26
gguf : fix "general.alignment" type in gguf_reader.py (#5136)
Riceball LEE
2024-01-26
readme : update hot topics
Georgi Gerganov
2024-01-26
Another bucket sort (#5109)
Kawrakow
2024-01-25
readme : add MobileVLM 1.7B/3B to the supported models list (#5107)
XiaotaoChen
2024-01-25
llama : dynamic temperature sampling (#4972)
l3utterfly
2024-01-25
examples : make pydantic scripts pass mypy and support py3.8 (#5099)
Jared Van Bortel
2024-01-25
android : use release cmake build type by default (#5123)
Valentin Konovalov
2024-01-25
Fix Q3_K_XS for MoE models (#5113)
Kawrakow
2024-01-25
metal : show compile log messages
Georgi Gerganov
2024-01-24
cuda : fix 2-bit quants on amd hip (#5105)
Engininja2
2024-01-24
nix-shell: use addToSearchPath
Michael Hueschen
2024-01-24
nix: add cc to devShell LD_LIBRARY_PATH
Michael Hueschen
2024-01-24
llama : pre-allocate input tensors in a separate buffer (#5100)
slaren
2024-01-23
metal : disable support for MUL_MAT F32 x F16
Georgi Gerganov
2024-01-23
Additional KL-divergence statistics (#5081)
Kawrakow
2024-01-23
CUDA: more info when no device code (#5088)
Johannes Gäßler
2024-01-23
minor : clean-up some warnings and style (#5094)
Georgi Gerganov
2024-01-23
devops : add intel oneapi dockerfile (#5068)
Xuan Son Nguyen
2024-01-23
llama.vim : added api key support (#5090)
Michael Coppola
2024-01-22
llama : fix not enough space in buffer with Qwen (#5086)
slaren
2024-01-22
KL-divergence (#5076)
Kawrakow
2024-01-22
ggml : parallelize FP32 conversion when using BLAS (#5045)
Reinforce-II
2024-01-22
llava : MobileVLM support (#4954)
XiaotaoChen
2024-01-22
flake.nix: add a comment about flakes vs nix
Someone Serge
2024-01-22
nix: add a comment on the many nixpkgs-with-cuda instances
Someone Serge
2024-01-22
nix: add a comment about makeScope
Someone Serge
2024-01-22
nix: refactor the cleanSource rules
Someone Serge
2024-01-22
workflows: nix-ci: drop the redundant "paths" filter
Someone Serge
2024-01-22
workflows: nix-build-aarch64: rate limit
Someone Serge
2024-01-22
workflows: nix-ci: rebuild on flake.lock updates
Someone Serge
2024-01-22
imatrix : keep intermediate imatrix results (#5077)
Kawrakow
2024-01-22
llama : support StableLM 2 1.6B (#5052)
compilade
2024-01-22
finetune : print sample-start/include-sample-start (#5072)
Daniel Bevenius
2024-01-22
llama : add Q3_K_XS (#5060)
Kawrakow
2024-01-22
ci : fix Windows CI by updating Intel SDE version (#5053)
bobqianic
2024-01-22
llama : add more qwen2 models (#5071)
Shijie
2024-01-21
Revert LLAMA_NATIVE to OFF in flake.nix (#5066)
iSma
2024-01-21
add safetensors support to convert-lora-to-ggml.py (#5062)
kuronekosaiko
2024-01-21
add `#include <string>` to unicode.h (#5051)
bobqianic
2024-01-21
Add ability to evauate multiple choice tasks (#5047)
Kawrakow
[next]