index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
Age
Commit message (
Expand
)
Author
2024-04-24
convert : add support of codeqwen due to tokenizer (#6707)
Junyang Lin
2024-04-24
llama : add phi3 support (#6852)
liuwei-git
2024-04-23
[SYCL] Windows default build instructions without -DLLAMA_SYCL_F16 flag activ...
Anas Ahouzi
2024-04-22
llamafile : improve sgemm.cpp (#6796)
Justine Tunney
2024-04-22
ggml : fix calloc argument ordering. (#6820)
Dave Airlie
2024-04-22
llama : fix typo in <|im_end|> token text (#6745)
Georgi Gerganov
2024-04-22
ci: fix job are cancelling each other (#6781)
Pierrick Hymbert
2024-04-22
flake.lock: Update
github-actions[bot]
2024-04-21
`build`: generate hex dump of server assets during build (#6661)
Olivier Chafik
2024-04-21
llama : add option to render special/control tokens (#6807)
Georgi Gerganov
2024-04-21
ggml : fix ggml_backend_cpu_supports_op() for CPY (#0)
Georgi Gerganov
2024-04-21
llama : add llama-3 chat template (#6751)
Wouter
2024-04-21
gguf-py : add IQ1_M to GGML_QUANT_SIZES (#6761)
pmysl
2024-04-21
doc : add link to falcon (#6789)
Jan Boon
2024-04-21
readme : add Fedora instructions (#6783)
Mohammadreza Hendiani
2024-04-21
llava : use logger in llava-cli (#6797)
Justine Tunney
2024-04-21
llama : support Llama 3 HF conversion (#6745)
Pedro Cuenca
2024-04-20
doc : server tests require llama to be built with curl enabled (#6788)
Jan Boon
2024-04-20
common : try to fix Android CI (#6780)
Georgi Gerganov
2024-04-19
ci: add ubuntu latest release and fix missing build number (mac & ubuntu) (#6...
loonerin
2024-04-19
server: static: upstream upgrade (#6765)
Pierrick Hymbert
2024-04-19
Implement the OLMo architecture (#6741)
nopperl
2024-04-19
train : add general name (#6752)
Austin
2024-04-19
fix wrong parameter in cmd in readme-sycl.md (#6755)
Neo Zhang
2024-04-18
ggml : group all experts in a single ggml_mul_mat_id (#6505)
slaren
2024-04-18
convert : support models with multiple chat templates (#6588)
Sigbjørn Skjæret
2024-04-18
Qwen2 : assume tied weights if lm_head/output weights is missing (#6738)
Ren Xuancheng
2024-04-18
llama : fix compatibility with old 2 expert models (#6735)
slaren
2024-04-17
llamafile : tmp disable + build sgemm.o when needed (#6716)
Georgi Gerganov
2024-04-17
readme : add UI (#6724)
Yaroslav
2024-04-16
convert : fix autoawq gemma (#6704)
Zheng.Deng
2024-04-16
llama : make general.name optional (#6709)
Georgi Gerganov
2024-04-16
ggml : fix llamafile sgemm wdata offsets (#6710)
Georgi Gerganov
2024-04-16
ggml : add llamafile sgemm (#6414)
Justine Tunney
2024-04-16
llama : add StableLM2 12B (#6635)
Ashish
2024-04-16
llama : add qwen2moe (#6074)
Shijie
2024-04-16
gritlm : add --outdir option to hf.sh script (#6699)
Daniel Bevenius
2024-04-16
perplexity : require positive --ctx-size arg (#6695)
Georgi Gerganov
2024-04-16
gguf : add special tokens metadata for FIM/Infill (#6689)
Daniel Bevenius
2024-04-15
`main`: add --json-schema / -j flag (#6659)
Olivier Chafik
2024-04-15
llama : fix restoring the number of outputs from state files (#6687)
compilade
2024-04-15
server : revert "minor layout improvements" (#6684)
Pierrick Hymbert
2024-04-15
swift : linux support (#6590)
Steven Prichard
2024-04-15
fix mul_mat_id() for new input, make the ut pass (#6682)
Neo Zhang Jianyu
2024-04-14
llama : add missing kv clear in llama_beam_search (#6664)
David Renshaw
2024-04-14
Add Command R chat template (#6650)
Chao Jiang
2024-04-14
flake.lock: Update (#6669)
Georgi Gerganov
2024-04-14
Added support for GGML_OP_CLAMP in Metal (#6662)
Dave
2024-04-14
Fix --split-max-size (#6655)
Sigbjørn Skjæret
2024-04-14
[bug fix] convert github repository_owner to lowercase (#6673)
Jaemin Son
[next]