summaryrefslogtreecommitdiff
AgeCommit message (Expand)Author
2024-04-24convert : add support of codeqwen due to tokenizer (#6707)Junyang Lin
2024-04-24llama : add phi3 support (#6852)liuwei-git
2024-04-23[SYCL] Windows default build instructions without -DLLAMA_SYCL_F16 flag activ...Anas Ahouzi
2024-04-22llamafile : improve sgemm.cpp (#6796)Justine Tunney
2024-04-22ggml : fix calloc argument ordering. (#6820)Dave Airlie
2024-04-22llama : fix typo in <|im_end|> token text (#6745)Georgi Gerganov
2024-04-22ci: fix job are cancelling each other (#6781)Pierrick Hymbert
2024-04-22flake.lock: Updategithub-actions[bot]
2024-04-21`build`: generate hex dump of server assets during build (#6661)Olivier Chafik
2024-04-21llama : add option to render special/control tokens (#6807)Georgi Gerganov
2024-04-21ggml : fix ggml_backend_cpu_supports_op() for CPY (#0)Georgi Gerganov
2024-04-21llama : add llama-3 chat template (#6751)Wouter
2024-04-21gguf-py : add IQ1_M to GGML_QUANT_SIZES (#6761)pmysl
2024-04-21doc : add link to falcon (#6789)Jan Boon
2024-04-21readme : add Fedora instructions (#6783)Mohammadreza Hendiani
2024-04-21llava : use logger in llava-cli (#6797)Justine Tunney
2024-04-21llama : support Llama 3 HF conversion (#6745)Pedro Cuenca
2024-04-20doc : server tests require llama to be built with curl enabled (#6788)Jan Boon
2024-04-20common : try to fix Android CI (#6780)Georgi Gerganov
2024-04-19ci: add ubuntu latest release and fix missing build number (mac & ubuntu) (#6...loonerin
2024-04-19server: static: upstream upgrade (#6765)Pierrick Hymbert
2024-04-19Implement the OLMo architecture (#6741)nopperl
2024-04-19train : add general name (#6752)Austin
2024-04-19fix wrong parameter in cmd in readme-sycl.md (#6755)Neo Zhang
2024-04-18ggml : group all experts in a single ggml_mul_mat_id (#6505)slaren
2024-04-18convert : support models with multiple chat templates (#6588)Sigbjørn Skjæret
2024-04-18Qwen2 : assume tied weights if lm_head/output weights is missing (#6738)Ren Xuancheng
2024-04-18llama : fix compatibility with old 2 expert models (#6735)slaren
2024-04-17llamafile : tmp disable + build sgemm.o when needed (#6716)Georgi Gerganov
2024-04-17readme : add UI (#6724)Yaroslav
2024-04-16convert : fix autoawq gemma (#6704)Zheng.Deng
2024-04-16llama : make general.name optional (#6709)Georgi Gerganov
2024-04-16ggml : fix llamafile sgemm wdata offsets (#6710)Georgi Gerganov
2024-04-16ggml : add llamafile sgemm (#6414)Justine Tunney
2024-04-16llama : add StableLM2 12B (#6635)Ashish
2024-04-16llama : add qwen2moe (#6074)Shijie
2024-04-16gritlm : add --outdir option to hf.sh script (#6699)Daniel Bevenius
2024-04-16perplexity : require positive --ctx-size arg (#6695)Georgi Gerganov
2024-04-16gguf : add special tokens metadata for FIM/Infill (#6689)Daniel Bevenius
2024-04-15`main`: add --json-schema / -j flag (#6659)Olivier Chafik
2024-04-15llama : fix restoring the number of outputs from state files (#6687)compilade
2024-04-15server : revert "minor layout improvements" (#6684)Pierrick Hymbert
2024-04-15swift : linux support (#6590)Steven Prichard
2024-04-15fix mul_mat_id() for new input, make the ut pass (#6682)Neo Zhang Jianyu
2024-04-14llama : add missing kv clear in llama_beam_search (#6664)David Renshaw
2024-04-14Add Command R chat template (#6650)Chao Jiang
2024-04-14flake.lock: Update (#6669)Georgi Gerganov
2024-04-14Added support for GGML_OP_CLAMP in Metal (#6662)Dave
2024-04-14Fix --split-max-size (#6655)Sigbjørn Skjæret
2024-04-14[bug fix] convert github repository_owner to lowercase (#6673)Jaemin Son