summaryrefslogtreecommitdiff
path: root/llama.cpp
AgeCommit message (Expand)Author
2024-04-13model: support arch `DbrxForCausalLM` (#6515)Pierrick Hymbert
2024-04-12llama : add gguf_remove_key + remove split meta during quantize (#6591)jiez
2024-04-12Correct free memory and total memory. (#6630)MasterYi1024
2024-04-11Optimization: eliminate addition of redundant stacks when advancing grammar. ...Clint Herron
2024-04-11grammars: 1.5x faster inference w/ complex grammars (vector reserves / reuses...Olivier Chafik
2024-04-11eval-callback: Example how to use eval callback for debugging (#6576)Pierrick Hymbert
2024-04-10llama : add model types for mixtral (#6589)slaren
2024-04-09BERT tokenizer fixes (#6498)Jared Van Bortel
2024-04-09llama : add Command R Plus support (#6491)Carolinabanana
2024-04-08llama : fix attention layer count sanity check (#6550)Georgi Gerganov
2024-04-08quantize : fix precedence of cli args (#6541)Georgi Gerganov
2024-04-08llama : support negative ith in llama_get_ API (#6519)Rick G
2024-04-08llama : save and restore kv cache for single seq id (#6341)Jan Boon
2024-04-05gguf.py : add licence and version to gguf writer (#6504)Brian
2024-04-04examples : add GBNF validator program (#5948)Clint Herron
2024-04-03llama : add SEA-LION support (#6448)bryanSwk
2024-04-03Add OpenChat, Alpaca, Vicuna chat templates (#6397)kaizau
2024-04-03ggml : mul_mat_id use the same tensor for all the experts (#6387)slaren
2024-03-29Vulkan k-quant mmq and ggml-backend offload functionality (#6155)0cc4m
2024-03-29[Model] Add support for xverse (#6301)hxer7963
2024-03-29llama : remove redundant reshape in build_kv_store (#6369)Daniel Bevenius
2024-03-28llama : fix command-r inference when omitting outputs (#6367)compilade
2024-03-26wpm : portable unicode tolower (#6305)Jared Van Bortel
2024-03-26llama : greatly reduce output buffer memory usage (#6122)compilade
2024-03-26IQ1_M: 1.75 bpw quantization (#6302)Kawrakow
2024-03-26quantize : be able to override metadata by key (#6321)Kawrakow
2024-03-26cuda : rename build flag to LLAMA_CUDA (#6299)slaren
2024-03-24[SYCL] offload op (#6217)Meng, Hengyu
2024-03-23use _wfopen instead of fopen on Windows (#6248)Jared Van Bortel
2024-03-23common: llama_load_model_from_url split support (#6192)Pierrick Hymbert
2024-03-23llama : add grok-1 support (#6204)Julius Arkenberg
2024-03-22quantize: options for output and token embedding tensors qtype (#6239)Kawrakow
2024-03-22llama_model_loader: support multiple split/shard GGUFs (#6187)Pierrick Hymbert
2024-03-22llama : correction of the attn.v.weight quantization for IQ3_XS (#6209)Nexesenex
2024-03-22metal : pad n_ctx by 32 (#6177)Georgi Gerganov
2024-03-18mpt : implement backwards compatiblity with duped output tensor (#6139)Jared Van Bortel
2024-03-18backend : offload large batches to GPU (#6083)slaren
2024-03-15llama : fix Baichuan2 13B (#6092)slaren
2024-03-15llama : add support for control vectors (#5970)Theia Vogel
2024-03-15llama : add Command-R support (#6033)Andrew Canis
2024-03-15fix set main gpu error (#6073)Neo Zhang Jianyu
2024-03-15llama : add Orion chat template (#6066)Xuan Son Nguyen
2024-03-14llama : fix integer overflow during quantization (#6063)Georgi Gerganov
2024-03-14llama : support models without vocabulary (#5798)Michael Podvitskiy
2024-03-14llama : fix typoGeorgi Gerganov
2024-03-14llama : optimize defrag moves + fix fragmentation calculation (#6037)Michael Podvitskiy
2024-03-13llama : add pipeline parallelism support (#6017)slaren
2024-03-11grammar : fix unnecessarily retained pointer to rules (#6003)gliptic
2024-03-11llama : more consistent names of count variables (#5994)Georgi Gerganov
2024-03-11llama : refactor unicode stuff (#5992)Georgi Gerganov