summaryrefslogtreecommitdiff
path: root/llama.h
AgeCommit message (Expand)Author
2024-03-03llama : allow for user specified embedding pooling type (#5849)Douglas Hanley
2024-03-02llama : add abort_callback to interrupt computation (#5409)Michael Podvitskiy
2024-03-01llama : cleanup unused mmq flags (#5772)Pierrick Hymbert
2024-02-29llama : constified `llama_set_state_data`'s `src` (#5774)Marcus Dunn
2024-02-28llama : remove deprecated API (#5770)Georgi Gerganov
2024-02-27IQ4_XS: a 4.25 bpw quantization (#5747)Kawrakow
2024-02-27llama : fix defrag bugs + add parameter (#5735)Georgi Gerganov
2024-02-26Adding IQ2_S and IQ2_M to complete coverage of the 2-3 bit quantization range...Kawrakow
2024-02-25llama : refactor k-shift implementation + KV defragmentation (#5691)Georgi Gerganov
2024-02-25code : normalize enum names (#5697)Georgi Gerganov
2024-02-24IQ3_S: a much better alternative to Q3_K (#5676)Kawrakow
2024-02-22Add docs for llama_chat_apply_template (#5645)Xuan Son Nguyen
2024-02-21IQ4_NL: 4-bit non-linear quants with blocks of 32 (#5590)Kawrakow
2024-02-19llama : add llama_chat_apply_template() (#5538)Xuan Son Nguyen
2024-02-181.5 bit quantization (#5453)Kawrakow
2024-02-16ggml : add numa options (#5377)bmwl
2024-02-15Use correct type of pooling for embedding models (#5500)Douglas Hanley
2024-02-13llama : support batched embeddings (#5466)Douglas Hanley
2024-02-11Add support for BERT embedding models (#5423)Douglas Hanley
2024-02-03YaRN : store rope scaling type as int32_t in memory (#5285)Jared Van Bortel
2024-01-31llama : remove LLAMA_MAX_DEVICES and LLAMA_SUPPORTS_GPU_OFFLOAD (#5240)Georgi Gerganov
2024-01-30SOTA 3-bit quants (#5196)Kawrakow
2024-01-29Nomic Vulkan backend (#4456)Jared Van Bortel
2024-01-28ggml : add Vulkan backend (#2059)0cc4m
2024-01-28ggml : add unified SYCL backend for Intel GPUs (#2690)Abhilash Majumder
2024-01-25llama : dynamic temperature sampling (#4972)l3utterfly
2024-01-22llama : add Q3_K_XS (#5060)Kawrakow
2024-01-17backend : add eval callback (#4935)Georgi Gerganov
2024-01-15llama : apply classifier-free guidance to logits directly (#4951)David Friehs
2024-01-142-bit quantizations (#4897)Kawrakow
2024-01-13llama : minimize size used for state save/load (#4820)David Friehs
2024-01-12llama : ggml-backend integration (#4766)slaren
2024-01-11llama : restore intended k-quants mixes for MoE models (#4872)Kawrakow
2024-01-11ggml : SOTA 2-bit quants (add IQ2_XS) (#4856)Kawrakow
2024-01-08SOTA 2-bit quants (#4773)Kawrakow
2024-01-08main : add self-extend support (#4815)Georgi Gerganov
2024-01-08examples : add passkey test (#3856)Georgi Gerganov
2024-01-02llama : replace all API facing `int`'s with `int32_t` (#4577)Marcus Dunn
2023-12-22llama : add ability to cancel model loading (#4462)crasm
2023-12-21llama : allow getting n_batch from llama_context in c api (#4540)Marcus Dunn
2023-12-16lora : add support for non-llama models (#3333)slaren
2023-12-12llama : document logits_all deprecation (#4418)crasm
2023-12-07llama : per-layer KV cache + quantum K cache (#4309)Georgi Gerganov
2023-12-05llama : allow overriding GGUF metadata when loading model (#4092)Kerfuffle
2023-11-25Update docs for yarn_ext_factor <0.0 as unspecified instead of NaN (#4189)crasm
2023-11-23llama : KV cache view API + better KV cache management (#4170)Georgi Gerganov
2023-11-17llama : add functions to get the model's metadata (#4013)slaren
2023-11-16Respect tokenizer.ggml.add_bos_token value when tokenizing (#4040)Kerfuffle
2023-11-03common : YAYF (yet another YARN fix) (#3925)Georgi Gerganov
2023-11-01llama : implement YaRN RoPE scaling (#2268)cebtenzzre