summaryrefslogtreecommitdiff
path: root/examples
AgeCommit message (Expand)Author
2024-03-26IQ1_M: 1.75 bpw quantization (#6302)Kawrakow
2024-03-26quantize : be able to override metadata by key (#6321)Kawrakow
2024-03-26embedding : adjust `n_ubatch` value (#6296)Minsoo Cheong
2024-03-26server : add `n_discard` parameter (#6300)Jan Boon
2024-03-26cuda : rename build flag to LLAMA_CUDA (#6299)slaren
2024-03-25Server: clean up OAI params parsing function (#6284)Xuan Son Nguyen
2024-03-25[SYCL] fix SYCL backend build on windows is break by LOG() error (#6290)Neo Zhang Jianyu
2024-03-25examples : add "retrieval" (#6193)Minsoo Cheong
2024-03-24imatrix : fix wname for mul_mat_id ops (#6271)Georgi Gerganov
2024-03-24sampling : deduplicated code for probability distribution access (#6240)Minsoo Cheong
2024-03-23common: llama_load_model_from_url split support (#6192)Pierrick Hymbert
2024-03-23server: docs: `--threads` and `--threads`, `--ubatch-size`, `--log-disable` (...Pierrick Hymbert
2024-03-23server: flush stdout after logging in both text and json layout (#6253)Pierrick Hymbert
2024-03-23lookup: complement data from context with general text statistics (#5479)Johannes Gäßler
2024-03-22convert-llama2c-to-ggml : enable conversion of GQA models (#6237)fraxy-v
2024-03-22quantize: options for output and token embedding tensors qtype (#6239)Kawrakow
2024-03-22llama_model_loader: support multiple split/shard GGUFs (#6187)Pierrick Hymbert
2024-03-22json-schema-to-grammar : fix order of props + non-str const/enum (#6232)Olivier Chafik
2024-03-22server : fix n_keep always showing as 0 in response (#6211)Jan Boon
2024-03-22server : enable continuous batching by default (#6231)Georgi Gerganov
2024-03-22metal : pad n_ctx by 32 (#6177)Georgi Gerganov
2024-03-21server : update readme doc from `slot_id` to `id_slot` (#6213)Jan Boon
2024-03-21json-schema-to-grammar improvements (+ added to server) (#5978)Olivier Chafik
2024-03-21Add ability to use Q5_0, Q5_1, and IQ4_NL for quantized K cache (#6183)Kawrakow
2024-03-20llava : update MobileVLM-README.md (#6180)Ziang Wu
2024-03-20llava : add MobileVLM_V2 backup (#6175)Ziang Wu
2024-03-20Server: version bump for httplib and json (#6169)Xuan Son Nguyen
2024-03-20server : allow to override -ngl in tests (#6170)Georgi Gerganov
2024-03-20Revert "llava : add a MobileVLM_V2-1.7B backup (#6152)"Georgi Gerganov
2024-03-20llava : add a MobileVLM_V2-1.7B backup (#6152)Ziang Wu
2024-03-20Server: Handle n_keep parameter in the request (#6174)Karthick
2024-03-20server tests : more pythonic process management; fix bare `except:` (#6146)Jared Van Bortel
2024-03-20update readme sycl for new update (#6151)Neo Zhang Jianyu
2024-03-19Remove undeed header file. (#6158)DAN™
2024-03-19gguf-split: split and merge gguf per batch of tensors (#6135)Pierrick Hymbert
2024-03-18clip : fix memory leak (#6138)Felix
2024-03-18backend : offload large batches to GPU (#6083)slaren
2024-03-17common: llama_load_model_from_url using --model-url (#6098)Pierrick Hymbert
2024-03-16gritlm : add initial README.md (#6086)Daniel Bevenius
2024-03-15llava : change API to pure C style for Rust FFI bindgen (#6079)Ting Lou
2024-03-15fix set main gpu error (#6073)Neo Zhang Jianyu
2024-03-15llama-bench : use random tokens to improve accuracy with mixtral (#6069)slaren
2024-03-14gguf : fix resource leaks (#6061)Steve Grubb
2024-03-14embedding : add EOS token if not present (#899)Georgi Gerganov
2024-03-14readme : improve readme for Llava-1.6 example (#6044)Jian Liao
2024-03-14server: disable debug release type sanitizer, simplify trigger (#6047)Pierrick Hymbert
2024-03-14embedding : print all resulting embeddings (#899)Georgi Gerganov
2024-03-14embedding : print cosine similarity (#899)Georgi Gerganov
2024-03-13llama : add pipeline parallelism support (#6017)slaren
2024-03-13Server: Use multi-task for embeddings endpoint (#6001)Xuan Son Nguyen