summaryrefslogtreecommitdiff
path: root/examples/CMakeLists.txt
AgeCommit message (Expand)Author
2025-02-23Add new sweep-bench benchmark (#225)saood06
2024-07-27Merge mainline llama.cpp (#3)Kawrakow
2024-06-15Add `cvector-generator` example (#7514)Xuan Son Nguyen
2024-06-13`build`: rename main → llama-cli, server → llama-server, llava-cli → ll...Olivier Chafik
2024-06-04llama : remove beam search (#7736)Georgi Gerganov
2024-05-14ggml : add RPC backend (#6829)Radoslav Gerganov
2024-04-11eval-callback: Example how to use eval callback for debugging (#6576)Pierrick Hymbert
2024-03-25examples : add "retrieval" (#6193)Minsoo Cheong
2024-03-19gguf-split: split and merge gguf per batch of tensors (#6135)Pierrick Hymbert
2024-03-10llama : add support for GritLM (#5959)DAN™
2024-02-13gguf : add python reader example (#5216)John
2024-01-28ggml : add unified SYCL backend for Intel GPUs (#2690)Abhilash Majumder
2024-01-13metal : remove old API (#4919)Georgi Gerganov
2024-01-12Importance Matrix calculation (#4861)Kawrakow
2024-01-08examples : add passkey test (#3856)Georgi Gerganov
2023-12-22lookup : add prompt lookup decoding example (#4484)LeonEricsson
2023-11-26lookahead : add example for lookahead decoding (#4207)Georgi Gerganov
2023-11-17examples : add tokenize (#4039)zakkor
2023-10-20sampling : refactor init to use llama_sampling_params (#3696)Georgi Gerganov
2023-10-12examples: support LLaVA v1.5 (multimodal model) (#3436)M. Yusuf Sarıgöz
2023-10-11batched : add bench tool (#3545)Georgi Gerganov
2023-09-28train : finetune LORA (#2632)xaedes
2023-09-28llama : custom attention mask + parallel decoding + no context swaps (#3228)Georgi Gerganov
2023-09-03speculative : PoC for speeding-up inference via speculative sampling (#2926)Georgi Gerganov
2023-08-30examples : fix underscore in beam-search + .gitignore (close #2900)Georgi Gerganov
2023-08-25llama : add llama_beam_search() (#2267)Matt Pulver
2023-08-21gguf : new file format with flexible meta data (beta) (#2398)Georgi Gerganov
2023-08-18llama : add benchmark example (#2626)slaren
2023-08-12Adding support for llama2.c models (#2559)byte-6174
2023-08-04Add --simple-io option for subprocesses and break out console.h and cpp (#1558)DannyDaemonic
2023-07-23llama : add grammar-based sampling (#1773)Evan Jones
2023-06-28llama : support input embeddings directly (#1910)ningshanwutuobang
2023-06-17llama : fix kv_cache `n` init (close #1903)Georgi Gerganov
2023-06-13train : improved training-from-scratch example (#1652)xaedes
2023-06-04llama : Metal inference (#1642)Georgi Gerganov
2023-05-21examples : add server example with REST API (#1443)Steward Garcia
2023-05-13ggml : implement backward pass for llama + small training-llama-from-scratch ...xaedes
2023-04-30Various fixes to mat_mul benchmark (#1253)Stephan Walter
2023-04-24examples : add save_load_state example (#1150)xaedes
2023-04-08Add quantize-stats command for testing quantization (#728)unbounded
2023-03-25Overhaul the examples structureGeorgi Gerganov