summaryrefslogtreecommitdiff
path: root/examples/llama-bench/llama-bench.cpp
AgeCommit message (Expand)Author
2023-11-02build : link against build info instead of compiling against it (#3879)cebtenzzre
2023-10-29Extend llama_kv_cache_seq_rm to allow matching any sequence (#3843)Kerfuffle
2023-10-23llama : remove token functions with `context` args in favor of `model` (#3720)Marcus Dunn
2023-09-28build : enable more non-default compiler warnings (#3200)Cebtenzzre
2023-09-28llama.cpp : split llama_context_params into model and context params (#3301)slaren
2023-09-28llama : custom attention mask + parallel decoding + no context swaps (#3228)Georgi Gerganov
2023-09-27metal : reusing llama.cpp logging (#3152)Rickard Hallerbäck
2023-09-15sync : ggml (Metal F32 support + reduce ggml-alloc size) (#3192)Georgi Gerganov
2023-09-07llama-bench : use two tokens in the warmup run for prompt evals (#3059)slaren
2023-09-05examples : replace fprintf to stdout with printf (#3017)Cebtenzzre
2023-09-04llama-bench : make cpp file non-executable (#2999)Cebtenzzre
2023-08-28llama-bench : set locale to utf8 (#2832)slaren
2023-08-25llama-bench : add model sizes (#2771)slaren
2023-08-25ROCm Port (#1087)Henri Vasserman
2023-08-22llama-bench : minor fixes (#2695)slaren
2023-08-21gguf : new file format with flexible meta data (beta) (#2398)Georgi Gerganov
2023-08-18llama : add benchmark example (#2626)slaren