index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
/
batched-bench
/
batched-bench.cpp
Age
Commit message (
Expand
)
Author
2025-01-12
MoE fix for R4 quants (#170)
Kawrakow
2024-08-12
Merge mainline - Aug 12 2024 (#17)
Kawrakow
2024-06-04
common : refactor cli arg parsing (#7675)
Georgi Gerganov
2024-04-30
ggml : add Flash Attention (#5021)
Georgi Gerganov
2024-04-05
bench : make n_batch and n_ubatch configurable in Batched bench (#6500)
Ting Sun
2024-03-13
llama : add pipeline parallelism support (#6017)
slaren
2024-03-11
llama : more consistent names of count variables (#5994)
Georgi Gerganov
2024-03-08
llama : support Mamba Selective State Space Models (#5328)
compilade
2024-03-01
llama : cleanup unused mmq flags (#5772)
Pierrick Hymbert
2024-02-18
ggml, common, examples, tests : fixed type arguments in printf (#5528)
Herman Semenov
2024-02-16
ggml : add numa options (#5377)
bmwl
2024-01-31
llama : remove LLAMA_MAX_DEVICES and LLAMA_SUPPORTS_GPU_OFFLOAD (#5240)
Georgi Gerganov
2024-01-12
llama : ggml-backend integration (#4766)
slaren
2023-12-01
ggml : add ggml_soft_max_ext (#4256)
Georgi Gerganov
2023-10-29
Extend llama_kv_cache_seq_rm to allow matching any sequence (#3843)
Kerfuffle
2023-10-25
batched-bench : print params at start
Georgi Gerganov
2023-10-18
speculative : add tree-based sampling example (#3624)
Georgi Gerganov
2023-10-11
batched : add bench tool (#3545)
Georgi Gerganov