index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
/
batched
/
batched.cpp
Age
Commit message (
Expand
)
Author
2024-07-27
Merge mainline llama.cpp (#3)
Kawrakow
2024-06-04
common : refactor cli arg parsing (#7675)
Georgi Gerganov
2024-05-22
common : normalize naming style (#7462)
Georgi Gerganov
2024-04-21
llama : support Llama 3 HF conversion (#6745)
Pedro Cuenca
2024-03-22
metal : pad n_ctx by 32 (#6177)
Georgi Gerganov
2024-03-11
llama : more consistent names of count variables (#5994)
Georgi Gerganov
2024-03-08
llama : support Mamba Selective State Space Models (#5328)
compilade
2024-02-18
ggml, common, examples, tests : fixed type arguments in printf (#5528)
Herman Semenov
2024-02-16
ggml : add numa options (#5377)
bmwl
2024-01-08
examples : add passkey test (#3856)
Georgi Gerganov
2023-10-24
cuda : add batched cuBLAS GEMM for faster attention (#3749)
Georgi Gerganov
2023-10-23
llama : remove token functions with `context` args in favor of `model` (#3720)
Marcus Dunn
2023-10-22
batched : add len CLI argument
Georgi Gerganov
2023-10-18
speculative : add tree-based sampling example (#3624)
Georgi Gerganov
2023-10-11
batched : add bench tool (#3545)
Georgi Gerganov
2023-09-28
llama.cpp : split llama_context_params into model and context params (#3301)
slaren
2023-09-28
llama : custom attention mask + parallel decoding + no context swaps (#3228)
Georgi Gerganov