index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
/
main
/
README.md
Age
Commit message (
Expand
)
Author
2024-05-07
docs: fix typos (#7124)
omahs
2024-04-30
Improve usability of --model-url & related flags (#6930)
Olivier Chafik
2024-04-15
`main`: add --json-schema / -j flag (#6659)
Olivier Chafik
2024-04-12
chore: Fix markdown warnings (#6625)
Rene Leonhardt
2024-03-28
doc: fix outdated default value of batch size (#6336)
Ting Sun
2024-03-26
cuda : rename build flag to LLAMA_CUDA (#6299)
slaren
2024-03-17
common: llama_load_model_from_url using --model-url (#6098)
Pierrick Hymbert
2024-02-16
ggml : add numa options (#5377)
bmwl
2023-11-11
Fix some documentation typos/grammar mistakes (#4032)
Richard Kiss
2023-10-31
samplers : Min-P sampler implementation [alternative to Top P/Top K] (#3841)
kalomaze
2023-09-28
llama.cpp : split llama_context_params into model and context params (#3301)
slaren
2023-09-15
llama : remove mtest (#3177)
Roland
2023-09-01
readme : quick start command fix (#2908)
ZHAOKAI WANG
2023-08-22
docs : add grammar docs (#2701)
Evan Jones
2023-08-10
Add --n-predict -2 for stopping generation on full context (#2565)
Christian Demsar
2023-08-07
Add --rope-scale parameter (#2544)
klosax
2023-07-28
readme : fix the description of the Tail free sampling (TFS) method (#2431)
Weird Constructor
2023-07-13
Revert "Support using mmap when applying LoRA (#2095)" (#2206)
Howard Su
2023-07-11
Support using mmap when applying LoRA (#2095)
Howard Su
2023-06-29
Use unsigned for random seed (#2006)
Howard Su
2023-06-26
ggml : add NUMA support (#1556)
zrm
2023-06-14
CUDA full GPU acceleration, KV cache in VRAM (#1827)
Johannes Gäßler
2023-06-06
Multi GPU support, CUDA refactor, CUDA scratch buffer (#1703)
Johannes Gäßler
2023-05-28
Only show -ngl option when relevant + other doc/arg handling updates (#1625)
Kerfuffle
2023-05-25
Some improvements to loading the session with --prompt-cache (#1550)
Kerfuffle
2023-05-10
main : add option to save full output to session (#1338)
Evan Jones
2023-05-04
main : add --in-suffix option (#1318)
44670
2023-05-04
Only escape prompts when used with `-e` (#1311)
DannyDaemonic
2023-05-04
Update main's README.md with new features (#1296)
DannyDaemonic
2023-05-02
llama : allow 0 as a seed number. (#1275)
Robert Brisita
2023-04-24
examples/main README improvements and some light refactoring (#1131)
mgroeber9110
2023-04-23
Fix LoRA acronym (#1145)
slaren
2023-04-23
Added README.md for main with examples and explanations (#1139)
DannyDaemonic
2023-04-11
Fix whitespace, add .editorconfig, add GitHub workflow (#883)
Pavol Rusnak
2023-03-25
Overhaul the examples structure
Georgi Gerganov