index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
/
main
Age
Commit message (
Expand
)
Author
2024-05-10
Fix memory bug in grammar parser (#7194)
Justine Tunney
2024-05-10
Main+: optionally allow special tokens from user in interactive mode (#7097)
HanishKVC
2024-05-08
main : add --conversation / -cnv flag (#7108)
Dawid Potocki
2024-05-07
main : update log text (EOS to EOG) (#7104)
RhinoDevel
2024-05-07
docs: fix typos (#7124)
omahs
2024-05-01
main : fix off by one error for context shift (#6921)
l3utterfly
2024-04-30
Improve usability of --model-url & related flags (#6930)
Olivier Chafik
2024-04-29
main : fix typo in comment in main.cpp (#6985)
Daniel Bevenius
2024-04-24
Server: fix seed for multiple slots (#6835)
Johannes Gäßler
2024-04-21
llama : support Llama 3 HF conversion (#6745)
Pedro Cuenca
2024-04-15
`main`: add --json-schema / -j flag (#6659)
Olivier Chafik
2024-04-12
chore: Fix markdown warnings (#6625)
Rene Leonhardt
2024-04-09
BERT tokenizer fixes (#6498)
Jared Van Bortel
2024-04-08
llama : save and restore kv cache for single seq id (#6341)
Jan Boon
2024-03-28
doc: fix outdated default value of batch size (#6336)
Ting Sun
2024-03-26
cuda : rename build flag to LLAMA_CUDA (#6299)
slaren
2024-03-17
common: llama_load_model_from_url using --model-url (#6098)
Pierrick Hymbert
2024-03-11
llama : more consistent names of count variables (#5994)
Georgi Gerganov
2024-03-04
main : support special tokens as reverse/anti prompt (#5847)
DAN™
2024-02-25
llama : refactor k-shift implementation + KV defragmentation (#5691)
Georgi Gerganov
2024-02-21
examples : do not assume BOS when shifting context (#5622)
Jared Van Bortel
2024-02-16
ggml : add numa options (#5377)
bmwl
2024-02-11
main : ctrl+C print timing in non-interactive mode (#3873)
Georgi Gerganov
2024-02-03
refactor : switch to emplace_back to avoid extra object (#5291)
Michael Klimenko
2024-01-30
main : allow empty --prompt-cache file (#5176)
divinity76
2024-01-13
main : add parameter --no-display-prompt (#4541)
Yann Follet
2024-01-11
main : better name for variable n_print (#4874)
Georgi Gerganov
2024-01-11
main : disable token count by default (#4874)
Georgi Gerganov
2024-01-11
main : print total token count and tokens consumed so far (#4874)
pudepiedj
2024-01-08
main : add self-extend support (#4815)
Georgi Gerganov
2023-12-05
sampling : custom samplers order (#4285)
MaggotHATE
2023-11-30
main : pass LOG_TEE callback to llama.cpp log (#4033)
Andrew Godfrey
2023-11-20
main : Add ChatML functionality to main example (#4046)
Seb C
2023-11-16
Respect tokenizer.ggml.add_bos_token value when tokenizing (#4040)
Kerfuffle
2023-11-11
Fix some documentation typos/grammar mistakes (#4032)
Richard Kiss
2023-11-02
build : link against build info instead of compiling against it (#3879)
cebtenzzre
2023-10-31
samplers : Min-P sampler implementation [alternative to Top P/Top K] (#3841)
kalomaze
2023-10-29
Extend llama_kv_cache_seq_rm to allow matching any sequence (#3843)
Kerfuffle
2023-10-23
llama : remove token functions with `context` args in favor of `model` (#3720)
Marcus Dunn
2023-10-22
main : escape prompt for cfg_negative_prompt and consecutive inputs in main w...
vvhg1
2023-10-20
sampling : refactor init to use llama_sampling_params (#3696)
Georgi Gerganov
2023-10-18
speculative : add tree-based sampling example (#3624)
Georgi Gerganov
2023-10-17
llama : avoid fprintf in favor of LLAMA_LOG (#3538)
Georgi Gerganov
2023-10-17
tokenizer : special token handling (#3538)
staviq
2023-10-11
main : fix session loading bug (#3400)
Georgi Gerganov
2023-10-11
common : fix mirostat state when using multiple sequences (#3543)
Kerfuffle
2023-10-03
main : consistent prefix/suffix coloring (#3425)
h-h-h-h
2023-10-03
llama : fix session saving/loading (#3400)
Georgi Gerganov
2023-09-28
build : enable more non-default compiler warnings (#3200)
Cebtenzzre
2023-09-28
llama.cpp : split llama_context_params into model and context params (#3301)
slaren
[next]