index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
Age
Commit message (
Expand
)
Author
2023-09-03
examples : fix gpt-neox (#2943)
momonga
2023-09-02
server : avoid aniprompt in probabilities of final response (#2849)
Jhen-Jie Hong
2023-09-01
readme : quick start command fix (#2908)
ZHAOKAI WANG
2023-09-01
Allow quantize to only copy tensors, some other improvements (#2931)
Kerfuffle
2023-09-01
llama2c : rename function
Georgi Gerganov
2023-09-01
minor : add const qualifiers (#2853)
m3ndax
2023-09-01
build : fix most gcc and clang warnings (#2861)
Cebtenzzre
2023-09-01
llama2c : fix segfault and alloc-dealloc-mismatch (#2913)
Cebtenzzre
2023-08-31
scripts: Use local gguf package when running from repo (#2927)
Kerfuffle
2023-08-30
examples : fix underscore in beam-search + .gitignore (close #2900)
Georgi Gerganov
2023-08-30
llm.vim : stop generation at multiple linebreaks, bind to <F2> (#2879)
chaihahaha
2023-08-30
main : log file (#2748)
staviq
2023-08-29
Tell users attmepting to run perplexity with too few tokens to use more (#2882)
Kawrakow
2023-08-28
train : mem usage and other improvements (#2439)
xaedes
2023-08-28
llama-bench : set locale to utf8 (#2832)
slaren
2023-08-28
YAML result logging + preset script (#2657)
Johannes Gäßler
2023-08-28
quantize : make output filename optional again (#2823)
Cebtenzzre
2023-08-27
examples : update llama2.c converter to read vocab and write models in GGUF f...
Olivier Chafik
2023-08-27
llama : speedup tokenization (#2831)
Kawrakow
2023-08-27
gguf : add 64-bit support (GGUF v2) (#2821)
Georgi Gerganov
2023-08-27
llama : more tokenizer fixes (#2810)
Georgi Gerganov
2023-08-27
server : add `/detokenize` endpoint (#2802)
Bruce MacDonald
2023-08-26
main : fix bug (penalize_nl=false doesn't work) + suppress warning on mingw (...
Dr. Tom Murphy VII Ph.D
2023-08-26
Fix HellaSwag (#2805)
Kawrakow
2023-08-26
Fix spm whitespaces (#2806)
klosax
2023-08-26
examples : skip unnecessary external lib in server README.md how-to (#2804)
lon
2023-08-25
Faster perplexity computation (#2786)
Kawrakow
2023-08-25
llama : add llama_beam_search() (#2267)
Matt Pulver
2023-08-25
llama-bench : add model sizes (#2771)
slaren
2023-08-25
server : display token probabilities in the UI (#2489)
Jhen-Jie Hong
2023-08-25
ROCm Port (#1087)
Henri Vasserman
2023-08-24
Fix for main example getting stuck when -n -2 and --interactive (#2767)
Kerfuffle
2023-08-23
llm : add Falcon support (#2717)
Georgi Gerganov
2023-08-23
minor : fix trailing whitespace
Georgi Gerganov
2023-08-23
examples : restore the functionality to import llama2.c models (#2685)
Olivier Chafik
2023-08-23
main : insert bos if no tokens (#2727)
klosax
2023-08-23
chmod : make scripts executable (#2675)
Cebtenzzre
2023-08-23
Fix values shown in the quantize tool help (#2735)
Kawrakow
2023-08-23
Strided perplexity (#2714)
Kawrakow
2023-08-23
server : allow json array in prompt or content for direct token input (#2306)
Xiao-Yong Jin
2023-08-22
docs : add grammar docs (#2701)
Evan Jones
2023-08-22
CUDA: use mul_mat_q kernels by default (#2683)
Johannes Gäßler
2023-08-22
embedding : evaluate prompt in batches (#2713)
slaren
2023-08-22
ggml : sync latest (SAM + SD operators, CUDA alibi) (#2709)
Georgi Gerganov
2023-08-22
llama-bench : minor fixes (#2695)
slaren
2023-08-22
server : fallback to default if client param is null (#2688)
Jhen-Jie Hong
2023-08-21
gguf : new file format with flexible meta data (beta) (#2398)
Georgi Gerganov
2023-08-21
HellaSwag: split token evaluation into batches if needed (#2681)
Kawrakow
2023-08-20
More efficient Hellaswag implementation (#2677)
Kawrakow
2023-08-19
server : better default prompt (#2646)
Georgi Gerganov
[next]