index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
convert-hf-to-gguf.py
Age
Commit message (
Expand
)
Author
2024-06-17
update: support Qwen2-57B-A14B (#7835)
Ștefan-Gabriel Muscalu
2024-06-14
convert : add Poro-34B-chat tokenizer support (#7713)
Elaine
2024-06-09
convert-hf : set the model name based on cli arg, if present (#7693)
sasha0552
2024-06-09
convert-hf : match model part name prefix and suffix (#7687)
compilade
2024-06-09
gguf-py : decouple adding metadata from writing in GGUFWriter (#7827)
compilade
2024-06-06
llama : add jina v2 base code (#7596)
Joan Fontanals
2024-06-06
Fix encoding in python scripts (#7733)
Galunid
2024-05-31
convert-hf : Handle NotImplementedError in convert-hf-to-gguf (#7660)
Galunid
2024-05-30
Move convert.py to examples/convert-legacy-llama.py (#7430)
Galunid
2024-05-28
llama : support small Granite models (#7481)
Giuseppe Scrivano
2024-05-28
Add support for DeepseekV2ForCausalLM (#7519)
fairydreaming
2024-05-26
Fix aya-23 conversion scripts (#7539)
Galunid
2024-05-26
llama : add Smaug 70B support (#7402)
Bartowski
2024-05-25
gguf-py : fix and simplify quantized shape round-trip (#7483)
compilade
2024-05-24
Add support for ArcticForCausalLM (#7020)
fairydreaming
2024-05-23
Add missing inference support for GPTNeoXForCausalLM (Pythia and GPT-NeoX bas...
fairydreaming
2024-05-21
llama : add phi3 128K model support (#7225)
liuwei-git
2024-05-21
tests : test-tokenizer-0.sh print more info (#7402)
Georgi Gerganov
2024-05-21
Tokenizer SPM fixes for phi-3 and llama-spm (bugfix) (#7425)
jaime-m-p
2024-05-20
Tokenizer SPM fixes for phi-3 and llama-spm (#7375)
jaime-m-p
2024-05-21
llama : remove Persimmon (#7408)
Georgi Gerganov
2024-05-19
Add StableLM2 pre-tokenizer (#7349)
Anas Ahouzi
2024-05-18
convert : fix set_vocab_sentencepiece (#6866)
Georgi Gerganov
2024-05-17
py : convert-hf-to-gguf-update improvements (#7340)
Aarni Koskela
2024-05-17
convert : fix Qwen/Qwen-7b conversion (#7308)
amd-lalithnc
2024-05-13
convert-hf : support direct Q8_0 conversion (#7234)
compilade
2024-05-13
llama : rename jina tokenizers to v2 (#7249)
Joan Fontanals
2024-05-11
convert-hf : support bfloat16 conversion (#7158)
compilade
2024-05-11
llama : add Jina Embeddings architecture (#6826)
Joan Fontanals
2024-05-11
ggml : full ALiBi support (#7192)
Georgi Gerganov
2024-05-08
convert-hf : save memory with lazy evaluation (#7075)
compilade
2024-05-08
llama : add BPE pre-tokenization for Qwen2 (#7114)
Ren Xuancheng
2024-05-08
convert : add BPE pre-tokenization for DBRX (#7132)
DAN™
2024-05-07
Fix OLMo HF to GGUF conversion (#6910)
nopperl
2024-05-05
command-r : add BPE pre-tokenization (#7063)
DAN™
2024-05-04
tests : add test-tokenizer-0.sh + fix some tokenizers (#7036)
Georgi Gerganov
2024-05-03
convert.py : add python logging instead of print() (#6511)
Brian
2024-05-03
Remove .attention from skipped tensors to match more accurately (#7051)
Bartowski
2024-04-30
convert : use utf8 encoding (#7000)
Georgi Gerganov
2024-04-29
llama : fix BPE pre-tokenization (#6920)
Georgi Gerganov
2024-04-29
convert : fix conversion of some BERT embedding models (#6937)
Christian Zhou-Zheng
2024-04-24
convert : add support of codeqwen due to tokenizer (#6707)
Junyang Lin
2024-04-24
llama : add phi3 support (#6852)
liuwei-git
2024-04-21
llama : support Llama 3 HF conversion (#6745)
Pedro Cuenca
2024-04-19
Implement the OLMo architecture (#6741)
nopperl
2024-04-16
convert : fix autoawq gemma (#6704)
Zheng.Deng
2024-04-16
llama : add StableLM2 12B (#6635)
Ashish
2024-04-16
llama : add qwen2moe (#6074)
Shijie
2024-04-16
gguf : add special tokens metadata for FIM/Infill (#6689)
Daniel Bevenius
2024-04-14
convert : enable the `--use-temp-file` cli flag (#6645)
James A Capozzoli
[next]