summaryrefslogtreecommitdiff
path: root/convert-hf-to-gguf.py
AgeCommit message (Expand)Author
2024-06-17update: support Qwen2-57B-A14B (#7835)Ștefan-Gabriel Muscalu
2024-06-14convert : add Poro-34B-chat tokenizer support (#7713)Elaine
2024-06-09convert-hf : set the model name based on cli arg, if present (#7693)sasha0552
2024-06-09convert-hf : match model part name prefix and suffix (#7687)compilade
2024-06-09gguf-py : decouple adding metadata from writing in GGUFWriter (#7827)compilade
2024-06-06llama : add jina v2 base code (#7596)Joan Fontanals
2024-06-06Fix encoding in python scripts (#7733)Galunid
2024-05-31convert-hf : Handle NotImplementedError in convert-hf-to-gguf (#7660)Galunid
2024-05-30Move convert.py to examples/convert-legacy-llama.py (#7430)Galunid
2024-05-28llama : support small Granite models (#7481)Giuseppe Scrivano
2024-05-28Add support for DeepseekV2ForCausalLM (#7519)fairydreaming
2024-05-26Fix aya-23 conversion scripts (#7539)Galunid
2024-05-26llama : add Smaug 70B support (#7402)Bartowski
2024-05-25gguf-py : fix and simplify quantized shape round-trip (#7483)compilade
2024-05-24Add support for ArcticForCausalLM (#7020)fairydreaming
2024-05-23Add missing inference support for GPTNeoXForCausalLM (Pythia and GPT-NeoX bas...fairydreaming
2024-05-21llama : add phi3 128K model support (#7225)liuwei-git
2024-05-21tests : test-tokenizer-0.sh print more info (#7402)Georgi Gerganov
2024-05-21Tokenizer SPM fixes for phi-3 and llama-spm (bugfix) (#7425)jaime-m-p
2024-05-20Tokenizer SPM fixes for phi-3 and llama-spm (#7375)jaime-m-p
2024-05-21llama : remove Persimmon (#7408)Georgi Gerganov
2024-05-19Add StableLM2 pre-tokenizer (#7349)Anas Ahouzi
2024-05-18convert : fix set_vocab_sentencepiece (#6866)Georgi Gerganov
2024-05-17py : convert-hf-to-gguf-update improvements (#7340)Aarni Koskela
2024-05-17convert : fix Qwen/Qwen-7b conversion (#7308)amd-lalithnc
2024-05-13convert-hf : support direct Q8_0 conversion (#7234)compilade
2024-05-13llama : rename jina tokenizers to v2 (#7249)Joan Fontanals
2024-05-11convert-hf : support bfloat16 conversion (#7158)compilade
2024-05-11llama : add Jina Embeddings architecture (#6826)Joan Fontanals
2024-05-11ggml : full ALiBi support (#7192)Georgi Gerganov
2024-05-08convert-hf : save memory with lazy evaluation (#7075)compilade
2024-05-08llama : add BPE pre-tokenization for Qwen2 (#7114)Ren Xuancheng
2024-05-08convert : add BPE pre-tokenization for DBRX (#7132)DAN™
2024-05-07Fix OLMo HF to GGUF conversion (#6910)nopperl
2024-05-05command-r : add BPE pre-tokenization (#7063)DAN™
2024-05-04tests : add test-tokenizer-0.sh + fix some tokenizers (#7036)Georgi Gerganov
2024-05-03convert.py : add python logging instead of print() (#6511)Brian
2024-05-03Remove .attention from skipped tensors to match more accurately (#7051)Bartowski
2024-04-30convert : use utf8 encoding (#7000)Georgi Gerganov
2024-04-29llama : fix BPE pre-tokenization (#6920)Georgi Gerganov
2024-04-29convert : fix conversion of some BERT embedding models (#6937)Christian Zhou-Zheng
2024-04-24convert : add support of codeqwen due to tokenizer (#6707)Junyang Lin
2024-04-24llama : add phi3 support (#6852)liuwei-git
2024-04-21llama : support Llama 3 HF conversion (#6745)Pedro Cuenca
2024-04-19Implement the OLMo architecture (#6741)nopperl
2024-04-16convert : fix autoawq gemma (#6704)Zheng.Deng
2024-04-16llama : add StableLM2 12B (#6635)Ashish
2024-04-16llama : add qwen2moe (#6074)Shijie
2024-04-16gguf : add special tokens metadata for FIM/Infill (#6689)Daniel Bevenius
2024-04-14convert : enable the `--use-temp-file` cli flag (#6645)James A Capozzoli