summaryrefslogtreecommitdiff
path: root/convert-hf-to-gguf.py
AgeCommit message (Expand)Author
2024-05-21tests : test-tokenizer-0.sh print more info (#7402)Georgi Gerganov
2024-05-21Tokenizer SPM fixes for phi-3 and llama-spm (bugfix) (#7425)jaime-m-p
2024-05-20Tokenizer SPM fixes for phi-3 and llama-spm (#7375)jaime-m-p
2024-05-21llama : remove Persimmon (#7408)Georgi Gerganov
2024-05-19Add StableLM2 pre-tokenizer (#7349)Anas Ahouzi
2024-05-18convert : fix set_vocab_sentencepiece (#6866)Georgi Gerganov
2024-05-17py : convert-hf-to-gguf-update improvements (#7340)Aarni Koskela
2024-05-17convert : fix Qwen/Qwen-7b conversion (#7308)amd-lalithnc
2024-05-13convert-hf : support direct Q8_0 conversion (#7234)compilade
2024-05-13llama : rename jina tokenizers to v2 (#7249)Joan Fontanals
2024-05-11convert-hf : support bfloat16 conversion (#7158)compilade
2024-05-11llama : add Jina Embeddings architecture (#6826)Joan Fontanals
2024-05-11ggml : full ALiBi support (#7192)Georgi Gerganov
2024-05-08convert-hf : save memory with lazy evaluation (#7075)compilade
2024-05-08llama : add BPE pre-tokenization for Qwen2 (#7114)Ren Xuancheng
2024-05-08convert : add BPE pre-tokenization for DBRX (#7132)DAN™
2024-05-07Fix OLMo HF to GGUF conversion (#6910)nopperl
2024-05-05command-r : add BPE pre-tokenization (#7063)DAN™
2024-05-04tests : add test-tokenizer-0.sh + fix some tokenizers (#7036)Georgi Gerganov
2024-05-03convert.py : add python logging instead of print() (#6511)Brian
2024-05-03Remove .attention from skipped tensors to match more accurately (#7051)Bartowski
2024-04-30convert : use utf8 encoding (#7000)Georgi Gerganov
2024-04-29llama : fix BPE pre-tokenization (#6920)Georgi Gerganov
2024-04-29convert : fix conversion of some BERT embedding models (#6937)Christian Zhou-Zheng
2024-04-24convert : add support of codeqwen due to tokenizer (#6707)Junyang Lin
2024-04-24llama : add phi3 support (#6852)liuwei-git
2024-04-21llama : support Llama 3 HF conversion (#6745)Pedro Cuenca
2024-04-19Implement the OLMo architecture (#6741)nopperl
2024-04-16convert : fix autoawq gemma (#6704)Zheng.Deng
2024-04-16llama : add StableLM2 12B (#6635)Ashish
2024-04-16llama : add qwen2moe (#6074)Shijie
2024-04-16gguf : add special tokens metadata for FIM/Infill (#6689)Daniel Bevenius
2024-04-14convert : enable the `--use-temp-file` cli flag (#6645)James A Capozzoli
2024-04-13model: support arch `DbrxForCausalLM` (#6515)Pierrick Hymbert
2024-04-09BERT tokenizer fixes (#6498)Jared Van Bortel
2024-04-09llama : add Command R Plus support (#6491)Carolinabanana
2024-04-04convert : fix for lint error complaining of bare except (#6470)Clint Herron
2024-04-03llama : add SEA-LION support (#6448)bryanSwk
2024-04-03Missing tokenizer.model error during gguf conversion (#6443)Abhishek Gopinath K
2024-04-03ggml : mul_mat_id use the same tensor for all the experts (#6387)slaren
2024-03-29[Model] Add support for xverse (#6301)hxer7963
2024-03-29convert : allow conversion of Mistral HF models (#6144)Pedro Cuenca
2024-03-28convert : refactor vocab selection logic (#6355)Jared Van Bortel
2024-03-26convert-hf : fix exception in sentencepiece with added tokens (#6320)Pedro Cuenca
2024-03-23llama : add grok-1 support (#6204)Julius Arkenberg
2024-03-18convert : add support for CamembertModel architecture (#6119)Thérence
2024-03-15llama : add Command-R support (#6033)Andrew Canis
2024-03-08llama : support Mamba Selective State Space Models (#5328)compilade
2024-03-04flake : fixGeorgi Gerganov
2024-03-03llama : allow for user specified embedding pooling type (#5849)Douglas Hanley