summaryrefslogtreecommitdiff
path: root/convert-hf-to-gguf.py
AgeCommit message (Expand)Author
2024-05-13convert-hf : support direct Q8_0 conversion (#7234)compilade
2024-05-13llama : rename jina tokenizers to v2 (#7249)Joan Fontanals
2024-05-11convert-hf : support bfloat16 conversion (#7158)compilade
2024-05-11llama : add Jina Embeddings architecture (#6826)Joan Fontanals
2024-05-11ggml : full ALiBi support (#7192)Georgi Gerganov
2024-05-08convert-hf : save memory with lazy evaluation (#7075)compilade
2024-05-08llama : add BPE pre-tokenization for Qwen2 (#7114)Ren Xuancheng
2024-05-08convert : add BPE pre-tokenization for DBRX (#7132)DAN™
2024-05-07Fix OLMo HF to GGUF conversion (#6910)nopperl
2024-05-05command-r : add BPE pre-tokenization (#7063)DAN™
2024-05-04tests : add test-tokenizer-0.sh + fix some tokenizers (#7036)Georgi Gerganov
2024-05-03convert.py : add python logging instead of print() (#6511)Brian
2024-05-03Remove .attention from skipped tensors to match more accurately (#7051)Bartowski
2024-04-30convert : use utf8 encoding (#7000)Georgi Gerganov
2024-04-29llama : fix BPE pre-tokenization (#6920)Georgi Gerganov
2024-04-29convert : fix conversion of some BERT embedding models (#6937)Christian Zhou-Zheng
2024-04-24convert : add support of codeqwen due to tokenizer (#6707)Junyang Lin
2024-04-24llama : add phi3 support (#6852)liuwei-git
2024-04-21llama : support Llama 3 HF conversion (#6745)Pedro Cuenca
2024-04-19Implement the OLMo architecture (#6741)nopperl
2024-04-16convert : fix autoawq gemma (#6704)Zheng.Deng
2024-04-16llama : add StableLM2 12B (#6635)Ashish
2024-04-16llama : add qwen2moe (#6074)Shijie
2024-04-16gguf : add special tokens metadata for FIM/Infill (#6689)Daniel Bevenius
2024-04-14convert : enable the `--use-temp-file` cli flag (#6645)James A Capozzoli
2024-04-13model: support arch `DbrxForCausalLM` (#6515)Pierrick Hymbert
2024-04-09BERT tokenizer fixes (#6498)Jared Van Bortel
2024-04-09llama : add Command R Plus support (#6491)Carolinabanana
2024-04-04convert : fix for lint error complaining of bare except (#6470)Clint Herron
2024-04-03llama : add SEA-LION support (#6448)bryanSwk
2024-04-03Missing tokenizer.model error during gguf conversion (#6443)Abhishek Gopinath K
2024-04-03ggml : mul_mat_id use the same tensor for all the experts (#6387)slaren
2024-03-29[Model] Add support for xverse (#6301)hxer7963
2024-03-29convert : allow conversion of Mistral HF models (#6144)Pedro Cuenca
2024-03-28convert : refactor vocab selection logic (#6355)Jared Van Bortel
2024-03-26convert-hf : fix exception in sentencepiece with added tokens (#6320)Pedro Cuenca
2024-03-23llama : add grok-1 support (#6204)Julius Arkenberg
2024-03-18convert : add support for CamembertModel architecture (#6119)Thérence
2024-03-15llama : add Command-R support (#6033)Andrew Canis
2024-03-08llama : support Mamba Selective State Space Models (#5328)compilade
2024-03-04flake : fixGeorgi Gerganov
2024-03-03llama : allow for user specified embedding pooling type (#5849)Douglas Hanley
2024-03-02convert-hf : make model class definitions self-contained (#5825)Jared Van Bortel
2024-03-01llama : add StarCoder2 support (#5795)Sourab Mangrulkar
2024-03-01gemma : fix bfloat16 -> float16 conversion issue (#5810)kunal-vaishnavi
2024-02-25py : fix StableLM conversion after config.json changes (#5703)Anas Ahouzi
2024-02-23convert : fix missing ftype for gemma (#5690)Jared Van Bortel
2024-02-22mpt : do not duplicate token_embd.weight on disk (#5670)Jared Van Bortel
2024-02-22py : add Gemma conversion from HF models (#5647)Georgi Gerganov
2024-02-22py : minor fixes (#5668)Georgi Gerganov