summaryrefslogtreecommitdiff
path: root/convert-hf-to-gguf.py
AgeCommit message (Expand)Author
2024-04-24convert : add support of codeqwen due to tokenizer (#6707)Junyang Lin
2024-04-24llama : add phi3 support (#6852)liuwei-git
2024-04-21llama : support Llama 3 HF conversion (#6745)Pedro Cuenca
2024-04-19Implement the OLMo architecture (#6741)nopperl
2024-04-16convert : fix autoawq gemma (#6704)Zheng.Deng
2024-04-16llama : add StableLM2 12B (#6635)Ashish
2024-04-16llama : add qwen2moe (#6074)Shijie
2024-04-16gguf : add special tokens metadata for FIM/Infill (#6689)Daniel Bevenius
2024-04-14convert : enable the `--use-temp-file` cli flag (#6645)James A Capozzoli
2024-04-13model: support arch `DbrxForCausalLM` (#6515)Pierrick Hymbert
2024-04-09BERT tokenizer fixes (#6498)Jared Van Bortel
2024-04-09llama : add Command R Plus support (#6491)Carolinabanana
2024-04-04convert : fix for lint error complaining of bare except (#6470)Clint Herron
2024-04-03llama : add SEA-LION support (#6448)bryanSwk
2024-04-03Missing tokenizer.model error during gguf conversion (#6443)Abhishek Gopinath K
2024-04-03ggml : mul_mat_id use the same tensor for all the experts (#6387)slaren
2024-03-29[Model] Add support for xverse (#6301)hxer7963
2024-03-29convert : allow conversion of Mistral HF models (#6144)Pedro Cuenca
2024-03-28convert : refactor vocab selection logic (#6355)Jared Van Bortel
2024-03-26convert-hf : fix exception in sentencepiece with added tokens (#6320)Pedro Cuenca
2024-03-23llama : add grok-1 support (#6204)Julius Arkenberg
2024-03-18convert : add support for CamembertModel architecture (#6119)Thérence
2024-03-15llama : add Command-R support (#6033)Andrew Canis
2024-03-08llama : support Mamba Selective State Space Models (#5328)compilade
2024-03-04flake : fixGeorgi Gerganov
2024-03-03llama : allow for user specified embedding pooling type (#5849)Douglas Hanley
2024-03-02convert-hf : make model class definitions self-contained (#5825)Jared Van Bortel
2024-03-01llama : add StarCoder2 support (#5795)Sourab Mangrulkar
2024-03-01gemma : fix bfloat16 -> float16 conversion issue (#5810)kunal-vaishnavi
2024-02-25py : fix StableLM conversion after config.json changes (#5703)Anas Ahouzi
2024-02-23convert : fix missing ftype for gemma (#5690)Jared Van Bortel
2024-02-22mpt : do not duplicate token_embd.weight on disk (#5670)Jared Van Bortel
2024-02-22py : add Gemma conversion from HF models (#5647)Georgi Gerganov
2024-02-22py : minor fixes (#5668)Georgi Gerganov
2024-02-15Use correct type of pooling for embedding models (#5500)Douglas Hanley
2024-02-13llama : add support for Nomic Embed (#5468)Jared Van Bortel
2024-02-13llama : support batched embeddings (#5466)Douglas Hanley
2024-02-11Add support for BERT embedding models (#5423)Douglas Hanley
2024-02-08llama : fix MiniCPM (#5392)runfuture
2024-02-07llama : add MiniCPM support (#5346)runfuture
2024-02-05py : fix internlm2-hf convert to gguf (#5305)Guoteng
2024-02-02py : add check for '.attn.masked_bias' layers to GPT2model (#5281)Mirror Azure
2024-02-01llama : support InternLM2 (#5184)Guoteng
2024-01-28llama : add support for Orion-14B (#5118)sharpHL
2024-01-22llama : support StableLM 2 1.6B (#5052)compilade
2024-01-20convert : partially revert PR #4818 (#5041)Jared Van Bortel
2024-01-19llama : support upcoming Qwen2 (#5037)Shijie
2024-01-19py : fix flake8 lintGeorgi Gerganov
2024-01-19llama : add CodeShell support (#5016)chiranko
2024-01-16py : remove unnecessary hasattr (#4903)Georgi Gerganov