diff options
author | ubergarm <leimgrub@gmail.com> | 2025-07-15 13:54:04 -0400 |
---|---|---|
committer | GitHub <noreply@github.com> | 2025-07-15 19:54:04 +0200 |
commit | 13b2f193723486f46efe34297cf797186ab14bc2 (patch) | |
tree | bda8a4b50adb20a564302e16dc42bed45ea798d4 /ggml/include/ggml.h | |
parent | 2081b3fccb9923699bf4d5e926d8719fc1d12c39 (diff) |
kimi-k2 convert script and chat template (#612)
* convert_hf_to_gguf for Kimi-K2-Instruct
Adapt mainline `PR14653` for tokenizer while maintaining proper MLA
tensors. Tested with this workflow using deepseek fp8_cast_bf16.py and
triton-cpu to upcast the fp8 safetensors to bf16 safetensors then used
this convert_hf_to_gguf.
* Add Kimi-K2 chat template
moonshotai/Kimi-K2-Instruct
https://github.com/ikawrakow/ik_llama.cpp/pull/609#issuecomment-3071259454
* kimi-k2 add ass to template to get response
Diffstat (limited to 'ggml/include/ggml.h')
0 files changed, 0 insertions, 0 deletions