πŸ“‹ Model Description


base_model:
  • moonshotai/Kimi-K2.5

Updates

02/11/2026: Vision support for K2.5 has been merged into llama.cpp's master branch and no longer needs to use the PR branch.

02/08/2026: I've updated the PR code to address feedback and updated the mmproj files here to be compatible with the new PR code.

02/01/2026: moonshotai has published an updated chat_template.jinja, I have updated the GGUFs in this repository so please re-download the first shard (00001) for your desired quant.
- The default system prompt might cause confusion to users and unexpected behaviours, so we remove it.
- The token <|mediastart|> is incorrect; it has been replaced with <|mediabegin|> in the chat template.

Model

This is a text-and-image-only GGUF quantization of moonshotai/Kimi-K2.5. This means that video input is not present in this GGUF, and will not be available until support is added upstream in llama.cpp.

MMPROJ files for image vision input have been provided, and support has been merged into the llama.cpp master branch recently.

This Q4X quant is the "full quality" equivalent since the conditional experts are natively INT4 quantized directly from the original model, and the rest of the model is Q80. I also produced and tested a Q80 / Q4K quant, the model size was identical and the PPL was barely higher. Their performance was about the same so I've only uploaded the Q4_X variant.

QuantSizeMixturePPLUploaded?
Q4X543.62 GiB (4.55 BPW)Q80 / Q4_01.8248 +/- 0.00699βœ…
Q4K543.62 GiB (4.55 BPW)Q80 / Q4_K1.8256 +/- 0.00700❌

πŸ“‚ GGUF File List

πŸ“ Filename πŸ“¦ Size ⚑ Download
mmproj-Kimi-K2.5-BF16.gguf
Recommended LFS FP16
909.74 MB Download
mmproj-Kimi-K2.5-F16.gguf
LFS FP16
908.44 MB Download
mmproj-Kimi-K2.5-F32.gguf
LFS
1.76 GB Download
mmproj-Kimi-K2.5-Q8_0.gguf
LFS Q8
613.38 MB Download