Notes

  • 05/05/26: I've updated all the quants to use the fused QKV conversion. The PR branch supports both fused + unfused so it's not necessary to download the new quants, but it may provide a small speed boost.
  • 05/03/26: WIP vision support on this branch: https://github.com/AesSedai/llama.cpp/tree/mimo-v2.5-vision (if it's broken with F16 mmproj, pull the latest commit and recompile) and uploaded mmproj files
  • 05/01/26: This branch includes CUDA flash attention, should speed up PP / TG: https://github.com/AesSedai/llama.cpp/tree/mimo-v2.5-fattn
  • 04/28/26: While this model should run on the llama.cpp master branch, there was a small change to the inference code to support the attention_value_scale parameter. For the best accuacy/performance, I recommend pulling and compiling from this PR branch: https://github.com/ggml-org/llama.cpp/pull/22493.

Model

This is a text-only GGUF quantization of XiaomiMiMo/MiMo-V2.5. This means that image and audio input is not present in this GGUF, and will not be available until support is added upstream in llama.cpp.

This repo contains specialized MoE-quants for MiMo-V2.5. The idea being that given the huge size of the FFN tensors compared to the rest of the tensors in the model, it should be possible to achieve a better quality while keeping the overall size of the entire model smaller compared to a similar naive quantization. To that end, the quantization type default is kept in high quality and the FFN UP + FFN GATE tensors are quanted down along with the FFN DOWN tensors.

Quant Size Mixture PPL 1-(Mean PPL(Q)/PPL(base)) KLD
Q8_0 305.68 GiB (8.50 BPW) Q8_0 5.135595 ± 0.030275 +0.1271% 0.012539 ± 0.000329
Q5_K_M 212.42 GiB (5.91 BPW) Q8_0 / Q5_K / Q5_K / Q6_K 5.148091 ± 0.030387 +0.3708% 0.014915 ± 0.000309
Q4_K_M 176.43 GiB (4.91 BPW) Q8_0 / Q4_K / Q4_K / Q5_K 0.000000 ± 0.000000 +0.0000% 0.000000 ± 0.000000
IQ4_XS 136.78 GiB (3.80 BPW) Q8_0 / IQ3_S / IQ3_S / IQ4_XS 5.271397 ± 0.031170 +2.7748% 0.041177 ± 0.000349
IQ3_S 105.33 GiB (2.93 BPW) Q6_K / IQ2_S / IQ2_S / IQ3_S 5.552710 ± 0.033286 +8.2595% 0.092639 ± 0.000604

kld_graph ppl_graph

Downloads last month
5,997
GGUF
Model size
309B params
Architecture
mimo2
Hardware compatibility
Log In to add your hardware

3-bit

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for AesSedai/MiMo-V2.5-GGUF

Quantized
(5)
this model