This is currently a STATIC quant, because the imatrix tool seems to be broken with Gemma 4 (>100 ppl). I will update with an imatrix once I can verify correctness.

6.52 bpw, a mixture of Q6_K, Q5_K, and Q8_0

Fits ~75k F16 CTX + MMPROJ on a 24GiB GPU, or ~150k CTX without vision. Measured on a DE, not headless.

Downloads last month
557
GGUF
Model size
25B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Beinsezii/gemma-4-26B-A4B-it-GGUF-6.52BPW-static

Quantized
(153)
this model