This is currently a STATIC quant, because the imatrix tool seems to be broken with Gemma 4 (>100 ppl). I will update with an imatrix once I can verify correctness.
6.52 bpw, a mixture of Q6_K, Q5_K, and Q8_0
Fits ~75k F16 CTX + MMPROJ on a 24GiB GPU, or ~150k CTX without vision. Measured on a DE, not headless.
- Downloads last month
- 557
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
Model tree for Beinsezii/gemma-4-26B-A4B-it-GGUF-6.52BPW-static
Base model
google/gemma-4-26B-A4B-it