gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking-GGUF

This is a Deckard(qx) experimental quant.

Brainwaves

         arc   arc/e boolq hswag obkqa piqa  wino
qx86-hi  0.431,0.505,0.426,0.670,0.376,0.766,0.710

Base model

gemma-4-31B-it (Instruct)
qx86-hi  0.496,0.653,0.901,0.624,0.380,0.732,0.653

Previous model

gemma-3-27b-it-heretic
q8       0.557,0.711,0.868,0.533,0.452,0.706,0.695

As I don't have an easy way to test this until LMStudio supports it, please Like it only if you had a good experience.

Thank you,

-G

Downloads last month
2,544
Safetensors
Model size
8B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for CountCandy/gemma-4-31B-it-The-DECKARD-HERETIC-UNCENSORED-Thinking-GGUFs