GGUF available β€” Cerebellum v4 (ablation-guided mixed-precision)

#28
by deucebucket - opened

Ablation-guided mixed-precision GGUF quant for running this model in llama.cpp / ollama:

deucebucket/Qwen3.6-27B-Cerebellum-v4-GGUF

Instead of treating every tensor the same, Cerebellum runs individual ablation experiments to measure which tensors are sensitive vs. tolerant and assigns precision accordingly. Details and benchmarks in the model card.

Sign up or log in to comment