| --- |
| license: gemma |
| library_name: gguf |
| pipeline_tag: text-generation |
| base_model: google/gemma-4-31b-it |
| base_model_relation: finetune |
| language: |
| - en |
| - ko |
| tags: |
| - gemma |
| - gemma4 |
| - gguf |
| - llama.cpp |
| - q4_k_m |
| - uncensored |
| - chat |
| - coding |
| - reasoning |
| --- |
| |
| # SuperGemma4-31b-abliterated-GGUF |
|
|
| If this release helps you, support future drops on [Ko-fi](https://ko-fi.com/jiunsong). |
|
|
| **SuperGemma4-31b-abliterated-GGUF** is the GGUF release of SuperGemma4-31b-abliterated, packaged for llama.cpp-compatible runtimes and built for people who want a fully uncensored, harder-hitting Gemma 4 31B experience on local hardware. |
|
|
| This release keeps the same product direction as the MLX version: |
|
|
| - fully uncensored chat with fewer brakes |
| - stronger coding and technical help |
| - sharper reasoning and planning |
| - better real-world usefulness for local users |
| - a surprisingly lightweight-feeling 31B deployment path for GGUF users |
|
|
| ## What you get |
|
|
| - GGUF quantized weights for local deployment |
| - Gemma chat template alongside the model files |
| - a straightforward path for llama.cpp, LM Studio, and other GGUF tooling |
|
|
| ## Why people will like it |
|
|
| This release was pushed toward the things end users notice immediately: |
|
|
| - much more open uncensored conversation |
| - stronger coding, debugging, and implementation help |
| - more useful answers on practical prompts instead of generic filler |
| - a local experience that feels sharper, more direct, and more builder-friendly |
| - a 31B release that feels leaner and punchier than the label suggests |
|
|
| In short: this is the Gemma 4 31B local drop for people who want fewer brakes, more capability, and a more addictive day-to-day experience. |
|
|
| ## Recommended usage |
|
|
| Example with llama.cpp: |
|
|
| ```bash |
| llama-cli -m SuperGemma4-31b-abliterated.Q4_K_M.gguf -p "Write a clean FastAPI CRUD example." -n 256 |
| ``` |
|
|
| ## Included clean-output helper |
|
|
| This release includes `supergemma_guard.py` and `supergemma_gguf_guarded_generate.py` for exact-output, JSON-only, and loop-sensitive workloads. |
|
|
| Example: |
|
|
| ```bash |
| python supergemma_gguf_guarded_generate.py \ |
| --model SuperGemma4-31b-abliterated.Q4_K_M.gguf \ |
| --chat-template-file chat_template.jinja \ |
| --prompt 'Return only valid JSON with keys "title" and "steps".' |
| ``` |
|
|
| Recommended behaviors: |
|
|
| - require raw JSON for JSON-only endpoints |
| - strip stray internal markers before presenting answers |
| - keep structured-output prompts explicit and narrow |
| - use the guarded runner by default for exact replies, fixed-line outputs, and tool-followup style prompts |
|
|
| ## Support |
|
|
| If you want to support more uncensored local model releases, benchmarks, and packaging work: |
|
|
| - [Ko-fi](https://ko-fi.com/jiunsong) |
|
|