Gemma 4 26B MoE AWQ 4-bit
AWQ 4-bit quantization of Gemma 4 26B-A4B-it optimized for AMD RDNA4 (gfx1201) inference with SGLang.
Model Details
| Base model | google/gemma-4-26b-a4b-it |
| Architecture | MoE (128 experts, top-8) |
| Parameters | 26B total / 4B active |
| Layers | 30 |
| Context | 4K (tested) |
| Quantization | AWQ 4-bit, group_size=32. Forced-routing GPTQ calibration covers all 128 experts (standard GPTQ only calibrates ~1/128). |
Performance (2x AMD Radeon AI PRO R9700, TP=2)
- Decode speed: 30 tok/s single-user on 2x R9700
- Launch:
scripts/launch.sh gemma4
Notes
Standard community GPTQ under-calibrates rare experts due to routing imbalance. This model uses forced-routing calibration to ensure all 128 experts are properly quantized.
Known Limitations
- Vision: UNTESTABLE — Vision encoder layers (
embed_vision.*) were quantized to INT4, which likely degrades vision quality. Server crashes on first request (pre-existing RDNA4 triton issue with this model's SWA configuration, not vision-specific). Text-only inference recommended. A future version should add vision layers tomodules_to_not_convert.
Usage with SGLang
git clone https://github.com/mattbucci/2x-R9700-RDNA4-GFX1201-sglang-inference
cd 2x-R9700-RDNA4-GFX1201-sglang-inference
./scripts/setup.sh
scripts/launch.sh gemma4
See the RDNA4 Inference Repository for full setup instructions, patches, and benchmarks.
Hardware
Tested on 2x AMD Radeon AI PRO R9700 (gfx1201, RDNA4, 32+34 GB VRAM) with ROCm 7.2 and SGLang v0.5.10 + RDNA4 patches.
- Downloads last month
- 2,827
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support