Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
mattbucci
/
gemma-4-26B-AWQ
like
0
Safetensors
gemma4
awq
4-bit precision
rdna4
gfx1201
rocm
sglang
quantized
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
main
gemma-4-26B-AWQ
16.8 GB
Ctrl+K
Ctrl+K
1 contributor
History:
11 commits
mattbucci
add quantization_config.ignore=1 (downstream audit fix)
06b09ed
verified
9 days ago
.gitattributes
Safe
1.57 kB
Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4
23 days ago
README.md
Safe
1.99 kB
Update vision status: untestable due to server crash
23 days ago
chat_template.jinja
Safe
16.4 kB
Add files using upload-large-folder tool
17 days ago
config.json
4.08 kB
add quantization_config.ignore=1 (downstream audit fix)
9 days ago
generation_config.json
Safe
203 Bytes
Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4
23 days ago
model-00001-of-00001.safetensors
Safe
15.7 GB
xet
Replace INT4 vision with BF16 originals (model-00001-of-00001.safetensors)
23 days ago
model-vision.safetensors
Safe
1.15 GB
xet
Replace INT4 vision with BF16 originals (model-vision.safetensors)
23 days ago
model.safetensors.index.json
Safe
3.64 MB
Replace INT4 vision with BF16 originals (model.safetensors.index.json)
23 days ago
processor_config.json
Safe
1.69 kB
Add files using upload-large-folder tool
17 days ago
tokenizer.json
Safe
32.2 MB
xet
Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4
23 days ago
tokenizer_config.json
Safe
2.69 kB
Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4
23 days ago