Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Buckets new
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

mattbucci
/
gemma-4-26B-AWQ

Safetensors
gemma4
awq
4-bit precision
rdna4
gfx1201
rocm
sglang
quantized
Model card Files Files and versions
xet
Community
gemma-4-26B-AWQ
16.8 GB
Ctrl+K
Ctrl+K
  • 1 contributor
History: 11 commits
mattbucci's picture
mattbucci
add quantization_config.ignore=1 (downstream audit fix)
06b09ed verified 9 days ago
  • .gitattributes
    1.57 kB
    Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4 23 days ago
  • README.md
    1.99 kB
    Update vision status: untestable due to server crash 23 days ago
  • chat_template.jinja
    16.4 kB
    Add files using upload-large-folder tool 17 days ago
  • config.json
    4.08 kB
    add quantization_config.ignore=1 (downstream audit fix) 9 days ago
  • generation_config.json
    203 Bytes
    Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4 23 days ago
  • model-00001-of-00001.safetensors
    15.7 GB
    xet
    Replace INT4 vision with BF16 originals (model-00001-of-00001.safetensors) 23 days ago
  • model-vision.safetensors
    1.15 GB
    xet
    Replace INT4 vision with BF16 originals (model-vision.safetensors) 23 days ago
  • model.safetensors.index.json
    3.64 MB
    Replace INT4 vision with BF16 originals (model.safetensors.index.json) 23 days ago
  • processor_config.json
    1.69 kB
    Add files using upload-large-folder tool 17 days ago
  • tokenizer.json
    32.2 MB
    xet
    Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4 23 days ago
  • tokenizer_config.json
    2.69 kB
    Gemma 4 26B MoE AWQ: forced-routing GPTQ calibration for all 128 experts, 30 tok/s on RDNA4 23 days ago