MamayLM Gemma 3 12B โ€” FP8 Dynamic

FP8 static quantized from INSAIT-Institute/MamayLM-Gemma-3-12B-IT-v1.0 using llmcompressor.

Usage

  • Designed for vLLM inference (H100).
  • trust_remote_code=True.
Downloads last month
16
Safetensors
Model size
12B params
Tensor type
BF16
ยท
F8_E4M3
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for oshkorinova/MamayLM-Gemma-3-12B-IT-v1.0-FP8-Dynamic

Quantized
(6)
this model