GemmaTranslate-v3-12B-GGUF

This repository contains the GGUF quantized versions of GemmaTranslate-v3-12B, a multimodal hybrid model merged from Gemma-3-12B-IT and TranslateGemma-12B-IT.

Available Formats

File Size Quantization Description
GemmaTranslate-v3-BF16.gguf ~23.5GB BF16 Full precision (source of truth).
GemmaTranslate-v3-Q8_0.gguf ~12.5GB Q8_0 High fidelity, recommended for 16GB+ RAM.
GemmaTranslate-v3-Q4_K_M.gguf ~7.3GB Q4_K_M Balanced performance/size, recommended for 8GB+ RAM.

Usage with llama.cpp

./llama-cli -m GemmaTranslate-v3-Q4_K_M.gguf -p "Translate 'AI is the future' to French:"

Usage with Ollama

ollama run hf.co/SpongeBOB9684/GemmaTranslate-v3-12B-GGUF:Q4_K_M

License

This model is subject to the Gemma Terms of Use.

Downloads last month
76
GGUF
Model size
12B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for SpongeBOB9684/GemmaTranslate-v3-12B-GGUF

Quantized
(3)
this model