Gemma-3-4B Q4_K_M GGUF

This is a Q4_K_M quantized GGUF conversion of google/gemma-3-4b-it optimized for on-device inference with llama.cpp.

Model Details

Property Value
Original Model Gemma 3 4B Instruct
Parameters 4 billion
Quantization Q4_K_M (4-bit, medium quality)
File Size ~2.5 GB
Context Window 8,192 tokens
Architecture Gemma 3

Intended Use

This model is optimized for:

  • Mobile/Edge Deployment: Runs efficiently on iOS devices with 8GB+ RAM (iPhone 15 Pro, 16, etc.)
  • llama.cpp Integration: Compatible with llama.cpp and its bindings
  • On-Device AI: Private, offline inference without cloud dependencies

Capabilities

  • General conversation and Q&A
  • Code assistance
  • Creative writing
  • Reasoning and analysis
  • Multilingual support

Usage with llama.cpp

./llama-cli -m Gemma-3-4B-Q4_K_M.gguf -p "Your prompt here" -n 512

License

This model is subject to Google's Gemma Terms of Use. Please review the license terms before use.

Attribution

  • Original Model: Gemma 3 4B by Google
  • Quantization: jc-builds
Downloads last month
55
GGUF
Model size
4B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for jc-builds/Gemma-3-4B-Q4_K_M-GGUF

Quantized
(209)
this model