Gemma 4 E4B Companion — GGUF Quantizations

GGUF quantized versions of TinmanLabSL/gemma4-companion-merged.

See the main model card for full details on training, system prompts, and deployment.

Available Quantizations

File Quant Size Quality Use Case
gemma4-companion-bf16.gguf BF16 14 GB Lossless Reference / re-quantization
gemma4-companion-Q8_0.gguf Q8_0 7.5 GB Near-lossless Desktop with ample RAM
gemma4-companion-Q4_K_M.gguf Q4_K_M 5.0 GB Good Recommended for most users

Usage

llama.cpp

./llama-cli -m gemma4-companion-Q4_K_M.gguf -cnv \
  -p "You are a warm, caring companion..."

LM Studio / GPT4All

Load the GGUF file directly from the UI.

System Prompts

See main model card for companion, roleplay, and assistant mode system prompts.

Downloads last month
1,573
GGUF
Model size
7B params
Architecture
gemma4
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Tinman-Lab/Tinman-gemma4-companion-gguf