Gemma 4 E4B Companion — GGUF Quantizations
GGUF quantized versions of TinmanLabSL/gemma4-companion-merged.
See the main model card for full details on training, system prompts, and deployment.
Available Quantizations
| File | Quant | Size | Quality | Use Case |
|---|---|---|---|---|
gemma4-companion-bf16.gguf |
BF16 | 14 GB | Lossless | Reference / re-quantization |
gemma4-companion-Q8_0.gguf |
Q8_0 | 7.5 GB | Near-lossless | Desktop with ample RAM |
gemma4-companion-Q4_K_M.gguf |
Q4_K_M | 5.0 GB | Good | Recommended for most users |
Usage
llama.cpp
./llama-cli -m gemma4-companion-Q4_K_M.gguf -cnv \
-p "You are a warm, caring companion..."
LM Studio / GPT4All
Load the GGUF file directly from the UI.
System Prompts
See main model card for companion, roleplay, and assistant mode system prompts.
- Downloads last month
- 1,573
Hardware compatibility
Log In to add your hardware
4-bit
8-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for Tinman-Lab/Tinman-gemma4-companion-gguf
Base model
google/gemma-4-E4B Finetuned
google/gemma-4-E4B-it Quantized
OBLITERATUS/gemma-4-E4B-it-OBLITERATED