qwen3-davesoma-v2-gguf

GGUF export for local inference (llama.cpp / Ollama workflows).

  • Quantization: Q4_K_M
  • File: qwen3_8b_davesoma_v2.Q4_K_M.gguf

Suggested use

  • Fast local drafting in the davesoma style
  • Opinion framing and concise technical responses

Notes

  • This is a quantized inference artifact. For reproducible training workflows, use the LoRA repo.
  • Pair with retrieval for factual tasks.

Related repos

Downloads last month
76
GGUF
Model size
8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support