qwen3-0.6b-codeforces-sft-gguf
This is a GGUF conversion of albertlieadrian/qwen3-0.6b-codeforces-sft, which is a LoRA fine-tuned version of Qwen/Qwen3-0.6B.
Model Details
- Base Model: Qwen/Qwen3-0.6B
- Fine-tuned Model: albertlieadrian/qwen3-0.6b-codeforces-sft
- Training: Supervised Fine-Tuning (SFT) with TRL
- Format: GGUF (for llama.cpp, Ollama, LM Studio, etc.)
Quantization
| File | Quant | Size | Description |
|---|---|---|---|
| qwen3-0.6b-codeforces-sft-q4_k_m.gguf | Q4_K_M | ~462MB | 4-bit medium (recommended) |
Usage
With llama.cpp
huggingface-cli download albertlieadrian/qwen3-0.6b-codeforces-sft-gguf qwen3-0.6b-codeforces-sft-q4_k_m.gguf
./llama-cli -m qwen3-0.6b-codeforces-sft-q4_k_m.gguf -p "Your prompt"
With Ollama
- Create a
Modelfile:
FROM ./qwen3-0.6b-codeforces-sft-q4_k_m.gguf
- Create and run:
ollama create my-model -f Modelfile
ollama run my-model
With LM Studio
- Download the
.gguffile - Import into LM Studio
- Start chatting!
License
Inherits the license from the base model: Qwen/Qwen3-0.6B
Converted to GGUF format using llama.cpp
- Downloads last month
- 7
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support