qwen3-0.6b-codeforces-sft-gguf

This is a GGUF conversion of albertlieadrian/qwen3-0.6b-codeforces-sft, which is a LoRA fine-tuned version of Qwen/Qwen3-0.6B.

Model Details

  • Base Model: Qwen/Qwen3-0.6B
  • Fine-tuned Model: albertlieadrian/qwen3-0.6b-codeforces-sft
  • Training: Supervised Fine-Tuning (SFT) with TRL
  • Format: GGUF (for llama.cpp, Ollama, LM Studio, etc.)

Quantization

File Quant Size Description
qwen3-0.6b-codeforces-sft-q4_k_m.gguf Q4_K_M ~462MB 4-bit medium (recommended)

Usage

With llama.cpp

huggingface-cli download albertlieadrian/qwen3-0.6b-codeforces-sft-gguf qwen3-0.6b-codeforces-sft-q4_k_m.gguf
./llama-cli -m qwen3-0.6b-codeforces-sft-q4_k_m.gguf -p "Your prompt"

With Ollama

  1. Create a Modelfile:
FROM ./qwen3-0.6b-codeforces-sft-q4_k_m.gguf
  1. Create and run:
ollama create my-model -f Modelfile
ollama run my-model

With LM Studio

  1. Download the .gguf file
  2. Import into LM Studio
  3. Start chatting!

License

Inherits the license from the base model: Qwen/Qwen3-0.6B


Converted to GGUF format using llama.cpp

Downloads last month
7
GGUF
Model size
0.8B params
Architecture
qwen3
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for albertlieadrian/qwen3-0.6b-codeforces-sft-gguf

Finetuned
Qwen/Qwen3-0.6B
Quantized
(288)
this model