GGUF conversion of Qwen/Qwen3.5-4B for llama.cpp.
Qwen/Qwen3.5-4B
llama.cpp
Qwen3.5-4B-Q4_K_M.gguf
Qwen3.5-4B-f16.gguf
llama-cli -m Qwen3.5-4B-Q4_K_M.gguf -p "Hello"
Chat template
4-bit
16-bit
Base model