Auto-Quantized GGUF Model

This repository contains automated GGUF quantization files for nbeerbower/Qwen3.5-9B-Writing-DPO.

The calibration data for the imatrix is targeted at Chinese novels and role-playing (RP), while preserving logic and common sense.
imatrix ็š„ๆ กๅ‡†ๆ•ฐๆฎไปฅไธญๆ–‡็š„ๅฐ่ฏดใ€่ง’่‰ฒๆ‰ฎๆผ”ไธบ็›ฎๆ ‡๏ผŒๅŒๆ—ถไฟ็•™้€ป่พ‘ๅ’Œๅธธ่ฏ†ใ€‚

๐Ÿ“Š Perplexity Evaluation

(Tested against the provided calibration dataset)

  • Base (F16/BF16): PPL = 15.6960 +/- 0.13226
  • Q5_K: PPL = 14.7557 +/- 0.12028
Downloads last month
1,305
GGUF
Model size
9B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for nuofang/Qwen3.5-9B-Writing-DPO-GGUF

Finetuned
Qwen/Qwen3.5-9B
Quantized
(4)
this model