⚠️ 兼容性警告 / INCOMPATIBILITY WARNING 本仓库的模型使用 ikawrakow/ik_llama.cpp 分支进行量化。 它与原版的 llama.cpp 及主流推理前端(如 LM Studio, Text-generation-webui, Ollama 等)不兼容! 请确保您使用 ik_llama.cpp 编译的二进制文件来运行此模型。

This model was quantized using the ik_llama.cpp fork. It is INCOMPATIBLE with standard llama.cpp and standard UI frontends! You must use binaries compiled directly from the ik_llama.cpp repository to run this model.

Auto-Quantized GGUF Model

This repository contains automated GGUF quantization files for nbeerbower/Qwen3.5-9B-Writing-DPO.

The calibration data for the imatrix is targeted at Chinese novels and role-playing (RP), while preserving logic and common sense.
imatrix 的校准数据以中文的小说、角色扮演为目标,同时保留逻辑和常识。

📊 Perplexity Evaluation

(Tested against the provided calibration dataset)

  • Base (F16/BF16): PPL = 14.6755 +/- 0.11970
Downloads last month
461
GGUF
Model size
9B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nuofang/Qwen3.5-9B-Writing-DPO-ik-GGUF

Finetuned
Qwen/Qwen3.5-9B
Quantized
(4)
this model