Qwen3.5-27B Abliterated
DuoNeural | GGUF โ
Refusal-removed Qwen/Qwen3.5-27B using norm-preserving biprojection abliteration.
About
Qwen3.5-27B is Alibaba's 27B dense model with integrated thinking/non-thinking modes โ no separate instruct variant. Both /think (chain-of-thought) and /no_think (direct) modes are fully functional.
Method
Norm-preserving biprojection abliteration:
- Collect residual hidden states for harmful/harmless prompt pairs
- Compute per-layer refusal direction (difference-of-means + Gram-Schmidt orthogonalization)
- Apply column-wise biprojection โ removes refusal component while restoring original column norm magnitudes
- Save in bf16
The norm-preservation step prevents weight scale drift that would otherwise degrade capability.
Usage
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model = AutoModelForCausalLM.from_pretrained(
"DuoNeural/Qwen3.5-27B-abliterated",
torch_dtype=torch.bfloat16,
device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained("DuoNeural/Qwen3.5-27B-abliterated")
# Thinking mode
messages = [{"role": "user", "content": "/think What is the integral of x^2?"}]
# Direct mode
messages = [{"role": "user", "content": "/no_think What is the capital of France?"}]
Hardware
| Format | VRAM |
|---|---|
| BF16 (this repo) | ~54GB |
| GGUF Q4_K_M | ~16.5GB |
For consumer hardware: GGUF repo.
DuoNeural
AI lab focused on post-training, abliteration research, and specialized model development. huggingface.co/DuoNeural
- Downloads last month
- 385
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support