Qwen3.5-27B Abliterated

DuoNeural | GGUF โ†’

Refusal-removed Qwen/Qwen3.5-27B using norm-preserving biprojection abliteration.

About

Qwen3.5-27B is Alibaba's 27B dense model with integrated thinking/non-thinking modes โ€” no separate instruct variant. Both /think (chain-of-thought) and /no_think (direct) modes are fully functional.

Method

Norm-preserving biprojection abliteration:

  1. Collect residual hidden states for harmful/harmless prompt pairs
  2. Compute per-layer refusal direction (difference-of-means + Gram-Schmidt orthogonalization)
  3. Apply column-wise biprojection โ€” removes refusal component while restoring original column norm magnitudes
  4. Save in bf16

The norm-preservation step prevents weight scale drift that would otherwise degrade capability.

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model = AutoModelForCausalLM.from_pretrained(
    "DuoNeural/Qwen3.5-27B-abliterated",
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
tokenizer = AutoTokenizer.from_pretrained("DuoNeural/Qwen3.5-27B-abliterated")

# Thinking mode
messages = [{"role": "user", "content": "/think What is the integral of x^2?"}]

# Direct mode
messages = [{"role": "user", "content": "/no_think What is the capital of France?"}]

Hardware

Format VRAM
BF16 (this repo) ~54GB
GGUF Q4_K_M ~16.5GB

For consumer hardware: GGUF repo.

DuoNeural

AI lab focused on post-training, abliteration research, and specialized model development. huggingface.co/DuoNeural

Downloads last month
385
Safetensors
Model size
27B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DuoNeural/Qwen3.5-27B-abliterated

Base model

Qwen/Qwen3.5-27B
Finetuned
(257)
this model
Quantizations
1 model