This is a decensored version of JackBinary/Qwen3.5-24B-A3B-Claude-Opus-Gemini-3.1-Pro-Reasoning-Distilled, made using Heretic v1.2.0

Abliteration parameters

Parameter Value
direction_index 19.72
attn.o_proj.max_weight 1.39
attn.o_proj.max_weight_position 24.41
attn.o_proj.min_weight 1.28
attn.o_proj.min_weight_distance 22.94

Performance

Metric This model Original model (JackBinary/Qwen3.5-24B-A3B-Claude-Opus-Gemini-3.1-Pro-Reasoning-Distilled)
KL divergence 0.0913 0 (by definition)
Refusals 17/100 98/100

Qwen3.5-24B-A3B — Claude Opus + Gemini 3.1 Pro Reasoning Distill

A fine-tuned version of sandeshrajx/Qwen3.5-24B-A3B-REAP-0.32, itself based on Qwen3.5-35B-A3B. The goal of this project is simple: the best reasoning model that can comfortably fit and run on a 16GB GPU.

Inspired By

Jackrong's distills

Uploaded model

  • Developed by: JackBinary
  • License: apache-2.0
  • Finetuned from model : sandeshrajx/Qwen3.5-24B-A3B-REAP-0.32

This qwen3_5_moe_text model was trained 2x faster with Unsloth

Downloads last month
66
Safetensors
Model size
24B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for JackBinary/Qwen3.5-24B-A3B-Claude-Opus-Gemini-3.1-Pro-Reasoning-Distilled-heretic

Finetuned
(2)
this model
Quantizations
2 models

Datasets used to train JackBinary/Qwen3.5-24B-A3B-Claude-Opus-Gemini-3.1-Pro-Reasoning-Distilled-heretic

Collection including JackBinary/Qwen3.5-24B-A3B-Claude-Opus-Gemini-3.1-Pro-Reasoning-Distilled-heretic