Huihui-Qwen3.5-27B-Opus-abliterated-YaRN-1M
Abliterated Qwen3.5 27B Opus with YaRN extension and reasoning, supporting up to 1M context.
Model Details
- Base: HuihuiAI/Huihui-Qwen3.5-27B-Opus-abliterated
- Context: 262K active / 1M max (YaRN)
- Quantization: BF16
- VRAM: ~60GB (262K) / ~90GB (1M)
- Reasoning: Enabled (Chain-of-Thought)
Usage
262K Context (80GB GPU)
1M Context (100GB+ GPU or Q8_0 model)
With Vision (mmproj)
Files
- GGUF: 53.8GB (BF16)
- mmproj: 931MB (BF16, vision)
- Tokenizer: 13MB
Performance
- 262K context: ~60GB VRAM (BF16 cache)
- 1M context: ~90GB VRAM (BF16 cache)
- Max output: 81,920 tokens
- Temperature: 0.6 (reasoning optimized)
Credits
- Base: HuihuiAI (abliterated Opus)
- Original: Qwen Team
- YaRN: arxiv.org/abs/2309.00071