Qwanko3.5-27B-V2

Qwopus added depth of thinking, Anko added creativity the results is a model that retain 75% ~ 98% of parent skills in both logical and creative task.

Tool calling seems to improve, i'm using hermes tho. 262k context seems stable. sometimes its ponder on the internal chamber for a while but pretty quick once it's done pondering often do rapid tool calling i must say it is an improvement IMO.

My daily driver is using Q6_K with SSM at F32 for stability in long context and first 4 input and output at BF16, check it out here

With ik_llama.cpp -sm graph really helpfull speed things up and lower vram usage fit 262k context with f16 kv cache.

It ponder with itself for a long time dunno why. But it get the job done so i'm not complaining.

Please disable thinking mode for immersive story experience.

I feels like... disable thinking is really good its instantenous compared to chain of thought. Still you gonna need to intervene before the model do some thing but... so far as long as i tell exactly what todo or what happened, non-thinking work like a charm.

Merge Details

This is a merge of pre-trained language models created using mergekit.

Merge Method

This model was merged using the SLERP merge method.

Models Merged

The following models were included in the merge:

  • allura-org/Qwen3.5-27B-Anko
  • huihui-ai/Huihui-Qwopus3.5-27B-v3-abliterated

Configuration

The following YAML configuration was used to produce this model:

merge_method: slerp
base_model: huihui-ai/Huihui-Qwopus3.5-27B-v3-abliterated
models:
  - model: allura-org/Qwen3.5-27B-Anko
parameters:
  t: 0.5
Downloads last month
1,268
Safetensors
Model size
27B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for h34v7/Qwanko3.5-27B-V2