Qwanko3.5-27B-V2
Qwopus added depth of thinking, Anko added creativity the results is a model that retain 75% ~ 98% of parent skills in both logical and creative task.
Tool calling seems to improve, i'm using hermes tho. 262k context seems stable. sometimes its ponder on the internal chamber for a while but pretty quick once it's done pondering often do rapid tool calling i must say it is an improvement IMO.
My daily driver is using Q6_K with SSM at F32 for stability in long context and first 4 input and output at BF16, check it out here
With ik_llama.cpp -sm graph really helpfull speed things up and lower vram usage fit 262k context with f16 kv cache.
It ponder with itself for a long time dunno why. But it get the job done so i'm not complaining.
Please disable thinking mode for immersive story experience.
I feels like... disable thinking is really good its instantenous compared to chain of thought. Still you gonna need to intervene before the model do some thing but... so far as long as i tell exactly what todo or what happened, non-thinking work like a charm.
Merge Details
This is a merge of pre-trained language models created using mergekit.
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
- allura-org/Qwen3.5-27B-Anko
- huihui-ai/Huihui-Qwopus3.5-27B-v3-abliterated
Configuration
The following YAML configuration was used to produce this model:
merge_method: slerp
base_model: huihui-ai/Huihui-Qwopus3.5-27B-v3-abliterated
models:
- model: allura-org/Qwen3.5-27B-Anko
parameters:
t: 0.5
- Downloads last month
- 1,268