New Models
Collection
Quants created recently.. where time is relative • 53 items • Updated
This model is Text only: vision was removed, otherwise no changes.
Brainwaves
arc arc/e boolq hswag obkqa piqa wino
qx86-hi 0.443,0.498,0.857,0.701,0.372,0.770,0.752
mxfp4 0.460,0.527,0.871,0.694,0.370,0.772,0.752
Similar models
TeichAI/Qwen3.5-27b-Opus-4.6-Distill
qx64-hi 0.459,0.542,0.724,0.764,0.402,0.790,0.783
DavidAU/Qwen3.5-27B-Polaris-Advanced-Thinking-Alpha
mxfp4 0.473,0.548,0.709,0.728,0.396,0.777,0.753
DavidAU/Qwen3.5-27B-Claude-4.6-OS-Auto-Variable-Thinking
mxfp8 0.485,0.566,0.875,0.746,0.408,0.789,0.730
Instruct models
DavidAU/Qwen3.5-27B-Claude-4.6-OS-INSTRUCT
mxfp8 0.675,0.827,0.900,0.750,0.496,0.800,0.721
qx86-hi 0.667,0.822,0.900
qx64-hi 0.664,0.820,0.902
mxfp4 0.653,0.815,0.899
This model Qwen3.5-27B-Text-mxfp4-mlx was converted to MLX format from Qwen/Qwen3.5-27B using mlx-lm version 0.30.8.
pip install mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("Qwen3.5-27B-Text-mxfp4-mlx")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True, return_dict=False,
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)
4-bit
Base model
Qwen/Qwen3.5-27B