Qwen3.5 MLX — Apple Silicon Optimized
Collection
4 items • Updated
MLX quantization of Qwen/Qwen3.5-27B optimized for Apple Silicon.
from mlx_lm import load, generate
model, tokenizer = load("TxemAI/Qwen3.5-27B-4bit-MLX")
response = generate(model, tokenizer, prompt="Your prompt here", verbose=True)
Use 4-bit if you want faster inference and lower RAM usage. Use 8-bit version if you prioritize output quality.
8-bit
Base model
Qwen/Qwen3.5-27B