This model was converted to MLX format and quantized from Qwen3.6-35B-A3B using oMLX.

What is "oQ"?

See "oQ: oMLX Universal Dynamic Quantization" for details.

Quantizations

See "Evaluation of various MLX quantizations" for details:

Qwen3.6-35B-A3B KLD/RAM chart

What is "VL"?

"VL" is Vision-Language, meaning quantization preserves the original model's multimodality.

No "VL" means quantization is Text-Only.

What is "FP16"?

"FP16" is an M1/M2 Apple Silicon tweak that delivers a very noticeable prompt processing boost, because older M-series lack native BF16 hardware support. See jundot/omlx/issues/604 for details.

No "FP16" means quantization is better suited for M3+ Apple Silicon.

Downloads last month
4,299
Safetensors
Model size
6B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for deepsweet/Qwen3.6-35B-A3B-MLX-oQ4

Quantized
(305)
this model

Collection including deepsweet/Qwen3.6-35B-A3B-MLX-oQ4