Huihui-Qwen3.5-2B-abliterated — MXFP4 MLX

MLX-converted version of huihui-ai/Huihui-Qwen3.5-2B-abliterated for Apple Silicon, quantized with MXFP4.

Model Details

Property Value
Base model Qwen3.5-2B
Type Vision-Language Model (VLM)
Format MLX MXFP4 (~6.0 bits/weight)
Size ~1.7 GB
Abliterated Yes — censorship layers removed

MXFP4 provides maximum compression (~3x vs fp16) while retaining good quality. Best for memory-constrained devices.

Variants

Variant Size Quality Link
fp16 ~4.5 GB Highest fp16
MXFP8 ~2.6 GB Near-native mxfp8
MXFP4 ~1.7 GB Good This repo

Usage

pip install mlx-vlm

# Text generation
python -m mlx_vlm.generate \
  --model AITRADER/Huihui-Qwen3.5-2B-abliterated-mxfp4-MLX \
  --prompt "Describe this image in detail" \
  --image <path-or-url>

# Chat UI
python -m mlx_vlm.chat_ui \
  --model AITRADER/Huihui-Qwen3.5-2B-abliterated-mxfp4-MLX

Credits

Downloads last month
134
Safetensors
Model size
0.7B params
Tensor type
U8
·
U32
·
BF16
·
F32
·
MLX
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for AITRADER/Huihui-Qwen3.5-2B-abliterated-mxfp4-MLX

Finetuned
Qwen/Qwen3.5-2B
Quantized
(83)
this model