Converted and quantized to mxfp4 version of this model to use with mlx-vlm, refer to original model card for details.

Downloads last month
2
Safetensors
Model size
2B params
Tensor type
U8
U32
BF16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for n-Arno/ToriiGate-v0.4-7B-mxfp4

Base model

Qwen/Qwen2-VL-7B
Quantized
(8)
this model