EXL3 quantization of Jan-v2-VL-high, 8 bits per weight, including output layers.

Downloads last month
1
Safetensors
Model size
5B params
Tensor type
F16
I16
BF16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for isogen/Jan-v2-VL-high-exl3-8bpw-h8

Quantized
(14)
this model