MiMo-VL-7B Quantize
Collection
XiaomiMiMo/MiMo-VL-7B Quantized Model • 5 items • Updated
This is a 4-bit quantized version of XiaomiMiMo/MiMo-VL-7B-SFT,
using the BitsAndBytes library.
Quantization reduces memory usage and makes it possible to run this model on consumer GPUs
(≤ 12 GB VRAM), at the cost of a small reduction in generation quality.
nf4) safetensorsUnable to build the model tree, the base model loops to the model itself. Learn more.