This model finding1/LongCat-Flash-Chat-MLX-5.5bpw was converted to MLX format from meituan-longcat/LongCat-Flash-Chat using mlx-lm version 0.27.1 by running mlx_lm.convert --quantize --q-bits 5 --mlx-path MLX-5.5bpw --hf-path meituan-longcat/LongCat-Flash-Chat until it crashed with a KeyError; adding "model_type": "longcat_flash", to the downloaded config.json, then running the command again.

Downloads last month
6
Safetensors
Model size
561B params
Tensor type
BF16
U32
F32
MLX
Hardware compatibility
Log In to add your hardware

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for finding1/LongCat-Flash-Chat-MLX-5.5bpw

Quantized
(3)
this model