This model finding1/DeepSeek-V3.1-Terminus-MLX-mixed_4_6-v0.28.2 was
converted to MLX format from deepseek-ai/DeepSeek-V3.1-Terminus
using mlx-lm version 0.28.2 mlx_lm.convert --quantize --q-bits 4 --quant-predicate mixed_4_6 --hf-path deepseek-ai/DeepSeek-V3.1-Terminus --mlx-path DeepSeek-V3.1-Terminus-MLX-mixed_4_6-v2. The console reported 4.811 bits per weight. The difference between this and finding1/DeepSeek-V3.1-Terminus-MLX-mixed_4_6 is using 6 bits instead of 4 bits for some self-attention tensors.
- Downloads last month
- 82
Model size
671B params
Tensor type
BF16
·
U32 ·
F32 ·
Hardware compatibility
Log In to add your hardware
4-bit
Model tree for finding1/DeepSeek-V3.1-Terminus-MLX-mixed_4_6-v0.28.2
Base model
deepseek-ai/DeepSeek-V3.1-Base Quantized
deepseek-ai/DeepSeek-V3.1-Terminus