dphn/Dolphin-Mistral-24B-Venice-Edition, UQFF quantization

Run with mistral.rs. Documentation: UQFF docs.

  1. Flexible ๐ŸŒ€: Multiple quantization formats in one file format with one framework to run them all.
  2. Reliable ๐Ÿ”’: Compatibility ensured with embedded and checked semantic versioning information from day 1.
  3. Easy ๐Ÿค—: Download UQFF models easily and quickly from Hugging Face, or use a local file.
  4. Customizable ๐Ÿ› : Make and publish your own UQFF files in minutes.

Examples

Quantization type(s) Example
Q2K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f Dolphin-Mistral-24B-Venice-Edition-v2-Q2K-0.uqff
Q3K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f Dolphin-Mistral-24B-Venice-Edition-v2-Q3K-0.uqff
Q4_0 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q4_0-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q4_0-1.uqff"
Q4_1 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q4_1-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q4_1-1.uqff"
Q4K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q4K-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q4K-1.uqff"
Q5_0 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q5_0-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q5_0-1.uqff"
Q5_1 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q5_1-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q5_1-1.uqff"
Q5K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q5K-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q5K-1.uqff"
Q6K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q6K-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q6K-1.uqff"
Q8_0 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q8_0-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q8_0-1.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q8_0-2.uqff"
Q8K ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-Q8K-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q8K-1.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-Q8K-2.uqff"
HQQ4 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-HQQ4-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-HQQ4-1.uqff"
HQQ8 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-HQQ8-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-HQQ8-1.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-HQQ8-2.uqff"
FP8 ./mistralrs-server -i plain -m sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF -f "Dolphin-Mistral-24B-Venice-Edition-v2-FP8-0.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-FP8-1.uqff;Dolphin-Mistral-24B-Venice-Edition-v2-FP8-2.uqff"
Downloads last month
1
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for sonicrules1234/dphn_Dolphin-Mistral-24B-Venice-Edition-UQFF