Quantized using the default exllamav3 (0.0.4) quantization process.
- Original model:
icefog72/IceMedovukhaRP-7b- refer for more details on the model. - exllamav3: https://github.com/turboderp-org/exllamav3
EXL3 quants available:
- 4.0bpw, 5.0bpw, 6.0bpw
- Go to "Files and versions", then click on "Main" to choose your quant
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for s1arsky/IceMedovukhaRP-7b-EXL3
Base model
icefog72/IceMedovukhaRP-7b