Q4_K_M quantized version using llama.cpp of this model.
iMatrix file used coming from this repository.
Many thanks to Felldude for his work on this.
- Downloads last month
- 22
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support