Using turboderp-org's ExLlamaV3 v0.0.1 for quantization.
2.00 bits per weight
2.25 bits per weight
2.50 bits per weight
3.00 bits per weight
3.50 bits per weight
4.00 bits per weight
5.00 bits per weight
6.00 bits per weight
8.00 bits per weight
8.00 bits per weight / H8
ELYZA-Thinking-1.0-Qwen-32B-exl3
- Model creator: elyza
- Original model: ELYZA-Thinking-1.0-Qwen-32B
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for RioShiina/ELYZA-Thinking-1.0-Qwen-32B-exl3
Base model
Qwen/Qwen2.5-32B Finetuned
Qwen/Qwen2.5-32B-Instruct Finetuned
elyza/ELYZA-Thinking-1.0-Qwen-32B