Exllamav2 package compiled with:
CUDA: 12.8.0
TORCH: 2.7.0
TORCH_CUDA_ARCH_LIST="6.0 6.1 7.0 7.5 8.0 8.6 8.9 9.0+PTX Blackwell"
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Exllamav2 package compiled with:
CUDA: 12.8.0
TORCH: 2.7.0
TORCH_CUDA_ARCH_LIST="6.0 6.1 7.0 7.5 8.0 8.6 8.9 9.0+PTX Blackwell"