Exllamav2 package compiled with:
CUDA: 12.8.0
TORCH: 2.7.0
TORCH_CUDA_ARCH_LIST="6.0 6.1 7.0 7.5 8.0 8.6 8.9 9.0+PTX Blackwell"
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support