GGUFs for moonshotai/Kimi-VL-A3B-Instruct
Didn't see any GGUFs for this model, which is a legit model, so baked a couple. Hopefully useful to someone. Just straight llama-quantize off a BF16 convert_hf_to_gguf.py run. Sanity checked.
- Base model: moonshotai/Kimi-VL-A3B-Instruct
- GGUFs for Thinking version: ssweens/Kimi-VL-A3B-Thinking-2506-GGUF
- Downloads last month
- 873
Hardware compatibility
Log In
to view the estimation
4-bit
6-bit
8-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for ssweens/Kimi-VL-A3B-Instruct-GGUF
Base model
moonshotai/Moonlight-16B-A3B
Finetuned
moonshotai/Kimi-VL-A3B-Instruct