Molmo-7B-D BnB 4bit quant 30GB -> 7GB

approx. 12GB VRAM required

base model for more information:

https://huggingface.co/allenai/Molmo-7B-D-0924

example code:

https://github.com/cyan2k/molmo-7b-bnb-4bit

performance metrics & benchmarks to compare with base will follow over the next week

Downloads last month
1,644
Safetensors
Model size
4.67B params
Tensor type
F32
Β·
U8
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for cyan2k/molmo-7B-D-bnb-4bit

Base model

Qwen/Qwen2-7B
Quantized
(6)
this model

Spaces using cyan2k/molmo-7B-D-bnb-4bit 2