4-bit LLaDA model
#9
by
chentianqi
- opened
Thank you for your great works. We have released the 4-bit GPTQ quantized LLaDA model on Hugging Face:
Based on the published evaluation code, we have evaluated the quantized base model. The results are as follows:
Dataset | GPTQ-4bit | FP16 |
---|---|---|
MMLU | 65.20 | 65.90 |
CMMLU | 69.23 | 69.90 |
ARC-Challenge | 45.48 | 47.90 |