4-bit LLaDA model

#9
by chentianqi - opened

Thank you for your great works. We have released the 4-bit GPTQ quantized LLaDA model on Hugging Face:

Based on the published evaluation code, we have evaluated the quantized base model. The results are as follows:

Dataset GPTQ-4bit FP16
MMLU 65.20 65.90
CMMLU 69.23 69.90
ARC-Challenge 45.48 47.90
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment