quantizing into int8 precision

#1
by tanvij - opened

I'm looking into quantizing this model to int8 precision and I'm wondering if I should manually quantize the weights or use an automated technique like AWQ or bitsandbytes. Any recommendations on which method works best for this model? Thanks! @klldmofashi

tanvij changed discussion title from quantizing to quantizing into int8 precision
Efficient-Large-Model org
This comment has been hidden (marked as Off-Topic)
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment