Mixtral-8x7B-v0.1-W8A16

This model is an INT8 weight-only quantized version of Mixtral-8x7B-v0.1.

Downloads last month
13
Safetensors
Model size
46.7B params
Tensor type
F32
BF16
I8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for Lynawa/Mixtral-8x7B-v0.1-W8A16

Finetuned
(62)
this model