Update README.md
Browse files
README.md
CHANGED
@@ -124,7 +124,7 @@ Use the following code to get the quantized model:
|
|
124 |
import torch
|
125 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TorchAoConfig
|
126 |
|
127 |
-
model_id = "
|
128 |
|
129 |
from torchao.quantization import Int4WeightOnlyConfig
|
130 |
quant_config = Int4WeightOnlyConfig(group_size=128, use_hqq=True)
|
|
|
124 |
import torch
|
125 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TorchAoConfig
|
126 |
|
127 |
+
model_id = "Qwen/Qwen3-8B"
|
128 |
|
129 |
from torchao.quantization import Int4WeightOnlyConfig
|
130 |
quant_config = Int4WeightOnlyConfig(group_size=128, use_hqq=True)
|