Update README.md
Browse files
README.md
CHANGED
@@ -14,6 +14,10 @@ A DL-worthy model, merge of Nemotron Abliterated and Llama 3.3 Instruct, with L3
|
|
14 |
And it's quite smart and verbose, albeit quite a bit prudish.
|
15 |
One of my favorites ATM.
|
16 |
|
|
|
|
|
|
|
|
|
17 |
---
|
18 |
# merge
|
19 |
|
@@ -44,7 +48,7 @@ models:
|
|
44 |
- model: meta-llama/Llama-3.3-70B-Instruct
|
45 |
parameters:
|
46 |
weight: 1.0
|
47 |
-
dtype: bfloat16
|
48 |
out_dtype: bfloat16
|
49 |
parameters:
|
50 |
int8_mask: true
|
|
|
14 |
And it's quite smart and verbose, albeit quite a bit prudish.
|
15 |
One of my favorites ATM.
|
16 |
|
17 |
+
Note : if someone could reproduce the recipe with huihui's Llama 3.3 70b Instruct abliterated (normal version, not the "finetuned") instead of Llama 3.3 70b Instruct, it'd be worth a try.
|
18 |
+
|
19 |
+
I don't have access to a mergekit rig atm, now that Arcee's space is down.
|
20 |
+
|
21 |
---
|
22 |
# merge
|
23 |
|
|
|
48 |
- model: meta-llama/Llama-3.3-70B-Instruct
|
49 |
parameters:
|
50 |
weight: 1.0
|
51 |
+
dtype: bfloat16 -> replace with float32 if possible
|
52 |
out_dtype: bfloat16
|
53 |
parameters:
|
54 |
int8_mask: true
|