Triangle104 commited on
Commit
cc66e1f
·
verified ·
1 Parent(s): 9e080b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md CHANGED
@@ -45,6 +45,54 @@ pipeline_tag: text-generation
45
  This model was converted to GGUF format from [`DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B`](https://huggingface.co/DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
46
  Refer to the [original model card](https://huggingface.co/DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B) for more details on the model.
47
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  ## Use with llama.cpp
49
  Install llama.cpp through brew (works on Mac and Linux)
50
 
 
45
  This model was converted to GGUF format from [`DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B`](https://huggingface.co/DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
46
  Refer to the [original model card](https://huggingface.co/DavidAU/L3.1-MOE-4X8B-Dark-Reasoning-Super-Nova-RP-Hermes-R1-Uncensored-25B) for more details on the model.
47
 
48
+ ---
49
+ Context : 128k.
50
+
51
+
52
+ Required: Llama 3 Instruct template.
53
+
54
+
55
+ "Dark Reasoning" is a variable control reasoning model that is uncensored and operates at all temps/settings and
56
+ is for creative uses cases and general usage.
57
+
58
+
59
+ This is version 2, MOE.
60
+
61
+
62
+ This one is... wild - 4 example generations below.
63
+
64
+
65
+ This is the MOE version - 32B (4X8B) - consisting of four 8B models
66
+ converted to DeepHermes reasoning in a MOE (Mixture of Experts) config
67
+ which results in a 25B "weight" model, that actually has 32B parameters.
68
+ All four reasoning models / experts are activated.
69
+
70
+
71
+ This version's "thinking"/"reasoning" has been "darkened" by FOUR
72
+ source models (see model tree) model's DNA and will also be shorter
73
+ and more compressed. Additional system prompts below to take this a lot
74
+ further - a lot darker, a lot more ... evil.
75
+
76
+
77
+ Higher temps will result in deeper, richer "thoughts"... and frankly more interesting ones too.
78
+
79
+
80
+ With the MOE setup, this model's thinking/output is even stronger.
81
+
82
+
83
+ The "thinking/reasoning" tech (for the model at this repo) is from the original Llama 3.1 "DeepHermes" model from NousResearch:
84
+
85
+
86
+ [ https://huggingface.co/NousResearch/DeepHermes-3-Llama-3-8B-Preview ]
87
+
88
+
89
+ This version will retain all the functions and features of the
90
+ original "DeepHermes" model at about 50%-67% of original reasoning
91
+ power.
92
+ Please visit their repo for all information on features, test results
93
+ and so on.
94
+
95
+ ---
96
  ## Use with llama.cpp
97
  Install llama.cpp through brew (works on Mac and Linux)
98