Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
base_model:
|
6 |
+
- Qwen/Qwen2.5-Coder-32B
|
7 |
+
- open-r1/OlympicCoder-32B
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
tags:
|
10 |
+
- merge
|
11 |
+
- programming
|
12 |
+
- code generation
|
13 |
+
- code
|
14 |
+
- qwen2
|
15 |
+
- codeqwen
|
16 |
+
- chat
|
17 |
+
- qwen
|
18 |
+
- qwen-coder
|
19 |
+
library_name: transformers
|
20 |
+
---
|
21 |
+
|
22 |
+
<h2>Qwen2.5-Godzilla-Coder-51B-gguf</h2>
|
23 |
+
|
24 |
+
<img src="godzilla-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
|
25 |
+
|
26 |
+
"It will pound your programming problems into the pavement... perfectly."
|
27 |
+
|
28 |
+
Tipping the scales at 101 layers and 1215 tensors... the monster lives.
|
29 |
+
|
30 |
+
Two monsters in fact.
|
31 |
+
|
32 |
+
Each model generates stronger, more compact code with an enhanced understanding of your instructions and follows what you tell them to the letter.
|
33 |
+
|
34 |
+
And then some.
|
35 |
+
|
36 |
+
These overpowered CODING ENGINEs are based on two of the best coder AIs:
|
37 |
+
|
38 |
+
"Qwen2.5-Coder-32B-Instruct"
|
39 |
+
|
40 |
+
[ https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct ]
|
41 |
+
|
42 |
+
and
|
43 |
+
|
44 |
+
"OlympicCoder-32B"
|
45 |
+
|
46 |
+
[ https://huggingface.co/open-r1/OlympicCoder-32B ]
|
47 |
+
|
48 |
+
These two models are stuffed into one MASSSIVE 51B merge that is stronger in performance and understanding than both donor models.
|
49 |
+
|
50 |
+
CONFIGS:
|
51 |
+
- #1 -> Qwen2.5-Coder-32B-Instruct primary/start, with OlympicCoder-32B as "finalizer".
|
52 |
+
- #2 -> OlympicCoder-32B as primary/start, with Qwen2.5-Coder-32B-Instruct as "finalizer".
|
53 |
+
|
54 |
+
NOTES:
|
55 |
+
- Each config/version will be very different from each other.
|
56 |
+
- Tool Calling is supported in both versions.
|
57 |
+
- Source(s) / full quanting to follow // full repos to follow.
|
58 |
+
- Model is fully operational at Q2k - both versions - and stronger than the base donor models in terms of raw performance.
|
59 |
+
- Final model size (including layers/tensors) / config subject to change.
|
60 |
+
|
61 |
+
---
|
62 |
+
|
63 |
+
Config / Settings
|
64 |
+
|
65 |
+
---
|
66 |
+
|
67 |
+
Model is set at 32k/32768 context for these GGUFS, full quants/full repos will be 128k/131072.
|
68 |
+
|
69 |
+
Requirements [Qwen 2.5 32B Coder default settings]:
|
70 |
+
- Temp .5 to .7 (or lower)
|
71 |
+
- topk: 20, topp: .8, minp: .05
|
72 |
+
- rep pen: 1.1 (can be lower)
|
73 |
+
- Jinja Template (embedded) or CHATML template.
|
74 |
+
- A System Prompt is not required. (ran tests with blank system prompt)
|
75 |
+
|
76 |
+
Refer to either "Qwen2.5-Coder-32B-Instruct" and/or "OlympicCoder-32B" repos (above) for additional settings, benchmarks and usage.
|
77 |
+
|
78 |
+
---
|