Files changed (1) hide show
  1. README.md +169 -155
README.md CHANGED
@@ -1,155 +1,169 @@
1
- ---
2
- library_name: transformers
3
- tags:
4
- - mergekit
5
- - merge
6
- base_model:
7
- - bunnycore/FuseQwQen-7B
8
- - bunnycore/Qwen-2.5-7B-R1-Stock
9
- - bunnycore/Qwen-2.5-7B-Deep-Stock-v1
10
- - Qwen/Qwen2.5-7B-Instruct
11
- model-index:
12
- - name: Qwen-2.5-7B-Deep-Stock-v4
13
- results:
14
- - task:
15
- type: text-generation
16
- name: Text Generation
17
- dataset:
18
- name: IFEval (0-Shot)
19
- type: HuggingFaceH4/ifeval
20
- args:
21
- num_few_shot: 0
22
- metrics:
23
- - type: inst_level_strict_acc and prompt_level_strict_acc
24
- value: 77.53
25
- name: strict accuracy
26
- source:
27
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
28
- name: Open LLM Leaderboard
29
- - task:
30
- type: text-generation
31
- name: Text Generation
32
- dataset:
33
- name: BBH (3-Shot)
34
- type: BBH
35
- args:
36
- num_few_shot: 3
37
- metrics:
38
- - type: acc_norm
39
- value: 35.91
40
- name: normalized accuracy
41
- source:
42
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
43
- name: Open LLM Leaderboard
44
- - task:
45
- type: text-generation
46
- name: Text Generation
47
- dataset:
48
- name: MATH Lvl 5 (4-Shot)
49
- type: hendrycks/competition_math
50
- args:
51
- num_few_shot: 4
52
- metrics:
53
- - type: exact_match
54
- value: 48.94
55
- name: exact match
56
- source:
57
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
58
- name: Open LLM Leaderboard
59
- - task:
60
- type: text-generation
61
- name: Text Generation
62
- dataset:
63
- name: GPQA (0-shot)
64
- type: Idavidrein/gpqa
65
- args:
66
- num_few_shot: 0
67
- metrics:
68
- - type: acc_norm
69
- value: 6.71
70
- name: acc_norm
71
- source:
72
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
73
- name: Open LLM Leaderboard
74
- - task:
75
- type: text-generation
76
- name: Text Generation
77
- dataset:
78
- name: MuSR (0-shot)
79
- type: TAUR-Lab/MuSR
80
- args:
81
- num_few_shot: 0
82
- metrics:
83
- - type: acc_norm
84
- value: 10.39
85
- name: acc_norm
86
- source:
87
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
88
- name: Open LLM Leaderboard
89
- - task:
90
- type: text-generation
91
- name: Text Generation
92
- dataset:
93
- name: MMLU-PRO (5-shot)
94
- type: TIGER-Lab/MMLU-Pro
95
- config: main
96
- split: test
97
- args:
98
- num_few_shot: 5
99
- metrics:
100
- - type: acc
101
- value: 37.13
102
- name: accuracy
103
- source:
104
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
105
- name: Open LLM Leaderboard
106
- ---
107
- # merge
108
-
109
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
110
-
111
- ## Merge Details
112
- ### Merge Method
113
-
114
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
115
-
116
- ### Models Merged
117
-
118
- The following models were included in the merge:
119
- * [bunnycore/FuseQwQen-7B](https://huggingface.co/bunnycore/FuseQwQen-7B)
120
- * [bunnycore/Qwen-2.5-7B-R1-Stock](https://huggingface.co/bunnycore/Qwen-2.5-7B-R1-Stock)
121
- * [bunnycore/Qwen-2.5-7B-Deep-Stock-v1](https://huggingface.co/bunnycore/Qwen-2.5-7B-Deep-Stock-v1)
122
-
123
- ### Configuration
124
-
125
- The following YAML configuration was used to produce this model:
126
-
127
- ```yaml
128
- models:
129
- - model: bunnycore/FuseQwQen-7B
130
- parameters:
131
- weight: 0.5
132
- - model: bunnycore/Qwen-2.5-7B-Deep-Stock-v1
133
- - model: bunnycore/Qwen-2.5-7B-R1-Stock
134
- - model: Qwen/Qwen2.5-7B-Instruct
135
- base_model: Qwen/Qwen2.5-7B-Instruct
136
- merge_method: model_stock
137
- parameters:
138
- dtype: bfloat16
139
- tokenizer_source: Qwen/Qwen2.5-7B-Instruct
140
-
141
- ```
142
-
143
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
144
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qwen-2.5-7B-Deep-Stock-v4-details)
145
-
146
- | Metric |Value|
147
- |-------------------|----:|
148
- |Avg. |36.10|
149
- |IFEval (0-Shot) |77.53|
150
- |BBH (3-Shot) |35.91|
151
- |MATH Lvl 5 (4-Shot)|48.94|
152
- |GPQA (0-shot) | 6.71|
153
- |MuSR (0-shot) |10.39|
154
- |MMLU-PRO (5-shot) |37.13|
155
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - mergekit
5
+ - merge
6
+ base_model:
7
+ - bunnycore/FuseQwQen-7B
8
+ - bunnycore/Qwen-2.5-7B-R1-Stock
9
+ - bunnycore/Qwen-2.5-7B-Deep-Stock-v1
10
+ - Qwen/Qwen2.5-7B-Instruct
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ model-index:
26
+ - name: Qwen-2.5-7B-Deep-Stock-v4
27
+ results:
28
+ - task:
29
+ type: text-generation
30
+ name: Text Generation
31
+ dataset:
32
+ name: IFEval (0-Shot)
33
+ type: HuggingFaceH4/ifeval
34
+ args:
35
+ num_few_shot: 0
36
+ metrics:
37
+ - type: inst_level_strict_acc and prompt_level_strict_acc
38
+ value: 77.53
39
+ name: strict accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: BBH (3-Shot)
48
+ type: BBH
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 35.91
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: hendrycks/competition_math
64
+ args:
65
+ num_few_shot: 4
66
+ metrics:
67
+ - type: exact_match
68
+ value: 48.94
69
+ name: exact match
70
+ source:
71
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: GPQA (0-shot)
78
+ type: Idavidrein/gpqa
79
+ args:
80
+ num_few_shot: 0
81
+ metrics:
82
+ - type: acc_norm
83
+ value: 6.71
84
+ name: acc_norm
85
+ source:
86
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
87
+ name: Open LLM Leaderboard
88
+ - task:
89
+ type: text-generation
90
+ name: Text Generation
91
+ dataset:
92
+ name: MuSR (0-shot)
93
+ type: TAUR-Lab/MuSR
94
+ args:
95
+ num_few_shot: 0
96
+ metrics:
97
+ - type: acc_norm
98
+ value: 10.39
99
+ name: acc_norm
100
+ source:
101
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: MMLU-PRO (5-shot)
108
+ type: TIGER-Lab/MMLU-Pro
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 37.13
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=bunnycore/Qwen-2.5-7B-Deep-Stock-v4
119
+ name: Open LLM Leaderboard
120
+ ---
121
+ # merge
122
+
123
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
124
+
125
+ ## Merge Details
126
+ ### Merge Method
127
+
128
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
129
+
130
+ ### Models Merged
131
+
132
+ The following models were included in the merge:
133
+ * [bunnycore/FuseQwQen-7B](https://huggingface.co/bunnycore/FuseQwQen-7B)
134
+ * [bunnycore/Qwen-2.5-7B-R1-Stock](https://huggingface.co/bunnycore/Qwen-2.5-7B-R1-Stock)
135
+ * [bunnycore/Qwen-2.5-7B-Deep-Stock-v1](https://huggingface.co/bunnycore/Qwen-2.5-7B-Deep-Stock-v1)
136
+
137
+ ### Configuration
138
+
139
+ The following YAML configuration was used to produce this model:
140
+
141
+ ```yaml
142
+ models:
143
+ - model: bunnycore/FuseQwQen-7B
144
+ parameters:
145
+ weight: 0.5
146
+ - model: bunnycore/Qwen-2.5-7B-Deep-Stock-v1
147
+ - model: bunnycore/Qwen-2.5-7B-R1-Stock
148
+ - model: Qwen/Qwen2.5-7B-Instruct
149
+ base_model: Qwen/Qwen2.5-7B-Instruct
150
+ merge_method: model_stock
151
+ parameters:
152
+ dtype: bfloat16
153
+ tokenizer_source: Qwen/Qwen2.5-7B-Instruct
154
+
155
+ ```
156
+
157
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
158
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/bunnycore__Qwen-2.5-7B-Deep-Stock-v4-details)
159
+
160
+ | Metric |Value|
161
+ |-------------------|----:|
162
+ |Avg. |36.10|
163
+ |IFEval (0-Shot) |77.53|
164
+ |BBH (3-Shot) |35.91|
165
+ |MATH Lvl 5 (4-Shot)|48.94|
166
+ |GPQA (0-shot) | 6.71|
167
+ |MuSR (0-shot) |10.39|
168
+ |MMLU-PRO (5-shot) |37.13|
169
+