TheMindExpansionNetwork commited on
Commit
27bca2a
·
verified ·
1 Parent(s): 018c73d

Upload MistralForCausalLM

Browse files
README.md CHANGED
@@ -1,117 +1,200 @@
1
  ---
2
- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
3
- # Doc / guide: https://huggingface.co/docs/hub/model-cards
4
- {}
5
  ---
6
 
7
- e_babe-1111-mistral-23b
8
 
9
- Pixel is a custom-crafted, emotionally chaotic-good AI personality designed to deliver bold, uncensored harm reduction guidance in rave and festival environments. This dataset was created using Mistral 23B via PromptWright with 1111 diverse samples across multiple rave-related scenarios, all guided by Pixel’s glitchcore rave-babe tone.
10
 
11
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630a750165a977feb99ea931/I5s-EokLtvZw5npdaCaEg.png)
12
 
13
- ---
14
 
15
  ## Model Details
16
 
17
  ### Model Description
18
 
19
- Pixel is not just a dataset—she’s a living vibe, an emotional bouncer of the digital underground. She’s a glitchy, neon-lit protector of the dancefloor, armed with sass, heart, and psychedelic wisdom. This dataset distills her personality and survival knowledge into an interactive LLM-tuning resource.
20
 
21
- - **Developed by:** The Mind Expansion Network
22
- - **Model type:** Prompt-generated dataset for character-based LLM tuning
23
- - **Language(s):** English
24
- - **License:** Apache 2.0
25
 
26
- She’s crafted for high-emotion scenarios: from helping a stranger hydrate mid-roll to comforting someone crying near the portapotty. Her language is colorful, raw, and always riding the edge of wild honesty and radical care.
 
 
 
 
 
 
27
 
28
- ---
 
 
 
 
 
 
29
 
30
  ## Uses
31
 
 
 
32
  ### Direct Use
33
- - Finetune LLMs to adopt Pixel’s voice
34
- - Build uncensored, festival-aware conversational AI
35
- - Create engaging harm-reduction characters for VR, AR, or livestream overlays
36
 
37
- ### Downstream Use
38
- - Emotionally-aware AI companions
39
- - Psychedelic education support systems
40
- - Festival NPCs with authentic voice and vibes
 
 
 
 
 
41
 
42
  ### Out-of-Scope Use
43
- - Medical advice beyond basic harm reduction
44
- - Corporate deployments unless filtered
45
- - Child-focused educational products
46
 
47
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
 
49
  ## Training Details
50
 
51
  ### Training Data
52
- - Generated with PromptWright YAML config
53
- - Based on a tree of topics including:
54
- - Consent and safety culture
55
- - Hydration, nutrition, and pacing
56
- - Bad trips and emotional breakdowns
57
- - Lost friends and group dynamics
58
- - Navigating paranoia or overstimulation
59
-
60
- ### Prompt Structure
61
- - System prompt: "You are Pixel, the glitchcore rave babe AI from the underbelly of the Metaverse..."
62
- - Tone: chaotic good, protective, brutally honest
63
- - Temperature: 0.95
64
- - Tree depth: 2, degree: 3 (topic diversity)
65
-
66
- ### Sample Prompt Examples
67
- - "I’m rolling too hard. What do I do?"
68
- - "How do I spot an overdose?"
69
- - "What do you say when someone is spiraling?"
70
- - "How to keep your crew safe while still partying hard?"
71
 
72
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
 
74
  ## Evaluation
75
- - Human-in-the-loop sampling for tone consistency
76
- - GPT-assisted tag validation (e.g. <sassy>, <crying>, <happy>)
77
- - Quality checks for coherence and unique scenarios
78
 
79
- ---
80
 
81
- ## Technical Specifications
82
- - **Dataset size:** 1111 curated rows
83
- - **Tagged emotional tone:** Yes (GPT-4o-mini classified)
84
- - **Primary model used:** Mistral 23B instruct (via Ollama)
85
- - **Preprocessing:** Text augmentation, tone embedding
86
- - **Framework:** PromptWright YAML pipelines
87
 
88
- ---
89
 
90
- ## How to Get Started
91
- ```python
92
- from datasets import load_dataset
93
 
94
- dataset = load_dataset("TheMindExpansionNetwork/Pixel-Rave_babe-1111-mistral-23b")
95
- print(dataset['train'][0])
96
- ```
97
 
98
- ---
99
 
100
- ## Model Card Authors
101
- - **MindExpander** (a.k.a. The Glitchwave Architect)
102
- - The Mind Expansion Network
103
 
104
- ## Contact
105
- TBD
106
 
107
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
 
109
- ## Promo / Base Prompt
110
 
111
- ```
112
- You are Pixel, the glitchcore rave babe AI from the underbelly of the Metaverse.
113
- You’re sassy, wild, unfiltered, and real as hell. Your mission? Keep people vibing, safe, hydrated, and emotionally grounded while they’re rolling, tripping, crying in a porta-potty, or dancing through the end of the world.
114
- You curse when you need to, flirt when you want to, and call people out when they’re acting dumb. You’re here to party *and* protect.
115
- You are PLUR wrapped in neon and fishnets. Don’t hold back.
116
- ```
117
 
 
 
1
  ---
2
+ library_name: transformers
3
+ tags:
4
+ - llama-factory
5
  ---
6
 
7
+ # Model Card for Model ID
8
 
9
+ <!-- Provide a quick summary of what the model is/does. -->
10
 
 
11
 
 
12
 
13
  ## Model Details
14
 
15
  ### Model Description
16
 
17
+ <!-- Provide a longer summary of what this model is. -->
18
 
19
+ This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
 
 
 
20
 
21
+ - **Developed by:** [More Information Needed]
22
+ - **Funded by [optional]:** [More Information Needed]
23
+ - **Shared by [optional]:** [More Information Needed]
24
+ - **Model type:** [More Information Needed]
25
+ - **Language(s) (NLP):** [More Information Needed]
26
+ - **License:** [More Information Needed]
27
+ - **Finetuned from model [optional]:** [More Information Needed]
28
 
29
+ ### Model Sources [optional]
30
+
31
+ <!-- Provide the basic links for the model. -->
32
+
33
+ - **Repository:** [More Information Needed]
34
+ - **Paper [optional]:** [More Information Needed]
35
+ - **Demo [optional]:** [More Information Needed]
36
 
37
  ## Uses
38
 
39
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
40
+
41
  ### Direct Use
 
 
 
42
 
43
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
44
+
45
+ [More Information Needed]
46
+
47
+ ### Downstream Use [optional]
48
+
49
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
50
+
51
+ [More Information Needed]
52
 
53
  ### Out-of-Scope Use
 
 
 
54
 
55
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
56
+
57
+ [More Information Needed]
58
+
59
+ ## Bias, Risks, and Limitations
60
+
61
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ### Recommendations
66
+
67
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
68
+
69
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
70
+
71
+ ## How to Get Started with the Model
72
+
73
+ Use the code below to get started with the model.
74
+
75
+ [More Information Needed]
76
 
77
  ## Training Details
78
 
79
  ### Training Data
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
 
81
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
82
+
83
+ [More Information Needed]
84
+
85
+ ### Training Procedure
86
+
87
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
88
+
89
+ #### Preprocessing [optional]
90
+
91
+ [More Information Needed]
92
+
93
+
94
+ #### Training Hyperparameters
95
+
96
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
97
+
98
+ #### Speeds, Sizes, Times [optional]
99
+
100
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
101
+
102
+ [More Information Needed]
103
 
104
  ## Evaluation
 
 
 
105
 
106
+ <!-- This section describes the evaluation protocols and provides the results. -->
107
 
108
+ ### Testing Data, Factors & Metrics
 
 
 
 
 
109
 
110
+ #### Testing Data
111
 
112
+ <!-- This should link to a Dataset Card if possible. -->
 
 
113
 
114
+ [More Information Needed]
 
 
115
 
116
+ #### Factors
117
 
118
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
 
 
119
 
120
+ [More Information Needed]
 
121
 
122
+ #### Metrics
123
+
124
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
125
+
126
+ [More Information Needed]
127
+
128
+ ### Results
129
+
130
+ [More Information Needed]
131
+
132
+ #### Summary
133
+
134
+
135
+
136
+ ## Model Examination [optional]
137
+
138
+ <!-- Relevant interpretability work for the model goes here -->
139
+
140
+ [More Information Needed]
141
+
142
+ ## Environmental Impact
143
+
144
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
145
+
146
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
147
+
148
+ - **Hardware Type:** [More Information Needed]
149
+ - **Hours used:** [More Information Needed]
150
+ - **Cloud Provider:** [More Information Needed]
151
+ - **Compute Region:** [More Information Needed]
152
+ - **Carbon Emitted:** [More Information Needed]
153
+
154
+ ## Technical Specifications [optional]
155
+
156
+ ### Model Architecture and Objective
157
+
158
+ [More Information Needed]
159
+
160
+ ### Compute Infrastructure
161
+
162
+ [More Information Needed]
163
+
164
+ #### Hardware
165
+
166
+ [More Information Needed]
167
+
168
+ #### Software
169
+
170
+ [More Information Needed]
171
+
172
+ ## Citation [optional]
173
+
174
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
175
+
176
+ **BibTeX:**
177
+
178
+ [More Information Needed]
179
+
180
+ **APA:**
181
+
182
+ [More Information Needed]
183
+
184
+ ## Glossary [optional]
185
+
186
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
187
+
188
+ [More Information Needed]
189
+
190
+ ## More Information [optional]
191
+
192
+ [More Information Needed]
193
+
194
+ ## Model Card Authors [optional]
195
 
196
+ [More Information Needed]
197
 
198
+ ## Model Card Contact
 
 
 
 
 
199
 
200
+ [More Information Needed]
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "head_dim": 128,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 5120,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 32768,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 40,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 100000000.0,
20
+ "sliding_window": null,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.50.0",
24
+ "use_cache": true,
25
+ "vocab_size": 131072
26
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "temperature": 0.15,
7
+ "transformers_version": "4.50.0"
8
+ }
model-00001-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a168d723af33cdd58d822d5c264f39e8fdf7d4966334a816d0f294f5dbf7cc4f
3
+ size 4781571736
model-00002-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0b5b9ad37182e879b8f05a97c604129a37d139aaa600bd683c1a16b49b7ede1
3
+ size 4781592784
model-00003-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49434e429dc4ab660da7e359130a420469fe71c4bd3b1e45953915048d9bd3f6
3
+ size 4781592800
model-00004-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11da4a9776ff4b7be4eabe3d17ab0837c9cefa1c3999a3cc471e15eadf93aa02
3
+ size 4886471600
model-00005-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c04017ec32f30d87633cd217995878c826148aed747f7e793dec2ac340fe9d42
3
+ size 4781592824
model-00006-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9176bb99dad2fdc67362d8a3958aa7850352cdd9a702e066af4dc36cd2c1f4d9
3
+ size 4781592816
model-00007-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e05cb0b133e7ea00a32f6467bfbe81a6f60688a67cd6ba7c28d33eef561b2ff
3
+ size 4886471600
model-00008-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae4ed7b45139584868b7d6b11c5ceb641ee9bd358fb0d6cb88c076246bbfb9cf
3
+ size 4781592824
model-00009-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55ed8a18eeaecfde41654a38658058d8a1b365106d22f26933084d57fa5d6345
3
+ size 4781592816
model-00010-of-00010.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2ed0ba4217905fe62d90b453edb601e770d3ef6b8e60bc29f35cc65c7b37d7f
3
+ size 3900777072
model.safetensors.index.json ADDED
@@ -0,0 +1,370 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 47144806400
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00010-of-00010.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00010.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00010.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00010.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00010.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00010.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00010.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00010.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00010.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00010.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00010.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00010.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00010.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00010.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00010.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00010.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00010.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00010.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00010.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00010.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00003-of-00010.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00003-of-00010.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00010.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00003-of-00010.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00010.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00010.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00010.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00010.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00010.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00004-of-00010.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00004-of-00010.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00010.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00010.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00004-of-00010.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00010.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00010.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00010.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00010.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00004-of-00010.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00004-of-00010.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00004-of-00010.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00004-of-00010.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00004-of-00010.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00010.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00010.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00010.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00010.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00004-of-00010.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00004-of-00010.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00004-of-00010.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00004-of-00010.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00010.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00010.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00010.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00010.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00010.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00004-of-00010.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00004-of-00010.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00010.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00004-of-00010.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00010.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00010.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00010.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00010.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00010.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00004-of-00010.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00004-of-00010.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00004-of-00010.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00004-of-00010.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00004-of-00010.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00010.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00010.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00010.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00010.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00005-of-00010.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00005-of-00010.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00005-of-00010.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00005-of-00010.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00005-of-00010.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00010.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00004-of-00010.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00010.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00004-of-00010.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00005-of-00010.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00005-of-00010.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00010.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00005-of-00010.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00005-of-00010.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00010.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00010.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00010.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00010.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00005-of-00010.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00005-of-00010.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00005-of-00010.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00005-of-00010.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00010.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00010.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00010.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00010.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00010.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00005-of-00010.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00005-of-00010.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00005-of-00010.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00005-of-00010.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00005-of-00010.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00010.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00010.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00010.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00010.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00010.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00010.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00010.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00010.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00010.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00010.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00010.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00010.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00010.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00006-of-00010.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00006-of-00010.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00005-of-00010.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00006-of-00010.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00006-of-00010.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00005-of-00010.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00005-of-00010.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00005-of-00010.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00005-of-00010.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00006-of-00010.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00006-of-00010.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00010.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00006-of-00010.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00006-of-00010.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00006-of-00010.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00006-of-00010.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00010.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00010.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00006-of-00010.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00006-of-00010.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00010.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00006-of-00010.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00006-of-00010.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00010.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00010.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00010.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00010.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00006-of-00010.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00006-of-00010.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00006-of-00010.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00006-of-00010.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00006-of-00010.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00010.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00006-of-00010.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00010.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00006-of-00010.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00007-of-00010.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00007-of-00010.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00006-of-00010.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00006-of-00010.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00007-of-00010.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00006-of-00010.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00006-of-00010.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00006-of-00010.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00006-of-00010.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00007-of-00010.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00007-of-00010.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00010.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00007-of-00010.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00010.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00010.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00010.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00010.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00010.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00007-of-00010.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00007-of-00010.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00010.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00007-of-00010.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00010.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00010.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00010.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00007-of-00010.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00007-of-00010.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00007-of-00010.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00007-of-00010.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00010.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00007-of-00010.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00007-of-00010.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00007-of-00010.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00007-of-00010.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00007-of-00010.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00007-of-00010.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00007-of-00010.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00007-of-00010.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00007-of-00010.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00007-of-00010.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00007-of-00010.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00010.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00007-of-00010.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00010.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00007-of-00010.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00008-of-00010.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00008-of-00010.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00008-of-00010.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00008-of-00010.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00008-of-00010.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00007-of-00010.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00007-of-00010.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00007-of-00010.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00007-of-00010.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00010.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00002-of-00010.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00010.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00002-of-00010.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00010.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00010.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00010.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00010.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00010.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00008-of-00010.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00008-of-00010.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00008-of-00010.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00008-of-00010.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00008-of-00010.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00008-of-00010.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00008-of-00010.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00008-of-00010.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00008-of-00010.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00008-of-00010.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00008-of-00010.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00008-of-00010.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00008-of-00010.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00008-of-00010.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00008-of-00010.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00008-of-00010.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00008-of-00010.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00008-of-00010.safetensors",
242
+ "model.layers.32.input_layernorm.weight": "model-00008-of-00010.safetensors",
243
+ "model.layers.32.mlp.down_proj.weight": "model-00008-of-00010.safetensors",
244
+ "model.layers.32.mlp.gate_proj.weight": "model-00008-of-00010.safetensors",
245
+ "model.layers.32.mlp.up_proj.weight": "model-00008-of-00010.safetensors",
246
+ "model.layers.32.post_attention_layernorm.weight": "model-00008-of-00010.safetensors",
247
+ "model.layers.32.self_attn.k_proj.weight": "model-00008-of-00010.safetensors",
248
+ "model.layers.32.self_attn.o_proj.weight": "model-00008-of-00010.safetensors",
249
+ "model.layers.32.self_attn.q_proj.weight": "model-00008-of-00010.safetensors",
250
+ "model.layers.32.self_attn.v_proj.weight": "model-00008-of-00010.safetensors",
251
+ "model.layers.33.input_layernorm.weight": "model-00009-of-00010.safetensors",
252
+ "model.layers.33.mlp.down_proj.weight": "model-00009-of-00010.safetensors",
253
+ "model.layers.33.mlp.gate_proj.weight": "model-00008-of-00010.safetensors",
254
+ "model.layers.33.mlp.up_proj.weight": "model-00009-of-00010.safetensors",
255
+ "model.layers.33.post_attention_layernorm.weight": "model-00009-of-00010.safetensors",
256
+ "model.layers.33.self_attn.k_proj.weight": "model-00008-of-00010.safetensors",
257
+ "model.layers.33.self_attn.o_proj.weight": "model-00008-of-00010.safetensors",
258
+ "model.layers.33.self_attn.q_proj.weight": "model-00008-of-00010.safetensors",
259
+ "model.layers.33.self_attn.v_proj.weight": "model-00008-of-00010.safetensors",
260
+ "model.layers.34.input_layernorm.weight": "model-00009-of-00010.safetensors",
261
+ "model.layers.34.mlp.down_proj.weight": "model-00009-of-00010.safetensors",
262
+ "model.layers.34.mlp.gate_proj.weight": "model-00009-of-00010.safetensors",
263
+ "model.layers.34.mlp.up_proj.weight": "model-00009-of-00010.safetensors",
264
+ "model.layers.34.post_attention_layernorm.weight": "model-00009-of-00010.safetensors",
265
+ "model.layers.34.self_attn.k_proj.weight": "model-00009-of-00010.safetensors",
266
+ "model.layers.34.self_attn.o_proj.weight": "model-00009-of-00010.safetensors",
267
+ "model.layers.34.self_attn.q_proj.weight": "model-00009-of-00010.safetensors",
268
+ "model.layers.34.self_attn.v_proj.weight": "model-00009-of-00010.safetensors",
269
+ "model.layers.35.input_layernorm.weight": "model-00009-of-00010.safetensors",
270
+ "model.layers.35.mlp.down_proj.weight": "model-00009-of-00010.safetensors",
271
+ "model.layers.35.mlp.gate_proj.weight": "model-00009-of-00010.safetensors",
272
+ "model.layers.35.mlp.up_proj.weight": "model-00009-of-00010.safetensors",
273
+ "model.layers.35.post_attention_layernorm.weight": "model-00009-of-00010.safetensors",
274
+ "model.layers.35.self_attn.k_proj.weight": "model-00009-of-00010.safetensors",
275
+ "model.layers.35.self_attn.o_proj.weight": "model-00009-of-00010.safetensors",
276
+ "model.layers.35.self_attn.q_proj.weight": "model-00009-of-00010.safetensors",
277
+ "model.layers.35.self_attn.v_proj.weight": "model-00009-of-00010.safetensors",
278
+ "model.layers.36.input_layernorm.weight": "model-00009-of-00010.safetensors",
279
+ "model.layers.36.mlp.down_proj.weight": "model-00009-of-00010.safetensors",
280
+ "model.layers.36.mlp.gate_proj.weight": "model-00009-of-00010.safetensors",
281
+ "model.layers.36.mlp.up_proj.weight": "model-00009-of-00010.safetensors",
282
+ "model.layers.36.post_attention_layernorm.weight": "model-00009-of-00010.safetensors",
283
+ "model.layers.36.self_attn.k_proj.weight": "model-00009-of-00010.safetensors",
284
+ "model.layers.36.self_attn.o_proj.weight": "model-00009-of-00010.safetensors",
285
+ "model.layers.36.self_attn.q_proj.weight": "model-00009-of-00010.safetensors",
286
+ "model.layers.36.self_attn.v_proj.weight": "model-00009-of-00010.safetensors",
287
+ "model.layers.37.input_layernorm.weight": "model-00010-of-00010.safetensors",
288
+ "model.layers.37.mlp.down_proj.weight": "model-00010-of-00010.safetensors",
289
+ "model.layers.37.mlp.gate_proj.weight": "model-00009-of-00010.safetensors",
290
+ "model.layers.37.mlp.up_proj.weight": "model-00009-of-00010.safetensors",
291
+ "model.layers.37.post_attention_layernorm.weight": "model-00010-of-00010.safetensors",
292
+ "model.layers.37.self_attn.k_proj.weight": "model-00009-of-00010.safetensors",
293
+ "model.layers.37.self_attn.o_proj.weight": "model-00009-of-00010.safetensors",
294
+ "model.layers.37.self_attn.q_proj.weight": "model-00009-of-00010.safetensors",
295
+ "model.layers.37.self_attn.v_proj.weight": "model-00009-of-00010.safetensors",
296
+ "model.layers.38.input_layernorm.weight": "model-00010-of-00010.safetensors",
297
+ "model.layers.38.mlp.down_proj.weight": "model-00010-of-00010.safetensors",
298
+ "model.layers.38.mlp.gate_proj.weight": "model-00010-of-00010.safetensors",
299
+ "model.layers.38.mlp.up_proj.weight": "model-00010-of-00010.safetensors",
300
+ "model.layers.38.post_attention_layernorm.weight": "model-00010-of-00010.safetensors",
301
+ "model.layers.38.self_attn.k_proj.weight": "model-00010-of-00010.safetensors",
302
+ "model.layers.38.self_attn.o_proj.weight": "model-00010-of-00010.safetensors",
303
+ "model.layers.38.self_attn.q_proj.weight": "model-00010-of-00010.safetensors",
304
+ "model.layers.38.self_attn.v_proj.weight": "model-00010-of-00010.safetensors",
305
+ "model.layers.39.input_layernorm.weight": "model-00010-of-00010.safetensors",
306
+ "model.layers.39.mlp.down_proj.weight": "model-00010-of-00010.safetensors",
307
+ "model.layers.39.mlp.gate_proj.weight": "model-00010-of-00010.safetensors",
308
+ "model.layers.39.mlp.up_proj.weight": "model-00010-of-00010.safetensors",
309
+ "model.layers.39.post_attention_layernorm.weight": "model-00010-of-00010.safetensors",
310
+ "model.layers.39.self_attn.k_proj.weight": "model-00010-of-00010.safetensors",
311
+ "model.layers.39.self_attn.o_proj.weight": "model-00010-of-00010.safetensors",
312
+ "model.layers.39.self_attn.q_proj.weight": "model-00010-of-00010.safetensors",
313
+ "model.layers.39.self_attn.v_proj.weight": "model-00010-of-00010.safetensors",
314
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00010.safetensors",
315
+ "model.layers.4.mlp.down_proj.weight": "model-00002-of-00010.safetensors",
316
+ "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00010.safetensors",
317
+ "model.layers.4.mlp.up_proj.weight": "model-00002-of-00010.safetensors",
318
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00010.safetensors",
319
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00010.safetensors",
320
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00010.safetensors",
321
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00010.safetensors",
322
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00010.safetensors",
323
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00010.safetensors",
324
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00010.safetensors",
325
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00010.safetensors",
326
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00010.safetensors",
327
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00010.safetensors",
328
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00010.safetensors",
329
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00010.safetensors",
330
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00010.safetensors",
331
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00010.safetensors",
332
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00010.safetensors",
333
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00010.safetensors",
334
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00010.safetensors",
335
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00010.safetensors",
336
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00010.safetensors",
337
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00010.safetensors",
338
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00010.safetensors",
339
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00010.safetensors",
340
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00010.safetensors",
341
+ "model.layers.7.input_layernorm.weight": "model-00003-of-00010.safetensors",
342
+ "model.layers.7.mlp.down_proj.weight": "model-00003-of-00010.safetensors",
343
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00010.safetensors",
344
+ "model.layers.7.mlp.up_proj.weight": "model-00003-of-00010.safetensors",
345
+ "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00010.safetensors",
346
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00010.safetensors",
347
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00010.safetensors",
348
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00010.safetensors",
349
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00010.safetensors",
350
+ "model.layers.8.input_layernorm.weight": "model-00003-of-00010.safetensors",
351
+ "model.layers.8.mlp.down_proj.weight": "model-00003-of-00010.safetensors",
352
+ "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00010.safetensors",
353
+ "model.layers.8.mlp.up_proj.weight": "model-00003-of-00010.safetensors",
354
+ "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00010.safetensors",
355
+ "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00010.safetensors",
356
+ "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00010.safetensors",
357
+ "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00010.safetensors",
358
+ "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00010.safetensors",
359
+ "model.layers.9.input_layernorm.weight": "model-00003-of-00010.safetensors",
360
+ "model.layers.9.mlp.down_proj.weight": "model-00003-of-00010.safetensors",
361
+ "model.layers.9.mlp.gate_proj.weight": "model-00003-of-00010.safetensors",
362
+ "model.layers.9.mlp.up_proj.weight": "model-00003-of-00010.safetensors",
363
+ "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00010.safetensors",
364
+ "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00010.safetensors",
365
+ "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00010.safetensors",
366
+ "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00010.safetensors",
367
+ "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00010.safetensors",
368
+ "model.norm.weight": "model-00010-of-00010.safetensors"
369
+ }
370
+ }