Commit
·
4b67b56
verified
·
0
Parent(s):
Duplicate from HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1
Browse filesCo-authored-by: Lewis Tunstall <lewtun@users.noreply.huggingface.co>
This view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +35 -0
- README.md +155 -0
- all_results.json +8 -0
- config.json +31 -0
- generation_config.json +6 -0
- logo.png +0 -0
- model-00001-of-00059.safetensors +3 -0
- model-00002-of-00059.safetensors +3 -0
- model-00003-of-00059.safetensors +3 -0
- model-00004-of-00059.safetensors +3 -0
- model-00005-of-00059.safetensors +3 -0
- model-00006-of-00059.safetensors +3 -0
- model-00007-of-00059.safetensors +3 -0
- model-00008-of-00059.safetensors +3 -0
- model-00009-of-00059.safetensors +3 -0
- model-00010-of-00059.safetensors +3 -0
- model-00011-of-00059.safetensors +3 -0
- model-00012-of-00059.safetensors +3 -0
- model-00013-of-00059.safetensors +3 -0
- model-00014-of-00059.safetensors +3 -0
- model-00015-of-00059.safetensors +3 -0
- model-00016-of-00059.safetensors +3 -0
- model-00017-of-00059.safetensors +3 -0
- model-00018-of-00059.safetensors +3 -0
- model-00019-of-00059.safetensors +3 -0
- model-00020-of-00059.safetensors +3 -0
- model-00021-of-00059.safetensors +3 -0
- model-00022-of-00059.safetensors +3 -0
- model-00023-of-00059.safetensors +3 -0
- model-00024-of-00059.safetensors +3 -0
- model-00025-of-00059.safetensors +3 -0
- model-00026-of-00059.safetensors +3 -0
- model-00027-of-00059.safetensors +3 -0
- model-00028-of-00059.safetensors +3 -0
- model-00029-of-00059.safetensors +3 -0
- model-00030-of-00059.safetensors +3 -0
- model-00031-of-00059.safetensors +3 -0
- model-00032-of-00059.safetensors +3 -0
- model-00033-of-00059.safetensors +3 -0
- model-00034-of-00059.safetensors +3 -0
- model-00035-of-00059.safetensors +3 -0
- model-00036-of-00059.safetensors +3 -0
- model-00037-of-00059.safetensors +3 -0
- model-00038-of-00059.safetensors +3 -0
- model-00039-of-00059.safetensors +3 -0
- model-00040-of-00059.safetensors +3 -0
- model-00041-of-00059.safetensors +3 -0
- model-00042-of-00059.safetensors +3 -0
- model-00043-of-00059.safetensors +3 -0
- model-00044-of-00059.safetensors +3 -0
.gitattributes
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
base_model: mistral-community/Mixtral-8x22B-v0.1
|
4 |
+
tags:
|
5 |
+
- trl
|
6 |
+
- orpo
|
7 |
+
- generated_from_trainer
|
8 |
+
datasets:
|
9 |
+
- argilla/distilabel-capybara-dpo-7k-binarized
|
10 |
+
model-index:
|
11 |
+
- name: zephyr-orpo-141b-A35b-v0.1
|
12 |
+
results: []
|
13 |
+
inference:
|
14 |
+
parameters:
|
15 |
+
temperature: 0.7
|
16 |
+
---
|
17 |
+
|
18 |
+
<img src="https://huggingface.co/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1/resolve/main/logo.png" alt="Zephyr 141B Logo" width="400" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
|
19 |
+
|
20 |
+
|
21 |
+
# Model Card for Zephyr 141B-A35B
|
22 |
+
|
23 |
+
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr 141B-A35B is the latest model in the series, and is a fine-tuned version of [mistral-community/Mixtral-8x22B-v0.1](https://huggingface.co/mistral-community/Mixtral-8x22B-v0.1) that was trained using a novel alignment algorithm called [Odds Ratio Preference Optimization (ORPO)](https://huggingface.co/papers/2403.07691) with **7k instances** for **1.3 hours** on 4 nodes of 8 x H100s. ORPO does not require an SFT step to achieve high performance and is thus much more computationally efficient than methods like DPO and PPO. To train Zephyr-141B-A35B, we used the [`argilla/distilabel-capybara-dpo-7k-binarized`](https://huggingface.co/datasets/argilla/distilabel-capybara-dpo-7k-binarized) preference dataset, which consists of synthetic, high-quality, multi-turn preferences that have been scored via LLMs.
|
24 |
+
|
25 |
+
> [!NOTE]
|
26 |
+
> This model was trained collaboratively between Argilla, KAIST, and Hugging Face
|
27 |
+
|
28 |
+
## Model Details
|
29 |
+
|
30 |
+
### Model Description
|
31 |
+
|
32 |
+
<!-- Provide a longer summary of what this model is. -->
|
33 |
+
|
34 |
+
- **Model type:** A Mixture of Experts (MoE) model with 141B total parameters and 35B active parameters. Fine-tuned on a mix of publicly available, synthetic datasets.
|
35 |
+
- **Language(s) (NLP):** Primarily English.
|
36 |
+
- **License:** Apache 2.0
|
37 |
+
- **Finetuned from model:** [mistral-community/Mixtral-8x22B-v0.1](https://huggingface.co/mistral-community/Mixtral-8x22B-v0.1)
|
38 |
+
|
39 |
+
### Model Sources
|
40 |
+
|
41 |
+
<!-- Provide the basic links for the model. -->
|
42 |
+
|
43 |
+
- **Repository:** https://github.com/huggingface/alignment-handbook
|
44 |
+
- **Dataset:** https://huggingface.co/datasets/argilla/distilabel-capybara-dpo-7k-binarized
|
45 |
+
|
46 |
+
## Performance
|
47 |
+
|
48 |
+
Zephyr 141B-A35B was trained to test the effectiveness of ORPO at scale and the underlying dataset contains a mix of general chat capabilities. It achieves strong performance on chat benchmarks like [MT Bench](https://huggingface.co/spaces/lmsys/mt-bench) and [IFEval](https://arxiv.org/abs/2311.07911). The scores reported below were obtained using the [LightEval](https://github.com/huggingface/lighteval) evaluation suite and each prompt has been formatted with the model's corresponding chat template to simulate real-world usage. This is why some scores may differ from those reported in technical reports or on the Open LLM Leaderboard.
|
49 |
+
|
50 |
+
| Model | MT Bench | IFEval | BBH | AGIEval |
|
51 |
+
|-----------------------------------------------------------------------------------------------------|---------:|-------:|------:|--------:|
|
52 |
+
| [zephyr-orpo-141b-A35b-v0.1](https://huggingface.co/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1) | 8.17 | 65.06 | 58.96 | 44.16 |
|
53 |
+
| [databricks/dbrx-instruct](https://huggingface.co/databricks/dbrx-instruct) | 8.26 | 52.13 | 48.50 | 41.16 |
|
54 |
+
| [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) | 8.30 | 55.08 | 45.31 | 47.68 |
|
55 |
+
|
56 |
+
|
57 |
+
## Intended uses & limitations
|
58 |
+
|
59 |
+
The model was fine-tuned on a blend of chat, code, math, and reasoning data. Here's how you can run the model using the `pipeline()` function from 🤗 Transformers:
|
60 |
+
|
61 |
+
```python
|
62 |
+
# pip install 'transformers>=4.39.3'
|
63 |
+
# pip install accelerate
|
64 |
+
|
65 |
+
import torch
|
66 |
+
from transformers import pipeline
|
67 |
+
|
68 |
+
pipe = pipeline(
|
69 |
+
"text-generation",
|
70 |
+
model="HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1",
|
71 |
+
device_map="auto",
|
72 |
+
torch_dtype=torch.bfloat16,
|
73 |
+
)
|
74 |
+
messages = [
|
75 |
+
{
|
76 |
+
"role": "system",
|
77 |
+
"content": "You are Zephyr, a helpful assistant.",
|
78 |
+
},
|
79 |
+
{"role": "user", "content": "Explain how Mixture of Experts work in language a child would understand."},
|
80 |
+
]
|
81 |
+
outputs = pipe(
|
82 |
+
messages,
|
83 |
+
max_new_tokens=512,
|
84 |
+
do_sample=True,
|
85 |
+
temperature=0.7,
|
86 |
+
top_k=50,
|
87 |
+
top_p=0.95,
|
88 |
+
)
|
89 |
+
print(outputs[0]["generated_text"][-1]["content"])
|
90 |
+
```
|
91 |
+
|
92 |
+
## Bias, Risks, and Limitations
|
93 |
+
|
94 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
95 |
+
|
96 |
+
Zephyr 141B-A35B has not been aligned to human preferences for safety within the RLHF phase or deployed with in-the-loop filtering of responses like ChatGPT, so the model can produce problematic outputs (especially when prompted to do so).
|
97 |
+
It is also unknown what the size and composition of the corpus was used to train the base model (`mistral-community/Mixtral-8x22B-v0.1`), however it is likely to have included a mix of Web data and technical sources like books and code. See the [Falcon 180B model card](https://huggingface.co/tiiuae/falcon-180B#training-data) for an example of this.
|
98 |
+
|
99 |
+
|
100 |
+
## Training procedure
|
101 |
+
|
102 |
+
### Training hyperparameters
|
103 |
+
|
104 |
+
The following hyperparameters were used during training:
|
105 |
+
- learning_rate: 5e-06
|
106 |
+
- train_batch_size: 1
|
107 |
+
- eval_batch_size: 8
|
108 |
+
- seed: 42
|
109 |
+
- distributed_type: multi-GPU
|
110 |
+
- num_devices: 32
|
111 |
+
- total_train_batch_size: 32
|
112 |
+
- total_eval_batch_size: 256
|
113 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
114 |
+
- lr_scheduler_type: inverse_sqrt
|
115 |
+
- lr_scheduler_warmup_steps: 100
|
116 |
+
- num_epochs: 3
|
117 |
+
|
118 |
+
### Training results
|
119 |
+
|
120 |
+
|
121 |
+
|
122 |
+
### Framework versions
|
123 |
+
|
124 |
+
- Transformers 4.39.3
|
125 |
+
- Pytorch 2.1.2+cu121
|
126 |
+
- Datasets 2.18.0
|
127 |
+
- Tokenizers 0.15.1
|
128 |
+
|
129 |
+
## Citation
|
130 |
+
|
131 |
+
If you find Zephyr 141B-A35B is useful in your work, please cite the ORPO paper:
|
132 |
+
|
133 |
+
```
|
134 |
+
@misc{hong2024orpo,
|
135 |
+
title={ORPO: Monolithic Preference Optimization without Reference Model},
|
136 |
+
author={Jiwoo Hong and Noah Lee and James Thorne},
|
137 |
+
year={2024},
|
138 |
+
eprint={2403.07691},
|
139 |
+
archivePrefix={arXiv},
|
140 |
+
primaryClass={cs.CL}
|
141 |
+
}
|
142 |
+
```
|
143 |
+
|
144 |
+
You may also wish to cite the creators of this model:
|
145 |
+
|
146 |
+
```
|
147 |
+
@misc{zephyr_141b,
|
148 |
+
author = {Alvaro Bartolome and Jiwoo Hong and Noah Lee and Kashif Rasul and Lewis Tunstall},
|
149 |
+
title = {Zephyr 141B A35B},
|
150 |
+
year = {2024},
|
151 |
+
publisher = {Hugging Face},
|
152 |
+
journal = {Hugging Face repository},
|
153 |
+
howpublished = {\url{https://huggingface.co/HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1}}
|
154 |
+
}
|
155 |
+
```
|
all_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 3.0,
|
3 |
+
"train_loss": 0.812556631554107,
|
4 |
+
"train_runtime": 4771.9621,
|
5 |
+
"train_samples": 6932,
|
6 |
+
"train_samples_per_second": 4.358,
|
7 |
+
"train_steps_per_second": 0.136
|
8 |
+
}
|
config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "mistral-community/Mixtral-8x22B-v0.1",
|
3 |
+
"architectures": [
|
4 |
+
"MixtralForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 1,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 6144,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 16384,
|
13 |
+
"max_position_embeddings": 65536,
|
14 |
+
"model_type": "mixtral",
|
15 |
+
"num_attention_heads": 48,
|
16 |
+
"num_experts_per_tok": 2,
|
17 |
+
"num_hidden_layers": 56,
|
18 |
+
"num_key_value_heads": 8,
|
19 |
+
"num_local_experts": 8,
|
20 |
+
"output_router_logits": false,
|
21 |
+
"rms_norm_eps": 1e-05,
|
22 |
+
"rope_theta": 1000000,
|
23 |
+
"router_aux_loss_coef": 0.001,
|
24 |
+
"router_jitter_noise": 0.0,
|
25 |
+
"sliding_window": null,
|
26 |
+
"tie_word_embeddings": false,
|
27 |
+
"torch_dtype": "bfloat16",
|
28 |
+
"transformers_version": "4.39.3",
|
29 |
+
"use_cache": true,
|
30 |
+
"vocab_size": 32000
|
31 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.39.3"
|
6 |
+
}
|
logo.png
ADDED
![]() |
model-00001-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f623d7c12dc3c39e3f011ac10e2894bb97885723967b646db09aa69a467b84e2
|
3 |
+
size 4998663696
|
model-00002-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79894cf86c8a1c824eae0fb8dce3dda2f88192d16a04ad617250f5b557e6418d
|
3 |
+
size 4806799120
|
model-00003-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4737356e72afbd1d899e51183a7a921c926ea3e2fa1aa03f6c24949ed54c806c
|
3 |
+
size 4806799120
|
model-00004-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07b1544d0594c8eace33d383d5f0c8cc80186b57d2d5924f1f62fbf37edb1850
|
3 |
+
size 4806799120
|
model-00005-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14dedde5e9178654af879895c81d86cdb984d5d2f3ab773c8982026d69c0e8eb
|
3 |
+
size 4806799120
|
model-00006-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f8522fe23daed9d69673009e8c97081e37a98b1c9d8d08fe5d4c94969e40fda
|
3 |
+
size 4806799120
|
model-00007-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:282b30abd7f2acafc3b3e4089a6a832655b9013c3fe4d65a6d3ad7f145a2a762
|
3 |
+
size 4806799120
|
model-00008-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26e9e370572be671609cfc77675d8f89c59afaf9ad69723d9f29a70e6fbd5a02
|
3 |
+
size 4806799120
|
model-00009-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2a8c348a223a71b871c2557157eeb11e229c52359ce68a571070df11d9fd40ea
|
3 |
+
size 4806799120
|
model-00010-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a25a5ca1d0e4674c7c3548f4b4c113ed811ccd14c1cc908b6fd88c5f456975f9
|
3 |
+
size 4806799120
|
model-00011-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02b5f57f8a14c8081a5a87b0d51d807070dfa5888387b7320815044b93a60e33
|
3 |
+
size 4806799136
|
model-00012-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b87b4d457274faef5cdb2849c36b8c2a1fd92be55a7b6b9ae498d823aabd2b9
|
3 |
+
size 4806799152
|
model-00013-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:698074e43a07c2d1fc1eded8ad2fe359c157ef8a157e917e7701d9d38094a55a
|
3 |
+
size 4806799152
|
model-00014-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:efc7c5861083c569421c04e0606631140f1982c7a88ce661291583e751016a2e
|
3 |
+
size 4806799152
|
model-00015-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:149f45ece2a1ade674f568b9f14b3a93ed271998593f9a9a598d108420e00bde
|
3 |
+
size 4806799152
|
model-00016-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c1076aed472ed5117df7922e0597993c620f8b2add0679463c9a1570eb444ffe
|
3 |
+
size 4806799152
|
model-00017-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db4e673a4df8c01b7728cc7200c0c514863702481d6191a6c06916f13a2221e1
|
3 |
+
size 4806799152
|
model-00018-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6797d2776235c46859055ec897edbb42ec6570d73955393bbc338f3c019afb0e
|
3 |
+
size 4806799152
|
model-00019-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:175150f016318ac04287e9f94d087b9ed82baf237ef577717bb7896656cf7085
|
3 |
+
size 4806799152
|
model-00020-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dbc3c4fa6b4616f70e9d19d89ecc381c88ea7e53aab09b55815037452c65e022
|
3 |
+
size 4806799152
|
model-00021-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2424009021647ebb60abce2b3ca56a66cc40bc52b1c5cb15ca9df94d212738ae
|
3 |
+
size 4806799152
|
model-00022-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:270cba731b5358645cf62b8675245a4e1bc66a520d292e75bc4f81d465b34188
|
3 |
+
size 4806799152
|
model-00023-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f450ec6f7509868e884ca80ef96a5d5f970297c57720a8d467c3ffd3292817e2
|
3 |
+
size 4806799152
|
model-00024-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2dbad77697665ec3a81523d8164c2a5eae4e362978be300e6efa1ee0af7a103
|
3 |
+
size 4932529864
|
model-00025-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9a7c3feac72afea74735cc8d372b1f4d0e107bdab3cdea39861e1ad489b00db
|
3 |
+
size 4995542848
|
model-00026-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f57ec4d67f24ade47ef0a8b09adb0f4dd8f6fc1685d2890141e02d11ca5d739
|
3 |
+
size 4995542848
|
model-00027-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ebe99b1f40c4b1740579d274ae2da8a732424380c851a36ede3a0142b080056
|
3 |
+
size 4932628288
|
model-00028-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d34357259ab1d1fd4b2e3857c3f7dbfc68ad8a1ec559fea465953616bb9857b
|
3 |
+
size 4806774344
|
model-00029-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cc629634bf49988aaadac27c1daff0d8fa3350aea7aa98cb8d810f7e748d549
|
3 |
+
size 4806799144
|
model-00030-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf21f65b23fcdc9cb726fe257a545181ec652fb416815670ce6113f18c15adce
|
3 |
+
size 4806799144
|
model-00031-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:526d7e75300daf3a153b6611e3b2a53820f3e59137f9443d3c6c50804009978c
|
3 |
+
size 4806799144
|
model-00032-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9fd1858ccb1779be8e4b4b79466d015a453052703a766a772f54b0f086fb8a2e
|
3 |
+
size 4806799144
|
model-00033-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c18ca349d07cc427ce05b15b05073f3b8020715260780c80697766290dbe3c95
|
3 |
+
size 4806799152
|
model-00034-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5801d046f2895b838c5e29ed0a1d92c9fd5acffc43be55086aff34aa4d353eba
|
3 |
+
size 4806799152
|
model-00035-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aef8078724b01727ea5f3fc6feae0251e058b813a1d5b3e2813d4e76f39d46c9
|
3 |
+
size 4806799152
|
model-00036-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65b826deb2e5a99114667e8317517b0ea244e90a277631fc257512cf7ea65fa6
|
3 |
+
size 4806799152
|
model-00037-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b63df2d0a8e042275ccd0725e7c2b799184e52e550fcc75dfddd29bf3b573a8b
|
3 |
+
size 4806799152
|
model-00038-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3327092547270cb9f680b3e9061243833d8cda2fa7866038e5c3d9c8e387237
|
3 |
+
size 4806799152
|
model-00039-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:180fbadb648dc8150c4171199cb1f0bfb490b78b2f997870e357cab00d8eb3bb
|
3 |
+
size 4806799152
|
model-00040-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c9cef16e5582f606decaf7e7d6c76f29c393b3db016f2959d755688d656f56d
|
3 |
+
size 4806799152
|
model-00041-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff7e2363129f1fc3797d65bdc2db8d969c2901aa0bf50c0697de1143c3c86d52
|
3 |
+
size 4806799152
|
model-00042-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3ad6e4d5dce397995a5db3453cb4266a36e5ac60c8a4b04f7c093670807ceeb
|
3 |
+
size 4806799152
|
model-00043-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9485220622e654a2b950d758102cef11fb04207f691199215198d2c0cc58d87
|
3 |
+
size 4806799152
|
model-00044-of-00059.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a03e4cea8cc668179e8ca768eeabb83c0f7d7f432208bed002bfa068387e66bb
|
3 |
+
size 4806799152
|