RealDragonMA commited on
Commit
d4eda59
·
verified ·
1 Parent(s): 892af04

Model save

Browse files
README.md CHANGED
@@ -27,17 +27,17 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/realdragonma-pelliculum/huggingface/runs/wcdoae50)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.15.0
38
  - Transformers: 4.49.0
39
  - Pytorch: 2.6.0+cu124
40
- - Datasets: 3.3.1
41
  - Tokenizers: 0.21.0
42
 
43
  ## Citations
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/realdragonma-pelliculum/huggingface/runs/b6toipkz)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.15.1
38
  - Transformers: 4.49.0
39
  - Pytorch: 2.6.0+cu124
40
+ - Datasets: 3.3.2
41
  - Tokenizers: 0.21.0
42
 
43
  ## Citations
final_checkpoint/adapter_config.json CHANGED
@@ -23,10 +23,10 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
27
  "q_proj",
28
- "o_proj",
29
- "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "q_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
final_checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf7f8b223f35118ded6f308ce0488d12a93f556f0cc1e31b007ce46ae8940916
3
  size 29523136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0311848f9173527ab70dfbf14762c966af168adf772a6ffd48b22bd118383a0
3
  size 29523136
final_checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19ffad974e6c86ba8d10a67fd65eeaea81497bd8351104df21e2567e618973a0
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:331137efe5ccd44e95d85312fee12e6c9731ee1374a1dd15abc3ffea6a3702a8
3
  size 5624