Jonasbukhave commited on
Commit
4277e2a
·
verified ·
1 Parent(s): 64cd9d8

shawgpt-ft-LoRA-rank-5

Browse files
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 3.8763
20
 
21
  ## Model description
22
 
@@ -51,22 +51,22 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
- | 8.5018 | 0.5714 | 1 | 4.2401 |
55
- | 8.605 | 1.5714 | 2 | 4.2222 |
56
- | 8.3604 | 2.5714 | 3 | 4.1702 |
57
- | 8.3194 | 3.5714 | 4 | 4.1137 |
58
- | 8.1514 | 4.5714 | 5 | 4.0556 |
59
- | 8.0551 | 5.5714 | 6 | 3.9996 |
60
- | 7.975 | 6.5714 | 7 | 3.9512 |
61
- | 7.8857 | 7.5714 | 8 | 3.9139 |
62
- | 7.7969 | 8.5714 | 9 | 3.8888 |
63
- | 3.7972 | 9.5714 | 10 | 3.8763 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - PEFT 0.14.0
69
- - Transformers 4.48.3
70
- - Pytorch 2.5.1+cu124
71
  - Datasets 3.3.1
72
  - Tokenizers 0.21.0
 
16
 
17
  This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 3.8777
20
 
21
  ## Model description
22
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
+ | 25.5434 | 0.5714 | 1 | 4.2401 |
55
+ | 25.8372 | 1.5714 | 2 | 4.2219 |
56
+ | 25.3504 | 2.5714 | 3 | 4.1705 |
57
+ | 25.0908 | 3.5714 | 4 | 4.1147 |
58
+ | 24.6234 | 4.5714 | 5 | 4.0568 |
59
+ | 24.2941 | 5.5714 | 6 | 4.0007 |
60
+ | 23.9362 | 6.5714 | 7 | 3.9525 |
61
+ | 23.8004 | 7.5714 | 8 | 3.9153 |
62
+ | 23.5882 | 8.5714 | 9 | 3.8904 |
63
+ | 15.2731 | 9.5714 | 10 | 3.8777 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - PEFT 0.14.0
69
+ - Transformers 4.47.1
70
+ - Pytorch 2.5.1+cu121
71
  - Datasets 3.3.1
72
  - Tokenizers 0.21.0
runs/Feb18_19-12-19_9a3887f9873e/events.out.tfevents.1739905939.9a3887f9873e.3474.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e90c92d7daa20bb3778423cdca53d23a47867cd8feb0240a045757150180ddf
3
+ size 10835
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95efa2d9cf70417546c061228fd9172190b76a5289daa454fb882325d9843a4e
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6132a688e48c1ef65025b50eac5e7d9f092c393723332a295b19872ed52788ee
3
  size 5368