Gangwoo commited on
Commit
a727476
·
1 Parent(s): dea0537

Model save

Browse files
README.md CHANGED
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 1.5184
18
 
19
  ## Model description
20
 
@@ -47,12 +47,12 @@ The following hyperparameters were used during training:
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:----:|:---------------:|
50
- | 0.3033 | 1.0 | 1 | 1.5184 |
51
 
52
 
53
  ### Framework versions
54
 
55
  - Transformers 4.35.0
56
- - Pytorch 2.1.0
57
  - Datasets 2.14.6
58
  - Tokenizers 0.14.1
 
14
 
15
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 1.5189
18
 
19
  ## Model description
20
 
 
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:----:|:---------------:|
50
+ | 0.3033 | 1.0 | 1 | 1.5189 |
51
 
52
 
53
  ### Framework versions
54
 
55
  - Transformers 4.35.0
56
+ - Pytorch 2.1.1
57
  - Datasets 2.14.6
58
  - Tokenizers 0.14.1
adapter_config.json CHANGED
@@ -16,8 +16,8 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "o_proj",
20
  "q_proj",
 
21
  "k_proj",
22
  "v_proj"
23
  ],
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "q_proj",
20
+ "o_proj",
21
  "k_proj",
22
  "v_proj"
23
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a6e588007baf58e516ad55d8255d8e94df1294163ebb59aa343e576e2054ae3
3
  size 268470272
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10694f26aba773f17fb2e3b021d0bc81023d360200fdca5c854a2b66f011d177
3
  size 268470272
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 1.5184062719345093,
4
- "eval_runtime": 17.7955,
5
  "eval_samples": 101,
6
- "eval_samples_per_second": 5.676,
7
- "eval_steps_per_second": 0.731,
8
- "train_loss": 0.30328860878944397,
9
- "train_runtime": 67.0626,
10
  "train_samples": 101,
11
- "train_samples_per_second": 1.506,
12
- "train_steps_per_second": 0.015
13
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 1.518870234489441,
4
+ "eval_runtime": 9.6585,
5
  "eval_samples": 101,
6
+ "eval_samples_per_second": 10.457,
7
+ "eval_steps_per_second": 1.346,
8
+ "train_loss": 0.30327314138412476,
9
+ "train_runtime": 39.3282,
10
  "train_samples": 101,
11
+ "train_samples_per_second": 2.568,
12
+ "train_steps_per_second": 0.025
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 1.5184062719345093,
4
- "eval_runtime": 17.7955,
5
  "eval_samples": 101,
6
- "eval_samples_per_second": 5.676,
7
- "eval_steps_per_second": 0.731
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 1.518870234489441,
4
+ "eval_runtime": 9.6585,
5
  "eval_samples": 101,
6
+ "eval_samples_per_second": 10.457,
7
+ "eval_steps_per_second": 1.346
8
  }
runs/Jan04_15-52-04_bdata-login01/events.out.tfevents.1704351270.bdata-login01.89932.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87639f5d14b92804c4f951f2cae5ef35f1d5a008503c08ca6aba51c0e6851986
3
+ size 5100
runs/Jan04_15-52-04_bdata-login01/events.out.tfevents.1704351319.bdata-login01.89932.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bee66e36f9b70b50eb951b8aeec586813651d3e692abba6e8dbcae9b5350cfe
3
+ size 354
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.30328860878944397,
4
- "train_runtime": 67.0626,
5
  "train_samples": 101,
6
- "train_samples_per_second": 1.506,
7
- "train_steps_per_second": 0.015
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.30327314138412476,
4
+ "train_runtime": 39.3282,
5
  "train_samples": 101,
6
+ "train_samples_per_second": 2.568,
7
+ "train_steps_per_second": 0.025
8
  }
trainer_state.json CHANGED
@@ -16,20 +16,20 @@
16
  },
17
  {
18
  "epoch": 1.0,
19
- "eval_loss": 1.5184062719345093,
20
- "eval_runtime": 18.0579,
21
- "eval_samples_per_second": 5.593,
22
- "eval_steps_per_second": 0.72,
23
  "step": 1
24
  },
25
  {
26
  "epoch": 1.0,
27
  "step": 1,
28
  "total_flos": 2285025504337920.0,
29
- "train_loss": 0.30328860878944397,
30
- "train_runtime": 67.0626,
31
- "train_samples_per_second": 1.506,
32
- "train_steps_per_second": 0.015
33
  }
34
  ],
35
  "logging_steps": 5,
 
16
  },
17
  {
18
  "epoch": 1.0,
19
+ "eval_loss": 1.518870234489441,
20
+ "eval_runtime": 9.6429,
21
+ "eval_samples_per_second": 10.474,
22
+ "eval_steps_per_second": 1.348,
23
  "step": 1
24
  },
25
  {
26
  "epoch": 1.0,
27
  "step": 1,
28
  "total_flos": 2285025504337920.0,
29
+ "train_loss": 0.30327314138412476,
30
+ "train_runtime": 39.3282,
31
+ "train_samples_per_second": 2.568,
32
+ "train_steps_per_second": 0.025
33
  }
34
  ],
35
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47ed542bd6bc35271666c4eba73e060bc73a44314bda10d405a69eef10e78d89
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6297c4020386090a5c357653c2298e45fbf900aa81265a079438e05349abd0f
3
  size 4664