CharlesLi commited on
Commit
e9887e4
·
verified ·
1 Parent(s): cea2316

Model save

Browse files
Files changed (4) hide show
  1. README.md +12 -7
  2. all_results.json +6 -11
  3. train_results.json +6 -6
  4. trainer_state.json +132 -42
README.md CHANGED
@@ -3,7 +3,6 @@ base_model: meta-llama/Llama-2-7b-chat-hf
3
  library_name: peft
4
  license: llama2
5
  tags:
6
- - alignment-handbook
7
  - trl
8
  - sft
9
  - generated_from_trainer
@@ -19,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.7346
23
 
24
  ## Model description
25
 
@@ -50,16 +49,22 @@ The following hyperparameters were used during training:
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: cosine
52
  - lr_scheduler_warmup_ratio: 0.1
53
- - training_steps: 20
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss |
58
  |:-------------:|:------:|:----:|:---------------:|
59
- | 1.3704 | 0.7692 | 5 | 0.9995 |
60
- | 0.8977 | 1.5385 | 10 | 0.8262 |
61
- | 0.7547 | 2.3077 | 15 | 0.7461 |
62
- | 0.6981 | 3.0769 | 20 | 0.7346 |
 
 
 
 
 
 
63
 
64
 
65
  ### Framework versions
 
3
  library_name: peft
4
  license: llama2
5
  tags:
 
6
  - trl
7
  - sft
8
  - generated_from_trainer
 
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-chat-hf](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6129
22
 
23
  ## Model description
24
 
 
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: cosine
51
  - lr_scheduler_warmup_ratio: 0.1
52
+ - training_steps: 50
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss |
57
  |:-------------:|:------:|:----:|:---------------:|
58
+ | 1.4804 | 0.7692 | 5 | 1.1583 |
59
+ | 0.9708 | 1.5385 | 10 | 0.8724 |
60
+ | 0.7685 | 2.3077 | 15 | 0.7143 |
61
+ | 0.6454 | 3.0769 | 20 | 0.6602 |
62
+ | 0.5826 | 3.8462 | 25 | 0.6317 |
63
+ | 0.5458 | 4.6154 | 30 | 0.6193 |
64
+ | 0.4844 | 5.3846 | 35 | 0.6151 |
65
+ | 0.506 | 6.1538 | 40 | 0.6133 |
66
+ | 0.4766 | 6.9231 | 45 | 0.6126 |
67
+ | 0.4639 | 7.6923 | 50 | 0.6129 |
68
 
69
 
70
  ### Framework versions
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
- "epoch": 3.076923076923077,
3
- "eval_loss": 0.7345929145812988,
4
- "eval_runtime": 1.4047,
5
- "eval_samples": 20,
6
- "eval_samples_per_second": 14.237,
7
- "eval_steps_per_second": 2.136,
8
- "total_flos": 469873164288.0,
9
- "train_loss": 0.9361693680286407,
10
- "train_runtime": 84.6814,
11
  "train_samples": 100,
12
- "train_samples_per_second": 3.779,
13
- "train_steps_per_second": 0.236
14
  }
 
1
  {
2
+ "epoch": 7.6923076923076925,
3
+ "total_flos": 1205290795008.0,
4
+ "train_loss": 0.692623188495636,
5
+ "train_runtime": 177.9895,
 
 
 
 
 
6
  "train_samples": 100,
7
+ "train_samples_per_second": 4.495,
8
+ "train_steps_per_second": 0.281
9
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.076923076923077,
3
- "total_flos": 469873164288.0,
4
- "train_loss": 0.9361693680286407,
5
- "train_runtime": 84.6814,
6
  "train_samples": 100,
7
- "train_samples_per_second": 3.779,
8
- "train_steps_per_second": 0.236
9
  }
 
1
  {
2
+ "epoch": 7.6923076923076925,
3
+ "total_flos": 1205290795008.0,
4
+ "train_loss": 0.692623188495636,
5
+ "train_runtime": 177.9895,
6
  "train_samples": 100,
7
+ "train_samples_per_second": 4.495,
8
+ "train_steps_per_second": 0.281
9
  }
trainer_state.json CHANGED
@@ -1,94 +1,184 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.076923076923077,
5
  "eval_steps": 5,
6
- "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.15384615384615385,
13
- "grad_norm": 0.7335724302830133,
14
- "learning_rate": 0.0001,
15
  "loss": 1.4896,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.7692307692307693,
20
- "grad_norm": 0.5632629853760255,
21
- "learning_rate": 0.00018660254037844388,
22
- "loss": 1.3704,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.7692307692307693,
27
- "eval_loss": 0.9995014071464539,
28
- "eval_runtime": 2.648,
29
- "eval_samples_per_second": 7.553,
30
- "eval_steps_per_second": 1.133,
31
  "step": 5
32
  },
33
  {
34
  "epoch": 1.5384615384615383,
35
- "grad_norm": 0.35539514680445433,
36
- "learning_rate": 0.00011736481776669306,
37
- "loss": 0.8977,
38
  "step": 10
39
  },
40
  {
41
  "epoch": 1.5384615384615383,
42
- "eval_loss": 0.8261862993240356,
43
- "eval_runtime": 1.4047,
44
- "eval_samples_per_second": 14.238,
45
- "eval_steps_per_second": 2.136,
46
  "step": 10
47
  },
48
  {
49
  "epoch": 2.3076923076923075,
50
- "grad_norm": 0.23338309624627512,
51
- "learning_rate": 3.5721239031346066e-05,
52
- "loss": 0.7547,
53
  "step": 15
54
  },
55
  {
56
  "epoch": 2.3076923076923075,
57
- "eval_loss": 0.7461313605308533,
58
- "eval_runtime": 1.4009,
59
- "eval_samples_per_second": 14.277,
60
- "eval_steps_per_second": 2.141,
61
  "step": 15
62
  },
63
  {
64
  "epoch": 3.076923076923077,
65
- "grad_norm": 0.22429457553641297,
66
- "learning_rate": 0.0,
67
- "loss": 0.6981,
68
  "step": 20
69
  },
70
  {
71
  "epoch": 3.076923076923077,
72
- "eval_loss": 0.7345929145812988,
73
- "eval_runtime": 1.3788,
74
- "eval_samples_per_second": 14.506,
75
- "eval_steps_per_second": 2.176,
76
  "step": 20
77
  },
78
  {
79
- "epoch": 3.076923076923077,
80
- "step": 20,
81
- "total_flos": 469873164288.0,
82
- "train_loss": 0.9361693680286407,
83
- "train_runtime": 84.6814,
84
- "train_samples_per_second": 3.779,
85
- "train_steps_per_second": 0.236
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
86
  }
87
  ],
88
  "logging_steps": 5,
89
- "max_steps": 20,
90
  "num_input_tokens_seen": 0,
91
- "num_train_epochs": 4,
92
  "save_steps": 1000,
93
  "stateful_callbacks": {
94
  "TrainerControl": {
@@ -102,7 +192,7 @@
102
  "attributes": {}
103
  }
104
  },
105
- "total_flos": 469873164288.0,
106
  "train_batch_size": 4,
107
  "trial_name": null,
108
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 7.6923076923076925,
5
  "eval_steps": 5,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.15384615384615385,
13
+ "grad_norm": 0.7337189811685318,
14
+ "learning_rate": 4e-05,
15
  "loss": 1.4896,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.7692307692307693,
20
+ "grad_norm": 0.788344053257058,
21
+ "learning_rate": 0.0002,
22
+ "loss": 1.4804,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.7692307692307693,
27
+ "eval_loss": 1.1582854986190796,
28
+ "eval_runtime": 2.6274,
29
+ "eval_samples_per_second": 7.612,
30
+ "eval_steps_per_second": 1.142,
31
  "step": 5
32
  },
33
  {
34
  "epoch": 1.5384615384615383,
35
+ "grad_norm": 0.4371059352187654,
36
+ "learning_rate": 0.00019396926207859084,
37
+ "loss": 0.9708,
38
  "step": 10
39
  },
40
  {
41
  "epoch": 1.5384615384615383,
42
+ "eval_loss": 0.8723791241645813,
43
+ "eval_runtime": 1.3731,
44
+ "eval_samples_per_second": 14.565,
45
+ "eval_steps_per_second": 2.185,
46
  "step": 10
47
  },
48
  {
49
  "epoch": 2.3076923076923075,
50
+ "grad_norm": 0.25625240789847736,
51
+ "learning_rate": 0.0001766044443118978,
52
+ "loss": 0.7685,
53
  "step": 15
54
  },
55
  {
56
  "epoch": 2.3076923076923075,
57
+ "eval_loss": 0.7142642140388489,
58
+ "eval_runtime": 1.3722,
59
+ "eval_samples_per_second": 14.575,
60
+ "eval_steps_per_second": 2.186,
61
  "step": 15
62
  },
63
  {
64
  "epoch": 3.076923076923077,
65
+ "grad_norm": 0.37095540648653375,
66
+ "learning_rate": 0.00015000000000000001,
67
+ "loss": 0.6454,
68
  "step": 20
69
  },
70
  {
71
  "epoch": 3.076923076923077,
72
+ "eval_loss": 0.660171389579773,
73
+ "eval_runtime": 1.3569,
74
+ "eval_samples_per_second": 14.739,
75
+ "eval_steps_per_second": 2.211,
76
  "step": 20
77
  },
78
  {
79
+ "epoch": 3.8461538461538463,
80
+ "grad_norm": 0.269490348868028,
81
+ "learning_rate": 0.00011736481776669306,
82
+ "loss": 0.5826,
83
+ "step": 25
84
+ },
85
+ {
86
+ "epoch": 3.8461538461538463,
87
+ "eval_loss": 0.6317092776298523,
88
+ "eval_runtime": 1.3548,
89
+ "eval_samples_per_second": 14.762,
90
+ "eval_steps_per_second": 2.214,
91
+ "step": 25
92
+ },
93
+ {
94
+ "epoch": 4.615384615384615,
95
+ "grad_norm": 0.1367860640067106,
96
+ "learning_rate": 8.263518223330697e-05,
97
+ "loss": 0.5458,
98
+ "step": 30
99
+ },
100
+ {
101
+ "epoch": 4.615384615384615,
102
+ "eval_loss": 0.6193430423736572,
103
+ "eval_runtime": 1.3508,
104
+ "eval_samples_per_second": 14.806,
105
+ "eval_steps_per_second": 2.221,
106
+ "step": 30
107
+ },
108
+ {
109
+ "epoch": 5.384615384615385,
110
+ "grad_norm": 0.14497012519202485,
111
+ "learning_rate": 5.000000000000002e-05,
112
+ "loss": 0.4844,
113
+ "step": 35
114
+ },
115
+ {
116
+ "epoch": 5.384615384615385,
117
+ "eval_loss": 0.6151253581047058,
118
+ "eval_runtime": 1.3512,
119
+ "eval_samples_per_second": 14.802,
120
+ "eval_steps_per_second": 2.22,
121
+ "step": 35
122
+ },
123
+ {
124
+ "epoch": 6.153846153846154,
125
+ "grad_norm": 0.1438702628497647,
126
+ "learning_rate": 2.339555568810221e-05,
127
+ "loss": 0.506,
128
+ "step": 40
129
+ },
130
+ {
131
+ "epoch": 6.153846153846154,
132
+ "eval_loss": 0.6133347749710083,
133
+ "eval_runtime": 1.358,
134
+ "eval_samples_per_second": 14.728,
135
+ "eval_steps_per_second": 2.209,
136
+ "step": 40
137
+ },
138
+ {
139
+ "epoch": 6.923076923076923,
140
+ "grad_norm": 0.15810290565530724,
141
+ "learning_rate": 6.030737921409169e-06,
142
+ "loss": 0.4766,
143
+ "step": 45
144
+ },
145
+ {
146
+ "epoch": 6.923076923076923,
147
+ "eval_loss": 0.6125746965408325,
148
+ "eval_runtime": 1.3498,
149
+ "eval_samples_per_second": 14.817,
150
+ "eval_steps_per_second": 2.223,
151
+ "step": 45
152
+ },
153
+ {
154
+ "epoch": 7.6923076923076925,
155
+ "grad_norm": 0.16553701241746885,
156
+ "learning_rate": 0.0,
157
+ "loss": 0.4639,
158
+ "step": 50
159
+ },
160
+ {
161
+ "epoch": 7.6923076923076925,
162
+ "eval_loss": 0.6129434108734131,
163
+ "eval_runtime": 1.3475,
164
+ "eval_samples_per_second": 14.842,
165
+ "eval_steps_per_second": 2.226,
166
+ "step": 50
167
+ },
168
+ {
169
+ "epoch": 7.6923076923076925,
170
+ "step": 50,
171
+ "total_flos": 1205290795008.0,
172
+ "train_loss": 0.692623188495636,
173
+ "train_runtime": 177.9895,
174
+ "train_samples_per_second": 4.495,
175
+ "train_steps_per_second": 0.281
176
  }
177
  ],
178
  "logging_steps": 5,
179
+ "max_steps": 50,
180
  "num_input_tokens_seen": 0,
181
+ "num_train_epochs": 9,
182
  "save_steps": 1000,
183
  "stateful_callbacks": {
184
  "TrainerControl": {
 
192
  "attributes": {}
193
  }
194
  },
195
+ "total_flos": 1205290795008.0,
196
  "train_batch_size": 4,
197
  "trial_name": null,
198
  "trial_params": null