leixa commited on
Commit
911a683
·
verified ·
1 Parent(s): 5c06d3f

Training in progress, step 96, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:611732e8b60b82f65ca3fb621af02d76e06289cda1042b9cb51448095f7f99f2
3
  size 201892112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10dd6e59839697983ac7357e410711a9437ba43312590efd839cd9b3c1eaf5ee
3
  size 201892112
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b78fd360fd0837a0e24e78ff53bcb54596b905619d3f1795815dbe338f7fdc0
3
  size 102864548
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87674fb31f651703f67324edeb4965dd0bab54531a95f1888f2f0b1f8a781f10
3
  size 102864548
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bea3603c9a5ac93aadff8538e8c1f7b7097253fb14e88f0b33e0bdddfa5087f6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20741c0db4af8011c95f93fd02a96b812116cb014b8f60028a24e84ef0e0f0a6
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87bd97bc98e39f7007b5f2750d097c7395ce4bde9cd90085b1ec2b6c643faaf8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80ffdba0827548f3b413b86d873a6fdae4989ec2c782588463bc4ec0efdb484e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.75,
5
  "eval_steps": 24,
6
- "global_step": 72,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -207,6 +207,70 @@
207
  "eval_samples_per_second": 48.555,
208
  "eval_steps_per_second": 6.294,
209
  "step": 72
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
210
  }
211
  ],
212
  "logging_steps": 3,
@@ -226,7 +290,7 @@
226
  "attributes": {}
227
  }
228
  },
229
- "total_flos": 1.6211923678789632e+16,
230
  "train_batch_size": 8,
231
  "trial_name": null,
232
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
  "eval_steps": 24,
6
+ "global_step": 96,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
207
  "eval_samples_per_second": 48.555,
208
  "eval_steps_per_second": 6.294,
209
  "step": 72
210
+ },
211
+ {
212
+ "epoch": 0.78125,
213
+ "grad_norm": 0.3321720063686371,
214
+ "learning_rate": 4.355339409154788e-05,
215
+ "loss": 1.4901,
216
+ "step": 75
217
+ },
218
+ {
219
+ "epoch": 0.8125,
220
+ "grad_norm": 0.349568635225296,
221
+ "learning_rate": 4.297477038394368e-05,
222
+ "loss": 1.5304,
223
+ "step": 78
224
+ },
225
+ {
226
+ "epoch": 0.84375,
227
+ "grad_norm": 0.38185468316078186,
228
+ "learning_rate": 4.237548933539718e-05,
229
+ "loss": 1.502,
230
+ "step": 81
231
+ },
232
+ {
233
+ "epoch": 0.875,
234
+ "grad_norm": 0.37284818291664124,
235
+ "learning_rate": 4.1756239664188273e-05,
236
+ "loss": 1.4548,
237
+ "step": 84
238
+ },
239
+ {
240
+ "epoch": 0.90625,
241
+ "grad_norm": 0.3401886522769928,
242
+ "learning_rate": 4.1117733037354314e-05,
243
+ "loss": 1.49,
244
+ "step": 87
245
+ },
246
+ {
247
+ "epoch": 0.9375,
248
+ "grad_norm": 0.31730660796165466,
249
+ "learning_rate": 4.0460703252813326e-05,
250
+ "loss": 1.485,
251
+ "step": 90
252
+ },
253
+ {
254
+ "epoch": 0.96875,
255
+ "grad_norm": 0.34687182307243347,
256
+ "learning_rate": 3.978590539605338e-05,
257
+ "loss": 1.5021,
258
+ "step": 93
259
+ },
260
+ {
261
+ "epoch": 1.0,
262
+ "grad_norm": 0.4444548487663269,
263
+ "learning_rate": 3.9094114972357524e-05,
264
+ "loss": 1.4711,
265
+ "step": 96
266
+ },
267
+ {
268
+ "epoch": 1.0,
269
+ "eval_loss": 1.452917218208313,
270
+ "eval_runtime": 3.3544,
271
+ "eval_samples_per_second": 48.294,
272
+ "eval_steps_per_second": 6.26,
273
+ "step": 96
274
  }
275
  ],
276
  "logging_steps": 3,
 
290
  "attributes": {}
291
  }
292
  },
293
+ "total_flos": 2.1558125484048384e+16,
294
  "train_batch_size": 8,
295
  "trial_name": null,
296
  "trial_params": null