ben81828 commited on
Commit
b94b820
·
verified ·
1 Parent(s): 0e8eee8

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f299ed065d076e986e51213d7bf889152797a26591dce7e23c56b2cecc05e88
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9637897eed1592d2d482e3090638e19cf1f19ebaae7ca0874cfe9fa7a0f34af2
3
  size 29034840
last-checkpoint/global_step200/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b01b9e430f2710b81c6b92378e0c3ce44b48055454229f6cb73b50aedc64d47
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7556b8d4bd28531abff384f75c5d5213c595b1ae36b8228de71ead7ec3ca7977
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d45b3f8a44ffc528838848732ad254738d53cfcf9cd5e442df7e0b0b87a05259
3
+ size 43429616
last-checkpoint/global_step200/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:121e763f96625efd69ec96d2af62ed25b0b8c9571a49971dbd882a8a20f8bcbb
3
+ size 43429616
last-checkpoint/global_step200/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c8aed679b2d5c060599974f1ab45390fb83c42d37e4ed63003977054cda112
3
+ size 637299
last-checkpoint/global_step200/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7ceb1154d5065a4632b1731df2acb5e51080a1f0388344212eaec5524452314
3
+ size 637171
last-checkpoint/global_step200/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30465b90efa0eb73fc80dc870a336c0cc691a34eaf088821a381a9f71ee7d652
3
+ size 637171
last-checkpoint/global_step200/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab524daca08828de12f8863c12ac5a32a74f179700a4a2850b77f7debcf459b3
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step150
 
1
+ global_step200
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70cc56408014c410353d4dd58ae9b03f4be043f5f800324f66fd8e20e99b840e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c5e18f922d0af74d820247ae97bee506ab412554a58345ddf2558abc94ee3e3
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49d1438e98cc9c53a6852464635ce62e9788e61eb3646b73e33813f487c4b6ae
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a2dcca6d9741f46592359768ea2212b9321da6408d1fd7d3a80b017bf37f434
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4388add9cec90932f8ff0100d27a0574d98e1bad52ff89d44e31967d2b4fbfde
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69420ece2c255923c5cbb3c6c9c4a6b9cb38fb57e5d3033c8b7d436a1faf6f13
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a705d6dfaae4f2c1b4b2be6b25a6eb521ffae6fcba21cc1531e97b60037ed079
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f278b40a1e23b88a657c4e5d03afa8dbbbe14dfeb16f6b4beedaece6cdd0b9
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c587ef55825cdfdbac47c29c3bd8a2996263c16833a5bd6aaa2fb014bea1e9d1
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fdde69e82f45c9368fe627c2082d010eb77a8ceb0b2354f60caa089407a428a6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7309949994087219,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-150",
4
- "epoch": 0.03862992531547772,
5
  "eval_steps": 50,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -274,11 +274,100 @@
274
  "eval_steps_per_second": 0.785,
275
  "num_input_tokens_seen": 1573376,
276
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
277
  }
278
  ],
279
  "logging_steps": 5,
280
  "max_steps": 3400,
281
- "num_input_tokens_seen": 1573376,
282
  "num_train_epochs": 1,
283
  "save_steps": 50,
284
  "stateful_callbacks": {
@@ -293,7 +382,7 @@
293
  "attributes": {}
294
  }
295
  },
296
- "total_flos": 103777530806272.0,
297
  "train_batch_size": 1,
298
  "trial_name": null,
299
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7274295687675476,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-200",
4
+ "epoch": 0.05150656708730363,
5
  "eval_steps": 50,
6
+ "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
274
  "eval_steps_per_second": 0.785,
275
  "num_input_tokens_seen": 1573376,
276
  "step": 150
277
+ },
278
+ {
279
+ "epoch": 0.03991758949266031,
280
+ "grad_norm": 0.5536144453733772,
281
+ "learning_rate": 9.11764705882353e-05,
282
+ "loss": 0.738,
283
+ "num_input_tokens_seen": 1626136,
284
+ "step": 155
285
+ },
286
+ {
287
+ "epoch": 0.04120525366984291,
288
+ "grad_norm": 0.5151715191704441,
289
+ "learning_rate": 9.411764705882353e-05,
290
+ "loss": 0.7579,
291
+ "num_input_tokens_seen": 1678760,
292
+ "step": 160
293
+ },
294
+ {
295
+ "epoch": 0.042492917847025496,
296
+ "grad_norm": 0.5209077394596254,
297
+ "learning_rate": 9.705882352941177e-05,
298
+ "loss": 0.7502,
299
+ "num_input_tokens_seen": 1731240,
300
+ "step": 165
301
+ },
302
+ {
303
+ "epoch": 0.043780582024208085,
304
+ "grad_norm": 0.721213601237688,
305
+ "learning_rate": 0.0001,
306
+ "loss": 0.7448,
307
+ "num_input_tokens_seen": 1783816,
308
+ "step": 170
309
+ },
310
+ {
311
+ "epoch": 0.045068246201390674,
312
+ "grad_norm": 0.48666007914879555,
313
+ "learning_rate": 9.999940874631277e-05,
314
+ "loss": 0.6648,
315
+ "num_input_tokens_seen": 1834592,
316
+ "step": 175
317
+ },
318
+ {
319
+ "epoch": 0.04635591037857327,
320
+ "grad_norm": 0.5136600613696797,
321
+ "learning_rate": 9.999763499923432e-05,
322
+ "loss": 0.7759,
323
+ "num_input_tokens_seen": 1888176,
324
+ "step": 180
325
+ },
326
+ {
327
+ "epoch": 0.04764357455575586,
328
+ "grad_norm": 0.6706281530046975,
329
+ "learning_rate": 9.999467880071402e-05,
330
+ "loss": 0.7167,
331
+ "num_input_tokens_seen": 1940280,
332
+ "step": 185
333
+ },
334
+ {
335
+ "epoch": 0.04893123873293845,
336
+ "grad_norm": 0.5159139445497618,
337
+ "learning_rate": 9.999054022066641e-05,
338
+ "loss": 0.7483,
339
+ "num_input_tokens_seen": 1993096,
340
+ "step": 190
341
+ },
342
+ {
343
+ "epoch": 0.050218902910121044,
344
+ "grad_norm": 0.40251006129746847,
345
+ "learning_rate": 9.998521935696953e-05,
346
+ "loss": 0.7464,
347
+ "num_input_tokens_seen": 2045648,
348
+ "step": 195
349
+ },
350
+ {
351
+ "epoch": 0.05150656708730363,
352
+ "grad_norm": 0.4811730853311867,
353
+ "learning_rate": 9.997871633546257e-05,
354
+ "loss": 0.7594,
355
+ "num_input_tokens_seen": 2099008,
356
+ "step": 200
357
+ },
358
+ {
359
+ "epoch": 0.05150656708730363,
360
+ "eval_loss": 0.7274295687675476,
361
+ "eval_runtime": 38.079,
362
+ "eval_samples_per_second": 3.151,
363
+ "eval_steps_per_second": 0.788,
364
+ "num_input_tokens_seen": 2099008,
365
+ "step": 200
366
  }
367
  ],
368
  "logging_steps": 5,
369
  "max_steps": 3400,
370
+ "num_input_tokens_seen": 2099008,
371
  "num_train_epochs": 1,
372
  "save_steps": 50,
373
  "stateful_callbacks": {
 
382
  "attributes": {}
383
  }
384
  },
385
+ "total_flos": 138478836776960.0,
386
  "train_batch_size": 1,
387
  "trial_name": null,
388
  "trial_params": null