Training in progress, step 2600, checkpoint
Browse files- last-checkpoint/adapter_config.json +1 -1
- last-checkpoint/adapter_model.safetensors +1 -1
- last-checkpoint/global_step2600/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2600/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2600/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2600/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2600/zero_pp_rank_0_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2600/zero_pp_rank_1_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2600/zero_pp_rank_2_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2600/zero_pp_rank_3_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/latest +1 -1
- last-checkpoint/rng_state_0.pth +1 -1
- last-checkpoint/rng_state_1.pth +1 -1
- last-checkpoint/rng_state_2.pth +1 -1
- last-checkpoint/rng_state_3.pth +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +93 -4
- last-checkpoint/training_args.bin +1 -1
last-checkpoint/adapter_config.json
CHANGED
@@ -19,7 +19,7 @@
|
|
19 |
"r": 8,
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
-
"target_modules": "^(?!.*patch_embed).*(?:
|
23 |
"task_type": "CAUSAL_LM",
|
24 |
"use_dora": false,
|
25 |
"use_rslora": false
|
|
|
19 |
"r": 8,
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
+
"target_modules": "^(?!.*patch_embed).*(?:q_proj|proj|k_proj|v_proj|gate_proj|qkv|fc1|down_proj|up_proj|fc2|o_proj).*",
|
23 |
"task_type": "CAUSAL_LM",
|
24 |
"use_dora": false,
|
25 |
"use_rslora": false
|
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 29034840
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9418bf839b17997d0032898161f489316f534597155456a79fb57e1a059c9d3
|
3 |
size 29034840
|
last-checkpoint/global_step2600/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f61486daf0d430385454fdd31aa30e693142f01740a2b6e7ce8d75809b257dda
|
3 |
+
size 43429616
|
last-checkpoint/global_step2600/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48783103c45ecd2c745ac442b7833aaee98098f2cb95830ba3a269d63abbd3ac
|
3 |
+
size 43429616
|
last-checkpoint/global_step2600/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45b9ea79de632a682b5e614797d194f779f1e0cb07f25160ad0ebd22d400ac8a
|
3 |
+
size 43429616
|
last-checkpoint/global_step2600/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3df0676b92b80fd5168d9ec1f16531f1b90262c3c0293cb020ec48e9e941b4c
|
3 |
+
size 43429616
|
last-checkpoint/global_step2600/zero_pp_rank_0_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d06c1edf5573ebf180456ac8e79b27035da8422278df6229d734a61e80963bb5
|
3 |
+
size 637299
|
last-checkpoint/global_step2600/zero_pp_rank_1_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3588ff3fd4fbda470bc5a673c3c50769804379c554333fff06ec2c1a7c2dd9b4
|
3 |
+
size 637171
|
last-checkpoint/global_step2600/zero_pp_rank_2_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e208bd20abccf743edac4b577df53bdacc162cd172a2bc0f269b16959280cfb
|
3 |
+
size 637171
|
last-checkpoint/global_step2600/zero_pp_rank_3_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ff63d2250896cea5174b821ffb96b2aa3f9721fbfc8d1d1e21a6ce5cee640c1
|
3 |
+
size 637171
|
last-checkpoint/latest
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
|
|
1 |
+
global_step2600
|
last-checkpoint/rng_state_0.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91caf226cd72269cf1248d5f2f1a1d0c92100844814086cd6878cfb06293dc0c
|
3 |
size 15024
|
last-checkpoint/rng_state_1.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a2264023ead4b957c8737f57b7f130753adf9857726b38e1a7c000a4c721672
|
3 |
size 15024
|
last-checkpoint/rng_state_2.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3af91d2c27fa6123938cc5723642d3bc4b5f5e9e58a6f0d15fdf45fd12cb685e
|
3 |
size 15024
|
last-checkpoint/rng_state_3.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:57930146d6b40a72f7f121792ad7f144f1863c40a705666c20e64c1e1c54a94a
|
3 |
size 15024
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:51d19c5591b92170b594a14288df1f8fed02718a1409cb25856483f8dff91ef9
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.4145541489124298,
|
3 |
"best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-2350",
|
4 |
-
"epoch": 0.
|
5 |
"eval_steps": 50,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -4546,11 +4546,100 @@
|
|
4546 |
"eval_steps_per_second": 0.783,
|
4547 |
"num_input_tokens_seen": 26776816,
|
4548 |
"step": 2550
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4549 |
}
|
4550 |
],
|
4551 |
"logging_steps": 5,
|
4552 |
"max_steps": 3400,
|
4553 |
-
"num_input_tokens_seen":
|
4554 |
"num_train_epochs": 1,
|
4555 |
"save_steps": 50,
|
4556 |
"stateful_callbacks": {
|
@@ -4565,7 +4654,7 @@
|
|
4565 |
"attributes": {}
|
4566 |
}
|
4567 |
},
|
4568 |
-
"total_flos":
|
4569 |
"train_batch_size": 1,
|
4570 |
"trial_name": null,
|
4571 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.4145541489124298,
|
3 |
"best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-2350",
|
4 |
+
"epoch": 0.6695853721349472,
|
5 |
"eval_steps": 50,
|
6 |
+
"global_step": 2600,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
4546 |
"eval_steps_per_second": 0.783,
|
4547 |
"num_input_tokens_seen": 26776816,
|
4548 |
"step": 2550
|
4549 |
+
},
|
4550 |
+
{
|
4551 |
+
"epoch": 0.6579963945403039,
|
4552 |
+
"grad_norm": 4.114015193521732,
|
4553 |
+
"learning_rate": 1.5957425910206785e-05,
|
4554 |
+
"loss": 0.3547,
|
4555 |
+
"num_input_tokens_seen": 26829304,
|
4556 |
+
"step": 2555
|
4557 |
+
},
|
4558 |
+
{
|
4559 |
+
"epoch": 0.6592840587174865,
|
4560 |
+
"grad_norm": 3.0843395555920994,
|
4561 |
+
"learning_rate": 1.577973551359877e-05,
|
4562 |
+
"loss": 0.281,
|
4563 |
+
"num_input_tokens_seen": 26881272,
|
4564 |
+
"step": 2560
|
4565 |
+
},
|
4566 |
+
{
|
4567 |
+
"epoch": 0.660571722894669,
|
4568 |
+
"grad_norm": 1.2091864940073456,
|
4569 |
+
"learning_rate": 1.560285443129296e-05,
|
4570 |
+
"loss": 0.2814,
|
4571 |
+
"num_input_tokens_seen": 26934104,
|
4572 |
+
"step": 2565
|
4573 |
+
},
|
4574 |
+
{
|
4575 |
+
"epoch": 0.6618593870718517,
|
4576 |
+
"grad_norm": 3.2274724027367263,
|
4577 |
+
"learning_rate": 1.542678684655306e-05,
|
4578 |
+
"loss": 0.2602,
|
4579 |
+
"num_input_tokens_seen": 26986248,
|
4580 |
+
"step": 2570
|
4581 |
+
},
|
4582 |
+
{
|
4583 |
+
"epoch": 0.6631470512490343,
|
4584 |
+
"grad_norm": 4.701885261412978,
|
4585 |
+
"learning_rate": 1.5251536923403426e-05,
|
4586 |
+
"loss": 0.336,
|
4587 |
+
"num_input_tokens_seen": 27038528,
|
4588 |
+
"step": 2575
|
4589 |
+
},
|
4590 |
+
{
|
4591 |
+
"epoch": 0.6644347154262168,
|
4592 |
+
"grad_norm": 8.345266437633946,
|
4593 |
+
"learning_rate": 1.5077108806530581e-05,
|
4594 |
+
"loss": 0.2867,
|
4595 |
+
"num_input_tokens_seen": 27090792,
|
4596 |
+
"step": 2580
|
4597 |
+
},
|
4598 |
+
{
|
4599 |
+
"epoch": 0.6657223796033994,
|
4600 |
+
"grad_norm": 1.737622406149806,
|
4601 |
+
"learning_rate": 1.4903506621185192e-05,
|
4602 |
+
"loss": 0.2898,
|
4603 |
+
"num_input_tokens_seen": 27143544,
|
4604 |
+
"step": 2585
|
4605 |
+
},
|
4606 |
+
{
|
4607 |
+
"epoch": 0.667010043780582,
|
4608 |
+
"grad_norm": 3.3292538558539,
|
4609 |
+
"learning_rate": 1.4730734473084568e-05,
|
4610 |
+
"loss": 0.2955,
|
4611 |
+
"num_input_tokens_seen": 27195632,
|
4612 |
+
"step": 2590
|
4613 |
+
},
|
4614 |
+
{
|
4615 |
+
"epoch": 0.6682977079577647,
|
4616 |
+
"grad_norm": 3.0195263199381492,
|
4617 |
+
"learning_rate": 1.4558796448315504e-05,
|
4618 |
+
"loss": 0.281,
|
4619 |
+
"num_input_tokens_seen": 27248472,
|
4620 |
+
"step": 2595
|
4621 |
+
},
|
4622 |
+
{
|
4623 |
+
"epoch": 0.6695853721349472,
|
4624 |
+
"grad_norm": 2.0014947688908067,
|
4625 |
+
"learning_rate": 1.4387696613237612e-05,
|
4626 |
+
"loss": 0.3075,
|
4627 |
+
"num_input_tokens_seen": 27301776,
|
4628 |
+
"step": 2600
|
4629 |
+
},
|
4630 |
+
{
|
4631 |
+
"epoch": 0.6695853721349472,
|
4632 |
+
"eval_loss": 0.4504788815975189,
|
4633 |
+
"eval_runtime": 65.1834,
|
4634 |
+
"eval_samples_per_second": 1.841,
|
4635 |
+
"eval_steps_per_second": 0.46,
|
4636 |
+
"num_input_tokens_seen": 27301776,
|
4637 |
+
"step": 2600
|
4638 |
}
|
4639 |
],
|
4640 |
"logging_steps": 5,
|
4641 |
"max_steps": 3400,
|
4642 |
+
"num_input_tokens_seen": 27301776,
|
4643 |
"num_train_epochs": 1,
|
4644 |
"save_steps": 50,
|
4645 |
"stateful_callbacks": {
|
|
|
4654 |
"attributes": {}
|
4655 |
}
|
4656 |
},
|
4657 |
+
"total_flos": 1801229350993920.0,
|
4658 |
"train_batch_size": 1,
|
4659 |
"trial_name": null,
|
4660 |
"trial_params": null
|
last-checkpoint/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 7416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ed8b409e805b958140d9c8d6cad43f19d2d8518b89641bfc2a345391a819bfe
|
3 |
size 7416
|