ben81828 commited on
Commit
35c7cc3
·
verified ·
1 Parent(s): 360ee15

Training in progress, step 1750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40488b690d3cbb62aef973e623da597fed2f1f503eac48f0d6e9e0140377208a
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95cb389d8c3bb8b023cd12cb33d7b6158b8a28b87779d6ac67f1897a141b861b
3
  size 29034840
last-checkpoint/global_step1750/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e8347056f1292ff97b1d73659503a0507b0653d32cc7a12e8ff9a1335f37028
3
+ size 43429616
last-checkpoint/global_step1750/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cbf33845645e53961a3c6bbb0a1d7cf348afeabd1d176afadd7cdfbca377cca
3
+ size 43429616
last-checkpoint/global_step1750/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e35335e72d59742db35a88656b081cd053a9d171496abe9eee765199e679e519
3
+ size 43429616
last-checkpoint/global_step1750/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1296baea3815cacda5acc196065ffa65a15808a01253fad978a91e114aff5889
3
+ size 43429616
last-checkpoint/global_step1750/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32f8c5b1da88eb2c3bd967f43f6931692786200c22aa97a202ebf66082557677
3
+ size 637299
last-checkpoint/global_step1750/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:037fcd6af5c00ed77baab88df849e8698d197844c05cc017806d8e74ae2f1b95
3
+ size 637171
last-checkpoint/global_step1750/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99970103805cf54854383f6ea3849fdf6038c8a68a2a992a04bd9ebaf6f4251d
3
+ size 637171
last-checkpoint/global_step1750/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd3d1f71573606393b3cb5f1deddf23fe9aa63fa2404d1a9570e793efcfa8328
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1700
 
1
+ global_step1750
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e7c17922709137dd2f358be87cc431f7959a56821e2d051582e81679e2d388e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:222e5f04f66dfcca4efb2d648f5a480d56c2a07755d7a1bae8232d01d4b479ce
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eaf0ec3cc431efdc1cb595b27e74fc020021d53f5c11850d7490a79bf42b71d6
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:150a9cfa07bd33135b1b8b22033907e44137689de662dda0a482f3af84c5a1ff
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:148fd0a663c121acf74d815964bc0d08d07065e8503b03adfd967fdaf63a6abc
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0c0d7619e94c90efa0c89a1f208db53b48c726519761710da6fbc31a80651d8
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56589f6c34527656fe0dc752874cfc1460efbb5c0c3f874c4cd09d6415dd16c1
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa66a4b2f749b1ec7e8868668dc670ff3f6d8df765a5249122f980ae5aae6a54
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcafe96450339d5b6e4ff3a8d909c473a6afaab9bb34bc798805cded569190ca
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a1c59a2217a7993f4a6258f4be38869d3dc42c6bc2349bde5d4858846c79cb4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.4339977502822876,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-1250",
4
- "epoch": 0.43780582024208087,
5
  "eval_steps": 50,
6
- "global_step": 1700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3033,11 +3033,100 @@
3033
  "eval_steps_per_second": 0.786,
3034
  "num_input_tokens_seen": 17845600,
3035
  "step": 1700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3036
  }
3037
  ],
3038
  "logging_steps": 5,
3039
  "max_steps": 3400,
3040
- "num_input_tokens_seen": 17845600,
3041
  "num_train_epochs": 1,
3042
  "save_steps": 50,
3043
  "stateful_callbacks": {
@@ -3052,7 +3141,7 @@
3052
  "attributes": {}
3053
  }
3054
  },
3055
- "total_flos": 1177420775555072.0,
3056
  "train_batch_size": 1,
3057
  "trial_name": null,
3058
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.4234265685081482,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_detect_scale4/lora/sft/checkpoint-1750",
4
+ "epoch": 0.45068246201390677,
5
  "eval_steps": 50,
6
+ "global_step": 1750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3033
  "eval_steps_per_second": 0.786,
3034
  "num_input_tokens_seen": 17845600,
3035
  "step": 1700
3036
+ },
3037
+ {
3038
+ "epoch": 0.43909348441926344,
3039
+ "grad_norm": 6.805905061208508,
3040
+ "learning_rate": 5.388659258287102e-05,
3041
+ "loss": 0.4066,
3042
+ "num_input_tokens_seen": 17897920,
3043
+ "step": 1705
3044
+ },
3045
+ {
3046
+ "epoch": 0.44038114859644606,
3047
+ "grad_norm": 6.964220035587859,
3048
+ "learning_rate": 5.364412597363759e-05,
3049
+ "loss": 0.3599,
3050
+ "num_input_tokens_seen": 17950920,
3051
+ "step": 1710
3052
+ },
3053
+ {
3054
+ "epoch": 0.4416688127736286,
3055
+ "grad_norm": 6.2080800188283956,
3056
+ "learning_rate": 5.3401573180287426e-05,
3057
+ "loss": 0.3681,
3058
+ "num_input_tokens_seen": 18003280,
3059
+ "step": 1715
3060
+ },
3061
+ {
3062
+ "epoch": 0.44295647695081125,
3063
+ "grad_norm": 6.3073464125282195,
3064
+ "learning_rate": 5.315893993922986e-05,
3065
+ "loss": 0.4005,
3066
+ "num_input_tokens_seen": 18056296,
3067
+ "step": 1720
3068
+ },
3069
+ {
3070
+ "epoch": 0.4442441411279938,
3071
+ "grad_norm": 4.274996116585604,
3072
+ "learning_rate": 5.29162319887768e-05,
3073
+ "loss": 0.3513,
3074
+ "num_input_tokens_seen": 18108904,
3075
+ "step": 1725
3076
+ },
3077
+ {
3078
+ "epoch": 0.4455318053051764,
3079
+ "grad_norm": 2.3776496789610224,
3080
+ "learning_rate": 5.26734550690071e-05,
3081
+ "loss": 0.373,
3082
+ "num_input_tokens_seen": 18160696,
3083
+ "step": 1730
3084
+ },
3085
+ {
3086
+ "epoch": 0.446819469482359,
3087
+ "grad_norm": 5.321325318314331,
3088
+ "learning_rate": 5.243061492163073e-05,
3089
+ "loss": 0.4246,
3090
+ "num_input_tokens_seen": 18213760,
3091
+ "step": 1735
3092
+ },
3093
+ {
3094
+ "epoch": 0.4481071336595416,
3095
+ "grad_norm": 4.658819445189394,
3096
+ "learning_rate": 5.2187717289852955e-05,
3097
+ "loss": 0.3703,
3098
+ "num_input_tokens_seen": 18266424,
3099
+ "step": 1740
3100
+ },
3101
+ {
3102
+ "epoch": 0.4493947978367242,
3103
+ "grad_norm": 4.197790185757161,
3104
+ "learning_rate": 5.1944767918238624e-05,
3105
+ "loss": 0.3763,
3106
+ "num_input_tokens_seen": 18318984,
3107
+ "step": 1745
3108
+ },
3109
+ {
3110
+ "epoch": 0.45068246201390677,
3111
+ "grad_norm": 8.378009104226413,
3112
+ "learning_rate": 5.170177255257618e-05,
3113
+ "loss": 0.3767,
3114
+ "num_input_tokens_seen": 18371928,
3115
+ "step": 1750
3116
+ },
3117
+ {
3118
+ "epoch": 0.45068246201390677,
3119
+ "eval_loss": 0.4234265685081482,
3120
+ "eval_runtime": 38.3269,
3121
+ "eval_samples_per_second": 3.131,
3122
+ "eval_steps_per_second": 0.783,
3123
+ "num_input_tokens_seen": 18371928,
3124
+ "step": 1750
3125
  }
3126
  ],
3127
  "logging_steps": 5,
3128
  "max_steps": 3400,
3129
+ "num_input_tokens_seen": 18371928,
3130
  "num_train_epochs": 1,
3131
  "save_steps": 50,
3132
  "stateful_callbacks": {
 
3141
  "attributes": {}
3142
  }
3143
  },
3144
+ "total_flos": 1212133561860096.0,
3145
  "train_batch_size": 1,
3146
  "trial_name": null,
3147
  "trial_params": null