llama3_based_claim_verifier / trainer_state.json
SYX's picture
Upload folder using huggingface_hub
d733135 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 537,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0186219739292365,
"grad_norm": 1.3874553442001343,
"learning_rate": 9.98134328358209e-05,
"loss": 1.8326,
"step": 10
},
{
"epoch": 0.037243947858473,
"grad_norm": 0.7816306352615356,
"learning_rate": 9.94402985074627e-05,
"loss": 1.314,
"step": 20
},
{
"epoch": 0.055865921787709494,
"grad_norm": 0.748516857624054,
"learning_rate": 9.906716417910448e-05,
"loss": 1.2913,
"step": 30
},
{
"epoch": 0.074487895716946,
"grad_norm": 0.7102469205856323,
"learning_rate": 9.869402985074628e-05,
"loss": 1.2041,
"step": 40
},
{
"epoch": 0.0931098696461825,
"grad_norm": 0.7098786234855652,
"learning_rate": 9.832089552238806e-05,
"loss": 1.1642,
"step": 50
},
{
"epoch": 0.11173184357541899,
"grad_norm": 0.6812120676040649,
"learning_rate": 9.794776119402985e-05,
"loss": 1.1717,
"step": 60
},
{
"epoch": 0.1303538175046555,
"grad_norm": 0.6887907981872559,
"learning_rate": 9.757462686567165e-05,
"loss": 1.169,
"step": 70
},
{
"epoch": 0.148975791433892,
"grad_norm": 0.685250461101532,
"learning_rate": 9.720149253731343e-05,
"loss": 1.0856,
"step": 80
},
{
"epoch": 0.16759776536312848,
"grad_norm": 0.7008183002471924,
"learning_rate": 9.682835820895523e-05,
"loss": 1.0955,
"step": 90
},
{
"epoch": 0.186219739292365,
"grad_norm": 0.6370360255241394,
"learning_rate": 9.645522388059703e-05,
"loss": 1.0581,
"step": 100
},
{
"epoch": 0.2048417132216015,
"grad_norm": 0.7338399887084961,
"learning_rate": 9.608208955223881e-05,
"loss": 1.0287,
"step": 110
},
{
"epoch": 0.22346368715083798,
"grad_norm": 0.6552614569664001,
"learning_rate": 9.57089552238806e-05,
"loss": 1.0256,
"step": 120
},
{
"epoch": 0.24208566108007448,
"grad_norm": 0.6847032308578491,
"learning_rate": 9.533582089552238e-05,
"loss": 0.9911,
"step": 130
},
{
"epoch": 0.260707635009311,
"grad_norm": 0.7163254618644714,
"learning_rate": 9.496268656716418e-05,
"loss": 1.0316,
"step": 140
},
{
"epoch": 0.27932960893854747,
"grad_norm": 0.6374409794807434,
"learning_rate": 9.458955223880598e-05,
"loss": 1.0137,
"step": 150
},
{
"epoch": 0.297951582867784,
"grad_norm": 1.027390718460083,
"learning_rate": 9.421641791044776e-05,
"loss": 0.9583,
"step": 160
},
{
"epoch": 0.3165735567970205,
"grad_norm": 0.6466365456581116,
"learning_rate": 9.384328358208956e-05,
"loss": 0.9578,
"step": 170
},
{
"epoch": 0.33519553072625696,
"grad_norm": 0.6674121022224426,
"learning_rate": 9.347014925373135e-05,
"loss": 0.9662,
"step": 180
},
{
"epoch": 0.3538175046554935,
"grad_norm": 0.6532049179077148,
"learning_rate": 9.309701492537313e-05,
"loss": 0.9258,
"step": 190
},
{
"epoch": 0.37243947858473,
"grad_norm": 0.6357593536376953,
"learning_rate": 9.272388059701493e-05,
"loss": 0.8891,
"step": 200
},
{
"epoch": 0.39106145251396646,
"grad_norm": 0.6210165619850159,
"learning_rate": 9.235074626865672e-05,
"loss": 0.8979,
"step": 210
},
{
"epoch": 0.409683426443203,
"grad_norm": 0.6325780749320984,
"learning_rate": 9.197761194029851e-05,
"loss": 0.9028,
"step": 220
},
{
"epoch": 0.42830540037243947,
"grad_norm": 0.6363133788108826,
"learning_rate": 9.16044776119403e-05,
"loss": 0.9283,
"step": 230
},
{
"epoch": 0.44692737430167595,
"grad_norm": 0.5902472138404846,
"learning_rate": 9.12313432835821e-05,
"loss": 0.8592,
"step": 240
},
{
"epoch": 0.4655493482309125,
"grad_norm": 0.6462955474853516,
"learning_rate": 9.08582089552239e-05,
"loss": 0.8595,
"step": 250
},
{
"epoch": 0.48417132216014896,
"grad_norm": 0.6341489553451538,
"learning_rate": 9.048507462686568e-05,
"loss": 0.8372,
"step": 260
},
{
"epoch": 0.5027932960893855,
"grad_norm": 0.6504695415496826,
"learning_rate": 9.011194029850746e-05,
"loss": 0.8458,
"step": 270
},
{
"epoch": 0.521415270018622,
"grad_norm": 0.6562509536743164,
"learning_rate": 8.973880597014925e-05,
"loss": 0.8211,
"step": 280
},
{
"epoch": 0.5400372439478585,
"grad_norm": 0.622184693813324,
"learning_rate": 8.936567164179105e-05,
"loss": 0.8278,
"step": 290
},
{
"epoch": 0.5586592178770949,
"grad_norm": 0.6273349523544312,
"learning_rate": 8.899253731343285e-05,
"loss": 0.8188,
"step": 300
},
{
"epoch": 0.5772811918063314,
"grad_norm": 0.6405043601989746,
"learning_rate": 8.861940298507463e-05,
"loss": 0.8646,
"step": 310
},
{
"epoch": 0.595903165735568,
"grad_norm": 0.6616275310516357,
"learning_rate": 8.824626865671643e-05,
"loss": 0.8606,
"step": 320
},
{
"epoch": 0.6145251396648045,
"grad_norm": 0.6546701192855835,
"learning_rate": 8.787313432835821e-05,
"loss": 0.8271,
"step": 330
},
{
"epoch": 0.633147113594041,
"grad_norm": 0.6446258425712585,
"learning_rate": 8.75e-05,
"loss": 0.805,
"step": 340
},
{
"epoch": 0.6517690875232774,
"grad_norm": 0.6694822311401367,
"learning_rate": 8.71268656716418e-05,
"loss": 0.8237,
"step": 350
},
{
"epoch": 0.6703910614525139,
"grad_norm": 0.5971983075141907,
"learning_rate": 8.675373134328358e-05,
"loss": 0.7784,
"step": 360
},
{
"epoch": 0.6890130353817505,
"grad_norm": 0.5651050209999084,
"learning_rate": 8.638059701492538e-05,
"loss": 0.7345,
"step": 370
},
{
"epoch": 0.707635009310987,
"grad_norm": 0.6789554357528687,
"learning_rate": 8.600746268656717e-05,
"loss": 0.8024,
"step": 380
},
{
"epoch": 0.7262569832402235,
"grad_norm": 0.6168780326843262,
"learning_rate": 8.563432835820896e-05,
"loss": 0.7387,
"step": 390
},
{
"epoch": 0.74487895716946,
"grad_norm": 0.6000941395759583,
"learning_rate": 8.526119402985075e-05,
"loss": 0.7403,
"step": 400
},
{
"epoch": 0.7635009310986964,
"grad_norm": 0.6785574555397034,
"learning_rate": 8.488805970149253e-05,
"loss": 0.7653,
"step": 410
},
{
"epoch": 0.7821229050279329,
"grad_norm": 0.6262904405593872,
"learning_rate": 8.451492537313433e-05,
"loss": 0.7421,
"step": 420
},
{
"epoch": 0.8007448789571695,
"grad_norm": 0.6499343514442444,
"learning_rate": 8.414179104477612e-05,
"loss": 0.6831,
"step": 430
},
{
"epoch": 0.819366852886406,
"grad_norm": 0.5795383453369141,
"learning_rate": 8.376865671641791e-05,
"loss": 0.7189,
"step": 440
},
{
"epoch": 0.8379888268156425,
"grad_norm": 0.6257463097572327,
"learning_rate": 8.339552238805971e-05,
"loss": 0.7052,
"step": 450
},
{
"epoch": 0.8566108007448789,
"grad_norm": 0.6415144801139832,
"learning_rate": 8.30223880597015e-05,
"loss": 0.703,
"step": 460
},
{
"epoch": 0.8752327746741154,
"grad_norm": 0.6050721406936646,
"learning_rate": 8.26492537313433e-05,
"loss": 0.7093,
"step": 470
},
{
"epoch": 0.8938547486033519,
"grad_norm": 0.63200443983078,
"learning_rate": 8.227611940298508e-05,
"loss": 0.711,
"step": 480
},
{
"epoch": 0.9124767225325885,
"grad_norm": 0.7535350918769836,
"learning_rate": 8.190298507462687e-05,
"loss": 0.7101,
"step": 490
},
{
"epoch": 0.931098696461825,
"grad_norm": 0.5603737831115723,
"learning_rate": 8.152985074626866e-05,
"loss": 0.6888,
"step": 500
},
{
"epoch": 0.9497206703910615,
"grad_norm": 0.6010568141937256,
"learning_rate": 8.115671641791045e-05,
"loss": 0.702,
"step": 510
},
{
"epoch": 0.9683426443202979,
"grad_norm": 0.6261228322982788,
"learning_rate": 8.078358208955225e-05,
"loss": 0.6634,
"step": 520
},
{
"epoch": 0.9869646182495344,
"grad_norm": 0.5894924998283386,
"learning_rate": 8.041044776119403e-05,
"loss": 0.6782,
"step": 530
},
{
"epoch": 1.0,
"eval_loss": 0.8664104342460632,
"eval_runtime": 40.1805,
"eval_samples_per_second": 10.005,
"eval_steps_per_second": 1.269,
"step": 537
}
],
"logging_steps": 10,
"max_steps": 2685,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"total_flos": 8.800341942707159e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}