codeGemma-7bFeed-70 / trainer_state.json
afrias5's picture
Upload 15 files
af70239 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 13.285714285714286,
"eval_steps": 500,
"global_step": 70,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.19047619047619047,
"grad_norm": 11.114676475524902,
"learning_rate": 2e-05,
"loss": 4.9155,
"step": 1
},
{
"epoch": 0.38095238095238093,
"grad_norm": 10.886661529541016,
"learning_rate": 4e-05,
"loss": 5.2012,
"step": 2
},
{
"epoch": 0.5714285714285714,
"grad_norm": 10.0223388671875,
"learning_rate": 6e-05,
"loss": 4.4341,
"step": 3
},
{
"epoch": 0.7619047619047619,
"grad_norm": 8.667222023010254,
"learning_rate": 8e-05,
"loss": 4.6939,
"step": 4
},
{
"epoch": 0.9523809523809523,
"grad_norm": 6.8060832023620605,
"learning_rate": 0.0001,
"loss": 3.9619,
"step": 5
},
{
"epoch": 1.1428571428571428,
"grad_norm": 10.650321006774902,
"learning_rate": 0.00012,
"loss": 6.2246,
"step": 6
},
{
"epoch": 1.3333333333333333,
"grad_norm": 5.177021026611328,
"learning_rate": 0.00014,
"loss": 3.1182,
"step": 7
},
{
"epoch": 1.5238095238095237,
"grad_norm": 4.4564313888549805,
"learning_rate": 0.00016,
"loss": 3.4089,
"step": 8
},
{
"epoch": 1.7142857142857144,
"grad_norm": 3.8821651935577393,
"learning_rate": 0.00018,
"loss": 2.8449,
"step": 9
},
{
"epoch": 1.9047619047619047,
"grad_norm": 4.001584529876709,
"learning_rate": 0.0002,
"loss": 2.541,
"step": 10
},
{
"epoch": 2.0476190476190474,
"grad_norm": 5.516446590423584,
"learning_rate": 0.0001998629534754574,
"loss": 3.0792,
"step": 11
},
{
"epoch": 2.238095238095238,
"grad_norm": 3.4709484577178955,
"learning_rate": 0.00019945218953682734,
"loss": 2.7096,
"step": 12
},
{
"epoch": 2.4285714285714284,
"grad_norm": 2.045745372772217,
"learning_rate": 0.00019876883405951377,
"loss": 1.6279,
"step": 13
},
{
"epoch": 2.619047619047619,
"grad_norm": 2.5490376949310303,
"learning_rate": 0.00019781476007338058,
"loss": 1.9507,
"step": 14
},
{
"epoch": 2.8095238095238093,
"grad_norm": 2.9036448001861572,
"learning_rate": 0.00019659258262890683,
"loss": 1.9104,
"step": 15
},
{
"epoch": 3.0,
"grad_norm": 4.653842449188232,
"learning_rate": 0.00019510565162951537,
"loss": 3.1175,
"step": 16
},
{
"epoch": 3.1904761904761907,
"grad_norm": 2.397826671600342,
"learning_rate": 0.00019335804264972018,
"loss": 1.7851,
"step": 17
},
{
"epoch": 3.380952380952381,
"grad_norm": 2.003856897354126,
"learning_rate": 0.0001913545457642601,
"loss": 1.3112,
"step": 18
},
{
"epoch": 3.571428571428571,
"grad_norm": 1.5474413633346558,
"learning_rate": 0.0001891006524188368,
"loss": 1.4914,
"step": 19
},
{
"epoch": 3.761904761904762,
"grad_norm": 1.5208486318588257,
"learning_rate": 0.00018660254037844388,
"loss": 1.2772,
"step": 20
},
{
"epoch": 3.9523809523809526,
"grad_norm": 1.8079063892364502,
"learning_rate": 0.00018386705679454242,
"loss": 1.3938,
"step": 21
},
{
"epoch": 4.142857142857143,
"grad_norm": 3.1666009426116943,
"learning_rate": 0.00018090169943749476,
"loss": 1.5114,
"step": 22
},
{
"epoch": 4.333333333333333,
"grad_norm": 1.6823382377624512,
"learning_rate": 0.0001777145961456971,
"loss": 1.4207,
"step": 23
},
{
"epoch": 4.523809523809524,
"grad_norm": 1.3504416942596436,
"learning_rate": 0.00017431448254773944,
"loss": 0.9105,
"step": 24
},
{
"epoch": 4.714285714285714,
"grad_norm": 1.3855412006378174,
"learning_rate": 0.00017071067811865476,
"loss": 1.0832,
"step": 25
},
{
"epoch": 4.904761904761905,
"grad_norm": 1.378572940826416,
"learning_rate": 0.00016691306063588583,
"loss": 1.1635,
"step": 26
},
{
"epoch": 5.095238095238095,
"grad_norm": 2.8164830207824707,
"learning_rate": 0.00016293203910498376,
"loss": 1.4419,
"step": 27
},
{
"epoch": 5.285714285714286,
"grad_norm": 1.147581696510315,
"learning_rate": 0.00015877852522924732,
"loss": 0.6387,
"step": 28
},
{
"epoch": 5.476190476190476,
"grad_norm": 1.5747478008270264,
"learning_rate": 0.00015446390350150273,
"loss": 1.0435,
"step": 29
},
{
"epoch": 5.666666666666667,
"grad_norm": 1.3611525297164917,
"learning_rate": 0.00015000000000000001,
"loss": 0.9718,
"step": 30
},
{
"epoch": 5.857142857142857,
"grad_norm": 1.1871294975280762,
"learning_rate": 0.00014539904997395468,
"loss": 0.7409,
"step": 31
},
{
"epoch": 6.0476190476190474,
"grad_norm": 4.008199214935303,
"learning_rate": 0.00014067366430758004,
"loss": 1.6548,
"step": 32
},
{
"epoch": 6.238095238095238,
"grad_norm": 1.111243486404419,
"learning_rate": 0.00013583679495453,
"loss": 0.5545,
"step": 33
},
{
"epoch": 6.428571428571429,
"grad_norm": 1.1197227239608765,
"learning_rate": 0.00013090169943749476,
"loss": 0.5296,
"step": 34
},
{
"epoch": 6.619047619047619,
"grad_norm": 1.1618750095367432,
"learning_rate": 0.00012588190451025207,
"loss": 0.5923,
"step": 35
},
{
"epoch": 6.809523809523809,
"grad_norm": 1.0321106910705566,
"learning_rate": 0.00012079116908177593,
"loss": 0.6033,
"step": 36
},
{
"epoch": 7.0,
"grad_norm": 3.302760601043701,
"learning_rate": 0.0001156434465040231,
"loss": 1.269,
"step": 37
},
{
"epoch": 7.190476190476191,
"grad_norm": 1.3132500648498535,
"learning_rate": 0.00011045284632676536,
"loss": 0.5142,
"step": 38
},
{
"epoch": 7.380952380952381,
"grad_norm": 0.8678253293037415,
"learning_rate": 0.0001052335956242944,
"loss": 0.4287,
"step": 39
},
{
"epoch": 7.571428571428571,
"grad_norm": 1.0537868738174438,
"learning_rate": 0.0001,
"loss": 0.5592,
"step": 40
},
{
"epoch": 7.761904761904762,
"grad_norm": 0.9256157279014587,
"learning_rate": 9.476640437570562e-05,
"loss": 0.3902,
"step": 41
},
{
"epoch": 7.9523809523809526,
"grad_norm": 0.938805341720581,
"learning_rate": 8.954715367323468e-05,
"loss": 0.442,
"step": 42
},
{
"epoch": 8.142857142857142,
"grad_norm": 1.98072350025177,
"learning_rate": 8.435655349597689e-05,
"loss": 0.6278,
"step": 43
},
{
"epoch": 8.333333333333334,
"grad_norm": 0.8120888471603394,
"learning_rate": 7.920883091822408e-05,
"loss": 0.2643,
"step": 44
},
{
"epoch": 8.523809523809524,
"grad_norm": 0.8413763642311096,
"learning_rate": 7.411809548974792e-05,
"loss": 0.3042,
"step": 45
},
{
"epoch": 8.714285714285714,
"grad_norm": 0.7817604541778564,
"learning_rate": 6.909830056250527e-05,
"loss": 0.3137,
"step": 46
},
{
"epoch": 8.904761904761905,
"grad_norm": 0.7885234355926514,
"learning_rate": 6.416320504546997e-05,
"loss": 0.2808,
"step": 47
},
{
"epoch": 9.095238095238095,
"grad_norm": 1.5161521434783936,
"learning_rate": 5.9326335692419995e-05,
"loss": 0.5427,
"step": 48
},
{
"epoch": 9.285714285714286,
"grad_norm": 0.5991838574409485,
"learning_rate": 5.4600950026045326e-05,
"loss": 0.2195,
"step": 49
},
{
"epoch": 9.476190476190476,
"grad_norm": 0.7848747968673706,
"learning_rate": 5.000000000000002e-05,
"loss": 0.2632,
"step": 50
},
{
"epoch": 9.666666666666666,
"grad_norm": 0.6151169538497925,
"learning_rate": 4.5536096498497295e-05,
"loss": 0.2032,
"step": 51
},
{
"epoch": 9.857142857142858,
"grad_norm": 1.1136269569396973,
"learning_rate": 4.12214747707527e-05,
"loss": 0.2819,
"step": 52
},
{
"epoch": 10.095238095238095,
"grad_norm": 0.659062922000885,
"learning_rate": 3.7067960895016275e-05,
"loss": 0.1995,
"step": 53
},
{
"epoch": 10.285714285714286,
"grad_norm": 0.6866421699523926,
"learning_rate": 3.308693936411421e-05,
"loss": 0.18,
"step": 54
},
{
"epoch": 10.476190476190476,
"grad_norm": 0.6459683775901794,
"learning_rate": 2.9289321881345254e-05,
"loss": 0.2022,
"step": 55
},
{
"epoch": 10.666666666666666,
"grad_norm": 0.6483227014541626,
"learning_rate": 2.5685517452260567e-05,
"loss": 0.1876,
"step": 56
},
{
"epoch": 10.857142857142858,
"grad_norm": 0.5956599116325378,
"learning_rate": 2.2285403854302912e-05,
"loss": 0.1969,
"step": 57
},
{
"epoch": 11.047619047619047,
"grad_norm": 2.297029733657837,
"learning_rate": 1.9098300562505266e-05,
"loss": 0.3517,
"step": 58
},
{
"epoch": 11.238095238095237,
"grad_norm": 0.5542204976081848,
"learning_rate": 1.6132943205457606e-05,
"loss": 0.1701,
"step": 59
},
{
"epoch": 11.428571428571429,
"grad_norm": 0.5855220556259155,
"learning_rate": 1.339745962155613e-05,
"loss": 0.1588,
"step": 60
},
{
"epoch": 11.619047619047619,
"grad_norm": 0.6266547441482544,
"learning_rate": 1.0899347581163221e-05,
"loss": 0.1762,
"step": 61
},
{
"epoch": 11.80952380952381,
"grad_norm": 0.5979227423667908,
"learning_rate": 8.645454235739903e-06,
"loss": 0.1604,
"step": 62
},
{
"epoch": 12.0,
"grad_norm": 1.2808066606521606,
"learning_rate": 6.6419573502798374e-06,
"loss": 0.2592,
"step": 63
},
{
"epoch": 12.19047619047619,
"grad_norm": 0.5678392052650452,
"learning_rate": 4.8943483704846475e-06,
"loss": 0.1635,
"step": 64
},
{
"epoch": 12.380952380952381,
"grad_norm": 0.746486246585846,
"learning_rate": 3.40741737109318e-06,
"loss": 0.1914,
"step": 65
},
{
"epoch": 12.571428571428571,
"grad_norm": 0.6837660074234009,
"learning_rate": 2.1852399266194314e-06,
"loss": 0.1632,
"step": 66
},
{
"epoch": 12.761904761904763,
"grad_norm": 0.6320339441299438,
"learning_rate": 1.231165940486234e-06,
"loss": 0.1514,
"step": 67
},
{
"epoch": 12.952380952380953,
"grad_norm": 0.6253515481948853,
"learning_rate": 5.478104631726711e-07,
"loss": 0.1538,
"step": 68
},
{
"epoch": 13.095238095238095,
"grad_norm": 0.8237464427947998,
"learning_rate": 1.3704652454261668e-07,
"loss": 0.1993,
"step": 69
},
{
"epoch": 13.285714285714286,
"grad_norm": 0.6406182646751404,
"learning_rate": 0.0,
"loss": 0.1828,
"step": 70
}
],
"logging_steps": 1,
"max_steps": 70,
"num_input_tokens_seen": 0,
"num_train_epochs": 14,
"save_steps": 5,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 62994788646912.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}