mariagrandury's picture
End of training
1c7d2cb verified
{
"best_metric": 0.3499479191731958,
"best_model_checkpoint": "/content/drive/MyDrive/model_outputs/trial_2/checkpoint-1752",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1752,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.017123287671232876,
"grad_norm": 11.892024993896484,
"learning_rate": 7.592988034631688e-07,
"loss": 2.215,
"step": 10
},
{
"epoch": 0.03424657534246575,
"grad_norm": 11.759550094604492,
"learning_rate": 1.5185976069263376e-06,
"loss": 2.2173,
"step": 20
},
{
"epoch": 0.05136986301369863,
"grad_norm": 11.668200492858887,
"learning_rate": 2.2778964103895065e-06,
"loss": 2.2151,
"step": 30
},
{
"epoch": 0.0684931506849315,
"grad_norm": 10.476677894592285,
"learning_rate": 3.0371952138526753e-06,
"loss": 2.1641,
"step": 40
},
{
"epoch": 0.08561643835616438,
"grad_norm": 9.34625244140625,
"learning_rate": 3.7964940173158444e-06,
"loss": 2.1653,
"step": 50
},
{
"epoch": 0.10273972602739725,
"grad_norm": 9.768696784973145,
"learning_rate": 4.555792820779013e-06,
"loss": 2.1354,
"step": 60
},
{
"epoch": 0.11986301369863013,
"grad_norm": 8.377198219299316,
"learning_rate": 5.315091624242182e-06,
"loss": 2.1244,
"step": 70
},
{
"epoch": 0.136986301369863,
"grad_norm": 11.591525077819824,
"learning_rate": 6.0743904277053505e-06,
"loss": 2.0764,
"step": 80
},
{
"epoch": 0.1541095890410959,
"grad_norm": 9.914732933044434,
"learning_rate": 6.83368923116852e-06,
"loss": 2.1816,
"step": 90
},
{
"epoch": 0.17123287671232876,
"grad_norm": 11.065092086791992,
"learning_rate": 7.592988034631689e-06,
"loss": 2.0771,
"step": 100
},
{
"epoch": 0.18835616438356165,
"grad_norm": 11.352612495422363,
"learning_rate": 8.352286838094857e-06,
"loss": 2.0857,
"step": 110
},
{
"epoch": 0.2054794520547945,
"grad_norm": 11.720881462097168,
"learning_rate": 9.111585641558026e-06,
"loss": 2.0766,
"step": 120
},
{
"epoch": 0.2226027397260274,
"grad_norm": 10.508756637573242,
"learning_rate": 9.870884445021195e-06,
"loss": 2.1546,
"step": 130
},
{
"epoch": 0.23972602739726026,
"grad_norm": 11.728009223937988,
"learning_rate": 1.0630183248484364e-05,
"loss": 2.0462,
"step": 140
},
{
"epoch": 0.2568493150684932,
"grad_norm": 9.81351089477539,
"learning_rate": 1.1389482051947532e-05,
"loss": 2.0667,
"step": 150
},
{
"epoch": 0.273972602739726,
"grad_norm": 8.76012134552002,
"learning_rate": 1.2148780855410701e-05,
"loss": 2.0924,
"step": 160
},
{
"epoch": 0.2910958904109589,
"grad_norm": 9.990751266479492,
"learning_rate": 1.290807965887387e-05,
"loss": 2.0672,
"step": 170
},
{
"epoch": 0.3082191780821918,
"grad_norm": 10.576468467712402,
"learning_rate": 1.366737846233704e-05,
"loss": 2.0109,
"step": 180
},
{
"epoch": 0.3253424657534247,
"grad_norm": 9.048011779785156,
"learning_rate": 1.4426677265800208e-05,
"loss": 2.0392,
"step": 190
},
{
"epoch": 0.3424657534246575,
"grad_norm": 11.787674903869629,
"learning_rate": 1.5185976069263378e-05,
"loss": 2.034,
"step": 200
},
{
"epoch": 0.3595890410958904,
"grad_norm": 10.221675872802734,
"learning_rate": 1.5945274872726547e-05,
"loss": 1.9762,
"step": 210
},
{
"epoch": 0.3767123287671233,
"grad_norm": 9.597102165222168,
"learning_rate": 1.6704573676189714e-05,
"loss": 2.0765,
"step": 220
},
{
"epoch": 0.3938356164383562,
"grad_norm": 11.017401695251465,
"learning_rate": 1.7463872479652885e-05,
"loss": 2.0997,
"step": 230
},
{
"epoch": 0.410958904109589,
"grad_norm": 10.641432762145996,
"learning_rate": 1.8223171283116052e-05,
"loss": 2.0985,
"step": 240
},
{
"epoch": 0.4280821917808219,
"grad_norm": 8.237184524536133,
"learning_rate": 1.8982470086579223e-05,
"loss": 2.056,
"step": 250
},
{
"epoch": 0.4452054794520548,
"grad_norm": 8.095678329467773,
"learning_rate": 1.974176889004239e-05,
"loss": 2.0807,
"step": 260
},
{
"epoch": 0.4623287671232877,
"grad_norm": 9.02453899383545,
"learning_rate": 2.050106769350556e-05,
"loss": 2.102,
"step": 270
},
{
"epoch": 0.4794520547945205,
"grad_norm": 9.920637130737305,
"learning_rate": 2.126036649696873e-05,
"loss": 1.9904,
"step": 280
},
{
"epoch": 0.4965753424657534,
"grad_norm": 10.004680633544922,
"learning_rate": 2.2019665300431896e-05,
"loss": 2.0086,
"step": 290
},
{
"epoch": 0.5136986301369864,
"grad_norm": 8.83816909790039,
"learning_rate": 2.2140715952718072e-05,
"loss": 1.9748,
"step": 300
},
{
"epoch": 0.5308219178082192,
"grad_norm": 8.490972518920898,
"learning_rate": 2.1988231683071254e-05,
"loss": 2.0026,
"step": 310
},
{
"epoch": 0.547945205479452,
"grad_norm": 9.475632667541504,
"learning_rate": 2.1835747413424435e-05,
"loss": 2.0512,
"step": 320
},
{
"epoch": 0.565068493150685,
"grad_norm": 9.864670753479004,
"learning_rate": 2.1683263143777617e-05,
"loss": 2.0304,
"step": 330
},
{
"epoch": 0.5821917808219178,
"grad_norm": 10.569999694824219,
"learning_rate": 2.15307788741308e-05,
"loss": 1.9682,
"step": 340
},
{
"epoch": 0.5993150684931506,
"grad_norm": 9.928328514099121,
"learning_rate": 2.137829460448398e-05,
"loss": 2.0027,
"step": 350
},
{
"epoch": 0.6164383561643836,
"grad_norm": 8.040825843811035,
"learning_rate": 2.1225810334837162e-05,
"loss": 1.9177,
"step": 360
},
{
"epoch": 0.6335616438356164,
"grad_norm": 10.714200019836426,
"learning_rate": 2.1073326065190344e-05,
"loss": 2.0449,
"step": 370
},
{
"epoch": 0.6506849315068494,
"grad_norm": 8.81334114074707,
"learning_rate": 2.0920841795543526e-05,
"loss": 2.0512,
"step": 380
},
{
"epoch": 0.6678082191780822,
"grad_norm": 10.270297050476074,
"learning_rate": 2.0768357525896707e-05,
"loss": 1.9713,
"step": 390
},
{
"epoch": 0.684931506849315,
"grad_norm": 10.366915702819824,
"learning_rate": 2.061587325624989e-05,
"loss": 1.9791,
"step": 400
},
{
"epoch": 0.702054794520548,
"grad_norm": 10.001801490783691,
"learning_rate": 2.0463388986603067e-05,
"loss": 1.9183,
"step": 410
},
{
"epoch": 0.7191780821917808,
"grad_norm": 11.41006088256836,
"learning_rate": 2.031090471695625e-05,
"loss": 1.9395,
"step": 420
},
{
"epoch": 0.7363013698630136,
"grad_norm": 9.574875831604004,
"learning_rate": 2.015842044730943e-05,
"loss": 1.9356,
"step": 430
},
{
"epoch": 0.7534246575342466,
"grad_norm": 11.301505088806152,
"learning_rate": 2.0005936177662612e-05,
"loss": 2.059,
"step": 440
},
{
"epoch": 0.7705479452054794,
"grad_norm": 11.491235733032227,
"learning_rate": 1.9853451908015794e-05,
"loss": 2.0022,
"step": 450
},
{
"epoch": 0.7876712328767124,
"grad_norm": 9.770740509033203,
"learning_rate": 1.9700967638368976e-05,
"loss": 1.9812,
"step": 460
},
{
"epoch": 0.8047945205479452,
"grad_norm": 9.801970481872559,
"learning_rate": 1.9548483368722154e-05,
"loss": 2.0287,
"step": 470
},
{
"epoch": 0.821917808219178,
"grad_norm": 10.885910034179688,
"learning_rate": 1.9395999099075336e-05,
"loss": 2.0574,
"step": 480
},
{
"epoch": 0.839041095890411,
"grad_norm": 10.215595245361328,
"learning_rate": 1.9243514829428517e-05,
"loss": 2.0379,
"step": 490
},
{
"epoch": 0.8561643835616438,
"grad_norm": 10.391200065612793,
"learning_rate": 1.9091030559781702e-05,
"loss": 1.9701,
"step": 500
},
{
"epoch": 0.8732876712328768,
"grad_norm": 10.24344253540039,
"learning_rate": 1.893854629013488e-05,
"loss": 1.9256,
"step": 510
},
{
"epoch": 0.8904109589041096,
"grad_norm": 11.228612899780273,
"learning_rate": 1.8786062020488062e-05,
"loss": 2.029,
"step": 520
},
{
"epoch": 0.9075342465753424,
"grad_norm": 13.008833885192871,
"learning_rate": 1.8633577750841244e-05,
"loss": 1.9837,
"step": 530
},
{
"epoch": 0.9246575342465754,
"grad_norm": 10.637398719787598,
"learning_rate": 1.8481093481194426e-05,
"loss": 1.8445,
"step": 540
},
{
"epoch": 0.9417808219178082,
"grad_norm": 13.296565055847168,
"learning_rate": 1.8328609211547607e-05,
"loss": 1.9631,
"step": 550
},
{
"epoch": 0.958904109589041,
"grad_norm": 11.811957359313965,
"learning_rate": 1.817612494190079e-05,
"loss": 1.961,
"step": 560
},
{
"epoch": 0.976027397260274,
"grad_norm": 11.553853988647461,
"learning_rate": 1.8023640672253967e-05,
"loss": 1.9842,
"step": 570
},
{
"epoch": 0.9931506849315068,
"grad_norm": 12.016478538513184,
"learning_rate": 1.787115640260715e-05,
"loss": 1.9183,
"step": 580
},
{
"epoch": 1.0,
"eval_classification_report": {
"accuracy": 0.273,
"ar": {
"f1-score": 0.22297297297297297,
"precision": 0.36666666666666664,
"recall": 0.16019417475728157,
"support": 206.0
},
"cl": {
"f1-score": 0.06876790830945559,
"precision": 0.2033898305084746,
"recall": 0.041379310344827586,
"support": 290.0
},
"co": {
"f1-score": 0.3118081180811808,
"precision": 0.21311475409836064,
"recall": 0.5807560137457045,
"support": 291.0
},
"es": {
"f1-score": 0.33212996389891697,
"precision": 0.33454545454545453,
"recall": 0.32974910394265233,
"support": 279.0
},
"macro avg": {
"f1-score": 0.2506678702739225,
"precision": 0.29977219111270315,
"recall": 0.2607198556236934,
"support": 2000.0
},
"mx": {
"f1-score": 0.26535626535626533,
"precision": 0.46551724137931033,
"recall": 0.18556701030927836,
"support": 291.0
},
"pe": {
"f1-score": 0.21484375,
"precision": 0.248868778280543,
"recall": 0.18900343642611683,
"support": 291.0
},
"pr": {
"f1-score": 0.574585635359116,
"precision": 0.65,
"recall": 0.5148514851485149,
"support": 101.0
},
"uy": {
"f1-score": 0.26554621848739496,
"precision": 0.21584699453551912,
"recall": 0.34497816593886466,
"support": 229.0
},
"ve": {
"f1-score": 0.0,
"precision": 0.0,
"recall": 0.0,
"support": 22.0
},
"weighted avg": {
"f1-score": 0.2539284929025767,
"precision": 0.30641812645562344,
"recall": 0.273,
"support": 2000.0
}
},
"eval_f1": 0.2506678702739225,
"eval_loss": 1.8817169666290283,
"eval_runtime": 2.8277,
"eval_samples_per_second": 707.278,
"eval_steps_per_second": 88.41,
"step": 584
},
{
"epoch": 1.0102739726027397,
"grad_norm": 11.690625190734863,
"learning_rate": 1.771867213296033e-05,
"loss": 1.8131,
"step": 590
},
{
"epoch": 1.0273972602739727,
"grad_norm": 12.48835563659668,
"learning_rate": 1.7566187863313512e-05,
"loss": 1.7617,
"step": 600
},
{
"epoch": 1.0445205479452055,
"grad_norm": 15.49386978149414,
"learning_rate": 1.7413703593666694e-05,
"loss": 1.7815,
"step": 610
},
{
"epoch": 1.0616438356164384,
"grad_norm": 17.888931274414062,
"learning_rate": 1.7261219324019876e-05,
"loss": 1.7337,
"step": 620
},
{
"epoch": 1.0787671232876712,
"grad_norm": 15.020529747009277,
"learning_rate": 1.7108735054373058e-05,
"loss": 1.6467,
"step": 630
},
{
"epoch": 1.095890410958904,
"grad_norm": 13.33208179473877,
"learning_rate": 1.695625078472624e-05,
"loss": 1.8363,
"step": 640
},
{
"epoch": 1.1130136986301369,
"grad_norm": 18.24432373046875,
"learning_rate": 1.680376651507942e-05,
"loss": 1.9009,
"step": 650
},
{
"epoch": 1.13013698630137,
"grad_norm": 13.63119125366211,
"learning_rate": 1.6651282245432603e-05,
"loss": 1.7742,
"step": 660
},
{
"epoch": 1.1472602739726028,
"grad_norm": 15.010072708129883,
"learning_rate": 1.6498797975785784e-05,
"loss": 1.6258,
"step": 670
},
{
"epoch": 1.1643835616438356,
"grad_norm": 11.895198822021484,
"learning_rate": 1.6346313706138963e-05,
"loss": 1.8729,
"step": 680
},
{
"epoch": 1.1815068493150684,
"grad_norm": 14.77038288116455,
"learning_rate": 1.6193829436492144e-05,
"loss": 1.7128,
"step": 690
},
{
"epoch": 1.1986301369863013,
"grad_norm": 11.579257011413574,
"learning_rate": 1.6041345166845326e-05,
"loss": 1.8186,
"step": 700
},
{
"epoch": 1.2157534246575343,
"grad_norm": 13.447906494140625,
"learning_rate": 1.5888860897198508e-05,
"loss": 1.7262,
"step": 710
},
{
"epoch": 1.2328767123287672,
"grad_norm": 16.56707763671875,
"learning_rate": 1.573637662755169e-05,
"loss": 1.8165,
"step": 720
},
{
"epoch": 1.25,
"grad_norm": 12.793381690979004,
"learning_rate": 1.558389235790487e-05,
"loss": 1.7352,
"step": 730
},
{
"epoch": 1.2671232876712328,
"grad_norm": 14.821900367736816,
"learning_rate": 1.543140808825805e-05,
"loss": 1.8093,
"step": 740
},
{
"epoch": 1.2842465753424657,
"grad_norm": 13.674068450927734,
"learning_rate": 1.527892381861123e-05,
"loss": 1.7997,
"step": 750
},
{
"epoch": 1.3013698630136985,
"grad_norm": 21.423152923583984,
"learning_rate": 1.5126439548964413e-05,
"loss": 1.8501,
"step": 760
},
{
"epoch": 1.3184931506849316,
"grad_norm": 13.842451095581055,
"learning_rate": 1.4973955279317594e-05,
"loss": 1.6396,
"step": 770
},
{
"epoch": 1.3356164383561644,
"grad_norm": 12.872172355651855,
"learning_rate": 1.4821471009670778e-05,
"loss": 1.7377,
"step": 780
},
{
"epoch": 1.3527397260273972,
"grad_norm": 18.74566650390625,
"learning_rate": 1.466898674002396e-05,
"loss": 1.7576,
"step": 790
},
{
"epoch": 1.36986301369863,
"grad_norm": 15.825615882873535,
"learning_rate": 1.451650247037714e-05,
"loss": 1.738,
"step": 800
},
{
"epoch": 1.3869863013698631,
"grad_norm": 13.375591278076172,
"learning_rate": 1.4364018200730321e-05,
"loss": 1.6459,
"step": 810
},
{
"epoch": 1.404109589041096,
"grad_norm": 17.51374053955078,
"learning_rate": 1.4211533931083503e-05,
"loss": 1.7292,
"step": 820
},
{
"epoch": 1.4212328767123288,
"grad_norm": 15.895259857177734,
"learning_rate": 1.4059049661436683e-05,
"loss": 1.8173,
"step": 830
},
{
"epoch": 1.4383561643835616,
"grad_norm": 24.653060913085938,
"learning_rate": 1.3906565391789864e-05,
"loss": 1.7303,
"step": 840
},
{
"epoch": 1.4554794520547945,
"grad_norm": 19.205734252929688,
"learning_rate": 1.3754081122143046e-05,
"loss": 1.5581,
"step": 850
},
{
"epoch": 1.4726027397260273,
"grad_norm": 13.14504337310791,
"learning_rate": 1.3601596852496226e-05,
"loss": 1.7028,
"step": 860
},
{
"epoch": 1.4897260273972603,
"grad_norm": 19.409351348876953,
"learning_rate": 1.3449112582849408e-05,
"loss": 1.7864,
"step": 870
},
{
"epoch": 1.5068493150684932,
"grad_norm": 20.153881072998047,
"learning_rate": 1.329662831320259e-05,
"loss": 1.5898,
"step": 880
},
{
"epoch": 1.523972602739726,
"grad_norm": 13.584912300109863,
"learning_rate": 1.314414404355577e-05,
"loss": 1.6459,
"step": 890
},
{
"epoch": 1.541095890410959,
"grad_norm": 20.62996482849121,
"learning_rate": 1.2991659773908951e-05,
"loss": 1.8207,
"step": 900
},
{
"epoch": 1.558219178082192,
"grad_norm": 16.82609748840332,
"learning_rate": 1.2839175504262133e-05,
"loss": 1.6256,
"step": 910
},
{
"epoch": 1.5753424657534247,
"grad_norm": 19.674978256225586,
"learning_rate": 1.2686691234615316e-05,
"loss": 1.6743,
"step": 920
},
{
"epoch": 1.5924657534246576,
"grad_norm": 23.367334365844727,
"learning_rate": 1.2534206964968498e-05,
"loss": 1.602,
"step": 930
},
{
"epoch": 1.6095890410958904,
"grad_norm": 19.46713638305664,
"learning_rate": 1.2381722695321678e-05,
"loss": 1.6916,
"step": 940
},
{
"epoch": 1.6267123287671232,
"grad_norm": 24.35736083984375,
"learning_rate": 1.222923842567486e-05,
"loss": 1.6058,
"step": 950
},
{
"epoch": 1.643835616438356,
"grad_norm": 24.921846389770508,
"learning_rate": 1.2076754156028041e-05,
"loss": 1.7183,
"step": 960
},
{
"epoch": 1.660958904109589,
"grad_norm": 15.05435848236084,
"learning_rate": 1.1924269886381221e-05,
"loss": 1.6577,
"step": 970
},
{
"epoch": 1.678082191780822,
"grad_norm": 15.788996696472168,
"learning_rate": 1.1771785616734403e-05,
"loss": 1.6803,
"step": 980
},
{
"epoch": 1.6952054794520548,
"grad_norm": 18.309133529663086,
"learning_rate": 1.1619301347087585e-05,
"loss": 1.5949,
"step": 990
},
{
"epoch": 1.7123287671232876,
"grad_norm": 20.53139877319336,
"learning_rate": 1.1466817077440765e-05,
"loss": 1.6097,
"step": 1000
},
{
"epoch": 1.7294520547945207,
"grad_norm": 23.069534301757812,
"learning_rate": 1.1314332807793946e-05,
"loss": 1.6068,
"step": 1010
},
{
"epoch": 1.7465753424657535,
"grad_norm": 22.394153594970703,
"learning_rate": 1.1161848538147128e-05,
"loss": 1.6641,
"step": 1020
},
{
"epoch": 1.7636986301369864,
"grad_norm": 20.755950927734375,
"learning_rate": 1.100936426850031e-05,
"loss": 1.4966,
"step": 1030
},
{
"epoch": 1.7808219178082192,
"grad_norm": 16.427976608276367,
"learning_rate": 1.0856879998853491e-05,
"loss": 1.7212,
"step": 1040
},
{
"epoch": 1.797945205479452,
"grad_norm": 33.8179817199707,
"learning_rate": 1.0704395729206671e-05,
"loss": 1.7811,
"step": 1050
},
{
"epoch": 1.8150684931506849,
"grad_norm": 20.979110717773438,
"learning_rate": 1.0551911459559853e-05,
"loss": 1.583,
"step": 1060
},
{
"epoch": 1.8321917808219177,
"grad_norm": 17.085895538330078,
"learning_rate": 1.0399427189913035e-05,
"loss": 1.7014,
"step": 1070
},
{
"epoch": 1.8493150684931505,
"grad_norm": 18.183605194091797,
"learning_rate": 1.0246942920266215e-05,
"loss": 1.5758,
"step": 1080
},
{
"epoch": 1.8664383561643836,
"grad_norm": 16.887876510620117,
"learning_rate": 1.0094458650619396e-05,
"loss": 1.4666,
"step": 1090
},
{
"epoch": 1.8835616438356164,
"grad_norm": 18.789949417114258,
"learning_rate": 9.94197438097258e-06,
"loss": 1.7645,
"step": 1100
},
{
"epoch": 1.9006849315068495,
"grad_norm": 18.971622467041016,
"learning_rate": 9.78949011132576e-06,
"loss": 1.6875,
"step": 1110
},
{
"epoch": 1.9178082191780823,
"grad_norm": 24.858366012573242,
"learning_rate": 9.637005841678942e-06,
"loss": 1.6846,
"step": 1120
},
{
"epoch": 1.9349315068493151,
"grad_norm": 13.931944847106934,
"learning_rate": 9.484521572032123e-06,
"loss": 1.6025,
"step": 1130
},
{
"epoch": 1.952054794520548,
"grad_norm": 17.908315658569336,
"learning_rate": 9.332037302385303e-06,
"loss": 1.4988,
"step": 1140
},
{
"epoch": 1.9691780821917808,
"grad_norm": 16.93858528137207,
"learning_rate": 9.179553032738485e-06,
"loss": 1.7157,
"step": 1150
},
{
"epoch": 1.9863013698630136,
"grad_norm": 19.648006439208984,
"learning_rate": 9.027068763091667e-06,
"loss": 1.5337,
"step": 1160
},
{
"epoch": 2.0,
"eval_classification_report": {
"accuracy": 0.3535,
"ar": {
"f1-score": 0.31309904153354634,
"precision": 0.45794392523364486,
"recall": 0.23786407766990292,
"support": 206.0
},
"cl": {
"f1-score": 0.23689320388349513,
"precision": 0.27111111111111114,
"recall": 0.2103448275862069,
"support": 290.0
},
"co": {
"f1-score": 0.36268068331143233,
"precision": 0.2936170212765957,
"recall": 0.4742268041237113,
"support": 291.0
},
"es": {
"f1-score": 0.38976377952755903,
"precision": 0.43231441048034935,
"recall": 0.3548387096774194,
"support": 279.0
},
"macro avg": {
"f1-score": 0.3353596555098448,
"precision": 0.3519727641328481,
"recall": 0.3389548790003148,
"support": 2000.0
},
"mx": {
"f1-score": 0.36254980079681276,
"precision": 0.4312796208530806,
"recall": 0.3127147766323024,
"support": 291.0
},
"pe": {
"f1-score": 0.31550068587105623,
"precision": 0.2625570776255708,
"recall": 0.3951890034364261,
"support": 291.0
},
"pr": {
"f1-score": 0.6543778801843319,
"precision": 0.6120689655172413,
"recall": 0.7029702970297029,
"support": 101.0
},
"uy": {
"f1-score": 0.3833718244803695,
"precision": 0.4068627450980392,
"recall": 0.3624454148471616,
"support": 229.0
},
"ve": {
"f1-score": 0.0,
"precision": 0.0,
"recall": 0.0,
"support": 22.0
},
"weighted avg": {
"f1-score": 0.34933930516945594,
"precision": 0.3679569789689199,
"recall": 0.3535,
"support": 2000.0
}
},
"eval_f1": 0.3353596555098448,
"eval_loss": 1.793221116065979,
"eval_runtime": 2.7925,
"eval_samples_per_second": 716.208,
"eval_steps_per_second": 89.526,
"step": 1168
},
{
"epoch": 2.0034246575342465,
"grad_norm": 20.4486083984375,
"learning_rate": 8.874584493444848e-06,
"loss": 1.5645,
"step": 1170
},
{
"epoch": 2.0205479452054793,
"grad_norm": 17.64771270751953,
"learning_rate": 8.72210022379803e-06,
"loss": 1.3974,
"step": 1180
},
{
"epoch": 2.037671232876712,
"grad_norm": 13.944244384765625,
"learning_rate": 8.56961595415121e-06,
"loss": 1.2199,
"step": 1190
},
{
"epoch": 2.0547945205479454,
"grad_norm": 16.506717681884766,
"learning_rate": 8.417131684504392e-06,
"loss": 1.2971,
"step": 1200
},
{
"epoch": 2.0719178082191783,
"grad_norm": 22.669170379638672,
"learning_rate": 8.264647414857573e-06,
"loss": 1.2465,
"step": 1210
},
{
"epoch": 2.089041095890411,
"grad_norm": 20.12689971923828,
"learning_rate": 8.112163145210753e-06,
"loss": 1.1901,
"step": 1220
},
{
"epoch": 2.106164383561644,
"grad_norm": 16.720157623291016,
"learning_rate": 7.959678875563935e-06,
"loss": 1.1476,
"step": 1230
},
{
"epoch": 2.1232876712328768,
"grad_norm": 13.676715850830078,
"learning_rate": 7.807194605917117e-06,
"loss": 1.2803,
"step": 1240
},
{
"epoch": 2.1404109589041096,
"grad_norm": 16.257658004760742,
"learning_rate": 7.654710336270298e-06,
"loss": 1.148,
"step": 1250
},
{
"epoch": 2.1575342465753424,
"grad_norm": 20.69976234436035,
"learning_rate": 7.50222606662348e-06,
"loss": 1.3063,
"step": 1260
},
{
"epoch": 2.1746575342465753,
"grad_norm": 21.628828048706055,
"learning_rate": 7.349741796976661e-06,
"loss": 1.3074,
"step": 1270
},
{
"epoch": 2.191780821917808,
"grad_norm": 18.611053466796875,
"learning_rate": 7.197257527329842e-06,
"loss": 1.2918,
"step": 1280
},
{
"epoch": 2.208904109589041,
"grad_norm": 19.288227081298828,
"learning_rate": 7.044773257683023e-06,
"loss": 1.1346,
"step": 1290
},
{
"epoch": 2.2260273972602738,
"grad_norm": 27.32809066772461,
"learning_rate": 6.892288988036204e-06,
"loss": 1.4171,
"step": 1300
},
{
"epoch": 2.243150684931507,
"grad_norm": 26.971960067749023,
"learning_rate": 6.739804718389387e-06,
"loss": 1.0316,
"step": 1310
},
{
"epoch": 2.26027397260274,
"grad_norm": 30.222131729125977,
"learning_rate": 6.587320448742568e-06,
"loss": 1.1692,
"step": 1320
},
{
"epoch": 2.2773972602739727,
"grad_norm": 20.51207160949707,
"learning_rate": 6.434836179095749e-06,
"loss": 1.0899,
"step": 1330
},
{
"epoch": 2.2945205479452055,
"grad_norm": 19.466644287109375,
"learning_rate": 6.28235190944893e-06,
"loss": 1.2698,
"step": 1340
},
{
"epoch": 2.3116438356164384,
"grad_norm": 21.64017677307129,
"learning_rate": 6.129867639802111e-06,
"loss": 1.244,
"step": 1350
},
{
"epoch": 2.328767123287671,
"grad_norm": 20.915922164916992,
"learning_rate": 5.977383370155293e-06,
"loss": 1.0774,
"step": 1360
},
{
"epoch": 2.345890410958904,
"grad_norm": 24.89851188659668,
"learning_rate": 5.8248991005084735e-06,
"loss": 1.2141,
"step": 1370
},
{
"epoch": 2.363013698630137,
"grad_norm": 24.925865173339844,
"learning_rate": 5.672414830861656e-06,
"loss": 1.2304,
"step": 1380
},
{
"epoch": 2.3801369863013697,
"grad_norm": 27.173961639404297,
"learning_rate": 5.519930561214836e-06,
"loss": 1.1165,
"step": 1390
},
{
"epoch": 2.3972602739726026,
"grad_norm": 19.15679931640625,
"learning_rate": 5.367446291568018e-06,
"loss": 1.3572,
"step": 1400
},
{
"epoch": 2.4143835616438354,
"grad_norm": 20.516983032226562,
"learning_rate": 5.214962021921199e-06,
"loss": 1.2764,
"step": 1410
},
{
"epoch": 2.4315068493150687,
"grad_norm": 19.259366989135742,
"learning_rate": 5.06247775227438e-06,
"loss": 1.2762,
"step": 1420
},
{
"epoch": 2.4486301369863015,
"grad_norm": 25.636960983276367,
"learning_rate": 4.909993482627562e-06,
"loss": 1.1584,
"step": 1430
},
{
"epoch": 2.4657534246575343,
"grad_norm": 17.860868453979492,
"learning_rate": 4.757509212980744e-06,
"loss": 0.9371,
"step": 1440
},
{
"epoch": 2.482876712328767,
"grad_norm": 31.28371238708496,
"learning_rate": 4.6050249433339244e-06,
"loss": 1.1262,
"step": 1450
},
{
"epoch": 2.5,
"grad_norm": 19.444276809692383,
"learning_rate": 4.452540673687105e-06,
"loss": 1.2705,
"step": 1460
},
{
"epoch": 2.517123287671233,
"grad_norm": 16.713764190673828,
"learning_rate": 4.300056404040287e-06,
"loss": 1.1658,
"step": 1470
},
{
"epoch": 2.5342465753424657,
"grad_norm": 24.59725570678711,
"learning_rate": 4.147572134393469e-06,
"loss": 1.303,
"step": 1480
},
{
"epoch": 2.5513698630136985,
"grad_norm": 28.553564071655273,
"learning_rate": 3.9950878647466495e-06,
"loss": 1.2434,
"step": 1490
},
{
"epoch": 2.5684931506849313,
"grad_norm": 25.321197509765625,
"learning_rate": 3.84260359509983e-06,
"loss": 1.2506,
"step": 1500
},
{
"epoch": 2.5856164383561646,
"grad_norm": 29.10701560974121,
"learning_rate": 3.6901193254530124e-06,
"loss": 1.3008,
"step": 1510
},
{
"epoch": 2.602739726027397,
"grad_norm": 23.615144729614258,
"learning_rate": 3.5376350558061937e-06,
"loss": 1.0486,
"step": 1520
},
{
"epoch": 2.6198630136986303,
"grad_norm": 28.57093620300293,
"learning_rate": 3.3851507861593745e-06,
"loss": 1.2169,
"step": 1530
},
{
"epoch": 2.636986301369863,
"grad_norm": 18.41623306274414,
"learning_rate": 3.2326665165125566e-06,
"loss": 1.2028,
"step": 1540
},
{
"epoch": 2.654109589041096,
"grad_norm": 23.891878128051758,
"learning_rate": 3.0801822468657375e-06,
"loss": 1.2732,
"step": 1550
},
{
"epoch": 2.671232876712329,
"grad_norm": 23.33771514892578,
"learning_rate": 2.9276979772189187e-06,
"loss": 1.2175,
"step": 1560
},
{
"epoch": 2.6883561643835616,
"grad_norm": 17.881248474121094,
"learning_rate": 2.7752137075721e-06,
"loss": 1.2555,
"step": 1570
},
{
"epoch": 2.7054794520547945,
"grad_norm": 19.803930282592773,
"learning_rate": 2.6227294379252813e-06,
"loss": 1.2051,
"step": 1580
},
{
"epoch": 2.7226027397260273,
"grad_norm": 22.89762306213379,
"learning_rate": 2.470245168278463e-06,
"loss": 1.1713,
"step": 1590
},
{
"epoch": 2.73972602739726,
"grad_norm": 16.94407844543457,
"learning_rate": 2.3177608986316442e-06,
"loss": 1.1175,
"step": 1600
},
{
"epoch": 2.756849315068493,
"grad_norm": 23.772077560424805,
"learning_rate": 2.1652766289848255e-06,
"loss": 1.1266,
"step": 1610
},
{
"epoch": 2.7739726027397262,
"grad_norm": 38.49664306640625,
"learning_rate": 2.0127923593380067e-06,
"loss": 1.3238,
"step": 1620
},
{
"epoch": 2.791095890410959,
"grad_norm": 27.785099029541016,
"learning_rate": 1.860308089691188e-06,
"loss": 1.1356,
"step": 1630
},
{
"epoch": 2.808219178082192,
"grad_norm": 34.19268798828125,
"learning_rate": 1.7078238200443693e-06,
"loss": 1.3087,
"step": 1640
},
{
"epoch": 2.8253424657534247,
"grad_norm": 24.389307022094727,
"learning_rate": 1.5553395503975505e-06,
"loss": 1.3934,
"step": 1650
},
{
"epoch": 2.8424657534246576,
"grad_norm": 41.80609130859375,
"learning_rate": 1.402855280750732e-06,
"loss": 1.2443,
"step": 1660
},
{
"epoch": 2.8595890410958904,
"grad_norm": 25.80476188659668,
"learning_rate": 1.2503710111039133e-06,
"loss": 1.3135,
"step": 1670
},
{
"epoch": 2.8767123287671232,
"grad_norm": 22.425390243530273,
"learning_rate": 1.0978867414570945e-06,
"loss": 1.0686,
"step": 1680
},
{
"epoch": 2.893835616438356,
"grad_norm": 35.875511169433594,
"learning_rate": 9.454024718102758e-07,
"loss": 1.2933,
"step": 1690
},
{
"epoch": 2.910958904109589,
"grad_norm": 43.623722076416016,
"learning_rate": 7.929182021634573e-07,
"loss": 1.2489,
"step": 1700
},
{
"epoch": 2.928082191780822,
"grad_norm": 29.01580047607422,
"learning_rate": 6.404339325166384e-07,
"loss": 1.2077,
"step": 1710
},
{
"epoch": 2.9452054794520546,
"grad_norm": 19.448177337646484,
"learning_rate": 4.879496628698198e-07,
"loss": 1.0898,
"step": 1720
},
{
"epoch": 2.962328767123288,
"grad_norm": 29.721946716308594,
"learning_rate": 3.354653932230011e-07,
"loss": 1.1101,
"step": 1730
},
{
"epoch": 2.9794520547945207,
"grad_norm": 23.694095611572266,
"learning_rate": 1.8298112357618242e-07,
"loss": 1.0485,
"step": 1740
},
{
"epoch": 2.9965753424657535,
"grad_norm": 27.28858757019043,
"learning_rate": 3.049685392936374e-08,
"loss": 1.1386,
"step": 1750
},
{
"epoch": 3.0,
"eval_classification_report": {
"accuracy": 0.368,
"ar": {
"f1-score": 0.34759358288770054,
"precision": 0.3869047619047619,
"recall": 0.3155339805825243,
"support": 206.0
},
"cl": {
"f1-score": 0.23636363636363636,
"precision": 0.25,
"recall": 0.22413793103448276,
"support": 290.0
},
"co": {
"f1-score": 0.38351822503961963,
"precision": 0.3558823529411765,
"recall": 0.41580756013745707,
"support": 291.0
},
"es": {
"f1-score": 0.39919354838709675,
"precision": 0.45622119815668205,
"recall": 0.3548387096774194,
"support": 279.0
},
"macro avg": {
"f1-score": 0.3499479191731958,
"precision": 0.3551406608650635,
"recall": 0.3502956304842808,
"support": 2000.0
},
"mx": {
"f1-score": 0.38202247191011235,
"precision": 0.35843373493975905,
"recall": 0.40893470790378006,
"support": 291.0
},
"pe": {
"f1-score": 0.3446244477172312,
"precision": 0.3015463917525773,
"recall": 0.4020618556701031,
"support": 291.0
},
"pr": {
"f1-score": 0.6666666666666666,
"precision": 0.6601941747572816,
"recall": 0.6732673267326733,
"support": 101.0
},
"uy": {
"f1-score": 0.38954869358669836,
"precision": 0.4270833333333333,
"recall": 0.35807860262008734,
"support": 229.0
},
"ve": {
"f1-score": 0.0,
"precision": 0.0,
"recall": 0.0,
"support": 22.0
},
"weighted avg": {
"f1-score": 0.3655613869415472,
"precision": 0.36979288589763315,
"recall": 0.368,
"support": 2000.0
}
},
"eval_f1": 0.3499479191731958,
"eval_loss": 1.8162137269973755,
"eval_runtime": 2.8908,
"eval_samples_per_second": 691.846,
"eval_steps_per_second": 86.481,
"step": 1752
}
],
"logging_steps": 10,
"max_steps": 1752,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 230203751995008.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}