{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.1737089201877935, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0011737089201877935, "grad_norm": 0.27773135900497437, "learning_rate": 4e-05, "loss": 1.1957, "step": 1 }, { "epoch": 0.002347417840375587, "grad_norm": 0.26547771692276, "learning_rate": 8e-05, "loss": 1.1284, "step": 2 }, { "epoch": 0.0035211267605633804, "grad_norm": 0.236787810921669, "learning_rate": 0.00012, "loss": 1.1823, "step": 3 }, { "epoch": 0.004694835680751174, "grad_norm": 0.2459038347005844, "learning_rate": 0.00016, "loss": 1.1409, "step": 4 }, { "epoch": 0.005868544600938967, "grad_norm": 0.2526487410068512, "learning_rate": 0.0002, "loss": 1.12, "step": 5 }, { "epoch": 0.007042253521126761, "grad_norm": 0.2795103192329407, "learning_rate": 0.00019976387249114524, "loss": 1.1579, "step": 6 }, { "epoch": 0.008215962441314555, "grad_norm": 0.2876183092594147, "learning_rate": 0.00019952774498229045, "loss": 1.1211, "step": 7 }, { "epoch": 0.009389671361502348, "grad_norm": 0.3014296293258667, "learning_rate": 0.00019929161747343565, "loss": 1.1118, "step": 8 }, { "epoch": 0.01056338028169014, "grad_norm": 0.29106494784355164, "learning_rate": 0.00019905548996458088, "loss": 1.1787, "step": 9 }, { "epoch": 0.011737089201877934, "grad_norm": 0.3211474120616913, "learning_rate": 0.00019881936245572609, "loss": 1.1004, "step": 10 }, { "epoch": 0.012910798122065728, "grad_norm": 0.3358176350593567, "learning_rate": 0.00019858323494687132, "loss": 1.1099, "step": 11 }, { "epoch": 0.014084507042253521, "grad_norm": 0.3236922323703766, "learning_rate": 0.00019834710743801655, "loss": 1.048, "step": 12 }, { "epoch": 0.015258215962441314, "grad_norm": 0.31388312578201294, "learning_rate": 0.00019811097992916175, "loss": 1.0532, "step": 13 }, { "epoch": 0.01643192488262911, "grad_norm": 0.320402055978775, "learning_rate": 0.00019787485242030696, "loss": 1.0757, "step": 14 }, { "epoch": 0.017605633802816902, "grad_norm": 0.32999494671821594, "learning_rate": 0.0001976387249114522, "loss": 1.122, "step": 15 }, { "epoch": 0.018779342723004695, "grad_norm": 0.30936214327812195, "learning_rate": 0.00019740259740259742, "loss": 1.1156, "step": 16 }, { "epoch": 0.01995305164319249, "grad_norm": 0.2863931953907013, "learning_rate": 0.00019716646989374263, "loss": 1.0414, "step": 17 }, { "epoch": 0.02112676056338028, "grad_norm": 0.29143351316452026, "learning_rate": 0.00019693034238488786, "loss": 1.0379, "step": 18 }, { "epoch": 0.022300469483568074, "grad_norm": 0.28874626755714417, "learning_rate": 0.0001966942148760331, "loss": 1.0388, "step": 19 }, { "epoch": 0.023474178403755867, "grad_norm": 0.30588293075561523, "learning_rate": 0.00019645808736717827, "loss": 1.0515, "step": 20 }, { "epoch": 0.02464788732394366, "grad_norm": 0.29231536388397217, "learning_rate": 0.0001962219598583235, "loss": 1.0472, "step": 21 }, { "epoch": 0.025821596244131457, "grad_norm": 0.2783581614494324, "learning_rate": 0.00019598583234946873, "loss": 1.0608, "step": 22 }, { "epoch": 0.02699530516431925, "grad_norm": 0.29816293716430664, "learning_rate": 0.00019574970484061393, "loss": 1.0986, "step": 23 }, { "epoch": 0.028169014084507043, "grad_norm": 0.27919578552246094, "learning_rate": 0.00019551357733175916, "loss": 1.0265, "step": 24 }, { "epoch": 0.029342723004694836, "grad_norm": 0.3144524097442627, "learning_rate": 0.00019527744982290437, "loss": 1.0699, "step": 25 }, { "epoch": 0.03051643192488263, "grad_norm": 0.3090282678604126, "learning_rate": 0.0001950413223140496, "loss": 1.0601, "step": 26 }, { "epoch": 0.03169014084507042, "grad_norm": 0.30304697155952454, "learning_rate": 0.0001948051948051948, "loss": 1.0926, "step": 27 }, { "epoch": 0.03286384976525822, "grad_norm": 0.29015883803367615, "learning_rate": 0.00019456906729634004, "loss": 1.0, "step": 28 }, { "epoch": 0.03403755868544601, "grad_norm": 0.29359501600265503, "learning_rate": 0.00019433293978748527, "loss": 0.988, "step": 29 }, { "epoch": 0.035211267605633804, "grad_norm": 0.2772333323955536, "learning_rate": 0.00019409681227863047, "loss": 0.9758, "step": 30 }, { "epoch": 0.036384976525821594, "grad_norm": 0.2761421799659729, "learning_rate": 0.00019386068476977568, "loss": 0.9926, "step": 31 }, { "epoch": 0.03755868544600939, "grad_norm": NaN, "learning_rate": 0.00019386068476977568, "loss": 1.0944, "step": 32 }, { "epoch": 0.03873239436619718, "grad_norm": 0.2766799330711365, "learning_rate": 0.0001936245572609209, "loss": 0.9813, "step": 33 }, { "epoch": 0.03990610328638498, "grad_norm": 0.28922533988952637, "learning_rate": 0.0001933884297520661, "loss": 0.9839, "step": 34 }, { "epoch": 0.04107981220657277, "grad_norm": 0.28271371126174927, "learning_rate": 0.00019315230224321134, "loss": 1.0125, "step": 35 }, { "epoch": 0.04225352112676056, "grad_norm": 0.2955509424209595, "learning_rate": 0.00019291617473435658, "loss": 1.0049, "step": 36 }, { "epoch": 0.04342723004694836, "grad_norm": 0.2909109592437744, "learning_rate": 0.00019268004722550178, "loss": 1.0015, "step": 37 }, { "epoch": 0.04460093896713615, "grad_norm": 0.29657021164894104, "learning_rate": 0.00019244391971664698, "loss": 1.0107, "step": 38 }, { "epoch": 0.045774647887323945, "grad_norm": 0.29010507464408875, "learning_rate": 0.00019220779220779222, "loss": 0.9918, "step": 39 }, { "epoch": 0.046948356807511735, "grad_norm": 0.2906627058982849, "learning_rate": 0.00019197166469893745, "loss": 0.9843, "step": 40 }, { "epoch": 0.04812206572769953, "grad_norm": 0.2919193208217621, "learning_rate": 0.00019173553719008265, "loss": 0.9889, "step": 41 }, { "epoch": 0.04929577464788732, "grad_norm": 0.3219091296195984, "learning_rate": 0.00019149940968122788, "loss": 0.9979, "step": 42 }, { "epoch": 0.05046948356807512, "grad_norm": 0.29512301087379456, "learning_rate": 0.0001912632821723731, "loss": 0.989, "step": 43 }, { "epoch": 0.051643192488262914, "grad_norm": 0.3190619647502899, "learning_rate": 0.0001910271546635183, "loss": 0.9563, "step": 44 }, { "epoch": 0.0528169014084507, "grad_norm": 0.310253381729126, "learning_rate": 0.00019079102715466352, "loss": 1.037, "step": 45 }, { "epoch": 0.0539906103286385, "grad_norm": 0.3140093684196472, "learning_rate": 0.00019055489964580876, "loss": 0.9687, "step": 46 }, { "epoch": 0.05516431924882629, "grad_norm": 0.2816644310951233, "learning_rate": 0.00019031877213695396, "loss": 0.9372, "step": 47 }, { "epoch": 0.056338028169014086, "grad_norm": 0.3012441396713257, "learning_rate": 0.0001900826446280992, "loss": 0.9968, "step": 48 }, { "epoch": 0.057511737089201875, "grad_norm": 0.29789185523986816, "learning_rate": 0.0001898465171192444, "loss": 0.9143, "step": 49 }, { "epoch": 0.05868544600938967, "grad_norm": 0.29454007744789124, "learning_rate": 0.00018961038961038963, "loss": 0.9837, "step": 50 }, { "epoch": 0.05985915492957746, "grad_norm": 0.321218341588974, "learning_rate": 0.00018937426210153483, "loss": 1.0135, "step": 51 }, { "epoch": 0.06103286384976526, "grad_norm": 0.30039164423942566, "learning_rate": 0.00018913813459268006, "loss": 0.9639, "step": 52 }, { "epoch": 0.062206572769953054, "grad_norm": 0.3052615225315094, "learning_rate": 0.0001889020070838253, "loss": 0.9401, "step": 53 }, { "epoch": 0.06338028169014084, "grad_norm": 0.3177138864994049, "learning_rate": 0.00018866587957497047, "loss": 0.9626, "step": 54 }, { "epoch": 0.06455399061032864, "grad_norm": 0.3098903298377991, "learning_rate": 0.0001884297520661157, "loss": 0.9535, "step": 55 }, { "epoch": 0.06572769953051644, "grad_norm": 0.33165299892425537, "learning_rate": 0.00018819362455726094, "loss": 1.0475, "step": 56 }, { "epoch": 0.06690140845070422, "grad_norm": 0.3054540455341339, "learning_rate": 0.00018795749704840614, "loss": 0.988, "step": 57 }, { "epoch": 0.06807511737089202, "grad_norm": 0.3412969708442688, "learning_rate": 0.00018772136953955137, "loss": 0.9531, "step": 58 }, { "epoch": 0.06924882629107981, "grad_norm": 0.3173505663871765, "learning_rate": 0.0001874852420306966, "loss": 1.0037, "step": 59 }, { "epoch": 0.07042253521126761, "grad_norm": 0.29377281665802, "learning_rate": 0.0001872491145218418, "loss": 0.9205, "step": 60 }, { "epoch": 0.0715962441314554, "grad_norm": 0.2970433831214905, "learning_rate": 0.000187012987012987, "loss": 0.8902, "step": 61 }, { "epoch": 0.07276995305164319, "grad_norm": 0.3081493675708771, "learning_rate": 0.00018677685950413224, "loss": 0.9498, "step": 62 }, { "epoch": 0.07394366197183098, "grad_norm": 0.31438371539115906, "learning_rate": 0.00018654073199527747, "loss": 0.9406, "step": 63 }, { "epoch": 0.07511737089201878, "grad_norm": 0.29640915989875793, "learning_rate": 0.00018630460448642268, "loss": 0.8948, "step": 64 }, { "epoch": 0.07629107981220658, "grad_norm": 0.33342233300209045, "learning_rate": 0.00018606847697756788, "loss": 0.941, "step": 65 }, { "epoch": 0.07746478873239436, "grad_norm": 0.31546634435653687, "learning_rate": 0.00018583234946871312, "loss": 0.9392, "step": 66 }, { "epoch": 0.07863849765258216, "grad_norm": 0.31528937816619873, "learning_rate": 0.00018559622195985832, "loss": 0.9293, "step": 67 }, { "epoch": 0.07981220657276995, "grad_norm": 0.33473101258277893, "learning_rate": 0.00018536009445100355, "loss": 0.9214, "step": 68 }, { "epoch": 0.08098591549295775, "grad_norm": 0.6588060259819031, "learning_rate": 0.00018512396694214878, "loss": 0.944, "step": 69 }, { "epoch": 0.08215962441314555, "grad_norm": 0.30120280385017395, "learning_rate": 0.000184887839433294, "loss": 0.9171, "step": 70 }, { "epoch": 0.08333333333333333, "grad_norm": 0.3417011499404907, "learning_rate": 0.0001846517119244392, "loss": 0.9382, "step": 71 }, { "epoch": 0.08450704225352113, "grad_norm": 0.3202987313270569, "learning_rate": 0.00018441558441558442, "loss": 0.931, "step": 72 }, { "epoch": 0.08568075117370892, "grad_norm": 0.3390517234802246, "learning_rate": 0.00018417945690672965, "loss": 0.9218, "step": 73 }, { "epoch": 0.08685446009389672, "grad_norm": 0.32109472155570984, "learning_rate": 0.00018394332939787486, "loss": 0.9226, "step": 74 }, { "epoch": 0.0880281690140845, "grad_norm": 0.3435365855693817, "learning_rate": 0.0001837072018890201, "loss": 0.9402, "step": 75 }, { "epoch": 0.0892018779342723, "grad_norm": 0.3335697054862976, "learning_rate": 0.00018347107438016532, "loss": 0.9385, "step": 76 }, { "epoch": 0.0903755868544601, "grad_norm": 0.32050758600234985, "learning_rate": 0.0001832349468713105, "loss": 0.8992, "step": 77 }, { "epoch": 0.09154929577464789, "grad_norm": 0.32620421051979065, "learning_rate": 0.00018299881936245573, "loss": 0.9476, "step": 78 }, { "epoch": 0.09272300469483569, "grad_norm": 0.33306750655174255, "learning_rate": 0.00018276269185360096, "loss": 0.9458, "step": 79 }, { "epoch": 0.09389671361502347, "grad_norm": 0.3500649034976959, "learning_rate": 0.00018252656434474617, "loss": 0.9612, "step": 80 }, { "epoch": 0.09507042253521127, "grad_norm": 0.3186359405517578, "learning_rate": 0.0001822904368358914, "loss": 0.9527, "step": 81 }, { "epoch": 0.09624413145539906, "grad_norm": 0.3317716717720032, "learning_rate": 0.0001820543093270366, "loss": 0.9648, "step": 82 }, { "epoch": 0.09741784037558686, "grad_norm": 0.3196907639503479, "learning_rate": 0.00018181818181818183, "loss": 0.9643, "step": 83 }, { "epoch": 0.09859154929577464, "grad_norm": 0.3195818066596985, "learning_rate": 0.00018158205430932704, "loss": 0.9121, "step": 84 }, { "epoch": 0.09976525821596244, "grad_norm": 0.33151793479919434, "learning_rate": 0.00018134592680047227, "loss": 0.9051, "step": 85 }, { "epoch": 0.10093896713615023, "grad_norm": 0.3110804259777069, "learning_rate": 0.00018110979929161747, "loss": 0.9241, "step": 86 }, { "epoch": 0.10211267605633803, "grad_norm": 0.34278568625450134, "learning_rate": 0.0001808736717827627, "loss": 0.9634, "step": 87 }, { "epoch": 0.10328638497652583, "grad_norm": 0.34013500809669495, "learning_rate": 0.0001806375442739079, "loss": 0.8822, "step": 88 }, { "epoch": 0.10446009389671361, "grad_norm": 0.3449755012989044, "learning_rate": 0.00018040141676505314, "loss": 0.969, "step": 89 }, { "epoch": 0.1056338028169014, "grad_norm": 0.3166862726211548, "learning_rate": 0.00018016528925619835, "loss": 0.885, "step": 90 }, { "epoch": 0.1068075117370892, "grad_norm": 0.3260084092617035, "learning_rate": 0.00017992916174734358, "loss": 0.8908, "step": 91 }, { "epoch": 0.107981220657277, "grad_norm": 0.32791605591773987, "learning_rate": 0.0001796930342384888, "loss": 0.8822, "step": 92 }, { "epoch": 0.10915492957746478, "grad_norm": 0.31909653544425964, "learning_rate": 0.000179456906729634, "loss": 0.8463, "step": 93 }, { "epoch": 0.11032863849765258, "grad_norm": 0.3413308262825012, "learning_rate": 0.00017922077922077922, "loss": 0.9232, "step": 94 }, { "epoch": 0.11150234741784038, "grad_norm": 0.32644134759902954, "learning_rate": 0.00017898465171192445, "loss": 0.9113, "step": 95 }, { "epoch": 0.11267605633802817, "grad_norm": 0.33090126514434814, "learning_rate": 0.00017874852420306965, "loss": 0.9286, "step": 96 }, { "epoch": 0.11384976525821597, "grad_norm": 0.37200361490249634, "learning_rate": 0.00017851239669421489, "loss": 0.9239, "step": 97 }, { "epoch": 0.11502347417840375, "grad_norm": 0.3274000585079193, "learning_rate": 0.00017827626918536012, "loss": 0.9038, "step": 98 }, { "epoch": 0.11619718309859155, "grad_norm": 0.3768482506275177, "learning_rate": 0.00017804014167650532, "loss": 0.8558, "step": 99 }, { "epoch": 0.11737089201877934, "grad_norm": 0.32970595359802246, "learning_rate": 0.00017780401416765053, "loss": 0.9057, "step": 100 }, { "epoch": 0.11854460093896714, "grad_norm": 0.37230944633483887, "learning_rate": 0.00017756788665879576, "loss": 0.9211, "step": 101 }, { "epoch": 0.11971830985915492, "grad_norm": 0.352201372385025, "learning_rate": 0.000177331759149941, "loss": 0.9497, "step": 102 }, { "epoch": 0.12089201877934272, "grad_norm": 0.363364577293396, "learning_rate": 0.0001770956316410862, "loss": 0.9535, "step": 103 }, { "epoch": 0.12206572769953052, "grad_norm": 0.3388724625110626, "learning_rate": 0.00017685950413223143, "loss": 0.8908, "step": 104 }, { "epoch": 0.12323943661971831, "grad_norm": 0.34684258699417114, "learning_rate": 0.00017662337662337663, "loss": 0.8981, "step": 105 }, { "epoch": 0.12441314553990611, "grad_norm": 0.31892621517181396, "learning_rate": 0.00017638724911452183, "loss": 0.8461, "step": 106 }, { "epoch": 0.1255868544600939, "grad_norm": 0.32913845777511597, "learning_rate": 0.00017615112160566707, "loss": 0.9087, "step": 107 }, { "epoch": 0.1267605633802817, "grad_norm": 0.3695410490036011, "learning_rate": 0.0001759149940968123, "loss": 0.8899, "step": 108 }, { "epoch": 0.12793427230046947, "grad_norm": 0.3455798923969269, "learning_rate": 0.0001756788665879575, "loss": 0.9045, "step": 109 }, { "epoch": 0.12910798122065728, "grad_norm": 0.3612275719642639, "learning_rate": 0.0001754427390791027, "loss": 0.8861, "step": 110 }, { "epoch": 0.13028169014084506, "grad_norm": 0.4106651544570923, "learning_rate": 0.00017520661157024794, "loss": 0.9152, "step": 111 }, { "epoch": 0.13145539906103287, "grad_norm": 0.3604993224143982, "learning_rate": 0.00017497048406139317, "loss": 0.9141, "step": 112 }, { "epoch": 0.13262910798122066, "grad_norm": 0.3496919870376587, "learning_rate": 0.00017473435655253837, "loss": 0.9061, "step": 113 }, { "epoch": 0.13380281690140844, "grad_norm": 0.33643972873687744, "learning_rate": 0.0001744982290436836, "loss": 0.8877, "step": 114 }, { "epoch": 0.13497652582159625, "grad_norm": 0.33064204454421997, "learning_rate": 0.00017426210153482884, "loss": 0.8967, "step": 115 }, { "epoch": 0.13615023474178403, "grad_norm": 0.37868356704711914, "learning_rate": 0.00017402597402597401, "loss": 0.8957, "step": 116 }, { "epoch": 0.13732394366197184, "grad_norm": 0.34379109740257263, "learning_rate": 0.00017378984651711925, "loss": 0.9332, "step": 117 }, { "epoch": 0.13849765258215962, "grad_norm": 0.37193912267684937, "learning_rate": 0.00017355371900826448, "loss": 0.9513, "step": 118 }, { "epoch": 0.1396713615023474, "grad_norm": 0.33701232075691223, "learning_rate": 0.00017331759149940968, "loss": 0.8946, "step": 119 }, { "epoch": 0.14084507042253522, "grad_norm": 0.35765206813812256, "learning_rate": 0.0001730814639905549, "loss": 0.8931, "step": 120 }, { "epoch": 0.142018779342723, "grad_norm": 0.3511311411857605, "learning_rate": 0.00017284533648170012, "loss": 0.9042, "step": 121 }, { "epoch": 0.1431924882629108, "grad_norm": 0.33516445755958557, "learning_rate": 0.00017260920897284535, "loss": 0.8564, "step": 122 }, { "epoch": 0.1443661971830986, "grad_norm": 0.385959267616272, "learning_rate": 0.00017237308146399055, "loss": 0.963, "step": 123 }, { "epoch": 0.14553990610328638, "grad_norm": 0.34608641266822815, "learning_rate": 0.00017213695395513578, "loss": 0.8666, "step": 124 }, { "epoch": 0.1467136150234742, "grad_norm": 0.3705556392669678, "learning_rate": 0.00017190082644628102, "loss": 0.7783, "step": 125 }, { "epoch": 0.14788732394366197, "grad_norm": 0.3213210701942444, "learning_rate": 0.00017166469893742622, "loss": 0.8428, "step": 126 }, { "epoch": 0.14906103286384975, "grad_norm": 0.3903498351573944, "learning_rate": 0.00017142857142857143, "loss": 0.8418, "step": 127 }, { "epoch": 0.15023474178403756, "grad_norm": 0.3556365668773651, "learning_rate": 0.00017119244391971666, "loss": 0.8612, "step": 128 }, { "epoch": 0.15140845070422534, "grad_norm": 0.3734995424747467, "learning_rate": 0.00017095631641086186, "loss": 0.8845, "step": 129 }, { "epoch": 0.15258215962441316, "grad_norm": 0.33735260367393494, "learning_rate": 0.0001707201889020071, "loss": 0.8752, "step": 130 }, { "epoch": 0.15375586854460094, "grad_norm": 0.38340267539024353, "learning_rate": 0.00017048406139315232, "loss": 0.8847, "step": 131 }, { "epoch": 0.15492957746478872, "grad_norm": 0.3654419779777527, "learning_rate": 0.00017024793388429753, "loss": 0.8448, "step": 132 }, { "epoch": 0.15610328638497653, "grad_norm": 0.3601568341255188, "learning_rate": 0.00017001180637544273, "loss": 0.8981, "step": 133 }, { "epoch": 0.1572769953051643, "grad_norm": 0.40733832120895386, "learning_rate": 0.00016977567886658796, "loss": 0.9135, "step": 134 }, { "epoch": 0.15845070422535212, "grad_norm": 0.34627673029899597, "learning_rate": 0.0001695395513577332, "loss": 0.9164, "step": 135 }, { "epoch": 0.1596244131455399, "grad_norm": 0.3865872621536255, "learning_rate": 0.0001693034238488784, "loss": 0.9222, "step": 136 }, { "epoch": 0.1607981220657277, "grad_norm": 0.4011456072330475, "learning_rate": 0.00016906729634002363, "loss": 0.8843, "step": 137 }, { "epoch": 0.1619718309859155, "grad_norm": 0.32259878516197205, "learning_rate": 0.00016883116883116884, "loss": 0.8427, "step": 138 }, { "epoch": 0.16314553990610328, "grad_norm": 0.3807618319988251, "learning_rate": 0.00016859504132231404, "loss": 0.8684, "step": 139 }, { "epoch": 0.1643192488262911, "grad_norm": 0.3658106327056885, "learning_rate": 0.00016835891381345927, "loss": 0.9024, "step": 140 }, { "epoch": 0.16549295774647887, "grad_norm": 0.3638787865638733, "learning_rate": 0.0001681227863046045, "loss": 0.8582, "step": 141 }, { "epoch": 0.16666666666666666, "grad_norm": 0.3839091360569, "learning_rate": 0.0001678866587957497, "loss": 0.8543, "step": 142 }, { "epoch": 0.16784037558685447, "grad_norm": 0.33579927682876587, "learning_rate": 0.00016765053128689494, "loss": 0.8765, "step": 143 }, { "epoch": 0.16901408450704225, "grad_norm": 0.35091203451156616, "learning_rate": 0.00016741440377804014, "loss": 0.8504, "step": 144 }, { "epoch": 0.17018779342723006, "grad_norm": 0.35823047161102295, "learning_rate": 0.00016717827626918538, "loss": 0.8534, "step": 145 }, { "epoch": 0.17136150234741784, "grad_norm": 0.37154486775398254, "learning_rate": 0.00016694214876033058, "loss": 0.851, "step": 146 }, { "epoch": 0.17253521126760563, "grad_norm": 0.33140066266059875, "learning_rate": 0.0001667060212514758, "loss": 0.8136, "step": 147 }, { "epoch": 0.17370892018779344, "grad_norm": 0.37408292293548584, "learning_rate": 0.00016646989374262104, "loss": 0.8933, "step": 148 }, { "epoch": 0.17488262910798122, "grad_norm": 0.36203357577323914, "learning_rate": 0.00016623376623376625, "loss": 0.8747, "step": 149 }, { "epoch": 0.176056338028169, "grad_norm": 0.35033532977104187, "learning_rate": 0.00016599763872491145, "loss": 0.8273, "step": 150 }, { "epoch": 0.1772300469483568, "grad_norm": 0.345048189163208, "learning_rate": 0.00016576151121605668, "loss": 0.8698, "step": 151 }, { "epoch": 0.1784037558685446, "grad_norm": 0.3592989146709442, "learning_rate": 0.0001655253837072019, "loss": 0.8483, "step": 152 }, { "epoch": 0.1795774647887324, "grad_norm": 0.3685864806175232, "learning_rate": 0.00016528925619834712, "loss": 0.915, "step": 153 }, { "epoch": 0.1807511737089202, "grad_norm": 0.3427909314632416, "learning_rate": 0.00016505312868949235, "loss": 0.8321, "step": 154 }, { "epoch": 0.18192488262910797, "grad_norm": 0.34697192907333374, "learning_rate": 0.00016481700118063756, "loss": 0.8801, "step": 155 }, { "epoch": 0.18309859154929578, "grad_norm": 0.3387276530265808, "learning_rate": 0.00016458087367178276, "loss": 0.8237, "step": 156 }, { "epoch": 0.18427230046948356, "grad_norm": 0.3547775447368622, "learning_rate": 0.000164344746162928, "loss": 0.8645, "step": 157 }, { "epoch": 0.18544600938967137, "grad_norm": 0.3342725932598114, "learning_rate": 0.00016410861865407322, "loss": 0.82, "step": 158 }, { "epoch": 0.18661971830985916, "grad_norm": 0.4317960739135742, "learning_rate": 0.00016387249114521843, "loss": 0.8614, "step": 159 }, { "epoch": 0.18779342723004694, "grad_norm": 0.35031062364578247, "learning_rate": 0.00016363636363636366, "loss": 0.8193, "step": 160 }, { "epoch": 0.18896713615023475, "grad_norm": 0.3616986572742462, "learning_rate": 0.00016340023612750886, "loss": 0.8571, "step": 161 }, { "epoch": 0.19014084507042253, "grad_norm": 0.36284518241882324, "learning_rate": 0.00016316410861865407, "loss": 0.8555, "step": 162 }, { "epoch": 0.19131455399061034, "grad_norm": 0.42962291836738586, "learning_rate": 0.0001629279811097993, "loss": 0.8574, "step": 163 }, { "epoch": 0.19248826291079812, "grad_norm": 0.330268532037735, "learning_rate": 0.00016269185360094453, "loss": 0.8952, "step": 164 }, { "epoch": 0.1936619718309859, "grad_norm": 0.33917295932769775, "learning_rate": 0.00016245572609208974, "loss": 0.8588, "step": 165 }, { "epoch": 0.19483568075117372, "grad_norm": 0.3963412046432495, "learning_rate": 0.00016221959858323494, "loss": 0.8451, "step": 166 }, { "epoch": 0.1960093896713615, "grad_norm": 0.33864182233810425, "learning_rate": 0.00016198347107438017, "loss": 0.8734, "step": 167 }, { "epoch": 0.19718309859154928, "grad_norm": 0.3751653730869293, "learning_rate": 0.00016174734356552538, "loss": 0.8786, "step": 168 }, { "epoch": 0.1983568075117371, "grad_norm": 0.4138842821121216, "learning_rate": 0.0001615112160566706, "loss": 0.8608, "step": 169 }, { "epoch": 0.19953051643192488, "grad_norm": 0.3747748136520386, "learning_rate": 0.00016127508854781584, "loss": 0.8901, "step": 170 }, { "epoch": 0.2007042253521127, "grad_norm": 0.3302014172077179, "learning_rate": 0.00016103896103896104, "loss": 0.8538, "step": 171 }, { "epoch": 0.20187793427230047, "grad_norm": 0.36144372820854187, "learning_rate": 0.00016080283353010625, "loss": 0.8634, "step": 172 }, { "epoch": 0.20305164319248825, "grad_norm": 0.3579455018043518, "learning_rate": 0.00016056670602125148, "loss": 0.8536, "step": 173 }, { "epoch": 0.20422535211267606, "grad_norm": 0.3475671410560608, "learning_rate": 0.0001603305785123967, "loss": 0.8304, "step": 174 }, { "epoch": 0.20539906103286384, "grad_norm": 0.34114810824394226, "learning_rate": 0.00016009445100354192, "loss": 0.8276, "step": 175 }, { "epoch": 0.20657276995305165, "grad_norm": 0.32198190689086914, "learning_rate": 0.00015985832349468715, "loss": 0.815, "step": 176 }, { "epoch": 0.20774647887323944, "grad_norm": 0.4003874361515045, "learning_rate": 0.00015962219598583238, "loss": 0.8523, "step": 177 }, { "epoch": 0.20892018779342722, "grad_norm": 0.32290229201316833, "learning_rate": 0.00015938606847697756, "loss": 0.8465, "step": 178 }, { "epoch": 0.21009389671361503, "grad_norm": 0.35729506611824036, "learning_rate": 0.0001591499409681228, "loss": 0.8437, "step": 179 }, { "epoch": 0.2112676056338028, "grad_norm": 0.33743324875831604, "learning_rate": 0.00015891381345926802, "loss": 0.8351, "step": 180 }, { "epoch": 0.21244131455399062, "grad_norm": 0.34673774242401123, "learning_rate": 0.00015867768595041322, "loss": 0.8146, "step": 181 }, { "epoch": 0.2136150234741784, "grad_norm": 0.37883323431015015, "learning_rate": 0.00015844155844155845, "loss": 0.8889, "step": 182 }, { "epoch": 0.2147887323943662, "grad_norm": 0.34172534942626953, "learning_rate": 0.00015820543093270366, "loss": 0.8479, "step": 183 }, { "epoch": 0.215962441314554, "grad_norm": 0.39948219060897827, "learning_rate": 0.0001579693034238489, "loss": 0.8383, "step": 184 }, { "epoch": 0.21713615023474178, "grad_norm": 0.33746814727783203, "learning_rate": 0.0001577331759149941, "loss": 0.8713, "step": 185 }, { "epoch": 0.21830985915492956, "grad_norm": 0.34141069650650024, "learning_rate": 0.00015749704840613933, "loss": 0.8303, "step": 186 }, { "epoch": 0.21948356807511737, "grad_norm": 0.35994264483451843, "learning_rate": 0.00015726092089728456, "loss": 0.7919, "step": 187 }, { "epoch": 0.22065727699530516, "grad_norm": 0.34234684705734253, "learning_rate": 0.00015702479338842976, "loss": 0.8225, "step": 188 }, { "epoch": 0.22183098591549297, "grad_norm": 0.3601793050765991, "learning_rate": 0.00015678866587957497, "loss": 0.8395, "step": 189 }, { "epoch": 0.22300469483568075, "grad_norm": 0.3154338002204895, "learning_rate": 0.0001565525383707202, "loss": 0.7735, "step": 190 }, { "epoch": 0.22417840375586853, "grad_norm": 0.3758296072483063, "learning_rate": 0.0001563164108618654, "loss": 0.8241, "step": 191 }, { "epoch": 0.22535211267605634, "grad_norm": 0.3732200264930725, "learning_rate": 0.00015608028335301063, "loss": 0.8116, "step": 192 }, { "epoch": 0.22652582159624413, "grad_norm": 0.3601556718349457, "learning_rate": 0.00015584415584415587, "loss": 0.8242, "step": 193 }, { "epoch": 0.22769953051643194, "grad_norm": 0.360442191362381, "learning_rate": 0.00015560802833530107, "loss": 0.832, "step": 194 }, { "epoch": 0.22887323943661972, "grad_norm": 0.35598254203796387, "learning_rate": 0.00015537190082644627, "loss": 0.8938, "step": 195 }, { "epoch": 0.2300469483568075, "grad_norm": 0.3962613046169281, "learning_rate": 0.0001551357733175915, "loss": 0.8409, "step": 196 }, { "epoch": 0.2312206572769953, "grad_norm": 0.3521510064601898, "learning_rate": 0.00015489964580873674, "loss": 0.8298, "step": 197 }, { "epoch": 0.2323943661971831, "grad_norm": 0.34407946467399597, "learning_rate": 0.00015466351829988194, "loss": 0.7921, "step": 198 }, { "epoch": 0.2335680751173709, "grad_norm": 0.3572155237197876, "learning_rate": 0.00015442739079102717, "loss": 0.8997, "step": 199 }, { "epoch": 0.2347417840375587, "grad_norm": 0.345745712518692, "learning_rate": 0.00015419126328217238, "loss": 0.8563, "step": 200 }, { "epoch": 0.23591549295774647, "grad_norm": 0.3741077780723572, "learning_rate": 0.00015395513577331758, "loss": 0.8334, "step": 201 }, { "epoch": 0.23708920187793428, "grad_norm": 0.36866459250450134, "learning_rate": 0.00015371900826446281, "loss": 0.8398, "step": 202 }, { "epoch": 0.23826291079812206, "grad_norm": 0.3834739625453949, "learning_rate": 0.00015348288075560805, "loss": 0.8181, "step": 203 }, { "epoch": 0.23943661971830985, "grad_norm": 0.373045951128006, "learning_rate": 0.00015324675324675325, "loss": 0.8044, "step": 204 }, { "epoch": 0.24061032863849766, "grad_norm": 0.3418562412261963, "learning_rate": 0.00015301062573789848, "loss": 0.8454, "step": 205 }, { "epoch": 0.24178403755868544, "grad_norm": 0.36289098858833313, "learning_rate": 0.00015277449822904369, "loss": 0.8478, "step": 206 }, { "epoch": 0.24295774647887325, "grad_norm": 0.38806968927383423, "learning_rate": 0.00015253837072018892, "loss": 0.804, "step": 207 }, { "epoch": 0.24413145539906103, "grad_norm": 0.34217599034309387, "learning_rate": 0.00015230224321133412, "loss": 0.8391, "step": 208 }, { "epoch": 0.24530516431924881, "grad_norm": 0.3738957643508911, "learning_rate": 0.00015206611570247935, "loss": 0.9026, "step": 209 }, { "epoch": 0.24647887323943662, "grad_norm": 0.3481609523296356, "learning_rate": 0.00015182998819362458, "loss": 0.8674, "step": 210 }, { "epoch": 0.2476525821596244, "grad_norm": 0.38967254757881165, "learning_rate": 0.00015159386068476976, "loss": 0.8796, "step": 211 }, { "epoch": 0.24882629107981222, "grad_norm": 0.34841835498809814, "learning_rate": 0.000151357733175915, "loss": 0.7913, "step": 212 }, { "epoch": 0.25, "grad_norm": 0.33826395869255066, "learning_rate": 0.00015112160566706023, "loss": 0.8539, "step": 213 }, { "epoch": 0.2511737089201878, "grad_norm": 0.35131266713142395, "learning_rate": 0.00015088547815820543, "loss": 0.8072, "step": 214 }, { "epoch": 0.25234741784037557, "grad_norm": 0.3298250734806061, "learning_rate": 0.00015064935064935066, "loss": 0.7688, "step": 215 }, { "epoch": 0.2535211267605634, "grad_norm": 0.33808133006095886, "learning_rate": 0.0001504132231404959, "loss": 0.7609, "step": 216 }, { "epoch": 0.2546948356807512, "grad_norm": 0.37146687507629395, "learning_rate": 0.0001501770956316411, "loss": 0.843, "step": 217 }, { "epoch": 0.25586854460093894, "grad_norm": 0.33817118406295776, "learning_rate": 0.0001499409681227863, "loss": 0.7828, "step": 218 }, { "epoch": 0.25704225352112675, "grad_norm": 0.35203686356544495, "learning_rate": 0.00014970484061393153, "loss": 0.8236, "step": 219 }, { "epoch": 0.25821596244131456, "grad_norm": 0.34176716208457947, "learning_rate": 0.00014946871310507676, "loss": 0.8191, "step": 220 }, { "epoch": 0.25938967136150237, "grad_norm": 0.34649035334587097, "learning_rate": 0.00014923258559622197, "loss": 0.8284, "step": 221 }, { "epoch": 0.2605633802816901, "grad_norm": 0.35891467332839966, "learning_rate": 0.00014899645808736717, "loss": 0.8149, "step": 222 }, { "epoch": 0.26173708920187794, "grad_norm": 0.3408451974391937, "learning_rate": 0.0001487603305785124, "loss": 0.8049, "step": 223 }, { "epoch": 0.26291079812206575, "grad_norm": 0.36554664373397827, "learning_rate": 0.0001485242030696576, "loss": 0.8478, "step": 224 }, { "epoch": 0.2640845070422535, "grad_norm": 0.3355228304862976, "learning_rate": 0.00014828807556080284, "loss": 0.815, "step": 225 }, { "epoch": 0.2652582159624413, "grad_norm": 0.3500598669052124, "learning_rate": 0.00014805194805194807, "loss": 0.8571, "step": 226 }, { "epoch": 0.2664319248826291, "grad_norm": 0.3362652659416199, "learning_rate": 0.00014781582054309328, "loss": 0.8363, "step": 227 }, { "epoch": 0.2676056338028169, "grad_norm": 0.34258243441581726, "learning_rate": 0.00014757969303423848, "loss": 0.7648, "step": 228 }, { "epoch": 0.2687793427230047, "grad_norm": 0.34023317694664, "learning_rate": 0.0001473435655253837, "loss": 0.8373, "step": 229 }, { "epoch": 0.2699530516431925, "grad_norm": 0.35829535126686096, "learning_rate": 0.00014710743801652894, "loss": 0.8255, "step": 230 }, { "epoch": 0.2711267605633803, "grad_norm": 0.3499360978603363, "learning_rate": 0.00014687131050767415, "loss": 0.8514, "step": 231 }, { "epoch": 0.27230046948356806, "grad_norm": 0.3703480362892151, "learning_rate": 0.00014663518299881938, "loss": 0.8615, "step": 232 }, { "epoch": 0.2734741784037559, "grad_norm": 0.3460928499698639, "learning_rate": 0.0001463990554899646, "loss": 0.7891, "step": 233 }, { "epoch": 0.2746478873239437, "grad_norm": 0.34184372425079346, "learning_rate": 0.0001461629279811098, "loss": 0.8168, "step": 234 }, { "epoch": 0.27582159624413144, "grad_norm": 0.34520068764686584, "learning_rate": 0.00014592680047225502, "loss": 0.8271, "step": 235 }, { "epoch": 0.27699530516431925, "grad_norm": 0.3415423631668091, "learning_rate": 0.00014569067296340025, "loss": 0.783, "step": 236 }, { "epoch": 0.27816901408450706, "grad_norm": 0.34584441781044006, "learning_rate": 0.00014545454545454546, "loss": 0.8488, "step": 237 }, { "epoch": 0.2793427230046948, "grad_norm": 0.33898866176605225, "learning_rate": 0.0001452184179456907, "loss": 0.8786, "step": 238 }, { "epoch": 0.2805164319248826, "grad_norm": 0.3591814339160919, "learning_rate": 0.0001449822904368359, "loss": 0.8081, "step": 239 }, { "epoch": 0.28169014084507044, "grad_norm": 0.34305432438850403, "learning_rate": 0.0001447461629279811, "loss": 0.7911, "step": 240 }, { "epoch": 0.2828638497652582, "grad_norm": 0.35866865515708923, "learning_rate": 0.00014451003541912633, "loss": 0.8393, "step": 241 }, { "epoch": 0.284037558685446, "grad_norm": 0.3422331213951111, "learning_rate": 0.00014427390791027156, "loss": 0.848, "step": 242 }, { "epoch": 0.2852112676056338, "grad_norm": 0.33504337072372437, "learning_rate": 0.00014403778040141676, "loss": 0.7782, "step": 243 }, { "epoch": 0.2863849765258216, "grad_norm": 0.3509252667427063, "learning_rate": 0.000143801652892562, "loss": 0.8535, "step": 244 }, { "epoch": 0.2875586854460094, "grad_norm": 0.3254059851169586, "learning_rate": 0.0001435655253837072, "loss": 0.7642, "step": 245 }, { "epoch": 0.2887323943661972, "grad_norm": 0.33594879508018494, "learning_rate": 0.00014332939787485243, "loss": 0.814, "step": 246 }, { "epoch": 0.289906103286385, "grad_norm": 0.3620656132698059, "learning_rate": 0.00014309327036599764, "loss": 0.8248, "step": 247 }, { "epoch": 0.29107981220657275, "grad_norm": 0.3325202167034149, "learning_rate": 0.00014285714285714287, "loss": 0.7408, "step": 248 }, { "epoch": 0.29225352112676056, "grad_norm": 0.33905264735221863, "learning_rate": 0.0001426210153482881, "loss": 0.8446, "step": 249 }, { "epoch": 0.2934272300469484, "grad_norm": 0.3577309548854828, "learning_rate": 0.0001423848878394333, "loss": 0.784, "step": 250 }, { "epoch": 0.29460093896713613, "grad_norm": 0.3840247392654419, "learning_rate": 0.0001421487603305785, "loss": 0.8068, "step": 251 }, { "epoch": 0.29577464788732394, "grad_norm": 0.3539847433567047, "learning_rate": 0.00014191263282172374, "loss": 0.8232, "step": 252 }, { "epoch": 0.29694835680751175, "grad_norm": 0.33225932717323303, "learning_rate": 0.00014167650531286894, "loss": 0.7946, "step": 253 }, { "epoch": 0.2981220657276995, "grad_norm": 0.3429291546344757, "learning_rate": 0.00014144037780401418, "loss": 0.816, "step": 254 }, { "epoch": 0.2992957746478873, "grad_norm": 0.3584197163581848, "learning_rate": 0.0001412042502951594, "loss": 0.8351, "step": 255 }, { "epoch": 0.3004694835680751, "grad_norm": 0.35585007071495056, "learning_rate": 0.0001409681227863046, "loss": 0.8255, "step": 256 }, { "epoch": 0.30164319248826293, "grad_norm": 0.3510012924671173, "learning_rate": 0.00014073199527744982, "loss": 0.7889, "step": 257 }, { "epoch": 0.3028169014084507, "grad_norm": 0.36646419763565063, "learning_rate": 0.00014049586776859505, "loss": 0.8161, "step": 258 }, { "epoch": 0.3039906103286385, "grad_norm": 0.35207659006118774, "learning_rate": 0.00014025974025974028, "loss": 0.8151, "step": 259 }, { "epoch": 0.3051643192488263, "grad_norm": 0.33348143100738525, "learning_rate": 0.00014002361275088548, "loss": 0.8108, "step": 260 }, { "epoch": 0.30633802816901406, "grad_norm": 0.3474767506122589, "learning_rate": 0.00013978748524203072, "loss": 0.8105, "step": 261 }, { "epoch": 0.3075117370892019, "grad_norm": 0.37046462297439575, "learning_rate": 0.00013955135773317592, "loss": 0.867, "step": 262 }, { "epoch": 0.3086854460093897, "grad_norm": 0.3426377475261688, "learning_rate": 0.00013931523022432112, "loss": 0.8281, "step": 263 }, { "epoch": 0.30985915492957744, "grad_norm": 0.3340952694416046, "learning_rate": 0.00013907910271546636, "loss": 0.7805, "step": 264 }, { "epoch": 0.31103286384976525, "grad_norm": 0.3546634316444397, "learning_rate": 0.0001388429752066116, "loss": 0.824, "step": 265 }, { "epoch": 0.31220657276995306, "grad_norm": 0.3211507499217987, "learning_rate": 0.0001386068476977568, "loss": 0.7572, "step": 266 }, { "epoch": 0.31338028169014087, "grad_norm": 0.3440265357494354, "learning_rate": 0.000138370720188902, "loss": 0.8247, "step": 267 }, { "epoch": 0.3145539906103286, "grad_norm": 0.34174132347106934, "learning_rate": 0.00013813459268004723, "loss": 0.7939, "step": 268 }, { "epoch": 0.31572769953051644, "grad_norm": 0.3415057361125946, "learning_rate": 0.00013789846517119246, "loss": 0.8184, "step": 269 }, { "epoch": 0.31690140845070425, "grad_norm": 0.3313206732273102, "learning_rate": 0.00013766233766233766, "loss": 0.7936, "step": 270 }, { "epoch": 0.318075117370892, "grad_norm": 0.35693395137786865, "learning_rate": 0.0001374262101534829, "loss": 0.7738, "step": 271 }, { "epoch": 0.3192488262910798, "grad_norm": 0.3530910313129425, "learning_rate": 0.00013719008264462813, "loss": 0.7901, "step": 272 }, { "epoch": 0.3204225352112676, "grad_norm": 0.34867924451828003, "learning_rate": 0.0001369539551357733, "loss": 0.8281, "step": 273 }, { "epoch": 0.3215962441314554, "grad_norm": 0.34141889214515686, "learning_rate": 0.00013671782762691854, "loss": 0.7987, "step": 274 }, { "epoch": 0.3227699530516432, "grad_norm": 0.3511849045753479, "learning_rate": 0.00013648170011806377, "loss": 0.8306, "step": 275 }, { "epoch": 0.323943661971831, "grad_norm": 0.343523770570755, "learning_rate": 0.00013624557260920897, "loss": 0.7813, "step": 276 }, { "epoch": 0.32511737089201875, "grad_norm": 0.3539726138114929, "learning_rate": 0.0001360094451003542, "loss": 0.8258, "step": 277 }, { "epoch": 0.32629107981220656, "grad_norm": 0.35628989338874817, "learning_rate": 0.00013577331759149943, "loss": 0.829, "step": 278 }, { "epoch": 0.3274647887323944, "grad_norm": 0.3531114459037781, "learning_rate": 0.00013553719008264464, "loss": 0.8475, "step": 279 }, { "epoch": 0.3286384976525822, "grad_norm": 0.35344576835632324, "learning_rate": 0.00013530106257378984, "loss": 0.8343, "step": 280 }, { "epoch": 0.32981220657276994, "grad_norm": 0.37604016065597534, "learning_rate": 0.00013506493506493507, "loss": 0.7598, "step": 281 }, { "epoch": 0.33098591549295775, "grad_norm": 0.35646241903305054, "learning_rate": 0.0001348288075560803, "loss": 0.83, "step": 282 }, { "epoch": 0.33215962441314556, "grad_norm": 0.36084675788879395, "learning_rate": 0.0001345926800472255, "loss": 0.7465, "step": 283 }, { "epoch": 0.3333333333333333, "grad_norm": 0.3514406085014343, "learning_rate": 0.00013435655253837071, "loss": 0.7979, "step": 284 }, { "epoch": 0.3345070422535211, "grad_norm": 0.3554603159427643, "learning_rate": 0.00013412042502951595, "loss": 0.8487, "step": 285 }, { "epoch": 0.33568075117370894, "grad_norm": 0.3360341787338257, "learning_rate": 0.00013388429752066115, "loss": 0.7787, "step": 286 }, { "epoch": 0.3368544600938967, "grad_norm": 0.35026323795318604, "learning_rate": 0.00013364817001180638, "loss": 0.7845, "step": 287 }, { "epoch": 0.3380281690140845, "grad_norm": 0.3419228494167328, "learning_rate": 0.00013341204250295161, "loss": 0.7971, "step": 288 }, { "epoch": 0.3392018779342723, "grad_norm": 0.3314400315284729, "learning_rate": 0.00013317591499409682, "loss": 0.7899, "step": 289 }, { "epoch": 0.3403755868544601, "grad_norm": 0.3434331715106964, "learning_rate": 0.00013293978748524202, "loss": 0.827, "step": 290 }, { "epoch": 0.3415492957746479, "grad_norm": 0.34718382358551025, "learning_rate": 0.00013270365997638725, "loss": 0.7835, "step": 291 }, { "epoch": 0.3427230046948357, "grad_norm": 0.3585168421268463, "learning_rate": 0.00013246753246753249, "loss": 0.8728, "step": 292 }, { "epoch": 0.3438967136150235, "grad_norm": 0.3508673906326294, "learning_rate": 0.0001322314049586777, "loss": 0.836, "step": 293 }, { "epoch": 0.34507042253521125, "grad_norm": 0.40241560339927673, "learning_rate": 0.00013199527744982292, "loss": 0.8043, "step": 294 }, { "epoch": 0.34624413145539906, "grad_norm": 0.33775267004966736, "learning_rate": 0.00013175914994096813, "loss": 0.8047, "step": 295 }, { "epoch": 0.3474178403755869, "grad_norm": 0.3423898220062256, "learning_rate": 0.00013152302243211333, "loss": 0.7894, "step": 296 }, { "epoch": 0.3485915492957746, "grad_norm": 0.3472992479801178, "learning_rate": 0.00013128689492325856, "loss": 0.8198, "step": 297 }, { "epoch": 0.34976525821596244, "grad_norm": 0.3425481915473938, "learning_rate": 0.0001310507674144038, "loss": 0.8178, "step": 298 }, { "epoch": 0.35093896713615025, "grad_norm": 0.3459112048149109, "learning_rate": 0.000130814639905549, "loss": 0.7749, "step": 299 }, { "epoch": 0.352112676056338, "grad_norm": 0.353595495223999, "learning_rate": 0.00013057851239669423, "loss": 0.7886, "step": 300 }, { "epoch": 0.3532863849765258, "grad_norm": 0.35495465993881226, "learning_rate": 0.00013034238488783943, "loss": 0.771, "step": 301 }, { "epoch": 0.3544600938967136, "grad_norm": 0.34812483191490173, "learning_rate": 0.00013010625737898467, "loss": 0.8335, "step": 302 }, { "epoch": 0.35563380281690143, "grad_norm": 0.3655085861682892, "learning_rate": 0.00012987012987012987, "loss": 0.8117, "step": 303 }, { "epoch": 0.3568075117370892, "grad_norm": 0.35925915837287903, "learning_rate": 0.0001296340023612751, "loss": 0.8147, "step": 304 }, { "epoch": 0.357981220657277, "grad_norm": 0.3293222486972809, "learning_rate": 0.00012939787485242033, "loss": 0.7602, "step": 305 }, { "epoch": 0.3591549295774648, "grad_norm": 0.3486446738243103, "learning_rate": 0.00012916174734356554, "loss": 0.7857, "step": 306 }, { "epoch": 0.36032863849765256, "grad_norm": 0.382565975189209, "learning_rate": 0.00012892561983471074, "loss": 0.863, "step": 307 }, { "epoch": 0.3615023474178404, "grad_norm": 0.32544344663619995, "learning_rate": 0.00012868949232585597, "loss": 0.781, "step": 308 }, { "epoch": 0.3626760563380282, "grad_norm": 0.38700491189956665, "learning_rate": 0.00012845336481700118, "loss": 0.8102, "step": 309 }, { "epoch": 0.36384976525821594, "grad_norm": 0.3503759503364563, "learning_rate": 0.0001282172373081464, "loss": 0.7699, "step": 310 }, { "epoch": 0.36502347417840375, "grad_norm": 0.3323630094528198, "learning_rate": 0.00012798110979929164, "loss": 0.7511, "step": 311 }, { "epoch": 0.36619718309859156, "grad_norm": 0.3668995797634125, "learning_rate": 0.00012774498229043685, "loss": 0.7374, "step": 312 }, { "epoch": 0.3673708920187793, "grad_norm": 0.37373387813568115, "learning_rate": 0.00012750885478158205, "loss": 0.8077, "step": 313 }, { "epoch": 0.3685446009389671, "grad_norm": 0.3601135015487671, "learning_rate": 0.00012727272727272728, "loss": 0.7991, "step": 314 }, { "epoch": 0.36971830985915494, "grad_norm": 0.3527435064315796, "learning_rate": 0.00012703659976387249, "loss": 0.7971, "step": 315 }, { "epoch": 0.37089201877934275, "grad_norm": 0.3584372401237488, "learning_rate": 0.00012680047225501772, "loss": 0.7513, "step": 316 }, { "epoch": 0.3720657276995305, "grad_norm": 0.3517726957798004, "learning_rate": 0.00012656434474616295, "loss": 0.8206, "step": 317 }, { "epoch": 0.3732394366197183, "grad_norm": 0.3655302822589874, "learning_rate": 0.00012632821723730815, "loss": 0.771, "step": 318 }, { "epoch": 0.3744131455399061, "grad_norm": 0.3659893274307251, "learning_rate": 0.00012609208972845336, "loss": 0.8048, "step": 319 }, { "epoch": 0.3755868544600939, "grad_norm": 0.36364591121673584, "learning_rate": 0.0001258559622195986, "loss": 0.7832, "step": 320 }, { "epoch": 0.3767605633802817, "grad_norm": 0.37528395652770996, "learning_rate": 0.00012561983471074382, "loss": 0.7926, "step": 321 }, { "epoch": 0.3779342723004695, "grad_norm": 0.37137654423713684, "learning_rate": 0.00012538370720188903, "loss": 0.8486, "step": 322 }, { "epoch": 0.37910798122065725, "grad_norm": 0.3466728925704956, "learning_rate": 0.00012514757969303423, "loss": 0.7961, "step": 323 }, { "epoch": 0.38028169014084506, "grad_norm": 0.38629114627838135, "learning_rate": 0.00012491145218417946, "loss": 0.8071, "step": 324 }, { "epoch": 0.3814553990610329, "grad_norm": 0.34686383605003357, "learning_rate": 0.00012467532467532467, "loss": 0.7698, "step": 325 }, { "epoch": 0.3826291079812207, "grad_norm": 0.36625292897224426, "learning_rate": 0.0001244391971664699, "loss": 0.8486, "step": 326 }, { "epoch": 0.38380281690140844, "grad_norm": 0.38903650641441345, "learning_rate": 0.00012420306965761513, "loss": 0.8031, "step": 327 }, { "epoch": 0.38497652582159625, "grad_norm": 0.3456287980079651, "learning_rate": 0.00012396694214876033, "loss": 0.7887, "step": 328 }, { "epoch": 0.38615023474178406, "grad_norm": 0.36374613642692566, "learning_rate": 0.00012373081463990554, "loss": 0.7588, "step": 329 }, { "epoch": 0.3873239436619718, "grad_norm": 0.360626220703125, "learning_rate": 0.00012349468713105077, "loss": 0.8239, "step": 330 }, { "epoch": 0.3884976525821596, "grad_norm": 0.40213796496391296, "learning_rate": 0.000123258559622196, "loss": 0.8029, "step": 331 }, { "epoch": 0.38967136150234744, "grad_norm": 0.3273613750934601, "learning_rate": 0.0001230224321133412, "loss": 0.7567, "step": 332 }, { "epoch": 0.3908450704225352, "grad_norm": 0.34953057765960693, "learning_rate": 0.00012278630460448644, "loss": 0.7512, "step": 333 }, { "epoch": 0.392018779342723, "grad_norm": 0.34772762656211853, "learning_rate": 0.00012255017709563167, "loss": 0.7551, "step": 334 }, { "epoch": 0.3931924882629108, "grad_norm": 0.34170207381248474, "learning_rate": 0.00012231404958677685, "loss": 0.7884, "step": 335 }, { "epoch": 0.39436619718309857, "grad_norm": 0.3696103096008301, "learning_rate": 0.00012207792207792208, "loss": 0.8658, "step": 336 }, { "epoch": 0.3955399061032864, "grad_norm": 0.3513827621936798, "learning_rate": 0.00012184179456906731, "loss": 0.8199, "step": 337 }, { "epoch": 0.3967136150234742, "grad_norm": 0.3454856872558594, "learning_rate": 0.00012160566706021253, "loss": 0.7627, "step": 338 }, { "epoch": 0.397887323943662, "grad_norm": 0.3246639370918274, "learning_rate": 0.00012136953955135774, "loss": 0.7454, "step": 339 }, { "epoch": 0.39906103286384975, "grad_norm": 0.33567938208580017, "learning_rate": 0.00012113341204250295, "loss": 0.7611, "step": 340 }, { "epoch": 0.40023474178403756, "grad_norm": 0.33728334307670593, "learning_rate": 0.00012089728453364817, "loss": 0.7575, "step": 341 }, { "epoch": 0.4014084507042254, "grad_norm": 0.35161352157592773, "learning_rate": 0.0001206611570247934, "loss": 0.8117, "step": 342 }, { "epoch": 0.4025821596244131, "grad_norm": 0.3425585925579071, "learning_rate": 0.00012042502951593862, "loss": 0.8019, "step": 343 }, { "epoch": 0.40375586854460094, "grad_norm": 0.3406507968902588, "learning_rate": 0.00012018890200708383, "loss": 0.8235, "step": 344 }, { "epoch": 0.40492957746478875, "grad_norm": 0.37840309739112854, "learning_rate": 0.00011995277449822907, "loss": 0.7866, "step": 345 }, { "epoch": 0.4061032863849765, "grad_norm": 0.35816213488578796, "learning_rate": 0.00011971664698937426, "loss": 0.8425, "step": 346 }, { "epoch": 0.4072769953051643, "grad_norm": 0.3441546559333801, "learning_rate": 0.00011948051948051949, "loss": 0.8094, "step": 347 }, { "epoch": 0.4084507042253521, "grad_norm": 0.34275054931640625, "learning_rate": 0.0001192443919716647, "loss": 0.7244, "step": 348 }, { "epoch": 0.4096244131455399, "grad_norm": 0.33207401633262634, "learning_rate": 0.00011900826446280992, "loss": 0.8108, "step": 349 }, { "epoch": 0.4107981220657277, "grad_norm": 0.3412252962589264, "learning_rate": 0.00011877213695395516, "loss": 0.7818, "step": 350 }, { "epoch": 0.4119718309859155, "grad_norm": 0.36701643466949463, "learning_rate": 0.00011853600944510035, "loss": 0.8293, "step": 351 }, { "epoch": 0.4131455399061033, "grad_norm": 0.34462520480155945, "learning_rate": 0.00011829988193624558, "loss": 0.7603, "step": 352 }, { "epoch": 0.41431924882629106, "grad_norm": 0.35232508182525635, "learning_rate": 0.0001180637544273908, "loss": 0.7616, "step": 353 }, { "epoch": 0.4154929577464789, "grad_norm": 0.37428373098373413, "learning_rate": 0.00011782762691853601, "loss": 0.7919, "step": 354 }, { "epoch": 0.4166666666666667, "grad_norm": 0.3429507911205292, "learning_rate": 0.00011759149940968123, "loss": 0.7859, "step": 355 }, { "epoch": 0.41784037558685444, "grad_norm": 0.3584844470024109, "learning_rate": 0.00011735537190082646, "loss": 0.7934, "step": 356 }, { "epoch": 0.41901408450704225, "grad_norm": 0.356391578912735, "learning_rate": 0.00011711924439197165, "loss": 0.8222, "step": 357 }, { "epoch": 0.42018779342723006, "grad_norm": 0.3663417100906372, "learning_rate": 0.00011688311688311689, "loss": 0.7507, "step": 358 }, { "epoch": 0.4213615023474178, "grad_norm": 0.3388553559780121, "learning_rate": 0.0001166469893742621, "loss": 0.8263, "step": 359 }, { "epoch": 0.4225352112676056, "grad_norm": 0.34876593947410583, "learning_rate": 0.00011641086186540732, "loss": 0.7969, "step": 360 }, { "epoch": 0.42370892018779344, "grad_norm": 0.3500271737575531, "learning_rate": 0.00011617473435655255, "loss": 0.7789, "step": 361 }, { "epoch": 0.42488262910798125, "grad_norm": 0.3554798662662506, "learning_rate": 0.00011593860684769777, "loss": 0.7681, "step": 362 }, { "epoch": 0.426056338028169, "grad_norm": 0.34559762477874756, "learning_rate": 0.00011570247933884298, "loss": 0.7676, "step": 363 }, { "epoch": 0.4272300469483568, "grad_norm": 0.3520505726337433, "learning_rate": 0.0001154663518299882, "loss": 0.7494, "step": 364 }, { "epoch": 0.4284037558685446, "grad_norm": 0.35454803705215454, "learning_rate": 0.00011523022432113341, "loss": 0.7516, "step": 365 }, { "epoch": 0.4295774647887324, "grad_norm": 0.36526602506637573, "learning_rate": 0.00011499409681227864, "loss": 0.7789, "step": 366 }, { "epoch": 0.4307511737089202, "grad_norm": 0.34084445238113403, "learning_rate": 0.00011475796930342386, "loss": 0.7446, "step": 367 }, { "epoch": 0.431924882629108, "grad_norm": 0.3405500054359436, "learning_rate": 0.00011452184179456907, "loss": 0.8217, "step": 368 }, { "epoch": 0.43309859154929575, "grad_norm": 0.3523256182670593, "learning_rate": 0.00011428571428571428, "loss": 0.7311, "step": 369 }, { "epoch": 0.43427230046948356, "grad_norm": 0.3336530327796936, "learning_rate": 0.0001140495867768595, "loss": 0.7806, "step": 370 }, { "epoch": 0.4354460093896714, "grad_norm": 0.3268769383430481, "learning_rate": 0.00011381345926800473, "loss": 0.7945, "step": 371 }, { "epoch": 0.43661971830985913, "grad_norm": 0.35258617997169495, "learning_rate": 0.00011357733175914995, "loss": 0.7468, "step": 372 }, { "epoch": 0.43779342723004694, "grad_norm": 0.3546913266181946, "learning_rate": 0.00011334120425029517, "loss": 0.7921, "step": 373 }, { "epoch": 0.43896713615023475, "grad_norm": 0.36266180872917175, "learning_rate": 0.00011310507674144037, "loss": 0.7623, "step": 374 }, { "epoch": 0.44014084507042256, "grad_norm": 0.3355543613433838, "learning_rate": 0.00011286894923258559, "loss": 0.7436, "step": 375 }, { "epoch": 0.4413145539906103, "grad_norm": 0.33666127920150757, "learning_rate": 0.00011263282172373082, "loss": 0.7609, "step": 376 }, { "epoch": 0.4424882629107981, "grad_norm": 0.3505670428276062, "learning_rate": 0.00011239669421487604, "loss": 0.7868, "step": 377 }, { "epoch": 0.44366197183098594, "grad_norm": 0.3446255028247833, "learning_rate": 0.00011216056670602126, "loss": 0.765, "step": 378 }, { "epoch": 0.4448356807511737, "grad_norm": 0.3761040270328522, "learning_rate": 0.00011192443919716649, "loss": 0.8104, "step": 379 }, { "epoch": 0.4460093896713615, "grad_norm": 0.35692986845970154, "learning_rate": 0.00011168831168831168, "loss": 0.7896, "step": 380 }, { "epoch": 0.4471830985915493, "grad_norm": 0.34384050965309143, "learning_rate": 0.00011145218417945691, "loss": 0.7716, "step": 381 }, { "epoch": 0.44835680751173707, "grad_norm": 0.3477395176887512, "learning_rate": 0.00011121605667060213, "loss": 0.8146, "step": 382 }, { "epoch": 0.4495305164319249, "grad_norm": 0.35172998905181885, "learning_rate": 0.00011097992916174735, "loss": 0.7844, "step": 383 }, { "epoch": 0.4507042253521127, "grad_norm": 0.33881857991218567, "learning_rate": 0.00011074380165289258, "loss": 0.7528, "step": 384 }, { "epoch": 0.4518779342723005, "grad_norm": 0.3429534137248993, "learning_rate": 0.00011050767414403777, "loss": 0.7826, "step": 385 }, { "epoch": 0.45305164319248825, "grad_norm": 0.34472665190696716, "learning_rate": 0.000110271546635183, "loss": 0.7153, "step": 386 }, { "epoch": 0.45422535211267606, "grad_norm": 0.3572479486465454, "learning_rate": 0.00011003541912632822, "loss": 0.7811, "step": 387 }, { "epoch": 0.45539906103286387, "grad_norm": 0.3531682789325714, "learning_rate": 0.00010979929161747344, "loss": 0.8016, "step": 388 }, { "epoch": 0.4565727699530516, "grad_norm": 0.3845299780368805, "learning_rate": 0.00010956316410861867, "loss": 0.7817, "step": 389 }, { "epoch": 0.45774647887323944, "grad_norm": 0.35217660665512085, "learning_rate": 0.00010932703659976389, "loss": 0.7495, "step": 390 }, { "epoch": 0.45892018779342725, "grad_norm": 0.35103702545166016, "learning_rate": 0.00010909090909090909, "loss": 0.7602, "step": 391 }, { "epoch": 0.460093896713615, "grad_norm": 0.3511259853839874, "learning_rate": 0.00010885478158205431, "loss": 0.7923, "step": 392 }, { "epoch": 0.4612676056338028, "grad_norm": 0.33732983469963074, "learning_rate": 0.00010861865407319953, "loss": 0.7875, "step": 393 }, { "epoch": 0.4624413145539906, "grad_norm": 0.35035955905914307, "learning_rate": 0.00010838252656434476, "loss": 0.7737, "step": 394 }, { "epoch": 0.4636150234741784, "grad_norm": 0.3277076482772827, "learning_rate": 0.00010814639905548998, "loss": 0.7619, "step": 395 }, { "epoch": 0.4647887323943662, "grad_norm": 0.34461456537246704, "learning_rate": 0.00010791027154663518, "loss": 0.7394, "step": 396 }, { "epoch": 0.465962441314554, "grad_norm": 0.36000820994377136, "learning_rate": 0.0001076741440377804, "loss": 0.8004, "step": 397 }, { "epoch": 0.4671361502347418, "grad_norm": 0.3291054666042328, "learning_rate": 0.00010743801652892562, "loss": 0.721, "step": 398 }, { "epoch": 0.46830985915492956, "grad_norm": 0.37541574239730835, "learning_rate": 0.00010720188902007085, "loss": 0.7673, "step": 399 }, { "epoch": 0.4694835680751174, "grad_norm": 0.33268067240715027, "learning_rate": 0.00010696576151121607, "loss": 0.7439, "step": 400 }, { "epoch": 0.4706572769953052, "grad_norm": 0.34383484721183777, "learning_rate": 0.00010672963400236129, "loss": 0.7453, "step": 401 }, { "epoch": 0.47183098591549294, "grad_norm": 0.3543702960014343, "learning_rate": 0.00010649350649350649, "loss": 0.7544, "step": 402 }, { "epoch": 0.47300469483568075, "grad_norm": 0.34553685784339905, "learning_rate": 0.00010625737898465171, "loss": 0.7656, "step": 403 }, { "epoch": 0.47417840375586856, "grad_norm": 0.3437071144580841, "learning_rate": 0.00010602125147579694, "loss": 0.773, "step": 404 }, { "epoch": 0.4753521126760563, "grad_norm": 0.34917253255844116, "learning_rate": 0.00010578512396694216, "loss": 0.7607, "step": 405 }, { "epoch": 0.4765258215962441, "grad_norm": 0.33429262042045593, "learning_rate": 0.00010554899645808738, "loss": 0.768, "step": 406 }, { "epoch": 0.47769953051643194, "grad_norm": 0.33842045068740845, "learning_rate": 0.00010531286894923261, "loss": 0.7665, "step": 407 }, { "epoch": 0.4788732394366197, "grad_norm": 0.3419265151023865, "learning_rate": 0.0001050767414403778, "loss": 0.7717, "step": 408 }, { "epoch": 0.4800469483568075, "grad_norm": 0.3458483815193176, "learning_rate": 0.00010484061393152303, "loss": 0.8031, "step": 409 }, { "epoch": 0.4812206572769953, "grad_norm": 0.37077274918556213, "learning_rate": 0.00010460448642266825, "loss": 0.8009, "step": 410 }, { "epoch": 0.4823943661971831, "grad_norm": 0.35040315985679626, "learning_rate": 0.00010436835891381347, "loss": 0.7545, "step": 411 }, { "epoch": 0.4835680751173709, "grad_norm": 0.3503456115722656, "learning_rate": 0.0001041322314049587, "loss": 0.8515, "step": 412 }, { "epoch": 0.4847417840375587, "grad_norm": 0.34627342224121094, "learning_rate": 0.00010389610389610389, "loss": 0.716, "step": 413 }, { "epoch": 0.4859154929577465, "grad_norm": 0.3596992790699005, "learning_rate": 0.00010365997638724912, "loss": 0.7636, "step": 414 }, { "epoch": 0.48708920187793425, "grad_norm": 0.3346829116344452, "learning_rate": 0.00010342384887839434, "loss": 0.7635, "step": 415 }, { "epoch": 0.48826291079812206, "grad_norm": 0.37179237604141235, "learning_rate": 0.00010318772136953956, "loss": 0.7642, "step": 416 }, { "epoch": 0.4894366197183099, "grad_norm": 0.34897381067276, "learning_rate": 0.00010295159386068479, "loss": 0.7792, "step": 417 }, { "epoch": 0.49061032863849763, "grad_norm": 0.3820830285549164, "learning_rate": 0.00010271546635183, "loss": 0.7722, "step": 418 }, { "epoch": 0.49178403755868544, "grad_norm": 0.3688552677631378, "learning_rate": 0.00010247933884297521, "loss": 0.7927, "step": 419 }, { "epoch": 0.49295774647887325, "grad_norm": 0.35100415349006653, "learning_rate": 0.00010224321133412043, "loss": 0.7848, "step": 420 }, { "epoch": 0.49413145539906106, "grad_norm": 0.3596225082874298, "learning_rate": 0.00010200708382526565, "loss": 0.7383, "step": 421 }, { "epoch": 0.4953051643192488, "grad_norm": 0.36203423142433167, "learning_rate": 0.00010177095631641088, "loss": 0.769, "step": 422 }, { "epoch": 0.4964788732394366, "grad_norm": 0.3776590824127197, "learning_rate": 0.0001015348288075561, "loss": 0.8007, "step": 423 }, { "epoch": 0.49765258215962443, "grad_norm": 0.36009421944618225, "learning_rate": 0.0001012987012987013, "loss": 0.7557, "step": 424 }, { "epoch": 0.4988262910798122, "grad_norm": 0.3442706763744354, "learning_rate": 0.00010106257378984652, "loss": 0.7488, "step": 425 }, { "epoch": 0.5, "grad_norm": 0.3635407090187073, "learning_rate": 0.00010082644628099174, "loss": 0.7922, "step": 426 }, { "epoch": 0.5011737089201878, "grad_norm": 0.3766370117664337, "learning_rate": 0.00010059031877213697, "loss": 0.7818, "step": 427 }, { "epoch": 0.5023474178403756, "grad_norm": 0.34344202280044556, "learning_rate": 0.00010035419126328218, "loss": 0.8308, "step": 428 }, { "epoch": 0.5035211267605634, "grad_norm": 0.3495674133300781, "learning_rate": 0.0001001180637544274, "loss": 0.799, "step": 429 }, { "epoch": 0.5046948356807511, "grad_norm": 0.36545464396476746, "learning_rate": 9.988193624557262e-05, "loss": 0.7453, "step": 430 }, { "epoch": 0.505868544600939, "grad_norm": 0.3482630252838135, "learning_rate": 9.964580873671782e-05, "loss": 0.7422, "step": 431 }, { "epoch": 0.5070422535211268, "grad_norm": 0.3745418190956116, "learning_rate": 9.940968122786304e-05, "loss": 0.7333, "step": 432 }, { "epoch": 0.5082159624413145, "grad_norm": 0.3470025062561035, "learning_rate": 9.917355371900827e-05, "loss": 0.7907, "step": 433 }, { "epoch": 0.5093896713615024, "grad_norm": 0.38251325488090515, "learning_rate": 9.893742621015348e-05, "loss": 0.7629, "step": 434 }, { "epoch": 0.5105633802816901, "grad_norm": 0.3829626739025116, "learning_rate": 9.870129870129871e-05, "loss": 0.7939, "step": 435 }, { "epoch": 0.5117370892018779, "grad_norm": 0.35726287961006165, "learning_rate": 9.846517119244393e-05, "loss": 0.755, "step": 436 }, { "epoch": 0.5129107981220657, "grad_norm": 0.38168108463287354, "learning_rate": 9.822904368358913e-05, "loss": 0.7396, "step": 437 }, { "epoch": 0.5140845070422535, "grad_norm": 0.35728660225868225, "learning_rate": 9.799291617473436e-05, "loss": 0.7568, "step": 438 }, { "epoch": 0.5152582159624414, "grad_norm": 0.37819668650627136, "learning_rate": 9.775678866587958e-05, "loss": 0.8046, "step": 439 }, { "epoch": 0.5164319248826291, "grad_norm": 0.4106784760951996, "learning_rate": 9.75206611570248e-05, "loss": 0.7116, "step": 440 }, { "epoch": 0.5176056338028169, "grad_norm": 0.3476578891277313, "learning_rate": 9.728453364817002e-05, "loss": 0.7824, "step": 441 }, { "epoch": 0.5187793427230047, "grad_norm": 0.36705800890922546, "learning_rate": 9.704840613931524e-05, "loss": 0.7631, "step": 442 }, { "epoch": 0.5199530516431925, "grad_norm": 0.3880864977836609, "learning_rate": 9.681227863046045e-05, "loss": 0.7608, "step": 443 }, { "epoch": 0.5211267605633803, "grad_norm": 0.3610959053039551, "learning_rate": 9.657615112160567e-05, "loss": 0.7909, "step": 444 }, { "epoch": 0.5223004694835681, "grad_norm": 0.33494657278060913, "learning_rate": 9.634002361275089e-05, "loss": 0.7108, "step": 445 }, { "epoch": 0.5234741784037559, "grad_norm": 0.352055162191391, "learning_rate": 9.610389610389611e-05, "loss": 0.7177, "step": 446 }, { "epoch": 0.5246478873239436, "grad_norm": 0.35466742515563965, "learning_rate": 9.586776859504133e-05, "loss": 0.7762, "step": 447 }, { "epoch": 0.5258215962441315, "grad_norm": 0.34477657079696655, "learning_rate": 9.563164108618654e-05, "loss": 0.7583, "step": 448 }, { "epoch": 0.5269953051643192, "grad_norm": 0.37008315324783325, "learning_rate": 9.539551357733176e-05, "loss": 0.7954, "step": 449 }, { "epoch": 0.528169014084507, "grad_norm": 0.34141793847084045, "learning_rate": 9.515938606847698e-05, "loss": 0.7444, "step": 450 }, { "epoch": 0.5293427230046949, "grad_norm": 0.3429400622844696, "learning_rate": 9.49232585596222e-05, "loss": 0.7499, "step": 451 }, { "epoch": 0.5305164319248826, "grad_norm": 0.3666730225086212, "learning_rate": 9.468713105076742e-05, "loss": 0.7704, "step": 452 }, { "epoch": 0.5316901408450704, "grad_norm": 0.34185874462127686, "learning_rate": 9.445100354191265e-05, "loss": 0.7446, "step": 453 }, { "epoch": 0.5328638497652582, "grad_norm": 0.3718375861644745, "learning_rate": 9.421487603305785e-05, "loss": 0.7316, "step": 454 }, { "epoch": 0.534037558685446, "grad_norm": 0.35064697265625, "learning_rate": 9.397874852420307e-05, "loss": 0.7651, "step": 455 }, { "epoch": 0.5352112676056338, "grad_norm": 0.3724139630794525, "learning_rate": 9.37426210153483e-05, "loss": 0.7639, "step": 456 }, { "epoch": 0.5363849765258216, "grad_norm": 0.3420800566673279, "learning_rate": 9.35064935064935e-05, "loss": 0.7578, "step": 457 }, { "epoch": 0.5375586854460094, "grad_norm": 0.3437943160533905, "learning_rate": 9.327036599763874e-05, "loss": 0.7898, "step": 458 }, { "epoch": 0.5387323943661971, "grad_norm": 0.3799413740634918, "learning_rate": 9.303423848878394e-05, "loss": 0.7216, "step": 459 }, { "epoch": 0.539906103286385, "grad_norm": 0.35702013969421387, "learning_rate": 9.279811097992916e-05, "loss": 0.7509, "step": 460 }, { "epoch": 0.5410798122065728, "grad_norm": 0.36074140667915344, "learning_rate": 9.256198347107439e-05, "loss": 0.7448, "step": 461 }, { "epoch": 0.5422535211267606, "grad_norm": 0.34211182594299316, "learning_rate": 9.23258559622196e-05, "loss": 0.7143, "step": 462 }, { "epoch": 0.5434272300469484, "grad_norm": 0.3816893398761749, "learning_rate": 9.208972845336483e-05, "loss": 0.7178, "step": 463 }, { "epoch": 0.5446009389671361, "grad_norm": 0.36033767461776733, "learning_rate": 9.185360094451005e-05, "loss": 0.7406, "step": 464 }, { "epoch": 0.545774647887324, "grad_norm": 0.38050010800361633, "learning_rate": 9.161747343565525e-05, "loss": 0.7528, "step": 465 }, { "epoch": 0.5469483568075117, "grad_norm": 0.3648395240306854, "learning_rate": 9.138134592680048e-05, "loss": 0.7802, "step": 466 }, { "epoch": 0.5481220657276995, "grad_norm": 0.35185542702674866, "learning_rate": 9.11452184179457e-05, "loss": 0.7489, "step": 467 }, { "epoch": 0.5492957746478874, "grad_norm": 0.3487717807292938, "learning_rate": 9.090909090909092e-05, "loss": 0.7742, "step": 468 }, { "epoch": 0.5504694835680751, "grad_norm": 0.36121654510498047, "learning_rate": 9.067296340023614e-05, "loss": 0.7974, "step": 469 }, { "epoch": 0.5516431924882629, "grad_norm": 0.3470339775085449, "learning_rate": 9.043683589138135e-05, "loss": 0.723, "step": 470 }, { "epoch": 0.5528169014084507, "grad_norm": 0.33549764752388, "learning_rate": 9.020070838252657e-05, "loss": 0.7334, "step": 471 }, { "epoch": 0.5539906103286385, "grad_norm": 0.36101868748664856, "learning_rate": 8.996458087367179e-05, "loss": 0.6817, "step": 472 }, { "epoch": 0.5551643192488263, "grad_norm": 0.36847153306007385, "learning_rate": 8.9728453364817e-05, "loss": 0.7942, "step": 473 }, { "epoch": 0.5563380281690141, "grad_norm": 0.3564891815185547, "learning_rate": 8.949232585596222e-05, "loss": 0.7071, "step": 474 }, { "epoch": 0.5575117370892019, "grad_norm": 0.36866652965545654, "learning_rate": 8.925619834710744e-05, "loss": 0.7685, "step": 475 }, { "epoch": 0.5586854460093896, "grad_norm": 0.370924711227417, "learning_rate": 8.902007083825266e-05, "loss": 0.7313, "step": 476 }, { "epoch": 0.5598591549295775, "grad_norm": 0.3611142039299011, "learning_rate": 8.878394332939788e-05, "loss": 0.7666, "step": 477 }, { "epoch": 0.5610328638497653, "grad_norm": 0.3418121635913849, "learning_rate": 8.85478158205431e-05, "loss": 0.7194, "step": 478 }, { "epoch": 0.562206572769953, "grad_norm": 0.3478650748729706, "learning_rate": 8.831168831168831e-05, "loss": 0.7145, "step": 479 }, { "epoch": 0.5633802816901409, "grad_norm": 0.3567008078098297, "learning_rate": 8.807556080283353e-05, "loss": 0.7591, "step": 480 }, { "epoch": 0.5645539906103286, "grad_norm": 0.3629607558250427, "learning_rate": 8.783943329397875e-05, "loss": 0.7856, "step": 481 }, { "epoch": 0.5657276995305164, "grad_norm": 0.37257978320121765, "learning_rate": 8.760330578512397e-05, "loss": 0.709, "step": 482 }, { "epoch": 0.5669014084507042, "grad_norm": 0.3570626676082611, "learning_rate": 8.736717827626919e-05, "loss": 0.7639, "step": 483 }, { "epoch": 0.568075117370892, "grad_norm": 0.34790506958961487, "learning_rate": 8.713105076741442e-05, "loss": 0.7375, "step": 484 }, { "epoch": 0.5692488262910798, "grad_norm": 0.3525756895542145, "learning_rate": 8.689492325855962e-05, "loss": 0.7274, "step": 485 }, { "epoch": 0.5704225352112676, "grad_norm": 0.3545394837856293, "learning_rate": 8.665879574970484e-05, "loss": 0.7531, "step": 486 }, { "epoch": 0.5715962441314554, "grad_norm": 0.35677066445350647, "learning_rate": 8.642266824085006e-05, "loss": 0.7682, "step": 487 }, { "epoch": 0.5727699530516432, "grad_norm": 0.3439461290836334, "learning_rate": 8.618654073199528e-05, "loss": 0.7176, "step": 488 }, { "epoch": 0.573943661971831, "grad_norm": 0.3622515797615051, "learning_rate": 8.595041322314051e-05, "loss": 0.7004, "step": 489 }, { "epoch": 0.5751173708920188, "grad_norm": 0.36056646704673767, "learning_rate": 8.571428571428571e-05, "loss": 0.74, "step": 490 }, { "epoch": 0.5762910798122066, "grad_norm": 0.3509630262851715, "learning_rate": 8.547815820543093e-05, "loss": 0.8006, "step": 491 }, { "epoch": 0.5774647887323944, "grad_norm": 0.3422422707080841, "learning_rate": 8.524203069657616e-05, "loss": 0.7162, "step": 492 }, { "epoch": 0.5786384976525821, "grad_norm": 0.35553744435310364, "learning_rate": 8.500590318772137e-05, "loss": 0.7554, "step": 493 }, { "epoch": 0.57981220657277, "grad_norm": 0.3443603813648224, "learning_rate": 8.47697756788666e-05, "loss": 0.7128, "step": 494 }, { "epoch": 0.5809859154929577, "grad_norm": 0.3314555883407593, "learning_rate": 8.453364817001182e-05, "loss": 0.7123, "step": 495 }, { "epoch": 0.5821596244131455, "grad_norm": 0.33951112627983093, "learning_rate": 8.429752066115702e-05, "loss": 0.7501, "step": 496 }, { "epoch": 0.5833333333333334, "grad_norm": 0.327809602022171, "learning_rate": 8.406139315230225e-05, "loss": 0.7543, "step": 497 }, { "epoch": 0.5845070422535211, "grad_norm": 0.33205023407936096, "learning_rate": 8.382526564344747e-05, "loss": 0.7395, "step": 498 }, { "epoch": 0.5856807511737089, "grad_norm": 0.3762659430503845, "learning_rate": 8.358913813459269e-05, "loss": 0.7424, "step": 499 }, { "epoch": 0.5868544600938967, "grad_norm": 0.3421575427055359, "learning_rate": 8.33530106257379e-05, "loss": 0.7167, "step": 500 }, { "epoch": 0.5880281690140845, "grad_norm": 0.3560996353626251, "learning_rate": 8.311688311688312e-05, "loss": 0.7464, "step": 501 }, { "epoch": 0.5892018779342723, "grad_norm": 0.3566039800643921, "learning_rate": 8.288075560802834e-05, "loss": 0.715, "step": 502 }, { "epoch": 0.5903755868544601, "grad_norm": 0.3481593430042267, "learning_rate": 8.264462809917356e-05, "loss": 0.7506, "step": 503 }, { "epoch": 0.5915492957746479, "grad_norm": 0.34428590536117554, "learning_rate": 8.240850059031878e-05, "loss": 0.7272, "step": 504 }, { "epoch": 0.5927230046948356, "grad_norm": 0.35629555583000183, "learning_rate": 8.2172373081464e-05, "loss": 0.7334, "step": 505 }, { "epoch": 0.5938967136150235, "grad_norm": 0.37292811274528503, "learning_rate": 8.193624557260921e-05, "loss": 0.7505, "step": 506 }, { "epoch": 0.5950704225352113, "grad_norm": 0.359614759683609, "learning_rate": 8.170011806375443e-05, "loss": 0.8006, "step": 507 }, { "epoch": 0.596244131455399, "grad_norm": 0.3388945460319519, "learning_rate": 8.146399055489965e-05, "loss": 0.7542, "step": 508 }, { "epoch": 0.5974178403755869, "grad_norm": 0.3528054356575012, "learning_rate": 8.122786304604487e-05, "loss": 0.7412, "step": 509 }, { "epoch": 0.5985915492957746, "grad_norm": 0.3354608416557312, "learning_rate": 8.099173553719009e-05, "loss": 0.7062, "step": 510 }, { "epoch": 0.5997652582159625, "grad_norm": 0.35168859362602234, "learning_rate": 8.07556080283353e-05, "loss": 0.7653, "step": 511 }, { "epoch": 0.6009389671361502, "grad_norm": 0.33843398094177246, "learning_rate": 8.051948051948052e-05, "loss": 0.7339, "step": 512 }, { "epoch": 0.602112676056338, "grad_norm": 0.32910212874412537, "learning_rate": 8.028335301062574e-05, "loss": 0.6966, "step": 513 }, { "epoch": 0.6032863849765259, "grad_norm": 0.3462936580181122, "learning_rate": 8.004722550177096e-05, "loss": 0.7386, "step": 514 }, { "epoch": 0.6044600938967136, "grad_norm": 0.3483426868915558, "learning_rate": 7.981109799291619e-05, "loss": 0.7548, "step": 515 }, { "epoch": 0.6056338028169014, "grad_norm": 0.3555918335914612, "learning_rate": 7.95749704840614e-05, "loss": 0.7144, "step": 516 }, { "epoch": 0.6068075117370892, "grad_norm": 0.3545628786087036, "learning_rate": 7.933884297520661e-05, "loss": 0.7601, "step": 517 }, { "epoch": 0.607981220657277, "grad_norm": 0.3554907441139221, "learning_rate": 7.910271546635183e-05, "loss": 0.7464, "step": 518 }, { "epoch": 0.6091549295774648, "grad_norm": 0.3457619547843933, "learning_rate": 7.886658795749705e-05, "loss": 0.7372, "step": 519 }, { "epoch": 0.6103286384976526, "grad_norm": 0.3450148105621338, "learning_rate": 7.863046044864228e-05, "loss": 0.7265, "step": 520 }, { "epoch": 0.6115023474178404, "grad_norm": 0.3475225567817688, "learning_rate": 7.839433293978748e-05, "loss": 0.798, "step": 521 }, { "epoch": 0.6126760563380281, "grad_norm": 0.34560921788215637, "learning_rate": 7.81582054309327e-05, "loss": 0.7583, "step": 522 }, { "epoch": 0.613849765258216, "grad_norm": 0.33480820059776306, "learning_rate": 7.792207792207793e-05, "loss": 0.7658, "step": 523 }, { "epoch": 0.6150234741784038, "grad_norm": 0.34581395983695984, "learning_rate": 7.768595041322314e-05, "loss": 0.7368, "step": 524 }, { "epoch": 0.6161971830985915, "grad_norm": 0.35383906960487366, "learning_rate": 7.744982290436837e-05, "loss": 0.7963, "step": 525 }, { "epoch": 0.6173708920187794, "grad_norm": 0.352117121219635, "learning_rate": 7.721369539551359e-05, "loss": 0.7589, "step": 526 }, { "epoch": 0.6185446009389671, "grad_norm": 0.34420257806777954, "learning_rate": 7.697756788665879e-05, "loss": 0.7209, "step": 527 }, { "epoch": 0.6197183098591549, "grad_norm": 0.3449562191963196, "learning_rate": 7.674144037780402e-05, "loss": 0.7526, "step": 528 }, { "epoch": 0.6208920187793427, "grad_norm": 0.37377694249153137, "learning_rate": 7.650531286894924e-05, "loss": 0.7348, "step": 529 }, { "epoch": 0.6220657276995305, "grad_norm": 0.32662031054496765, "learning_rate": 7.626918536009446e-05, "loss": 0.7125, "step": 530 }, { "epoch": 0.6232394366197183, "grad_norm": 0.3551415801048279, "learning_rate": 7.603305785123968e-05, "loss": 0.7497, "step": 531 }, { "epoch": 0.6244131455399061, "grad_norm": 0.3519802689552307, "learning_rate": 7.579693034238488e-05, "loss": 0.7864, "step": 532 }, { "epoch": 0.6255868544600939, "grad_norm": 0.3773750364780426, "learning_rate": 7.556080283353011e-05, "loss": 0.7681, "step": 533 }, { "epoch": 0.6267605633802817, "grad_norm": 0.3558037281036377, "learning_rate": 7.532467532467533e-05, "loss": 0.7392, "step": 534 }, { "epoch": 0.6279342723004695, "grad_norm": 0.33910447359085083, "learning_rate": 7.508854781582055e-05, "loss": 0.7036, "step": 535 }, { "epoch": 0.6291079812206573, "grad_norm": 0.35620275139808655, "learning_rate": 7.485242030696577e-05, "loss": 0.7272, "step": 536 }, { "epoch": 0.6302816901408451, "grad_norm": 0.3377542495727539, "learning_rate": 7.461629279811098e-05, "loss": 0.7244, "step": 537 }, { "epoch": 0.6314553990610329, "grad_norm": 0.35217198729515076, "learning_rate": 7.43801652892562e-05, "loss": 0.7655, "step": 538 }, { "epoch": 0.6326291079812206, "grad_norm": 0.34656718373298645, "learning_rate": 7.414403778040142e-05, "loss": 0.7474, "step": 539 }, { "epoch": 0.6338028169014085, "grad_norm": 0.34429579973220825, "learning_rate": 7.390791027154664e-05, "loss": 0.7333, "step": 540 }, { "epoch": 0.6349765258215962, "grad_norm": 0.374262273311615, "learning_rate": 7.367178276269186e-05, "loss": 0.7876, "step": 541 }, { "epoch": 0.636150234741784, "grad_norm": 0.363299161195755, "learning_rate": 7.343565525383707e-05, "loss": 0.7784, "step": 542 }, { "epoch": 0.6373239436619719, "grad_norm": 0.36767125129699707, "learning_rate": 7.31995277449823e-05, "loss": 0.7329, "step": 543 }, { "epoch": 0.6384976525821596, "grad_norm": 0.3338686525821686, "learning_rate": 7.296340023612751e-05, "loss": 0.7737, "step": 544 }, { "epoch": 0.6396713615023474, "grad_norm": 0.3493046164512634, "learning_rate": 7.272727272727273e-05, "loss": 0.7461, "step": 545 }, { "epoch": 0.6408450704225352, "grad_norm": 0.3691573441028595, "learning_rate": 7.249114521841795e-05, "loss": 0.765, "step": 546 }, { "epoch": 0.642018779342723, "grad_norm": 0.3573099374771118, "learning_rate": 7.225501770956316e-05, "loss": 0.7589, "step": 547 }, { "epoch": 0.6431924882629108, "grad_norm": 0.36218926310539246, "learning_rate": 7.201889020070838e-05, "loss": 0.7314, "step": 548 }, { "epoch": 0.6443661971830986, "grad_norm": 0.35753628611564636, "learning_rate": 7.17827626918536e-05, "loss": 0.7564, "step": 549 }, { "epoch": 0.6455399061032864, "grad_norm": 0.3394756615161896, "learning_rate": 7.154663518299882e-05, "loss": 0.7162, "step": 550 }, { "epoch": 0.6467136150234741, "grad_norm": 0.350090891122818, "learning_rate": 7.131050767414405e-05, "loss": 0.7561, "step": 551 }, { "epoch": 0.647887323943662, "grad_norm": 0.328924298286438, "learning_rate": 7.107438016528925e-05, "loss": 0.7143, "step": 552 }, { "epoch": 0.6490610328638498, "grad_norm": 0.3552818298339844, "learning_rate": 7.083825265643447e-05, "loss": 0.7264, "step": 553 }, { "epoch": 0.6502347417840375, "grad_norm": 0.3504960536956787, "learning_rate": 7.06021251475797e-05, "loss": 0.7512, "step": 554 }, { "epoch": 0.6514084507042254, "grad_norm": 0.33755823969841003, "learning_rate": 7.036599763872491e-05, "loss": 0.7621, "step": 555 }, { "epoch": 0.6525821596244131, "grad_norm": 0.35977354645729065, "learning_rate": 7.012987012987014e-05, "loss": 0.776, "step": 556 }, { "epoch": 0.653755868544601, "grad_norm": 0.37304726243019104, "learning_rate": 6.989374262101536e-05, "loss": 0.7601, "step": 557 }, { "epoch": 0.6549295774647887, "grad_norm": 0.3569071590900421, "learning_rate": 6.965761511216056e-05, "loss": 0.7303, "step": 558 }, { "epoch": 0.6561032863849765, "grad_norm": 0.348264217376709, "learning_rate": 6.94214876033058e-05, "loss": 0.759, "step": 559 }, { "epoch": 0.6572769953051644, "grad_norm": 0.3501366674900055, "learning_rate": 6.9185360094451e-05, "loss": 0.7588, "step": 560 }, { "epoch": 0.6584507042253521, "grad_norm": 0.3633224666118622, "learning_rate": 6.894923258559623e-05, "loss": 0.7741, "step": 561 }, { "epoch": 0.6596244131455399, "grad_norm": 0.35944506525993347, "learning_rate": 6.871310507674145e-05, "loss": 0.756, "step": 562 }, { "epoch": 0.6607981220657277, "grad_norm": 0.3479359745979309, "learning_rate": 6.847697756788665e-05, "loss": 0.7292, "step": 563 }, { "epoch": 0.6619718309859155, "grad_norm": 0.37013959884643555, "learning_rate": 6.824085005903188e-05, "loss": 0.7618, "step": 564 }, { "epoch": 0.6631455399061033, "grad_norm": 0.36679190397262573, "learning_rate": 6.80047225501771e-05, "loss": 0.7797, "step": 565 }, { "epoch": 0.6643192488262911, "grad_norm": 0.35092490911483765, "learning_rate": 6.776859504132232e-05, "loss": 0.705, "step": 566 }, { "epoch": 0.6654929577464789, "grad_norm": 0.3594275712966919, "learning_rate": 6.753246753246754e-05, "loss": 0.7215, "step": 567 }, { "epoch": 0.6666666666666666, "grad_norm": 0.3503059148788452, "learning_rate": 6.729634002361276e-05, "loss": 0.7248, "step": 568 }, { "epoch": 0.6678403755868545, "grad_norm": 0.35919633507728577, "learning_rate": 6.706021251475797e-05, "loss": 0.7718, "step": 569 }, { "epoch": 0.6690140845070423, "grad_norm": 0.36752262711524963, "learning_rate": 6.682408500590319e-05, "loss": 0.7738, "step": 570 }, { "epoch": 0.67018779342723, "grad_norm": 0.33812567591667175, "learning_rate": 6.658795749704841e-05, "loss": 0.7846, "step": 571 }, { "epoch": 0.6713615023474179, "grad_norm": 0.3429810404777527, "learning_rate": 6.635182998819363e-05, "loss": 0.7371, "step": 572 }, { "epoch": 0.6725352112676056, "grad_norm": 0.3457571864128113, "learning_rate": 6.611570247933885e-05, "loss": 0.7318, "step": 573 }, { "epoch": 0.6737089201877934, "grad_norm": 0.3476294279098511, "learning_rate": 6.587957497048406e-05, "loss": 0.7344, "step": 574 }, { "epoch": 0.6748826291079812, "grad_norm": 0.34464409947395325, "learning_rate": 6.564344746162928e-05, "loss": 0.7429, "step": 575 }, { "epoch": 0.676056338028169, "grad_norm": 0.34444373846054077, "learning_rate": 6.54073199527745e-05, "loss": 0.7663, "step": 576 }, { "epoch": 0.6772300469483568, "grad_norm": 0.3656728267669678, "learning_rate": 6.517119244391972e-05, "loss": 0.7068, "step": 577 }, { "epoch": 0.6784037558685446, "grad_norm": 0.3591727614402771, "learning_rate": 6.493506493506494e-05, "loss": 0.7481, "step": 578 }, { "epoch": 0.6795774647887324, "grad_norm": 0.38865676522254944, "learning_rate": 6.469893742621017e-05, "loss": 0.7659, "step": 579 }, { "epoch": 0.6807511737089202, "grad_norm": 0.3438194990158081, "learning_rate": 6.446280991735537e-05, "loss": 0.6748, "step": 580 }, { "epoch": 0.681924882629108, "grad_norm": 0.34979990124702454, "learning_rate": 6.422668240850059e-05, "loss": 0.7529, "step": 581 }, { "epoch": 0.6830985915492958, "grad_norm": 0.37309062480926514, "learning_rate": 6.399055489964582e-05, "loss": 0.7417, "step": 582 }, { "epoch": 0.6842723004694836, "grad_norm": 0.3737837076187134, "learning_rate": 6.375442739079102e-05, "loss": 0.773, "step": 583 }, { "epoch": 0.6854460093896714, "grad_norm": 0.3397013247013092, "learning_rate": 6.351829988193624e-05, "loss": 0.7093, "step": 584 }, { "epoch": 0.6866197183098591, "grad_norm": 0.37165701389312744, "learning_rate": 6.328217237308147e-05, "loss": 0.7078, "step": 585 }, { "epoch": 0.687793427230047, "grad_norm": 0.3533116579055786, "learning_rate": 6.304604486422668e-05, "loss": 0.7105, "step": 586 }, { "epoch": 0.6889671361502347, "grad_norm": 0.35352569818496704, "learning_rate": 6.280991735537191e-05, "loss": 0.7282, "step": 587 }, { "epoch": 0.6901408450704225, "grad_norm": 0.3754810094833374, "learning_rate": 6.257378984651711e-05, "loss": 0.7364, "step": 588 }, { "epoch": 0.6913145539906104, "grad_norm": 0.36235493421554565, "learning_rate": 6.233766233766233e-05, "loss": 0.7024, "step": 589 }, { "epoch": 0.6924882629107981, "grad_norm": 0.3446933627128601, "learning_rate": 6.210153482880756e-05, "loss": 0.7392, "step": 590 }, { "epoch": 0.6936619718309859, "grad_norm": 0.34918078780174255, "learning_rate": 6.186540731995277e-05, "loss": 0.6716, "step": 591 }, { "epoch": 0.6948356807511737, "grad_norm": 0.3438567519187927, "learning_rate": 6.1629279811098e-05, "loss": 0.7812, "step": 592 }, { "epoch": 0.6960093896713615, "grad_norm": 0.346626341342926, "learning_rate": 6.139315230224322e-05, "loss": 0.7538, "step": 593 }, { "epoch": 0.6971830985915493, "grad_norm": 0.3506343960762024, "learning_rate": 6.115702479338842e-05, "loss": 0.7434, "step": 594 }, { "epoch": 0.6983568075117371, "grad_norm": 0.35403555631637573, "learning_rate": 6.0920897284533654e-05, "loss": 0.7333, "step": 595 }, { "epoch": 0.6995305164319249, "grad_norm": 0.3391430377960205, "learning_rate": 6.068476977567887e-05, "loss": 0.7486, "step": 596 }, { "epoch": 0.7007042253521126, "grad_norm": 0.33783578872680664, "learning_rate": 6.044864226682408e-05, "loss": 0.7588, "step": 597 }, { "epoch": 0.7018779342723005, "grad_norm": 0.3333738446235657, "learning_rate": 6.021251475796931e-05, "loss": 0.7268, "step": 598 }, { "epoch": 0.7030516431924883, "grad_norm": 0.3494018316268921, "learning_rate": 5.997638724911453e-05, "loss": 0.7363, "step": 599 }, { "epoch": 0.704225352112676, "grad_norm": 0.34416642785072327, "learning_rate": 5.9740259740259744e-05, "loss": 0.7322, "step": 600 }, { "epoch": 0.7053990610328639, "grad_norm": 0.3523387312889099, "learning_rate": 5.950413223140496e-05, "loss": 0.6986, "step": 601 }, { "epoch": 0.7065727699530516, "grad_norm": 0.33000361919403076, "learning_rate": 5.926800472255017e-05, "loss": 0.7535, "step": 602 }, { "epoch": 0.7077464788732394, "grad_norm": 0.33932214975357056, "learning_rate": 5.90318772136954e-05, "loss": 0.7051, "step": 603 }, { "epoch": 0.7089201877934272, "grad_norm": 0.3373797833919525, "learning_rate": 5.8795749704840616e-05, "loss": 0.7022, "step": 604 }, { "epoch": 0.710093896713615, "grad_norm": 0.35239875316619873, "learning_rate": 5.855962219598583e-05, "loss": 0.7893, "step": 605 }, { "epoch": 0.7112676056338029, "grad_norm": 0.36973506212234497, "learning_rate": 5.832349468713105e-05, "loss": 0.7157, "step": 606 }, { "epoch": 0.7124413145539906, "grad_norm": 0.3447434604167938, "learning_rate": 5.8087367178276277e-05, "loss": 0.7306, "step": 607 }, { "epoch": 0.7136150234741784, "grad_norm": 0.36380118131637573, "learning_rate": 5.785123966942149e-05, "loss": 0.7238, "step": 608 }, { "epoch": 0.7147887323943662, "grad_norm": 0.33784252405166626, "learning_rate": 5.7615112160566706e-05, "loss": 0.6792, "step": 609 }, { "epoch": 0.715962441314554, "grad_norm": 0.34995025396347046, "learning_rate": 5.737898465171193e-05, "loss": 0.7158, "step": 610 }, { "epoch": 0.7171361502347418, "grad_norm": 0.3586655259132385, "learning_rate": 5.714285714285714e-05, "loss": 0.7345, "step": 611 }, { "epoch": 0.7183098591549296, "grad_norm": 0.3490711450576782, "learning_rate": 5.6906729634002366e-05, "loss": 0.759, "step": 612 }, { "epoch": 0.7194835680751174, "grad_norm": 0.3405636250972748, "learning_rate": 5.6670602125147584e-05, "loss": 0.7069, "step": 613 }, { "epoch": 0.7206572769953051, "grad_norm": 0.3362460136413574, "learning_rate": 5.6434474616292796e-05, "loss": 0.7413, "step": 614 }, { "epoch": 0.721830985915493, "grad_norm": 0.3571033775806427, "learning_rate": 5.619834710743802e-05, "loss": 0.7138, "step": 615 }, { "epoch": 0.7230046948356808, "grad_norm": 0.33801379799842834, "learning_rate": 5.5962219598583245e-05, "loss": 0.7004, "step": 616 }, { "epoch": 0.7241784037558685, "grad_norm": 0.350063294172287, "learning_rate": 5.5726092089728456e-05, "loss": 0.7342, "step": 617 }, { "epoch": 0.7253521126760564, "grad_norm": 0.3471220135688782, "learning_rate": 5.5489964580873674e-05, "loss": 0.7591, "step": 618 }, { "epoch": 0.7265258215962441, "grad_norm": 0.3600592613220215, "learning_rate": 5.5253837072018886e-05, "loss": 0.7427, "step": 619 }, { "epoch": 0.7276995305164319, "grad_norm": 0.34294822812080383, "learning_rate": 5.501770956316411e-05, "loss": 0.7085, "step": 620 }, { "epoch": 0.7288732394366197, "grad_norm": 0.3481101989746094, "learning_rate": 5.4781582054309335e-05, "loss": 0.7465, "step": 621 }, { "epoch": 0.7300469483568075, "grad_norm": 0.3402861952781677, "learning_rate": 5.4545454545454546e-05, "loss": 0.7613, "step": 622 }, { "epoch": 0.7312206572769953, "grad_norm": 0.3475019335746765, "learning_rate": 5.4309327036599764e-05, "loss": 0.775, "step": 623 }, { "epoch": 0.7323943661971831, "grad_norm": 0.34003034234046936, "learning_rate": 5.407319952774499e-05, "loss": 0.6817, "step": 624 }, { "epoch": 0.7335680751173709, "grad_norm": 0.33620044589042664, "learning_rate": 5.38370720188902e-05, "loss": 0.7392, "step": 625 }, { "epoch": 0.7347417840375586, "grad_norm": 0.34645119309425354, "learning_rate": 5.3600944510035425e-05, "loss": 0.717, "step": 626 }, { "epoch": 0.7359154929577465, "grad_norm": 0.3485560417175293, "learning_rate": 5.336481700118064e-05, "loss": 0.7361, "step": 627 }, { "epoch": 0.7370892018779343, "grad_norm": 0.36997392773628235, "learning_rate": 5.3128689492325854e-05, "loss": 0.7264, "step": 628 }, { "epoch": 0.7382629107981221, "grad_norm": 0.3379404842853546, "learning_rate": 5.289256198347108e-05, "loss": 0.7303, "step": 629 }, { "epoch": 0.7394366197183099, "grad_norm": 0.3385223150253296, "learning_rate": 5.2656434474616304e-05, "loss": 0.7174, "step": 630 }, { "epoch": 0.7406103286384976, "grad_norm": 0.38303306698799133, "learning_rate": 5.2420306965761515e-05, "loss": 0.7539, "step": 631 }, { "epoch": 0.7417840375586855, "grad_norm": 0.3544706404209137, "learning_rate": 5.218417945690673e-05, "loss": 0.7108, "step": 632 }, { "epoch": 0.7429577464788732, "grad_norm": 0.35137131810188293, "learning_rate": 5.1948051948051944e-05, "loss": 0.7184, "step": 633 }, { "epoch": 0.744131455399061, "grad_norm": 0.35326629877090454, "learning_rate": 5.171192443919717e-05, "loss": 0.7114, "step": 634 }, { "epoch": 0.7453051643192489, "grad_norm": 0.35051414370536804, "learning_rate": 5.1475796930342393e-05, "loss": 0.6966, "step": 635 }, { "epoch": 0.7464788732394366, "grad_norm": 0.37491628527641296, "learning_rate": 5.1239669421487605e-05, "loss": 0.7061, "step": 636 }, { "epoch": 0.7476525821596244, "grad_norm": 0.37242433428764343, "learning_rate": 5.100354191263282e-05, "loss": 0.6904, "step": 637 }, { "epoch": 0.7488262910798122, "grad_norm": 0.376429945230484, "learning_rate": 5.076741440377805e-05, "loss": 0.7203, "step": 638 }, { "epoch": 0.75, "grad_norm": 0.34106218814849854, "learning_rate": 5.053128689492326e-05, "loss": 0.6878, "step": 639 }, { "epoch": 0.7511737089201878, "grad_norm": 0.37987956404685974, "learning_rate": 5.029515938606848e-05, "loss": 0.7835, "step": 640 }, { "epoch": 0.7523474178403756, "grad_norm": 0.355932354927063, "learning_rate": 5.00590318772137e-05, "loss": 0.7382, "step": 641 }, { "epoch": 0.7535211267605634, "grad_norm": 0.33495378494262695, "learning_rate": 4.982290436835891e-05, "loss": 0.7244, "step": 642 }, { "epoch": 0.7546948356807511, "grad_norm": 0.36573663353919983, "learning_rate": 4.958677685950414e-05, "loss": 0.7339, "step": 643 }, { "epoch": 0.755868544600939, "grad_norm": 0.34233418107032776, "learning_rate": 4.9350649350649355e-05, "loss": 0.7303, "step": 644 }, { "epoch": 0.7570422535211268, "grad_norm": 0.36358365416526794, "learning_rate": 4.9114521841794566e-05, "loss": 0.7169, "step": 645 }, { "epoch": 0.7582159624413145, "grad_norm": 0.3423750400543213, "learning_rate": 4.887839433293979e-05, "loss": 0.7413, "step": 646 }, { "epoch": 0.7593896713615024, "grad_norm": 0.34080007672309875, "learning_rate": 4.864226682408501e-05, "loss": 0.7319, "step": 647 }, { "epoch": 0.7605633802816901, "grad_norm": 0.35408544540405273, "learning_rate": 4.840613931523023e-05, "loss": 0.6895, "step": 648 }, { "epoch": 0.7617370892018779, "grad_norm": 0.34515753388404846, "learning_rate": 4.8170011806375445e-05, "loss": 0.7181, "step": 649 }, { "epoch": 0.7629107981220657, "grad_norm": 0.3446560502052307, "learning_rate": 4.793388429752066e-05, "loss": 0.7156, "step": 650 }, { "epoch": 0.7640845070422535, "grad_norm": 0.3451150357723236, "learning_rate": 4.769775678866588e-05, "loss": 0.7232, "step": 651 }, { "epoch": 0.7652582159624414, "grad_norm": 0.357740193605423, "learning_rate": 4.74616292798111e-05, "loss": 0.6872, "step": 652 }, { "epoch": 0.7664319248826291, "grad_norm": 0.3685015141963959, "learning_rate": 4.7225501770956324e-05, "loss": 0.735, "step": 653 }, { "epoch": 0.7676056338028169, "grad_norm": 0.3503192961215973, "learning_rate": 4.6989374262101535e-05, "loss": 0.7336, "step": 654 }, { "epoch": 0.7687793427230047, "grad_norm": 0.33453887701034546, "learning_rate": 4.675324675324675e-05, "loss": 0.7101, "step": 655 }, { "epoch": 0.7699530516431925, "grad_norm": 0.3708442747592926, "learning_rate": 4.651711924439197e-05, "loss": 0.7153, "step": 656 }, { "epoch": 0.7711267605633803, "grad_norm": 0.3736172318458557, "learning_rate": 4.6280991735537196e-05, "loss": 0.7071, "step": 657 }, { "epoch": 0.7723004694835681, "grad_norm": 0.35988256335258484, "learning_rate": 4.6044864226682414e-05, "loss": 0.7285, "step": 658 }, { "epoch": 0.7734741784037559, "grad_norm": 0.34314337372779846, "learning_rate": 4.5808736717827625e-05, "loss": 0.7137, "step": 659 }, { "epoch": 0.7746478873239436, "grad_norm": 0.3723309338092804, "learning_rate": 4.557260920897285e-05, "loss": 0.7391, "step": 660 }, { "epoch": 0.7758215962441315, "grad_norm": 0.3581268787384033, "learning_rate": 4.533648170011807e-05, "loss": 0.7157, "step": 661 }, { "epoch": 0.7769953051643192, "grad_norm": 0.36784443259239197, "learning_rate": 4.5100354191263286e-05, "loss": 0.6865, "step": 662 }, { "epoch": 0.778169014084507, "grad_norm": 0.36377546191215515, "learning_rate": 4.48642266824085e-05, "loss": 0.7437, "step": 663 }, { "epoch": 0.7793427230046949, "grad_norm": 0.349101722240448, "learning_rate": 4.462809917355372e-05, "loss": 0.7226, "step": 664 }, { "epoch": 0.7805164319248826, "grad_norm": 0.36608216166496277, "learning_rate": 4.439197166469894e-05, "loss": 0.7543, "step": 665 }, { "epoch": 0.7816901408450704, "grad_norm": 0.3495696783065796, "learning_rate": 4.415584415584416e-05, "loss": 0.708, "step": 666 }, { "epoch": 0.7828638497652582, "grad_norm": 0.3664140999317169, "learning_rate": 4.3919716646989375e-05, "loss": 0.7225, "step": 667 }, { "epoch": 0.784037558685446, "grad_norm": 0.3560849726200104, "learning_rate": 4.368358913813459e-05, "loss": 0.6972, "step": 668 }, { "epoch": 0.7852112676056338, "grad_norm": 0.3571857511997223, "learning_rate": 4.344746162927981e-05, "loss": 0.694, "step": 669 }, { "epoch": 0.7863849765258216, "grad_norm": 0.37072160840034485, "learning_rate": 4.321133412042503e-05, "loss": 0.7202, "step": 670 }, { "epoch": 0.7875586854460094, "grad_norm": 0.354948490858078, "learning_rate": 4.2975206611570254e-05, "loss": 0.7481, "step": 671 }, { "epoch": 0.7887323943661971, "grad_norm": 0.3736347258090973, "learning_rate": 4.2739079102715465e-05, "loss": 0.7261, "step": 672 }, { "epoch": 0.789906103286385, "grad_norm": 0.3690294623374939, "learning_rate": 4.250295159386068e-05, "loss": 0.7529, "step": 673 }, { "epoch": 0.7910798122065728, "grad_norm": 0.354192316532135, "learning_rate": 4.226682408500591e-05, "loss": 0.7176, "step": 674 }, { "epoch": 0.7922535211267606, "grad_norm": 0.355185866355896, "learning_rate": 4.2030696576151126e-05, "loss": 0.7099, "step": 675 }, { "epoch": 0.7934272300469484, "grad_norm": 0.3503565490245819, "learning_rate": 4.1794569067296344e-05, "loss": 0.7072, "step": 676 }, { "epoch": 0.7946009389671361, "grad_norm": 0.3727845549583435, "learning_rate": 4.155844155844156e-05, "loss": 0.7334, "step": 677 }, { "epoch": 0.795774647887324, "grad_norm": 0.33894312381744385, "learning_rate": 4.132231404958678e-05, "loss": 0.6946, "step": 678 }, { "epoch": 0.7969483568075117, "grad_norm": 0.3385523855686188, "learning_rate": 4.1086186540732e-05, "loss": 0.7096, "step": 679 }, { "epoch": 0.7981220657276995, "grad_norm": 0.3488437235355377, "learning_rate": 4.0850059031877216e-05, "loss": 0.6942, "step": 680 }, { "epoch": 0.7992957746478874, "grad_norm": 0.34666576981544495, "learning_rate": 4.0613931523022434e-05, "loss": 0.7329, "step": 681 }, { "epoch": 0.8004694835680751, "grad_norm": 0.3557136356830597, "learning_rate": 4.037780401416765e-05, "loss": 0.7655, "step": 682 }, { "epoch": 0.8016431924882629, "grad_norm": 0.3647683262825012, "learning_rate": 4.014167650531287e-05, "loss": 0.7578, "step": 683 }, { "epoch": 0.8028169014084507, "grad_norm": 0.3452191650867462, "learning_rate": 3.9905548996458095e-05, "loss": 0.7145, "step": 684 }, { "epoch": 0.8039906103286385, "grad_norm": 0.3540481925010681, "learning_rate": 3.9669421487603306e-05, "loss": 0.7347, "step": 685 }, { "epoch": 0.8051643192488263, "grad_norm": 0.3536418378353119, "learning_rate": 3.9433293978748524e-05, "loss": 0.7103, "step": 686 }, { "epoch": 0.8063380281690141, "grad_norm": 0.34728798270225525, "learning_rate": 3.919716646989374e-05, "loss": 0.7376, "step": 687 }, { "epoch": 0.8075117370892019, "grad_norm": 0.354643851518631, "learning_rate": 3.8961038961038966e-05, "loss": 0.7223, "step": 688 }, { "epoch": 0.8086854460093896, "grad_norm": 0.3438583016395569, "learning_rate": 3.8724911452184184e-05, "loss": 0.6906, "step": 689 }, { "epoch": 0.8098591549295775, "grad_norm": 0.34713107347488403, "learning_rate": 3.8488783943329396e-05, "loss": 0.7361, "step": 690 }, { "epoch": 0.8110328638497653, "grad_norm": 0.3483150005340576, "learning_rate": 3.825265643447462e-05, "loss": 0.7016, "step": 691 }, { "epoch": 0.812206572769953, "grad_norm": 0.34848445653915405, "learning_rate": 3.801652892561984e-05, "loss": 0.6966, "step": 692 }, { "epoch": 0.8133802816901409, "grad_norm": 0.34223318099975586, "learning_rate": 3.7780401416765056e-05, "loss": 0.7088, "step": 693 }, { "epoch": 0.8145539906103286, "grad_norm": 0.33693239092826843, "learning_rate": 3.7544273907910274e-05, "loss": 0.7108, "step": 694 }, { "epoch": 0.8157276995305164, "grad_norm": 0.34613272547721863, "learning_rate": 3.730814639905549e-05, "loss": 0.7075, "step": 695 }, { "epoch": 0.8169014084507042, "grad_norm": 0.3430733382701874, "learning_rate": 3.707201889020071e-05, "loss": 0.7246, "step": 696 }, { "epoch": 0.818075117370892, "grad_norm": 0.35237351059913635, "learning_rate": 3.683589138134593e-05, "loss": 0.6918, "step": 697 }, { "epoch": 0.8192488262910798, "grad_norm": 0.3375650644302368, "learning_rate": 3.659976387249115e-05, "loss": 0.6978, "step": 698 }, { "epoch": 0.8204225352112676, "grad_norm": 0.3585062026977539, "learning_rate": 3.6363636363636364e-05, "loss": 0.7241, "step": 699 }, { "epoch": 0.8215962441314554, "grad_norm": 0.35660460591316223, "learning_rate": 3.612750885478158e-05, "loss": 0.6946, "step": 700 }, { "epoch": 0.8227699530516432, "grad_norm": 0.3468845784664154, "learning_rate": 3.58913813459268e-05, "loss": 0.7535, "step": 701 }, { "epoch": 0.823943661971831, "grad_norm": 0.365291029214859, "learning_rate": 3.5655253837072025e-05, "loss": 0.7438, "step": 702 }, { "epoch": 0.8251173708920188, "grad_norm": 0.353506863117218, "learning_rate": 3.5419126328217236e-05, "loss": 0.7359, "step": 703 }, { "epoch": 0.8262910798122066, "grad_norm": 0.381610244512558, "learning_rate": 3.5182998819362454e-05, "loss": 0.7821, "step": 704 }, { "epoch": 0.8274647887323944, "grad_norm": 0.37710806727409363, "learning_rate": 3.494687131050768e-05, "loss": 0.7349, "step": 705 }, { "epoch": 0.8286384976525821, "grad_norm": 0.361545592546463, "learning_rate": 3.47107438016529e-05, "loss": 0.7229, "step": 706 }, { "epoch": 0.82981220657277, "grad_norm": 0.3615299463272095, "learning_rate": 3.4474616292798115e-05, "loss": 0.748, "step": 707 }, { "epoch": 0.8309859154929577, "grad_norm": 0.3437252342700958, "learning_rate": 3.4238488783943326e-05, "loss": 0.7165, "step": 708 }, { "epoch": 0.8321596244131455, "grad_norm": 0.35603129863739014, "learning_rate": 3.400236127508855e-05, "loss": 0.7373, "step": 709 }, { "epoch": 0.8333333333333334, "grad_norm": 0.3586898446083069, "learning_rate": 3.376623376623377e-05, "loss": 0.7056, "step": 710 }, { "epoch": 0.8345070422535211, "grad_norm": 0.3558507263660431, "learning_rate": 3.353010625737899e-05, "loss": 0.742, "step": 711 }, { "epoch": 0.8356807511737089, "grad_norm": 0.3359735608100891, "learning_rate": 3.3293978748524205e-05, "loss": 0.6994, "step": 712 }, { "epoch": 0.8368544600938967, "grad_norm": 0.34250345826148987, "learning_rate": 3.305785123966942e-05, "loss": 0.6762, "step": 713 }, { "epoch": 0.8380281690140845, "grad_norm": 0.38417667150497437, "learning_rate": 3.282172373081464e-05, "loss": 0.7213, "step": 714 }, { "epoch": 0.8392018779342723, "grad_norm": 0.3643978536128998, "learning_rate": 3.258559622195986e-05, "loss": 0.6884, "step": 715 }, { "epoch": 0.8403755868544601, "grad_norm": 0.3544299602508545, "learning_rate": 3.234946871310508e-05, "loss": 0.712, "step": 716 }, { "epoch": 0.8415492957746479, "grad_norm": 0.36903661489486694, "learning_rate": 3.2113341204250294e-05, "loss": 0.7227, "step": 717 }, { "epoch": 0.8427230046948356, "grad_norm": 0.3557377755641937, "learning_rate": 3.187721369539551e-05, "loss": 0.6904, "step": 718 }, { "epoch": 0.8438967136150235, "grad_norm": 0.36762547492980957, "learning_rate": 3.164108618654074e-05, "loss": 0.7469, "step": 719 }, { "epoch": 0.8450704225352113, "grad_norm": 0.35805556178092957, "learning_rate": 3.1404958677685955e-05, "loss": 0.7443, "step": 720 }, { "epoch": 0.846244131455399, "grad_norm": 0.38130536675453186, "learning_rate": 3.1168831168831166e-05, "loss": 0.7664, "step": 721 }, { "epoch": 0.8474178403755869, "grad_norm": 0.3599521219730377, "learning_rate": 3.0932703659976384e-05, "loss": 0.7065, "step": 722 }, { "epoch": 0.8485915492957746, "grad_norm": 0.3531062602996826, "learning_rate": 3.069657615112161e-05, "loss": 0.7451, "step": 723 }, { "epoch": 0.8497652582159625, "grad_norm": 0.36916878819465637, "learning_rate": 3.0460448642266827e-05, "loss": 0.7077, "step": 724 }, { "epoch": 0.8509389671361502, "grad_norm": 0.38139578700065613, "learning_rate": 3.022432113341204e-05, "loss": 0.7452, "step": 725 }, { "epoch": 0.852112676056338, "grad_norm": 0.337944358587265, "learning_rate": 2.9988193624557266e-05, "loss": 0.6596, "step": 726 }, { "epoch": 0.8532863849765259, "grad_norm": 0.36196213960647583, "learning_rate": 2.975206611570248e-05, "loss": 0.7081, "step": 727 }, { "epoch": 0.8544600938967136, "grad_norm": 0.34913602471351624, "learning_rate": 2.95159386068477e-05, "loss": 0.6901, "step": 728 }, { "epoch": 0.8556338028169014, "grad_norm": 0.343414843082428, "learning_rate": 2.9279811097992914e-05, "loss": 0.675, "step": 729 }, { "epoch": 0.8568075117370892, "grad_norm": 0.3704102039337158, "learning_rate": 2.9043683589138138e-05, "loss": 0.7566, "step": 730 }, { "epoch": 0.857981220657277, "grad_norm": 0.3464911878108978, "learning_rate": 2.8807556080283353e-05, "loss": 0.6872, "step": 731 }, { "epoch": 0.8591549295774648, "grad_norm": 0.3615940511226654, "learning_rate": 2.857142857142857e-05, "loss": 0.7755, "step": 732 }, { "epoch": 0.8603286384976526, "grad_norm": 0.35284191370010376, "learning_rate": 2.8335301062573792e-05, "loss": 0.7483, "step": 733 }, { "epoch": 0.8615023474178404, "grad_norm": 0.3469059467315674, "learning_rate": 2.809917355371901e-05, "loss": 0.6902, "step": 734 }, { "epoch": 0.8626760563380281, "grad_norm": 0.35148003697395325, "learning_rate": 2.7863046044864228e-05, "loss": 0.732, "step": 735 }, { "epoch": 0.863849765258216, "grad_norm": 0.3533206880092621, "learning_rate": 2.7626918536009443e-05, "loss": 0.7287, "step": 736 }, { "epoch": 0.8650234741784038, "grad_norm": 0.383095383644104, "learning_rate": 2.7390791027154668e-05, "loss": 0.8017, "step": 737 }, { "epoch": 0.8661971830985915, "grad_norm": 0.3541397452354431, "learning_rate": 2.7154663518299882e-05, "loss": 0.7291, "step": 738 }, { "epoch": 0.8673708920187794, "grad_norm": 0.35989582538604736, "learning_rate": 2.69185360094451e-05, "loss": 0.7211, "step": 739 }, { "epoch": 0.8685446009389671, "grad_norm": 0.34245404601097107, "learning_rate": 2.668240850059032e-05, "loss": 0.7062, "step": 740 }, { "epoch": 0.8697183098591549, "grad_norm": 0.3396112024784088, "learning_rate": 2.644628099173554e-05, "loss": 0.6946, "step": 741 }, { "epoch": 0.8708920187793427, "grad_norm": 0.34901162981987, "learning_rate": 2.6210153482880757e-05, "loss": 0.7742, "step": 742 }, { "epoch": 0.8720657276995305, "grad_norm": 0.3654363453388214, "learning_rate": 2.5974025974025972e-05, "loss": 0.7894, "step": 743 }, { "epoch": 0.8732394366197183, "grad_norm": 0.3478833734989166, "learning_rate": 2.5737898465171197e-05, "loss": 0.6909, "step": 744 }, { "epoch": 0.8744131455399061, "grad_norm": 0.3447161912918091, "learning_rate": 2.550177095631641e-05, "loss": 0.7166, "step": 745 }, { "epoch": 0.8755868544600939, "grad_norm": 0.35436901450157166, "learning_rate": 2.526564344746163e-05, "loss": 0.6962, "step": 746 }, { "epoch": 0.8767605633802817, "grad_norm": 0.3359661400318146, "learning_rate": 2.502951593860685e-05, "loss": 0.7345, "step": 747 }, { "epoch": 0.8779342723004695, "grad_norm": 0.35876211524009705, "learning_rate": 2.479338842975207e-05, "loss": 0.6723, "step": 748 }, { "epoch": 0.8791079812206573, "grad_norm": 0.35507625341415405, "learning_rate": 2.4557260920897283e-05, "loss": 0.6744, "step": 749 }, { "epoch": 0.8802816901408451, "grad_norm": 0.3504907786846161, "learning_rate": 2.4321133412042505e-05, "loss": 0.7281, "step": 750 }, { "epoch": 0.8814553990610329, "grad_norm": 0.3498130440711975, "learning_rate": 2.4085005903187723e-05, "loss": 0.7079, "step": 751 }, { "epoch": 0.8826291079812206, "grad_norm": 0.36793026328086853, "learning_rate": 2.384887839433294e-05, "loss": 0.747, "step": 752 }, { "epoch": 0.8838028169014085, "grad_norm": 0.3484232723712921, "learning_rate": 2.3612750885478162e-05, "loss": 0.7347, "step": 753 }, { "epoch": 0.8849765258215962, "grad_norm": 0.34402692317962646, "learning_rate": 2.3376623376623376e-05, "loss": 0.6717, "step": 754 }, { "epoch": 0.886150234741784, "grad_norm": 0.377380907535553, "learning_rate": 2.3140495867768598e-05, "loss": 0.7642, "step": 755 }, { "epoch": 0.8873239436619719, "grad_norm": 0.361382395029068, "learning_rate": 2.2904368358913812e-05, "loss": 0.7081, "step": 756 }, { "epoch": 0.8884976525821596, "grad_norm": 0.3643784821033478, "learning_rate": 2.2668240850059034e-05, "loss": 0.7219, "step": 757 }, { "epoch": 0.8896713615023474, "grad_norm": 0.3974801301956177, "learning_rate": 2.243211334120425e-05, "loss": 0.712, "step": 758 }, { "epoch": 0.8908450704225352, "grad_norm": 0.35573598742485046, "learning_rate": 2.219598583234947e-05, "loss": 0.7335, "step": 759 }, { "epoch": 0.892018779342723, "grad_norm": 0.3532857596874237, "learning_rate": 2.1959858323494688e-05, "loss": 0.7013, "step": 760 }, { "epoch": 0.8931924882629108, "grad_norm": 0.33362728357315063, "learning_rate": 2.1723730814639906e-05, "loss": 0.6739, "step": 761 }, { "epoch": 0.8943661971830986, "grad_norm": 0.3325813412666321, "learning_rate": 2.1487603305785127e-05, "loss": 0.7099, "step": 762 }, { "epoch": 0.8955399061032864, "grad_norm": 0.3451225459575653, "learning_rate": 2.125147579693034e-05, "loss": 0.6959, "step": 763 }, { "epoch": 0.8967136150234741, "grad_norm": 0.3604796528816223, "learning_rate": 2.1015348288075563e-05, "loss": 0.737, "step": 764 }, { "epoch": 0.897887323943662, "grad_norm": 0.34980282187461853, "learning_rate": 2.077922077922078e-05, "loss": 0.7206, "step": 765 }, { "epoch": 0.8990610328638498, "grad_norm": 0.35130617022514343, "learning_rate": 2.0543093270366e-05, "loss": 0.7153, "step": 766 }, { "epoch": 0.9002347417840375, "grad_norm": 0.34524810314178467, "learning_rate": 2.0306965761511217e-05, "loss": 0.7237, "step": 767 }, { "epoch": 0.9014084507042254, "grad_norm": 0.35661572217941284, "learning_rate": 2.0070838252656435e-05, "loss": 0.6831, "step": 768 }, { "epoch": 0.9025821596244131, "grad_norm": 0.35206255316734314, "learning_rate": 1.9834710743801653e-05, "loss": 0.7721, "step": 769 }, { "epoch": 0.903755868544601, "grad_norm": 0.35439351201057434, "learning_rate": 1.959858323494687e-05, "loss": 0.7142, "step": 770 }, { "epoch": 0.9049295774647887, "grad_norm": 0.33722493052482605, "learning_rate": 1.9362455726092092e-05, "loss": 0.6968, "step": 771 }, { "epoch": 0.9061032863849765, "grad_norm": 0.3573172092437744, "learning_rate": 1.912632821723731e-05, "loss": 0.7301, "step": 772 }, { "epoch": 0.9072769953051644, "grad_norm": 0.3347008526325226, "learning_rate": 1.8890200708382528e-05, "loss": 0.6721, "step": 773 }, { "epoch": 0.9084507042253521, "grad_norm": 0.3563063144683838, "learning_rate": 1.8654073199527746e-05, "loss": 0.7233, "step": 774 }, { "epoch": 0.9096244131455399, "grad_norm": 0.35159915685653687, "learning_rate": 1.8417945690672964e-05, "loss": 0.7184, "step": 775 }, { "epoch": 0.9107981220657277, "grad_norm": 0.35826948285102844, "learning_rate": 1.8181818181818182e-05, "loss": 0.7301, "step": 776 }, { "epoch": 0.9119718309859155, "grad_norm": 0.3533133268356323, "learning_rate": 1.79456906729634e-05, "loss": 0.7373, "step": 777 }, { "epoch": 0.9131455399061033, "grad_norm": 0.3495820164680481, "learning_rate": 1.7709563164108618e-05, "loss": 0.7379, "step": 778 }, { "epoch": 0.9143192488262911, "grad_norm": 0.33082085847854614, "learning_rate": 1.747343565525384e-05, "loss": 0.6789, "step": 779 }, { "epoch": 0.9154929577464789, "grad_norm": 0.34669029712677, "learning_rate": 1.7237308146399057e-05, "loss": 0.6962, "step": 780 }, { "epoch": 0.9166666666666666, "grad_norm": 0.3401969373226166, "learning_rate": 1.7001180637544275e-05, "loss": 0.717, "step": 781 }, { "epoch": 0.9178403755868545, "grad_norm": 0.3488728702068329, "learning_rate": 1.6765053128689493e-05, "loss": 0.7087, "step": 782 }, { "epoch": 0.9190140845070423, "grad_norm": 0.39244547486305237, "learning_rate": 1.652892561983471e-05, "loss": 0.7331, "step": 783 }, { "epoch": 0.92018779342723, "grad_norm": 0.33185505867004395, "learning_rate": 1.629279811097993e-05, "loss": 0.6821, "step": 784 }, { "epoch": 0.9213615023474179, "grad_norm": 0.34186288714408875, "learning_rate": 1.6056670602125147e-05, "loss": 0.6766, "step": 785 }, { "epoch": 0.9225352112676056, "grad_norm": 0.34512627124786377, "learning_rate": 1.582054309327037e-05, "loss": 0.6837, "step": 786 }, { "epoch": 0.9237089201877934, "grad_norm": 0.34042122960090637, "learning_rate": 1.5584415584415583e-05, "loss": 0.7266, "step": 787 }, { "epoch": 0.9248826291079812, "grad_norm": 0.34173402190208435, "learning_rate": 1.5348288075560805e-05, "loss": 0.6998, "step": 788 }, { "epoch": 0.926056338028169, "grad_norm": 0.34008073806762695, "learning_rate": 1.511216056670602e-05, "loss": 0.7211, "step": 789 }, { "epoch": 0.9272300469483568, "grad_norm": 0.3400252163410187, "learning_rate": 1.487603305785124e-05, "loss": 0.6771, "step": 790 }, { "epoch": 0.9284037558685446, "grad_norm": 0.3393029570579529, "learning_rate": 1.4639905548996457e-05, "loss": 0.7274, "step": 791 }, { "epoch": 0.9295774647887324, "grad_norm": 0.3489772379398346, "learning_rate": 1.4403778040141676e-05, "loss": 0.7195, "step": 792 }, { "epoch": 0.9307511737089202, "grad_norm": 0.3434072732925415, "learning_rate": 1.4167650531286896e-05, "loss": 0.6806, "step": 793 }, { "epoch": 0.931924882629108, "grad_norm": 0.35593146085739136, "learning_rate": 1.3931523022432114e-05, "loss": 0.7026, "step": 794 }, { "epoch": 0.9330985915492958, "grad_norm": 0.33654287457466125, "learning_rate": 1.3695395513577334e-05, "loss": 0.6655, "step": 795 }, { "epoch": 0.9342723004694836, "grad_norm": 0.35049983859062195, "learning_rate": 1.345926800472255e-05, "loss": 0.686, "step": 796 }, { "epoch": 0.9354460093896714, "grad_norm": 0.3442087471485138, "learning_rate": 1.322314049586777e-05, "loss": 0.7048, "step": 797 }, { "epoch": 0.9366197183098591, "grad_norm": 0.3569439649581909, "learning_rate": 1.2987012987012986e-05, "loss": 0.7271, "step": 798 }, { "epoch": 0.937793427230047, "grad_norm": 0.3418942391872406, "learning_rate": 1.2750885478158206e-05, "loss": 0.7132, "step": 799 }, { "epoch": 0.9389671361502347, "grad_norm": 0.3399513363838196, "learning_rate": 1.2514757969303425e-05, "loss": 0.7046, "step": 800 }, { "epoch": 0.9401408450704225, "grad_norm": 0.34055379033088684, "learning_rate": 0.00010641553855208948, "loss": 0.7293, "step": 801 }, { "epoch": 0.9413145539906104, "grad_norm": 0.3299119770526886, "learning_rate": 0.0001062978222483814, "loss": 0.6779, "step": 802 }, { "epoch": 0.9424882629107981, "grad_norm": 0.3833242356777191, "learning_rate": 0.00010618010594467334, "loss": 0.6909, "step": 803 }, { "epoch": 0.9436619718309859, "grad_norm": 0.39958855509757996, "learning_rate": 0.00010606238964096529, "loss": 0.7307, "step": 804 }, { "epoch": 0.9448356807511737, "grad_norm": 0.38618725538253784, "learning_rate": 0.00010594467333725722, "loss": 0.6984, "step": 805 }, { "epoch": 0.9460093896713615, "grad_norm": 0.4084942936897278, "learning_rate": 0.00010582695703354914, "loss": 0.7456, "step": 806 }, { "epoch": 0.9471830985915493, "grad_norm": 0.4109421372413635, "learning_rate": 0.00010570924072984109, "loss": 0.6991, "step": 807 }, { "epoch": 0.9483568075117371, "grad_norm": 0.382415771484375, "learning_rate": 0.00010559152442613303, "loss": 0.726, "step": 808 }, { "epoch": 0.9495305164319249, "grad_norm": 0.4036392867565155, "learning_rate": 0.00010547380812242496, "loss": 0.7264, "step": 809 }, { "epoch": 0.9507042253521126, "grad_norm": 0.38903331756591797, "learning_rate": 0.00010535609181871691, "loss": 0.691, "step": 810 }, { "epoch": 0.9518779342723005, "grad_norm": 0.3803318440914154, "learning_rate": 0.00010523837551500883, "loss": 0.7271, "step": 811 }, { "epoch": 0.9530516431924883, "grad_norm": 0.3850460350513458, "learning_rate": 0.00010512065921130076, "loss": 0.7111, "step": 812 }, { "epoch": 0.954225352112676, "grad_norm": 0.4110994040966034, "learning_rate": 0.00010500294290759271, "loss": 0.7282, "step": 813 }, { "epoch": 0.9553990610328639, "grad_norm": 0.3853722810745239, "learning_rate": 0.00010488522660388465, "loss": 0.7194, "step": 814 }, { "epoch": 0.9565727699530516, "grad_norm": 0.37440797686576843, "learning_rate": 0.00010476751030017658, "loss": 0.7116, "step": 815 }, { "epoch": 0.9577464788732394, "grad_norm": 0.42637899518013, "learning_rate": 0.00010464979399646853, "loss": 0.7189, "step": 816 }, { "epoch": 0.9589201877934272, "grad_norm": 0.4067356288433075, "learning_rate": 0.00010453207769276045, "loss": 0.7509, "step": 817 }, { "epoch": 0.960093896713615, "grad_norm": 0.3854503929615021, "learning_rate": 0.00010441436138905238, "loss": 0.7426, "step": 818 }, { "epoch": 0.9612676056338029, "grad_norm": 0.4298991858959198, "learning_rate": 0.00010429664508534433, "loss": 0.7528, "step": 819 }, { "epoch": 0.9624413145539906, "grad_norm": 0.3748774826526642, "learning_rate": 0.00010417892878163627, "loss": 0.6512, "step": 820 }, { "epoch": 0.9636150234741784, "grad_norm": 0.38448989391326904, "learning_rate": 0.00010406121247792819, "loss": 0.6929, "step": 821 }, { "epoch": 0.9647887323943662, "grad_norm": 0.42416030168533325, "learning_rate": 0.00010394349617422015, "loss": 0.7312, "step": 822 }, { "epoch": 0.965962441314554, "grad_norm": 0.3875625729560852, "learning_rate": 0.00010382577987051207, "loss": 0.7121, "step": 823 }, { "epoch": 0.9671361502347418, "grad_norm": 0.4241638481616974, "learning_rate": 0.000103708063566804, "loss": 0.7248, "step": 824 }, { "epoch": 0.9683098591549296, "grad_norm": 0.4026165306568146, "learning_rate": 0.00010359034726309595, "loss": 0.7224, "step": 825 }, { "epoch": 0.9694835680751174, "grad_norm": 0.39895206689834595, "learning_rate": 0.00010347263095938789, "loss": 0.7193, "step": 826 }, { "epoch": 0.9706572769953051, "grad_norm": 0.395463228225708, "learning_rate": 0.00010335491465567981, "loss": 0.7673, "step": 827 }, { "epoch": 0.971830985915493, "grad_norm": 0.4351494312286377, "learning_rate": 0.00010323719835197174, "loss": 0.7684, "step": 828 }, { "epoch": 0.9730046948356808, "grad_norm": 0.4378681182861328, "learning_rate": 0.00010311948204826369, "loss": 0.7277, "step": 829 }, { "epoch": 0.9741784037558685, "grad_norm": 0.4214630722999573, "learning_rate": 0.00010300176574455563, "loss": 0.7107, "step": 830 }, { "epoch": 0.9753521126760564, "grad_norm": 0.41999107599258423, "learning_rate": 0.00010288404944084755, "loss": 0.7328, "step": 831 }, { "epoch": 0.9765258215962441, "grad_norm": 0.49026909470558167, "learning_rate": 0.00010276633313713951, "loss": 0.7345, "step": 832 }, { "epoch": 0.9776995305164319, "grad_norm": 0.4068211317062378, "learning_rate": 0.00010264861683343143, "loss": 0.701, "step": 833 }, { "epoch": 0.9788732394366197, "grad_norm": 0.42514288425445557, "learning_rate": 0.00010253090052972336, "loss": 0.729, "step": 834 }, { "epoch": 0.9800469483568075, "grad_norm": 0.4883005619049072, "learning_rate": 0.00010241318422601531, "loss": 0.7183, "step": 835 }, { "epoch": 0.9812206572769953, "grad_norm": 0.38146787881851196, "learning_rate": 0.00010229546792230725, "loss": 0.6977, "step": 836 }, { "epoch": 0.9823943661971831, "grad_norm": 0.3898909389972687, "learning_rate": 0.00010217775161859917, "loss": 0.7131, "step": 837 }, { "epoch": 0.9835680751173709, "grad_norm": 0.39693424105644226, "learning_rate": 0.00010206003531489112, "loss": 0.7184, "step": 838 }, { "epoch": 0.9847417840375586, "grad_norm": 0.3968975841999054, "learning_rate": 0.00010194231901118305, "loss": 0.7536, "step": 839 }, { "epoch": 0.9859154929577465, "grad_norm": 0.4030087888240814, "learning_rate": 0.00010182460270747499, "loss": 0.7156, "step": 840 }, { "epoch": 0.9870892018779343, "grad_norm": 0.37477344274520874, "learning_rate": 0.00010170688640376693, "loss": 0.6815, "step": 841 }, { "epoch": 0.9882629107981221, "grad_norm": 0.40929409861564636, "learning_rate": 0.00010158917010005887, "loss": 0.6827, "step": 842 }, { "epoch": 0.9894366197183099, "grad_norm": 0.36350882053375244, "learning_rate": 0.00010147145379635079, "loss": 0.6927, "step": 843 }, { "epoch": 0.9906103286384976, "grad_norm": 0.3828059434890747, "learning_rate": 0.00010135373749264274, "loss": 0.7254, "step": 844 }, { "epoch": 0.9917840375586855, "grad_norm": 0.4095743000507355, "learning_rate": 0.00010123602118893467, "loss": 0.719, "step": 845 }, { "epoch": 0.9929577464788732, "grad_norm": 0.37418296933174133, "learning_rate": 0.0001011183048852266, "loss": 0.682, "step": 846 }, { "epoch": 0.994131455399061, "grad_norm": 0.39427751302719116, "learning_rate": 0.00010100058858151855, "loss": 0.7742, "step": 847 }, { "epoch": 0.9953051643192489, "grad_norm": 0.3696395754814148, "learning_rate": 0.00010088287227781048, "loss": 0.7377, "step": 848 }, { "epoch": 0.9964788732394366, "grad_norm": 0.36249879002571106, "learning_rate": 0.00010076515597410241, "loss": 0.7237, "step": 849 }, { "epoch": 0.9976525821596244, "grad_norm": 0.3712272047996521, "learning_rate": 0.00010064743967039436, "loss": 0.6737, "step": 850 }, { "epoch": 0.9988262910798122, "grad_norm": 0.37550613284111023, "learning_rate": 0.00010052972336668629, "loss": 0.7147, "step": 851 }, { "epoch": 1.0, "grad_norm": 0.405351459980011, "learning_rate": 0.00010041200706297821, "loss": 0.7364, "step": 852 }, { "epoch": 1.0011737089201878, "grad_norm": 0.39747750759124756, "learning_rate": 0.00010029429075927018, "loss": 0.6934, "step": 853 }, { "epoch": 1.0023474178403755, "grad_norm": 0.3695623576641083, "learning_rate": 0.0001001765744555621, "loss": 0.6971, "step": 854 }, { "epoch": 1.0035211267605635, "grad_norm": 0.3880208134651184, "learning_rate": 0.00010005885815185403, "loss": 0.7219, "step": 855 }, { "epoch": 1.0046948356807512, "grad_norm": 0.40131011605262756, "learning_rate": 9.994114184814597e-05, "loss": 0.6925, "step": 856 }, { "epoch": 1.005868544600939, "grad_norm": 0.38630256056785583, "learning_rate": 9.982342554443791e-05, "loss": 0.7412, "step": 857 }, { "epoch": 1.0070422535211268, "grad_norm": 0.39141979813575745, "learning_rate": 9.970570924072985e-05, "loss": 0.7089, "step": 858 }, { "epoch": 1.0082159624413145, "grad_norm": 0.3811167776584625, "learning_rate": 9.958799293702178e-05, "loss": 0.6979, "step": 859 }, { "epoch": 1.0093896713615023, "grad_norm": 0.38177528977394104, "learning_rate": 9.947027663331372e-05, "loss": 0.7181, "step": 860 }, { "epoch": 1.0105633802816902, "grad_norm": 0.36225804686546326, "learning_rate": 9.935256032960567e-05, "loss": 0.6495, "step": 861 }, { "epoch": 1.011737089201878, "grad_norm": 0.3796376585960388, "learning_rate": 9.923484402589759e-05, "loss": 0.6661, "step": 862 }, { "epoch": 1.0129107981220657, "grad_norm": 0.3896029591560364, "learning_rate": 9.911712772218953e-05, "loss": 0.6705, "step": 863 }, { "epoch": 1.0140845070422535, "grad_norm": 0.35688912868499756, "learning_rate": 9.899941141848147e-05, "loss": 0.6835, "step": 864 }, { "epoch": 1.0152582159624413, "grad_norm": 0.3919657766819, "learning_rate": 9.88816951147734e-05, "loss": 0.6771, "step": 865 }, { "epoch": 1.016431924882629, "grad_norm": 0.390311062335968, "learning_rate": 9.876397881106534e-05, "loss": 0.7208, "step": 866 }, { "epoch": 1.017605633802817, "grad_norm": 0.3857402205467224, "learning_rate": 9.864626250735727e-05, "loss": 0.7321, "step": 867 }, { "epoch": 1.0187793427230047, "grad_norm": 0.3688738942146301, "learning_rate": 9.852854620364921e-05, "loss": 0.6853, "step": 868 }, { "epoch": 1.0199530516431925, "grad_norm": 0.3814820647239685, "learning_rate": 9.841082989994114e-05, "loss": 0.664, "step": 869 }, { "epoch": 1.0211267605633803, "grad_norm": 0.3849344253540039, "learning_rate": 9.829311359623309e-05, "loss": 0.6844, "step": 870 }, { "epoch": 1.022300469483568, "grad_norm": 0.36203038692474365, "learning_rate": 9.817539729252502e-05, "loss": 0.7201, "step": 871 }, { "epoch": 1.0234741784037558, "grad_norm": 0.36614471673965454, "learning_rate": 9.805768098881696e-05, "loss": 0.659, "step": 872 }, { "epoch": 1.0246478873239437, "grad_norm": 0.3908173143863678, "learning_rate": 9.79399646851089e-05, "loss": 0.6638, "step": 873 }, { "epoch": 1.0258215962441315, "grad_norm": 0.35966452956199646, "learning_rate": 9.782224838140083e-05, "loss": 0.7187, "step": 874 }, { "epoch": 1.0269953051643192, "grad_norm": 0.40878093242645264, "learning_rate": 9.770453207769276e-05, "loss": 0.691, "step": 875 }, { "epoch": 1.028169014084507, "grad_norm": 0.38903382420539856, "learning_rate": 9.75868157739847e-05, "loss": 0.718, "step": 876 }, { "epoch": 1.0293427230046948, "grad_norm": 0.3865324556827545, "learning_rate": 9.746909947027663e-05, "loss": 0.7331, "step": 877 }, { "epoch": 1.0305164319248827, "grad_norm": 0.37417513132095337, "learning_rate": 9.735138316656858e-05, "loss": 0.677, "step": 878 }, { "epoch": 1.0316901408450705, "grad_norm": 0.38043439388275146, "learning_rate": 9.72336668628605e-05, "loss": 0.6932, "step": 879 }, { "epoch": 1.0328638497652582, "grad_norm": 0.37418729066848755, "learning_rate": 9.711595055915245e-05, "loss": 0.7119, "step": 880 }, { "epoch": 1.034037558685446, "grad_norm": 0.4013047218322754, "learning_rate": 9.699823425544438e-05, "loss": 0.7041, "step": 881 }, { "epoch": 1.0352112676056338, "grad_norm": 0.38462570309638977, "learning_rate": 9.688051795173632e-05, "loss": 0.6861, "step": 882 }, { "epoch": 1.0363849765258215, "grad_norm": 0.3900148868560791, "learning_rate": 9.676280164802825e-05, "loss": 0.6382, "step": 883 }, { "epoch": 1.0375586854460095, "grad_norm": 0.3882652819156647, "learning_rate": 9.66450853443202e-05, "loss": 0.6948, "step": 884 }, { "epoch": 1.0387323943661972, "grad_norm": 0.36546608805656433, "learning_rate": 9.652736904061212e-05, "loss": 0.7064, "step": 885 }, { "epoch": 1.039906103286385, "grad_norm": 0.3788559138774872, "learning_rate": 9.640965273690407e-05, "loss": 0.7129, "step": 886 }, { "epoch": 1.0410798122065728, "grad_norm": 0.3979467451572418, "learning_rate": 9.6291936433196e-05, "loss": 0.7196, "step": 887 }, { "epoch": 1.0422535211267605, "grad_norm": 0.3777488172054291, "learning_rate": 9.617422012948794e-05, "loss": 0.6922, "step": 888 }, { "epoch": 1.0434272300469483, "grad_norm": 0.39730504155158997, "learning_rate": 9.605650382577987e-05, "loss": 0.6529, "step": 889 }, { "epoch": 1.0446009389671362, "grad_norm": 0.39619576930999756, "learning_rate": 9.593878752207182e-05, "loss": 0.6505, "step": 890 }, { "epoch": 1.045774647887324, "grad_norm": 0.3763888478279114, "learning_rate": 9.582107121836374e-05, "loss": 0.638, "step": 891 }, { "epoch": 1.0469483568075117, "grad_norm": 0.3947450518608093, "learning_rate": 9.570335491465569e-05, "loss": 0.7099, "step": 892 }, { "epoch": 1.0481220657276995, "grad_norm": 0.43239885568618774, "learning_rate": 9.558563861094763e-05, "loss": 0.7112, "step": 893 }, { "epoch": 1.0492957746478873, "grad_norm": 0.37725165486335754, "learning_rate": 9.546792230723956e-05, "loss": 0.6775, "step": 894 }, { "epoch": 1.050469483568075, "grad_norm": 0.3807140290737152, "learning_rate": 9.53502060035315e-05, "loss": 0.7201, "step": 895 }, { "epoch": 1.051643192488263, "grad_norm": 0.40270236134529114, "learning_rate": 9.523248969982343e-05, "loss": 0.6908, "step": 896 }, { "epoch": 1.0528169014084507, "grad_norm": 0.38907137513160706, "learning_rate": 9.511477339611536e-05, "loss": 0.7274, "step": 897 }, { "epoch": 1.0539906103286385, "grad_norm": 0.35074397921562195, "learning_rate": 9.49970570924073e-05, "loss": 0.6765, "step": 898 }, { "epoch": 1.0551643192488263, "grad_norm": 0.37548649311065674, "learning_rate": 9.487934078869925e-05, "loss": 0.7258, "step": 899 }, { "epoch": 1.056338028169014, "grad_norm": 0.3947518467903137, "learning_rate": 9.476162448499118e-05, "loss": 0.7142, "step": 900 }, { "epoch": 1.057511737089202, "grad_norm": 0.36888387799263, "learning_rate": 9.464390818128312e-05, "loss": 0.664, "step": 901 }, { "epoch": 1.0586854460093897, "grad_norm": 0.3735831379890442, "learning_rate": 9.452619187757505e-05, "loss": 0.6914, "step": 902 }, { "epoch": 1.0598591549295775, "grad_norm": 0.3840358257293701, "learning_rate": 9.440847557386698e-05, "loss": 0.663, "step": 903 }, { "epoch": 1.0610328638497653, "grad_norm": 0.408840537071228, "learning_rate": 9.429075927015892e-05, "loss": 0.7225, "step": 904 }, { "epoch": 1.062206572769953, "grad_norm": 0.36408165097236633, "learning_rate": 9.417304296645085e-05, "loss": 0.6744, "step": 905 }, { "epoch": 1.0633802816901408, "grad_norm": 0.4005196690559387, "learning_rate": 9.405532666274279e-05, "loss": 0.7285, "step": 906 }, { "epoch": 1.0645539906103287, "grad_norm": 0.3824830949306488, "learning_rate": 9.393761035903474e-05, "loss": 0.6978, "step": 907 }, { "epoch": 1.0657276995305165, "grad_norm": 0.38410818576812744, "learning_rate": 9.381989405532666e-05, "loss": 0.6725, "step": 908 }, { "epoch": 1.0669014084507042, "grad_norm": 0.37026217579841614, "learning_rate": 9.37021777516186e-05, "loss": 0.6908, "step": 909 }, { "epoch": 1.068075117370892, "grad_norm": 0.37652963399887085, "learning_rate": 9.358446144791054e-05, "loss": 0.6674, "step": 910 }, { "epoch": 1.0692488262910798, "grad_norm": 0.40584585070610046, "learning_rate": 9.346674514420247e-05, "loss": 0.7087, "step": 911 }, { "epoch": 1.0704225352112675, "grad_norm": 0.3777616620063782, "learning_rate": 9.334902884049441e-05, "loss": 0.6633, "step": 912 }, { "epoch": 1.0715962441314555, "grad_norm": 0.35584181547164917, "learning_rate": 9.323131253678636e-05, "loss": 0.6484, "step": 913 }, { "epoch": 1.0727699530516432, "grad_norm": 0.40920573472976685, "learning_rate": 9.311359623307828e-05, "loss": 0.6781, "step": 914 }, { "epoch": 1.073943661971831, "grad_norm": 0.37617766857147217, "learning_rate": 9.299587992937023e-05, "loss": 0.6785, "step": 915 }, { "epoch": 1.0751173708920188, "grad_norm": 0.36854755878448486, "learning_rate": 9.287816362566216e-05, "loss": 0.6805, "step": 916 }, { "epoch": 1.0762910798122065, "grad_norm": 0.3820021152496338, "learning_rate": 9.27604473219541e-05, "loss": 0.7413, "step": 917 }, { "epoch": 1.0774647887323943, "grad_norm": 0.3654205799102783, "learning_rate": 9.264273101824603e-05, "loss": 0.6996, "step": 918 }, { "epoch": 1.0786384976525822, "grad_norm": 0.36847448348999023, "learning_rate": 9.252501471453798e-05, "loss": 0.6593, "step": 919 }, { "epoch": 1.07981220657277, "grad_norm": 0.4072454571723938, "learning_rate": 9.24072984108299e-05, "loss": 0.7062, "step": 920 }, { "epoch": 1.0809859154929577, "grad_norm": 0.37201663851737976, "learning_rate": 9.228958210712185e-05, "loss": 0.7188, "step": 921 }, { "epoch": 1.0821596244131455, "grad_norm": 0.40708494186401367, "learning_rate": 9.217186580341378e-05, "loss": 0.6984, "step": 922 }, { "epoch": 1.0833333333333333, "grad_norm": 0.37668758630752563, "learning_rate": 9.205414949970572e-05, "loss": 0.6856, "step": 923 }, { "epoch": 1.084507042253521, "grad_norm": 0.41518712043762207, "learning_rate": 9.193643319599765e-05, "loss": 0.7093, "step": 924 }, { "epoch": 1.085680751173709, "grad_norm": 0.3661474883556366, "learning_rate": 9.181871689228958e-05, "loss": 0.6765, "step": 925 }, { "epoch": 1.0868544600938967, "grad_norm": 0.3910673260688782, "learning_rate": 9.170100058858152e-05, "loss": 0.6778, "step": 926 }, { "epoch": 1.0880281690140845, "grad_norm": 0.3851100206375122, "learning_rate": 9.158328428487345e-05, "loss": 0.7188, "step": 927 }, { "epoch": 1.0892018779342723, "grad_norm": 0.36254799365997314, "learning_rate": 9.14655679811654e-05, "loss": 0.7182, "step": 928 }, { "epoch": 1.09037558685446, "grad_norm": 0.39364567399024963, "learning_rate": 9.134785167745734e-05, "loss": 0.7208, "step": 929 }, { "epoch": 1.091549295774648, "grad_norm": 0.3755466639995575, "learning_rate": 9.123013537374927e-05, "loss": 0.6771, "step": 930 }, { "epoch": 1.0927230046948357, "grad_norm": 0.361087828874588, "learning_rate": 9.11124190700412e-05, "loss": 0.6541, "step": 931 }, { "epoch": 1.0938967136150235, "grad_norm": 0.37327754497528076, "learning_rate": 9.099470276633314e-05, "loss": 0.698, "step": 932 }, { "epoch": 1.0950704225352113, "grad_norm": 0.38413748145103455, "learning_rate": 9.087698646262507e-05, "loss": 0.6933, "step": 933 }, { "epoch": 1.096244131455399, "grad_norm": 0.4182147681713104, "learning_rate": 9.075927015891701e-05, "loss": 0.6776, "step": 934 }, { "epoch": 1.0974178403755868, "grad_norm": 0.3987724483013153, "learning_rate": 9.064155385520894e-05, "loss": 0.694, "step": 935 }, { "epoch": 1.0985915492957747, "grad_norm": 0.37629225850105286, "learning_rate": 9.052383755150089e-05, "loss": 0.6565, "step": 936 }, { "epoch": 1.0997652582159625, "grad_norm": 0.38973352313041687, "learning_rate": 9.040612124779281e-05, "loss": 0.6739, "step": 937 }, { "epoch": 1.1009389671361502, "grad_norm": 0.3845914900302887, "learning_rate": 9.028840494408476e-05, "loss": 0.6788, "step": 938 }, { "epoch": 1.102112676056338, "grad_norm": 0.3861023485660553, "learning_rate": 9.01706886403767e-05, "loss": 0.6763, "step": 939 }, { "epoch": 1.1032863849765258, "grad_norm": 0.37565183639526367, "learning_rate": 9.005297233666863e-05, "loss": 0.6478, "step": 940 }, { "epoch": 1.1044600938967135, "grad_norm": 0.4068315029144287, "learning_rate": 8.993525603296056e-05, "loss": 0.6752, "step": 941 }, { "epoch": 1.1056338028169015, "grad_norm": 0.37796974182128906, "learning_rate": 8.981753972925251e-05, "loss": 0.7355, "step": 942 }, { "epoch": 1.1068075117370892, "grad_norm": 0.4024117887020111, "learning_rate": 8.969982342554443e-05, "loss": 0.6648, "step": 943 }, { "epoch": 1.107981220657277, "grad_norm": 0.404442697763443, "learning_rate": 8.958210712183638e-05, "loss": 0.7, "step": 944 }, { "epoch": 1.1091549295774648, "grad_norm": 0.35948899388313293, "learning_rate": 8.946439081812832e-05, "loss": 0.6859, "step": 945 }, { "epoch": 1.1103286384976525, "grad_norm": 0.4014012813568115, "learning_rate": 8.934667451442025e-05, "loss": 0.7294, "step": 946 }, { "epoch": 1.1115023474178405, "grad_norm": 0.38261109590530396, "learning_rate": 8.922895821071219e-05, "loss": 0.6965, "step": 947 }, { "epoch": 1.1126760563380282, "grad_norm": 0.39297208189964294, "learning_rate": 8.911124190700413e-05, "loss": 0.7153, "step": 948 }, { "epoch": 1.113849765258216, "grad_norm": 0.3710176348686218, "learning_rate": 8.899352560329605e-05, "loss": 0.7085, "step": 949 }, { "epoch": 1.1150234741784038, "grad_norm": 0.3750080168247223, "learning_rate": 8.8875809299588e-05, "loss": 0.6739, "step": 950 }, { "epoch": 1.1161971830985915, "grad_norm": 0.3672105073928833, "learning_rate": 8.875809299587994e-05, "loss": 0.7097, "step": 951 }, { "epoch": 1.1173708920187793, "grad_norm": 0.3663265109062195, "learning_rate": 8.864037669217187e-05, "loss": 0.6594, "step": 952 }, { "epoch": 1.1185446009389672, "grad_norm": 0.4023442268371582, "learning_rate": 8.85226603884638e-05, "loss": 0.7186, "step": 953 }, { "epoch": 1.119718309859155, "grad_norm": 0.36602139472961426, "learning_rate": 8.840494408475574e-05, "loss": 0.67, "step": 954 }, { "epoch": 1.1208920187793427, "grad_norm": 0.36866381764411926, "learning_rate": 8.828722778104768e-05, "loss": 0.6954, "step": 955 }, { "epoch": 1.1220657276995305, "grad_norm": 0.38905832171440125, "learning_rate": 8.816951147733961e-05, "loss": 0.7214, "step": 956 }, { "epoch": 1.1232394366197183, "grad_norm": 0.3806670010089874, "learning_rate": 8.805179517363156e-05, "loss": 0.6679, "step": 957 }, { "epoch": 1.124413145539906, "grad_norm": 0.3796343505382538, "learning_rate": 8.793407886992349e-05, "loss": 0.6334, "step": 958 }, { "epoch": 1.125586854460094, "grad_norm": 0.4143288731575012, "learning_rate": 8.781636256621543e-05, "loss": 0.7484, "step": 959 }, { "epoch": 1.1267605633802817, "grad_norm": 0.3692832887172699, "learning_rate": 8.769864626250736e-05, "loss": 0.6581, "step": 960 }, { "epoch": 1.1279342723004695, "grad_norm": 0.39971667528152466, "learning_rate": 8.75809299587993e-05, "loss": 0.7252, "step": 961 }, { "epoch": 1.1291079812206573, "grad_norm": 0.391924113035202, "learning_rate": 8.746321365509123e-05, "loss": 0.673, "step": 962 }, { "epoch": 1.130281690140845, "grad_norm": 0.39626866579055786, "learning_rate": 8.734549735138317e-05, "loss": 0.7161, "step": 963 }, { "epoch": 1.131455399061033, "grad_norm": 0.3812800347805023, "learning_rate": 8.72277810476751e-05, "loss": 0.6735, "step": 964 }, { "epoch": 1.1326291079812207, "grad_norm": 0.36054447293281555, "learning_rate": 8.711006474396705e-05, "loss": 0.6861, "step": 965 }, { "epoch": 1.1338028169014085, "grad_norm": 0.41179588437080383, "learning_rate": 8.699234844025897e-05, "loss": 0.7151, "step": 966 }, { "epoch": 1.1349765258215962, "grad_norm": 0.3688051998615265, "learning_rate": 8.687463213655092e-05, "loss": 0.6608, "step": 967 }, { "epoch": 1.136150234741784, "grad_norm": 0.3877013325691223, "learning_rate": 8.675691583284285e-05, "loss": 0.6826, "step": 968 }, { "epoch": 1.1373239436619718, "grad_norm": 0.38986387848854065, "learning_rate": 8.663919952913479e-05, "loss": 0.6915, "step": 969 }, { "epoch": 1.1384976525821595, "grad_norm": 0.41986656188964844, "learning_rate": 8.652148322542672e-05, "loss": 0.7471, "step": 970 }, { "epoch": 1.1396713615023475, "grad_norm": 0.3977747857570648, "learning_rate": 8.640376692171867e-05, "loss": 0.6844, "step": 971 }, { "epoch": 1.1408450704225352, "grad_norm": 0.3956218361854553, "learning_rate": 8.628605061801059e-05, "loss": 0.6586, "step": 972 }, { "epoch": 1.142018779342723, "grad_norm": 0.3789028227329254, "learning_rate": 8.616833431430254e-05, "loss": 0.7415, "step": 973 }, { "epoch": 1.1431924882629108, "grad_norm": 0.3878764808177948, "learning_rate": 8.605061801059447e-05, "loss": 0.6559, "step": 974 }, { "epoch": 1.1443661971830985, "grad_norm": 0.37901559472084045, "learning_rate": 8.593290170688641e-05, "loss": 0.6685, "step": 975 }, { "epoch": 1.1455399061032865, "grad_norm": 0.40399041771888733, "learning_rate": 8.581518540317834e-05, "loss": 0.6602, "step": 976 }, { "epoch": 1.1467136150234742, "grad_norm": 0.38144391775131226, "learning_rate": 8.569746909947029e-05, "loss": 0.6683, "step": 977 }, { "epoch": 1.147887323943662, "grad_norm": 0.3610433042049408, "learning_rate": 8.557975279576221e-05, "loss": 0.6579, "step": 978 }, { "epoch": 1.1490610328638498, "grad_norm": 0.42147722840309143, "learning_rate": 8.546203649205416e-05, "loss": 0.6997, "step": 979 }, { "epoch": 1.1502347417840375, "grad_norm": 0.3799455761909485, "learning_rate": 8.53443201883461e-05, "loss": 0.7096, "step": 980 }, { "epoch": 1.1514084507042253, "grad_norm": 0.4173739552497864, "learning_rate": 8.522660388463803e-05, "loss": 0.6708, "step": 981 }, { "epoch": 1.1525821596244132, "grad_norm": 0.3997640013694763, "learning_rate": 8.510888758092996e-05, "loss": 0.6514, "step": 982 }, { "epoch": 1.153755868544601, "grad_norm": 0.3758656978607178, "learning_rate": 8.49911712772219e-05, "loss": 0.6442, "step": 983 }, { "epoch": 1.1549295774647887, "grad_norm": 0.37429675459861755, "learning_rate": 8.487345497351383e-05, "loss": 0.6619, "step": 984 }, { "epoch": 1.1561032863849765, "grad_norm": 0.3747265934944153, "learning_rate": 8.475573866980577e-05, "loss": 0.7107, "step": 985 }, { "epoch": 1.1572769953051643, "grad_norm": 0.37782514095306396, "learning_rate": 8.463802236609771e-05, "loss": 0.7241, "step": 986 }, { "epoch": 1.158450704225352, "grad_norm": 0.3703122138977051, "learning_rate": 8.452030606238965e-05, "loss": 0.6952, "step": 987 }, { "epoch": 1.15962441314554, "grad_norm": 0.37990477681159973, "learning_rate": 8.440258975868158e-05, "loss": 0.7364, "step": 988 }, { "epoch": 1.1607981220657277, "grad_norm": 0.42046844959259033, "learning_rate": 8.428487345497352e-05, "loss": 0.695, "step": 989 }, { "epoch": 1.1619718309859155, "grad_norm": 0.3745966851711273, "learning_rate": 8.416715715126545e-05, "loss": 0.6875, "step": 990 }, { "epoch": 1.1631455399061033, "grad_norm": 0.3496320843696594, "learning_rate": 8.404944084755739e-05, "loss": 0.6826, "step": 991 }, { "epoch": 1.164319248826291, "grad_norm": 0.39181873202323914, "learning_rate": 8.393172454384934e-05, "loss": 0.6937, "step": 992 }, { "epoch": 1.165492957746479, "grad_norm": 0.3910543620586395, "learning_rate": 8.381400824014126e-05, "loss": 0.749, "step": 993 }, { "epoch": 1.1666666666666667, "grad_norm": 0.3770748972892761, "learning_rate": 8.36962919364332e-05, "loss": 0.6743, "step": 994 }, { "epoch": 1.1678403755868545, "grad_norm": 0.3675018846988678, "learning_rate": 8.357857563272513e-05, "loss": 0.6499, "step": 995 }, { "epoch": 1.1690140845070423, "grad_norm": 0.36867639422416687, "learning_rate": 8.346085932901707e-05, "loss": 0.6642, "step": 996 }, { "epoch": 1.17018779342723, "grad_norm": 0.3860320746898651, "learning_rate": 8.334314302530901e-05, "loss": 0.6947, "step": 997 }, { "epoch": 1.1713615023474178, "grad_norm": 0.36680731177330017, "learning_rate": 8.322542672160094e-05, "loss": 0.7111, "step": 998 }, { "epoch": 1.1725352112676055, "grad_norm": 0.38997524976730347, "learning_rate": 8.310771041789288e-05, "loss": 0.6842, "step": 999 }, { "epoch": 1.1737089201877935, "grad_norm": 0.3883102834224701, "learning_rate": 8.298999411418483e-05, "loss": 0.6655, "step": 1000 } ], "logging_steps": 1, "max_steps": 1704, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 100, "total_flos": 1.4844237184940114e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }