Flor-6.3B-Instruct / trainer_state.json
Ankush
add all files
f95fcc5
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9996375498368975,
"eval_steps": 345,
"global_step": 1379,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0007249003262051468,
"grad_norm": 44.75750397895495,
"learning_rate": 4.7619047619047623e-07,
"loss": 2.3443,
"step": 1
},
{
"epoch": 0.0014498006524102935,
"grad_norm": 45.205818482220735,
"learning_rate": 9.523809523809525e-07,
"loss": 2.1445,
"step": 2
},
{
"epoch": 0.0021747009786154403,
"grad_norm": 56.724033240854,
"learning_rate": 1.4285714285714286e-06,
"loss": 2.5876,
"step": 3
},
{
"epoch": 0.002899601304820587,
"grad_norm": 53.17611331537967,
"learning_rate": 1.904761904761905e-06,
"loss": 2.3983,
"step": 4
},
{
"epoch": 0.003624501631025734,
"grad_norm": 57.8664359648513,
"learning_rate": 2.380952380952381e-06,
"loss": 2.1637,
"step": 5
},
{
"epoch": 0.004349401957230881,
"grad_norm": 14.246129965796554,
"learning_rate": 2.8571428571428573e-06,
"loss": 1.9145,
"step": 6
},
{
"epoch": 0.005074302283436027,
"grad_norm": 11.159860869255928,
"learning_rate": 3.3333333333333333e-06,
"loss": 1.7268,
"step": 7
},
{
"epoch": 0.005799202609641174,
"grad_norm": 8.887448463820203,
"learning_rate": 3.80952380952381e-06,
"loss": 1.5522,
"step": 8
},
{
"epoch": 0.006524102935846321,
"grad_norm": 6.753347075425067,
"learning_rate": 4.2857142857142855e-06,
"loss": 1.5248,
"step": 9
},
{
"epoch": 0.007249003262051468,
"grad_norm": 11.059402659834483,
"learning_rate": 4.761904761904762e-06,
"loss": 1.5304,
"step": 10
},
{
"epoch": 0.007973903588256614,
"grad_norm": 9.28275165512847,
"learning_rate": 5.2380952380952384e-06,
"loss": 1.4316,
"step": 11
},
{
"epoch": 0.008698803914461761,
"grad_norm": 4.714004999293759,
"learning_rate": 5.7142857142857145e-06,
"loss": 1.331,
"step": 12
},
{
"epoch": 0.009423704240666908,
"grad_norm": 4.708567041689236,
"learning_rate": 6.1904761904761914e-06,
"loss": 1.4035,
"step": 13
},
{
"epoch": 0.010148604566872055,
"grad_norm": 7.8340994412560665,
"learning_rate": 6.666666666666667e-06,
"loss": 1.4499,
"step": 14
},
{
"epoch": 0.010873504893077202,
"grad_norm": 5.7926127365771105,
"learning_rate": 7.1428571428571436e-06,
"loss": 1.2915,
"step": 15
},
{
"epoch": 0.011598405219282348,
"grad_norm": 4.338940169766625,
"learning_rate": 7.61904761904762e-06,
"loss": 1.2546,
"step": 16
},
{
"epoch": 0.012323305545487495,
"grad_norm": 5.3158492494198875,
"learning_rate": 8.095238095238097e-06,
"loss": 1.3437,
"step": 17
},
{
"epoch": 0.013048205871692642,
"grad_norm": 4.267050487387928,
"learning_rate": 8.571428571428571e-06,
"loss": 1.2076,
"step": 18
},
{
"epoch": 0.013773106197897789,
"grad_norm": 4.289501496589146,
"learning_rate": 9.047619047619049e-06,
"loss": 1.2095,
"step": 19
},
{
"epoch": 0.014498006524102935,
"grad_norm": 4.475865685170405,
"learning_rate": 9.523809523809525e-06,
"loss": 1.2499,
"step": 20
},
{
"epoch": 0.015222906850308082,
"grad_norm": 4.341512982701414,
"learning_rate": 1e-05,
"loss": 1.3725,
"step": 21
},
{
"epoch": 0.01594780717651323,
"grad_norm": 3.9777632208757443,
"learning_rate": 1.0476190476190477e-05,
"loss": 1.2679,
"step": 22
},
{
"epoch": 0.016672707502718376,
"grad_norm": 3.971780990603446,
"learning_rate": 1.0952380952380955e-05,
"loss": 1.3049,
"step": 23
},
{
"epoch": 0.017397607828923523,
"grad_norm": 4.355148301567439,
"learning_rate": 1.1428571428571429e-05,
"loss": 1.1696,
"step": 24
},
{
"epoch": 0.01812250815512867,
"grad_norm": 5.700836460033174,
"learning_rate": 1.1904761904761905e-05,
"loss": 1.3116,
"step": 25
},
{
"epoch": 0.018847408481333816,
"grad_norm": 4.266676186480929,
"learning_rate": 1.2380952380952383e-05,
"loss": 1.2132,
"step": 26
},
{
"epoch": 0.019572308807538963,
"grad_norm": 4.685688044042141,
"learning_rate": 1.2857142857142859e-05,
"loss": 1.2743,
"step": 27
},
{
"epoch": 0.02029720913374411,
"grad_norm": 3.8278571816897977,
"learning_rate": 1.3333333333333333e-05,
"loss": 1.3062,
"step": 28
},
{
"epoch": 0.021022109459949256,
"grad_norm": 4.207466080048964,
"learning_rate": 1.3809523809523811e-05,
"loss": 1.2238,
"step": 29
},
{
"epoch": 0.021747009786154403,
"grad_norm": 3.772755917601343,
"learning_rate": 1.4285714285714287e-05,
"loss": 1.2333,
"step": 30
},
{
"epoch": 0.02247191011235955,
"grad_norm": 3.2903338340660095,
"learning_rate": 1.4761904761904763e-05,
"loss": 1.1223,
"step": 31
},
{
"epoch": 0.023196810438564697,
"grad_norm": 3.2688442989444746,
"learning_rate": 1.523809523809524e-05,
"loss": 1.1476,
"step": 32
},
{
"epoch": 0.023921710764769843,
"grad_norm": 3.6401775542496284,
"learning_rate": 1.5714285714285715e-05,
"loss": 1.1805,
"step": 33
},
{
"epoch": 0.02464661109097499,
"grad_norm": 3.4329875974688004,
"learning_rate": 1.6190476190476193e-05,
"loss": 1.1406,
"step": 34
},
{
"epoch": 0.025371511417180137,
"grad_norm": 3.457570825026417,
"learning_rate": 1.6666666666666667e-05,
"loss": 1.1447,
"step": 35
},
{
"epoch": 0.026096411743385284,
"grad_norm": 3.503397413974539,
"learning_rate": 1.7142857142857142e-05,
"loss": 1.13,
"step": 36
},
{
"epoch": 0.02682131206959043,
"grad_norm": 2.880316116737165,
"learning_rate": 1.761904761904762e-05,
"loss": 1.2275,
"step": 37
},
{
"epoch": 0.027546212395795577,
"grad_norm": 3.4491978928897686,
"learning_rate": 1.8095238095238097e-05,
"loss": 1.2675,
"step": 38
},
{
"epoch": 0.028271112722000724,
"grad_norm": 3.4801420565131695,
"learning_rate": 1.8571428571428575e-05,
"loss": 1.1656,
"step": 39
},
{
"epoch": 0.02899601304820587,
"grad_norm": 3.154782654672038,
"learning_rate": 1.904761904761905e-05,
"loss": 1.157,
"step": 40
},
{
"epoch": 0.029720913374411018,
"grad_norm": 3.0915926527583633,
"learning_rate": 1.9523809523809524e-05,
"loss": 1.1432,
"step": 41
},
{
"epoch": 0.030445813700616164,
"grad_norm": 2.772154104430272,
"learning_rate": 2e-05,
"loss": 1.0994,
"step": 42
},
{
"epoch": 0.03117071402682131,
"grad_norm": 3.0517198878965304,
"learning_rate": 1.9999972393793303e-05,
"loss": 1.2549,
"step": 43
},
{
"epoch": 0.03189561435302646,
"grad_norm": 2.6750901532162987,
"learning_rate": 1.9999889575325633e-05,
"loss": 1.143,
"step": 44
},
{
"epoch": 0.032620514679231605,
"grad_norm": 3.6101390652119516,
"learning_rate": 1.9999751545054247e-05,
"loss": 1.2469,
"step": 45
},
{
"epoch": 0.03334541500543675,
"grad_norm": 2.843300653538732,
"learning_rate": 1.9999558303741244e-05,
"loss": 1.061,
"step": 46
},
{
"epoch": 0.0340703153316419,
"grad_norm": 3.9392334069856907,
"learning_rate": 1.9999309852453556e-05,
"loss": 1.2809,
"step": 47
},
{
"epoch": 0.034795215657847045,
"grad_norm": 3.773339147888779,
"learning_rate": 1.9999006192562948e-05,
"loss": 1.3008,
"step": 48
},
{
"epoch": 0.03552011598405219,
"grad_norm": 2.7832080749846453,
"learning_rate": 1.9998647325745995e-05,
"loss": 1.1342,
"step": 49
},
{
"epoch": 0.03624501631025734,
"grad_norm": 3.7382629441207667,
"learning_rate": 1.9998233253984088e-05,
"loss": 1.2641,
"step": 50
},
{
"epoch": 0.036969916636462485,
"grad_norm": 2.799440280567736,
"learning_rate": 1.9997763979563418e-05,
"loss": 1.2242,
"step": 51
},
{
"epoch": 0.03769481696266763,
"grad_norm": 3.3316333443131696,
"learning_rate": 1.9997239505074954e-05,
"loss": 1.2312,
"step": 52
},
{
"epoch": 0.03841971728887278,
"grad_norm": 5.725082920380033,
"learning_rate": 1.999665983341446e-05,
"loss": 1.4632,
"step": 53
},
{
"epoch": 0.039144617615077926,
"grad_norm": 2.553768110728692,
"learning_rate": 1.9996024967782436e-05,
"loss": 1.2164,
"step": 54
},
{
"epoch": 0.03986951794128307,
"grad_norm": 3.041402503331704,
"learning_rate": 1.9995334911684127e-05,
"loss": 1.2156,
"step": 55
},
{
"epoch": 0.04059441826748822,
"grad_norm": 2.595475331293293,
"learning_rate": 1.99945896689295e-05,
"loss": 1.1247,
"step": 56
},
{
"epoch": 0.041319318593693366,
"grad_norm": 2.939590634487055,
"learning_rate": 1.9993789243633227e-05,
"loss": 1.2171,
"step": 57
},
{
"epoch": 0.04204421891989851,
"grad_norm": 2.6474772321680606,
"learning_rate": 1.999293364021464e-05,
"loss": 1.187,
"step": 58
},
{
"epoch": 0.04276911924610366,
"grad_norm": 2.901881436981748,
"learning_rate": 1.9992022863397737e-05,
"loss": 1.1224,
"step": 59
},
{
"epoch": 0.043494019572308806,
"grad_norm": 2.9320116081953835,
"learning_rate": 1.999105691821113e-05,
"loss": 1.1692,
"step": 60
},
{
"epoch": 0.04421891989851395,
"grad_norm": 3.5247973111889004,
"learning_rate": 1.9990035809988045e-05,
"loss": 1.1575,
"step": 61
},
{
"epoch": 0.0449438202247191,
"grad_norm": 2.669047070211342,
"learning_rate": 1.9988959544366265e-05,
"loss": 1.1433,
"step": 62
},
{
"epoch": 0.04566872055092425,
"grad_norm": 3.213177659539329,
"learning_rate": 1.9987828127288105e-05,
"loss": 1.2635,
"step": 63
},
{
"epoch": 0.04639362087712939,
"grad_norm": 3.3559565945133554,
"learning_rate": 1.99866415650004e-05,
"loss": 1.3022,
"step": 64
},
{
"epoch": 0.04711852120333454,
"grad_norm": 2.3670279547756894,
"learning_rate": 1.998539986405444e-05,
"loss": 1.0142,
"step": 65
},
{
"epoch": 0.04784342152953969,
"grad_norm": 3.037710920162538,
"learning_rate": 1.998410303130596e-05,
"loss": 1.2591,
"step": 66
},
{
"epoch": 0.048568321855744834,
"grad_norm": 2.865611039939188,
"learning_rate": 1.9982751073915087e-05,
"loss": 1.1993,
"step": 67
},
{
"epoch": 0.04929322218194998,
"grad_norm": 3.48123837358242,
"learning_rate": 1.9981343999346302e-05,
"loss": 1.2121,
"step": 68
},
{
"epoch": 0.05001812250815513,
"grad_norm": 3.2299518093080977,
"learning_rate": 1.9979881815368406e-05,
"loss": 1.2152,
"step": 69
},
{
"epoch": 0.050743022834360274,
"grad_norm": 3.0034023909101233,
"learning_rate": 1.9978364530054465e-05,
"loss": 1.1339,
"step": 70
},
{
"epoch": 0.05146792316056542,
"grad_norm": 3.4428257188696794,
"learning_rate": 1.9976792151781778e-05,
"loss": 1.2363,
"step": 71
},
{
"epoch": 0.05219282348677057,
"grad_norm": 2.9622416635006408,
"learning_rate": 1.997516468923183e-05,
"loss": 1.274,
"step": 72
},
{
"epoch": 0.052917723812975714,
"grad_norm": 2.5702004303735455,
"learning_rate": 1.9973482151390228e-05,
"loss": 1.1873,
"step": 73
},
{
"epoch": 0.05364262413918086,
"grad_norm": 3.3076211834809492,
"learning_rate": 1.9971744547546676e-05,
"loss": 1.228,
"step": 74
},
{
"epoch": 0.05436752446538601,
"grad_norm": 2.6091996272999243,
"learning_rate": 1.99699518872949e-05,
"loss": 1.0701,
"step": 75
},
{
"epoch": 0.055092424791591155,
"grad_norm": 2.8268323348853044,
"learning_rate": 1.996810418053261e-05,
"loss": 1.2859,
"step": 76
},
{
"epoch": 0.0558173251177963,
"grad_norm": 2.5281726771474244,
"learning_rate": 1.996620143746144e-05,
"loss": 1.1387,
"step": 77
},
{
"epoch": 0.05654222544400145,
"grad_norm": 2.300493492648369,
"learning_rate": 1.99642436685869e-05,
"loss": 1.0669,
"step": 78
},
{
"epoch": 0.057267125770206595,
"grad_norm": 3.493223728132168,
"learning_rate": 1.9962230884718296e-05,
"loss": 1.1832,
"step": 79
},
{
"epoch": 0.05799202609641174,
"grad_norm": 2.7188006009604293,
"learning_rate": 1.9960163096968702e-05,
"loss": 1.1356,
"step": 80
},
{
"epoch": 0.05871692642261689,
"grad_norm": 2.7364057030270956,
"learning_rate": 1.9958040316754866e-05,
"loss": 1.2446,
"step": 81
},
{
"epoch": 0.059441826748822035,
"grad_norm": 2.7307874925286475,
"learning_rate": 1.995586255579717e-05,
"loss": 1.1808,
"step": 82
},
{
"epoch": 0.06016672707502718,
"grad_norm": 2.4215808753597226,
"learning_rate": 1.9953629826119562e-05,
"loss": 1.1579,
"step": 83
},
{
"epoch": 0.06089162740123233,
"grad_norm": 2.631267713200398,
"learning_rate": 1.9951342140049483e-05,
"loss": 1.1941,
"step": 84
},
{
"epoch": 0.061616527727437476,
"grad_norm": 2.673115849546717,
"learning_rate": 1.994899951021779e-05,
"loss": 1.2284,
"step": 85
},
{
"epoch": 0.06234142805364262,
"grad_norm": 2.6934406037224807,
"learning_rate": 1.9946601949558722e-05,
"loss": 1.206,
"step": 86
},
{
"epoch": 0.06306632837984777,
"grad_norm": 3.7302320786317673,
"learning_rate": 1.9944149471309777e-05,
"loss": 1.2247,
"step": 87
},
{
"epoch": 0.06379122870605292,
"grad_norm": 2.4550165493932363,
"learning_rate": 1.9941642089011684e-05,
"loss": 1.0596,
"step": 88
},
{
"epoch": 0.06451612903225806,
"grad_norm": 2.868687433143366,
"learning_rate": 1.9939079816508312e-05,
"loss": 1.1435,
"step": 89
},
{
"epoch": 0.06524102935846321,
"grad_norm": 3.0984151506536586,
"learning_rate": 1.9936462667946574e-05,
"loss": 1.1143,
"step": 90
},
{
"epoch": 0.06596592968466836,
"grad_norm": 2.483528934101709,
"learning_rate": 1.993379065777639e-05,
"loss": 1.1165,
"step": 91
},
{
"epoch": 0.0666908300108735,
"grad_norm": 2.4080631755195925,
"learning_rate": 1.9931063800750564e-05,
"loss": 1.1144,
"step": 92
},
{
"epoch": 0.06741573033707865,
"grad_norm": 2.8973853952983375,
"learning_rate": 1.9928282111924736e-05,
"loss": 1.2365,
"step": 93
},
{
"epoch": 0.0681406306632838,
"grad_norm": 3.3833221119081656,
"learning_rate": 1.9925445606657286e-05,
"loss": 1.2146,
"step": 94
},
{
"epoch": 0.06886553098948894,
"grad_norm": 2.908934165124817,
"learning_rate": 1.9922554300609237e-05,
"loss": 1.1916,
"step": 95
},
{
"epoch": 0.06959043131569409,
"grad_norm": 2.512496806932867,
"learning_rate": 1.991960820974419e-05,
"loss": 1.0935,
"step": 96
},
{
"epoch": 0.07031533164189924,
"grad_norm": 2.825963677244521,
"learning_rate": 1.991660735032822e-05,
"loss": 1.0931,
"step": 97
},
{
"epoch": 0.07104023196810438,
"grad_norm": 2.8818264863814296,
"learning_rate": 1.9913551738929803e-05,
"loss": 1.2153,
"step": 98
},
{
"epoch": 0.07176513229430953,
"grad_norm": 2.7275750720550147,
"learning_rate": 1.9910441392419702e-05,
"loss": 1.193,
"step": 99
},
{
"epoch": 0.07249003262051468,
"grad_norm": 3.462468824391245,
"learning_rate": 1.9907276327970892e-05,
"loss": 1.1469,
"step": 100
},
{
"epoch": 0.07321493294671982,
"grad_norm": 3.3377754324869224,
"learning_rate": 1.990405656305846e-05,
"loss": 1.2357,
"step": 101
},
{
"epoch": 0.07393983327292497,
"grad_norm": 2.839009968658976,
"learning_rate": 1.9900782115459503e-05,
"loss": 1.1848,
"step": 102
},
{
"epoch": 0.07466473359913012,
"grad_norm": 2.1391099705614063,
"learning_rate": 1.9897453003253035e-05,
"loss": 1.1225,
"step": 103
},
{
"epoch": 0.07538963392533526,
"grad_norm": 2.7719189493972243,
"learning_rate": 1.9894069244819893e-05,
"loss": 1.1538,
"step": 104
},
{
"epoch": 0.07611453425154041,
"grad_norm": 2.3940421507214933,
"learning_rate": 1.9890630858842614e-05,
"loss": 1.2852,
"step": 105
},
{
"epoch": 0.07683943457774556,
"grad_norm": 2.341837519488398,
"learning_rate": 1.9887137864305365e-05,
"loss": 1.1116,
"step": 106
},
{
"epoch": 0.0775643349039507,
"grad_norm": 2.403845819495617,
"learning_rate": 1.9883590280493815e-05,
"loss": 1.1509,
"step": 107
},
{
"epoch": 0.07828923523015585,
"grad_norm": 3.348129496258059,
"learning_rate": 1.9879988126995023e-05,
"loss": 1.1837,
"step": 108
},
{
"epoch": 0.079014135556361,
"grad_norm": 2.591536638052907,
"learning_rate": 1.9876331423697345e-05,
"loss": 1.1223,
"step": 109
},
{
"epoch": 0.07973903588256614,
"grad_norm": 2.3592029882899648,
"learning_rate": 1.9872620190790334e-05,
"loss": 1.143,
"step": 110
},
{
"epoch": 0.08046393620877129,
"grad_norm": 2.1392569469392346,
"learning_rate": 1.9868854448764594e-05,
"loss": 1.0607,
"step": 111
},
{
"epoch": 0.08118883653497644,
"grad_norm": 2.863527881373612,
"learning_rate": 1.9865034218411698e-05,
"loss": 1.1717,
"step": 112
},
{
"epoch": 0.08191373686118159,
"grad_norm": 2.924899234680409,
"learning_rate": 1.986115952082406e-05,
"loss": 1.2318,
"step": 113
},
{
"epoch": 0.08263863718738673,
"grad_norm": 2.433809686371776,
"learning_rate": 1.985723037739482e-05,
"loss": 1.2658,
"step": 114
},
{
"epoch": 0.08336353751359188,
"grad_norm": 2.8856283656848905,
"learning_rate": 1.9853246809817725e-05,
"loss": 1.3125,
"step": 115
},
{
"epoch": 0.08408843783979703,
"grad_norm": 2.5486082662488627,
"learning_rate": 1.9849208840087018e-05,
"loss": 1.1655,
"step": 116
},
{
"epoch": 0.08481333816600217,
"grad_norm": 2.769678541141219,
"learning_rate": 1.98451164904973e-05,
"loss": 1.1953,
"step": 117
},
{
"epoch": 0.08553823849220732,
"grad_norm": 2.473525065101367,
"learning_rate": 1.984096978364342e-05,
"loss": 1.1201,
"step": 118
},
{
"epoch": 0.08626313881841247,
"grad_norm": 2.7533405016925583,
"learning_rate": 1.9836768742420355e-05,
"loss": 1.2437,
"step": 119
},
{
"epoch": 0.08698803914461761,
"grad_norm": 2.4288996024121605,
"learning_rate": 1.983251339002306e-05,
"loss": 1.2333,
"step": 120
},
{
"epoch": 0.08771293947082276,
"grad_norm": 2.8989781107762616,
"learning_rate": 1.9828203749946362e-05,
"loss": 1.1184,
"step": 121
},
{
"epoch": 0.0884378397970279,
"grad_norm": 2.3092483374926527,
"learning_rate": 1.982383984598483e-05,
"loss": 1.0504,
"step": 122
},
{
"epoch": 0.08916274012323305,
"grad_norm": 2.714849936506903,
"learning_rate": 1.9819421702232624e-05,
"loss": 1.2218,
"step": 123
},
{
"epoch": 0.0898876404494382,
"grad_norm": 2.8105161825117286,
"learning_rate": 1.9814949343083383e-05,
"loss": 1.0007,
"step": 124
},
{
"epoch": 0.09061254077564335,
"grad_norm": 4.147738573133108,
"learning_rate": 1.981042279323009e-05,
"loss": 1.2543,
"step": 125
},
{
"epoch": 0.0913374411018485,
"grad_norm": 2.8310310979565783,
"learning_rate": 1.9805842077664913e-05,
"loss": 1.146,
"step": 126
},
{
"epoch": 0.09206234142805364,
"grad_norm": 3.0754511920919376,
"learning_rate": 1.9801207221679085e-05,
"loss": 1.203,
"step": 127
},
{
"epoch": 0.09278724175425879,
"grad_norm": 2.7063407512582787,
"learning_rate": 1.979651825086277e-05,
"loss": 1.0783,
"step": 128
},
{
"epoch": 0.09351214208046393,
"grad_norm": 2.511083096485902,
"learning_rate": 1.97917751911049e-05,
"loss": 1.1531,
"step": 129
},
{
"epoch": 0.09423704240666908,
"grad_norm": 2.4741075426355614,
"learning_rate": 1.9786978068593062e-05,
"loss": 1.1428,
"step": 130
},
{
"epoch": 0.09496194273287423,
"grad_norm": 2.5626898456386815,
"learning_rate": 1.978212690981332e-05,
"loss": 1.1239,
"step": 131
},
{
"epoch": 0.09568684305907937,
"grad_norm": 2.2391021429874263,
"learning_rate": 1.9777221741550096e-05,
"loss": 1.0702,
"step": 132
},
{
"epoch": 0.09641174338528452,
"grad_norm": 2.4369139499586074,
"learning_rate": 1.9772262590886006e-05,
"loss": 1.1692,
"step": 133
},
{
"epoch": 0.09713664371148967,
"grad_norm": 2.7781914212064747,
"learning_rate": 1.976724948520172e-05,
"loss": 1.1412,
"step": 134
},
{
"epoch": 0.09786154403769481,
"grad_norm": 2.5311023412287623,
"learning_rate": 1.9762182452175806e-05,
"loss": 1.2879,
"step": 135
},
{
"epoch": 0.09858644436389996,
"grad_norm": 3.2212943074374327,
"learning_rate": 1.9757061519784568e-05,
"loss": 1.1971,
"step": 136
},
{
"epoch": 0.09931134469010511,
"grad_norm": 2.549155447627049,
"learning_rate": 1.9751886716301914e-05,
"loss": 1.1439,
"step": 137
},
{
"epoch": 0.10003624501631025,
"grad_norm": 2.59925115990486,
"learning_rate": 1.9746658070299188e-05,
"loss": 1.1126,
"step": 138
},
{
"epoch": 0.1007611453425154,
"grad_norm": 3.1442581997348364,
"learning_rate": 1.9741375610645e-05,
"loss": 1.166,
"step": 139
},
{
"epoch": 0.10148604566872055,
"grad_norm": 2.7390978310814456,
"learning_rate": 1.9736039366505087e-05,
"loss": 1.1593,
"step": 140
},
{
"epoch": 0.1022109459949257,
"grad_norm": 2.68099301776624,
"learning_rate": 1.9730649367342137e-05,
"loss": 1.091,
"step": 141
},
{
"epoch": 0.10293584632113084,
"grad_norm": 2.5055718294224474,
"learning_rate": 1.9725205642915645e-05,
"loss": 1.1567,
"step": 142
},
{
"epoch": 0.10366074664733599,
"grad_norm": 2.4806609688516064,
"learning_rate": 1.971970822328172e-05,
"loss": 1.1877,
"step": 143
},
{
"epoch": 0.10438564697354114,
"grad_norm": 2.446019560687823,
"learning_rate": 1.9714157138792946e-05,
"loss": 0.8962,
"step": 144
},
{
"epoch": 0.10511054729974628,
"grad_norm": 2.6145362990605006,
"learning_rate": 1.9708552420098196e-05,
"loss": 1.1652,
"step": 145
},
{
"epoch": 0.10583544762595143,
"grad_norm": 2.8215724838180303,
"learning_rate": 1.9702894098142476e-05,
"loss": 1.1084,
"step": 146
},
{
"epoch": 0.10656034795215658,
"grad_norm": 2.3497120281958503,
"learning_rate": 1.969718220416675e-05,
"loss": 1.167,
"step": 147
},
{
"epoch": 0.10728524827836172,
"grad_norm": 2.583947968920025,
"learning_rate": 1.9691416769707763e-05,
"loss": 1.1908,
"step": 148
},
{
"epoch": 0.10801014860456687,
"grad_norm": 2.751597734547836,
"learning_rate": 1.968559782659787e-05,
"loss": 1.2345,
"step": 149
},
{
"epoch": 0.10873504893077202,
"grad_norm": 2.6976314838665685,
"learning_rate": 1.9679725406964853e-05,
"loss": 1.1939,
"step": 150
},
{
"epoch": 0.10945994925697716,
"grad_norm": 2.1954623751957905,
"learning_rate": 1.967379954323177e-05,
"loss": 1.1555,
"step": 151
},
{
"epoch": 0.11018484958318231,
"grad_norm": 2.7043983207505624,
"learning_rate": 1.9667820268116732e-05,
"loss": 1.1899,
"step": 152
},
{
"epoch": 0.11090974990938746,
"grad_norm": 2.076332249665599,
"learning_rate": 1.966178761463277e-05,
"loss": 1.1009,
"step": 153
},
{
"epoch": 0.1116346502355926,
"grad_norm": 2.3441526940711097,
"learning_rate": 1.965570161608762e-05,
"loss": 1.1319,
"step": 154
},
{
"epoch": 0.11235955056179775,
"grad_norm": 2.2285061151426113,
"learning_rate": 1.964956230608354e-05,
"loss": 1.1211,
"step": 155
},
{
"epoch": 0.1130844508880029,
"grad_norm": 2.3200577550605064,
"learning_rate": 1.9643369718517152e-05,
"loss": 1.1515,
"step": 156
},
{
"epoch": 0.11380935121420804,
"grad_norm": 2.5448094549925844,
"learning_rate": 1.9637123887579217e-05,
"loss": 1.302,
"step": 157
},
{
"epoch": 0.11453425154041319,
"grad_norm": 2.1389116260296714,
"learning_rate": 1.963082484775448e-05,
"loss": 1.1653,
"step": 158
},
{
"epoch": 0.11525915186661834,
"grad_norm": 2.752841944773758,
"learning_rate": 1.9624472633821464e-05,
"loss": 1.1178,
"step": 159
},
{
"epoch": 0.11598405219282348,
"grad_norm": 2.3858753840621345,
"learning_rate": 1.961806728085227e-05,
"loss": 1.0138,
"step": 160
},
{
"epoch": 0.11670895251902863,
"grad_norm": 2.6783021219358694,
"learning_rate": 1.9611608824212395e-05,
"loss": 1.2371,
"step": 161
},
{
"epoch": 0.11743385284523378,
"grad_norm": 2.626777858203357,
"learning_rate": 1.9605097299560545e-05,
"loss": 1.0034,
"step": 162
},
{
"epoch": 0.11815875317143892,
"grad_norm": 2.333729825834433,
"learning_rate": 1.959853274284841e-05,
"loss": 1.1103,
"step": 163
},
{
"epoch": 0.11888365349764407,
"grad_norm": 2.5562196927508425,
"learning_rate": 1.95919151903205e-05,
"loss": 1.2076,
"step": 164
},
{
"epoch": 0.11960855382384922,
"grad_norm": 3.2522325367194425,
"learning_rate": 1.9585244678513913e-05,
"loss": 1.258,
"step": 165
},
{
"epoch": 0.12033345415005436,
"grad_norm": 3.792145378660367,
"learning_rate": 1.9578521244258162e-05,
"loss": 1.1639,
"step": 166
},
{
"epoch": 0.12105835447625951,
"grad_norm": 3.4172955897048207,
"learning_rate": 1.9571744924674943e-05,
"loss": 1.1899,
"step": 167
},
{
"epoch": 0.12178325480246466,
"grad_norm": 2.884712209279333,
"learning_rate": 1.9564915757177955e-05,
"loss": 1.1658,
"step": 168
},
{
"epoch": 0.1225081551286698,
"grad_norm": 2.584866024173676,
"learning_rate": 1.955803377947268e-05,
"loss": 1.1995,
"step": 169
},
{
"epoch": 0.12323305545487495,
"grad_norm": 2.608204074681373,
"learning_rate": 1.955109902955617e-05,
"loss": 1.1891,
"step": 170
},
{
"epoch": 0.1239579557810801,
"grad_norm": 3.021997715506911,
"learning_rate": 1.954411154571687e-05,
"loss": 1.1485,
"step": 171
},
{
"epoch": 0.12468285610728524,
"grad_norm": 2.136848876157592,
"learning_rate": 1.9537071366534352e-05,
"loss": 1.0598,
"step": 172
},
{
"epoch": 0.1254077564334904,
"grad_norm": 2.4835581901419332,
"learning_rate": 1.9529978530879144e-05,
"loss": 1.1085,
"step": 173
},
{
"epoch": 0.12613265675969554,
"grad_norm": 2.3827179769931823,
"learning_rate": 1.9522833077912512e-05,
"loss": 1.1484,
"step": 174
},
{
"epoch": 0.12685755708590069,
"grad_norm": 2.256838168033284,
"learning_rate": 1.951563504708622e-05,
"loss": 1.1174,
"step": 175
},
{
"epoch": 0.12758245741210583,
"grad_norm": 2.677664336590985,
"learning_rate": 1.950838447814233e-05,
"loss": 1.1908,
"step": 176
},
{
"epoch": 0.12830735773831098,
"grad_norm": 2.851750128481314,
"learning_rate": 1.9501081411112987e-05,
"loss": 1.2887,
"step": 177
},
{
"epoch": 0.12903225806451613,
"grad_norm": 2.299218668392714,
"learning_rate": 1.9493725886320193e-05,
"loss": 1.0708,
"step": 178
},
{
"epoch": 0.12975715839072127,
"grad_norm": 2.654535712782758,
"learning_rate": 1.9486317944375563e-05,
"loss": 1.1428,
"step": 179
},
{
"epoch": 0.13048205871692642,
"grad_norm": 3.1562180739485184,
"learning_rate": 1.947885762618014e-05,
"loss": 1.1446,
"step": 180
},
{
"epoch": 0.13120695904313157,
"grad_norm": 2.4022359841083762,
"learning_rate": 1.9471344972924142e-05,
"loss": 1.1913,
"step": 181
},
{
"epoch": 0.1319318593693367,
"grad_norm": 1.9874430354522221,
"learning_rate": 1.9463780026086735e-05,
"loss": 1.1536,
"step": 182
},
{
"epoch": 0.13265675969554186,
"grad_norm": 2.6028657856098762,
"learning_rate": 1.945616282743582e-05,
"loss": 1.0459,
"step": 183
},
{
"epoch": 0.133381660021747,
"grad_norm": 2.4961285769674584,
"learning_rate": 1.9448493419027795e-05,
"loss": 1.2184,
"step": 184
},
{
"epoch": 0.13410656034795215,
"grad_norm": 2.4545736340279607,
"learning_rate": 1.9440771843207305e-05,
"loss": 1.0966,
"step": 185
},
{
"epoch": 0.1348314606741573,
"grad_norm": 2.4064312104977796,
"learning_rate": 1.9432998142607036e-05,
"loss": 1.1427,
"step": 186
},
{
"epoch": 0.13555636100036245,
"grad_norm": 2.10260997961401,
"learning_rate": 1.9425172360147467e-05,
"loss": 1.153,
"step": 187
},
{
"epoch": 0.1362812613265676,
"grad_norm": 2.2184620974349,
"learning_rate": 1.9417294539036634e-05,
"loss": 1.1356,
"step": 188
},
{
"epoch": 0.13700616165277274,
"grad_norm": 2.715658298293896,
"learning_rate": 1.9409364722769882e-05,
"loss": 1.2261,
"step": 189
},
{
"epoch": 0.1377310619789779,
"grad_norm": 2.240649838269821,
"learning_rate": 1.9401382955129646e-05,
"loss": 1.0599,
"step": 190
},
{
"epoch": 0.13845596230518303,
"grad_norm": 3.0178674865097728,
"learning_rate": 1.9393349280185187e-05,
"loss": 1.2233,
"step": 191
},
{
"epoch": 0.13918086263138818,
"grad_norm": 2.132340826672663,
"learning_rate": 1.9385263742292368e-05,
"loss": 1.0894,
"step": 192
},
{
"epoch": 0.13990576295759333,
"grad_norm": 2.4647699353823733,
"learning_rate": 1.937712638609339e-05,
"loss": 1.2108,
"step": 193
},
{
"epoch": 0.14063066328379847,
"grad_norm": 2.4490272195884852,
"learning_rate": 1.9368937256516567e-05,
"loss": 1.1608,
"step": 194
},
{
"epoch": 0.14135556361000362,
"grad_norm": 2.1989447827390665,
"learning_rate": 1.9360696398776056e-05,
"loss": 1.1457,
"step": 195
},
{
"epoch": 0.14208046393620877,
"grad_norm": 2.068383189183824,
"learning_rate": 1.9352403858371618e-05,
"loss": 1.1081,
"step": 196
},
{
"epoch": 0.14280536426241391,
"grad_norm": 2.0311982289386474,
"learning_rate": 1.9344059681088372e-05,
"loss": 1.1035,
"step": 197
},
{
"epoch": 0.14353026458861906,
"grad_norm": 2.486963653282551,
"learning_rate": 1.933566391299654e-05,
"loss": 1.0822,
"step": 198
},
{
"epoch": 0.1442551649148242,
"grad_norm": 2.3367902123479314,
"learning_rate": 1.9327216600451177e-05,
"loss": 1.017,
"step": 199
},
{
"epoch": 0.14498006524102935,
"grad_norm": 2.390926126929064,
"learning_rate": 1.931871779009194e-05,
"loss": 1.0842,
"step": 200
},
{
"epoch": 0.1457049655672345,
"grad_norm": 3.3906627893855936,
"learning_rate": 1.9310167528842808e-05,
"loss": 1.2239,
"step": 201
},
{
"epoch": 0.14642986589343965,
"grad_norm": 2.36160531790082,
"learning_rate": 1.9301565863911836e-05,
"loss": 1.1343,
"step": 202
},
{
"epoch": 0.1471547662196448,
"grad_norm": 2.4172767818260885,
"learning_rate": 1.9292912842790893e-05,
"loss": 1.1009,
"step": 203
},
{
"epoch": 0.14787966654584994,
"grad_norm": 2.391111956498114,
"learning_rate": 1.92842085132554e-05,
"loss": 1.1245,
"step": 204
},
{
"epoch": 0.1486045668720551,
"grad_norm": 2.9512769517526527,
"learning_rate": 1.927545292336406e-05,
"loss": 1.1341,
"step": 205
},
{
"epoch": 0.14932946719826024,
"grad_norm": 2.247325705969668,
"learning_rate": 1.9266646121458597e-05,
"loss": 1.1099,
"step": 206
},
{
"epoch": 0.15005436752446538,
"grad_norm": 4.45573844838241,
"learning_rate": 1.9257788156163484e-05,
"loss": 1.2177,
"step": 207
},
{
"epoch": 0.15077926785067053,
"grad_norm": 2.308206270902384,
"learning_rate": 1.9248879076385696e-05,
"loss": 1.1115,
"step": 208
},
{
"epoch": 0.15150416817687568,
"grad_norm": 2.655262555875558,
"learning_rate": 1.9239918931314404e-05,
"loss": 1.1081,
"step": 209
},
{
"epoch": 0.15222906850308082,
"grad_norm": 2.939655738673358,
"learning_rate": 1.9230907770420737e-05,
"loss": 1.1192,
"step": 210
},
{
"epoch": 0.15295396882928597,
"grad_norm": 2.7252753733179564,
"learning_rate": 1.9221845643457485e-05,
"loss": 1.1471,
"step": 211
},
{
"epoch": 0.15367886915549112,
"grad_norm": 2.7708259653245677,
"learning_rate": 1.9212732600458845e-05,
"loss": 1.0776,
"step": 212
},
{
"epoch": 0.15440376948169626,
"grad_norm": 1.9749460870383841,
"learning_rate": 1.9203568691740115e-05,
"loss": 1.0709,
"step": 213
},
{
"epoch": 0.1551286698079014,
"grad_norm": 2.0889553423500353,
"learning_rate": 1.9194353967897453e-05,
"loss": 1.1986,
"step": 214
},
{
"epoch": 0.15585357013410656,
"grad_norm": 2.5179546631862624,
"learning_rate": 1.9185088479807577e-05,
"loss": 1.2635,
"step": 215
},
{
"epoch": 0.1565784704603117,
"grad_norm": 2.7700789430710424,
"learning_rate": 1.9175772278627477e-05,
"loss": 1.2196,
"step": 216
},
{
"epoch": 0.15730337078651685,
"grad_norm": 2.3103773826109313,
"learning_rate": 1.916640541579415e-05,
"loss": 1.1924,
"step": 217
},
{
"epoch": 0.158028271112722,
"grad_norm": 3.7179435673997587,
"learning_rate": 1.9156987943024305e-05,
"loss": 1.0714,
"step": 218
},
{
"epoch": 0.15875317143892714,
"grad_norm": 3.047980163715747,
"learning_rate": 1.9147519912314086e-05,
"loss": 1.0348,
"step": 219
},
{
"epoch": 0.1594780717651323,
"grad_norm": 2.3355941636809323,
"learning_rate": 1.9138001375938773e-05,
"loss": 1.0714,
"step": 220
},
{
"epoch": 0.16020297209133744,
"grad_norm": 2.4367237951314764,
"learning_rate": 1.9128432386452503e-05,
"loss": 1.0923,
"step": 221
},
{
"epoch": 0.16092787241754258,
"grad_norm": 2.534850993992721,
"learning_rate": 1.9118812996687975e-05,
"loss": 1.2025,
"step": 222
},
{
"epoch": 0.16165277274374773,
"grad_norm": 2.6850585710213637,
"learning_rate": 1.910914325975616e-05,
"loss": 1.2315,
"step": 223
},
{
"epoch": 0.16237767306995288,
"grad_norm": 2.6390954960986,
"learning_rate": 1.9099423229046015e-05,
"loss": 1.171,
"step": 224
},
{
"epoch": 0.16310257339615802,
"grad_norm": 3.674303992733298,
"learning_rate": 1.908965295822417e-05,
"loss": 1.22,
"step": 225
},
{
"epoch": 0.16382747372236317,
"grad_norm": 2.3198272055347413,
"learning_rate": 1.907983250123465e-05,
"loss": 0.9999,
"step": 226
},
{
"epoch": 0.16455237404856832,
"grad_norm": 2.367941394706144,
"learning_rate": 1.906996191229857e-05,
"loss": 1.1324,
"step": 227
},
{
"epoch": 0.16527727437477346,
"grad_norm": 2.2348457223954536,
"learning_rate": 1.906004124591383e-05,
"loss": 1.1695,
"step": 228
},
{
"epoch": 0.1660021747009786,
"grad_norm": 2.5357413278993373,
"learning_rate": 1.905007055685483e-05,
"loss": 1.2143,
"step": 229
},
{
"epoch": 0.16672707502718376,
"grad_norm": 2.197512694165324,
"learning_rate": 1.904004990017214e-05,
"loss": 1.059,
"step": 230
},
{
"epoch": 0.1674519753533889,
"grad_norm": 2.246714216934824,
"learning_rate": 1.902997933119223e-05,
"loss": 1.0306,
"step": 231
},
{
"epoch": 0.16817687567959405,
"grad_norm": 2.3508232503422257,
"learning_rate": 1.9019858905517146e-05,
"loss": 1.232,
"step": 232
},
{
"epoch": 0.1689017760057992,
"grad_norm": 3.2397999719968706,
"learning_rate": 1.900968867902419e-05,
"loss": 1.1642,
"step": 233
},
{
"epoch": 0.16962667633200434,
"grad_norm": 2.2082937384067134,
"learning_rate": 1.899946870786565e-05,
"loss": 1.0615,
"step": 234
},
{
"epoch": 0.1703515766582095,
"grad_norm": 2.6489946010926437,
"learning_rate": 1.8989199048468443e-05,
"loss": 1.2625,
"step": 235
},
{
"epoch": 0.17107647698441464,
"grad_norm": 2.243711710470245,
"learning_rate": 1.8978879757533838e-05,
"loss": 1.1316,
"step": 236
},
{
"epoch": 0.17180137731061979,
"grad_norm": 2.846092716410688,
"learning_rate": 1.8968510892037136e-05,
"loss": 1.2307,
"step": 237
},
{
"epoch": 0.17252627763682493,
"grad_norm": 2.7609914990765128,
"learning_rate": 1.8958092509227347e-05,
"loss": 1.1384,
"step": 238
},
{
"epoch": 0.17325117796303008,
"grad_norm": 1.8868238757930658,
"learning_rate": 1.8947624666626866e-05,
"loss": 1.0692,
"step": 239
},
{
"epoch": 0.17397607828923523,
"grad_norm": 2.5980341998176146,
"learning_rate": 1.893710742203119e-05,
"loss": 1.1284,
"step": 240
},
{
"epoch": 0.17470097861544037,
"grad_norm": 2.7513723878066982,
"learning_rate": 1.8926540833508557e-05,
"loss": 1.0836,
"step": 241
},
{
"epoch": 0.17542587894164552,
"grad_norm": 2.6788196765201535,
"learning_rate": 1.8915924959399656e-05,
"loss": 1.1034,
"step": 242
},
{
"epoch": 0.17615077926785067,
"grad_norm": 2.5118625400183023,
"learning_rate": 1.8905259858317287e-05,
"loss": 1.2339,
"step": 243
},
{
"epoch": 0.1768756795940558,
"grad_norm": 2.126239831693274,
"learning_rate": 1.889454558914605e-05,
"loss": 1.0637,
"step": 244
},
{
"epoch": 0.17760057992026096,
"grad_norm": 2.604602180586931,
"learning_rate": 1.888378221104201e-05,
"loss": 1.2855,
"step": 245
},
{
"epoch": 0.1783254802464661,
"grad_norm": 2.4753012194229562,
"learning_rate": 1.8872969783432376e-05,
"loss": 1.0602,
"step": 246
},
{
"epoch": 0.17905038057267125,
"grad_norm": 2.6860064054799584,
"learning_rate": 1.886210836601517e-05,
"loss": 1.1418,
"step": 247
},
{
"epoch": 0.1797752808988764,
"grad_norm": 2.5126660296525833,
"learning_rate": 1.8851198018758898e-05,
"loss": 1.1937,
"step": 248
},
{
"epoch": 0.18050018122508155,
"grad_norm": 1.7715675237469672,
"learning_rate": 1.884023880190222e-05,
"loss": 1.0683,
"step": 249
},
{
"epoch": 0.1812250815512867,
"grad_norm": 2.3549127021241048,
"learning_rate": 1.8829230775953616e-05,
"loss": 1.1331,
"step": 250
},
{
"epoch": 0.18194998187749184,
"grad_norm": 2.043319865990913,
"learning_rate": 1.8818174001691055e-05,
"loss": 1.009,
"step": 251
},
{
"epoch": 0.182674882203697,
"grad_norm": 2.5088376984194682,
"learning_rate": 1.880706854016166e-05,
"loss": 1.0838,
"step": 252
},
{
"epoch": 0.18339978252990213,
"grad_norm": 2.095009396399693,
"learning_rate": 1.8795914452681362e-05,
"loss": 1.0487,
"step": 253
},
{
"epoch": 0.18412468285610728,
"grad_norm": 2.3982978050324073,
"learning_rate": 1.8784711800834564e-05,
"loss": 1.0634,
"step": 254
},
{
"epoch": 0.18484958318231243,
"grad_norm": 2.1820663845429524,
"learning_rate": 1.877346064647382e-05,
"loss": 1.0621,
"step": 255
},
{
"epoch": 0.18557448350851757,
"grad_norm": 2.9905602073691147,
"learning_rate": 1.8762161051719464e-05,
"loss": 1.2322,
"step": 256
},
{
"epoch": 0.18629938383472272,
"grad_norm": 2.2435538559731505,
"learning_rate": 1.8750813078959282e-05,
"loss": 1.0788,
"step": 257
},
{
"epoch": 0.18702428416092787,
"grad_norm": 2.2367444808107906,
"learning_rate": 1.8739416790848177e-05,
"loss": 1.0598,
"step": 258
},
{
"epoch": 0.18774918448713301,
"grad_norm": 2.472646715254004,
"learning_rate": 1.87279722503078e-05,
"loss": 1.0482,
"step": 259
},
{
"epoch": 0.18847408481333816,
"grad_norm": 2.111989985490911,
"learning_rate": 1.8716479520526227e-05,
"loss": 1.1027,
"step": 260
},
{
"epoch": 0.1891989851395433,
"grad_norm": 2.510087294479276,
"learning_rate": 1.870493866495759e-05,
"loss": 1.1856,
"step": 261
},
{
"epoch": 0.18992388546574845,
"grad_norm": 2.5133146185218904,
"learning_rate": 1.8693349747321737e-05,
"loss": 1.1478,
"step": 262
},
{
"epoch": 0.1906487857919536,
"grad_norm": 2.3395156482299972,
"learning_rate": 1.868171283160388e-05,
"loss": 1.1815,
"step": 263
},
{
"epoch": 0.19137368611815875,
"grad_norm": 2.3764168972959707,
"learning_rate": 1.867002798205424e-05,
"loss": 1.2432,
"step": 264
},
{
"epoch": 0.1920985864443639,
"grad_norm": 2.559923113432804,
"learning_rate": 1.8658295263187692e-05,
"loss": 1.1154,
"step": 265
},
{
"epoch": 0.19282348677056904,
"grad_norm": 2.8783556981475544,
"learning_rate": 1.8646514739783404e-05,
"loss": 1.1236,
"step": 266
},
{
"epoch": 0.1935483870967742,
"grad_norm": 2.6502413673324985,
"learning_rate": 1.8634686476884497e-05,
"loss": 1.2676,
"step": 267
},
{
"epoch": 0.19427328742297933,
"grad_norm": 3.3520733595809857,
"learning_rate": 1.8622810539797656e-05,
"loss": 1.0714,
"step": 268
},
{
"epoch": 0.19499818774918448,
"grad_norm": 2.9157405583997353,
"learning_rate": 1.8610886994092802e-05,
"loss": 1.1882,
"step": 269
},
{
"epoch": 0.19572308807538963,
"grad_norm": 2.118069033541846,
"learning_rate": 1.8598915905602704e-05,
"loss": 1.1082,
"step": 270
},
{
"epoch": 0.19644798840159478,
"grad_norm": 2.3662623126360693,
"learning_rate": 1.858689734042263e-05,
"loss": 1.0517,
"step": 271
},
{
"epoch": 0.19717288872779992,
"grad_norm": 2.3124998904314658,
"learning_rate": 1.8574831364909988e-05,
"loss": 1.1954,
"step": 272
},
{
"epoch": 0.19789778905400507,
"grad_norm": 2.2466176708447576,
"learning_rate": 1.8562718045683933e-05,
"loss": 1.153,
"step": 273
},
{
"epoch": 0.19862268938021022,
"grad_norm": 3.186725132878171,
"learning_rate": 1.855055744962502e-05,
"loss": 1.0749,
"step": 274
},
{
"epoch": 0.19934758970641536,
"grad_norm": 1.9797977750729439,
"learning_rate": 1.8538349643874845e-05,
"loss": 1.1717,
"step": 275
},
{
"epoch": 0.2000724900326205,
"grad_norm": 2.650096534254559,
"learning_rate": 1.8526094695835647e-05,
"loss": 1.0998,
"step": 276
},
{
"epoch": 0.20079739035882566,
"grad_norm": 2.242222036665605,
"learning_rate": 1.851379267316995e-05,
"loss": 1.12,
"step": 277
},
{
"epoch": 0.2015222906850308,
"grad_norm": 3.0768668231596763,
"learning_rate": 1.8501443643800185e-05,
"loss": 1.2119,
"step": 278
},
{
"epoch": 0.20224719101123595,
"grad_norm": 2.4142131380970655,
"learning_rate": 1.8489047675908328e-05,
"loss": 1.1211,
"step": 279
},
{
"epoch": 0.2029720913374411,
"grad_norm": 2.494879063057067,
"learning_rate": 1.8476604837935515e-05,
"loss": 1.1135,
"step": 280
},
{
"epoch": 0.20369699166364624,
"grad_norm": 2.105871660944368,
"learning_rate": 1.846411519858165e-05,
"loss": 1.1952,
"step": 281
},
{
"epoch": 0.2044218919898514,
"grad_norm": 2.2793551122599847,
"learning_rate": 1.8451578826805046e-05,
"loss": 1.148,
"step": 282
},
{
"epoch": 0.20514679231605654,
"grad_norm": 2.3515542983789794,
"learning_rate": 1.843899579182204e-05,
"loss": 1.078,
"step": 283
},
{
"epoch": 0.20587169264226168,
"grad_norm": 2.3215932574696043,
"learning_rate": 1.8426366163106603e-05,
"loss": 0.9794,
"step": 284
},
{
"epoch": 0.20659659296846683,
"grad_norm": 2.9004041098455122,
"learning_rate": 1.841369001038997e-05,
"loss": 1.1092,
"step": 285
},
{
"epoch": 0.20732149329467198,
"grad_norm": 2.503251753777213,
"learning_rate": 1.8400967403660228e-05,
"loss": 1.1586,
"step": 286
},
{
"epoch": 0.20804639362087712,
"grad_norm": 2.6266283769571546,
"learning_rate": 1.8388198413161962e-05,
"loss": 1.1247,
"step": 287
},
{
"epoch": 0.20877129394708227,
"grad_norm": 2.6053044014908644,
"learning_rate": 1.837538310939586e-05,
"loss": 1.0471,
"step": 288
},
{
"epoch": 0.20949619427328742,
"grad_norm": 2.2521209365507766,
"learning_rate": 1.8362521563118298e-05,
"loss": 1.0282,
"step": 289
},
{
"epoch": 0.21022109459949256,
"grad_norm": 2.086975827241926,
"learning_rate": 1.834961384534098e-05,
"loss": 1.1105,
"step": 290
},
{
"epoch": 0.2109459949256977,
"grad_norm": 2.282332985227685,
"learning_rate": 1.8336660027330525e-05,
"loss": 1.1147,
"step": 291
},
{
"epoch": 0.21167089525190286,
"grad_norm": 2.4541255599254326,
"learning_rate": 1.83236601806081e-05,
"loss": 1.0968,
"step": 292
},
{
"epoch": 0.212395795578108,
"grad_norm": 2.5576128314214763,
"learning_rate": 1.8310614376948986e-05,
"loss": 1.2477,
"step": 293
},
{
"epoch": 0.21312069590431315,
"grad_norm": 2.073918446338343,
"learning_rate": 1.829752268838222e-05,
"loss": 1.0705,
"step": 294
},
{
"epoch": 0.2138455962305183,
"grad_norm": 3.0383016771932168,
"learning_rate": 1.8284385187190168e-05,
"loss": 1.2738,
"step": 295
},
{
"epoch": 0.21457049655672344,
"grad_norm": 2.5327376795510714,
"learning_rate": 1.8271201945908152e-05,
"loss": 1.0948,
"step": 296
},
{
"epoch": 0.2152953968829286,
"grad_norm": 2.913601708623835,
"learning_rate": 1.825797303732402e-05,
"loss": 1.1031,
"step": 297
},
{
"epoch": 0.21602029720913374,
"grad_norm": 2.7092322107658933,
"learning_rate": 1.8244698534477776e-05,
"loss": 1.0646,
"step": 298
},
{
"epoch": 0.21674519753533888,
"grad_norm": 2.2272013984412418,
"learning_rate": 1.823137851066115e-05,
"loss": 0.9988,
"step": 299
},
{
"epoch": 0.21747009786154403,
"grad_norm": 1.897477153962133,
"learning_rate": 1.821801303941721e-05,
"loss": 1.086,
"step": 300
},
{
"epoch": 0.21819499818774918,
"grad_norm": 2.055450943656558,
"learning_rate": 1.8204602194539948e-05,
"loss": 1.1499,
"step": 301
},
{
"epoch": 0.21891989851395433,
"grad_norm": 2.4909766282023016,
"learning_rate": 1.819114605007388e-05,
"loss": 1.1902,
"step": 302
},
{
"epoch": 0.21964479884015947,
"grad_norm": 2.155603573794576,
"learning_rate": 1.8177644680313618e-05,
"loss": 1.1291,
"step": 303
},
{
"epoch": 0.22036969916636462,
"grad_norm": 2.3903217112059845,
"learning_rate": 1.8164098159803485e-05,
"loss": 1.074,
"step": 304
},
{
"epoch": 0.22109459949256977,
"grad_norm": 2.4252393061914566,
"learning_rate": 1.8150506563337095e-05,
"loss": 1.1277,
"step": 305
},
{
"epoch": 0.2218194998187749,
"grad_norm": 2.390835505304097,
"learning_rate": 1.813686996595693e-05,
"loss": 1.0942,
"step": 306
},
{
"epoch": 0.22254440014498006,
"grad_norm": 2.3599664718135087,
"learning_rate": 1.8123188442953935e-05,
"loss": 1.1321,
"step": 307
},
{
"epoch": 0.2232693004711852,
"grad_norm": 3.4063853490696356,
"learning_rate": 1.81094620698671e-05,
"loss": 1.2013,
"step": 308
},
{
"epoch": 0.22399420079739035,
"grad_norm": 2.8215352516328682,
"learning_rate": 1.809569092248304e-05,
"loss": 1.1624,
"step": 309
},
{
"epoch": 0.2247191011235955,
"grad_norm": 2.692738685235975,
"learning_rate": 1.8081875076835587e-05,
"loss": 1.2877,
"step": 310
},
{
"epoch": 0.22544400144980065,
"grad_norm": 1.9994403713749573,
"learning_rate": 1.8068014609205363e-05,
"loss": 1.1385,
"step": 311
},
{
"epoch": 0.2261689017760058,
"grad_norm": 2.6062186167459864,
"learning_rate": 1.805410959611935e-05,
"loss": 1.17,
"step": 312
},
{
"epoch": 0.22689380210221094,
"grad_norm": 2.471082211880037,
"learning_rate": 1.804016011435048e-05,
"loss": 1.1468,
"step": 313
},
{
"epoch": 0.2276187024284161,
"grad_norm": 2.685017007295627,
"learning_rate": 1.802616624091721e-05,
"loss": 1.1734,
"step": 314
},
{
"epoch": 0.22834360275462123,
"grad_norm": 2.462530495353453,
"learning_rate": 1.8012128053083097e-05,
"loss": 1.1009,
"step": 315
},
{
"epoch": 0.22906850308082638,
"grad_norm": 2.2431993395283767,
"learning_rate": 1.7998045628356355e-05,
"loss": 1.1612,
"step": 316
},
{
"epoch": 0.22979340340703153,
"grad_norm": 2.3899511245334133,
"learning_rate": 1.7983919044489455e-05,
"loss": 1.1758,
"step": 317
},
{
"epoch": 0.23051830373323667,
"grad_norm": 2.259336195556954,
"learning_rate": 1.7969748379478675e-05,
"loss": 1.1089,
"step": 318
},
{
"epoch": 0.23124320405944182,
"grad_norm": 3.0578949924857013,
"learning_rate": 1.795553371156368e-05,
"loss": 1.2194,
"step": 319
},
{
"epoch": 0.23196810438564697,
"grad_norm": 2.4251956933972867,
"learning_rate": 1.7941275119227076e-05,
"loss": 1.1154,
"step": 320
},
{
"epoch": 0.2326930047118521,
"grad_norm": 2.3987899466375344,
"learning_rate": 1.7926972681193996e-05,
"loss": 1.1155,
"step": 321
},
{
"epoch": 0.23341790503805726,
"grad_norm": 2.0179495834873395,
"learning_rate": 1.7912626476431648e-05,
"loss": 1.1214,
"step": 322
},
{
"epoch": 0.2341428053642624,
"grad_norm": 2.2577973639628492,
"learning_rate": 1.78982365841489e-05,
"loss": 1.145,
"step": 323
},
{
"epoch": 0.23486770569046755,
"grad_norm": 2.5258056683443857,
"learning_rate": 1.788380308379581e-05,
"loss": 1.0736,
"step": 324
},
{
"epoch": 0.2355926060166727,
"grad_norm": 2.597512148415182,
"learning_rate": 1.7869326055063223e-05,
"loss": 1.0742,
"step": 325
},
{
"epoch": 0.23631750634287785,
"grad_norm": 2.4111146476669063,
"learning_rate": 1.7854805577882307e-05,
"loss": 1.1139,
"step": 326
},
{
"epoch": 0.237042406669083,
"grad_norm": 2.513312885502925,
"learning_rate": 1.784024173242412e-05,
"loss": 1.1792,
"step": 327
},
{
"epoch": 0.23776730699528814,
"grad_norm": 2.6103960099654446,
"learning_rate": 1.7825634599099167e-05,
"loss": 1.0752,
"step": 328
},
{
"epoch": 0.2384922073214933,
"grad_norm": 2.5153496087422136,
"learning_rate": 1.7810984258556955e-05,
"loss": 1.1024,
"step": 329
},
{
"epoch": 0.23921710764769843,
"grad_norm": 1.891594673564252,
"learning_rate": 1.779629079168556e-05,
"loss": 0.9794,
"step": 330
},
{
"epoch": 0.23994200797390358,
"grad_norm": 2.103561148427696,
"learning_rate": 1.7781554279611143e-05,
"loss": 0.9872,
"step": 331
},
{
"epoch": 0.24066690830010873,
"grad_norm": 2.258815019112651,
"learning_rate": 1.7766774803697555e-05,
"loss": 1.1095,
"step": 332
},
{
"epoch": 0.24139180862631388,
"grad_norm": 2.0559334388087938,
"learning_rate": 1.7751952445545846e-05,
"loss": 1.1,
"step": 333
},
{
"epoch": 0.24211670895251902,
"grad_norm": 2.382861239464527,
"learning_rate": 1.7737087286993832e-05,
"loss": 1.118,
"step": 334
},
{
"epoch": 0.24284160927872417,
"grad_norm": 2.0365687571462536,
"learning_rate": 1.7722179410115644e-05,
"loss": 1.1233,
"step": 335
},
{
"epoch": 0.24356650960492932,
"grad_norm": 2.5340895730401036,
"learning_rate": 1.770722889722126e-05,
"loss": 1.2206,
"step": 336
},
{
"epoch": 0.24429140993113446,
"grad_norm": 2.2828107095058865,
"learning_rate": 1.769223583085608e-05,
"loss": 1.144,
"step": 337
},
{
"epoch": 0.2450163102573396,
"grad_norm": 2.7476805734384495,
"learning_rate": 1.767720029380044e-05,
"loss": 1.0737,
"step": 338
},
{
"epoch": 0.24574121058354476,
"grad_norm": 2.1237162836325068,
"learning_rate": 1.7662122369069164e-05,
"loss": 1.0684,
"step": 339
},
{
"epoch": 0.2464661109097499,
"grad_norm": 2.5061834120195488,
"learning_rate": 1.764700213991111e-05,
"loss": 1.1518,
"step": 340
},
{
"epoch": 0.24719101123595505,
"grad_norm": 2.3826699101861353,
"learning_rate": 1.7631839689808724e-05,
"loss": 1.0676,
"step": 341
},
{
"epoch": 0.2479159115621602,
"grad_norm": 2.8748822951105235,
"learning_rate": 1.7616635102477545e-05,
"loss": 1.1453,
"step": 342
},
{
"epoch": 0.24864081188836534,
"grad_norm": 2.2078908610481434,
"learning_rate": 1.760138846186577e-05,
"loss": 0.9728,
"step": 343
},
{
"epoch": 0.2493657122145705,
"grad_norm": 2.750380120949678,
"learning_rate": 1.7586099852153778e-05,
"loss": 1.194,
"step": 344
},
{
"epoch": 0.25009061254077564,
"grad_norm": 1.9828736953069481,
"learning_rate": 1.7570769357753682e-05,
"loss": 1.122,
"step": 345
},
{
"epoch": 0.25009061254077564,
"eval_loss": 1.9538915157318115,
"eval_runtime": 1129.0599,
"eval_samples_per_second": 15.636,
"eval_steps_per_second": 0.489,
"step": 345
},
{
"epoch": 0.2508155128669808,
"grad_norm": 2.4573515309403184,
"learning_rate": 1.755539706330883e-05,
"loss": 1.1254,
"step": 346
},
{
"epoch": 0.25154041319318593,
"grad_norm": 2.382471730290679,
"learning_rate": 1.753998305369338e-05,
"loss": 1.0609,
"step": 347
},
{
"epoch": 0.2522653135193911,
"grad_norm": 2.0670497345618624,
"learning_rate": 1.752452741401179e-05,
"loss": 1.11,
"step": 348
},
{
"epoch": 0.2529902138455962,
"grad_norm": 2.826479622570429,
"learning_rate": 1.7509030229598387e-05,
"loss": 1.0688,
"step": 349
},
{
"epoch": 0.25371511417180137,
"grad_norm": 2.4410246210380797,
"learning_rate": 1.749349158601686e-05,
"loss": 1.0848,
"step": 350
},
{
"epoch": 0.2544400144980065,
"grad_norm": 2.1500648572615226,
"learning_rate": 1.7477911569059808e-05,
"loss": 1.1225,
"step": 351
},
{
"epoch": 0.25516491482421166,
"grad_norm": 2.303703660619773,
"learning_rate": 1.746229026474827e-05,
"loss": 1.0781,
"step": 352
},
{
"epoch": 0.2558898151504168,
"grad_norm": 2.6829872427143853,
"learning_rate": 1.744662775933123e-05,
"loss": 1.0742,
"step": 353
},
{
"epoch": 0.25661471547662196,
"grad_norm": 2.2217416215597448,
"learning_rate": 1.743092413928517e-05,
"loss": 1.142,
"step": 354
},
{
"epoch": 0.2573396158028271,
"grad_norm": 2.6300226414802594,
"learning_rate": 1.741517949131356e-05,
"loss": 1.0869,
"step": 355
},
{
"epoch": 0.25806451612903225,
"grad_norm": 2.094472680631636,
"learning_rate": 1.7399393902346403e-05,
"loss": 1.0797,
"step": 356
},
{
"epoch": 0.2587894164552374,
"grad_norm": 2.6696223277650097,
"learning_rate": 1.738356745953975e-05,
"loss": 1.1725,
"step": 357
},
{
"epoch": 0.25951431678144254,
"grad_norm": 2.560913641389733,
"learning_rate": 1.7367700250275196e-05,
"loss": 1.0524,
"step": 358
},
{
"epoch": 0.2602392171076477,
"grad_norm": 2.499190502628081,
"learning_rate": 1.735179236215945e-05,
"loss": 1.0864,
"step": 359
},
{
"epoch": 0.26096411743385284,
"grad_norm": 2.1752493153518127,
"learning_rate": 1.7335843883023792e-05,
"loss": 1.1682,
"step": 360
},
{
"epoch": 0.261689017760058,
"grad_norm": 2.193115771884426,
"learning_rate": 1.7319854900923627e-05,
"loss": 1.1538,
"step": 361
},
{
"epoch": 0.26241391808626313,
"grad_norm": 3.6037027339305228,
"learning_rate": 1.7303825504137985e-05,
"loss": 1.2803,
"step": 362
},
{
"epoch": 0.2631388184124683,
"grad_norm": 2.151895072296486,
"learning_rate": 1.728775578116903e-05,
"loss": 1.027,
"step": 363
},
{
"epoch": 0.2638637187386734,
"grad_norm": 2.0474799790136315,
"learning_rate": 1.7271645820741586e-05,
"loss": 1.0166,
"step": 364
},
{
"epoch": 0.26458861906487857,
"grad_norm": 1.9599613314591962,
"learning_rate": 1.7255495711802627e-05,
"loss": 1.022,
"step": 365
},
{
"epoch": 0.2653135193910837,
"grad_norm": 2.401594401551547,
"learning_rate": 1.723930554352081e-05,
"loss": 1.1917,
"step": 366
},
{
"epoch": 0.26603841971728887,
"grad_norm": 2.642087840213965,
"learning_rate": 1.7223075405285956e-05,
"loss": 1.2054,
"step": 367
},
{
"epoch": 0.266763320043494,
"grad_norm": 2.18604834719014,
"learning_rate": 1.7206805386708572e-05,
"loss": 1.1475,
"step": 368
},
{
"epoch": 0.26748822036969916,
"grad_norm": 2.0338413719031507,
"learning_rate": 1.7190495577619364e-05,
"loss": 1.1028,
"step": 369
},
{
"epoch": 0.2682131206959043,
"grad_norm": 2.1711132034769394,
"learning_rate": 1.717414606806872e-05,
"loss": 1.1044,
"step": 370
},
{
"epoch": 0.26893802102210945,
"grad_norm": 1.930408817282508,
"learning_rate": 1.715775694832623e-05,
"loss": 1.1185,
"step": 371
},
{
"epoch": 0.2696629213483146,
"grad_norm": 2.1748455201638643,
"learning_rate": 1.7141328308880178e-05,
"loss": 1.0437,
"step": 372
},
{
"epoch": 0.27038782167451975,
"grad_norm": 2.487178508999438,
"learning_rate": 1.712486024043705e-05,
"loss": 1.1261,
"step": 373
},
{
"epoch": 0.2711127220007249,
"grad_norm": 2.6620415854141077,
"learning_rate": 1.7108352833921022e-05,
"loss": 1.1125,
"step": 374
},
{
"epoch": 0.27183762232693004,
"grad_norm": 2.4779294920405333,
"learning_rate": 1.7091806180473472e-05,
"loss": 1.1376,
"step": 375
},
{
"epoch": 0.2725625226531352,
"grad_norm": 2.1914137672000433,
"learning_rate": 1.7075220371452465e-05,
"loss": 1.2255,
"step": 376
},
{
"epoch": 0.27328742297934033,
"grad_norm": 2.2406112575154404,
"learning_rate": 1.705859549843226e-05,
"loss": 1.1511,
"step": 377
},
{
"epoch": 0.2740123233055455,
"grad_norm": 2.0333226545718914,
"learning_rate": 1.7041931653202788e-05,
"loss": 1.1184,
"step": 378
},
{
"epoch": 0.2747372236317506,
"grad_norm": 2.8707122599293875,
"learning_rate": 1.702522892776916e-05,
"loss": 1.0367,
"step": 379
},
{
"epoch": 0.2754621239579558,
"grad_norm": 2.242723836798557,
"learning_rate": 1.7008487414351163e-05,
"loss": 1.1647,
"step": 380
},
{
"epoch": 0.2761870242841609,
"grad_norm": 3.052876110953187,
"learning_rate": 1.6991707205382723e-05,
"loss": 1.1574,
"step": 381
},
{
"epoch": 0.27691192461036607,
"grad_norm": 2.3302064287160955,
"learning_rate": 1.6974888393511425e-05,
"loss": 1.0833,
"step": 382
},
{
"epoch": 0.2776368249365712,
"grad_norm": 2.362447133340326,
"learning_rate": 1.695803107159799e-05,
"loss": 1.0142,
"step": 383
},
{
"epoch": 0.27836172526277636,
"grad_norm": 1.9262635584689904,
"learning_rate": 1.694113533271576e-05,
"loss": 0.9974,
"step": 384
},
{
"epoch": 0.2790866255889815,
"grad_norm": 2.6725693558247943,
"learning_rate": 1.6924201270150194e-05,
"loss": 1.1584,
"step": 385
},
{
"epoch": 0.27981152591518665,
"grad_norm": 2.3758190016297855,
"learning_rate": 1.6907228977398324e-05,
"loss": 1.1007,
"step": 386
},
{
"epoch": 0.2805364262413918,
"grad_norm": 2.02794227684942,
"learning_rate": 1.689021854816829e-05,
"loss": 0.9606,
"step": 387
},
{
"epoch": 0.28126132656759695,
"grad_norm": 2.182217806815362,
"learning_rate": 1.6873170076378764e-05,
"loss": 1.1012,
"step": 388
},
{
"epoch": 0.2819862268938021,
"grad_norm": 2.8274640981113714,
"learning_rate": 1.685608365615848e-05,
"loss": 1.1609,
"step": 389
},
{
"epoch": 0.28271112722000724,
"grad_norm": 2.0005395310858165,
"learning_rate": 1.683895938184569e-05,
"loss": 1.1083,
"step": 390
},
{
"epoch": 0.2834360275462124,
"grad_norm": 2.5469245481992195,
"learning_rate": 1.6821797347987637e-05,
"loss": 1.1813,
"step": 391
},
{
"epoch": 0.28416092787241753,
"grad_norm": 2.250275964552977,
"learning_rate": 1.680459764934006e-05,
"loss": 1.0717,
"step": 392
},
{
"epoch": 0.2848858281986227,
"grad_norm": 2.1541186966473314,
"learning_rate": 1.678736038086664e-05,
"loss": 1.0937,
"step": 393
},
{
"epoch": 0.28561072852482783,
"grad_norm": 2.2526083463595694,
"learning_rate": 1.67700856377385e-05,
"loss": 1.1692,
"step": 394
},
{
"epoch": 0.286335628851033,
"grad_norm": 2.360291496431212,
"learning_rate": 1.6752773515333665e-05,
"loss": 1.0956,
"step": 395
},
{
"epoch": 0.2870605291772381,
"grad_norm": 2.7484396524462693,
"learning_rate": 1.6735424109236538e-05,
"loss": 1.0348,
"step": 396
},
{
"epoch": 0.28778542950344327,
"grad_norm": 1.991944001354783,
"learning_rate": 1.671803751523738e-05,
"loss": 1.0371,
"step": 397
},
{
"epoch": 0.2885103298296484,
"grad_norm": 2.1121316874137794,
"learning_rate": 1.6700613829331778e-05,
"loss": 1.0483,
"step": 398
},
{
"epoch": 0.28923523015585356,
"grad_norm": 2.9850153218884405,
"learning_rate": 1.6683153147720098e-05,
"loss": 1.2002,
"step": 399
},
{
"epoch": 0.2899601304820587,
"grad_norm": 2.2268590912100987,
"learning_rate": 1.666565556680698e-05,
"loss": 1.1616,
"step": 400
},
{
"epoch": 0.29068503080826386,
"grad_norm": 2.135948670713283,
"learning_rate": 1.664812118320079e-05,
"loss": 0.9466,
"step": 401
},
{
"epoch": 0.291409931134469,
"grad_norm": 2.8208944014300603,
"learning_rate": 1.6630550093713096e-05,
"loss": 1.1814,
"step": 402
},
{
"epoch": 0.29213483146067415,
"grad_norm": 2.225373597938535,
"learning_rate": 1.661294239535812e-05,
"loss": 1.0444,
"step": 403
},
{
"epoch": 0.2928597317868793,
"grad_norm": 2.253251092388969,
"learning_rate": 1.6595298185352216e-05,
"loss": 1.092,
"step": 404
},
{
"epoch": 0.29358463211308444,
"grad_norm": 2.46062887701134,
"learning_rate": 1.6577617561113322e-05,
"loss": 1.0818,
"step": 405
},
{
"epoch": 0.2943095324392896,
"grad_norm": 3.129489147993457,
"learning_rate": 1.6559900620260435e-05,
"loss": 1.1676,
"step": 406
},
{
"epoch": 0.29503443276549474,
"grad_norm": 2.063039937217649,
"learning_rate": 1.654214746061306e-05,
"loss": 1.1648,
"step": 407
},
{
"epoch": 0.2957593330916999,
"grad_norm": 2.4465593098747256,
"learning_rate": 1.652435818019068e-05,
"loss": 1.0201,
"step": 408
},
{
"epoch": 0.29648423341790503,
"grad_norm": 2.399324464811924,
"learning_rate": 1.6506532877212195e-05,
"loss": 1.0413,
"step": 409
},
{
"epoch": 0.2972091337441102,
"grad_norm": 2.6482971890623546,
"learning_rate": 1.6488671650095416e-05,
"loss": 1.105,
"step": 410
},
{
"epoch": 0.2979340340703153,
"grad_norm": 3.034059815258785,
"learning_rate": 1.6470774597456482e-05,
"loss": 1.066,
"step": 411
},
{
"epoch": 0.29865893439652047,
"grad_norm": 2.0229310257745725,
"learning_rate": 1.6452841818109342e-05,
"loss": 1.0126,
"step": 412
},
{
"epoch": 0.2993838347227256,
"grad_norm": 2.09062623797439,
"learning_rate": 1.64348734110652e-05,
"loss": 1.0359,
"step": 413
},
{
"epoch": 0.30010873504893076,
"grad_norm": 2.261696086833709,
"learning_rate": 1.6416869475531964e-05,
"loss": 1.0387,
"step": 414
},
{
"epoch": 0.3008336353751359,
"grad_norm": 2.0194176082613686,
"learning_rate": 1.639883011091371e-05,
"loss": 1.0375,
"step": 415
},
{
"epoch": 0.30155853570134106,
"grad_norm": 3.0655992063503392,
"learning_rate": 1.6380755416810125e-05,
"loss": 1.1433,
"step": 416
},
{
"epoch": 0.3022834360275462,
"grad_norm": 2.200388293782514,
"learning_rate": 1.6362645493015955e-05,
"loss": 1.0384,
"step": 417
},
{
"epoch": 0.30300833635375135,
"grad_norm": 2.2340973013167646,
"learning_rate": 1.634450043952046e-05,
"loss": 1.1233,
"step": 418
},
{
"epoch": 0.3037332366799565,
"grad_norm": 2.4431864602307334,
"learning_rate": 1.6326320356506858e-05,
"loss": 1.1127,
"step": 419
},
{
"epoch": 0.30445813700616164,
"grad_norm": 2.5340829203747997,
"learning_rate": 1.6308105344351776e-05,
"loss": 1.1336,
"step": 420
},
{
"epoch": 0.3051830373323668,
"grad_norm": 2.4621749417333865,
"learning_rate": 1.6289855503624694e-05,
"loss": 1.18,
"step": 421
},
{
"epoch": 0.30590793765857194,
"grad_norm": 2.435171083757431,
"learning_rate": 1.6271570935087385e-05,
"loss": 1.1034,
"step": 422
},
{
"epoch": 0.3066328379847771,
"grad_norm": 2.56776282560331,
"learning_rate": 1.6253251739693366e-05,
"loss": 1.1227,
"step": 423
},
{
"epoch": 0.30735773831098223,
"grad_norm": 2.6606550385377172,
"learning_rate": 1.6234898018587336e-05,
"loss": 1.0606,
"step": 424
},
{
"epoch": 0.3080826386371874,
"grad_norm": 2.3905958945143584,
"learning_rate": 1.621650987310462e-05,
"loss": 1.0924,
"step": 425
},
{
"epoch": 0.3088075389633925,
"grad_norm": 2.16146889842216,
"learning_rate": 1.6198087404770602e-05,
"loss": 1.0924,
"step": 426
},
{
"epoch": 0.30953243928959767,
"grad_norm": 2.511397106934819,
"learning_rate": 1.617963071530018e-05,
"loss": 1.1366,
"step": 427
},
{
"epoch": 0.3102573396158028,
"grad_norm": 2.394538554344066,
"learning_rate": 1.616113990659719e-05,
"loss": 1.0009,
"step": 428
},
{
"epoch": 0.31098223994200797,
"grad_norm": 2.4259004458514006,
"learning_rate": 1.6142615080753846e-05,
"loss": 1.1315,
"step": 429
},
{
"epoch": 0.3117071402682131,
"grad_norm": 2.6037915854967726,
"learning_rate": 1.6124056340050186e-05,
"loss": 1.0235,
"step": 430
},
{
"epoch": 0.31243204059441826,
"grad_norm": 2.260088913738152,
"learning_rate": 1.6105463786953502e-05,
"loss": 1.016,
"step": 431
},
{
"epoch": 0.3131569409206234,
"grad_norm": 1.9542058053758606,
"learning_rate": 1.6086837524117755e-05,
"loss": 1.0546,
"step": 432
},
{
"epoch": 0.31388184124682855,
"grad_norm": 2.279040211460153,
"learning_rate": 1.6068177654383042e-05,
"loss": 1.1363,
"step": 433
},
{
"epoch": 0.3146067415730337,
"grad_norm": 2.4083685524014933,
"learning_rate": 1.6049484280775012e-05,
"loss": 1.0617,
"step": 434
},
{
"epoch": 0.31533164189923885,
"grad_norm": 2.3755540476745844,
"learning_rate": 1.6030757506504284e-05,
"loss": 1.0368,
"step": 435
},
{
"epoch": 0.316056542225444,
"grad_norm": 2.1752306039124565,
"learning_rate": 1.6011997434965906e-05,
"loss": 1.0696,
"step": 436
},
{
"epoch": 0.31678144255164914,
"grad_norm": 2.0792526403603557,
"learning_rate": 1.5993204169738757e-05,
"loss": 1.0583,
"step": 437
},
{
"epoch": 0.3175063428778543,
"grad_norm": 2.1158935057351838,
"learning_rate": 1.5974377814584986e-05,
"loss": 1.1241,
"step": 438
},
{
"epoch": 0.31823124320405943,
"grad_norm": 1.7878941702995481,
"learning_rate": 1.5955518473449448e-05,
"loss": 1.1129,
"step": 439
},
{
"epoch": 0.3189561435302646,
"grad_norm": 2.1238392656978404,
"learning_rate": 1.593662625045912e-05,
"loss": 1.0596,
"step": 440
},
{
"epoch": 0.3196810438564697,
"grad_norm": 2.1813159053603104,
"learning_rate": 1.591770124992252e-05,
"loss": 1.1104,
"step": 441
},
{
"epoch": 0.3204059441826749,
"grad_norm": 1.8368381622027763,
"learning_rate": 1.589874357632914e-05,
"loss": 1.0579,
"step": 442
},
{
"epoch": 0.32113084450888,
"grad_norm": 2.233702212234041,
"learning_rate": 1.587975333434888e-05,
"loss": 1.1181,
"step": 443
},
{
"epoch": 0.32185574483508517,
"grad_norm": 2.112351085565823,
"learning_rate": 1.586073062883144e-05,
"loss": 1.028,
"step": 444
},
{
"epoch": 0.3225806451612903,
"grad_norm": 3.4664697517256595,
"learning_rate": 1.5841675564805773e-05,
"loss": 1.0466,
"step": 445
},
{
"epoch": 0.32330554548749546,
"grad_norm": 1.9010894055087597,
"learning_rate": 1.582258824747949e-05,
"loss": 1.0831,
"step": 446
},
{
"epoch": 0.3240304458137006,
"grad_norm": 2.9306779332036714,
"learning_rate": 1.580346878223827e-05,
"loss": 1.2466,
"step": 447
},
{
"epoch": 0.32475534613990575,
"grad_norm": 1.9192009119584101,
"learning_rate": 1.5784317274645294e-05,
"loss": 1.1088,
"step": 448
},
{
"epoch": 0.3254802464661109,
"grad_norm": 2.242207905403273,
"learning_rate": 1.576513383044066e-05,
"loss": 1.1625,
"step": 449
},
{
"epoch": 0.32620514679231605,
"grad_norm": 2.224398160967933,
"learning_rate": 1.5745918555540796e-05,
"loss": 1.1488,
"step": 450
},
{
"epoch": 0.3269300471185212,
"grad_norm": 2.0861425073253503,
"learning_rate": 1.572667155603787e-05,
"loss": 0.9955,
"step": 451
},
{
"epoch": 0.32765494744472634,
"grad_norm": 1.92493109968563,
"learning_rate": 1.5707392938199208e-05,
"loss": 0.9444,
"step": 452
},
{
"epoch": 0.3283798477709315,
"grad_norm": 1.9283323786352506,
"learning_rate": 1.5688082808466714e-05,
"loss": 0.9664,
"step": 453
},
{
"epoch": 0.32910474809713663,
"grad_norm": 2.5055746794884093,
"learning_rate": 1.566874127345628e-05,
"loss": 1.0542,
"step": 454
},
{
"epoch": 0.3298296484233418,
"grad_norm": 2.4180849774657784,
"learning_rate": 1.5649368439957182e-05,
"loss": 1.1004,
"step": 455
},
{
"epoch": 0.33055454874954693,
"grad_norm": 3.1969300216088574,
"learning_rate": 1.562996441493151e-05,
"loss": 1.1926,
"step": 456
},
{
"epoch": 0.3312794490757521,
"grad_norm": 1.8450416397567433,
"learning_rate": 1.561052930551357e-05,
"loss": 0.9204,
"step": 457
},
{
"epoch": 0.3320043494019572,
"grad_norm": 2.1352988507915076,
"learning_rate": 1.5591063219009296e-05,
"loss": 1.1082,
"step": 458
},
{
"epoch": 0.33272924972816237,
"grad_norm": 2.0068487127258563,
"learning_rate": 1.5571566262895637e-05,
"loss": 1.0769,
"step": 459
},
{
"epoch": 0.3334541500543675,
"grad_norm": 2.158036051650137,
"learning_rate": 1.5552038544820008e-05,
"loss": 1.0965,
"step": 460
},
{
"epoch": 0.33417905038057266,
"grad_norm": 2.260378271679646,
"learning_rate": 1.5532480172599646e-05,
"loss": 1.08,
"step": 461
},
{
"epoch": 0.3349039507067778,
"grad_norm": 2.0768476079893317,
"learning_rate": 1.5512891254221046e-05,
"loss": 1.0548,
"step": 462
},
{
"epoch": 0.33562885103298296,
"grad_norm": 2.227359781654928,
"learning_rate": 1.549327189783935e-05,
"loss": 1.0364,
"step": 463
},
{
"epoch": 0.3363537513591881,
"grad_norm": 2.008303234643065,
"learning_rate": 1.5473622211777763e-05,
"loss": 1.0596,
"step": 464
},
{
"epoch": 0.33707865168539325,
"grad_norm": 2.1053343622780525,
"learning_rate": 1.5453942304526945e-05,
"loss": 1.0106,
"step": 465
},
{
"epoch": 0.3378035520115984,
"grad_norm": 2.2528412924953645,
"learning_rate": 1.543423228474441e-05,
"loss": 1.1579,
"step": 466
},
{
"epoch": 0.33852845233780354,
"grad_norm": 2.098527027839517,
"learning_rate": 1.5414492261253937e-05,
"loss": 1.0913,
"step": 467
},
{
"epoch": 0.3392533526640087,
"grad_norm": 2.3910084145103383,
"learning_rate": 1.539472234304496e-05,
"loss": 1.0481,
"step": 468
},
{
"epoch": 0.33997825299021384,
"grad_norm": 2.0843546569961937,
"learning_rate": 1.537492263927196e-05,
"loss": 1.1351,
"step": 469
},
{
"epoch": 0.340703153316419,
"grad_norm": 2.0698901517851294,
"learning_rate": 1.5355093259253897e-05,
"loss": 1.1095,
"step": 470
},
{
"epoch": 0.34142805364262413,
"grad_norm": 2.2916177664815867,
"learning_rate": 1.533523431247355e-05,
"loss": 1.094,
"step": 471
},
{
"epoch": 0.3421529539688293,
"grad_norm": 2.000471695373447,
"learning_rate": 1.531534590857696e-05,
"loss": 1.0647,
"step": 472
},
{
"epoch": 0.3428778542950344,
"grad_norm": 2.048745212856722,
"learning_rate": 1.529542815737281e-05,
"loss": 1.0448,
"step": 473
},
{
"epoch": 0.34360275462123957,
"grad_norm": 2.104486287385562,
"learning_rate": 1.52754811688318e-05,
"loss": 1.0525,
"step": 474
},
{
"epoch": 0.3443276549474447,
"grad_norm": 2.172054502687611,
"learning_rate": 1.525550505308608e-05,
"loss": 1.0437,
"step": 475
},
{
"epoch": 0.34505255527364986,
"grad_norm": 2.1612251883461355,
"learning_rate": 1.52354999204286e-05,
"loss": 1.1142,
"step": 476
},
{
"epoch": 0.345777455599855,
"grad_norm": 2.5598958893453987,
"learning_rate": 1.5215465881312526e-05,
"loss": 1.0597,
"step": 477
},
{
"epoch": 0.34650235592606016,
"grad_norm": 2.1402492854032875,
"learning_rate": 1.519540304635062e-05,
"loss": 1.0678,
"step": 478
},
{
"epoch": 0.3472272562522653,
"grad_norm": 2.386151874903598,
"learning_rate": 1.5175311526314642e-05,
"loss": 1.1297,
"step": 479
},
{
"epoch": 0.34795215657847045,
"grad_norm": 2.8147411958053734,
"learning_rate": 1.5155191432134721e-05,
"loss": 1.047,
"step": 480
},
{
"epoch": 0.3486770569046756,
"grad_norm": 2.185563612843593,
"learning_rate": 1.513504287489875e-05,
"loss": 1.0674,
"step": 481
},
{
"epoch": 0.34940195723088074,
"grad_norm": 2.1264088878183705,
"learning_rate": 1.5114865965851779e-05,
"loss": 1.1097,
"step": 482
},
{
"epoch": 0.3501268575570859,
"grad_norm": 2.294219589338203,
"learning_rate": 1.509466081639539e-05,
"loss": 0.9861,
"step": 483
},
{
"epoch": 0.35085175788329104,
"grad_norm": 2.2845015506312474,
"learning_rate": 1.507442753808709e-05,
"loss": 1.0646,
"step": 484
},
{
"epoch": 0.3515766582094962,
"grad_norm": 2.1560916077937606,
"learning_rate": 1.5054166242639694e-05,
"loss": 1.0056,
"step": 485
},
{
"epoch": 0.35230155853570133,
"grad_norm": 2.0570915443776654,
"learning_rate": 1.5033877041920703e-05,
"loss": 1.0926,
"step": 486
},
{
"epoch": 0.3530264588619065,
"grad_norm": 1.9130291874506817,
"learning_rate": 1.5013560047951689e-05,
"loss": 0.9783,
"step": 487
},
{
"epoch": 0.3537513591881116,
"grad_norm": 2.3476451826235625,
"learning_rate": 1.499321537290768e-05,
"loss": 1.0546,
"step": 488
},
{
"epoch": 0.35447625951431677,
"grad_norm": 2.3654378069877966,
"learning_rate": 1.4972843129116537e-05,
"loss": 1.0442,
"step": 489
},
{
"epoch": 0.3552011598405219,
"grad_norm": 1.9825769936244912,
"learning_rate": 1.4952443429058334e-05,
"loss": 1.0081,
"step": 490
},
{
"epoch": 0.35592606016672707,
"grad_norm": 2.2643797127980747,
"learning_rate": 1.493201638536474e-05,
"loss": 1.0317,
"step": 491
},
{
"epoch": 0.3566509604929322,
"grad_norm": 2.4516169299014465,
"learning_rate": 1.4911562110818392e-05,
"loss": 0.9924,
"step": 492
},
{
"epoch": 0.35737586081913736,
"grad_norm": 2.446561602895059,
"learning_rate": 1.4891080718352275e-05,
"loss": 1.0407,
"step": 493
},
{
"epoch": 0.3581007611453425,
"grad_norm": 2.0486524427646144,
"learning_rate": 1.4870572321049102e-05,
"loss": 1.0797,
"step": 494
},
{
"epoch": 0.35882566147154765,
"grad_norm": 2.485346536887422,
"learning_rate": 1.4850037032140683e-05,
"loss": 0.9983,
"step": 495
},
{
"epoch": 0.3595505617977528,
"grad_norm": 3.0439426290182094,
"learning_rate": 1.4829474965007303e-05,
"loss": 1.152,
"step": 496
},
{
"epoch": 0.36027546212395795,
"grad_norm": 2.540523235848105,
"learning_rate": 1.4808886233177096e-05,
"loss": 1.2146,
"step": 497
},
{
"epoch": 0.3610003624501631,
"grad_norm": 2.5194385485382207,
"learning_rate": 1.4788270950325424e-05,
"loss": 1.0989,
"step": 498
},
{
"epoch": 0.36172526277636824,
"grad_norm": 2.3200096809507693,
"learning_rate": 1.4767629230274236e-05,
"loss": 1.2226,
"step": 499
},
{
"epoch": 0.3624501631025734,
"grad_norm": 2.0411294097073363,
"learning_rate": 1.474696118699145e-05,
"loss": 1.0421,
"step": 500
},
{
"epoch": 0.36317506342877853,
"grad_norm": 2.3819141883855597,
"learning_rate": 1.4726266934590321e-05,
"loss": 1.0162,
"step": 501
},
{
"epoch": 0.3638999637549837,
"grad_norm": 1.9177330075859071,
"learning_rate": 1.4705546587328811e-05,
"loss": 1.0171,
"step": 502
},
{
"epoch": 0.3646248640811888,
"grad_norm": 2.06883784081889,
"learning_rate": 1.4684800259608958e-05,
"loss": 1.0567,
"step": 503
},
{
"epoch": 0.365349764407394,
"grad_norm": 2.796404840199928,
"learning_rate": 1.4664028065976245e-05,
"loss": 1.0567,
"step": 504
},
{
"epoch": 0.3660746647335991,
"grad_norm": 2.0344928788193433,
"learning_rate": 1.4643230121118965e-05,
"loss": 1.1192,
"step": 505
},
{
"epoch": 0.36679956505980427,
"grad_norm": 1.931575463278115,
"learning_rate": 1.4622406539867593e-05,
"loss": 1.0491,
"step": 506
},
{
"epoch": 0.3675244653860094,
"grad_norm": 2.5445474863300106,
"learning_rate": 1.460155743719414e-05,
"loss": 1.0213,
"step": 507
},
{
"epoch": 0.36824936571221456,
"grad_norm": 2.0670836939977564,
"learning_rate": 1.4580682928211545e-05,
"loss": 0.92,
"step": 508
},
{
"epoch": 0.3689742660384197,
"grad_norm": 2.575085198850707,
"learning_rate": 1.4559783128173e-05,
"loss": 1.0829,
"step": 509
},
{
"epoch": 0.36969916636462485,
"grad_norm": 2.7113350290504723,
"learning_rate": 1.4538858152471349e-05,
"loss": 1.1535,
"step": 510
},
{
"epoch": 0.37042406669083,
"grad_norm": 2.312117149467026,
"learning_rate": 1.4517908116638433e-05,
"loss": 1.0897,
"step": 511
},
{
"epoch": 0.37114896701703515,
"grad_norm": 2.3702995154115656,
"learning_rate": 1.4496933136344453e-05,
"loss": 1.0635,
"step": 512
},
{
"epoch": 0.3718738673432403,
"grad_norm": 2.4501757498607537,
"learning_rate": 1.4475933327397342e-05,
"loss": 1.0603,
"step": 513
},
{
"epoch": 0.37259876766944544,
"grad_norm": 2.023783253482957,
"learning_rate": 1.445490880574211e-05,
"loss": 1.0789,
"step": 514
},
{
"epoch": 0.3733236679956506,
"grad_norm": 3.2864848049242648,
"learning_rate": 1.4433859687460216e-05,
"loss": 1.0952,
"step": 515
},
{
"epoch": 0.37404856832185573,
"grad_norm": 2.4191595705564737,
"learning_rate": 1.4412786088768923e-05,
"loss": 1.1147,
"step": 516
},
{
"epoch": 0.3747734686480609,
"grad_norm": 2.4447277431874825,
"learning_rate": 1.439168812602065e-05,
"loss": 1.0393,
"step": 517
},
{
"epoch": 0.37549836897426603,
"grad_norm": 1.9206969249262023,
"learning_rate": 1.437056591570235e-05,
"loss": 1.0022,
"step": 518
},
{
"epoch": 0.3762232693004712,
"grad_norm": 2.359427663223057,
"learning_rate": 1.4349419574434838e-05,
"loss": 1.093,
"step": 519
},
{
"epoch": 0.3769481696266763,
"grad_norm": 2.3010718359012103,
"learning_rate": 1.4328249218972168e-05,
"loss": 1.0408,
"step": 520
},
{
"epoch": 0.37767306995288147,
"grad_norm": 2.2594128382319596,
"learning_rate": 1.4307054966200984e-05,
"loss": 0.9412,
"step": 521
},
{
"epoch": 0.3783979702790866,
"grad_norm": 2.1054269973126747,
"learning_rate": 1.4285836933139865e-05,
"loss": 1.0254,
"step": 522
},
{
"epoch": 0.37912287060529176,
"grad_norm": 2.168937643340505,
"learning_rate": 1.42645952369387e-05,
"loss": 1.0537,
"step": 523
},
{
"epoch": 0.3798477709314969,
"grad_norm": 2.061073976619602,
"learning_rate": 1.4243329994878017e-05,
"loss": 1.1168,
"step": 524
},
{
"epoch": 0.38057267125770206,
"grad_norm": 2.299186158706926,
"learning_rate": 1.4222041324368347e-05,
"loss": 1.1138,
"step": 525
},
{
"epoch": 0.3812975715839072,
"grad_norm": 2.251869960906443,
"learning_rate": 1.4200729342949577e-05,
"loss": 0.9248,
"step": 526
},
{
"epoch": 0.38202247191011235,
"grad_norm": 2.2506167583969856,
"learning_rate": 1.4179394168290309e-05,
"loss": 1.0757,
"step": 527
},
{
"epoch": 0.3827473722363175,
"grad_norm": 2.916531852779516,
"learning_rate": 1.4158035918187182e-05,
"loss": 1.1818,
"step": 528
},
{
"epoch": 0.38347227256252264,
"grad_norm": 2.3787214433218806,
"learning_rate": 1.4136654710564251e-05,
"loss": 1.0336,
"step": 529
},
{
"epoch": 0.3841971728887278,
"grad_norm": 2.4765072845885068,
"learning_rate": 1.4115250663472326e-05,
"loss": 1.0506,
"step": 530
},
{
"epoch": 0.38492207321493294,
"grad_norm": 2.407842302823844,
"learning_rate": 1.4093823895088315e-05,
"loss": 1.1,
"step": 531
},
{
"epoch": 0.3856469735411381,
"grad_norm": 2.3150861515619487,
"learning_rate": 1.4072374523714577e-05,
"loss": 1.0437,
"step": 532
},
{
"epoch": 0.38637187386734323,
"grad_norm": 2.0825864756582506,
"learning_rate": 1.4050902667778272e-05,
"loss": 1.1277,
"step": 533
},
{
"epoch": 0.3870967741935484,
"grad_norm": 2.243168272872563,
"learning_rate": 1.402940844583069e-05,
"loss": 1.0506,
"step": 534
},
{
"epoch": 0.3878216745197535,
"grad_norm": 2.185661024343008,
"learning_rate": 1.4007891976546627e-05,
"loss": 1.088,
"step": 535
},
{
"epoch": 0.38854657484595867,
"grad_norm": 2.2131084797482186,
"learning_rate": 1.3986353378723696e-05,
"loss": 1.0654,
"step": 536
},
{
"epoch": 0.3892714751721638,
"grad_norm": 2.4512297593144177,
"learning_rate": 1.3964792771281702e-05,
"loss": 1.2112,
"step": 537
},
{
"epoch": 0.38999637549836896,
"grad_norm": 1.9649805513289704,
"learning_rate": 1.394321027326195e-05,
"loss": 1.0487,
"step": 538
},
{
"epoch": 0.3907212758245741,
"grad_norm": 2.0806905976421985,
"learning_rate": 1.392160600382663e-05,
"loss": 1.1033,
"step": 539
},
{
"epoch": 0.39144617615077926,
"grad_norm": 2.117182296397933,
"learning_rate": 1.3899980082258122e-05,
"loss": 1.0957,
"step": 540
},
{
"epoch": 0.3921710764769844,
"grad_norm": 2.4101205011369777,
"learning_rate": 1.3878332627958365e-05,
"loss": 1.1324,
"step": 541
},
{
"epoch": 0.39289597680318955,
"grad_norm": 2.0554352528605255,
"learning_rate": 1.385666376044817e-05,
"loss": 1.0756,
"step": 542
},
{
"epoch": 0.3936208771293947,
"grad_norm": 2.8127997073836806,
"learning_rate": 1.3834973599366588e-05,
"loss": 1.203,
"step": 543
},
{
"epoch": 0.39434577745559984,
"grad_norm": 2.249229842195169,
"learning_rate": 1.3813262264470235e-05,
"loss": 1.0307,
"step": 544
},
{
"epoch": 0.395070677781805,
"grad_norm": 2.371447175626824,
"learning_rate": 1.3791529875632628e-05,
"loss": 1.1348,
"step": 545
},
{
"epoch": 0.39579557810801014,
"grad_norm": 2.2035574911903697,
"learning_rate": 1.3769776552843532e-05,
"loss": 1.1021,
"step": 546
},
{
"epoch": 0.3965204784342153,
"grad_norm": 2.749098638445283,
"learning_rate": 1.3748002416208292e-05,
"loss": 1.1599,
"step": 547
},
{
"epoch": 0.39724537876042043,
"grad_norm": 1.9103185758921004,
"learning_rate": 1.372620758594717e-05,
"loss": 0.9921,
"step": 548
},
{
"epoch": 0.3979702790866256,
"grad_norm": 2.1532622643710946,
"learning_rate": 1.3704392182394686e-05,
"loss": 1.0174,
"step": 549
},
{
"epoch": 0.3986951794128307,
"grad_norm": 2.0770483467925875,
"learning_rate": 1.3682556325998947e-05,
"loss": 1.1642,
"step": 550
},
{
"epoch": 0.39942007973903587,
"grad_norm": 2.0510030802176002,
"learning_rate": 1.3660700137320986e-05,
"loss": 1.0736,
"step": 551
},
{
"epoch": 0.400144980065241,
"grad_norm": 2.4143942712642583,
"learning_rate": 1.3638823737034095e-05,
"loss": 1.0679,
"step": 552
},
{
"epoch": 0.40086988039144617,
"grad_norm": 2.087827823927243,
"learning_rate": 1.3616927245923157e-05,
"loss": 1.077,
"step": 553
},
{
"epoch": 0.4015947807176513,
"grad_norm": 2.155634016832207,
"learning_rate": 1.359501078488399e-05,
"loss": 0.9344,
"step": 554
},
{
"epoch": 0.40231968104385646,
"grad_norm": 2.9410176032472104,
"learning_rate": 1.357307447492266e-05,
"loss": 0.9876,
"step": 555
},
{
"epoch": 0.4030445813700616,
"grad_norm": 2.3785327314059965,
"learning_rate": 1.3551118437154833e-05,
"loss": 0.8992,
"step": 556
},
{
"epoch": 0.40376948169626675,
"grad_norm": 2.206200464844771,
"learning_rate": 1.3529142792805087e-05,
"loss": 1.0472,
"step": 557
},
{
"epoch": 0.4044943820224719,
"grad_norm": 2.4336430655807058,
"learning_rate": 1.3507147663206257e-05,
"loss": 1.147,
"step": 558
},
{
"epoch": 0.40521928234867705,
"grad_norm": 2.0423284010474627,
"learning_rate": 1.348513316979877e-05,
"loss": 1.0467,
"step": 559
},
{
"epoch": 0.4059441826748822,
"grad_norm": 2.6261549064320215,
"learning_rate": 1.346309943412995e-05,
"loss": 0.954,
"step": 560
},
{
"epoch": 0.40666908300108734,
"grad_norm": 2.8231886882036386,
"learning_rate": 1.3441046577853371e-05,
"loss": 1.088,
"step": 561
},
{
"epoch": 0.4073939833272925,
"grad_norm": 2.5545065191928766,
"learning_rate": 1.3418974722728177e-05,
"loss": 1.1541,
"step": 562
},
{
"epoch": 0.40811888365349763,
"grad_norm": 2.714166874930808,
"learning_rate": 1.3396883990618404e-05,
"loss": 1.1053,
"step": 563
},
{
"epoch": 0.4088437839797028,
"grad_norm": 1.9784578326236464,
"learning_rate": 1.3374774503492316e-05,
"loss": 1.0065,
"step": 564
},
{
"epoch": 0.4095686843059079,
"grad_norm": 2.1486595608204357,
"learning_rate": 1.3352646383421728e-05,
"loss": 1.0184,
"step": 565
},
{
"epoch": 0.4102935846321131,
"grad_norm": 2.5242559447298483,
"learning_rate": 1.3330499752581331e-05,
"loss": 1.0812,
"step": 566
},
{
"epoch": 0.4110184849583182,
"grad_norm": 1.9089213008734895,
"learning_rate": 1.3308334733248019e-05,
"loss": 0.9932,
"step": 567
},
{
"epoch": 0.41174338528452337,
"grad_norm": 1.890310480638506,
"learning_rate": 1.3286151447800211e-05,
"loss": 0.9832,
"step": 568
},
{
"epoch": 0.4124682856107285,
"grad_norm": 2.038892051314124,
"learning_rate": 1.3263950018717184e-05,
"loss": 0.9662,
"step": 569
},
{
"epoch": 0.41319318593693366,
"grad_norm": 2.241417479572917,
"learning_rate": 1.3241730568578383e-05,
"loss": 1.0493,
"step": 570
},
{
"epoch": 0.4139180862631388,
"grad_norm": 2.260815460922618,
"learning_rate": 1.3219493220062756e-05,
"loss": 1.0101,
"step": 571
},
{
"epoch": 0.41464298658934395,
"grad_norm": 2.456266733687105,
"learning_rate": 1.3197238095948066e-05,
"loss": 1.0541,
"step": 572
},
{
"epoch": 0.4153678869155491,
"grad_norm": 2.5507405445700377,
"learning_rate": 1.3174965319110231e-05,
"loss": 1.0279,
"step": 573
},
{
"epoch": 0.41609278724175425,
"grad_norm": 2.847963262037451,
"learning_rate": 1.3152675012522629e-05,
"loss": 1.108,
"step": 574
},
{
"epoch": 0.4168176875679594,
"grad_norm": 2.1767458920508784,
"learning_rate": 1.313036729925541e-05,
"loss": 1.0479,
"step": 575
},
{
"epoch": 0.41754258789416454,
"grad_norm": 3.6818743459800234,
"learning_rate": 1.3108042302474858e-05,
"loss": 1.1633,
"step": 576
},
{
"epoch": 0.4182674882203697,
"grad_norm": 2.4735330534741595,
"learning_rate": 1.3085700145442653e-05,
"loss": 0.9817,
"step": 577
},
{
"epoch": 0.41899238854657483,
"grad_norm": 2.099259381744443,
"learning_rate": 1.3063340951515246e-05,
"loss": 0.9347,
"step": 578
},
{
"epoch": 0.41971728887278,
"grad_norm": 2.3256524631392916,
"learning_rate": 1.3040964844143144e-05,
"loss": 1.0972,
"step": 579
},
{
"epoch": 0.42044218919898513,
"grad_norm": 2.5886329967045385,
"learning_rate": 1.301857194687023e-05,
"loss": 1.1258,
"step": 580
},
{
"epoch": 0.4211670895251903,
"grad_norm": 2.0842353427178097,
"learning_rate": 1.2996162383333097e-05,
"loss": 1.0067,
"step": 581
},
{
"epoch": 0.4218919898513954,
"grad_norm": 2.0384216517802662,
"learning_rate": 1.2973736277260351e-05,
"loss": 1.0734,
"step": 582
},
{
"epoch": 0.42261689017760057,
"grad_norm": 2.4453213613902784,
"learning_rate": 1.295129375247194e-05,
"loss": 1.1793,
"step": 583
},
{
"epoch": 0.4233417905038057,
"grad_norm": 1.9025172070678824,
"learning_rate": 1.292883493287846e-05,
"loss": 1.083,
"step": 584
},
{
"epoch": 0.42406669083001086,
"grad_norm": 2.6363997339281595,
"learning_rate": 1.290635994248047e-05,
"loss": 1.0293,
"step": 585
},
{
"epoch": 0.424791591156216,
"grad_norm": 2.311748807183176,
"learning_rate": 1.2883868905367818e-05,
"loss": 1.1425,
"step": 586
},
{
"epoch": 0.42551649148242116,
"grad_norm": 2.357170974552937,
"learning_rate": 1.2861361945718947e-05,
"loss": 1.2254,
"step": 587
},
{
"epoch": 0.4262413918086263,
"grad_norm": 2.176179466256725,
"learning_rate": 1.2838839187800218e-05,
"loss": 1.0198,
"step": 588
},
{
"epoch": 0.42696629213483145,
"grad_norm": 2.9614059903538967,
"learning_rate": 1.2816300755965202e-05,
"loss": 1.1044,
"step": 589
},
{
"epoch": 0.4276911924610366,
"grad_norm": 2.089967789540345,
"learning_rate": 1.2793746774654034e-05,
"loss": 0.9743,
"step": 590
},
{
"epoch": 0.42841609278724174,
"grad_norm": 2.0512215889422114,
"learning_rate": 1.277117736839268e-05,
"loss": 1.0381,
"step": 591
},
{
"epoch": 0.4291409931134469,
"grad_norm": 3.3593984924680065,
"learning_rate": 1.2748592661792278e-05,
"loss": 1.0852,
"step": 592
},
{
"epoch": 0.42986589343965204,
"grad_norm": 1.8379649951016293,
"learning_rate": 1.2725992779548451e-05,
"loss": 1.0407,
"step": 593
},
{
"epoch": 0.4305907937658572,
"grad_norm": 2.0580648499060614,
"learning_rate": 1.2703377846440591e-05,
"loss": 0.9797,
"step": 594
},
{
"epoch": 0.43131569409206233,
"grad_norm": 2.547624173063941,
"learning_rate": 1.2680747987331215e-05,
"loss": 1.0976,
"step": 595
},
{
"epoch": 0.4320405944182675,
"grad_norm": 2.358740461402887,
"learning_rate": 1.2658103327165226e-05,
"loss": 0.9832,
"step": 596
},
{
"epoch": 0.4327654947444726,
"grad_norm": 2.2958356138544223,
"learning_rate": 1.2635443990969266e-05,
"loss": 1.0707,
"step": 597
},
{
"epoch": 0.43349039507067777,
"grad_norm": 2.2673074349057036,
"learning_rate": 1.2612770103850998e-05,
"loss": 1.155,
"step": 598
},
{
"epoch": 0.4342152953968829,
"grad_norm": 2.191734084847357,
"learning_rate": 1.2590081790998414e-05,
"loss": 1.1389,
"step": 599
},
{
"epoch": 0.43494019572308806,
"grad_norm": 2.2857227047676365,
"learning_rate": 1.2567379177679178e-05,
"loss": 0.9963,
"step": 600
},
{
"epoch": 0.4356650960492932,
"grad_norm": 2.1784697059963047,
"learning_rate": 1.254466238923989e-05,
"loss": 1.0586,
"step": 601
},
{
"epoch": 0.43638999637549836,
"grad_norm": 2.521226757264337,
"learning_rate": 1.2521931551105427e-05,
"loss": 1.1325,
"step": 602
},
{
"epoch": 0.4371148967017035,
"grad_norm": 2.061977865722029,
"learning_rate": 1.2499186788778223e-05,
"loss": 0.9365,
"step": 603
},
{
"epoch": 0.43783979702790865,
"grad_norm": 2.447645929802628,
"learning_rate": 1.2476428227837607e-05,
"loss": 1.0145,
"step": 604
},
{
"epoch": 0.4385646973541138,
"grad_norm": 2.267005489506538,
"learning_rate": 1.2453655993939088e-05,
"loss": 1.0399,
"step": 605
},
{
"epoch": 0.43928959768031894,
"grad_norm": 2.084407985939619,
"learning_rate": 1.2430870212813661e-05,
"loss": 0.9565,
"step": 606
},
{
"epoch": 0.4400144980065241,
"grad_norm": 2.6037410215898458,
"learning_rate": 1.2408071010267125e-05,
"loss": 1.0644,
"step": 607
},
{
"epoch": 0.44073939833272924,
"grad_norm": 2.2081346806030235,
"learning_rate": 1.2385258512179375e-05,
"loss": 0.9149,
"step": 608
},
{
"epoch": 0.4414642986589344,
"grad_norm": 2.3544684711758173,
"learning_rate": 1.2362432844503725e-05,
"loss": 1.0141,
"step": 609
},
{
"epoch": 0.44218919898513953,
"grad_norm": 2.0791622079935914,
"learning_rate": 1.233959413326619e-05,
"loss": 1.0256,
"step": 610
},
{
"epoch": 0.4429140993113447,
"grad_norm": 2.968477393667513,
"learning_rate": 1.2316742504564811e-05,
"loss": 1.1866,
"step": 611
},
{
"epoch": 0.4436389996375498,
"grad_norm": 2.076859528680039,
"learning_rate": 1.2293878084568944e-05,
"loss": 0.9385,
"step": 612
},
{
"epoch": 0.44436389996375497,
"grad_norm": 2.499714197915474,
"learning_rate": 1.2271000999518563e-05,
"loss": 1.1666,
"step": 613
},
{
"epoch": 0.4450888002899601,
"grad_norm": 2.003190354705559,
"learning_rate": 1.2248111375723585e-05,
"loss": 1.0057,
"step": 614
},
{
"epoch": 0.44581370061616526,
"grad_norm": 2.037354395792757,
"learning_rate": 1.2225209339563144e-05,
"loss": 0.9492,
"step": 615
},
{
"epoch": 0.4465386009423704,
"grad_norm": 2.586450689063457,
"learning_rate": 1.2202295017484911e-05,
"loss": 1.0752,
"step": 616
},
{
"epoch": 0.44726350126857556,
"grad_norm": 1.903194057594901,
"learning_rate": 1.2179368536004382e-05,
"loss": 0.9995,
"step": 617
},
{
"epoch": 0.4479884015947807,
"grad_norm": 2.3340177701835523,
"learning_rate": 1.2156430021704196e-05,
"loss": 1.0832,
"step": 618
},
{
"epoch": 0.44871330192098585,
"grad_norm": 2.181064156987608,
"learning_rate": 1.2133479601233431e-05,
"loss": 0.9726,
"step": 619
},
{
"epoch": 0.449438202247191,
"grad_norm": 2.1594915956868097,
"learning_rate": 1.2110517401306896e-05,
"loss": 0.9393,
"step": 620
},
{
"epoch": 0.45016310257339615,
"grad_norm": 2.5184901765324934,
"learning_rate": 1.2087543548704435e-05,
"loss": 1.1569,
"step": 621
},
{
"epoch": 0.4508880028996013,
"grad_norm": 2.2361203028149585,
"learning_rate": 1.2064558170270236e-05,
"loss": 1.0704,
"step": 622
},
{
"epoch": 0.45161290322580644,
"grad_norm": 1.9402238173723454,
"learning_rate": 1.2041561392912118e-05,
"loss": 1.0127,
"step": 623
},
{
"epoch": 0.4523378035520116,
"grad_norm": 2.6102129250515724,
"learning_rate": 1.2018553343600841e-05,
"loss": 1.1364,
"step": 624
},
{
"epoch": 0.45306270387821673,
"grad_norm": 2.0267045043502883,
"learning_rate": 1.1995534149369397e-05,
"loss": 1.0408,
"step": 625
},
{
"epoch": 0.4537876042044219,
"grad_norm": 2.0973318190957353,
"learning_rate": 1.1972503937312315e-05,
"loss": 0.994,
"step": 626
},
{
"epoch": 0.454512504530627,
"grad_norm": 1.9897533182088447,
"learning_rate": 1.194946283458495e-05,
"loss": 0.9902,
"step": 627
},
{
"epoch": 0.4552374048568322,
"grad_norm": 2.624450240222456,
"learning_rate": 1.1926410968402791e-05,
"loss": 1.1201,
"step": 628
},
{
"epoch": 0.4559623051830373,
"grad_norm": 1.9043526191630893,
"learning_rate": 1.1903348466040758e-05,
"loss": 1.0846,
"step": 629
},
{
"epoch": 0.45668720550924247,
"grad_norm": 1.9419173600857174,
"learning_rate": 1.1880275454832493e-05,
"loss": 1.0469,
"step": 630
},
{
"epoch": 0.4574121058354476,
"grad_norm": 2.2783826877349673,
"learning_rate": 1.1857192062169656e-05,
"loss": 0.9387,
"step": 631
},
{
"epoch": 0.45813700616165276,
"grad_norm": 2.093160188755417,
"learning_rate": 1.183409841550123e-05,
"loss": 1.1626,
"step": 632
},
{
"epoch": 0.4588619064878579,
"grad_norm": 2.203652115607489,
"learning_rate": 1.1810994642332808e-05,
"loss": 1.0042,
"step": 633
},
{
"epoch": 0.45958680681406305,
"grad_norm": 2.0594672175740008,
"learning_rate": 1.1787880870225905e-05,
"loss": 1.0613,
"step": 634
},
{
"epoch": 0.4603117071402682,
"grad_norm": 2.1686023194883055,
"learning_rate": 1.1764757226797233e-05,
"loss": 1.0283,
"step": 635
},
{
"epoch": 0.46103660746647335,
"grad_norm": 2.3713976995681327,
"learning_rate": 1.1741623839718002e-05,
"loss": 1.0412,
"step": 636
},
{
"epoch": 0.4617615077926785,
"grad_norm": 2.3932119999968005,
"learning_rate": 1.1718480836713228e-05,
"loss": 1.0953,
"step": 637
},
{
"epoch": 0.46248640811888364,
"grad_norm": 2.411654231766739,
"learning_rate": 1.169532834556102e-05,
"loss": 1.0146,
"step": 638
},
{
"epoch": 0.4632113084450888,
"grad_norm": 2.481106610277318,
"learning_rate": 1.167216649409187e-05,
"loss": 0.9666,
"step": 639
},
{
"epoch": 0.46393620877129393,
"grad_norm": 2.466302214254987,
"learning_rate": 1.164899541018794e-05,
"loss": 1.1808,
"step": 640
},
{
"epoch": 0.4646611090974991,
"grad_norm": 2.086736565800973,
"learning_rate": 1.1625815221782386e-05,
"loss": 1.066,
"step": 641
},
{
"epoch": 0.4653860094237042,
"grad_norm": 2.1129903671304344,
"learning_rate": 1.1602626056858617e-05,
"loss": 1.0851,
"step": 642
},
{
"epoch": 0.4661109097499094,
"grad_norm": 2.012359347751327,
"learning_rate": 1.1579428043449612e-05,
"loss": 0.9234,
"step": 643
},
{
"epoch": 0.4668358100761145,
"grad_norm": 1.731209443837663,
"learning_rate": 1.1556221309637204e-05,
"loss": 0.9405,
"step": 644
},
{
"epoch": 0.46756071040231967,
"grad_norm": 2.3738797051074587,
"learning_rate": 1.1533005983551361e-05,
"loss": 1.0442,
"step": 645
},
{
"epoch": 0.4682856107285248,
"grad_norm": 2.5124526996392333,
"learning_rate": 1.1509782193369514e-05,
"loss": 1.0362,
"step": 646
},
{
"epoch": 0.46901051105472996,
"grad_norm": 2.2245947511275403,
"learning_rate": 1.1486550067315803e-05,
"loss": 1.0992,
"step": 647
},
{
"epoch": 0.4697354113809351,
"grad_norm": 2.3135678458341484,
"learning_rate": 1.146330973366041e-05,
"loss": 1.0667,
"step": 648
},
{
"epoch": 0.47046031170714026,
"grad_norm": 1.8464379606767438,
"learning_rate": 1.1440061320718821e-05,
"loss": 1.094,
"step": 649
},
{
"epoch": 0.4711852120333454,
"grad_norm": 2.0764785227825864,
"learning_rate": 1.1416804956851138e-05,
"loss": 0.8967,
"step": 650
},
{
"epoch": 0.47191011235955055,
"grad_norm": 2.434020067387766,
"learning_rate": 1.1393540770461358e-05,
"loss": 0.9351,
"step": 651
},
{
"epoch": 0.4726350126857557,
"grad_norm": 2.2367136913648578,
"learning_rate": 1.1370268889996665e-05,
"loss": 0.9879,
"step": 652
},
{
"epoch": 0.47335991301196084,
"grad_norm": 2.16313174121554,
"learning_rate": 1.1346989443946734e-05,
"loss": 1.1194,
"step": 653
},
{
"epoch": 0.474084813338166,
"grad_norm": 2.165617974637285,
"learning_rate": 1.1323702560842998e-05,
"loss": 1.0268,
"step": 654
},
{
"epoch": 0.47480971366437114,
"grad_norm": 2.150563367555029,
"learning_rate": 1.1300408369257962e-05,
"loss": 1.1303,
"step": 655
},
{
"epoch": 0.4755346139905763,
"grad_norm": 2.1496140379016513,
"learning_rate": 1.127710699780448e-05,
"loss": 1.0181,
"step": 656
},
{
"epoch": 0.47625951431678143,
"grad_norm": 1.9308649623116936,
"learning_rate": 1.125379857513505e-05,
"loss": 1.0044,
"step": 657
},
{
"epoch": 0.4769844146429866,
"grad_norm": 2.520717730566298,
"learning_rate": 1.1230483229941092e-05,
"loss": 1.0142,
"step": 658
},
{
"epoch": 0.4777093149691917,
"grad_norm": 1.9969419284442884,
"learning_rate": 1.1207161090952255e-05,
"loss": 0.9586,
"step": 659
},
{
"epoch": 0.47843421529539687,
"grad_norm": 2.199551212478649,
"learning_rate": 1.1183832286935703e-05,
"loss": 1.0442,
"step": 660
},
{
"epoch": 0.479159115621602,
"grad_norm": 2.4903266005653433,
"learning_rate": 1.1160496946695388e-05,
"loss": 1.057,
"step": 661
},
{
"epoch": 0.47988401594780716,
"grad_norm": 2.0757922408522447,
"learning_rate": 1.1137155199071356e-05,
"loss": 1.0865,
"step": 662
},
{
"epoch": 0.4806089162740123,
"grad_norm": 1.9478487958266806,
"learning_rate": 1.1113807172939033e-05,
"loss": 0.9878,
"step": 663
},
{
"epoch": 0.48133381660021746,
"grad_norm": 2.5077394876515027,
"learning_rate": 1.1090452997208496e-05,
"loss": 1.001,
"step": 664
},
{
"epoch": 0.4820587169264226,
"grad_norm": 2.3619801940222542,
"learning_rate": 1.1067092800823798e-05,
"loss": 1.0786,
"step": 665
},
{
"epoch": 0.48278361725262775,
"grad_norm": 2.4608146249469485,
"learning_rate": 1.1043726712762213e-05,
"loss": 0.8398,
"step": 666
},
{
"epoch": 0.4835085175788329,
"grad_norm": 2.5682411824097273,
"learning_rate": 1.1020354862033553e-05,
"loss": 1.0639,
"step": 667
},
{
"epoch": 0.48423341790503804,
"grad_norm": 2.2377079055852636,
"learning_rate": 1.099697737767945e-05,
"loss": 1.0577,
"step": 668
},
{
"epoch": 0.4849583182312432,
"grad_norm": 2.4684218039338366,
"learning_rate": 1.097359438877263e-05,
"loss": 1.2059,
"step": 669
},
{
"epoch": 0.48568321855744834,
"grad_norm": 1.9986702143467359,
"learning_rate": 1.0950206024416228e-05,
"loss": 1.1099,
"step": 670
},
{
"epoch": 0.4864081188836535,
"grad_norm": 1.651255607571327,
"learning_rate": 1.0926812413743041e-05,
"loss": 1.0481,
"step": 671
},
{
"epoch": 0.48713301920985863,
"grad_norm": 2.1890746076308734,
"learning_rate": 1.0903413685914843e-05,
"loss": 0.9785,
"step": 672
},
{
"epoch": 0.4878579195360638,
"grad_norm": 1.983204547752865,
"learning_rate": 1.0880009970121655e-05,
"loss": 0.9511,
"step": 673
},
{
"epoch": 0.4885828198622689,
"grad_norm": 1.992684611945782,
"learning_rate": 1.0856601395581037e-05,
"loss": 1.0077,
"step": 674
},
{
"epoch": 0.48930772018847407,
"grad_norm": 2.155401888693332,
"learning_rate": 1.0833188091537386e-05,
"loss": 0.9072,
"step": 675
},
{
"epoch": 0.4900326205146792,
"grad_norm": 2.7840807715345406,
"learning_rate": 1.08097701872612e-05,
"loss": 1.0153,
"step": 676
},
{
"epoch": 0.49075752084088436,
"grad_norm": 2.3948835243364996,
"learning_rate": 1.0786347812048381e-05,
"loss": 1.0366,
"step": 677
},
{
"epoch": 0.4914824211670895,
"grad_norm": 2.3996815926961865,
"learning_rate": 1.0762921095219511e-05,
"loss": 1.1168,
"step": 678
},
{
"epoch": 0.49220732149329466,
"grad_norm": 2.186176414807414,
"learning_rate": 1.0739490166119155e-05,
"loss": 0.9772,
"step": 679
},
{
"epoch": 0.4929322218194998,
"grad_norm": 2.397287885437739,
"learning_rate": 1.0716055154115123e-05,
"loss": 1.0678,
"step": 680
},
{
"epoch": 0.49365712214570495,
"grad_norm": 2.257426550702264,
"learning_rate": 1.0692616188597775e-05,
"loss": 0.9975,
"step": 681
},
{
"epoch": 0.4943820224719101,
"grad_norm": 2.2392714970259324,
"learning_rate": 1.0669173398979292e-05,
"loss": 0.9476,
"step": 682
},
{
"epoch": 0.49510692279811525,
"grad_norm": 1.9243901111495054,
"learning_rate": 1.0645726914692973e-05,
"loss": 1.0169,
"step": 683
},
{
"epoch": 0.4958318231243204,
"grad_norm": 1.874543436460937,
"learning_rate": 1.062227686519252e-05,
"loss": 0.9925,
"step": 684
},
{
"epoch": 0.49655672345052554,
"grad_norm": 2.4206739807777744,
"learning_rate": 1.0598823379951315e-05,
"loss": 0.9653,
"step": 685
},
{
"epoch": 0.4972816237767307,
"grad_norm": 1.8452641728018429,
"learning_rate": 1.057536658846171e-05,
"loss": 1.1065,
"step": 686
},
{
"epoch": 0.49800652410293583,
"grad_norm": 2.615942811689107,
"learning_rate": 1.0551906620234312e-05,
"loss": 1.1092,
"step": 687
},
{
"epoch": 0.498731424429141,
"grad_norm": 2.7357084625917576,
"learning_rate": 1.0528443604797266e-05,
"loss": 1.1432,
"step": 688
},
{
"epoch": 0.4994563247553461,
"grad_norm": 2.2460612751605185,
"learning_rate": 1.0504977671695542e-05,
"loss": 0.9907,
"step": 689
},
{
"epoch": 0.5001812250815513,
"grad_norm": 2.753596180660813,
"learning_rate": 1.0481508950490222e-05,
"loss": 1.0457,
"step": 690
},
{
"epoch": 0.5001812250815513,
"eval_loss": 1.8697493076324463,
"eval_runtime": 1108.3592,
"eval_samples_per_second": 15.928,
"eval_steps_per_second": 0.498,
"step": 690
},
{
"epoch": 0.5009061254077565,
"grad_norm": 2.6999729213529817,
"learning_rate": 1.0458037570757781e-05,
"loss": 1.0595,
"step": 691
},
{
"epoch": 0.5016310257339616,
"grad_norm": 2.211696739230099,
"learning_rate": 1.0434563662089367e-05,
"loss": 0.9517,
"step": 692
},
{
"epoch": 0.5023559260601668,
"grad_norm": 2.4087448709687935,
"learning_rate": 1.04110873540901e-05,
"loss": 1.1667,
"step": 693
},
{
"epoch": 0.5030808263863719,
"grad_norm": 2.4196372563289645,
"learning_rate": 1.0387608776378337e-05,
"loss": 1.0019,
"step": 694
},
{
"epoch": 0.5038057267125771,
"grad_norm": 2.2271775487795424,
"learning_rate": 1.0364128058584974e-05,
"loss": 1.0789,
"step": 695
},
{
"epoch": 0.5045306270387822,
"grad_norm": 2.3449990407458,
"learning_rate": 1.0340645330352723e-05,
"loss": 1.0373,
"step": 696
},
{
"epoch": 0.5052555273649874,
"grad_norm": 2.0780920184441847,
"learning_rate": 1.0317160721335393e-05,
"loss": 1.0106,
"step": 697
},
{
"epoch": 0.5059804276911924,
"grad_norm": 2.5924431334903186,
"learning_rate": 1.0293674361197173e-05,
"loss": 1.0588,
"step": 698
},
{
"epoch": 0.5067053280173976,
"grad_norm": 2.1365135156985744,
"learning_rate": 1.0270186379611932e-05,
"loss": 1.023,
"step": 699
},
{
"epoch": 0.5074302283436027,
"grad_norm": 2.173447677361539,
"learning_rate": 1.0246696906262484e-05,
"loss": 1.0468,
"step": 700
},
{
"epoch": 0.5081551286698079,
"grad_norm": 2.245470686825682,
"learning_rate": 1.0223206070839878e-05,
"loss": 1.0525,
"step": 701
},
{
"epoch": 0.508880028996013,
"grad_norm": 2.2398665851941706,
"learning_rate": 1.0199714003042685e-05,
"loss": 1.0942,
"step": 702
},
{
"epoch": 0.5096049293222182,
"grad_norm": 2.5997096574935314,
"learning_rate": 1.017622083257628e-05,
"loss": 1.0766,
"step": 703
},
{
"epoch": 0.5103298296484233,
"grad_norm": 2.4146958737582644,
"learning_rate": 1.0152726689152135e-05,
"loss": 1.0928,
"step": 704
},
{
"epoch": 0.5110547299746285,
"grad_norm": 2.130801798872479,
"learning_rate": 1.0129231702487077e-05,
"loss": 0.972,
"step": 705
},
{
"epoch": 0.5117796303008336,
"grad_norm": 2.236352513835616,
"learning_rate": 1.01057360023026e-05,
"loss": 1.0103,
"step": 706
},
{
"epoch": 0.5125045306270388,
"grad_norm": 2.188186582922333,
"learning_rate": 1.0082239718324136e-05,
"loss": 1.0367,
"step": 707
},
{
"epoch": 0.5132294309532439,
"grad_norm": 2.198716617442744,
"learning_rate": 1.0058742980280341e-05,
"loss": 0.972,
"step": 708
},
{
"epoch": 0.5139543312794491,
"grad_norm": 2.294063462573355,
"learning_rate": 1.0035245917902376e-05,
"loss": 1.0778,
"step": 709
},
{
"epoch": 0.5146792316056542,
"grad_norm": 2.228364149473082,
"learning_rate": 1.001174866092319e-05,
"loss": 1.0271,
"step": 710
},
{
"epoch": 0.5154041319318594,
"grad_norm": 1.9300202276554854,
"learning_rate": 9.988251339076811e-06,
"loss": 0.9044,
"step": 711
},
{
"epoch": 0.5161290322580645,
"grad_norm": 2.3634423427900533,
"learning_rate": 9.964754082097626e-06,
"loss": 1.0843,
"step": 712
},
{
"epoch": 0.5168539325842697,
"grad_norm": 2.38754231763043,
"learning_rate": 9.941257019719662e-06,
"loss": 1.0854,
"step": 713
},
{
"epoch": 0.5175788329104748,
"grad_norm": 1.8044209443143941,
"learning_rate": 9.917760281675867e-06,
"loss": 0.9059,
"step": 714
},
{
"epoch": 0.51830373323668,
"grad_norm": 2.506057831976918,
"learning_rate": 9.894263997697405e-06,
"loss": 1.0033,
"step": 715
},
{
"epoch": 0.5190286335628851,
"grad_norm": 2.383694009514269,
"learning_rate": 9.870768297512924e-06,
"loss": 0.9435,
"step": 716
},
{
"epoch": 0.5197535338890903,
"grad_norm": 2.7218871091799195,
"learning_rate": 9.847273310847869e-06,
"loss": 1.013,
"step": 717
},
{
"epoch": 0.5204784342152954,
"grad_norm": 1.9488674447864536,
"learning_rate": 9.823779167423723e-06,
"loss": 1.064,
"step": 718
},
{
"epoch": 0.5212033345415006,
"grad_norm": 2.543494468466047,
"learning_rate": 9.80028599695732e-06,
"loss": 1.0396,
"step": 719
},
{
"epoch": 0.5219282348677057,
"grad_norm": 2.1642282148880976,
"learning_rate": 9.776793929160129e-06,
"loss": 0.9959,
"step": 720
},
{
"epoch": 0.5226531351939109,
"grad_norm": 1.9378193767391163,
"learning_rate": 9.753303093737518e-06,
"loss": 1.0852,
"step": 721
},
{
"epoch": 0.523378035520116,
"grad_norm": 2.5527944661940536,
"learning_rate": 9.729813620388071e-06,
"loss": 1.1522,
"step": 722
},
{
"epoch": 0.5241029358463212,
"grad_norm": 2.3217437760784074,
"learning_rate": 9.70632563880283e-06,
"loss": 1.0067,
"step": 723
},
{
"epoch": 0.5248278361725263,
"grad_norm": 3.1417602276811527,
"learning_rate": 9.682839278664614e-06,
"loss": 0.9952,
"step": 724
},
{
"epoch": 0.5255527364987315,
"grad_norm": 1.9148171819921405,
"learning_rate": 9.659354669647277e-06,
"loss": 0.9656,
"step": 725
},
{
"epoch": 0.5262776368249366,
"grad_norm": 2.449509825571267,
"learning_rate": 9.635871941415026e-06,
"loss": 0.9131,
"step": 726
},
{
"epoch": 0.5270025371511418,
"grad_norm": 2.1963443291334697,
"learning_rate": 9.612391223621667e-06,
"loss": 1.0032,
"step": 727
},
{
"epoch": 0.5277274374773469,
"grad_norm": 2.2338194520764074,
"learning_rate": 9.588912645909905e-06,
"loss": 1.0494,
"step": 728
},
{
"epoch": 0.528452337803552,
"grad_norm": 1.9959846236290153,
"learning_rate": 9.565436337910637e-06,
"loss": 0.947,
"step": 729
},
{
"epoch": 0.5291772381297571,
"grad_norm": 2.078163095936865,
"learning_rate": 9.54196242924222e-06,
"loss": 1.0311,
"step": 730
},
{
"epoch": 0.5299021384559623,
"grad_norm": 2.1980066549775024,
"learning_rate": 9.51849104950978e-06,
"loss": 1.0051,
"step": 731
},
{
"epoch": 0.5306270387821674,
"grad_norm": 2.4191287974234705,
"learning_rate": 9.495022328304461e-06,
"loss": 1.0469,
"step": 732
},
{
"epoch": 0.5313519391083726,
"grad_norm": 2.298052468185075,
"learning_rate": 9.47155639520274e-06,
"loss": 1.0115,
"step": 733
},
{
"epoch": 0.5320768394345777,
"grad_norm": 2.1232889455421406,
"learning_rate": 9.448093379765693e-06,
"loss": 1.0524,
"step": 734
},
{
"epoch": 0.5328017397607829,
"grad_norm": 2.44786208230678,
"learning_rate": 9.424633411538289e-06,
"loss": 1.0772,
"step": 735
},
{
"epoch": 0.533526640086988,
"grad_norm": 2.102354340691057,
"learning_rate": 9.401176620048687e-06,
"loss": 1.0344,
"step": 736
},
{
"epoch": 0.5342515404131932,
"grad_norm": 2.108353702137098,
"learning_rate": 9.377723134807482e-06,
"loss": 1.1669,
"step": 737
},
{
"epoch": 0.5349764407393983,
"grad_norm": 2.4286899167576412,
"learning_rate": 9.35427308530703e-06,
"loss": 1.0635,
"step": 738
},
{
"epoch": 0.5357013410656035,
"grad_norm": 2.2012613191456025,
"learning_rate": 9.330826601020713e-06,
"loss": 1.1047,
"step": 739
},
{
"epoch": 0.5364262413918086,
"grad_norm": 2.7290128431948646,
"learning_rate": 9.307383811402229e-06,
"loss": 1.0265,
"step": 740
},
{
"epoch": 0.5371511417180138,
"grad_norm": 2.3858907048476135,
"learning_rate": 9.283944845884878e-06,
"loss": 0.9387,
"step": 741
},
{
"epoch": 0.5378760420442189,
"grad_norm": 2.080475476085547,
"learning_rate": 9.260509833880848e-06,
"loss": 0.9617,
"step": 742
},
{
"epoch": 0.5386009423704241,
"grad_norm": 1.8615886294130846,
"learning_rate": 9.23707890478049e-06,
"loss": 0.8874,
"step": 743
},
{
"epoch": 0.5393258426966292,
"grad_norm": 1.9904190396103651,
"learning_rate": 9.213652187951624e-06,
"loss": 1.0746,
"step": 744
},
{
"epoch": 0.5400507430228344,
"grad_norm": 2.96232733164202,
"learning_rate": 9.190229812738802e-06,
"loss": 1.1089,
"step": 745
},
{
"epoch": 0.5407756433490395,
"grad_norm": 2.525943985627746,
"learning_rate": 9.166811908462616e-06,
"loss": 1.0642,
"step": 746
},
{
"epoch": 0.5415005436752447,
"grad_norm": 2.2418322634514167,
"learning_rate": 9.143398604418965e-06,
"loss": 1.1124,
"step": 747
},
{
"epoch": 0.5422254440014498,
"grad_norm": 1.9220381898458985,
"learning_rate": 9.11999002987835e-06,
"loss": 0.9473,
"step": 748
},
{
"epoch": 0.542950344327655,
"grad_norm": 2.6621473432063176,
"learning_rate": 9.096586314085162e-06,
"loss": 1.0125,
"step": 749
},
{
"epoch": 0.5436752446538601,
"grad_norm": 2.3340603683637156,
"learning_rate": 9.07318758625696e-06,
"loss": 1.0542,
"step": 750
},
{
"epoch": 0.5444001449800653,
"grad_norm": 2.1363055122068215,
"learning_rate": 9.049793975583775e-06,
"loss": 1.0586,
"step": 751
},
{
"epoch": 0.5451250453062704,
"grad_norm": 2.0326003541767803,
"learning_rate": 9.026405611227371e-06,
"loss": 0.9054,
"step": 752
},
{
"epoch": 0.5458499456324756,
"grad_norm": 1.6432957234320769,
"learning_rate": 9.003022622320555e-06,
"loss": 0.9625,
"step": 753
},
{
"epoch": 0.5465748459586807,
"grad_norm": 1.78897998635208,
"learning_rate": 8.97964513796645e-06,
"loss": 1.0138,
"step": 754
},
{
"epoch": 0.5472997462848859,
"grad_norm": 2.797841405037185,
"learning_rate": 8.95627328723779e-06,
"loss": 1.2344,
"step": 755
},
{
"epoch": 0.548024646611091,
"grad_norm": 2.0182212487321864,
"learning_rate": 8.932907199176206e-06,
"loss": 1.0249,
"step": 756
},
{
"epoch": 0.5487495469372962,
"grad_norm": 2.1185607535847586,
"learning_rate": 8.909547002791506e-06,
"loss": 1.0009,
"step": 757
},
{
"epoch": 0.5494744472635013,
"grad_norm": 2.5153165221241687,
"learning_rate": 8.886192827060974e-06,
"loss": 1.0271,
"step": 758
},
{
"epoch": 0.5501993475897065,
"grad_norm": 2.205784694298984,
"learning_rate": 8.862844800928645e-06,
"loss": 1.0508,
"step": 759
},
{
"epoch": 0.5509242479159115,
"grad_norm": 2.908166150479793,
"learning_rate": 8.839503053304614e-06,
"loss": 1.039,
"step": 760
},
{
"epoch": 0.5516491482421167,
"grad_norm": 1.9201627389676703,
"learning_rate": 8.8161677130643e-06,
"loss": 0.9532,
"step": 761
},
{
"epoch": 0.5523740485683218,
"grad_norm": 2.923787620312773,
"learning_rate": 8.792838909047747e-06,
"loss": 1.0188,
"step": 762
},
{
"epoch": 0.553098948894527,
"grad_norm": 3.157465367532413,
"learning_rate": 8.769516770058915e-06,
"loss": 1.0947,
"step": 763
},
{
"epoch": 0.5538238492207321,
"grad_norm": 2.060375498630697,
"learning_rate": 8.746201424864956e-06,
"loss": 0.9142,
"step": 764
},
{
"epoch": 0.5545487495469373,
"grad_norm": 1.9982518012339532,
"learning_rate": 8.722893002195523e-06,
"loss": 0.9624,
"step": 765
},
{
"epoch": 0.5552736498731424,
"grad_norm": 1.841122163960603,
"learning_rate": 8.699591630742042e-06,
"loss": 0.979,
"step": 766
},
{
"epoch": 0.5559985501993476,
"grad_norm": 2.0151604980838633,
"learning_rate": 8.676297439157007e-06,
"loss": 0.994,
"step": 767
},
{
"epoch": 0.5567234505255527,
"grad_norm": 2.5456721286611037,
"learning_rate": 8.653010556053271e-06,
"loss": 1.0495,
"step": 768
},
{
"epoch": 0.5574483508517579,
"grad_norm": 2.7097490860787183,
"learning_rate": 8.629731110003337e-06,
"loss": 1.124,
"step": 769
},
{
"epoch": 0.558173251177963,
"grad_norm": 1.949724143019753,
"learning_rate": 8.606459229538645e-06,
"loss": 0.997,
"step": 770
},
{
"epoch": 0.5588981515041682,
"grad_norm": 2.020801130677908,
"learning_rate": 8.583195043148864e-06,
"loss": 1.1449,
"step": 771
},
{
"epoch": 0.5596230518303733,
"grad_norm": 2.130589840488668,
"learning_rate": 8.55993867928118e-06,
"loss": 0.9983,
"step": 772
},
{
"epoch": 0.5603479521565785,
"grad_norm": 2.0836041709296316,
"learning_rate": 8.536690266339593e-06,
"loss": 0.9946,
"step": 773
},
{
"epoch": 0.5610728524827836,
"grad_norm": 2.3227053384824274,
"learning_rate": 8.513449932684198e-06,
"loss": 0.8824,
"step": 774
},
{
"epoch": 0.5617977528089888,
"grad_norm": 1.983623134191808,
"learning_rate": 8.490217806630489e-06,
"loss": 0.9447,
"step": 775
},
{
"epoch": 0.5625226531351939,
"grad_norm": 2.1427968901005725,
"learning_rate": 8.46699401644864e-06,
"loss": 1.0275,
"step": 776
},
{
"epoch": 0.5632475534613991,
"grad_norm": 2.263521727209926,
"learning_rate": 8.443778690362801e-06,
"loss": 0.9779,
"step": 777
},
{
"epoch": 0.5639724537876042,
"grad_norm": 1.784185254505844,
"learning_rate": 8.42057195655039e-06,
"loss": 0.9217,
"step": 778
},
{
"epoch": 0.5646973541138094,
"grad_norm": 2.5535009035102934,
"learning_rate": 8.397373943141384e-06,
"loss": 1.0259,
"step": 779
},
{
"epoch": 0.5654222544400145,
"grad_norm": 2.3274203754164873,
"learning_rate": 8.374184778217617e-06,
"loss": 1.0449,
"step": 780
},
{
"epoch": 0.5661471547662197,
"grad_norm": 2.751811026011905,
"learning_rate": 8.351004589812061e-06,
"loss": 1.0106,
"step": 781
},
{
"epoch": 0.5668720550924248,
"grad_norm": 2.0365364413353726,
"learning_rate": 8.327833505908135e-06,
"loss": 1.0282,
"step": 782
},
{
"epoch": 0.56759695541863,
"grad_norm": 2.3504702430826154,
"learning_rate": 8.30467165443898e-06,
"loss": 1.0629,
"step": 783
},
{
"epoch": 0.5683218557448351,
"grad_norm": 2.0575133385244646,
"learning_rate": 8.281519163286772e-06,
"loss": 1.0005,
"step": 784
},
{
"epoch": 0.5690467560710403,
"grad_norm": 2.5919383011426467,
"learning_rate": 8.258376160282001e-06,
"loss": 1.0555,
"step": 785
},
{
"epoch": 0.5697716563972454,
"grad_norm": 2.4060395092358102,
"learning_rate": 8.235242773202772e-06,
"loss": 1.0631,
"step": 786
},
{
"epoch": 0.5704965567234506,
"grad_norm": 1.9397501796747694,
"learning_rate": 8.212119129774098e-06,
"loss": 1.1304,
"step": 787
},
{
"epoch": 0.5712214570496557,
"grad_norm": 2.110530023102397,
"learning_rate": 8.189005357667191e-06,
"loss": 0.9567,
"step": 788
},
{
"epoch": 0.5719463573758609,
"grad_norm": 2.1870319191564658,
"learning_rate": 8.165901584498774e-06,
"loss": 1.0683,
"step": 789
},
{
"epoch": 0.572671257702066,
"grad_norm": 1.8495796212471047,
"learning_rate": 8.142807937830348e-06,
"loss": 0.9837,
"step": 790
},
{
"epoch": 0.5733961580282712,
"grad_norm": 2.64388618965451,
"learning_rate": 8.11972454516751e-06,
"loss": 0.9659,
"step": 791
},
{
"epoch": 0.5741210583544762,
"grad_norm": 2.0330246862165415,
"learning_rate": 8.096651533959244e-06,
"loss": 0.9591,
"step": 792
},
{
"epoch": 0.5748459586806814,
"grad_norm": 2.2631207779996827,
"learning_rate": 8.073589031597209e-06,
"loss": 0.9873,
"step": 793
},
{
"epoch": 0.5755708590068865,
"grad_norm": 2.3948632401763064,
"learning_rate": 8.050537165415053e-06,
"loss": 1.1371,
"step": 794
},
{
"epoch": 0.5762957593330917,
"grad_norm": 2.075607270336768,
"learning_rate": 8.027496062687688e-06,
"loss": 0.8908,
"step": 795
},
{
"epoch": 0.5770206596592968,
"grad_norm": 2.065746544851814,
"learning_rate": 8.004465850630605e-06,
"loss": 1.0025,
"step": 796
},
{
"epoch": 0.577745559985502,
"grad_norm": 2.2202432240869845,
"learning_rate": 7.981446656399162e-06,
"loss": 1.0429,
"step": 797
},
{
"epoch": 0.5784704603117071,
"grad_norm": 2.309075300778911,
"learning_rate": 7.958438607087884e-06,
"loss": 0.9851,
"step": 798
},
{
"epoch": 0.5791953606379123,
"grad_norm": 2.4033759237105166,
"learning_rate": 7.935441829729766e-06,
"loss": 1.032,
"step": 799
},
{
"epoch": 0.5799202609641174,
"grad_norm": 2.625985750403029,
"learning_rate": 7.912456451295567e-06,
"loss": 1.0102,
"step": 800
},
{
"epoch": 0.5806451612903226,
"grad_norm": 2.410099594821792,
"learning_rate": 7.889482598693108e-06,
"loss": 0.9639,
"step": 801
},
{
"epoch": 0.5813700616165277,
"grad_norm": 1.9619910669766136,
"learning_rate": 7.866520398766574e-06,
"loss": 0.9297,
"step": 802
},
{
"epoch": 0.5820949619427329,
"grad_norm": 2.930559344753908,
"learning_rate": 7.843569978295804e-06,
"loss": 1.0681,
"step": 803
},
{
"epoch": 0.582819862268938,
"grad_norm": 1.8579566089499246,
"learning_rate": 7.820631463995622e-06,
"loss": 0.9247,
"step": 804
},
{
"epoch": 0.5835447625951432,
"grad_norm": 1.783098743265068,
"learning_rate": 7.797704982515094e-06,
"loss": 0.9649,
"step": 805
},
{
"epoch": 0.5842696629213483,
"grad_norm": 1.9534992322071332,
"learning_rate": 7.774790660436857e-06,
"loss": 0.9234,
"step": 806
},
{
"epoch": 0.5849945632475535,
"grad_norm": 2.3775300305715428,
"learning_rate": 7.751888624276418e-06,
"loss": 1.0246,
"step": 807
},
{
"epoch": 0.5857194635737586,
"grad_norm": 2.0533804995629525,
"learning_rate": 7.728999000481436e-06,
"loss": 0.9705,
"step": 808
},
{
"epoch": 0.5864443638999638,
"grad_norm": 1.959728561512236,
"learning_rate": 7.706121915431059e-06,
"loss": 0.9359,
"step": 809
},
{
"epoch": 0.5871692642261689,
"grad_norm": 2.502678481308203,
"learning_rate": 7.68325749543519e-06,
"loss": 0.9802,
"step": 810
},
{
"epoch": 0.5878941645523741,
"grad_norm": 3.0385891335086352,
"learning_rate": 7.660405866733813e-06,
"loss": 1.1222,
"step": 811
},
{
"epoch": 0.5886190648785792,
"grad_norm": 2.040606391856475,
"learning_rate": 7.637567155496277e-06,
"loss": 1.02,
"step": 812
},
{
"epoch": 0.5893439652047844,
"grad_norm": 2.305055241806458,
"learning_rate": 7.614741487820626e-06,
"loss": 0.9964,
"step": 813
},
{
"epoch": 0.5900688655309895,
"grad_norm": 1.882568911490993,
"learning_rate": 7.5919289897328784e-06,
"loss": 0.9941,
"step": 814
},
{
"epoch": 0.5907937658571947,
"grad_norm": 2.5290386344366156,
"learning_rate": 7.569129787186342e-06,
"loss": 1.0734,
"step": 815
},
{
"epoch": 0.5915186661833998,
"grad_norm": 2.8819681752053454,
"learning_rate": 7.546344006060918e-06,
"loss": 1.0041,
"step": 816
},
{
"epoch": 0.592243566509605,
"grad_norm": 1.8804822924012536,
"learning_rate": 7.523571772162392e-06,
"loss": 0.9151,
"step": 817
},
{
"epoch": 0.5929684668358101,
"grad_norm": 2.255992660838785,
"learning_rate": 7.500813211221778e-06,
"loss": 1.0903,
"step": 818
},
{
"epoch": 0.5936933671620153,
"grad_norm": 2.9086758708738345,
"learning_rate": 7.478068448894577e-06,
"loss": 1.0814,
"step": 819
},
{
"epoch": 0.5944182674882204,
"grad_norm": 2.437481894199292,
"learning_rate": 7.455337610760114e-06,
"loss": 1.1113,
"step": 820
},
{
"epoch": 0.5951431678144256,
"grad_norm": 1.8660888898062387,
"learning_rate": 7.4326208223208274e-06,
"loss": 0.9695,
"step": 821
},
{
"epoch": 0.5958680681406306,
"grad_norm": 2.1229612134239995,
"learning_rate": 7.409918209001585e-06,
"loss": 1.0515,
"step": 822
},
{
"epoch": 0.5965929684668358,
"grad_norm": 2.3583730518901995,
"learning_rate": 7.387229896149006e-06,
"loss": 0.9714,
"step": 823
},
{
"epoch": 0.5973178687930409,
"grad_norm": 2.3153711669204413,
"learning_rate": 7.364556009030734e-06,
"loss": 1.059,
"step": 824
},
{
"epoch": 0.5980427691192461,
"grad_norm": 1.9967290233140527,
"learning_rate": 7.341896672834776e-06,
"loss": 1.047,
"step": 825
},
{
"epoch": 0.5987676694454512,
"grad_norm": 1.9339818611902988,
"learning_rate": 7.31925201266879e-06,
"loss": 0.972,
"step": 826
},
{
"epoch": 0.5994925697716564,
"grad_norm": 1.9457458272271317,
"learning_rate": 7.2966221535594085e-06,
"loss": 1.0206,
"step": 827
},
{
"epoch": 0.6002174700978615,
"grad_norm": 2.1792046814563495,
"learning_rate": 7.274007220451553e-06,
"loss": 1.0732,
"step": 828
},
{
"epoch": 0.6009423704240667,
"grad_norm": 2.029220407091385,
"learning_rate": 7.251407338207725e-06,
"loss": 0.8824,
"step": 829
},
{
"epoch": 0.6016672707502718,
"grad_norm": 1.915241829499402,
"learning_rate": 7.228822631607327e-06,
"loss": 0.9666,
"step": 830
},
{
"epoch": 0.602392171076477,
"grad_norm": 1.8518746441416718,
"learning_rate": 7.2062532253459714e-06,
"loss": 0.8742,
"step": 831
},
{
"epoch": 0.6031170714026821,
"grad_norm": 2.5929079919190547,
"learning_rate": 7.183699244034797e-06,
"loss": 1.1043,
"step": 832
},
{
"epoch": 0.6038419717288873,
"grad_norm": 2.2618493759552516,
"learning_rate": 7.161160812199785e-06,
"loss": 1.1689,
"step": 833
},
{
"epoch": 0.6045668720550924,
"grad_norm": 1.9916609519278592,
"learning_rate": 7.138638054281055e-06,
"loss": 1.0807,
"step": 834
},
{
"epoch": 0.6052917723812976,
"grad_norm": 1.8106802403260034,
"learning_rate": 7.116131094632186e-06,
"loss": 0.8694,
"step": 835
},
{
"epoch": 0.6060166727075027,
"grad_norm": 2.2006653848900015,
"learning_rate": 7.093640057519531e-06,
"loss": 1.0905,
"step": 836
},
{
"epoch": 0.6067415730337079,
"grad_norm": 2.3615298366429496,
"learning_rate": 7.0711650671215405e-06,
"loss": 1.0178,
"step": 837
},
{
"epoch": 0.607466473359913,
"grad_norm": 1.9549782714474917,
"learning_rate": 7.048706247528061e-06,
"loss": 1.0098,
"step": 838
},
{
"epoch": 0.6081913736861182,
"grad_norm": 2.2334786388231396,
"learning_rate": 7.02626372273965e-06,
"loss": 1.0028,
"step": 839
},
{
"epoch": 0.6089162740123233,
"grad_norm": 2.515375317083732,
"learning_rate": 7.003837616666906e-06,
"loss": 0.8934,
"step": 840
},
{
"epoch": 0.6096411743385285,
"grad_norm": 2.0714175972616347,
"learning_rate": 6.98142805312977e-06,
"loss": 1.008,
"step": 841
},
{
"epoch": 0.6103660746647336,
"grad_norm": 2.183667905495028,
"learning_rate": 6.959035155856857e-06,
"loss": 1.0119,
"step": 842
},
{
"epoch": 0.6110909749909388,
"grad_norm": 2.178274745085645,
"learning_rate": 6.936659048484755e-06,
"loss": 0.9791,
"step": 843
},
{
"epoch": 0.6118158753171439,
"grad_norm": 2.2299367836336903,
"learning_rate": 6.914299854557349e-06,
"loss": 1.0342,
"step": 844
},
{
"epoch": 0.6125407756433491,
"grad_norm": 1.870612790239624,
"learning_rate": 6.891957697525149e-06,
"loss": 1.0123,
"step": 845
},
{
"epoch": 0.6132656759695542,
"grad_norm": 2.2457853373416996,
"learning_rate": 6.869632700744588e-06,
"loss": 0.9599,
"step": 846
},
{
"epoch": 0.6139905762957594,
"grad_norm": 1.8964211823089199,
"learning_rate": 6.847324987477375e-06,
"loss": 1.0702,
"step": 847
},
{
"epoch": 0.6147154766219645,
"grad_norm": 2.2231409481558595,
"learning_rate": 6.82503468088977e-06,
"loss": 0.9252,
"step": 848
},
{
"epoch": 0.6154403769481697,
"grad_norm": 2.2158057236580575,
"learning_rate": 6.802761904051937e-06,
"loss": 0.8728,
"step": 849
},
{
"epoch": 0.6161652772743748,
"grad_norm": 1.984090771986885,
"learning_rate": 6.78050677993725e-06,
"loss": 0.9488,
"step": 850
},
{
"epoch": 0.61689017760058,
"grad_norm": 2.194669700982665,
"learning_rate": 6.758269431421618e-06,
"loss": 1.0144,
"step": 851
},
{
"epoch": 0.617615077926785,
"grad_norm": 2.391278122000099,
"learning_rate": 6.7360499812828195e-06,
"loss": 0.9911,
"step": 852
},
{
"epoch": 0.6183399782529903,
"grad_norm": 2.022013346401567,
"learning_rate": 6.713848552199791e-06,
"loss": 0.9898,
"step": 853
},
{
"epoch": 0.6190648785791953,
"grad_norm": 2.109272857624347,
"learning_rate": 6.6916652667519855e-06,
"loss": 0.9562,
"step": 854
},
{
"epoch": 0.6197897789054005,
"grad_norm": 2.829806104335344,
"learning_rate": 6.669500247418674e-06,
"loss": 1.0893,
"step": 855
},
{
"epoch": 0.6205146792316056,
"grad_norm": 2.2159865585012715,
"learning_rate": 6.647353616578274e-06,
"loss": 1.0566,
"step": 856
},
{
"epoch": 0.6212395795578108,
"grad_norm": 2.5699500401567272,
"learning_rate": 6.625225496507688e-06,
"loss": 0.9406,
"step": 857
},
{
"epoch": 0.6219644798840159,
"grad_norm": 1.9318589400149984,
"learning_rate": 6.6031160093816005e-06,
"loss": 0.9556,
"step": 858
},
{
"epoch": 0.6226893802102211,
"grad_norm": 2.046984039555971,
"learning_rate": 6.5810252772718276e-06,
"loss": 1.0834,
"step": 859
},
{
"epoch": 0.6234142805364262,
"grad_norm": 2.6992216221025087,
"learning_rate": 6.5589534221466324e-06,
"loss": 0.9793,
"step": 860
},
{
"epoch": 0.6241391808626314,
"grad_norm": 2.212396980492137,
"learning_rate": 6.536900565870052e-06,
"loss": 0.8663,
"step": 861
},
{
"epoch": 0.6248640811888365,
"grad_norm": 1.7999266895906552,
"learning_rate": 6.514866830201234e-06,
"loss": 0.8443,
"step": 862
},
{
"epoch": 0.6255889815150417,
"grad_norm": 2.1176385115675083,
"learning_rate": 6.4928523367937444e-06,
"loss": 0.9933,
"step": 863
},
{
"epoch": 0.6263138818412468,
"grad_norm": 2.2762631364540784,
"learning_rate": 6.470857207194918e-06,
"loss": 1.0051,
"step": 864
},
{
"epoch": 0.627038782167452,
"grad_norm": 2.0490123793497976,
"learning_rate": 6.448881562845169e-06,
"loss": 0.9627,
"step": 865
},
{
"epoch": 0.6277636824936571,
"grad_norm": 2.5600372472836117,
"learning_rate": 6.426925525077341e-06,
"loss": 1.0176,
"step": 866
},
{
"epoch": 0.6284885828198623,
"grad_norm": 2.8510918650806882,
"learning_rate": 6.404989215116012e-06,
"loss": 1.0215,
"step": 867
},
{
"epoch": 0.6292134831460674,
"grad_norm": 1.9769591446882206,
"learning_rate": 6.3830727540768445e-06,
"loss": 1.0687,
"step": 868
},
{
"epoch": 0.6299383834722726,
"grad_norm": 1.677393229029825,
"learning_rate": 6.361176262965909e-06,
"loss": 0.9384,
"step": 869
},
{
"epoch": 0.6306632837984777,
"grad_norm": 2.4132581003126705,
"learning_rate": 6.339299862679016e-06,
"loss": 0.973,
"step": 870
},
{
"epoch": 0.6313881841246829,
"grad_norm": 2.562250697801829,
"learning_rate": 6.317443674001055e-06,
"loss": 0.9105,
"step": 871
},
{
"epoch": 0.632113084450888,
"grad_norm": 2.0225187463112033,
"learning_rate": 6.295607817605316e-06,
"loss": 0.9315,
"step": 872
},
{
"epoch": 0.6328379847770932,
"grad_norm": 2.6962240997771167,
"learning_rate": 6.273792414052833e-06,
"loss": 1.058,
"step": 873
},
{
"epoch": 0.6335628851032983,
"grad_norm": 2.3293379147129705,
"learning_rate": 6.251997583791711e-06,
"loss": 1.0571,
"step": 874
},
{
"epoch": 0.6342877854295035,
"grad_norm": 2.17475653423949,
"learning_rate": 6.230223447156469e-06,
"loss": 1.0355,
"step": 875
},
{
"epoch": 0.6350126857557086,
"grad_norm": 2.293124315791374,
"learning_rate": 6.208470124367374e-06,
"loss": 1.0007,
"step": 876
},
{
"epoch": 0.6357375860819138,
"grad_norm": 2.1218951006607236,
"learning_rate": 6.186737735529769e-06,
"loss": 0.9075,
"step": 877
},
{
"epoch": 0.6364624864081189,
"grad_norm": 2.1947275964384114,
"learning_rate": 6.165026400633416e-06,
"loss": 1.0155,
"step": 878
},
{
"epoch": 0.6371873867343241,
"grad_norm": 1.8448468197236452,
"learning_rate": 6.143336239551836e-06,
"loss": 0.9704,
"step": 879
},
{
"epoch": 0.6379122870605292,
"grad_norm": 2.2203852100211496,
"learning_rate": 6.12166737204164e-06,
"loss": 1.0183,
"step": 880
},
{
"epoch": 0.6386371873867344,
"grad_norm": 1.9940936134783194,
"learning_rate": 6.10001991774188e-06,
"loss": 1.0362,
"step": 881
},
{
"epoch": 0.6393620877129395,
"grad_norm": 1.876178541905188,
"learning_rate": 6.078393996173375e-06,
"loss": 1.0958,
"step": 882
},
{
"epoch": 0.6400869880391447,
"grad_norm": 2.4300065422050174,
"learning_rate": 6.056789726738055e-06,
"loss": 1.0112,
"step": 883
},
{
"epoch": 0.6408118883653497,
"grad_norm": 2.182070128778044,
"learning_rate": 6.035207228718305e-06,
"loss": 1.0532,
"step": 884
},
{
"epoch": 0.641536788691555,
"grad_norm": 2.2790151746934826,
"learning_rate": 6.0136466212763055e-06,
"loss": 0.9994,
"step": 885
},
{
"epoch": 0.64226168901776,
"grad_norm": 2.0380124474521333,
"learning_rate": 5.992108023453376e-06,
"loss": 0.9901,
"step": 886
},
{
"epoch": 0.6429865893439652,
"grad_norm": 2.3256602048348234,
"learning_rate": 5.9705915541693114e-06,
"loss": 1.1098,
"step": 887
},
{
"epoch": 0.6437114896701703,
"grad_norm": 2.753858071131981,
"learning_rate": 5.949097332221734e-06,
"loss": 1.0406,
"step": 888
},
{
"epoch": 0.6444363899963755,
"grad_norm": 2.068578563869525,
"learning_rate": 5.927625476285426e-06,
"loss": 0.9272,
"step": 889
},
{
"epoch": 0.6451612903225806,
"grad_norm": 2.0943155971716947,
"learning_rate": 5.906176104911687e-06,
"loss": 1.0903,
"step": 890
},
{
"epoch": 0.6458861906487858,
"grad_norm": 2.244523642610638,
"learning_rate": 5.884749336527676e-06,
"loss": 1.0465,
"step": 891
},
{
"epoch": 0.6466110909749909,
"grad_norm": 2.2948461929009554,
"learning_rate": 5.8633452894357516e-06,
"loss": 0.9543,
"step": 892
},
{
"epoch": 0.6473359913011961,
"grad_norm": 2.316443702020551,
"learning_rate": 5.841964081812822e-06,
"loss": 1.0483,
"step": 893
},
{
"epoch": 0.6480608916274012,
"grad_norm": 2.210302183543952,
"learning_rate": 5.820605831709694e-06,
"loss": 0.8937,
"step": 894
},
{
"epoch": 0.6487857919536064,
"grad_norm": 2.1138527969671097,
"learning_rate": 5.799270657050422e-06,
"loss": 0.8968,
"step": 895
},
{
"epoch": 0.6495106922798115,
"grad_norm": 2.136178605345999,
"learning_rate": 5.777958675631657e-06,
"loss": 1.008,
"step": 896
},
{
"epoch": 0.6502355926060167,
"grad_norm": 1.9381106844139175,
"learning_rate": 5.756670005121987e-06,
"loss": 0.9595,
"step": 897
},
{
"epoch": 0.6509604929322218,
"grad_norm": 2.0173878555394604,
"learning_rate": 5.735404763061303e-06,
"loss": 0.9614,
"step": 898
},
{
"epoch": 0.651685393258427,
"grad_norm": 2.159725328780946,
"learning_rate": 5.714163066860132e-06,
"loss": 0.9797,
"step": 899
},
{
"epoch": 0.6524102935846321,
"grad_norm": 2.0644994410010713,
"learning_rate": 5.6929450337990175e-06,
"loss": 0.9167,
"step": 900
},
{
"epoch": 0.6531351939108373,
"grad_norm": 2.019906324854319,
"learning_rate": 5.671750781027836e-06,
"loss": 0.9609,
"step": 901
},
{
"epoch": 0.6538600942370424,
"grad_norm": 2.774207911473111,
"learning_rate": 5.650580425565166e-06,
"loss": 0.9905,
"step": 902
},
{
"epoch": 0.6545849945632476,
"grad_norm": 1.9230557138061348,
"learning_rate": 5.629434084297654e-06,
"loss": 1.053,
"step": 903
},
{
"epoch": 0.6553098948894527,
"grad_norm": 1.6916309404007928,
"learning_rate": 5.60831187397935e-06,
"loss": 0.9022,
"step": 904
},
{
"epoch": 0.6560347952156579,
"grad_norm": 2.292413850175932,
"learning_rate": 5.58721391123108e-06,
"loss": 0.9563,
"step": 905
},
{
"epoch": 0.656759695541863,
"grad_norm": 2.187934211851262,
"learning_rate": 5.566140312539786e-06,
"loss": 1.0045,
"step": 906
},
{
"epoch": 0.6574845958680682,
"grad_norm": 2.2765131164518726,
"learning_rate": 5.545091194257892e-06,
"loss": 1.121,
"step": 907
},
{
"epoch": 0.6582094961942733,
"grad_norm": 4.530518710071565,
"learning_rate": 5.524066672602662e-06,
"loss": 1.1574,
"step": 908
},
{
"epoch": 0.6589343965204785,
"grad_norm": 2.964421873486366,
"learning_rate": 5.503066863655546e-06,
"loss": 0.9753,
"step": 909
},
{
"epoch": 0.6596592968466836,
"grad_norm": 1.7482124392759353,
"learning_rate": 5.482091883361571e-06,
"loss": 0.9915,
"step": 910
},
{
"epoch": 0.6603841971728888,
"grad_norm": 2.224885148056625,
"learning_rate": 5.461141847528655e-06,
"loss": 0.9071,
"step": 911
},
{
"epoch": 0.6611090974990939,
"grad_norm": 3.186202824223461,
"learning_rate": 5.440216871827004e-06,
"loss": 1.0746,
"step": 912
},
{
"epoch": 0.6618339978252991,
"grad_norm": 1.7683023401240061,
"learning_rate": 5.419317071788461e-06,
"loss": 0.9636,
"step": 913
},
{
"epoch": 0.6625588981515041,
"grad_norm": 3.098972560392732,
"learning_rate": 5.398442562805859e-06,
"loss": 1.1527,
"step": 914
},
{
"epoch": 0.6632837984777094,
"grad_norm": 2.2445548524254946,
"learning_rate": 5.3775934601324094e-06,
"loss": 1.0563,
"step": 915
},
{
"epoch": 0.6640086988039144,
"grad_norm": 1.739464246848828,
"learning_rate": 5.3567698788810366e-06,
"loss": 1.0422,
"step": 916
},
{
"epoch": 0.6647335991301196,
"grad_norm": 2.0974710880958485,
"learning_rate": 5.335971934023757e-06,
"loss": 0.899,
"step": 917
},
{
"epoch": 0.6654584994563247,
"grad_norm": 2.029612124614373,
"learning_rate": 5.315199740391044e-06,
"loss": 0.9541,
"step": 918
},
{
"epoch": 0.6661833997825299,
"grad_norm": 2.2132638391788158,
"learning_rate": 5.294453412671192e-06,
"loss": 1.0131,
"step": 919
},
{
"epoch": 0.666908300108735,
"grad_norm": 2.5732318578474107,
"learning_rate": 5.273733065409683e-06,
"loss": 0.9976,
"step": 920
},
{
"epoch": 0.6676332004349402,
"grad_norm": 2.265338233038655,
"learning_rate": 5.2530388130085555e-06,
"loss": 0.9039,
"step": 921
},
{
"epoch": 0.6683581007611453,
"grad_norm": 2.006264114627049,
"learning_rate": 5.232370769725769e-06,
"loss": 0.9893,
"step": 922
},
{
"epoch": 0.6690830010873505,
"grad_norm": 2.0904960315209076,
"learning_rate": 5.2117290496745764e-06,
"loss": 0.9729,
"step": 923
},
{
"epoch": 0.6698079014135556,
"grad_norm": 1.943083190146926,
"learning_rate": 5.191113766822905e-06,
"loss": 0.9807,
"step": 924
},
{
"epoch": 0.6705328017397608,
"grad_norm": 2.661838718236442,
"learning_rate": 5.1705250349927004e-06,
"loss": 0.9773,
"step": 925
},
{
"epoch": 0.6712577020659659,
"grad_norm": 2.462190516757415,
"learning_rate": 5.149962967859321e-06,
"loss": 1.0769,
"step": 926
},
{
"epoch": 0.6719826023921711,
"grad_norm": 2.142823487160251,
"learning_rate": 5.129427678950902e-06,
"loss": 0.938,
"step": 927
},
{
"epoch": 0.6727075027183762,
"grad_norm": 1.972351101050343,
"learning_rate": 5.108919281647728e-06,
"loss": 0.9529,
"step": 928
},
{
"epoch": 0.6734324030445814,
"grad_norm": 1.9370633685514163,
"learning_rate": 5.088437889181612e-06,
"loss": 1.075,
"step": 929
},
{
"epoch": 0.6741573033707865,
"grad_norm": 1.975590280335086,
"learning_rate": 5.067983614635263e-06,
"loss": 0.9653,
"step": 930
},
{
"epoch": 0.6748822036969917,
"grad_norm": 2.187164922126263,
"learning_rate": 5.04755657094167e-06,
"loss": 1.1111,
"step": 931
},
{
"epoch": 0.6756071040231968,
"grad_norm": 1.9292793498955334,
"learning_rate": 5.027156870883468e-06,
"loss": 0.9759,
"step": 932
},
{
"epoch": 0.676332004349402,
"grad_norm": 1.9011667121841174,
"learning_rate": 5.006784627092322e-06,
"loss": 0.9786,
"step": 933
},
{
"epoch": 0.6770569046756071,
"grad_norm": 1.8408728752655816,
"learning_rate": 4.986439952048313e-06,
"loss": 0.8755,
"step": 934
},
{
"epoch": 0.6777818050018123,
"grad_norm": 2.2274071731605747,
"learning_rate": 4.9661229580792995e-06,
"loss": 0.9514,
"step": 935
},
{
"epoch": 0.6785067053280174,
"grad_norm": 2.11818671076986,
"learning_rate": 4.945833757360308e-06,
"loss": 0.9173,
"step": 936
},
{
"epoch": 0.6792316056542226,
"grad_norm": 2.244914641056037,
"learning_rate": 4.925572461912912e-06,
"loss": 0.9013,
"step": 937
},
{
"epoch": 0.6799565059804277,
"grad_norm": 2.175683562082561,
"learning_rate": 4.905339183604614e-06,
"loss": 1.0385,
"step": 938
},
{
"epoch": 0.6806814063066329,
"grad_norm": 1.8770079340690111,
"learning_rate": 4.885134034148225e-06,
"loss": 0.9151,
"step": 939
},
{
"epoch": 0.681406306632838,
"grad_norm": 2.071578077945829,
"learning_rate": 4.864957125101254e-06,
"loss": 1.0396,
"step": 940
},
{
"epoch": 0.6821312069590432,
"grad_norm": 2.6472781206388203,
"learning_rate": 4.844808567865283e-06,
"loss": 1.0126,
"step": 941
},
{
"epoch": 0.6828561072852483,
"grad_norm": 1.9436626348748187,
"learning_rate": 4.824688473685362e-06,
"loss": 1.0218,
"step": 942
},
{
"epoch": 0.6835810076114535,
"grad_norm": 2.986623329743054,
"learning_rate": 4.804596953649381e-06,
"loss": 1.0179,
"step": 943
},
{
"epoch": 0.6843059079376586,
"grad_norm": 2.0261869595811843,
"learning_rate": 4.784534118687477e-06,
"loss": 0.9552,
"step": 944
},
{
"epoch": 0.6850308082638638,
"grad_norm": 2.084576288901219,
"learning_rate": 4.764500079571403e-06,
"loss": 0.9645,
"step": 945
},
{
"epoch": 0.6857557085900688,
"grad_norm": 1.941612444910497,
"learning_rate": 4.744494946913923e-06,
"loss": 0.9273,
"step": 946
},
{
"epoch": 0.686480608916274,
"grad_norm": 2.3028054945708005,
"learning_rate": 4.724518831168203e-06,
"loss": 0.8653,
"step": 947
},
{
"epoch": 0.6872055092424791,
"grad_norm": 2.587974290597117,
"learning_rate": 4.704571842627196e-06,
"loss": 0.9328,
"step": 948
},
{
"epoch": 0.6879304095686843,
"grad_norm": 2.745061826423869,
"learning_rate": 4.684654091423043e-06,
"loss": 1.011,
"step": 949
},
{
"epoch": 0.6886553098948894,
"grad_norm": 2.084724138668754,
"learning_rate": 4.664765687526455e-06,
"loss": 1.031,
"step": 950
},
{
"epoch": 0.6893802102210946,
"grad_norm": 2.409381235205495,
"learning_rate": 4.644906740746108e-06,
"loss": 0.9139,
"step": 951
},
{
"epoch": 0.6901051105472997,
"grad_norm": 2.3347733533309905,
"learning_rate": 4.6250773607280375e-06,
"loss": 0.9462,
"step": 952
},
{
"epoch": 0.6908300108735049,
"grad_norm": 2.6210504366809095,
"learning_rate": 4.605277656955043e-06,
"loss": 1.0772,
"step": 953
},
{
"epoch": 0.69155491119971,
"grad_norm": 1.971545084083224,
"learning_rate": 4.585507738746064e-06,
"loss": 0.9594,
"step": 954
},
{
"epoch": 0.6922798115259152,
"grad_norm": 2.369021752491031,
"learning_rate": 4.5657677152555925e-06,
"loss": 1.0682,
"step": 955
},
{
"epoch": 0.6930047118521203,
"grad_norm": 2.714912007155751,
"learning_rate": 4.5460576954730585e-06,
"loss": 1.0267,
"step": 956
},
{
"epoch": 0.6937296121783255,
"grad_norm": 2.0355578348064727,
"learning_rate": 4.526377788222239e-06,
"loss": 1.0435,
"step": 957
},
{
"epoch": 0.6944545125045306,
"grad_norm": 2.1978211554878953,
"learning_rate": 4.506728102160653e-06,
"loss": 1.008,
"step": 958
},
{
"epoch": 0.6951794128307358,
"grad_norm": 2.007198364190231,
"learning_rate": 4.487108745778958e-06,
"loss": 1.0288,
"step": 959
},
{
"epoch": 0.6959043131569409,
"grad_norm": 2.019964605246266,
"learning_rate": 4.467519827400357e-06,
"loss": 1.1013,
"step": 960
},
{
"epoch": 0.6966292134831461,
"grad_norm": 2.0193639156586136,
"learning_rate": 4.447961455179995e-06,
"loss": 0.9918,
"step": 961
},
{
"epoch": 0.6973541138093512,
"grad_norm": 1.9471456399675218,
"learning_rate": 4.428433737104362e-06,
"loss": 0.9593,
"step": 962
},
{
"epoch": 0.6980790141355564,
"grad_norm": 2.1469665342189206,
"learning_rate": 4.408936780990708e-06,
"loss": 0.8953,
"step": 963
},
{
"epoch": 0.6988039144617615,
"grad_norm": 2.5540392388335453,
"learning_rate": 4.3894706944864305e-06,
"loss": 1.0014,
"step": 964
},
{
"epoch": 0.6995288147879667,
"grad_norm": 2.8104373863238177,
"learning_rate": 4.370035585068493e-06,
"loss": 1.1114,
"step": 965
},
{
"epoch": 0.7002537151141718,
"grad_norm": 2.30574506834477,
"learning_rate": 4.350631560042821e-06,
"loss": 0.9957,
"step": 966
},
{
"epoch": 0.700978615440377,
"grad_norm": 2.4431897548258523,
"learning_rate": 4.331258726543723e-06,
"loss": 1.007,
"step": 967
},
{
"epoch": 0.7017035157665821,
"grad_norm": 2.2915344664162918,
"learning_rate": 4.311917191533288e-06,
"loss": 1.022,
"step": 968
},
{
"epoch": 0.7024284160927873,
"grad_norm": 1.6307353345438238,
"learning_rate": 4.292607061800796e-06,
"loss": 0.9134,
"step": 969
},
{
"epoch": 0.7031533164189924,
"grad_norm": 2.094209654163443,
"learning_rate": 4.273328443962137e-06,
"loss": 1.0199,
"step": 970
},
{
"epoch": 0.7038782167451976,
"grad_norm": 2.6089783909619397,
"learning_rate": 4.254081444459209e-06,
"loss": 1.034,
"step": 971
},
{
"epoch": 0.7046031170714027,
"grad_norm": 2.756310969599043,
"learning_rate": 4.234866169559342e-06,
"loss": 0.9988,
"step": 972
},
{
"epoch": 0.7053280173976079,
"grad_norm": 2.1627058695633554,
"learning_rate": 4.2156827253547095e-06,
"loss": 0.8527,
"step": 973
},
{
"epoch": 0.706052917723813,
"grad_norm": 2.128748230219529,
"learning_rate": 4.1965312177617355e-06,
"loss": 0.9118,
"step": 974
},
{
"epoch": 0.7067778180500182,
"grad_norm": 2.3974561539624117,
"learning_rate": 4.177411752520514e-06,
"loss": 1.0645,
"step": 975
},
{
"epoch": 0.7075027183762232,
"grad_norm": 2.3317825972998016,
"learning_rate": 4.1583244351942275e-06,
"loss": 1.0144,
"step": 976
},
{
"epoch": 0.7082276187024285,
"grad_norm": 2.2782371205197585,
"learning_rate": 4.139269371168562e-06,
"loss": 0.9467,
"step": 977
},
{
"epoch": 0.7089525190286335,
"grad_norm": 1.9702225529384383,
"learning_rate": 4.1202466656511245e-06,
"loss": 0.9395,
"step": 978
},
{
"epoch": 0.7096774193548387,
"grad_norm": 2.056174409141514,
"learning_rate": 4.101256423670862e-06,
"loss": 0.942,
"step": 979
},
{
"epoch": 0.7104023196810438,
"grad_norm": 2.7573170909057474,
"learning_rate": 4.082298750077485e-06,
"loss": 0.8837,
"step": 980
},
{
"epoch": 0.711127220007249,
"grad_norm": 2.79548321777382,
"learning_rate": 4.06337374954088e-06,
"loss": 0.9387,
"step": 981
},
{
"epoch": 0.7118521203334541,
"grad_norm": 2.7061864575287875,
"learning_rate": 4.04448152655055e-06,
"loss": 1.114,
"step": 982
},
{
"epoch": 0.7125770206596593,
"grad_norm": 2.2826349591175488,
"learning_rate": 4.025622185415014e-06,
"loss": 0.9945,
"step": 983
},
{
"epoch": 0.7133019209858644,
"grad_norm": 2.2626152003002176,
"learning_rate": 4.006795830261247e-06,
"loss": 0.9539,
"step": 984
},
{
"epoch": 0.7140268213120696,
"grad_norm": 1.9616752319508053,
"learning_rate": 3.988002565034096e-06,
"loss": 0.9027,
"step": 985
},
{
"epoch": 0.7147517216382747,
"grad_norm": 2.2013038556275704,
"learning_rate": 3.9692424934957175e-06,
"loss": 0.9855,
"step": 986
},
{
"epoch": 0.7154766219644799,
"grad_norm": 2.2452819506033697,
"learning_rate": 3.950515719224991e-06,
"loss": 0.9616,
"step": 987
},
{
"epoch": 0.716201522290685,
"grad_norm": 2.298977632173436,
"learning_rate": 3.9318223456169604e-06,
"loss": 0.9073,
"step": 988
},
{
"epoch": 0.7169264226168902,
"grad_norm": 2.838544913848763,
"learning_rate": 3.91316247588225e-06,
"loss": 1.1417,
"step": 989
},
{
"epoch": 0.7176513229430953,
"grad_norm": 2.131182681333421,
"learning_rate": 3.8945362130465035e-06,
"loss": 0.9419,
"step": 990
},
{
"epoch": 0.7183762232693005,
"grad_norm": 2.065261545069331,
"learning_rate": 3.8759436599498125e-06,
"loss": 0.9205,
"step": 991
},
{
"epoch": 0.7191011235955056,
"grad_norm": 2.1618911363647086,
"learning_rate": 3.857384919246154e-06,
"loss": 0.9108,
"step": 992
},
{
"epoch": 0.7198260239217108,
"grad_norm": 2.7172592088098044,
"learning_rate": 3.838860093402813e-06,
"loss": 1.0987,
"step": 993
},
{
"epoch": 0.7205509242479159,
"grad_norm": 2.6032460817441203,
"learning_rate": 3.820369284699823e-06,
"loss": 0.9295,
"step": 994
},
{
"epoch": 0.7212758245741211,
"grad_norm": 2.0455131957130717,
"learning_rate": 3.8019125952294045e-06,
"loss": 0.8968,
"step": 995
},
{
"epoch": 0.7220007249003262,
"grad_norm": 2.4549976135759866,
"learning_rate": 3.783490126895384e-06,
"loss": 0.9866,
"step": 996
},
{
"epoch": 0.7227256252265314,
"grad_norm": 2.38725429694958,
"learning_rate": 3.7651019814126656e-06,
"loss": 1.0763,
"step": 997
},
{
"epoch": 0.7234505255527365,
"grad_norm": 2.447633489268824,
"learning_rate": 3.746748260306635e-06,
"loss": 1.0442,
"step": 998
},
{
"epoch": 0.7241754258789417,
"grad_norm": 2.015059323195586,
"learning_rate": 3.7284290649126176e-06,
"loss": 1.0369,
"step": 999
},
{
"epoch": 0.7249003262051468,
"grad_norm": 2.1439701734436944,
"learning_rate": 3.7101444963753096e-06,
"loss": 0.9838,
"step": 1000
},
{
"epoch": 0.725625226531352,
"grad_norm": 2.196165996622997,
"learning_rate": 3.691894655648225e-06,
"loss": 1.0422,
"step": 1001
},
{
"epoch": 0.7263501268575571,
"grad_norm": 2.1072970110294498,
"learning_rate": 3.6736796434931443e-06,
"loss": 1.0379,
"step": 1002
},
{
"epoch": 0.7270750271837623,
"grad_norm": 2.052889977594413,
"learning_rate": 3.6554995604795427e-06,
"loss": 0.9731,
"step": 1003
},
{
"epoch": 0.7277999275099674,
"grad_norm": 2.1849621710221605,
"learning_rate": 3.637354506984051e-06,
"loss": 0.9397,
"step": 1004
},
{
"epoch": 0.7285248278361726,
"grad_norm": 2.302003380708732,
"learning_rate": 3.619244583189877e-06,
"loss": 1.0793,
"step": 1005
},
{
"epoch": 0.7292497281623777,
"grad_norm": 2.4873231398689692,
"learning_rate": 3.6011698890862913e-06,
"loss": 0.9777,
"step": 1006
},
{
"epoch": 0.7299746284885829,
"grad_norm": 2.06017307338732,
"learning_rate": 3.583130524468038e-06,
"loss": 0.8838,
"step": 1007
},
{
"epoch": 0.730699528814788,
"grad_norm": 1.9994143545054046,
"learning_rate": 3.565126588934803e-06,
"loss": 0.9971,
"step": 1008
},
{
"epoch": 0.7314244291409931,
"grad_norm": 1.8151755246614727,
"learning_rate": 3.547158181890661e-06,
"loss": 0.9291,
"step": 1009
},
{
"epoch": 0.7321493294671982,
"grad_norm": 1.9292646278494385,
"learning_rate": 3.529225402543518e-06,
"loss": 1.0508,
"step": 1010
},
{
"epoch": 0.7328742297934034,
"grad_norm": 1.9667040655974664,
"learning_rate": 3.5113283499045846e-06,
"loss": 1.0381,
"step": 1011
},
{
"epoch": 0.7335991301196085,
"grad_norm": 2.490956546908503,
"learning_rate": 3.4934671227878046e-06,
"loss": 0.9943,
"step": 1012
},
{
"epoch": 0.7343240304458137,
"grad_norm": 1.7574463836197582,
"learning_rate": 3.475641819809327e-06,
"loss": 0.9338,
"step": 1013
},
{
"epoch": 0.7350489307720188,
"grad_norm": 2.8098119938777795,
"learning_rate": 3.4578525393869444e-06,
"loss": 1.0599,
"step": 1014
},
{
"epoch": 0.735773831098224,
"grad_norm": 2.1547229502003877,
"learning_rate": 3.4400993797395664e-06,
"loss": 0.9071,
"step": 1015
},
{
"epoch": 0.7364987314244291,
"grad_norm": 2.3529684580357673,
"learning_rate": 3.42238243888668e-06,
"loss": 0.8903,
"step": 1016
},
{
"epoch": 0.7372236317506343,
"grad_norm": 3.3656369614826076,
"learning_rate": 3.4047018146477873e-06,
"loss": 0.9104,
"step": 1017
},
{
"epoch": 0.7379485320768394,
"grad_norm": 2.3603322866432315,
"learning_rate": 3.3870576046418825e-06,
"loss": 0.9784,
"step": 1018
},
{
"epoch": 0.7386734324030446,
"grad_norm": 2.321872140130776,
"learning_rate": 3.369449906286907e-06,
"loss": 1.0385,
"step": 1019
},
{
"epoch": 0.7393983327292497,
"grad_norm": 2.1167495374440315,
"learning_rate": 3.351878816799209e-06,
"loss": 1.0955,
"step": 1020
},
{
"epoch": 0.7401232330554549,
"grad_norm": 1.786896281579396,
"learning_rate": 3.3343444331930207e-06,
"loss": 0.991,
"step": 1021
},
{
"epoch": 0.74084813338166,
"grad_norm": 2.2746758416106427,
"learning_rate": 3.316846852279907e-06,
"loss": 0.9766,
"step": 1022
},
{
"epoch": 0.7415730337078652,
"grad_norm": 2.6138994872082097,
"learning_rate": 3.2993861706682274e-06,
"loss": 1.1535,
"step": 1023
},
{
"epoch": 0.7422979340340703,
"grad_norm": 2.6165034172242176,
"learning_rate": 3.281962484762622e-06,
"loss": 0.9753,
"step": 1024
},
{
"epoch": 0.7430228343602755,
"grad_norm": 1.9155784149193085,
"learning_rate": 3.264575890763463e-06,
"loss": 0.9675,
"step": 1025
},
{
"epoch": 0.7437477346864806,
"grad_norm": 2.0624358920096215,
"learning_rate": 3.247226484666339e-06,
"loss": 1.0538,
"step": 1026
},
{
"epoch": 0.7444726350126858,
"grad_norm": 3.5846085824449987,
"learning_rate": 3.2299143622615036e-06,
"loss": 1.107,
"step": 1027
},
{
"epoch": 0.7451975353388909,
"grad_norm": 2.46962567912047,
"learning_rate": 3.2126396191333643e-06,
"loss": 0.9327,
"step": 1028
},
{
"epoch": 0.7459224356650961,
"grad_norm": 2.4139413136075225,
"learning_rate": 3.195402350659945e-06,
"loss": 1.098,
"step": 1029
},
{
"epoch": 0.7466473359913012,
"grad_norm": 2.017065330116569,
"learning_rate": 3.178202652012363e-06,
"loss": 0.9482,
"step": 1030
},
{
"epoch": 0.7473722363175064,
"grad_norm": 2.404901697722625,
"learning_rate": 3.161040618154315e-06,
"loss": 1.0855,
"step": 1031
},
{
"epoch": 0.7480971366437115,
"grad_norm": 1.8467098745774249,
"learning_rate": 3.143916343841523e-06,
"loss": 1.0062,
"step": 1032
},
{
"epoch": 0.7488220369699167,
"grad_norm": 2.5314983525811883,
"learning_rate": 3.1268299236212395e-06,
"loss": 0.8812,
"step": 1033
},
{
"epoch": 0.7495469372961218,
"grad_norm": 2.3038281210461697,
"learning_rate": 3.1097814518317127e-06,
"loss": 1.0849,
"step": 1034
},
{
"epoch": 0.750271837622327,
"grad_norm": 2.4284883217355846,
"learning_rate": 3.092771022601676e-06,
"loss": 1.0175,
"step": 1035
},
{
"epoch": 0.750271837622327,
"eval_loss": 1.7681310176849365,
"eval_runtime": 1124.5222,
"eval_samples_per_second": 15.699,
"eval_steps_per_second": 0.245,
"step": 1035
},
{
"epoch": 0.7509967379485321,
"grad_norm": 2.1284406550026986,
"learning_rate": 3.0757987298498106e-06,
"loss": 1.0076,
"step": 1036
},
{
"epoch": 0.7517216382747373,
"grad_norm": 2.22165220425758,
"learning_rate": 3.0588646672842404e-06,
"loss": 0.951,
"step": 1037
},
{
"epoch": 0.7524465386009423,
"grad_norm": 2.5711481889523804,
"learning_rate": 3.0419689284020126e-06,
"loss": 1.0273,
"step": 1038
},
{
"epoch": 0.7531714389271476,
"grad_norm": 2.0814034949241202,
"learning_rate": 3.0251116064885767e-06,
"loss": 0.8638,
"step": 1039
},
{
"epoch": 0.7538963392533526,
"grad_norm": 2.3755493898944664,
"learning_rate": 3.0082927946172837e-06,
"loss": 0.9276,
"step": 1040
},
{
"epoch": 0.7546212395795578,
"grad_norm": 1.860045261304839,
"learning_rate": 2.9915125856488424e-06,
"loss": 0.8793,
"step": 1041
},
{
"epoch": 0.7553461399057629,
"grad_norm": 2.166915926531366,
"learning_rate": 2.974771072230842e-06,
"loss": 1.035,
"step": 1042
},
{
"epoch": 0.7560710402319681,
"grad_norm": 2.1712284867002745,
"learning_rate": 2.958068346797217e-06,
"loss": 0.9518,
"step": 1043
},
{
"epoch": 0.7567959405581732,
"grad_norm": 2.888919414338719,
"learning_rate": 2.941404501567743e-06,
"loss": 1.0718,
"step": 1044
},
{
"epoch": 0.7575208408843784,
"grad_norm": 2.230289529665858,
"learning_rate": 2.9247796285475362e-06,
"loss": 1.0045,
"step": 1045
},
{
"epoch": 0.7582457412105835,
"grad_norm": 2.4680966008183534,
"learning_rate": 2.90819381952653e-06,
"loss": 1.1622,
"step": 1046
},
{
"epoch": 0.7589706415367887,
"grad_norm": 2.965957230775603,
"learning_rate": 2.8916471660789813e-06,
"loss": 0.9385,
"step": 1047
},
{
"epoch": 0.7596955418629938,
"grad_norm": 2.3702554177333104,
"learning_rate": 2.875139759562954e-06,
"loss": 0.9725,
"step": 1048
},
{
"epoch": 0.760420442189199,
"grad_norm": 2.1238051733926113,
"learning_rate": 2.8586716911198244e-06,
"loss": 0.9945,
"step": 1049
},
{
"epoch": 0.7611453425154041,
"grad_norm": 1.8445910518867181,
"learning_rate": 2.8422430516737733e-06,
"loss": 0.959,
"step": 1050
},
{
"epoch": 0.7618702428416093,
"grad_norm": 2.4547009955302177,
"learning_rate": 2.825853931931283e-06,
"loss": 0.9879,
"step": 1051
},
{
"epoch": 0.7625951431678144,
"grad_norm": 2.002478666013241,
"learning_rate": 2.80950442238064e-06,
"loss": 0.9518,
"step": 1052
},
{
"epoch": 0.7633200434940196,
"grad_norm": 2.3705241165870197,
"learning_rate": 2.793194613291431e-06,
"loss": 1.0771,
"step": 1053
},
{
"epoch": 0.7640449438202247,
"grad_norm": 2.5129264862370957,
"learning_rate": 2.7769245947140468e-06,
"loss": 1.0603,
"step": 1054
},
{
"epoch": 0.7647698441464299,
"grad_norm": 1.8777445092267904,
"learning_rate": 2.7606944564791914e-06,
"loss": 0.8863,
"step": 1055
},
{
"epoch": 0.765494744472635,
"grad_norm": 2.0066823550902164,
"learning_rate": 2.7445042881973727e-06,
"loss": 0.9708,
"step": 1056
},
{
"epoch": 0.7662196447988402,
"grad_norm": 1.9969442657507075,
"learning_rate": 2.7283541792584165e-06,
"loss": 1.0216,
"step": 1057
},
{
"epoch": 0.7669445451250453,
"grad_norm": 2.043307797498135,
"learning_rate": 2.712244218830973e-06,
"loss": 1.0059,
"step": 1058
},
{
"epoch": 0.7676694454512505,
"grad_norm": 2.05640044579258,
"learning_rate": 2.6961744958620195e-06,
"loss": 0.977,
"step": 1059
},
{
"epoch": 0.7683943457774556,
"grad_norm": 2.0705940087415353,
"learning_rate": 2.6801450990763767e-06,
"loss": 1.0993,
"step": 1060
},
{
"epoch": 0.7691192461036608,
"grad_norm": 2.2801073634969757,
"learning_rate": 2.664156116976212e-06,
"loss": 1.1062,
"step": 1061
},
{
"epoch": 0.7698441464298659,
"grad_norm": 1.8809467398522293,
"learning_rate": 2.648207637840554e-06,
"loss": 0.9475,
"step": 1062
},
{
"epoch": 0.7705690467560711,
"grad_norm": 1.9284658912112769,
"learning_rate": 2.6322997497248036e-06,
"loss": 0.8972,
"step": 1063
},
{
"epoch": 0.7712939470822762,
"grad_norm": 2.1543128958058384,
"learning_rate": 2.616432540460255e-06,
"loss": 0.9557,
"step": 1064
},
{
"epoch": 0.7720188474084814,
"grad_norm": 1.9290768627484227,
"learning_rate": 2.6006060976535974e-06,
"loss": 0.9678,
"step": 1065
},
{
"epoch": 0.7727437477346865,
"grad_norm": 3.3836187115327174,
"learning_rate": 2.58482050868644e-06,
"loss": 1.0705,
"step": 1066
},
{
"epoch": 0.7734686480608917,
"grad_norm": 2.379431503328967,
"learning_rate": 2.5690758607148304e-06,
"loss": 1.034,
"step": 1067
},
{
"epoch": 0.7741935483870968,
"grad_norm": 2.1978701615866667,
"learning_rate": 2.553372240668769e-06,
"loss": 0.9683,
"step": 1068
},
{
"epoch": 0.774918448713302,
"grad_norm": 2.334219947775647,
"learning_rate": 2.5377097352517332e-06,
"loss": 0.9329,
"step": 1069
},
{
"epoch": 0.775643349039507,
"grad_norm": 1.8982866415446444,
"learning_rate": 2.522088430940195e-06,
"loss": 0.843,
"step": 1070
},
{
"epoch": 0.7763682493657122,
"grad_norm": 1.8718401538017733,
"learning_rate": 2.5065084139831443e-06,
"loss": 0.872,
"step": 1071
},
{
"epoch": 0.7770931496919173,
"grad_norm": 2.1329592476566663,
"learning_rate": 2.4909697704016158e-06,
"loss": 0.9499,
"step": 1072
},
{
"epoch": 0.7778180500181225,
"grad_norm": 2.3116722160450855,
"learning_rate": 2.4754725859882077e-06,
"loss": 1.071,
"step": 1073
},
{
"epoch": 0.7785429503443276,
"grad_norm": 2.411486503556151,
"learning_rate": 2.4600169463066213e-06,
"loss": 0.9443,
"step": 1074
},
{
"epoch": 0.7792678506705328,
"grad_norm": 1.8955688908325088,
"learning_rate": 2.4446029366911708e-06,
"loss": 1.0264,
"step": 1075
},
{
"epoch": 0.7799927509967379,
"grad_norm": 2.2466514583663972,
"learning_rate": 2.4292306422463218e-06,
"loss": 1.0286,
"step": 1076
},
{
"epoch": 0.7807176513229431,
"grad_norm": 2.605174653703314,
"learning_rate": 2.4139001478462233e-06,
"loss": 1.0762,
"step": 1077
},
{
"epoch": 0.7814425516491482,
"grad_norm": 2.125905431943264,
"learning_rate": 2.3986115381342347e-06,
"loss": 1.0855,
"step": 1078
},
{
"epoch": 0.7821674519753534,
"grad_norm": 2.1131544647871685,
"learning_rate": 2.3833648975224588e-06,
"loss": 0.9731,
"step": 1079
},
{
"epoch": 0.7828923523015585,
"grad_norm": 2.1666492164090068,
"learning_rate": 2.3681603101912785e-06,
"loss": 1.1026,
"step": 1080
},
{
"epoch": 0.7836172526277637,
"grad_norm": 1.902794404555453,
"learning_rate": 2.3529978600888915e-06,
"loss": 0.9013,
"step": 1081
},
{
"epoch": 0.7843421529539688,
"grad_norm": 1.904459392353737,
"learning_rate": 2.3378776309308414e-06,
"loss": 0.9706,
"step": 1082
},
{
"epoch": 0.785067053280174,
"grad_norm": 1.6870600380072511,
"learning_rate": 2.3227997061995622e-06,
"loss": 0.8814,
"step": 1083
},
{
"epoch": 0.7857919536063791,
"grad_norm": 2.034799823983302,
"learning_rate": 2.307764169143919e-06,
"loss": 0.9905,
"step": 1084
},
{
"epoch": 0.7865168539325843,
"grad_norm": 2.148503285324599,
"learning_rate": 2.292771102778739e-06,
"loss": 1.0019,
"step": 1085
},
{
"epoch": 0.7872417542587894,
"grad_norm": 1.8765819842880695,
"learning_rate": 2.277820589884359e-06,
"loss": 0.9529,
"step": 1086
},
{
"epoch": 0.7879666545849946,
"grad_norm": 2.6372186463117666,
"learning_rate": 2.26291271300617e-06,
"loss": 0.9756,
"step": 1087
},
{
"epoch": 0.7886915549111997,
"grad_norm": 2.129327741523149,
"learning_rate": 2.248047554454157e-06,
"loss": 0.9456,
"step": 1088
},
{
"epoch": 0.7894164552374049,
"grad_norm": 1.9575868058617358,
"learning_rate": 2.2332251963024475e-06,
"loss": 0.9592,
"step": 1089
},
{
"epoch": 0.79014135556361,
"grad_norm": 2.239457152716971,
"learning_rate": 2.2184457203888598e-06,
"loss": 1.0179,
"step": 1090
},
{
"epoch": 0.7908662558898152,
"grad_norm": 2.0380730008597463,
"learning_rate": 2.2037092083144463e-06,
"loss": 1.0047,
"step": 1091
},
{
"epoch": 0.7915911562160203,
"grad_norm": 2.651283946960534,
"learning_rate": 2.1890157414430448e-06,
"loss": 0.9971,
"step": 1092
},
{
"epoch": 0.7923160565422255,
"grad_norm": 1.9224733064527182,
"learning_rate": 2.1743654009008362e-06,
"loss": 0.9758,
"step": 1093
},
{
"epoch": 0.7930409568684306,
"grad_norm": 2.042210473646339,
"learning_rate": 2.1597582675758833e-06,
"loss": 0.9157,
"step": 1094
},
{
"epoch": 0.7937658571946358,
"grad_norm": 1.8843195169105176,
"learning_rate": 2.1451944221176966e-06,
"loss": 0.8876,
"step": 1095
},
{
"epoch": 0.7944907575208409,
"grad_norm": 1.7432937559280743,
"learning_rate": 2.1306739449367796e-06,
"loss": 0.966,
"step": 1096
},
{
"epoch": 0.7952156578470461,
"grad_norm": 2.4904321442188073,
"learning_rate": 2.116196916204193e-06,
"loss": 1.0147,
"step": 1097
},
{
"epoch": 0.7959405581732512,
"grad_norm": 1.867849742985621,
"learning_rate": 2.1017634158511037e-06,
"loss": 0.8396,
"step": 1098
},
{
"epoch": 0.7966654584994564,
"grad_norm": 2.176037217510385,
"learning_rate": 2.0873735235683535e-06,
"loss": 1.0433,
"step": 1099
},
{
"epoch": 0.7973903588256614,
"grad_norm": 1.8093979197769543,
"learning_rate": 2.073027318806008e-06,
"loss": 1.0163,
"step": 1100
},
{
"epoch": 0.7981152591518667,
"grad_norm": 2.317013372213296,
"learning_rate": 2.058724880772929e-06,
"loss": 0.934,
"step": 1101
},
{
"epoch": 0.7988401594780717,
"grad_norm": 1.8950042969585994,
"learning_rate": 2.044466288436322e-06,
"loss": 0.9235,
"step": 1102
},
{
"epoch": 0.799565059804277,
"grad_norm": 2.26981624646591,
"learning_rate": 2.030251620521325e-06,
"loss": 0.9664,
"step": 1103
},
{
"epoch": 0.800289960130482,
"grad_norm": 1.9558043676284143,
"learning_rate": 2.0160809555105467e-06,
"loss": 1.0178,
"step": 1104
},
{
"epoch": 0.8010148604566872,
"grad_norm": 2.919669591456051,
"learning_rate": 2.0019543716436474e-06,
"loss": 0.9538,
"step": 1105
},
{
"epoch": 0.8017397607828923,
"grad_norm": 2.0335272734209924,
"learning_rate": 1.9878719469169104e-06,
"loss": 0.8791,
"step": 1106
},
{
"epoch": 0.8024646611090975,
"grad_norm": 2.171132136552404,
"learning_rate": 1.9738337590827918e-06,
"loss": 0.9331,
"step": 1107
},
{
"epoch": 0.8031895614353026,
"grad_norm": 1.942377740059999,
"learning_rate": 1.9598398856495227e-06,
"loss": 0.8755,
"step": 1108
},
{
"epoch": 0.8039144617615078,
"grad_norm": 2.135810469663323,
"learning_rate": 1.9458904038806537e-06,
"loss": 1.0203,
"step": 1109
},
{
"epoch": 0.8046393620877129,
"grad_norm": 2.3944483070855362,
"learning_rate": 1.9319853907946393e-06,
"loss": 0.9595,
"step": 1110
},
{
"epoch": 0.8053642624139181,
"grad_norm": 1.9622747573911234,
"learning_rate": 1.918124923164414e-06,
"loss": 0.9016,
"step": 1111
},
{
"epoch": 0.8060891627401232,
"grad_norm": 2.703351031078299,
"learning_rate": 1.9043090775169604e-06,
"loss": 1.0531,
"step": 1112
},
{
"epoch": 0.8068140630663284,
"grad_norm": 2.1339961008284063,
"learning_rate": 1.890537930132903e-06,
"loss": 1.0974,
"step": 1113
},
{
"epoch": 0.8075389633925335,
"grad_norm": 2.6633749557025896,
"learning_rate": 1.8768115570460677e-06,
"loss": 1.0197,
"step": 1114
},
{
"epoch": 0.8082638637187387,
"grad_norm": 1.7899870312478274,
"learning_rate": 1.8631300340430747e-06,
"loss": 1.0164,
"step": 1115
},
{
"epoch": 0.8089887640449438,
"grad_norm": 1.786288489816615,
"learning_rate": 1.8494934366629057e-06,
"loss": 1.0085,
"step": 1116
},
{
"epoch": 0.809713664371149,
"grad_norm": 2.213791559175714,
"learning_rate": 1.8359018401965156e-06,
"loss": 1.0711,
"step": 1117
},
{
"epoch": 0.8104385646973541,
"grad_norm": 2.227118578843932,
"learning_rate": 1.8223553196863852e-06,
"loss": 1.0198,
"step": 1118
},
{
"epoch": 0.8111634650235593,
"grad_norm": 1.9847568274305198,
"learning_rate": 1.808853949926125e-06,
"loss": 0.9065,
"step": 1119
},
{
"epoch": 0.8118883653497644,
"grad_norm": 1.9370018576926253,
"learning_rate": 1.795397805460053e-06,
"loss": 0.9994,
"step": 1120
},
{
"epoch": 0.8126132656759696,
"grad_norm": 2.2601268831652424,
"learning_rate": 1.781986960582789e-06,
"loss": 0.9872,
"step": 1121
},
{
"epoch": 0.8133381660021747,
"grad_norm": 2.3214803090685265,
"learning_rate": 1.7686214893388509e-06,
"loss": 0.9926,
"step": 1122
},
{
"epoch": 0.8140630663283799,
"grad_norm": 2.5777964336742647,
"learning_rate": 1.7553014655222255e-06,
"loss": 0.9657,
"step": 1123
},
{
"epoch": 0.814787966654585,
"grad_norm": 1.9895856551523463,
"learning_rate": 1.7420269626759822e-06,
"loss": 0.8018,
"step": 1124
},
{
"epoch": 0.8155128669807902,
"grad_norm": 2.071389074929151,
"learning_rate": 1.7287980540918548e-06,
"loss": 1.1168,
"step": 1125
},
{
"epoch": 0.8162377673069953,
"grad_norm": 2.3768284586683786,
"learning_rate": 1.7156148128098337e-06,
"loss": 1.0561,
"step": 1126
},
{
"epoch": 0.8169626676332005,
"grad_norm": 2.108829554647349,
"learning_rate": 1.7024773116177839e-06,
"loss": 1.0184,
"step": 1127
},
{
"epoch": 0.8176875679594056,
"grad_norm": 1.9145136763918424,
"learning_rate": 1.6893856230510152e-06,
"loss": 1.0269,
"step": 1128
},
{
"epoch": 0.8184124682856108,
"grad_norm": 1.984940928437078,
"learning_rate": 1.6763398193919034e-06,
"loss": 0.9718,
"step": 1129
},
{
"epoch": 0.8191373686118159,
"grad_norm": 2.491917383583085,
"learning_rate": 1.6633399726694755e-06,
"loss": 0.9837,
"step": 1130
},
{
"epoch": 0.819862268938021,
"grad_norm": 2.1889459447102295,
"learning_rate": 1.6503861546590216e-06,
"loss": 0.8859,
"step": 1131
},
{
"epoch": 0.8205871692642261,
"grad_norm": 2.2855735715808807,
"learning_rate": 1.6374784368817032e-06,
"loss": 0.8568,
"step": 1132
},
{
"epoch": 0.8213120695904313,
"grad_norm": 2.4119276770093245,
"learning_rate": 1.6246168906041405e-06,
"loss": 1.0357,
"step": 1133
},
{
"epoch": 0.8220369699166364,
"grad_norm": 2.2497941042194207,
"learning_rate": 1.6118015868380387e-06,
"loss": 1.0791,
"step": 1134
},
{
"epoch": 0.8227618702428416,
"grad_norm": 2.6543479685385036,
"learning_rate": 1.5990325963397779e-06,
"loss": 1.0276,
"step": 1135
},
{
"epoch": 0.8234867705690467,
"grad_norm": 1.9249733038058623,
"learning_rate": 1.5863099896100343e-06,
"loss": 0.966,
"step": 1136
},
{
"epoch": 0.8242116708952519,
"grad_norm": 2.2761621468500497,
"learning_rate": 1.5736338368933978e-06,
"loss": 0.8657,
"step": 1137
},
{
"epoch": 0.824936571221457,
"grad_norm": 1.9374428897603755,
"learning_rate": 1.561004208177962e-06,
"loss": 1.0411,
"step": 1138
},
{
"epoch": 0.8256614715476622,
"grad_norm": 2.13221264820333,
"learning_rate": 1.5484211731949572e-06,
"loss": 1.0386,
"step": 1139
},
{
"epoch": 0.8263863718738673,
"grad_norm": 2.1816254648738194,
"learning_rate": 1.5358848014183547e-06,
"loss": 1.0673,
"step": 1140
},
{
"epoch": 0.8271112722000725,
"grad_norm": 2.333905932739388,
"learning_rate": 1.523395162064486e-06,
"loss": 0.8971,
"step": 1141
},
{
"epoch": 0.8278361725262776,
"grad_norm": 1.9732510886779988,
"learning_rate": 1.5109523240916702e-06,
"loss": 0.9907,
"step": 1142
},
{
"epoch": 0.8285610728524828,
"grad_norm": 2.0768050361793606,
"learning_rate": 1.4985563561998185e-06,
"loss": 1.0126,
"step": 1143
},
{
"epoch": 0.8292859731786879,
"grad_norm": 2.0071930948425933,
"learning_rate": 1.4862073268300558e-06,
"loss": 0.9991,
"step": 1144
},
{
"epoch": 0.8300108735048931,
"grad_norm": 2.0782467893964593,
"learning_rate": 1.4739053041643536e-06,
"loss": 0.9887,
"step": 1145
},
{
"epoch": 0.8307357738310982,
"grad_norm": 2.171294999994677,
"learning_rate": 1.4616503561251538e-06,
"loss": 0.9402,
"step": 1146
},
{
"epoch": 0.8314606741573034,
"grad_norm": 2.407000985423698,
"learning_rate": 1.4494425503749788e-06,
"loss": 0.981,
"step": 1147
},
{
"epoch": 0.8321855744835085,
"grad_norm": 1.9223688184324994,
"learning_rate": 1.437281954316071e-06,
"loss": 0.9358,
"step": 1148
},
{
"epoch": 0.8329104748097137,
"grad_norm": 1.9968819662132635,
"learning_rate": 1.4251686350900152e-06,
"loss": 0.9298,
"step": 1149
},
{
"epoch": 0.8336353751359188,
"grad_norm": 2.136971059899701,
"learning_rate": 1.4131026595773689e-06,
"loss": 0.9215,
"step": 1150
},
{
"epoch": 0.834360275462124,
"grad_norm": 1.8337681676570194,
"learning_rate": 1.4010840943972982e-06,
"loss": 1.0268,
"step": 1151
},
{
"epoch": 0.8350851757883291,
"grad_norm": 2.1977568849453744,
"learning_rate": 1.3891130059072032e-06,
"loss": 0.9156,
"step": 1152
},
{
"epoch": 0.8358100761145343,
"grad_norm": 2.255344231343079,
"learning_rate": 1.3771894602023483e-06,
"loss": 1.087,
"step": 1153
},
{
"epoch": 0.8365349764407394,
"grad_norm": 2.2385193595399606,
"learning_rate": 1.3653135231155079e-06,
"loss": 1.0131,
"step": 1154
},
{
"epoch": 0.8372598767669446,
"grad_norm": 1.9074930266008068,
"learning_rate": 1.353485260216596e-06,
"loss": 1.046,
"step": 1155
},
{
"epoch": 0.8379847770931497,
"grad_norm": 2.3457699506131124,
"learning_rate": 1.3417047368123103e-06,
"loss": 1.0063,
"step": 1156
},
{
"epoch": 0.8387096774193549,
"grad_norm": 2.195479709372169,
"learning_rate": 1.329972017945761e-06,
"loss": 0.9382,
"step": 1157
},
{
"epoch": 0.83943457774556,
"grad_norm": 2.815541285577747,
"learning_rate": 1.3182871683961217e-06,
"loss": 0.9169,
"step": 1158
},
{
"epoch": 0.8401594780717652,
"grad_norm": 2.473680572699823,
"learning_rate": 1.3066502526782653e-06,
"loss": 0.984,
"step": 1159
},
{
"epoch": 0.8408843783979703,
"grad_norm": 2.3309597371635995,
"learning_rate": 1.2950613350424113e-06,
"loss": 0.9095,
"step": 1160
},
{
"epoch": 0.8416092787241755,
"grad_norm": 2.6521093626394148,
"learning_rate": 1.2835204794737755e-06,
"loss": 0.9806,
"step": 1161
},
{
"epoch": 0.8423341790503805,
"grad_norm": 2.2183368928655134,
"learning_rate": 1.272027749692203e-06,
"loss": 1.0063,
"step": 1162
},
{
"epoch": 0.8430590793765858,
"grad_norm": 2.335358593922603,
"learning_rate": 1.2605832091518277e-06,
"loss": 1.0009,
"step": 1163
},
{
"epoch": 0.8437839797027908,
"grad_norm": 3.1347973939628586,
"learning_rate": 1.2491869210407215e-06,
"loss": 1.0215,
"step": 1164
},
{
"epoch": 0.844508880028996,
"grad_norm": 2.2148770630869667,
"learning_rate": 1.2378389482805397e-06,
"loss": 0.9492,
"step": 1165
},
{
"epoch": 0.8452337803552011,
"grad_norm": 1.8250263747397901,
"learning_rate": 1.2265393535261827e-06,
"loss": 0.9507,
"step": 1166
},
{
"epoch": 0.8459586806814063,
"grad_norm": 2.5651175153179806,
"learning_rate": 1.2152881991654364e-06,
"loss": 1.0011,
"step": 1167
},
{
"epoch": 0.8466835810076114,
"grad_norm": 1.9418141541403875,
"learning_rate": 1.204085547318642e-06,
"loss": 1.0761,
"step": 1168
},
{
"epoch": 0.8474084813338166,
"grad_norm": 2.0646032693286833,
"learning_rate": 1.1929314598383423e-06,
"loss": 0.9781,
"step": 1169
},
{
"epoch": 0.8481333816600217,
"grad_norm": 1.8004575813318966,
"learning_rate": 1.1818259983089452e-06,
"loss": 0.9592,
"step": 1170
},
{
"epoch": 0.8488582819862269,
"grad_norm": 2.589648686714714,
"learning_rate": 1.1707692240463876e-06,
"loss": 1.1313,
"step": 1171
},
{
"epoch": 0.849583182312432,
"grad_norm": 2.0114408256271075,
"learning_rate": 1.1597611980977853e-06,
"loss": 0.9651,
"step": 1172
},
{
"epoch": 0.8503080826386372,
"grad_norm": 1.7560110108173013,
"learning_rate": 1.1488019812411067e-06,
"loss": 0.969,
"step": 1173
},
{
"epoch": 0.8510329829648423,
"grad_norm": 2.0223370350736958,
"learning_rate": 1.1378916339848323e-06,
"loss": 0.896,
"step": 1174
},
{
"epoch": 0.8517578832910475,
"grad_norm": 2.2941607223768075,
"learning_rate": 1.127030216567625e-06,
"loss": 1.1139,
"step": 1175
},
{
"epoch": 0.8524827836172526,
"grad_norm": 1.7519205543151821,
"learning_rate": 1.1162177889579906e-06,
"loss": 0.9145,
"step": 1176
},
{
"epoch": 0.8532076839434578,
"grad_norm": 2.0778773139552094,
"learning_rate": 1.1054544108539511e-06,
"loss": 0.8809,
"step": 1177
},
{
"epoch": 0.8539325842696629,
"grad_norm": 1.8224740916196427,
"learning_rate": 1.0947401416827152e-06,
"loss": 0.9089,
"step": 1178
},
{
"epoch": 0.8546574845958681,
"grad_norm": 3.369653427250154,
"learning_rate": 1.0840750406003486e-06,
"loss": 0.9113,
"step": 1179
},
{
"epoch": 0.8553823849220732,
"grad_norm": 3.0341804857536254,
"learning_rate": 1.0734591664914463e-06,
"loss": 1.0581,
"step": 1180
},
{
"epoch": 0.8561072852482784,
"grad_norm": 2.176890160993487,
"learning_rate": 1.0628925779688148e-06,
"loss": 0.8511,
"step": 1181
},
{
"epoch": 0.8568321855744835,
"grad_norm": 2.505948283667956,
"learning_rate": 1.0523753333731357e-06,
"loss": 1.023,
"step": 1182
},
{
"epoch": 0.8575570859006887,
"grad_norm": 2.8808186692504423,
"learning_rate": 1.041907490772658e-06,
"loss": 0.9036,
"step": 1183
},
{
"epoch": 0.8582819862268938,
"grad_norm": 2.378412246151113,
"learning_rate": 1.031489107962863e-06,
"loss": 0.951,
"step": 1184
},
{
"epoch": 0.859006886553099,
"grad_norm": 2.4071476437909594,
"learning_rate": 1.0211202424661604e-06,
"loss": 0.9586,
"step": 1185
},
{
"epoch": 0.8597317868793041,
"grad_norm": 2.6658079886381785,
"learning_rate": 1.0108009515315586e-06,
"loss": 1.1094,
"step": 1186
},
{
"epoch": 0.8604566872055093,
"grad_norm": 2.2215584462527596,
"learning_rate": 1.0005312921343523e-06,
"loss": 0.8504,
"step": 1187
},
{
"epoch": 0.8611815875317144,
"grad_norm": 2.0187808429022107,
"learning_rate": 9.903113209758098e-07,
"loss": 0.9252,
"step": 1188
},
{
"epoch": 0.8619064878579196,
"grad_norm": 2.4590642711919006,
"learning_rate": 9.801410944828572e-07,
"loss": 0.9505,
"step": 1189
},
{
"epoch": 0.8626313881841247,
"grad_norm": 1.9067553771838575,
"learning_rate": 9.700206688077707e-07,
"loss": 0.9373,
"step": 1190
},
{
"epoch": 0.8633562885103299,
"grad_norm": 2.206921164430631,
"learning_rate": 9.599500998278632e-07,
"loss": 1.0474,
"step": 1191
},
{
"epoch": 0.864081188836535,
"grad_norm": 2.2333129821917095,
"learning_rate": 9.499294431451755e-07,
"loss": 1.0328,
"step": 1192
},
{
"epoch": 0.8648060891627402,
"grad_norm": 2.482019605419004,
"learning_rate": 9.399587540861721e-07,
"loss": 0.9418,
"step": 1193
},
{
"epoch": 0.8655309894889452,
"grad_norm": 1.998307923726419,
"learning_rate": 9.300380877014315e-07,
"loss": 0.979,
"step": 1194
},
{
"epoch": 0.8662558898151504,
"grad_norm": 1.9225932853992822,
"learning_rate": 9.201674987653509e-07,
"loss": 0.9395,
"step": 1195
},
{
"epoch": 0.8669807901413555,
"grad_norm": 2.9091365349477796,
"learning_rate": 9.10347041775832e-07,
"loss": 0.9469,
"step": 1196
},
{
"epoch": 0.8677056904675607,
"grad_norm": 1.9149962880721147,
"learning_rate": 9.00576770953987e-07,
"loss": 1.0476,
"step": 1197
},
{
"epoch": 0.8684305907937658,
"grad_norm": 1.9887916713607694,
"learning_rate": 8.908567402438417e-07,
"loss": 0.8714,
"step": 1198
},
{
"epoch": 0.869155491119971,
"grad_norm": 2.245816854601276,
"learning_rate": 8.811870033120284e-07,
"loss": 0.9797,
"step": 1199
},
{
"epoch": 0.8698803914461761,
"grad_norm": 2.2713836769286595,
"learning_rate": 8.715676135475004e-07,
"loss": 0.9377,
"step": 1200
},
{
"epoch": 0.8706052917723813,
"grad_norm": 2.5205170405942905,
"learning_rate": 8.6199862406123e-07,
"loss": 0.9551,
"step": 1201
},
{
"epoch": 0.8713301920985864,
"grad_norm": 2.35062405333527,
"learning_rate": 8.524800876859162e-07,
"loss": 0.997,
"step": 1202
},
{
"epoch": 0.8720550924247916,
"grad_norm": 2.214767113139861,
"learning_rate": 8.430120569756949e-07,
"loss": 0.9209,
"step": 1203
},
{
"epoch": 0.8727799927509967,
"grad_norm": 1.8210526020082338,
"learning_rate": 8.335945842058524e-07,
"loss": 0.8505,
"step": 1204
},
{
"epoch": 0.8735048930772019,
"grad_norm": 3.169940456082299,
"learning_rate": 8.242277213725258e-07,
"loss": 1.0918,
"step": 1205
},
{
"epoch": 0.874229793403407,
"grad_norm": 2.466935646079598,
"learning_rate": 8.149115201924263e-07,
"loss": 0.9271,
"step": 1206
},
{
"epoch": 0.8749546937296122,
"grad_norm": 1.8036637263972901,
"learning_rate": 8.056460321025483e-07,
"loss": 0.9608,
"step": 1207
},
{
"epoch": 0.8756795940558173,
"grad_norm": 1.9639970375097486,
"learning_rate": 7.964313082598884e-07,
"loss": 0.863,
"step": 1208
},
{
"epoch": 0.8764044943820225,
"grad_norm": 1.8756276620751722,
"learning_rate": 7.872673995411606e-07,
"loss": 0.8938,
"step": 1209
},
{
"epoch": 0.8771293947082276,
"grad_norm": 2.2245625435715577,
"learning_rate": 7.781543565425153e-07,
"loss": 0.8795,
"step": 1210
},
{
"epoch": 0.8778542950344328,
"grad_norm": 2.5233957402711105,
"learning_rate": 7.690922295792647e-07,
"loss": 0.9564,
"step": 1211
},
{
"epoch": 0.8785791953606379,
"grad_norm": 2.325154347559063,
"learning_rate": 7.600810686855975e-07,
"loss": 0.9919,
"step": 1212
},
{
"epoch": 0.8793040956868431,
"grad_norm": 2.5256436615222437,
"learning_rate": 7.511209236143047e-07,
"loss": 1.0318,
"step": 1213
},
{
"epoch": 0.8800289960130482,
"grad_norm": 2.709568964312284,
"learning_rate": 7.422118438365156e-07,
"loss": 1.0083,
"step": 1214
},
{
"epoch": 0.8807538963392534,
"grad_norm": 2.2518723793524384,
"learning_rate": 7.333538785414062e-07,
"loss": 0.894,
"step": 1215
},
{
"epoch": 0.8814787966654585,
"grad_norm": 2.443491219954294,
"learning_rate": 7.24547076635942e-07,
"loss": 0.9463,
"step": 1216
},
{
"epoch": 0.8822036969916637,
"grad_norm": 2.134388443555549,
"learning_rate": 7.157914867446003e-07,
"loss": 0.933,
"step": 1217
},
{
"epoch": 0.8829285973178688,
"grad_norm": 2.512674514907188,
"learning_rate": 7.070871572091076e-07,
"loss": 0.893,
"step": 1218
},
{
"epoch": 0.883653497644074,
"grad_norm": 2.453936234251306,
"learning_rate": 6.984341360881664e-07,
"loss": 0.9721,
"step": 1219
},
{
"epoch": 0.8843783979702791,
"grad_norm": 2.1489162189409337,
"learning_rate": 6.898324711571958e-07,
"loss": 0.9212,
"step": 1220
},
{
"epoch": 0.8851032982964843,
"grad_norm": 2.510668864545298,
"learning_rate": 6.812822099080629e-07,
"loss": 0.9371,
"step": 1221
},
{
"epoch": 0.8858281986226894,
"grad_norm": 1.9545987636324476,
"learning_rate": 6.727833995488242e-07,
"loss": 0.9738,
"step": 1222
},
{
"epoch": 0.8865530989488946,
"grad_norm": 2.1706699023503604,
"learning_rate": 6.64336087003461e-07,
"loss": 1.1229,
"step": 1223
},
{
"epoch": 0.8872779992750996,
"grad_norm": 1.8682535430546066,
"learning_rate": 6.559403189116275e-07,
"loss": 0.8849,
"step": 1224
},
{
"epoch": 0.8880028996013049,
"grad_norm": 2.159689783926262,
"learning_rate": 6.475961416283838e-07,
"loss": 0.9401,
"step": 1225
},
{
"epoch": 0.8887277999275099,
"grad_norm": 2.7767173791052913,
"learning_rate": 6.393036012239473e-07,
"loss": 0.8924,
"step": 1226
},
{
"epoch": 0.8894527002537151,
"grad_norm": 2.4204337218238905,
"learning_rate": 6.310627434834349e-07,
"loss": 0.9524,
"step": 1227
},
{
"epoch": 0.8901776005799202,
"grad_norm": 1.8765448611106061,
"learning_rate": 6.228736139066105e-07,
"loss": 0.859,
"step": 1228
},
{
"epoch": 0.8909025009061254,
"grad_norm": 1.919204528359367,
"learning_rate": 6.147362577076343e-07,
"loss": 0.8362,
"step": 1229
},
{
"epoch": 0.8916274012323305,
"grad_norm": 1.9433910718170642,
"learning_rate": 6.066507198148142e-07,
"loss": 0.8749,
"step": 1230
},
{
"epoch": 0.8923523015585357,
"grad_norm": 2.9697875554755884,
"learning_rate": 5.986170448703577e-07,
"loss": 0.943,
"step": 1231
},
{
"epoch": 0.8930772018847408,
"grad_norm": 2.353044622771824,
"learning_rate": 5.906352772301193e-07,
"loss": 1.1478,
"step": 1232
},
{
"epoch": 0.893802102210946,
"grad_norm": 1.850720536456356,
"learning_rate": 5.827054609633686e-07,
"loss": 0.9644,
"step": 1233
},
{
"epoch": 0.8945270025371511,
"grad_norm": 1.8013108857133973,
"learning_rate": 5.748276398525332e-07,
"loss": 1.0246,
"step": 1234
},
{
"epoch": 0.8952519028633563,
"grad_norm": 2.469890038185596,
"learning_rate": 5.670018573929647e-07,
"loss": 0.8624,
"step": 1235
},
{
"epoch": 0.8959768031895614,
"grad_norm": 2.2167116711922175,
"learning_rate": 5.592281567926983e-07,
"loss": 0.9879,
"step": 1236
},
{
"epoch": 0.8967017035157666,
"grad_norm": 2.4449293680021236,
"learning_rate": 5.515065809722064e-07,
"loss": 0.9898,
"step": 1237
},
{
"epoch": 0.8974266038419717,
"grad_norm": 2.746110084978734,
"learning_rate": 5.438371725641778e-07,
"loss": 1.0036,
"step": 1238
},
{
"epoch": 0.8981515041681769,
"grad_norm": 2.4639823270799983,
"learning_rate": 5.362199739132656e-07,
"loss": 1.0811,
"step": 1239
},
{
"epoch": 0.898876404494382,
"grad_norm": 1.9704675962545042,
"learning_rate": 5.286550270758617e-07,
"loss": 0.8005,
"step": 1240
},
{
"epoch": 0.8996013048205872,
"grad_norm": 2.2146310697220546,
"learning_rate": 5.211423738198618e-07,
"loss": 0.9509,
"step": 1241
},
{
"epoch": 0.9003262051467923,
"grad_norm": 1.8720450871060204,
"learning_rate": 5.136820556244393e-07,
"loss": 0.9735,
"step": 1242
},
{
"epoch": 0.9010511054729975,
"grad_norm": 1.8982694817893693,
"learning_rate": 5.062741136798111e-07,
"loss": 0.9381,
"step": 1243
},
{
"epoch": 0.9017760057992026,
"grad_norm": 1.858714766448879,
"learning_rate": 4.98918588887013e-07,
"loss": 0.9523,
"step": 1244
},
{
"epoch": 0.9025009061254078,
"grad_norm": 2.4016041132668344,
"learning_rate": 4.916155218576734e-07,
"loss": 1.0073,
"step": 1245
},
{
"epoch": 0.9032258064516129,
"grad_norm": 2.347954663659916,
"learning_rate": 4.843649529137861e-07,
"loss": 1.0266,
"step": 1246
},
{
"epoch": 0.9039507067778181,
"grad_norm": 2.1409896869995193,
"learning_rate": 4.771669220874908e-07,
"loss": 0.9691,
"step": 1247
},
{
"epoch": 0.9046756071040232,
"grad_norm": 2.49959014733721,
"learning_rate": 4.70021469120856e-07,
"loss": 1.0186,
"step": 1248
},
{
"epoch": 0.9054005074302284,
"grad_norm": 2.1558155360425513,
"learning_rate": 4.629286334656502e-07,
"loss": 0.9444,
"step": 1249
},
{
"epoch": 0.9061254077564335,
"grad_norm": 1.6872092574160515,
"learning_rate": 4.55888454283131e-07,
"loss": 0.8999,
"step": 1250
},
{
"epoch": 0.9068503080826387,
"grad_norm": 2.4896678483642445,
"learning_rate": 4.489009704438274e-07,
"loss": 0.9269,
"step": 1251
},
{
"epoch": 0.9075752084088438,
"grad_norm": 2.2948377479813344,
"learning_rate": 4.419662205273234e-07,
"loss": 1.0014,
"step": 1252
},
{
"epoch": 0.908300108735049,
"grad_norm": 2.287777069195488,
"learning_rate": 4.35084242822047e-07,
"loss": 0.9587,
"step": 1253
},
{
"epoch": 0.909025009061254,
"grad_norm": 2.32897276572487,
"learning_rate": 4.282550753250603e-07,
"loss": 0.9474,
"step": 1254
},
{
"epoch": 0.9097499093874593,
"grad_norm": 2.314915920284867,
"learning_rate": 4.2147875574184095e-07,
"loss": 0.981,
"step": 1255
},
{
"epoch": 0.9104748097136643,
"grad_norm": 1.9516096065203952,
"learning_rate": 4.1475532148608665e-07,
"loss": 0.9547,
"step": 1256
},
{
"epoch": 0.9111997100398695,
"grad_norm": 2.1927356687145227,
"learning_rate": 4.08084809679502e-07,
"loss": 0.9606,
"step": 1257
},
{
"epoch": 0.9119246103660746,
"grad_norm": 2.0841597477886924,
"learning_rate": 4.0146725715159096e-07,
"loss": 0.9804,
"step": 1258
},
{
"epoch": 0.9126495106922798,
"grad_norm": 2.5613066364600914,
"learning_rate": 3.9490270043945787e-07,
"loss": 1.0065,
"step": 1259
},
{
"epoch": 0.9133744110184849,
"grad_norm": 2.4125648530924497,
"learning_rate": 3.883911757876058e-07,
"loss": 0.9775,
"step": 1260
},
{
"epoch": 0.9140993113446901,
"grad_norm": 1.939270842406947,
"learning_rate": 3.8193271914773287e-07,
"loss": 0.9695,
"step": 1261
},
{
"epoch": 0.9148242116708952,
"grad_norm": 2.2535052888130336,
"learning_rate": 3.7552736617853745e-07,
"loss": 0.8691,
"step": 1262
},
{
"epoch": 0.9155491119971004,
"grad_norm": 2.06134443104641,
"learning_rate": 3.6917515224552115e-07,
"loss": 0.9972,
"step": 1263
},
{
"epoch": 0.9162740123233055,
"grad_norm": 2.187718479642105,
"learning_rate": 3.628761124207858e-07,
"loss": 0.9614,
"step": 1264
},
{
"epoch": 0.9169989126495107,
"grad_norm": 2.728957931329134,
"learning_rate": 3.566302814828526e-07,
"loss": 1.0619,
"step": 1265
},
{
"epoch": 0.9177238129757158,
"grad_norm": 2.295000346358395,
"learning_rate": 3.504376939164611e-07,
"loss": 1.0473,
"step": 1266
},
{
"epoch": 0.918448713301921,
"grad_norm": 2.077400704239704,
"learning_rate": 3.442983839123826e-07,
"loss": 1.0098,
"step": 1267
},
{
"epoch": 0.9191736136281261,
"grad_norm": 2.3680349659705016,
"learning_rate": 3.382123853672303e-07,
"loss": 1.0091,
"step": 1268
},
{
"epoch": 0.9198985139543313,
"grad_norm": 2.392219243933128,
"learning_rate": 3.321797318832687e-07,
"loss": 1.0328,
"step": 1269
},
{
"epoch": 0.9206234142805364,
"grad_norm": 1.933874299943873,
"learning_rate": 3.262004567682342e-07,
"loss": 0.9634,
"step": 1270
},
{
"epoch": 0.9213483146067416,
"grad_norm": 2.1317697728223997,
"learning_rate": 3.20274593035147e-07,
"loss": 0.9766,
"step": 1271
},
{
"epoch": 0.9220732149329467,
"grad_norm": 1.7882374391836628,
"learning_rate": 3.144021734021352e-07,
"loss": 0.9166,
"step": 1272
},
{
"epoch": 0.9227981152591519,
"grad_norm": 2.053906284134757,
"learning_rate": 3.085832302922398e-07,
"loss": 0.9921,
"step": 1273
},
{
"epoch": 0.923523015585357,
"grad_norm": 2.130994971696479,
"learning_rate": 3.028177958332512e-07,
"loss": 0.9309,
"step": 1274
},
{
"epoch": 0.9242479159115622,
"grad_norm": 2.070911861584657,
"learning_rate": 2.971059018575262e-07,
"loss": 0.9546,
"step": 1275
},
{
"epoch": 0.9249728162377673,
"grad_norm": 2.180354582475891,
"learning_rate": 2.9144757990180685e-07,
"loss": 0.9385,
"step": 1276
},
{
"epoch": 0.9256977165639725,
"grad_norm": 2.101379134926334,
"learning_rate": 2.858428612070585e-07,
"loss": 0.9473,
"step": 1277
},
{
"epoch": 0.9264226168901776,
"grad_norm": 2.941363671739184,
"learning_rate": 2.802917767182822e-07,
"loss": 1.0476,
"step": 1278
},
{
"epoch": 0.9271475172163828,
"grad_norm": 2.099272282487294,
"learning_rate": 2.74794357084357e-07,
"loss": 1.135,
"step": 1279
},
{
"epoch": 0.9278724175425879,
"grad_norm": 2.0386416225655744,
"learning_rate": 2.693506326578632e-07,
"loss": 0.9461,
"step": 1280
},
{
"epoch": 0.9285973178687931,
"grad_norm": 2.158374614724584,
"learning_rate": 2.639606334949163e-07,
"loss": 0.8921,
"step": 1281
},
{
"epoch": 0.9293222181949982,
"grad_norm": 1.8765718966068676,
"learning_rate": 2.586243893550033e-07,
"loss": 0.9305,
"step": 1282
},
{
"epoch": 0.9300471185212034,
"grad_norm": 2.3160935821533144,
"learning_rate": 2.533419297008155e-07,
"loss": 0.9176,
"step": 1283
},
{
"epoch": 0.9307720188474085,
"grad_norm": 2.1878330344349313,
"learning_rate": 2.481132836980871e-07,
"loss": 0.955,
"step": 1284
},
{
"epoch": 0.9314969191736137,
"grad_norm": 2.195724525034424,
"learning_rate": 2.4293848021543443e-07,
"loss": 0.8916,
"step": 1285
},
{
"epoch": 0.9322218194998187,
"grad_norm": 1.7464619703681252,
"learning_rate": 2.378175478241984e-07,
"loss": 0.9268,
"step": 1286
},
{
"epoch": 0.932946719826024,
"grad_norm": 2.5869309705425465,
"learning_rate": 2.3275051479828115e-07,
"loss": 0.9683,
"step": 1287
},
{
"epoch": 0.933671620152229,
"grad_norm": 1.9334780055162688,
"learning_rate": 2.27737409113995e-07,
"loss": 0.9208,
"step": 1288
},
{
"epoch": 0.9343965204784342,
"grad_norm": 1.8929021569949063,
"learning_rate": 2.2277825844990607e-07,
"loss": 0.912,
"step": 1289
},
{
"epoch": 0.9351214208046393,
"grad_norm": 2.2127598690640564,
"learning_rate": 2.1787309018668213e-07,
"loss": 1.0383,
"step": 1290
},
{
"epoch": 0.9358463211308445,
"grad_norm": 2.210748449295597,
"learning_rate": 2.130219314069415e-07,
"loss": 0.9721,
"step": 1291
},
{
"epoch": 0.9365712214570496,
"grad_norm": 2.115265811779472,
"learning_rate": 2.0822480889510222e-07,
"loss": 1.0026,
"step": 1292
},
{
"epoch": 0.9372961217832548,
"grad_norm": 2.040211532276065,
"learning_rate": 2.034817491372354e-07,
"loss": 0.9751,
"step": 1293
},
{
"epoch": 0.9380210221094599,
"grad_norm": 2.3955059923006483,
"learning_rate": 1.9879277832091758e-07,
"loss": 1.0528,
"step": 1294
},
{
"epoch": 0.9387459224356651,
"grad_norm": 2.0551722313954297,
"learning_rate": 1.941579223350898e-07,
"loss": 0.9919,
"step": 1295
},
{
"epoch": 0.9394708227618702,
"grad_norm": 1.8511261527636662,
"learning_rate": 1.8957720676991087e-07,
"loss": 0.957,
"step": 1296
},
{
"epoch": 0.9401957230880754,
"grad_norm": 1.986865247006612,
"learning_rate": 1.850506569166155e-07,
"loss": 1.0243,
"step": 1297
},
{
"epoch": 0.9409206234142805,
"grad_norm": 1.9332096525646203,
"learning_rate": 1.8057829776737867e-07,
"loss": 1.0149,
"step": 1298
},
{
"epoch": 0.9416455237404857,
"grad_norm": 2.0914992940185724,
"learning_rate": 1.761601540151736e-07,
"loss": 1.0177,
"step": 1299
},
{
"epoch": 0.9423704240666908,
"grad_norm": 2.4468777213087205,
"learning_rate": 1.7179625005364076e-07,
"loss": 1.072,
"step": 1300
},
{
"epoch": 0.943095324392896,
"grad_norm": 1.982397014057623,
"learning_rate": 1.6748660997694454e-07,
"loss": 0.862,
"step": 1301
},
{
"epoch": 0.9438202247191011,
"grad_norm": 2.4635792769761005,
"learning_rate": 1.6323125757964799e-07,
"loss": 0.9996,
"step": 1302
},
{
"epoch": 0.9445451250453063,
"grad_norm": 2.1226326530971704,
"learning_rate": 1.5903021635658044e-07,
"loss": 1.0194,
"step": 1303
},
{
"epoch": 0.9452700253715114,
"grad_norm": 2.2271194959547116,
"learning_rate": 1.5488350950270237e-07,
"loss": 0.9331,
"step": 1304
},
{
"epoch": 0.9459949256977166,
"grad_norm": 2.472581051188092,
"learning_rate": 1.507911599129841e-07,
"loss": 0.9543,
"step": 1305
},
{
"epoch": 0.9467198260239217,
"grad_norm": 1.872694939056278,
"learning_rate": 1.4675319018227608e-07,
"loss": 1.0782,
"step": 1306
},
{
"epoch": 0.9474447263501269,
"grad_norm": 2.0981456286626297,
"learning_rate": 1.4276962260518223e-07,
"loss": 0.9838,
"step": 1307
},
{
"epoch": 0.948169626676332,
"grad_norm": 2.1938985740167483,
"learning_rate": 1.3884047917594236e-07,
"loss": 1.0023,
"step": 1308
},
{
"epoch": 0.9488945270025372,
"grad_norm": 2.342040537819885,
"learning_rate": 1.349657815883032e-07,
"loss": 0.9773,
"step": 1309
},
{
"epoch": 0.9496194273287423,
"grad_norm": 2.0373395983000795,
"learning_rate": 1.3114555123540762e-07,
"loss": 0.9412,
"step": 1310
},
{
"epoch": 0.9503443276549475,
"grad_norm": 2.3706302014168714,
"learning_rate": 1.2737980920966785e-07,
"loss": 1.0296,
"step": 1311
},
{
"epoch": 0.9510692279811526,
"grad_norm": 2.130493636831118,
"learning_rate": 1.236685763026546e-07,
"loss": 0.9575,
"step": 1312
},
{
"epoch": 0.9517941283073578,
"grad_norm": 2.257925035976561,
"learning_rate": 1.2001187300498153e-07,
"loss": 0.9083,
"step": 1313
},
{
"epoch": 0.9525190286335629,
"grad_norm": 2.1864022493275783,
"learning_rate": 1.1640971950618752e-07,
"loss": 0.923,
"step": 1314
},
{
"epoch": 0.9532439289597681,
"grad_norm": 2.5860606750284734,
"learning_rate": 1.1286213569463467e-07,
"loss": 1.0475,
"step": 1315
},
{
"epoch": 0.9539688292859732,
"grad_norm": 2.258441636325009,
"learning_rate": 1.0936914115738717e-07,
"loss": 0.8783,
"step": 1316
},
{
"epoch": 0.9546937296121784,
"grad_norm": 1.8730597641684372,
"learning_rate": 1.0593075518011143e-07,
"loss": 0.9545,
"step": 1317
},
{
"epoch": 0.9554186299383834,
"grad_norm": 2.3138211525804584,
"learning_rate": 1.0254699674696611e-07,
"loss": 1.016,
"step": 1318
},
{
"epoch": 0.9561435302645886,
"grad_norm": 1.9117655582358168,
"learning_rate": 9.921788454049896e-08,
"loss": 1.0075,
"step": 1319
},
{
"epoch": 0.9568684305907937,
"grad_norm": 2.0487380925988368,
"learning_rate": 9.594343694154129e-08,
"loss": 0.9545,
"step": 1320
},
{
"epoch": 0.9575933309169989,
"grad_norm": 1.808692209003071,
"learning_rate": 9.272367202910804e-08,
"loss": 0.8925,
"step": 1321
},
{
"epoch": 0.958318231243204,
"grad_norm": 2.7015235278624194,
"learning_rate": 8.9558607580299e-08,
"loss": 0.9318,
"step": 1322
},
{
"epoch": 0.9590431315694092,
"grad_norm": 2.2921044642024757,
"learning_rate": 8.644826107019888e-08,
"loss": 1.0238,
"step": 1323
},
{
"epoch": 0.9597680318956143,
"grad_norm": 1.9280021670942513,
"learning_rate": 8.339264967178074e-08,
"loss": 0.971,
"step": 1324
},
{
"epoch": 0.9604929322218195,
"grad_norm": 1.8067867464540794,
"learning_rate": 8.039179025581379e-08,
"loss": 0.8251,
"step": 1325
},
{
"epoch": 0.9612178325480246,
"grad_norm": 2.7036098510140016,
"learning_rate": 7.744569939076574e-08,
"loss": 1.0214,
"step": 1326
},
{
"epoch": 0.9619427328742298,
"grad_norm": 1.7075937015114786,
"learning_rate": 7.455439334271619e-08,
"loss": 0.9234,
"step": 1327
},
{
"epoch": 0.9626676332004349,
"grad_norm": 1.8777145955794914,
"learning_rate": 7.171788807526336e-08,
"loss": 0.8121,
"step": 1328
},
{
"epoch": 0.9633925335266401,
"grad_norm": 2.0244110553223695,
"learning_rate": 6.89361992494375e-08,
"loss": 0.9208,
"step": 1329
},
{
"epoch": 0.9641174338528452,
"grad_norm": 2.542844240335413,
"learning_rate": 6.62093422236132e-08,
"loss": 1.0046,
"step": 1330
},
{
"epoch": 0.9648423341790504,
"grad_norm": 2.543683044105189,
"learning_rate": 6.353733205342715e-08,
"loss": 1.0122,
"step": 1331
},
{
"epoch": 0.9655672345052555,
"grad_norm": 2.1697924274060245,
"learning_rate": 6.092018349169171e-08,
"loss": 1.0031,
"step": 1332
},
{
"epoch": 0.9662921348314607,
"grad_norm": 2.0531262536255985,
"learning_rate": 5.8357910988317e-08,
"loss": 0.8397,
"step": 1333
},
{
"epoch": 0.9670170351576658,
"grad_norm": 2.913557846619087,
"learning_rate": 5.585052869022556e-08,
"loss": 0.9297,
"step": 1334
},
{
"epoch": 0.967741935483871,
"grad_norm": 2.204845207370084,
"learning_rate": 5.339805044128121e-08,
"loss": 1.003,
"step": 1335
},
{
"epoch": 0.9684668358100761,
"grad_norm": 2.644048528912831,
"learning_rate": 5.1000489782210284e-08,
"loss": 1.0444,
"step": 1336
},
{
"epoch": 0.9691917361362813,
"grad_norm": 2.402838017572579,
"learning_rate": 4.8657859950520524e-08,
"loss": 0.9235,
"step": 1337
},
{
"epoch": 0.9699166364624864,
"grad_norm": 1.9654858262351746,
"learning_rate": 4.6370173880438964e-08,
"loss": 0.958,
"step": 1338
},
{
"epoch": 0.9706415367886916,
"grad_norm": 2.7707673392332746,
"learning_rate": 4.413744420283195e-08,
"loss": 0.8956,
"step": 1339
},
{
"epoch": 0.9713664371148967,
"grad_norm": 1.9709988493814223,
"learning_rate": 4.195968324513744e-08,
"loss": 0.9951,
"step": 1340
},
{
"epoch": 0.9720913374411019,
"grad_norm": 2.1055256777177958,
"learning_rate": 3.983690303130172e-08,
"loss": 0.9204,
"step": 1341
},
{
"epoch": 0.972816237767307,
"grad_norm": 2.218819032192361,
"learning_rate": 3.776911528170391e-08,
"loss": 0.8971,
"step": 1342
},
{
"epoch": 0.9735411380935122,
"grad_norm": 2.0858300430985146,
"learning_rate": 3.575633141310153e-08,
"loss": 0.897,
"step": 1343
},
{
"epoch": 0.9742660384197173,
"grad_norm": 2.3972281459872713,
"learning_rate": 3.379856253855951e-08,
"loss": 0.9454,
"step": 1344
},
{
"epoch": 0.9749909387459225,
"grad_norm": 2.488073794270114,
"learning_rate": 3.1895819467391286e-08,
"loss": 0.8999,
"step": 1345
},
{
"epoch": 0.9757158390721276,
"grad_norm": 1.988995529917583,
"learning_rate": 3.0048112705102215e-08,
"loss": 0.971,
"step": 1346
},
{
"epoch": 0.9764407393983328,
"grad_norm": 2.077919130586888,
"learning_rate": 2.8255452453326282e-08,
"loss": 1.003,
"step": 1347
},
{
"epoch": 0.9771656397245378,
"grad_norm": 2.028713109556211,
"learning_rate": 2.65178486097728e-08,
"loss": 1.049,
"step": 1348
},
{
"epoch": 0.977890540050743,
"grad_norm": 1.6908367188168048,
"learning_rate": 2.4835310768172026e-08,
"loss": 0.8732,
"step": 1349
},
{
"epoch": 0.9786154403769481,
"grad_norm": 2.4804402514448127,
"learning_rate": 2.3207848218222974e-08,
"loss": 0.9352,
"step": 1350
},
{
"epoch": 0.9793403407031533,
"grad_norm": 2.1542622951441857,
"learning_rate": 2.163546994553789e-08,
"loss": 0.9104,
"step": 1351
},
{
"epoch": 0.9800652410293584,
"grad_norm": 2.527194109345364,
"learning_rate": 2.011818463159676e-08,
"loss": 1.0832,
"step": 1352
},
{
"epoch": 0.9807901413555636,
"grad_norm": 2.317833340426504,
"learning_rate": 1.8656000653698436e-08,
"loss": 1.0094,
"step": 1353
},
{
"epoch": 0.9815150416817687,
"grad_norm": 2.343407752981275,
"learning_rate": 1.724892608491291e-08,
"loss": 1.0245,
"step": 1354
},
{
"epoch": 0.9822399420079739,
"grad_norm": 1.8989849927794842,
"learning_rate": 1.5896968694040228e-08,
"loss": 0.9417,
"step": 1355
},
{
"epoch": 0.982964842334179,
"grad_norm": 2.770976661865063,
"learning_rate": 1.4600135945560534e-08,
"loss": 0.99,
"step": 1356
},
{
"epoch": 0.9836897426603842,
"grad_norm": 1.7947478635260772,
"learning_rate": 1.3358434999602987e-08,
"loss": 0.892,
"step": 1357
},
{
"epoch": 0.9844146429865893,
"grad_norm": 2.310601190306124,
"learning_rate": 1.2171872711895794e-08,
"loss": 1.0248,
"step": 1358
},
{
"epoch": 0.9851395433127945,
"grad_norm": 2.287100614820235,
"learning_rate": 1.1040455633738457e-08,
"loss": 0.9974,
"step": 1359
},
{
"epoch": 0.9858644436389996,
"grad_norm": 2.107445997393385,
"learning_rate": 9.964190011955144e-09,
"loss": 0.9899,
"step": 1360
},
{
"epoch": 0.9865893439652048,
"grad_norm": 2.447737679629097,
"learning_rate": 8.943081788869157e-09,
"loss": 0.9687,
"step": 1361
},
{
"epoch": 0.9873142442914099,
"grad_norm": 2.151832291830881,
"learning_rate": 7.97713660226629e-09,
"loss": 1.0199,
"step": 1362
},
{
"epoch": 0.9880391446176151,
"grad_norm": 2.0385527802522976,
"learning_rate": 7.066359785362631e-09,
"loss": 1.021,
"step": 1363
},
{
"epoch": 0.9887640449438202,
"grad_norm": 2.3197757332833415,
"learning_rate": 6.210756366775705e-09,
"loss": 0.9591,
"step": 1364
},
{
"epoch": 0.9894889452700254,
"grad_norm": 2.3203612246956924,
"learning_rate": 5.410331070498931e-09,
"loss": 0.9674,
"step": 1365
},
{
"epoch": 0.9902138455962305,
"grad_norm": 1.8262826260586555,
"learning_rate": 4.665088315874977e-09,
"loss": 1.0296,
"step": 1366
},
{
"epoch": 0.9909387459224357,
"grad_norm": 1.961319160005979,
"learning_rate": 3.975032217566899e-09,
"loss": 0.9492,
"step": 1367
},
{
"epoch": 0.9916636462486408,
"grad_norm": 1.8126619362913512,
"learning_rate": 3.3401665855414824e-09,
"loss": 0.8814,
"step": 1368
},
{
"epoch": 0.992388546574846,
"grad_norm": 2.0882014952836916,
"learning_rate": 2.760494925045931e-09,
"loss": 0.8576,
"step": 1369
},
{
"epoch": 0.9931134469010511,
"grad_norm": 2.2450600007272037,
"learning_rate": 2.236020436586772e-09,
"loss": 0.9932,
"step": 1370
},
{
"epoch": 0.9938383472272563,
"grad_norm": 2.44028976343664,
"learning_rate": 1.7667460159143113e-09,
"loss": 0.8992,
"step": 1371
},
{
"epoch": 0.9945632475534614,
"grad_norm": 2.0146202920467293,
"learning_rate": 1.3526742540070913e-09,
"loss": 1.006,
"step": 1372
},
{
"epoch": 0.9952881478796666,
"grad_norm": 1.9601240017014594,
"learning_rate": 9.938074370530182e-10,
"loss": 0.9297,
"step": 1373
},
{
"epoch": 0.9960130482058717,
"grad_norm": 1.8087006446044052,
"learning_rate": 6.90147546443809e-10,
"loss": 0.8517,
"step": 1374
},
{
"epoch": 0.9967379485320769,
"grad_norm": 2.3390439919559936,
"learning_rate": 4.416962587583395e-10,
"loss": 0.9555,
"step": 1375
},
{
"epoch": 0.997462848858282,
"grad_norm": 1.9176591785146089,
"learning_rate": 2.484549457559826e-10,
"loss": 0.9623,
"step": 1376
},
{
"epoch": 0.9981877491844872,
"grad_norm": 2.215362657603315,
"learning_rate": 1.1042467436994664e-10,
"loss": 0.8946,
"step": 1377
},
{
"epoch": 0.9989126495106923,
"grad_norm": 2.000196078224186,
"learning_rate": 2.7606206697283755e-11,
"loss": 0.9503,
"step": 1378
},
{
"epoch": 0.9996375498368975,
"grad_norm": 2.2467269541867663,
"learning_rate": 0.0,
"loss": 0.9984,
"step": 1379
},
{
"epoch": 0.9996375498368975,
"step": 1379,
"total_flos": 1.3996327755055104e+16,
"train_loss": 0.49300233914940317,
"train_runtime": 19998.664,
"train_samples_per_second": 8.828,
"train_steps_per_second": 0.069
}
],
"logging_steps": 1.0,
"max_steps": 1379,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 345,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.3996327755055104e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}