{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 1250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0032, "grad_norm": 5.2346367835998535, "learning_rate": 3.1746031746031746e-06, "loss": 5.8721, "step": 2 }, { "epoch": 0.0064, "grad_norm": 6.022397994995117, "learning_rate": 6.349206349206349e-06, "loss": 5.775, "step": 4 }, { "epoch": 0.0096, "grad_norm": 4.282025337219238, "learning_rate": 9.523809523809523e-06, "loss": 5.6253, "step": 6 }, { "epoch": 0.0128, "grad_norm": 3.7371106147766113, "learning_rate": 1.2698412698412699e-05, "loss": 5.1153, "step": 8 }, { "epoch": 0.016, "grad_norm": 8.722667694091797, "learning_rate": 1.5873015873015872e-05, "loss": 5.3364, "step": 10 }, { "epoch": 0.0192, "grad_norm": 4.185220718383789, "learning_rate": 1.9047619047619046e-05, "loss": 5.3102, "step": 12 }, { "epoch": 0.0224, "grad_norm": 3.5511889457702637, "learning_rate": 2.2222222222222223e-05, "loss": 4.9739, "step": 14 }, { "epoch": 0.0256, "grad_norm": 3.152355194091797, "learning_rate": 2.5396825396825397e-05, "loss": 4.5878, "step": 16 }, { "epoch": 0.0288, "grad_norm": 2.925100326538086, "learning_rate": 2.857142857142857e-05, "loss": 4.7562, "step": 18 }, { "epoch": 0.032, "grad_norm": 3.192331552505493, "learning_rate": 3.1746031746031745e-05, "loss": 4.6627, "step": 20 }, { "epoch": 0.0352, "grad_norm": 2.958089590072632, "learning_rate": 3.492063492063492e-05, "loss": 4.6227, "step": 22 }, { "epoch": 0.0384, "grad_norm": 4.2280592918396, "learning_rate": 3.809523809523809e-05, "loss": 4.4283, "step": 24 }, { "epoch": 0.0416, "grad_norm": 2.748825788497925, "learning_rate": 4.126984126984127e-05, "loss": 4.2503, "step": 26 }, { "epoch": 0.0448, "grad_norm": 2.705292224884033, "learning_rate": 4.4444444444444447e-05, "loss": 4.4007, "step": 28 }, { "epoch": 0.048, "grad_norm": 2.5890371799468994, "learning_rate": 4.761904761904762e-05, "loss": 4.3219, "step": 30 }, { "epoch": 0.0512, "grad_norm": 2.485609769821167, "learning_rate": 5.0793650793650794e-05, "loss": 4.2298, "step": 32 }, { "epoch": 0.0544, "grad_norm": 2.5438082218170166, "learning_rate": 5.396825396825397e-05, "loss": 4.2862, "step": 34 }, { "epoch": 0.0576, "grad_norm": 2.5402183532714844, "learning_rate": 5.714285714285714e-05, "loss": 4.0492, "step": 36 }, { "epoch": 0.0608, "grad_norm": 2.5497469902038574, "learning_rate": 6.0317460317460316e-05, "loss": 4.0702, "step": 38 }, { "epoch": 0.064, "grad_norm": 2.3628146648406982, "learning_rate": 6.349206349206349e-05, "loss": 3.9895, "step": 40 }, { "epoch": 0.0672, "grad_norm": 2.201223611831665, "learning_rate": 6.666666666666667e-05, "loss": 3.9743, "step": 42 }, { "epoch": 0.0704, "grad_norm": 2.448514938354492, "learning_rate": 6.984126984126984e-05, "loss": 4.076, "step": 44 }, { "epoch": 0.0736, "grad_norm": 2.386176824569702, "learning_rate": 7.301587301587302e-05, "loss": 3.912, "step": 46 }, { "epoch": 0.0768, "grad_norm": 2.2783148288726807, "learning_rate": 7.619047619047618e-05, "loss": 3.8678, "step": 48 }, { "epoch": 0.08, "grad_norm": 2.472463369369507, "learning_rate": 7.936507936507937e-05, "loss": 3.9058, "step": 50 }, { "epoch": 0.0832, "grad_norm": 2.548892021179199, "learning_rate": 8.253968253968255e-05, "loss": 3.7761, "step": 52 }, { "epoch": 0.0864, "grad_norm": 2.4397833347320557, "learning_rate": 8.571428571428571e-05, "loss": 3.9437, "step": 54 }, { "epoch": 0.0896, "grad_norm": 2.532597303390503, "learning_rate": 8.888888888888889e-05, "loss": 3.8437, "step": 56 }, { "epoch": 0.0928, "grad_norm": 2.45221209526062, "learning_rate": 9.206349206349206e-05, "loss": 3.8429, "step": 58 }, { "epoch": 0.096, "grad_norm": 2.645132064819336, "learning_rate": 9.523809523809524e-05, "loss": 3.7965, "step": 60 }, { "epoch": 0.0992, "grad_norm": 2.385370969772339, "learning_rate": 9.841269841269841e-05, "loss": 3.7968, "step": 62 }, { "epoch": 0.1024, "grad_norm": 2.2052712440490723, "learning_rate": 9.99998248790669e-05, "loss": 3.7794, "step": 64 }, { "epoch": 0.1056, "grad_norm": 2.3219361305236816, "learning_rate": 9.999842391896222e-05, "loss": 3.6841, "step": 66 }, { "epoch": 0.1088, "grad_norm": 2.5465614795684814, "learning_rate": 9.999562203800676e-05, "loss": 3.5312, "step": 68 }, { "epoch": 0.112, "grad_norm": 2.497755527496338, "learning_rate": 9.999141931470729e-05, "loss": 3.6803, "step": 70 }, { "epoch": 0.1152, "grad_norm": 2.157752752304077, "learning_rate": 9.998581586682116e-05, "loss": 3.5856, "step": 72 }, { "epoch": 0.1184, "grad_norm": 2.1215410232543945, "learning_rate": 9.997881185135307e-05, "loss": 3.5429, "step": 74 }, { "epoch": 0.1216, "grad_norm": 3.784062385559082, "learning_rate": 9.997040746455062e-05, "loss": 3.6365, "step": 76 }, { "epoch": 0.1248, "grad_norm": 2.2366509437561035, "learning_rate": 9.996060294189887e-05, "loss": 3.5129, "step": 78 }, { "epoch": 0.128, "grad_norm": 2.240877628326416, "learning_rate": 9.994939855811362e-05, "loss": 3.6185, "step": 80 }, { "epoch": 0.1312, "grad_norm": 2.176579475402832, "learning_rate": 9.993679462713395e-05, "loss": 3.4999, "step": 82 }, { "epoch": 0.1344, "grad_norm": 2.3514060974121094, "learning_rate": 9.992279150211314e-05, "loss": 3.466, "step": 84 }, { "epoch": 0.1376, "grad_norm": 2.2598917484283447, "learning_rate": 9.990738957540896e-05, "loss": 3.6413, "step": 86 }, { "epoch": 0.1408, "grad_norm": 2.0476276874542236, "learning_rate": 9.989058927857263e-05, "loss": 3.5978, "step": 88 }, { "epoch": 0.144, "grad_norm": 2.1762521266937256, "learning_rate": 9.987239108233668e-05, "loss": 3.6667, "step": 90 }, { "epoch": 0.1472, "grad_norm": 2.274958848953247, "learning_rate": 9.985279549660185e-05, "loss": 3.6054, "step": 92 }, { "epoch": 0.1504, "grad_norm": 2.249992847442627, "learning_rate": 9.983180307042274e-05, "loss": 3.6287, "step": 94 }, { "epoch": 0.1536, "grad_norm": 2.23592209815979, "learning_rate": 9.980941439199246e-05, "loss": 3.5967, "step": 96 }, { "epoch": 0.1568, "grad_norm": 2.1270549297332764, "learning_rate": 9.97856300886261e-05, "loss": 3.4583, "step": 98 }, { "epoch": 0.16, "grad_norm": 2.140577554702759, "learning_rate": 9.976045082674319e-05, "loss": 3.4091, "step": 100 }, { "epoch": 0.1632, "grad_norm": 2.1698827743530273, "learning_rate": 9.973387731184902e-05, "loss": 3.5535, "step": 102 }, { "epoch": 0.1664, "grad_norm": 2.188966751098633, "learning_rate": 9.97059102885149e-05, "loss": 3.4673, "step": 104 }, { "epoch": 0.1696, "grad_norm": 2.015054702758789, "learning_rate": 9.967655054035727e-05, "loss": 3.5025, "step": 106 }, { "epoch": 0.1728, "grad_norm": 2.42785906791687, "learning_rate": 9.964579889001569e-05, "loss": 3.3789, "step": 108 }, { "epoch": 0.176, "grad_norm": 3.828245162963867, "learning_rate": 9.961365619912989e-05, "loss": 3.3673, "step": 110 }, { "epoch": 0.1792, "grad_norm": 2.0954813957214355, "learning_rate": 9.95801233683156e-05, "loss": 3.501, "step": 112 }, { "epoch": 0.1824, "grad_norm": 2.17081618309021, "learning_rate": 9.954520133713924e-05, "loss": 3.3926, "step": 114 }, { "epoch": 0.1856, "grad_norm": 2.04852557182312, "learning_rate": 9.950889108409172e-05, "loss": 3.391, "step": 116 }, { "epoch": 0.1888, "grad_norm": 2.426689386367798, "learning_rate": 9.947119362656092e-05, "loss": 3.4257, "step": 118 }, { "epoch": 0.192, "grad_norm": 3.680421829223633, "learning_rate": 9.94321100208032e-05, "loss": 3.2982, "step": 120 }, { "epoch": 0.1952, "grad_norm": 2.1409482955932617, "learning_rate": 9.939164136191384e-05, "loss": 3.4619, "step": 122 }, { "epoch": 0.1984, "grad_norm": 1.9399126768112183, "learning_rate": 9.934978878379636e-05, "loss": 3.3362, "step": 124 }, { "epoch": 0.2016, "grad_norm": 1.954500675201416, "learning_rate": 9.930655345913071e-05, "loss": 3.1957, "step": 126 }, { "epoch": 0.2048, "grad_norm": 2.1550300121307373, "learning_rate": 9.926193659934043e-05, "loss": 3.4578, "step": 128 }, { "epoch": 0.208, "grad_norm": 2.44838547706604, "learning_rate": 9.921593945455869e-05, "loss": 3.3975, "step": 130 }, { "epoch": 0.2112, "grad_norm": 2.087881565093994, "learning_rate": 9.916856331359335e-05, "loss": 3.3682, "step": 132 }, { "epoch": 0.2144, "grad_norm": 2.253127336502075, "learning_rate": 9.911980950389067e-05, "loss": 3.2451, "step": 134 }, { "epoch": 0.2176, "grad_norm": 2.3103411197662354, "learning_rate": 9.906967939149831e-05, "loss": 3.3999, "step": 136 }, { "epoch": 0.2208, "grad_norm": 2.2471373081207275, "learning_rate": 9.901817438102695e-05, "loss": 3.2925, "step": 138 }, { "epoch": 0.224, "grad_norm": 2.1333861351013184, "learning_rate": 9.896529591561093e-05, "loss": 3.3604, "step": 140 }, { "epoch": 0.2272, "grad_norm": 1.914016842842102, "learning_rate": 9.891104547686782e-05, "loss": 3.1728, "step": 142 }, { "epoch": 0.2304, "grad_norm": 2.061126232147217, "learning_rate": 9.8855424584857e-05, "loss": 3.2384, "step": 144 }, { "epoch": 0.2336, "grad_norm": 2.153684139251709, "learning_rate": 9.879843479803691e-05, "loss": 3.2965, "step": 146 }, { "epoch": 0.2368, "grad_norm": 2.109224557876587, "learning_rate": 9.874007771322151e-05, "loss": 3.2568, "step": 148 }, { "epoch": 0.24, "grad_norm": 3.588501214981079, "learning_rate": 9.868035496553546e-05, "loss": 3.223, "step": 150 }, { "epoch": 0.2432, "grad_norm": 2.2085981369018555, "learning_rate": 9.86192682283684e-05, "loss": 3.3506, "step": 152 }, { "epoch": 0.2464, "grad_norm": 1.9722400903701782, "learning_rate": 9.855681921332793e-05, "loss": 3.2184, "step": 154 }, { "epoch": 0.2496, "grad_norm": 2.0837275981903076, "learning_rate": 9.849300967019175e-05, "loss": 3.2083, "step": 156 }, { "epoch": 0.2528, "grad_norm": 1.799812912940979, "learning_rate": 9.84278413868586e-05, "loss": 3.2038, "step": 158 }, { "epoch": 0.256, "grad_norm": 2.0242528915405273, "learning_rate": 9.836131618929819e-05, "loss": 3.3193, "step": 160 }, { "epoch": 0.2592, "grad_norm": 1.948160171508789, "learning_rate": 9.82934359415e-05, "loss": 3.2132, "step": 162 }, { "epoch": 0.2624, "grad_norm": 2.1192405223846436, "learning_rate": 9.822420254542108e-05, "loss": 3.2444, "step": 164 }, { "epoch": 0.2656, "grad_norm": 1.9564788341522217, "learning_rate": 9.815361794093272e-05, "loss": 3.105, "step": 166 }, { "epoch": 0.2688, "grad_norm": 2.1899731159210205, "learning_rate": 9.808168410576617e-05, "loss": 3.1558, "step": 168 }, { "epoch": 0.272, "grad_norm": 2.0988922119140625, "learning_rate": 9.800840305545715e-05, "loss": 3.2485, "step": 170 }, { "epoch": 0.2752, "grad_norm": 2.2193591594696045, "learning_rate": 9.793377684328939e-05, "loss": 3.296, "step": 172 }, { "epoch": 0.2784, "grad_norm": 2.007413864135742, "learning_rate": 9.785780756023714e-05, "loss": 3.1287, "step": 174 }, { "epoch": 0.2816, "grad_norm": 1.9749376773834229, "learning_rate": 9.778049733490655e-05, "loss": 3.0076, "step": 176 }, { "epoch": 0.2848, "grad_norm": 2.059288263320923, "learning_rate": 9.770184833347606e-05, "loss": 3.1663, "step": 178 }, { "epoch": 0.288, "grad_norm": 1.96829354763031, "learning_rate": 9.762186275963563e-05, "loss": 3.2163, "step": 180 }, { "epoch": 0.2912, "grad_norm": 2.0208017826080322, "learning_rate": 9.754054285452506e-05, "loss": 3.2242, "step": 182 }, { "epoch": 0.2944, "grad_norm": 1.88913094997406, "learning_rate": 9.745789089667121e-05, "loss": 3.0072, "step": 184 }, { "epoch": 0.2976, "grad_norm": 1.996383547782898, "learning_rate": 9.737390920192408e-05, "loss": 3.2108, "step": 186 }, { "epoch": 0.3008, "grad_norm": 2.0667550563812256, "learning_rate": 9.7288600123392e-05, "loss": 3.108, "step": 188 }, { "epoch": 0.304, "grad_norm": 1.9526984691619873, "learning_rate": 9.720196605137565e-05, "loss": 3.067, "step": 190 }, { "epoch": 0.3072, "grad_norm": 2.0643012523651123, "learning_rate": 9.71140094133011e-05, "loss": 3.2143, "step": 192 }, { "epoch": 0.3104, "grad_norm": 2.187326431274414, "learning_rate": 9.702473267365182e-05, "loss": 3.1007, "step": 194 }, { "epoch": 0.3136, "grad_norm": 2.1299145221710205, "learning_rate": 9.693413833389956e-05, "loss": 3.1868, "step": 196 }, { "epoch": 0.3168, "grad_norm": 1.7644037008285522, "learning_rate": 9.684222893243431e-05, "loss": 2.9406, "step": 198 }, { "epoch": 0.32, "grad_norm": 2.112617015838623, "learning_rate": 9.674900704449324e-05, "loss": 3.1198, "step": 200 }, { "epoch": 0.3232, "grad_norm": 1.8327059745788574, "learning_rate": 9.665447528208836e-05, "loss": 3.1278, "step": 202 }, { "epoch": 0.3264, "grad_norm": 1.8569375276565552, "learning_rate": 9.655863629393351e-05, "loss": 3.2069, "step": 204 }, { "epoch": 0.3296, "grad_norm": 1.7960104942321777, "learning_rate": 9.64614927653701e-05, "loss": 3.0708, "step": 206 }, { "epoch": 0.3328, "grad_norm": 1.888593316078186, "learning_rate": 9.636304741829181e-05, "loss": 3.1365, "step": 208 }, { "epoch": 0.336, "grad_norm": 1.8564034700393677, "learning_rate": 9.626330301106837e-05, "loss": 3.0059, "step": 210 }, { "epoch": 0.3392, "grad_norm": 1.9591517448425293, "learning_rate": 9.616226233846828e-05, "loss": 2.9778, "step": 212 }, { "epoch": 0.3424, "grad_norm": 2.025777816772461, "learning_rate": 9.605992823158046e-05, "loss": 3.0969, "step": 214 }, { "epoch": 0.3456, "grad_norm": 1.8839352130889893, "learning_rate": 9.595630355773501e-05, "loss": 3.1342, "step": 216 }, { "epoch": 0.3488, "grad_norm": 5.388115882873535, "learning_rate": 9.585139122042274e-05, "loss": 3.1961, "step": 218 }, { "epoch": 0.352, "grad_norm": 2.056678533554077, "learning_rate": 9.574519415921396e-05, "loss": 3.1183, "step": 220 }, { "epoch": 0.3552, "grad_norm": 3.0575530529022217, "learning_rate": 9.5637715349676e-05, "loss": 3.1446, "step": 222 }, { "epoch": 0.3584, "grad_norm": 1.8165247440338135, "learning_rate": 9.552895780328987e-05, "loss": 3.0338, "step": 224 }, { "epoch": 0.3616, "grad_norm": 1.845023512840271, "learning_rate": 9.541892456736595e-05, "loss": 3.194, "step": 226 }, { "epoch": 0.3648, "grad_norm": 1.9389755725860596, "learning_rate": 9.530761872495849e-05, "loss": 3.0054, "step": 228 }, { "epoch": 0.368, "grad_norm": 1.9471769332885742, "learning_rate": 9.519504339477932e-05, "loss": 3.1499, "step": 230 }, { "epoch": 0.3712, "grad_norm": 1.9367070198059082, "learning_rate": 9.508120173111039e-05, "loss": 3.0068, "step": 232 }, { "epoch": 0.3744, "grad_norm": 2.018630027770996, "learning_rate": 9.496609692371548e-05, "loss": 3.1722, "step": 234 }, { "epoch": 0.3776, "grad_norm": 2.0086734294891357, "learning_rate": 9.484973219775074e-05, "loss": 3.2773, "step": 236 }, { "epoch": 0.3808, "grad_norm": 1.9771322011947632, "learning_rate": 9.473211081367436e-05, "loss": 3.0502, "step": 238 }, { "epoch": 0.384, "grad_norm": 1.9231762886047363, "learning_rate": 9.46132360671552e-05, "loss": 3.0415, "step": 240 }, { "epoch": 0.3872, "grad_norm": 1.924302101135254, "learning_rate": 9.449311128898049e-05, "loss": 3.0794, "step": 242 }, { "epoch": 0.3904, "grad_norm": 1.967323899269104, "learning_rate": 9.437173984496246e-05, "loss": 3.1527, "step": 244 }, { "epoch": 0.3936, "grad_norm": 1.9681285619735718, "learning_rate": 9.424912513584401e-05, "loss": 3.1767, "step": 246 }, { "epoch": 0.3968, "grad_norm": 1.908687710762024, "learning_rate": 9.412527059720352e-05, "loss": 2.9755, "step": 248 }, { "epoch": 0.4, "grad_norm": 1.9891773462295532, "learning_rate": 9.400017969935848e-05, "loss": 2.9644, "step": 250 }, { "epoch": 0.4032, "grad_norm": 1.8200337886810303, "learning_rate": 9.387385594726829e-05, "loss": 3.0334, "step": 252 }, { "epoch": 0.4064, "grad_norm": 1.9553104639053345, "learning_rate": 9.374630288043614e-05, "loss": 3.1011, "step": 254 }, { "epoch": 0.4096, "grad_norm": 3.5282905101776123, "learning_rate": 9.361752407280965e-05, "loss": 2.9859, "step": 256 }, { "epoch": 0.4128, "grad_norm": 1.9712797403335571, "learning_rate": 9.348752313268093e-05, "loss": 2.9472, "step": 258 }, { "epoch": 0.416, "grad_norm": 1.927635908126831, "learning_rate": 9.335630370258533e-05, "loss": 3.1396, "step": 260 }, { "epoch": 0.4192, "grad_norm": 1.9233123064041138, "learning_rate": 9.322386945919946e-05, "loss": 3.1889, "step": 262 }, { "epoch": 0.4224, "grad_norm": 1.870160460472107, "learning_rate": 9.309022411323816e-05, "loss": 3.0916, "step": 264 }, { "epoch": 0.4256, "grad_norm": 1.7860538959503174, "learning_rate": 9.295537140935049e-05, "loss": 3.1584, "step": 266 }, { "epoch": 0.4288, "grad_norm": 1.723097801208496, "learning_rate": 9.281931512601485e-05, "loss": 2.8587, "step": 268 }, { "epoch": 0.432, "grad_norm": 1.976706862449646, "learning_rate": 9.26820590754331e-05, "loss": 2.9942, "step": 270 }, { "epoch": 0.4352, "grad_norm": 1.8147152662277222, "learning_rate": 9.254360710342371e-05, "loss": 3.087, "step": 272 }, { "epoch": 0.4384, "grad_norm": 1.8946576118469238, "learning_rate": 9.240396308931407e-05, "loss": 3.0101, "step": 274 }, { "epoch": 0.4416, "grad_norm": 1.8432953357696533, "learning_rate": 9.226313094583173e-05, "loss": 3.0351, "step": 276 }, { "epoch": 0.4448, "grad_norm": 1.8600575923919678, "learning_rate": 9.212111461899479e-05, "loss": 3.0027, "step": 278 }, { "epoch": 0.448, "grad_norm": 1.7912688255310059, "learning_rate": 9.197791808800135e-05, "loss": 3.0568, "step": 280 }, { "epoch": 0.4512, "grad_norm": 2.005932569503784, "learning_rate": 9.183354536511803e-05, "loss": 2.9778, "step": 282 }, { "epoch": 0.4544, "grad_norm": 1.8989531993865967, "learning_rate": 9.168800049556747e-05, "loss": 2.9711, "step": 284 }, { "epoch": 0.4576, "grad_norm": 1.7888331413269043, "learning_rate": 9.154128755741509e-05, "loss": 2.9901, "step": 286 }, { "epoch": 0.4608, "grad_norm": 1.9094816446304321, "learning_rate": 9.139341066145472e-05, "loss": 3.0248, "step": 288 }, { "epoch": 0.464, "grad_norm": 1.7940737009048462, "learning_rate": 9.124437395109353e-05, "loss": 3.0141, "step": 290 }, { "epoch": 0.4672, "grad_norm": 1.7626845836639404, "learning_rate": 9.109418160223585e-05, "loss": 2.9531, "step": 292 }, { "epoch": 0.4704, "grad_norm": 1.9440515041351318, "learning_rate": 9.094283782316619e-05, "loss": 2.9732, "step": 294 }, { "epoch": 0.4736, "grad_norm": 1.7515082359313965, "learning_rate": 9.079034685443133e-05, "loss": 2.8, "step": 296 }, { "epoch": 0.4768, "grad_norm": 1.8595532178878784, "learning_rate": 9.063671296872149e-05, "loss": 2.9873, "step": 298 }, { "epoch": 0.48, "grad_norm": 1.9954842329025269, "learning_rate": 9.048194047075069e-05, "loss": 2.9793, "step": 300 }, { "epoch": 0.4832, "grad_norm": 1.8819364309310913, "learning_rate": 9.032603369713596e-05, "loss": 2.8904, "step": 302 }, { "epoch": 0.4864, "grad_norm": 1.75027596950531, "learning_rate": 9.016899701627604e-05, "loss": 2.9811, "step": 304 }, { "epoch": 0.4896, "grad_norm": 1.9617975950241089, "learning_rate": 9.00108348282288e-05, "loss": 3.0418, "step": 306 }, { "epoch": 0.4928, "grad_norm": 1.8097938299179077, "learning_rate": 8.985155156458811e-05, "loss": 3.0068, "step": 308 }, { "epoch": 0.496, "grad_norm": 2.008989095687866, "learning_rate": 8.969115168835954e-05, "loss": 2.8913, "step": 310 }, { "epoch": 0.4992, "grad_norm": 1.8119149208068848, "learning_rate": 8.952963969383538e-05, "loss": 3.004, "step": 312 }, { "epoch": 0.5024, "grad_norm": 1.9150359630584717, "learning_rate": 8.93670201064687e-05, "loss": 2.9404, "step": 314 }, { "epoch": 0.5056, "grad_norm": 1.8803378343582153, "learning_rate": 8.920329748274649e-05, "loss": 2.9279, "step": 316 }, { "epoch": 0.5088, "grad_norm": 1.7308014631271362, "learning_rate": 8.903847641006218e-05, "loss": 2.9482, "step": 318 }, { "epoch": 0.512, "grad_norm": 2.0764575004577637, "learning_rate": 8.887256150658684e-05, "loss": 2.9425, "step": 320 }, { "epoch": 0.5152, "grad_norm": 1.8219692707061768, "learning_rate": 8.870555742113998e-05, "loss": 3.0176, "step": 322 }, { "epoch": 0.5184, "grad_norm": 1.8642444610595703, "learning_rate": 8.85374688330592e-05, "loss": 3.0474, "step": 324 }, { "epoch": 0.5216, "grad_norm": 1.9277112483978271, "learning_rate": 8.836830045206911e-05, "loss": 2.9487, "step": 326 }, { "epoch": 0.5248, "grad_norm": 1.793232798576355, "learning_rate": 8.81980570181494e-05, "loss": 2.8907, "step": 328 }, { "epoch": 0.528, "grad_norm": 1.8513329029083252, "learning_rate": 8.802674330140192e-05, "loss": 2.9645, "step": 330 }, { "epoch": 0.5312, "grad_norm": 1.7978984117507935, "learning_rate": 8.785436410191714e-05, "loss": 2.9939, "step": 332 }, { "epoch": 0.5344, "grad_norm": 1.7157683372497559, "learning_rate": 8.76809242496396e-05, "loss": 2.8079, "step": 334 }, { "epoch": 0.5376, "grad_norm": 1.7450584173202515, "learning_rate": 8.750642860423262e-05, "loss": 2.9477, "step": 336 }, { "epoch": 0.5408, "grad_norm": 1.812904715538025, "learning_rate": 8.733088205494205e-05, "loss": 2.9842, "step": 338 }, { "epoch": 0.544, "grad_norm": 1.878509759902954, "learning_rate": 8.715428952045936e-05, "loss": 2.8992, "step": 340 }, { "epoch": 0.5472, "grad_norm": 1.8090356588363647, "learning_rate": 8.697665594878382e-05, "loss": 2.9507, "step": 342 }, { "epoch": 0.5504, "grad_norm": 1.8601405620574951, "learning_rate": 8.679798631708375e-05, "loss": 2.8263, "step": 344 }, { "epoch": 0.5536, "grad_norm": 1.8406038284301758, "learning_rate": 8.661828563155727e-05, "loss": 2.8991, "step": 346 }, { "epoch": 0.5568, "grad_norm": 1.7687346935272217, "learning_rate": 8.643755892729179e-05, "loss": 2.8437, "step": 348 }, { "epoch": 0.56, "grad_norm": 1.9318656921386719, "learning_rate": 8.625581126812312e-05, "loss": 3.0261, "step": 350 }, { "epoch": 0.5632, "grad_norm": 1.8095970153808594, "learning_rate": 8.607304774649349e-05, "loss": 2.9269, "step": 352 }, { "epoch": 0.5664, "grad_norm": 1.8424136638641357, "learning_rate": 8.588927348330887e-05, "loss": 2.7917, "step": 354 }, { "epoch": 0.5696, "grad_norm": 1.9472522735595703, "learning_rate": 8.57044936277955e-05, "loss": 2.7557, "step": 356 }, { "epoch": 0.5728, "grad_norm": 1.8774663209915161, "learning_rate": 8.551871335735565e-05, "loss": 2.8426, "step": 358 }, { "epoch": 0.576, "grad_norm": 2.0363616943359375, "learning_rate": 8.533193787742251e-05, "loss": 2.8605, "step": 360 }, { "epoch": 0.5792, "grad_norm": 1.802148699760437, "learning_rate": 8.51441724213143e-05, "loss": 2.8999, "step": 362 }, { "epoch": 0.5824, "grad_norm": 1.9002522230148315, "learning_rate": 8.495542225008771e-05, "loss": 2.854, "step": 364 }, { "epoch": 0.5856, "grad_norm": 1.715409755706787, "learning_rate": 8.476569265239046e-05, "loss": 2.8574, "step": 366 }, { "epoch": 0.5888, "grad_norm": 1.9336328506469727, "learning_rate": 8.457498894431311e-05, "loss": 2.7513, "step": 368 }, { "epoch": 0.592, "grad_norm": 1.8150614500045776, "learning_rate": 8.438331646924013e-05, "loss": 2.8648, "step": 370 }, { "epoch": 0.5952, "grad_norm": 1.745450735092163, "learning_rate": 8.419068059770011e-05, "loss": 2.8102, "step": 372 }, { "epoch": 0.5984, "grad_norm": 1.7649202346801758, "learning_rate": 8.399708672721539e-05, "loss": 2.943, "step": 374 }, { "epoch": 0.6016, "grad_norm": 1.9029461145401, "learning_rate": 8.380254028215076e-05, "loss": 2.9549, "step": 376 }, { "epoch": 0.6048, "grad_norm": 1.6569948196411133, "learning_rate": 8.360704671356145e-05, "loss": 2.762, "step": 378 }, { "epoch": 0.608, "grad_norm": 1.8082654476165771, "learning_rate": 8.341061149904045e-05, "loss": 2.8673, "step": 380 }, { "epoch": 0.6112, "grad_norm": 2.0130746364593506, "learning_rate": 8.321324014256504e-05, "loss": 2.8208, "step": 382 }, { "epoch": 0.6144, "grad_norm": 1.9243208169937134, "learning_rate": 8.30149381743425e-05, "loss": 2.8175, "step": 384 }, { "epoch": 0.6176, "grad_norm": 1.7602218389511108, "learning_rate": 8.28157111506552e-05, "loss": 2.8133, "step": 386 }, { "epoch": 0.6208, "grad_norm": 1.9011287689208984, "learning_rate": 8.261556465370493e-05, "loss": 2.915, "step": 388 }, { "epoch": 0.624, "grad_norm": 1.8260935544967651, "learning_rate": 8.24145042914565e-05, "loss": 2.7879, "step": 390 }, { "epoch": 0.6272, "grad_norm": 2.0104498863220215, "learning_rate": 8.221253569748055e-05, "loss": 2.9628, "step": 392 }, { "epoch": 0.6304, "grad_norm": 1.821583867073059, "learning_rate": 8.200966453079575e-05, "loss": 2.8391, "step": 394 }, { "epoch": 0.6336, "grad_norm": 1.8291980028152466, "learning_rate": 8.180589647571023e-05, "loss": 2.8921, "step": 396 }, { "epoch": 0.6368, "grad_norm": 1.8733659982681274, "learning_rate": 8.16012372416623e-05, "loss": 2.8988, "step": 398 }, { "epoch": 0.64, "grad_norm": 1.9064126014709473, "learning_rate": 8.13956925630605e-05, "loss": 2.8047, "step": 400 }, { "epoch": 0.6432, "grad_norm": 1.7694967985153198, "learning_rate": 8.118926819912287e-05, "loss": 2.8175, "step": 402 }, { "epoch": 0.6464, "grad_norm": 1.9278019666671753, "learning_rate": 8.098196993371565e-05, "loss": 2.7233, "step": 404 }, { "epoch": 0.6496, "grad_norm": 1.7496165037155151, "learning_rate": 8.077380357519115e-05, "loss": 2.8196, "step": 406 }, { "epoch": 0.6528, "grad_norm": 1.8212895393371582, "learning_rate": 8.056477495622511e-05, "loss": 2.8749, "step": 408 }, { "epoch": 0.656, "grad_norm": 1.8443467617034912, "learning_rate": 8.035488993365312e-05, "loss": 2.8373, "step": 410 }, { "epoch": 0.6592, "grad_norm": 1.804992437362671, "learning_rate": 8.014415438830667e-05, "loss": 2.8248, "step": 412 }, { "epoch": 0.6624, "grad_norm": 1.7744520902633667, "learning_rate": 7.993257422484826e-05, "loss": 2.8403, "step": 414 }, { "epoch": 0.6656, "grad_norm": 1.9854934215545654, "learning_rate": 7.972015537160602e-05, "loss": 2.9588, "step": 416 }, { "epoch": 0.6688, "grad_norm": 1.7830610275268555, "learning_rate": 7.950690378040758e-05, "loss": 2.7737, "step": 418 }, { "epoch": 0.672, "grad_norm": 1.8114231824874878, "learning_rate": 7.929282542641325e-05, "loss": 2.7003, "step": 420 }, { "epoch": 0.6752, "grad_norm": 1.931700348854065, "learning_rate": 7.907792630794876e-05, "loss": 2.8088, "step": 422 }, { "epoch": 0.6784, "grad_norm": 1.8684518337249756, "learning_rate": 7.886221244633703e-05, "loss": 2.878, "step": 424 }, { "epoch": 0.6816, "grad_norm": 1.8997987508773804, "learning_rate": 7.864568988572947e-05, "loss": 2.9374, "step": 426 }, { "epoch": 0.6848, "grad_norm": 1.7682809829711914, "learning_rate": 7.842836469293673e-05, "loss": 2.7694, "step": 428 }, { "epoch": 0.688, "grad_norm": 1.8019146919250488, "learning_rate": 7.821024295725865e-05, "loss": 2.8153, "step": 430 }, { "epoch": 0.6912, "grad_norm": 1.8119292259216309, "learning_rate": 7.79913307903136e-05, "loss": 2.8072, "step": 432 }, { "epoch": 0.6944, "grad_norm": 1.8016608953475952, "learning_rate": 7.777163432586734e-05, "loss": 2.7276, "step": 434 }, { "epoch": 0.6976, "grad_norm": 1.8160144090652466, "learning_rate": 7.755115971966104e-05, "loss": 2.8539, "step": 436 }, { "epoch": 0.7008, "grad_norm": 1.825020670890808, "learning_rate": 7.732991314923891e-05, "loss": 2.7796, "step": 438 }, { "epoch": 0.704, "grad_norm": 1.7632222175598145, "learning_rate": 7.710790081377502e-05, "loss": 2.7914, "step": 440 }, { "epoch": 0.7072, "grad_norm": 1.8498951196670532, "learning_rate": 7.688512893389964e-05, "loss": 2.6861, "step": 442 }, { "epoch": 0.7104, "grad_norm": 2.0129451751708984, "learning_rate": 7.666160375152496e-05, "loss": 2.8217, "step": 444 }, { "epoch": 0.7136, "grad_norm": 1.780062198638916, "learning_rate": 7.643733152967019e-05, "loss": 2.8554, "step": 446 }, { "epoch": 0.7168, "grad_norm": 1.726577877998352, "learning_rate": 7.621231855228604e-05, "loss": 2.831, "step": 448 }, { "epoch": 0.72, "grad_norm": 1.7651227712631226, "learning_rate": 7.598657112407865e-05, "loss": 2.8193, "step": 450 }, { "epoch": 0.7232, "grad_norm": 1.7961740493774414, "learning_rate": 7.576009557033304e-05, "loss": 2.8149, "step": 452 }, { "epoch": 0.7264, "grad_norm": 1.813366174697876, "learning_rate": 7.553289823673568e-05, "loss": 2.855, "step": 454 }, { "epoch": 0.7296, "grad_norm": 1.8143000602722168, "learning_rate": 7.530498548919693e-05, "loss": 2.8651, "step": 456 }, { "epoch": 0.7328, "grad_norm": 1.7585805654525757, "learning_rate": 7.507636371367246e-05, "loss": 3.0031, "step": 458 }, { "epoch": 0.736, "grad_norm": 2.2777135372161865, "learning_rate": 7.484703931598445e-05, "loss": 2.8548, "step": 460 }, { "epoch": 0.7392, "grad_norm": 1.8288154602050781, "learning_rate": 7.461701872164204e-05, "loss": 2.7425, "step": 462 }, { "epoch": 0.7424, "grad_norm": 1.8734841346740723, "learning_rate": 7.438630837566133e-05, "loss": 2.8703, "step": 464 }, { "epoch": 0.7456, "grad_norm": 1.742242455482483, "learning_rate": 7.415491474238475e-05, "loss": 2.772, "step": 466 }, { "epoch": 0.7488, "grad_norm": 1.7874287366867065, "learning_rate": 7.39228443053e-05, "loss": 2.7379, "step": 468 }, { "epoch": 0.752, "grad_norm": 1.8021794557571411, "learning_rate": 7.369010356685833e-05, "loss": 2.9262, "step": 470 }, { "epoch": 0.7552, "grad_norm": 1.7524378299713135, "learning_rate": 7.345669904829237e-05, "loss": 2.6676, "step": 472 }, { "epoch": 0.7584, "grad_norm": 1.8277724981307983, "learning_rate": 7.32226372894334e-05, "loss": 2.771, "step": 474 }, { "epoch": 0.7616, "grad_norm": 1.7385722398757935, "learning_rate": 7.298792484852808e-05, "loss": 2.7508, "step": 476 }, { "epoch": 0.7648, "grad_norm": 1.927331805229187, "learning_rate": 7.27525683020548e-05, "loss": 2.8279, "step": 478 }, { "epoch": 0.768, "grad_norm": 1.7165371179580688, "learning_rate": 7.251657424453928e-05, "loss": 2.6799, "step": 480 }, { "epoch": 0.7712, "grad_norm": 1.6585590839385986, "learning_rate": 7.227994928836988e-05, "loss": 2.6849, "step": 482 }, { "epoch": 0.7744, "grad_norm": 1.7066259384155273, "learning_rate": 7.204270006361228e-05, "loss": 2.7189, "step": 484 }, { "epoch": 0.7776, "grad_norm": 1.8811277151107788, "learning_rate": 7.180483321782374e-05, "loss": 2.7771, "step": 486 }, { "epoch": 0.7808, "grad_norm": 1.790667176246643, "learning_rate": 7.156635541586682e-05, "loss": 2.6777, "step": 488 }, { "epoch": 0.784, "grad_norm": 1.7074140310287476, "learning_rate": 7.132727333972265e-05, "loss": 2.7974, "step": 490 }, { "epoch": 0.7872, "grad_norm": 1.6692975759506226, "learning_rate": 7.108759368830371e-05, "loss": 2.7194, "step": 492 }, { "epoch": 0.7904, "grad_norm": 1.759386420249939, "learning_rate": 7.084732317726611e-05, "loss": 2.7475, "step": 494 }, { "epoch": 0.7936, "grad_norm": 1.7140787839889526, "learning_rate": 7.060646853882145e-05, "loss": 2.7576, "step": 496 }, { "epoch": 0.7968, "grad_norm": 1.6590884923934937, "learning_rate": 7.036503652154812e-05, "loss": 2.8157, "step": 498 }, { "epoch": 0.8, "grad_norm": 1.7052589654922485, "learning_rate": 7.012303389020234e-05, "loss": 2.7951, "step": 500 }, { "epoch": 0.8032, "grad_norm": 1.730635166168213, "learning_rate": 6.988046742552845e-05, "loss": 2.8279, "step": 502 }, { "epoch": 0.8064, "grad_norm": 1.7786180973052979, "learning_rate": 6.963734392406907e-05, "loss": 2.6559, "step": 504 }, { "epoch": 0.8096, "grad_norm": 1.901053547859192, "learning_rate": 6.93936701979746e-05, "loss": 2.8896, "step": 506 }, { "epoch": 0.8128, "grad_norm": 1.7321664094924927, "learning_rate": 6.914945307481228e-05, "loss": 2.795, "step": 508 }, { "epoch": 0.816, "grad_norm": 1.7901755571365356, "learning_rate": 6.890469939737506e-05, "loss": 2.7142, "step": 510 }, { "epoch": 0.8192, "grad_norm": 1.7946327924728394, "learning_rate": 6.865941602348966e-05, "loss": 2.781, "step": 512 }, { "epoch": 0.8224, "grad_norm": 1.8260494470596313, "learning_rate": 6.841360982582463e-05, "loss": 2.6868, "step": 514 }, { "epoch": 0.8256, "grad_norm": 1.8089832067489624, "learning_rate": 6.816728769169757e-05, "loss": 2.7845, "step": 516 }, { "epoch": 0.8288, "grad_norm": 1.711962342262268, "learning_rate": 6.792045652288234e-05, "loss": 2.7037, "step": 518 }, { "epoch": 0.832, "grad_norm": 1.7422336339950562, "learning_rate": 6.767312323541555e-05, "loss": 2.7938, "step": 520 }, { "epoch": 0.8352, "grad_norm": 1.7964292764663696, "learning_rate": 6.742529475940284e-05, "loss": 2.6584, "step": 522 }, { "epoch": 0.8384, "grad_norm": 1.7422146797180176, "learning_rate": 6.717697803882467e-05, "loss": 2.735, "step": 524 }, { "epoch": 0.8416, "grad_norm": 1.7856199741363525, "learning_rate": 6.692818003134184e-05, "loss": 2.7753, "step": 526 }, { "epoch": 0.8448, "grad_norm": 1.7519943714141846, "learning_rate": 6.667890770810035e-05, "loss": 2.7173, "step": 528 }, { "epoch": 0.848, "grad_norm": 1.705423355102539, "learning_rate": 6.64291680535363e-05, "loss": 2.7212, "step": 530 }, { "epoch": 0.8512, "grad_norm": 1.787747859954834, "learning_rate": 6.617896806518005e-05, "loss": 2.7499, "step": 532 }, { "epoch": 0.8544, "grad_norm": 1.652585744857788, "learning_rate": 6.592831475346018e-05, "loss": 2.5542, "step": 534 }, { "epoch": 0.8576, "grad_norm": 1.8116321563720703, "learning_rate": 6.56772151415071e-05, "loss": 2.8155, "step": 536 }, { "epoch": 0.8608, "grad_norm": 1.7901153564453125, "learning_rate": 6.542567626495619e-05, "loss": 2.7472, "step": 538 }, { "epoch": 0.864, "grad_norm": 1.7034342288970947, "learning_rate": 6.517370517175081e-05, "loss": 2.7116, "step": 540 }, { "epoch": 0.8672, "grad_norm": 1.832322597503662, "learning_rate": 6.492130892194461e-05, "loss": 2.7618, "step": 542 }, { "epoch": 0.8704, "grad_norm": 1.7125661373138428, "learning_rate": 6.466849458750394e-05, "loss": 2.6383, "step": 544 }, { "epoch": 0.8736, "grad_norm": 1.801355004310608, "learning_rate": 6.441526925210949e-05, "loss": 2.5274, "step": 546 }, { "epoch": 0.8768, "grad_norm": 1.7398046255111694, "learning_rate": 6.416164001095799e-05, "loss": 2.7207, "step": 548 }, { "epoch": 0.88, "grad_norm": 1.6901566982269287, "learning_rate": 6.390761397056328e-05, "loss": 2.7043, "step": 550 }, { "epoch": 0.8832, "grad_norm": 1.6404509544372559, "learning_rate": 6.365319824855727e-05, "loss": 2.6003, "step": 552 }, { "epoch": 0.8864, "grad_norm": 1.6571403741836548, "learning_rate": 6.339839997349045e-05, "loss": 2.7515, "step": 554 }, { "epoch": 0.8896, "grad_norm": 1.64620041847229, "learning_rate": 6.314322628463219e-05, "loss": 2.619, "step": 556 }, { "epoch": 0.8928, "grad_norm": 1.7471263408660889, "learning_rate": 6.288768433177068e-05, "loss": 2.6689, "step": 558 }, { "epoch": 0.896, "grad_norm": 1.7717193365097046, "learning_rate": 6.26317812750126e-05, "loss": 2.7607, "step": 560 }, { "epoch": 0.8992, "grad_norm": 1.6764250993728638, "learning_rate": 6.237552428458256e-05, "loss": 2.6201, "step": 562 }, { "epoch": 0.9024, "grad_norm": 1.6406267881393433, "learning_rate": 6.21189205406221e-05, "loss": 2.7044, "step": 564 }, { "epoch": 0.9056, "grad_norm": 1.7228118181228638, "learning_rate": 6.186197723298855e-05, "loss": 2.8106, "step": 566 }, { "epoch": 0.9088, "grad_norm": 1.6297177076339722, "learning_rate": 6.160470156105362e-05, "loss": 2.7442, "step": 568 }, { "epoch": 0.912, "grad_norm": 1.691156268119812, "learning_rate": 6.134710073350156e-05, "loss": 2.6915, "step": 570 }, { "epoch": 0.9152, "grad_norm": 1.7338085174560547, "learning_rate": 6.108918196812734e-05, "loss": 2.6418, "step": 572 }, { "epoch": 0.9184, "grad_norm": 1.7589360475540161, "learning_rate": 6.083095249163424e-05, "loss": 2.7577, "step": 574 }, { "epoch": 0.9216, "grad_norm": 1.6277837753295898, "learning_rate": 6.057241953943154e-05, "loss": 2.6624, "step": 576 }, { "epoch": 0.9248, "grad_norm": 1.7026268243789673, "learning_rate": 6.031359035543158e-05, "loss": 2.6401, "step": 578 }, { "epoch": 0.928, "grad_norm": 1.781171202659607, "learning_rate": 6.005447219184702e-05, "loss": 2.7218, "step": 580 }, { "epoch": 0.9312, "grad_norm": 1.7018693685531616, "learning_rate": 5.9795072308987485e-05, "loss": 2.5756, "step": 582 }, { "epoch": 0.9344, "grad_norm": 1.710750699043274, "learning_rate": 5.9535397975056154e-05, "loss": 2.738, "step": 584 }, { "epoch": 0.9376, "grad_norm": 1.7480794191360474, "learning_rate": 5.927545646594617e-05, "loss": 2.5714, "step": 586 }, { "epoch": 0.9408, "grad_norm": 1.7700692415237427, "learning_rate": 5.901525506503668e-05, "loss": 2.8079, "step": 588 }, { "epoch": 0.944, "grad_norm": 1.7823493480682373, "learning_rate": 5.87548010629889e-05, "loss": 2.6776, "step": 590 }, { "epoch": 0.9472, "grad_norm": 1.6947530508041382, "learning_rate": 5.8494101757541676e-05, "loss": 2.6956, "step": 592 }, { "epoch": 0.9504, "grad_norm": 3.6984357833862305, "learning_rate": 5.8233164453307156e-05, "loss": 2.7138, "step": 594 }, { "epoch": 0.9536, "grad_norm": 1.7368056774139404, "learning_rate": 5.797199646156596e-05, "loss": 2.6646, "step": 596 }, { "epoch": 0.9568, "grad_norm": 1.7944828271865845, "learning_rate": 5.7710605100062485e-05, "loss": 2.7675, "step": 598 }, { "epoch": 0.96, "grad_norm": 1.6918281316757202, "learning_rate": 5.7448997692799764e-05, "loss": 2.5737, "step": 600 }, { "epoch": 0.9632, "grad_norm": 1.6261357069015503, "learning_rate": 5.718718156983428e-05, "loss": 2.6307, "step": 602 }, { "epoch": 0.9664, "grad_norm": 1.7582976818084717, "learning_rate": 5.69251640670706e-05, "loss": 2.6687, "step": 604 }, { "epoch": 0.9696, "grad_norm": 1.6695537567138672, "learning_rate": 5.6662952526055793e-05, "loss": 2.5779, "step": 606 }, { "epoch": 0.9728, "grad_norm": 1.7661280632019043, "learning_rate": 5.6400554293773744e-05, "loss": 2.7142, "step": 608 }, { "epoch": 0.976, "grad_norm": 1.8265255689620972, "learning_rate": 5.61379767224393e-05, "loss": 2.749, "step": 610 }, { "epoch": 0.9792, "grad_norm": 1.6744146347045898, "learning_rate": 5.587522716929228e-05, "loss": 2.5499, "step": 612 }, { "epoch": 0.9824, "grad_norm": 1.6788204908370972, "learning_rate": 5.561231299639127e-05, "loss": 2.7511, "step": 614 }, { "epoch": 0.9856, "grad_norm": 1.8337587118148804, "learning_rate": 5.534924157040745e-05, "loss": 2.7018, "step": 616 }, { "epoch": 0.9888, "grad_norm": 1.706388235092163, "learning_rate": 5.508602026241807e-05, "loss": 2.6037, "step": 618 }, { "epoch": 0.992, "grad_norm": 1.699271559715271, "learning_rate": 5.482265644769998e-05, "loss": 2.7222, "step": 620 }, { "epoch": 0.9952, "grad_norm": 1.67972993850708, "learning_rate": 5.4559157505522985e-05, "loss": 2.4653, "step": 622 }, { "epoch": 0.9984, "grad_norm": 1.6633968353271484, "learning_rate": 5.429553081894304e-05, "loss": 2.6239, "step": 624 }, { "epoch": 1.0016, "grad_norm": 1.6871411800384521, "learning_rate": 5.4031783774595455e-05, "loss": 2.4794, "step": 626 }, { "epoch": 1.0048, "grad_norm": 1.5942273139953613, "learning_rate": 5.3767923762487824e-05, "loss": 2.1013, "step": 628 }, { "epoch": 1.008, "grad_norm": 1.6011604070663452, "learning_rate": 5.3503958175793055e-05, "loss": 2.139, "step": 630 }, { "epoch": 1.0112, "grad_norm": 1.642553687095642, "learning_rate": 5.323989441064216e-05, "loss": 2.0493, "step": 632 }, { "epoch": 1.0144, "grad_norm": 1.7297475337982178, "learning_rate": 5.2975739865917074e-05, "loss": 1.9964, "step": 634 }, { "epoch": 1.0176, "grad_norm": 1.7036877870559692, "learning_rate": 5.271150194304326e-05, "loss": 2.0491, "step": 636 }, { "epoch": 1.0208, "grad_norm": 1.6151326894760132, "learning_rate": 5.244718804578246e-05, "loss": 2.1586, "step": 638 }, { "epoch": 1.024, "grad_norm": 1.6182959079742432, "learning_rate": 5.218280558002506e-05, "loss": 2.0155, "step": 640 }, { "epoch": 1.0272, "grad_norm": 1.5469590425491333, "learning_rate": 5.191836195358278e-05, "loss": 2.0697, "step": 642 }, { "epoch": 1.0304, "grad_norm": 1.574875831604004, "learning_rate": 5.165386457598099e-05, "loss": 2.0498, "step": 644 }, { "epoch": 1.0336, "grad_norm": 1.6611751317977905, "learning_rate": 5.13893208582511e-05, "loss": 2.0674, "step": 646 }, { "epoch": 1.0368, "grad_norm": 1.622523546218872, "learning_rate": 5.1124738212722966e-05, "loss": 2.1196, "step": 648 }, { "epoch": 1.04, "grad_norm": 1.5410690307617188, "learning_rate": 5.086012405281717e-05, "loss": 2.1537, "step": 650 }, { "epoch": 1.0432, "grad_norm": 1.564372181892395, "learning_rate": 5.0595485792837305e-05, "loss": 1.9727, "step": 652 }, { "epoch": 1.0464, "grad_norm": 1.5646798610687256, "learning_rate": 5.033083084776222e-05, "loss": 2.1061, "step": 654 }, { "epoch": 1.0496, "grad_norm": 1.661833643913269, "learning_rate": 5.0066166633038305e-05, "loss": 2.075, "step": 656 }, { "epoch": 1.0528, "grad_norm": 1.6063848733901978, "learning_rate": 4.980150056437163e-05, "loss": 2.0384, "step": 658 }, { "epoch": 1.056, "grad_norm": 1.7570315599441528, "learning_rate": 4.9536840057520224e-05, "loss": 2.1477, "step": 660 }, { "epoch": 1.0592, "grad_norm": 1.6103792190551758, "learning_rate": 4.927219252808631e-05, "loss": 2.025, "step": 662 }, { "epoch": 1.0624, "grad_norm": 1.6376656293869019, "learning_rate": 4.900756539130846e-05, "loss": 2.0121, "step": 664 }, { "epoch": 1.0656, "grad_norm": 1.6290217638015747, "learning_rate": 4.874296606185387e-05, "loss": 1.9497, "step": 666 }, { "epoch": 1.0688, "grad_norm": 1.622162938117981, "learning_rate": 4.847840195361058e-05, "loss": 2.1003, "step": 668 }, { "epoch": 1.072, "grad_norm": 1.6676520109176636, "learning_rate": 4.821388047947979e-05, "loss": 2.0269, "step": 670 }, { "epoch": 1.0752, "grad_norm": 1.6200871467590332, "learning_rate": 4.7949409051168085e-05, "loss": 2.186, "step": 672 }, { "epoch": 1.0784, "grad_norm": 1.6458507776260376, "learning_rate": 4.768499507897981e-05, "loss": 2.1137, "step": 674 }, { "epoch": 1.0816, "grad_norm": 1.5950709581375122, "learning_rate": 4.742064597160948e-05, "loss": 1.9582, "step": 676 }, { "epoch": 1.0848, "grad_norm": 1.7151204347610474, "learning_rate": 4.715636913593404e-05, "loss": 2.0608, "step": 678 }, { "epoch": 1.088, "grad_norm": 1.6758900880813599, "learning_rate": 4.689217197680554e-05, "loss": 1.9813, "step": 680 }, { "epoch": 1.0912, "grad_norm": 1.7488733530044556, "learning_rate": 4.6628061896843474e-05, "loss": 1.983, "step": 682 }, { "epoch": 1.0944, "grad_norm": 1.6792566776275635, "learning_rate": 4.6364046296227484e-05, "loss": 2.0934, "step": 684 }, { "epoch": 1.0976, "grad_norm": 1.595950722694397, "learning_rate": 4.6100132572489915e-05, "loss": 2.0593, "step": 686 }, { "epoch": 1.1008, "grad_norm": 1.6133977174758911, "learning_rate": 4.5836328120308674e-05, "loss": 2.1856, "step": 688 }, { "epoch": 1.104, "grad_norm": 1.6371673345565796, "learning_rate": 4.5572640331299875e-05, "loss": 1.8256, "step": 690 }, { "epoch": 1.1072, "grad_norm": 1.6217482089996338, "learning_rate": 4.530907659381086e-05, "loss": 2.0228, "step": 692 }, { "epoch": 1.1104, "grad_norm": 1.8364999294281006, "learning_rate": 4.504564429271311e-05, "loss": 2.2268, "step": 694 }, { "epoch": 1.1136, "grad_norm": 1.6585414409637451, "learning_rate": 4.478235080919536e-05, "loss": 2.1736, "step": 696 }, { "epoch": 1.1168, "grad_norm": 1.7598919868469238, "learning_rate": 4.451920352055678e-05, "loss": 2.0017, "step": 698 }, { "epoch": 1.12, "grad_norm": 1.659216284751892, "learning_rate": 4.425620980000026e-05, "loss": 2.1184, "step": 700 }, { "epoch": 1.1232, "grad_norm": 1.6339902877807617, "learning_rate": 4.39933770164258e-05, "loss": 2.0396, "step": 702 }, { "epoch": 1.1264, "grad_norm": 1.713364839553833, "learning_rate": 4.373071253422408e-05, "loss": 2.0963, "step": 704 }, { "epoch": 1.1296, "grad_norm": 1.728036642074585, "learning_rate": 4.346822371307009e-05, "loss": 2.026, "step": 706 }, { "epoch": 1.1328, "grad_norm": 1.6496070623397827, "learning_rate": 4.320591790771691e-05, "loss": 2.0528, "step": 708 }, { "epoch": 1.1360000000000001, "grad_norm": 1.7058520317077637, "learning_rate": 4.294380246778966e-05, "loss": 1.9498, "step": 710 }, { "epoch": 1.1392, "grad_norm": 1.5882078409194946, "learning_rate": 4.2681884737579524e-05, "loss": 1.9813, "step": 712 }, { "epoch": 1.1424, "grad_norm": 1.7214953899383545, "learning_rate": 4.242017205583805e-05, "loss": 2.0316, "step": 714 }, { "epoch": 1.1456, "grad_norm": 1.6648966073989868, "learning_rate": 4.215867175557142e-05, "loss": 2.0019, "step": 716 }, { "epoch": 1.1488, "grad_norm": 1.6206002235412598, "learning_rate": 4.189739116383506e-05, "loss": 1.9073, "step": 718 }, { "epoch": 1.152, "grad_norm": 1.6015231609344482, "learning_rate": 4.163633760152834e-05, "loss": 2.013, "step": 720 }, { "epoch": 1.1552, "grad_norm": 1.6537818908691406, "learning_rate": 4.137551838318936e-05, "loss": 2.1472, "step": 722 }, { "epoch": 1.1584, "grad_norm": 1.9408031702041626, "learning_rate": 4.1114940816790135e-05, "loss": 2.1239, "step": 724 }, { "epoch": 1.1616, "grad_norm": 1.7310969829559326, "learning_rate": 4.08546122035317e-05, "loss": 2.1907, "step": 726 }, { "epoch": 1.1648, "grad_norm": 1.6550313234329224, "learning_rate": 4.059453983763967e-05, "loss": 2.0719, "step": 728 }, { "epoch": 1.168, "grad_norm": 1.8182940483093262, "learning_rate": 4.03347310061597e-05, "loss": 2.0437, "step": 730 }, { "epoch": 1.1712, "grad_norm": 1.6281683444976807, "learning_rate": 4.007519298875347e-05, "loss": 2.0488, "step": 732 }, { "epoch": 1.1743999999999999, "grad_norm": 3.110891342163086, "learning_rate": 3.98159330574946e-05, "loss": 2.0592, "step": 734 }, { "epoch": 1.1776, "grad_norm": 1.7148778438568115, "learning_rate": 3.955695847666494e-05, "loss": 2.0754, "step": 736 }, { "epoch": 1.1808, "grad_norm": 1.6069791316986084, "learning_rate": 3.929827650255104e-05, "loss": 1.9366, "step": 738 }, { "epoch": 1.184, "grad_norm": 1.7344155311584473, "learning_rate": 3.903989438324077e-05, "loss": 2.1932, "step": 740 }, { "epoch": 1.1872, "grad_norm": 1.7402360439300537, "learning_rate": 3.878181935842033e-05, "loss": 2.1352, "step": 742 }, { "epoch": 1.1904, "grad_norm": 1.7004384994506836, "learning_rate": 3.85240586591713e-05, "loss": 1.9779, "step": 744 }, { "epoch": 1.1936, "grad_norm": 1.8418961763381958, "learning_rate": 3.8266619507768126e-05, "loss": 2.2648, "step": 746 }, { "epoch": 1.1968, "grad_norm": 1.7444120645523071, "learning_rate": 3.800950911747565e-05, "loss": 2.0055, "step": 748 }, { "epoch": 1.2, "grad_norm": 1.709296703338623, "learning_rate": 3.775273469234712e-05, "loss": 2.1322, "step": 750 }, { "epoch": 1.2032, "grad_norm": 1.6528595685958862, "learning_rate": 3.749630342702221e-05, "loss": 2.0173, "step": 752 }, { "epoch": 1.2064, "grad_norm": 1.86497163772583, "learning_rate": 3.724022250652557e-05, "loss": 2.2187, "step": 754 }, { "epoch": 1.2096, "grad_norm": 1.6813569068908691, "learning_rate": 3.698449910606536e-05, "loss": 1.9149, "step": 756 }, { "epoch": 1.2128, "grad_norm": 1.4694783687591553, "learning_rate": 3.672914039083233e-05, "loss": 1.9928, "step": 758 }, { "epoch": 1.216, "grad_norm": 1.6934654712677002, "learning_rate": 3.6474153515799e-05, "loss": 2.0018, "step": 760 }, { "epoch": 1.2192, "grad_norm": 1.7023409605026245, "learning_rate": 3.6219545625519145e-05, "loss": 1.9795, "step": 762 }, { "epoch": 1.2224, "grad_norm": 1.5944123268127441, "learning_rate": 3.596532385392772e-05, "loss": 1.9711, "step": 764 }, { "epoch": 1.2256, "grad_norm": 1.665781855583191, "learning_rate": 3.5711495324140845e-05, "loss": 2.0697, "step": 766 }, { "epoch": 1.2288000000000001, "grad_norm": 1.6670796871185303, "learning_rate": 3.54580671482563e-05, "loss": 2.0016, "step": 768 }, { "epoch": 1.232, "grad_norm": 1.81361985206604, "learning_rate": 3.520504642715424e-05, "loss": 2.1093, "step": 770 }, { "epoch": 1.2352, "grad_norm": 1.6931331157684326, "learning_rate": 3.495244025029822e-05, "loss": 2.0959, "step": 772 }, { "epoch": 1.2384, "grad_norm": 1.6072664260864258, "learning_rate": 3.470025569553653e-05, "loss": 1.9525, "step": 774 }, { "epoch": 1.2416, "grad_norm": 1.5714666843414307, "learning_rate": 3.444849982890393e-05, "loss": 1.9701, "step": 776 }, { "epoch": 1.2448, "grad_norm": 1.7283622026443481, "learning_rate": 3.4197179704423656e-05, "loss": 2.1773, "step": 778 }, { "epoch": 1.248, "grad_norm": 1.5926114320755005, "learning_rate": 3.39463023639097e-05, "loss": 1.9916, "step": 780 }, { "epoch": 1.2511999999999999, "grad_norm": 1.6143758296966553, "learning_rate": 3.36958748367696e-05, "loss": 2.0623, "step": 782 }, { "epoch": 1.2544, "grad_norm": 1.5428004264831543, "learning_rate": 3.3445904139807405e-05, "loss": 1.966, "step": 784 }, { "epoch": 1.2576, "grad_norm": 1.7031697034835815, "learning_rate": 3.319639727702716e-05, "loss": 2.1779, "step": 786 }, { "epoch": 1.2608, "grad_norm": 1.5835998058319092, "learning_rate": 3.2947361239436525e-05, "loss": 1.9549, "step": 788 }, { "epoch": 1.264, "grad_norm": 1.7775769233703613, "learning_rate": 3.2698803004851026e-05, "loss": 1.9312, "step": 790 }, { "epoch": 1.2671999999999999, "grad_norm": 1.583977460861206, "learning_rate": 3.245072953769844e-05, "loss": 1.8452, "step": 792 }, { "epoch": 1.2704, "grad_norm": 1.5791876316070557, "learning_rate": 3.2203147788823764e-05, "loss": 2.0222, "step": 794 }, { "epoch": 1.2736, "grad_norm": 1.6620211601257324, "learning_rate": 3.1956064695294305e-05, "loss": 2.0612, "step": 796 }, { "epoch": 1.2768, "grad_norm": 1.6324514150619507, "learning_rate": 3.170948718020546e-05, "loss": 2.0428, "step": 798 }, { "epoch": 1.28, "grad_norm": 1.6789875030517578, "learning_rate": 3.1463422152486674e-05, "loss": 1.967, "step": 800 }, { "epoch": 1.2832, "grad_norm": 2.000178337097168, "learning_rate": 3.121787650670781e-05, "loss": 1.9851, "step": 802 }, { "epoch": 1.2864, "grad_norm": 1.7736194133758545, "learning_rate": 3.097285712288605e-05, "loss": 1.8732, "step": 804 }, { "epoch": 1.2896, "grad_norm": 1.7309801578521729, "learning_rate": 3.072837086629305e-05, "loss": 2.0212, "step": 806 }, { "epoch": 1.2928, "grad_norm": 1.7109277248382568, "learning_rate": 3.0484424587262655e-05, "loss": 2.0203, "step": 808 }, { "epoch": 1.296, "grad_norm": 1.6125346422195435, "learning_rate": 3.024102512099889e-05, "loss": 2.075, "step": 810 }, { "epoch": 1.2992, "grad_norm": 1.661042332649231, "learning_rate": 2.9998179287384485e-05, "loss": 2.1594, "step": 812 }, { "epoch": 1.3024, "grad_norm": 1.6753617525100708, "learning_rate": 2.975589389078975e-05, "loss": 2.0397, "step": 814 }, { "epoch": 1.3056, "grad_norm": 1.6913373470306396, "learning_rate": 2.9514175719881966e-05, "loss": 2.0017, "step": 816 }, { "epoch": 1.3088, "grad_norm": 1.6717087030410767, "learning_rate": 2.9273031547435114e-05, "loss": 2.1042, "step": 818 }, { "epoch": 1.312, "grad_norm": 1.7565637826919556, "learning_rate": 2.9032468130140168e-05, "loss": 2.1539, "step": 820 }, { "epoch": 1.3152, "grad_norm": 1.6280590295791626, "learning_rate": 2.8792492208415768e-05, "loss": 1.9744, "step": 822 }, { "epoch": 1.3184, "grad_norm": 1.5978885889053345, "learning_rate": 2.8553110506219283e-05, "loss": 2.0044, "step": 824 }, { "epoch": 1.3216, "grad_norm": 1.6840837001800537, "learning_rate": 2.831432973085848e-05, "loss": 2.062, "step": 826 }, { "epoch": 1.3248, "grad_norm": 1.5866726636886597, "learning_rate": 2.8076156572803635e-05, "loss": 1.9896, "step": 828 }, { "epoch": 1.328, "grad_norm": 1.5741474628448486, "learning_rate": 2.783859770549996e-05, "loss": 1.9465, "step": 830 }, { "epoch": 1.3312, "grad_norm": 1.683282732963562, "learning_rate": 2.760165978518067e-05, "loss": 2.0189, "step": 832 }, { "epoch": 1.3344, "grad_norm": 1.6024737358093262, "learning_rate": 2.7365349450680466e-05, "loss": 1.941, "step": 834 }, { "epoch": 1.3376000000000001, "grad_norm": 1.6115479469299316, "learning_rate": 2.7129673323249604e-05, "loss": 1.9948, "step": 836 }, { "epoch": 1.3408, "grad_norm": 1.6458511352539062, "learning_rate": 2.689463800636824e-05, "loss": 2.0941, "step": 838 }, { "epoch": 1.3439999999999999, "grad_norm": 1.5999996662139893, "learning_rate": 2.6660250085561457e-05, "loss": 2.081, "step": 840 }, { "epoch": 1.3472, "grad_norm": 1.6597297191619873, "learning_rate": 2.6426516128214807e-05, "loss": 1.8927, "step": 842 }, { "epoch": 1.3504, "grad_norm": 1.7434853315353394, "learning_rate": 2.619344268339021e-05, "loss": 2.0059, "step": 844 }, { "epoch": 1.3536000000000001, "grad_norm": 2.4289755821228027, "learning_rate": 2.5961036281642493e-05, "loss": 2.1133, "step": 846 }, { "epoch": 1.3568, "grad_norm": 1.623465657234192, "learning_rate": 2.572930343483637e-05, "loss": 1.9284, "step": 848 }, { "epoch": 1.3599999999999999, "grad_norm": 1.9334876537322998, "learning_rate": 2.54982506359641e-05, "loss": 2.0351, "step": 850 }, { "epoch": 1.3632, "grad_norm": 1.7356194257736206, "learning_rate": 2.526788435896339e-05, "loss": 2.0995, "step": 852 }, { "epoch": 1.3664, "grad_norm": 1.581726312637329, "learning_rate": 2.5038211058536133e-05, "loss": 1.9878, "step": 854 }, { "epoch": 1.3696, "grad_norm": 1.7448982000350952, "learning_rate": 2.4809237169967458e-05, "loss": 2.0996, "step": 856 }, { "epoch": 1.3728, "grad_norm": 1.6610596179962158, "learning_rate": 2.4580969108945533e-05, "loss": 1.9928, "step": 858 }, { "epoch": 1.376, "grad_norm": 1.6751465797424316, "learning_rate": 2.435341327138168e-05, "loss": 2.0282, "step": 860 }, { "epoch": 1.3792, "grad_norm": 1.7888554334640503, "learning_rate": 2.4126576033231208e-05, "loss": 1.9525, "step": 862 }, { "epoch": 1.3824, "grad_norm": 1.66560959815979, "learning_rate": 2.3900463750314834e-05, "loss": 1.984, "step": 864 }, { "epoch": 1.3856, "grad_norm": 1.6778205633163452, "learning_rate": 2.3675082758140475e-05, "loss": 1.8466, "step": 866 }, { "epoch": 1.3888, "grad_norm": 1.6759799718856812, "learning_rate": 2.3450439371725825e-05, "loss": 2.0435, "step": 868 }, { "epoch": 1.392, "grad_norm": 1.6375926733016968, "learning_rate": 2.3226539885421343e-05, "loss": 1.9285, "step": 870 }, { "epoch": 1.3952, "grad_norm": 1.7163138389587402, "learning_rate": 2.3003390572734006e-05, "loss": 2.0509, "step": 872 }, { "epoch": 1.3984, "grad_norm": 1.7290489673614502, "learning_rate": 2.2780997686151378e-05, "loss": 2.0219, "step": 874 }, { "epoch": 1.4016, "grad_norm": 1.5673465728759766, "learning_rate": 2.255936745696652e-05, "loss": 2.0706, "step": 876 }, { "epoch": 1.4048, "grad_norm": 1.7429379224777222, "learning_rate": 2.2338506095103334e-05, "loss": 1.9849, "step": 878 }, { "epoch": 1.408, "grad_norm": 1.6174511909484863, "learning_rate": 2.2118419788942672e-05, "loss": 1.958, "step": 880 }, { "epoch": 1.4112, "grad_norm": 1.653412938117981, "learning_rate": 2.189911470514881e-05, "loss": 1.9639, "step": 882 }, { "epoch": 1.4144, "grad_norm": 1.7575905323028564, "learning_rate": 2.1680596988496705e-05, "loss": 2.0275, "step": 884 }, { "epoch": 1.4176, "grad_norm": 1.4820189476013184, "learning_rate": 2.1462872761699905e-05, "loss": 1.8751, "step": 886 }, { "epoch": 1.4208, "grad_norm": 1.558645248413086, "learning_rate": 2.1245948125238867e-05, "loss": 1.8586, "step": 888 }, { "epoch": 1.424, "grad_norm": 1.6014423370361328, "learning_rate": 2.1029829157190117e-05, "loss": 1.9532, "step": 890 }, { "epoch": 1.4272, "grad_norm": 1.5966588258743286, "learning_rate": 2.081452191305587e-05, "loss": 1.9198, "step": 892 }, { "epoch": 1.4304000000000001, "grad_norm": 1.5865297317504883, "learning_rate": 2.06000324255945e-05, "loss": 1.907, "step": 894 }, { "epoch": 1.4336, "grad_norm": 1.6410948038101196, "learning_rate": 2.0386366704651315e-05, "loss": 2.1222, "step": 896 }, { "epoch": 1.4368, "grad_norm": 1.6076593399047852, "learning_rate": 2.0173530736990304e-05, "loss": 1.8874, "step": 898 }, { "epoch": 1.44, "grad_norm": 1.610167384147644, "learning_rate": 1.9961530486126327e-05, "loss": 2.074, "step": 900 }, { "epoch": 1.4432, "grad_norm": 1.7789791822433472, "learning_rate": 1.9750371892158103e-05, "loss": 2.0396, "step": 902 }, { "epoch": 1.4464000000000001, "grad_norm": 1.6089529991149902, "learning_rate": 1.9540060871601646e-05, "loss": 2.0706, "step": 904 }, { "epoch": 1.4496, "grad_norm": 1.6056288480758667, "learning_rate": 1.933060331722457e-05, "loss": 1.8598, "step": 906 }, { "epoch": 1.4527999999999999, "grad_norm": 1.5929008722305298, "learning_rate": 1.9122005097881014e-05, "loss": 1.9494, "step": 908 }, { "epoch": 1.456, "grad_norm": 1.6866933107376099, "learning_rate": 1.8914272058347088e-05, "loss": 1.8477, "step": 910 }, { "epoch": 1.4592, "grad_norm": 1.614665150642395, "learning_rate": 1.8707410019157196e-05, "loss": 1.907, "step": 912 }, { "epoch": 1.4624, "grad_norm": 1.5032063722610474, "learning_rate": 1.8501424776440907e-05, "loss": 1.9152, "step": 914 }, { "epoch": 1.4656, "grad_norm": 1.6080886125564575, "learning_rate": 1.829632210176061e-05, "loss": 1.9224, "step": 916 }, { "epoch": 1.4687999999999999, "grad_norm": 1.6803348064422607, "learning_rate": 1.809210774194971e-05, "loss": 1.9741, "step": 918 }, { "epoch": 1.472, "grad_norm": 1.7053711414337158, "learning_rate": 1.7888787418951645e-05, "loss": 2.0472, "step": 920 }, { "epoch": 1.4752, "grad_norm": 1.6643372774124146, "learning_rate": 1.7686366829659628e-05, "loss": 2.0729, "step": 922 }, { "epoch": 1.4784, "grad_norm": 1.6798537969589233, "learning_rate": 1.74848516457569e-05, "loss": 1.9989, "step": 924 }, { "epoch": 1.4816, "grad_norm": 1.6798433065414429, "learning_rate": 1.72842475135579e-05, "loss": 2.0111, "step": 926 }, { "epoch": 1.4848, "grad_norm": 1.5733171701431274, "learning_rate": 1.7084560053850024e-05, "loss": 2.0498, "step": 928 }, { "epoch": 1.488, "grad_norm": 1.6396392583847046, "learning_rate": 1.6885794861736183e-05, "loss": 1.9701, "step": 930 }, { "epoch": 1.4912, "grad_norm": 1.628389596939087, "learning_rate": 1.668795750647796e-05, "loss": 2.0453, "step": 932 }, { "epoch": 1.4944, "grad_norm": 1.580734372138977, "learning_rate": 1.6491053531339607e-05, "loss": 1.908, "step": 934 }, { "epoch": 1.4976, "grad_norm": 1.6267027854919434, "learning_rate": 1.62950884534327e-05, "loss": 1.949, "step": 936 }, { "epoch": 1.5008, "grad_norm": 1.5968527793884277, "learning_rate": 1.6100067763561626e-05, "loss": 1.9075, "step": 938 }, { "epoch": 1.504, "grad_norm": 1.6220203638076782, "learning_rate": 1.5905996926069628e-05, "loss": 1.8555, "step": 940 }, { "epoch": 1.5072, "grad_norm": 1.6823129653930664, "learning_rate": 1.5712881378685755e-05, "loss": 2.083, "step": 942 }, { "epoch": 1.5104, "grad_norm": 1.634656548500061, "learning_rate": 1.5520726532372537e-05, "loss": 1.9641, "step": 944 }, { "epoch": 1.5135999999999998, "grad_norm": 1.5676469802856445, "learning_rate": 1.532953777117429e-05, "loss": 1.9694, "step": 946 }, { "epoch": 1.5168, "grad_norm": 1.652653455734253, "learning_rate": 1.5139320452066313e-05, "loss": 1.9754, "step": 948 }, { "epoch": 1.52, "grad_norm": 1.5929044485092163, "learning_rate": 1.4950079904804759e-05, "loss": 1.9062, "step": 950 }, { "epoch": 1.5232, "grad_norm": 1.6427743434906006, "learning_rate": 1.4761821431777373e-05, "loss": 2.0017, "step": 952 }, { "epoch": 1.5264, "grad_norm": 1.5246950387954712, "learning_rate": 1.4574550307854817e-05, "loss": 1.9427, "step": 954 }, { "epoch": 1.5295999999999998, "grad_norm": 1.6303575038909912, "learning_rate": 1.4388271780242929e-05, "loss": 1.8824, "step": 956 }, { "epoch": 1.5328, "grad_norm": 1.690637469291687, "learning_rate": 1.4202991068335697e-05, "loss": 2.0095, "step": 958 }, { "epoch": 1.536, "grad_norm": 1.7552894353866577, "learning_rate": 1.4018713363569035e-05, "loss": 2.0734, "step": 960 }, { "epoch": 1.5392000000000001, "grad_norm": 1.633931279182434, "learning_rate": 1.3835443829275268e-05, "loss": 1.8231, "step": 962 }, { "epoch": 1.5424, "grad_norm": 1.586737036705017, "learning_rate": 1.365318760053848e-05, "loss": 1.9087, "step": 964 }, { "epoch": 1.5455999999999999, "grad_norm": 1.5858091115951538, "learning_rate": 1.3471949784050702e-05, "loss": 1.8308, "step": 966 }, { "epoch": 1.5488, "grad_norm": 1.6560001373291016, "learning_rate": 1.3291735457968701e-05, "loss": 2.0853, "step": 968 }, { "epoch": 1.552, "grad_norm": 1.6715130805969238, "learning_rate": 1.3112549671771796e-05, "loss": 2.0848, "step": 970 }, { "epoch": 1.5552000000000001, "grad_norm": 1.6843191385269165, "learning_rate": 1.2934397446120306e-05, "loss": 2.0182, "step": 972 }, { "epoch": 1.5584, "grad_norm": 1.616097092628479, "learning_rate": 1.2757283772714957e-05, "loss": 1.9008, "step": 974 }, { "epoch": 1.5615999999999999, "grad_norm": 1.5122194290161133, "learning_rate": 1.2581213614156928e-05, "loss": 1.7905, "step": 976 }, { "epoch": 1.5648, "grad_norm": 1.6870348453521729, "learning_rate": 1.2406191903808844e-05, "loss": 2.1301, "step": 978 }, { "epoch": 1.568, "grad_norm": 1.5509247779846191, "learning_rate": 1.2232223545656552e-05, "loss": 1.8702, "step": 980 }, { "epoch": 1.5712000000000002, "grad_norm": 1.5840715169906616, "learning_rate": 1.205931341417173e-05, "loss": 2.0163, "step": 982 }, { "epoch": 1.5744, "grad_norm": 1.659145712852478, "learning_rate": 1.1887466354175253e-05, "loss": 1.8171, "step": 984 }, { "epoch": 1.5776, "grad_norm": 1.7107597589492798, "learning_rate": 1.1716687180701474e-05, "loss": 2.0593, "step": 986 }, { "epoch": 1.5808, "grad_norm": 1.5526190996170044, "learning_rate": 1.1546980678863361e-05, "loss": 1.8897, "step": 988 }, { "epoch": 1.584, "grad_norm": 1.6252529621124268, "learning_rate": 1.1378351603718312e-05, "loss": 1.9001, "step": 990 }, { "epoch": 1.5872000000000002, "grad_norm": 1.64818274974823, "learning_rate": 1.1210804680135022e-05, "loss": 2.0044, "step": 992 }, { "epoch": 1.5904, "grad_norm": 1.6926898956298828, "learning_rate": 1.1044344602661034e-05, "loss": 2.1088, "step": 994 }, { "epoch": 1.5936, "grad_norm": 1.7737877368927002, "learning_rate": 1.0878976035391252e-05, "loss": 1.9949, "step": 996 }, { "epoch": 1.5968, "grad_norm": 1.6282092332839966, "learning_rate": 1.0714703611837201e-05, "loss": 1.9759, "step": 998 }, { "epoch": 1.6, "grad_norm": 1.5869003534317017, "learning_rate": 1.0551531934797243e-05, "loss": 2.0077, "step": 1000 }, { "epoch": 1.6032, "grad_norm": 1.5880354642868042, "learning_rate": 1.0389465576227558e-05, "loss": 2.0529, "step": 1002 }, { "epoch": 1.6064, "grad_norm": 1.796076774597168, "learning_rate": 1.0228509077114146e-05, "loss": 1.9833, "step": 1004 }, { "epoch": 1.6096, "grad_norm": 1.6073834896087646, "learning_rate": 1.0068666947345456e-05, "loss": 1.8307, "step": 1006 }, { "epoch": 1.6128, "grad_norm": 1.637624740600586, "learning_rate": 9.909943665586102e-06, "loss": 1.9477, "step": 1008 }, { "epoch": 1.616, "grad_norm": 1.6284286975860596, "learning_rate": 9.752343679151399e-06, "loss": 1.9163, "step": 1010 }, { "epoch": 1.6192, "grad_norm": 1.5366007089614868, "learning_rate": 9.595871403882661e-06, "loss": 1.8232, "step": 1012 }, { "epoch": 1.6223999999999998, "grad_norm": 1.5945929288864136, "learning_rate": 9.440531224023552e-06, "loss": 1.9313, "step": 1014 }, { "epoch": 1.6256, "grad_norm": 1.589713454246521, "learning_rate": 9.286327492097196e-06, "loss": 1.8511, "step": 1016 }, { "epoch": 1.6288, "grad_norm": 1.6770306825637817, "learning_rate": 9.133264528784274e-06, "loss": 2.1287, "step": 1018 }, { "epoch": 1.6320000000000001, "grad_norm": 1.5880622863769531, "learning_rate": 8.981346622801905e-06, "loss": 1.9876, "step": 1020 }, { "epoch": 1.6352, "grad_norm": 1.6425656080245972, "learning_rate": 8.830578030783493e-06, "loss": 1.9793, "step": 1022 }, { "epoch": 1.6383999999999999, "grad_norm": 1.7399017810821533, "learning_rate": 8.680962977159502e-06, "loss": 1.9023, "step": 1024 }, { "epoch": 1.6416, "grad_norm": 1.6427820920944214, "learning_rate": 8.53250565403903e-06, "loss": 1.897, "step": 1026 }, { "epoch": 1.6448, "grad_norm": 1.5444337129592896, "learning_rate": 8.385210221092382e-06, "loss": 1.8902, "step": 1028 }, { "epoch": 1.6480000000000001, "grad_norm": 1.6862332820892334, "learning_rate": 8.239080805434513e-06, "loss": 1.9339, "step": 1030 }, { "epoch": 1.6512, "grad_norm": 1.5926483869552612, "learning_rate": 8.094121501509399e-06, "loss": 1.8791, "step": 1032 }, { "epoch": 1.6543999999999999, "grad_norm": 1.634626865386963, "learning_rate": 7.950336370975304e-06, "loss": 1.9781, "step": 1034 }, { "epoch": 1.6576, "grad_norm": 1.6866514682769775, "learning_rate": 7.80772944259096e-06, "loss": 1.9539, "step": 1036 }, { "epoch": 1.6608, "grad_norm": 1.58961820602417, "learning_rate": 7.666304712102695e-06, "loss": 1.669, "step": 1038 }, { "epoch": 1.6640000000000001, "grad_norm": 1.800206184387207, "learning_rate": 7.526066142132521e-06, "loss": 1.9265, "step": 1040 }, { "epoch": 1.6672, "grad_norm": 1.6069746017456055, "learning_rate": 7.3870176620670194e-06, "loss": 1.8683, "step": 1042 }, { "epoch": 1.6703999999999999, "grad_norm": 1.639652132987976, "learning_rate": 7.249163167947287e-06, "loss": 1.8093, "step": 1044 }, { "epoch": 1.6736, "grad_norm": 1.6031360626220703, "learning_rate": 7.1125065223598076e-06, "loss": 2.0082, "step": 1046 }, { "epoch": 1.6768, "grad_norm": 1.6937788724899292, "learning_rate": 6.9770515543281455e-06, "loss": 2.0, "step": 1048 }, { "epoch": 1.6800000000000002, "grad_norm": 1.6093850135803223, "learning_rate": 6.842802059205727e-06, "loss": 2.0147, "step": 1050 }, { "epoch": 1.6832, "grad_norm": 1.541329026222229, "learning_rate": 6.709761798569442e-06, "loss": 1.8566, "step": 1052 }, { "epoch": 1.6864, "grad_norm": 1.5949456691741943, "learning_rate": 6.577934500114335e-06, "loss": 1.873, "step": 1054 }, { "epoch": 1.6896, "grad_norm": 1.6678704023361206, "learning_rate": 6.44732385754902e-06, "loss": 1.8975, "step": 1056 }, { "epoch": 1.6928, "grad_norm": 1.7647327184677124, "learning_rate": 6.3179335304923095e-06, "loss": 1.932, "step": 1058 }, { "epoch": 1.696, "grad_norm": 1.6539088487625122, "learning_rate": 6.189767144370645e-06, "loss": 1.9846, "step": 1060 }, { "epoch": 1.6992, "grad_norm": 2.910256862640381, "learning_rate": 6.062828290316469e-06, "loss": 2.0498, "step": 1062 }, { "epoch": 1.7024, "grad_norm": 1.6432437896728516, "learning_rate": 5.937120525067641e-06, "loss": 1.9539, "step": 1064 }, { "epoch": 1.7056, "grad_norm": 1.6407873630523682, "learning_rate": 5.812647370867763e-06, "loss": 1.9476, "step": 1066 }, { "epoch": 1.7088, "grad_norm": 1.617720603942871, "learning_rate": 5.689412315367543e-06, "loss": 1.9902, "step": 1068 }, { "epoch": 1.712, "grad_norm": 1.5690410137176514, "learning_rate": 5.567418811526981e-06, "loss": 1.8755, "step": 1070 }, { "epoch": 1.7151999999999998, "grad_norm": 1.602350115776062, "learning_rate": 5.4466702775186785e-06, "loss": 1.9992, "step": 1072 }, { "epoch": 1.7184, "grad_norm": 1.511695384979248, "learning_rate": 5.327170096632089e-06, "loss": 1.749, "step": 1074 }, { "epoch": 1.7216, "grad_norm": 1.547643780708313, "learning_rate": 5.208921617178641e-06, "loss": 1.8993, "step": 1076 }, { "epoch": 1.7248, "grad_norm": 1.7196307182312012, "learning_rate": 5.091928152397984e-06, "loss": 1.9801, "step": 1078 }, { "epoch": 1.728, "grad_norm": 1.7015469074249268, "learning_rate": 4.976192980365124e-06, "loss": 1.7825, "step": 1080 }, { "epoch": 1.7311999999999999, "grad_norm": 1.688491702079773, "learning_rate": 4.861719343898613e-06, "loss": 2.0279, "step": 1082 }, { "epoch": 1.7344, "grad_norm": 1.6017239093780518, "learning_rate": 4.748510450469623e-06, "loss": 1.8703, "step": 1084 }, { "epoch": 1.7376, "grad_norm": 1.6276230812072754, "learning_rate": 4.63656947211214e-06, "loss": 1.9359, "step": 1086 }, { "epoch": 1.7408000000000001, "grad_norm": 1.5692756175994873, "learning_rate": 4.525899545334023e-06, "loss": 1.9388, "step": 1088 }, { "epoch": 1.744, "grad_norm": 1.6018142700195312, "learning_rate": 4.416503771029201e-06, "loss": 1.8337, "step": 1090 }, { "epoch": 1.7471999999999999, "grad_norm": 1.6205910444259644, "learning_rate": 4.308385214390709e-06, "loss": 1.9551, "step": 1092 }, { "epoch": 1.7504, "grad_norm": 1.5405246019363403, "learning_rate": 4.2015469048248375e-06, "loss": 1.824, "step": 1094 }, { "epoch": 1.7536, "grad_norm": 1.6863925457000732, "learning_rate": 4.095991835866275e-06, "loss": 2.0388, "step": 1096 }, { "epoch": 1.7568000000000001, "grad_norm": 1.59629225730896, "learning_rate": 3.99172296509418e-06, "loss": 1.9307, "step": 1098 }, { "epoch": 1.76, "grad_norm": 1.6163777112960815, "learning_rate": 3.888743214049346e-06, "loss": 1.9041, "step": 1100 }, { "epoch": 1.7631999999999999, "grad_norm": 1.6829643249511719, "learning_rate": 3.7870554681523287e-06, "loss": 1.9971, "step": 1102 }, { "epoch": 1.7664, "grad_norm": 1.614225149154663, "learning_rate": 3.6866625766226293e-06, "loss": 1.9732, "step": 1104 }, { "epoch": 1.7696, "grad_norm": 1.631516933441162, "learning_rate": 3.587567352398796e-06, "loss": 1.8115, "step": 1106 }, { "epoch": 1.7728000000000002, "grad_norm": 1.5938745737075806, "learning_rate": 3.489772572059674e-06, "loss": 1.9675, "step": 1108 }, { "epoch": 1.776, "grad_norm": 1.580931544303894, "learning_rate": 3.393280975746588e-06, "loss": 1.9522, "step": 1110 }, { "epoch": 1.7792, "grad_norm": 1.5829232931137085, "learning_rate": 3.2980952670865317e-06, "loss": 1.7989, "step": 1112 }, { "epoch": 1.7824, "grad_norm": 1.695078730583191, "learning_rate": 3.2042181131164528e-06, "loss": 1.9354, "step": 1114 }, { "epoch": 1.7856, "grad_norm": 1.579819679260254, "learning_rate": 3.11165214420851e-06, "loss": 1.8468, "step": 1116 }, { "epoch": 1.7888, "grad_norm": 1.5718368291854858, "learning_rate": 3.020399953996389e-06, "loss": 1.8217, "step": 1118 }, { "epoch": 1.792, "grad_norm": 1.5997322797775269, "learning_rate": 2.9304640993025988e-06, "loss": 1.788, "step": 1120 }, { "epoch": 1.7952, "grad_norm": 1.5769128799438477, "learning_rate": 2.8418471000668523e-06, "loss": 1.8006, "step": 1122 }, { "epoch": 1.7984, "grad_norm": 1.602059006690979, "learning_rate": 2.7545514392754437e-06, "loss": 1.9391, "step": 1124 }, { "epoch": 1.8016, "grad_norm": 1.7443532943725586, "learning_rate": 2.6685795628917266e-06, "loss": 2.0419, "step": 1126 }, { "epoch": 1.8048, "grad_norm": 2.4460225105285645, "learning_rate": 2.5839338797875036e-06, "loss": 2.0401, "step": 1128 }, { "epoch": 1.808, "grad_norm": 1.5434504747390747, "learning_rate": 2.500616761675578e-06, "loss": 1.8181, "step": 1130 }, { "epoch": 1.8112, "grad_norm": 1.6132147312164307, "learning_rate": 2.41863054304331e-06, "loss": 2.0401, "step": 1132 }, { "epoch": 1.8144, "grad_norm": 1.6048423051834106, "learning_rate": 2.3379775210871648e-06, "loss": 1.8892, "step": 1134 }, { "epoch": 1.8176, "grad_norm": 1.5809746980667114, "learning_rate": 2.2586599556483734e-06, "loss": 1.8272, "step": 1136 }, { "epoch": 1.8208, "grad_norm": 1.7967019081115723, "learning_rate": 2.180680069149621e-06, "loss": 1.9311, "step": 1138 }, { "epoch": 1.8239999999999998, "grad_norm": 1.6240350008010864, "learning_rate": 2.104040046532768e-06, "loss": 1.9291, "step": 1140 }, { "epoch": 1.8272, "grad_norm": 1.6555086374282837, "learning_rate": 2.0287420351976063e-06, "loss": 1.9264, "step": 1142 }, { "epoch": 1.8304, "grad_norm": 1.6264870166778564, "learning_rate": 1.954788144941727e-06, "loss": 1.8148, "step": 1144 }, { "epoch": 1.8336000000000001, "grad_norm": 1.6269564628601074, "learning_rate": 1.8821804479013772e-06, "loss": 1.7891, "step": 1146 }, { "epoch": 1.8368, "grad_norm": 1.6172057390213013, "learning_rate": 1.81092097849343e-06, "loss": 2.0688, "step": 1148 }, { "epoch": 1.8399999999999999, "grad_norm": 1.6283061504364014, "learning_rate": 1.7410117333583498e-06, "loss": 1.9176, "step": 1150 }, { "epoch": 1.8432, "grad_norm": 1.6128772497177124, "learning_rate": 1.6724546713042577e-06, "loss": 1.7971, "step": 1152 }, { "epoch": 1.8464, "grad_norm": 1.603771686553955, "learning_rate": 1.6052517132520651e-06, "loss": 1.9016, "step": 1154 }, { "epoch": 1.8496000000000001, "grad_norm": 1.5155688524246216, "learning_rate": 1.5394047421816327e-06, "loss": 1.8073, "step": 1156 }, { "epoch": 1.8528, "grad_norm": 1.6564611196517944, "learning_rate": 1.4749156030790024e-06, "loss": 1.9445, "step": 1158 }, { "epoch": 1.8559999999999999, "grad_norm": 1.7882254123687744, "learning_rate": 1.4117861028847267e-06, "loss": 2.0257, "step": 1160 }, { "epoch": 1.8592, "grad_norm": 1.6352617740631104, "learning_rate": 1.3500180104432325e-06, "loss": 2.1369, "step": 1162 }, { "epoch": 1.8624, "grad_norm": 1.5756018161773682, "learning_rate": 1.2896130564532427e-06, "loss": 1.7884, "step": 1164 }, { "epoch": 1.8656000000000001, "grad_norm": 1.6122503280639648, "learning_rate": 1.2305729334192994e-06, "loss": 1.9276, "step": 1166 }, { "epoch": 1.8688, "grad_norm": 1.888556718826294, "learning_rate": 1.1728992956043238e-06, "loss": 2.0542, "step": 1168 }, { "epoch": 1.8719999999999999, "grad_norm": 1.6685359477996826, "learning_rate": 1.1165937589833087e-06, "loss": 1.9068, "step": 1170 }, { "epoch": 1.8752, "grad_norm": 1.647849202156067, "learning_rate": 1.061657901197971e-06, "loss": 1.905, "step": 1172 }, { "epoch": 1.8784, "grad_norm": 1.5802327394485474, "learning_rate": 1.008093261512616e-06, "loss": 1.8536, "step": 1174 }, { "epoch": 1.8816000000000002, "grad_norm": 1.6293805837631226, "learning_rate": 9.559013407709595e-07, "loss": 1.9312, "step": 1176 }, { "epoch": 1.8848, "grad_norm": 1.5595163106918335, "learning_rate": 9.050836013541009e-07, "loss": 1.9464, "step": 1178 }, { "epoch": 1.888, "grad_norm": 1.518751621246338, "learning_rate": 8.55641467139534e-07, "loss": 1.8809, "step": 1180 }, { "epoch": 1.8912, "grad_norm": 1.6101138591766357, "learning_rate": 8.075763234612622e-07, "loss": 1.8115, "step": 1182 }, { "epoch": 1.8944, "grad_norm": 1.6096612215042114, "learning_rate": 7.60889517070984e-07, "loss": 1.9574, "step": 1184 }, { "epoch": 1.8976, "grad_norm": 1.5596234798431396, "learning_rate": 7.155823561003361e-07, "loss": 1.9247, "step": 1186 }, { "epoch": 1.9008, "grad_norm": 1.63005530834198, "learning_rate": 6.716561100242658e-07, "loss": 1.9093, "step": 1188 }, { "epoch": 1.904, "grad_norm": 1.932976245880127, "learning_rate": 6.291120096254433e-07, "loss": 1.9061, "step": 1190 }, { "epoch": 1.9072, "grad_norm": 1.571678638458252, "learning_rate": 5.879512469598058e-07, "loss": 1.9505, "step": 1192 }, { "epoch": 1.9104, "grad_norm": 1.6436798572540283, "learning_rate": 5.481749753231124e-07, "loss": 1.977, "step": 1194 }, { "epoch": 1.9136, "grad_norm": 1.6677478551864624, "learning_rate": 5.097843092186583e-07, "loss": 1.9259, "step": 1196 }, { "epoch": 1.9167999999999998, "grad_norm": 1.6268632411956787, "learning_rate": 4.7278032432604425e-07, "loss": 1.9393, "step": 1198 }, { "epoch": 1.92, "grad_norm": 1.673553228378296, "learning_rate": 4.371640574710345e-07, "loss": 1.9702, "step": 1200 }, { "epoch": 1.9232, "grad_norm": 1.6786904335021973, "learning_rate": 4.0293650659650184e-07, "loss": 1.9319, "step": 1202 }, { "epoch": 1.9264000000000001, "grad_norm": 1.6758121252059937, "learning_rate": 3.7009863073446673e-07, "loss": 1.9173, "step": 1204 }, { "epoch": 1.9296, "grad_norm": 1.7078176736831665, "learning_rate": 3.386513499792354e-07, "loss": 2.1281, "step": 1206 }, { "epoch": 1.9327999999999999, "grad_norm": 1.6237930059432983, "learning_rate": 3.0859554546160965e-07, "loss": 1.8814, "step": 1208 }, { "epoch": 1.936, "grad_norm": 1.6799558401107788, "learning_rate": 2.7993205932420053e-07, "loss": 1.891, "step": 1210 }, { "epoch": 1.9392, "grad_norm": 1.6336385011672974, "learning_rate": 2.5266169469783105e-07, "loss": 1.9337, "step": 1212 }, { "epoch": 1.9424000000000001, "grad_norm": 1.6081047058105469, "learning_rate": 2.2678521567903176e-07, "loss": 1.8411, "step": 1214 }, { "epoch": 1.9456, "grad_norm": 1.64824640750885, "learning_rate": 2.023033473086411e-07, "loss": 1.9833, "step": 1216 }, { "epoch": 1.9487999999999999, "grad_norm": 1.6368495225906372, "learning_rate": 1.7921677555147177e-07, "loss": 1.8514, "step": 1218 }, { "epoch": 1.952, "grad_norm": 1.6810824871063232, "learning_rate": 1.5752614727712057e-07, "loss": 2.0633, "step": 1220 }, { "epoch": 1.9552, "grad_norm": 1.5862808227539062, "learning_rate": 1.3723207024180507e-07, "loss": 1.9975, "step": 1222 }, { "epoch": 1.9584000000000001, "grad_norm": 3.4000141620635986, "learning_rate": 1.1833511307136613e-07, "loss": 2.0093, "step": 1224 }, { "epoch": 1.9616, "grad_norm": 1.5674700736999512, "learning_rate": 1.0083580524531955e-07, "loss": 1.6954, "step": 1226 }, { "epoch": 1.9647999999999999, "grad_norm": 1.700727105140686, "learning_rate": 8.473463708202345e-08, "loss": 2.0088, "step": 1228 }, { "epoch": 1.968, "grad_norm": 1.6896897554397583, "learning_rate": 7.003205972494486e-08, "loss": 2.1039, "step": 1230 }, { "epoch": 1.9712, "grad_norm": 1.693572998046875, "learning_rate": 5.672848513000873e-08, "loss": 1.942, "step": 1232 }, { "epoch": 1.9744000000000002, "grad_norm": 1.662209153175354, "learning_rate": 4.482428605407374e-08, "loss": 1.8457, "step": 1234 }, { "epoch": 1.9776, "grad_norm": 1.6514451503753662, "learning_rate": 3.431979604445745e-08, "loss": 1.9565, "step": 1236 }, { "epoch": 1.9808, "grad_norm": 1.5132412910461426, "learning_rate": 2.521530942962702e-08, "loss": 1.66, "step": 1238 }, { "epoch": 1.984, "grad_norm": 1.609984040260315, "learning_rate": 1.7511081310922495e-08, "loss": 1.8459, "step": 1240 }, { "epoch": 1.9872, "grad_norm": 1.6186810731887817, "learning_rate": 1.1207327555429192e-08, "loss": 1.9105, "step": 1242 }, { "epoch": 1.9904, "grad_norm": 1.6339943408966064, "learning_rate": 6.304224789910329e-09, "loss": 1.867, "step": 1244 }, { "epoch": 1.9936, "grad_norm": 1.6734284162521362, "learning_rate": 2.801910395877627e-09, "loss": 1.936, "step": 1246 }, { "epoch": 1.9968, "grad_norm": 1.6463286876678467, "learning_rate": 7.004825057277398e-10, "loss": 2.0955, "step": 1248 }, { "epoch": 2.0, "grad_norm": 1.7489490509033203, "learning_rate": 0.0, "loss": 1.8566, "step": 1250 } ], "logging_steps": 2, "max_steps": 1250, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 625, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.1154802941978214e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }