{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9998780636507743, "eval_steps": 500, "global_step": 1025, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009754907938056335, "grad_norm": 4.439117513053022, "learning_rate": 9.615384615384617e-08, "loss": 0.3691, "step": 1 }, { "epoch": 0.001950981587611267, "grad_norm": 4.665976578311697, "learning_rate": 1.9230769230769234e-07, "loss": 0.3804, "step": 2 }, { "epoch": 0.0029264723814169005, "grad_norm": 4.458573970698933, "learning_rate": 2.884615384615385e-07, "loss": 0.3682, "step": 3 }, { "epoch": 0.003901963175222534, "grad_norm": 4.21755765446574, "learning_rate": 3.846153846153847e-07, "loss": 0.3567, "step": 4 }, { "epoch": 0.004877453969028167, "grad_norm": 4.1944148540591195, "learning_rate": 4.807692307692308e-07, "loss": 0.3773, "step": 5 }, { "epoch": 0.005852944762833801, "grad_norm": 3.9774353472382957, "learning_rate": 5.76923076923077e-07, "loss": 0.3502, "step": 6 }, { "epoch": 0.006828435556639434, "grad_norm": 3.553247347541323, "learning_rate": 6.730769230769231e-07, "loss": 0.3614, "step": 7 }, { "epoch": 0.007803926350445068, "grad_norm": 2.949302152008964, "learning_rate": 7.692307692307694e-07, "loss": 0.3463, "step": 8 }, { "epoch": 0.008779417144250701, "grad_norm": 2.799468109303019, "learning_rate": 8.653846153846154e-07, "loss": 0.3582, "step": 9 }, { "epoch": 0.009754907938056334, "grad_norm": 2.6971355767992464, "learning_rate": 9.615384615384617e-07, "loss": 0.3277, "step": 10 }, { "epoch": 0.010730398731861967, "grad_norm": 2.2435805544189256, "learning_rate": 1.0576923076923078e-06, "loss": 0.3411, "step": 11 }, { "epoch": 0.011705889525667602, "grad_norm": 2.3858778520973933, "learning_rate": 1.153846153846154e-06, "loss": 0.3323, "step": 12 }, { "epoch": 0.012681380319473235, "grad_norm": 2.325354329209325, "learning_rate": 1.25e-06, "loss": 0.3201, "step": 13 }, { "epoch": 0.013656871113278868, "grad_norm": 2.268837198744112, "learning_rate": 1.3461538461538462e-06, "loss": 0.3458, "step": 14 }, { "epoch": 0.014632361907084501, "grad_norm": 2.2354597650236263, "learning_rate": 1.4423076923076922e-06, "loss": 0.3057, "step": 15 }, { "epoch": 0.015607852700890136, "grad_norm": 2.4354294587165968, "learning_rate": 1.5384615384615387e-06, "loss": 0.2974, "step": 16 }, { "epoch": 0.016583343494695767, "grad_norm": 1.9801626818691969, "learning_rate": 1.6346153846153848e-06, "loss": 0.2738, "step": 17 }, { "epoch": 0.017558834288501402, "grad_norm": 1.7527920071956236, "learning_rate": 1.7307692307692308e-06, "loss": 0.268, "step": 18 }, { "epoch": 0.018534325082307037, "grad_norm": 1.6175943655326306, "learning_rate": 1.826923076923077e-06, "loss": 0.2851, "step": 19 }, { "epoch": 0.01950981587611267, "grad_norm": 1.8343780524193107, "learning_rate": 1.9230769230769234e-06, "loss": 0.2773, "step": 20 }, { "epoch": 0.020485306669918303, "grad_norm": 2.0687178014460055, "learning_rate": 2.0192307692307692e-06, "loss": 0.279, "step": 21 }, { "epoch": 0.021460797463723934, "grad_norm": 1.850604452835117, "learning_rate": 2.1153846153846155e-06, "loss": 0.2743, "step": 22 }, { "epoch": 0.02243628825752957, "grad_norm": 2.149994632477382, "learning_rate": 2.211538461538462e-06, "loss": 0.2694, "step": 23 }, { "epoch": 0.023411779051335204, "grad_norm": 1.8792389072954687, "learning_rate": 2.307692307692308e-06, "loss": 0.2644, "step": 24 }, { "epoch": 0.024387269845140835, "grad_norm": 1.457358625174777, "learning_rate": 2.403846153846154e-06, "loss": 0.2419, "step": 25 }, { "epoch": 0.02536276063894647, "grad_norm": 1.7014683746851735, "learning_rate": 2.5e-06, "loss": 0.2544, "step": 26 }, { "epoch": 0.026338251432752105, "grad_norm": 1.3815851595211681, "learning_rate": 2.5961538461538465e-06, "loss": 0.223, "step": 27 }, { "epoch": 0.027313742226557736, "grad_norm": 1.3007184555893962, "learning_rate": 2.6923076923076923e-06, "loss": 0.2456, "step": 28 }, { "epoch": 0.02828923302036337, "grad_norm": 1.4472785639522512, "learning_rate": 2.7884615384615386e-06, "loss": 0.2516, "step": 29 }, { "epoch": 0.029264723814169002, "grad_norm": 1.3652941522006306, "learning_rate": 2.8846153846153845e-06, "loss": 0.2416, "step": 30 }, { "epoch": 0.030240214607974637, "grad_norm": 1.2522050977996877, "learning_rate": 2.980769230769231e-06, "loss": 0.2329, "step": 31 }, { "epoch": 0.031215705401780272, "grad_norm": 1.1304756888064047, "learning_rate": 3.0769230769230774e-06, "loss": 0.2217, "step": 32 }, { "epoch": 0.03219119619558591, "grad_norm": 1.4050382297283845, "learning_rate": 3.1730769230769233e-06, "loss": 0.2309, "step": 33 }, { "epoch": 0.033166686989391535, "grad_norm": 1.0865416261086687, "learning_rate": 3.2692307692307696e-06, "loss": 0.2217, "step": 34 }, { "epoch": 0.03414217778319717, "grad_norm": 1.303187815201321, "learning_rate": 3.365384615384616e-06, "loss": 0.2338, "step": 35 }, { "epoch": 0.035117668577002804, "grad_norm": 1.2021321704753516, "learning_rate": 3.4615384615384617e-06, "loss": 0.2137, "step": 36 }, { "epoch": 0.03609315937080844, "grad_norm": 1.1561740561669396, "learning_rate": 3.557692307692308e-06, "loss": 0.2162, "step": 37 }, { "epoch": 0.037068650164614074, "grad_norm": 1.287983442354845, "learning_rate": 3.653846153846154e-06, "loss": 0.228, "step": 38 }, { "epoch": 0.0380441409584197, "grad_norm": 1.0789644716188591, "learning_rate": 3.7500000000000005e-06, "loss": 0.2106, "step": 39 }, { "epoch": 0.03901963175222534, "grad_norm": 1.1684419312705403, "learning_rate": 3.846153846153847e-06, "loss": 0.218, "step": 40 }, { "epoch": 0.03999512254603097, "grad_norm": 1.1195832124738565, "learning_rate": 3.942307692307692e-06, "loss": 0.2032, "step": 41 }, { "epoch": 0.040970613339836606, "grad_norm": 1.0236100570446467, "learning_rate": 4.0384615384615385e-06, "loss": 0.217, "step": 42 }, { "epoch": 0.04194610413364224, "grad_norm": 1.182743593377574, "learning_rate": 4.134615384615385e-06, "loss": 0.2215, "step": 43 }, { "epoch": 0.04292159492744787, "grad_norm": 1.1077853276119973, "learning_rate": 4.230769230769231e-06, "loss": 0.2138, "step": 44 }, { "epoch": 0.043897085721253504, "grad_norm": 1.1863700432181472, "learning_rate": 4.326923076923077e-06, "loss": 0.2113, "step": 45 }, { "epoch": 0.04487257651505914, "grad_norm": 1.1214318092951887, "learning_rate": 4.423076923076924e-06, "loss": 0.2106, "step": 46 }, { "epoch": 0.04584806730886477, "grad_norm": 1.0805493864146463, "learning_rate": 4.51923076923077e-06, "loss": 0.2104, "step": 47 }, { "epoch": 0.04682355810267041, "grad_norm": 1.0336295929522208, "learning_rate": 4.615384615384616e-06, "loss": 0.1981, "step": 48 }, { "epoch": 0.04779904889647604, "grad_norm": 1.119869757376067, "learning_rate": 4.711538461538462e-06, "loss": 0.2014, "step": 49 }, { "epoch": 0.04877453969028167, "grad_norm": 1.030817002409712, "learning_rate": 4.807692307692308e-06, "loss": 0.1996, "step": 50 }, { "epoch": 0.049750030484087306, "grad_norm": 1.0055854671916695, "learning_rate": 4.903846153846154e-06, "loss": 0.2052, "step": 51 }, { "epoch": 0.05072552127789294, "grad_norm": 1.107516098502413, "learning_rate": 5e-06, "loss": 0.2168, "step": 52 }, { "epoch": 0.051701012071698575, "grad_norm": 1.0591849818221413, "learning_rate": 4.9999869688212956e-06, "loss": 0.2099, "step": 53 }, { "epoch": 0.05267650286550421, "grad_norm": 1.0486177613460919, "learning_rate": 4.999947875421032e-06, "loss": 0.1978, "step": 54 }, { "epoch": 0.05365199365930984, "grad_norm": 1.0333038397384442, "learning_rate": 4.999882720206755e-06, "loss": 0.1934, "step": 55 }, { "epoch": 0.05462748445311547, "grad_norm": 1.0394851360953357, "learning_rate": 4.999791503857704e-06, "loss": 0.2019, "step": 56 }, { "epoch": 0.05560297524692111, "grad_norm": 1.1046443338767855, "learning_rate": 4.999674227324805e-06, "loss": 0.2044, "step": 57 }, { "epoch": 0.05657846604072674, "grad_norm": 1.2371807932970489, "learning_rate": 4.9995308918306595e-06, "loss": 0.1972, "step": 58 }, { "epoch": 0.05755395683453238, "grad_norm": 0.9905094942349434, "learning_rate": 4.99936149886953e-06, "loss": 0.19, "step": 59 }, { "epoch": 0.058529447628338005, "grad_norm": 1.005308785484947, "learning_rate": 4.999166050207331e-06, "loss": 0.1838, "step": 60 }, { "epoch": 0.05950493842214364, "grad_norm": 1.0197164862121335, "learning_rate": 4.9989445478816e-06, "loss": 0.1906, "step": 61 }, { "epoch": 0.060480429215949275, "grad_norm": 1.112996579159861, "learning_rate": 4.99869699420149e-06, "loss": 0.1949, "step": 62 }, { "epoch": 0.06145592000975491, "grad_norm": 0.9210560581443838, "learning_rate": 4.998423391747731e-06, "loss": 0.1869, "step": 63 }, { "epoch": 0.062431410803560544, "grad_norm": 0.994708396507738, "learning_rate": 4.998123743372615e-06, "loss": 0.1832, "step": 64 }, { "epoch": 0.06340690159736617, "grad_norm": 1.000252041680365, "learning_rate": 4.997798052199959e-06, "loss": 0.2011, "step": 65 }, { "epoch": 0.06438239239117181, "grad_norm": 1.001566090006205, "learning_rate": 4.9974463216250735e-06, "loss": 0.1851, "step": 66 }, { "epoch": 0.06535788318497744, "grad_norm": 1.0035711549242645, "learning_rate": 4.9970685553147316e-06, "loss": 0.1959, "step": 67 }, { "epoch": 0.06633337397878307, "grad_norm": 1.0321021956412775, "learning_rate": 4.996664757207124e-06, "loss": 0.1948, "step": 68 }, { "epoch": 0.06730886477258871, "grad_norm": 1.1196394074729383, "learning_rate": 4.996234931511823e-06, "loss": 0.1959, "step": 69 }, { "epoch": 0.06828435556639434, "grad_norm": 1.204621557285872, "learning_rate": 4.995779082709739e-06, "loss": 0.1962, "step": 70 }, { "epoch": 0.06925984636019998, "grad_norm": 1.0177518163557318, "learning_rate": 4.995297215553067e-06, "loss": 0.1806, "step": 71 }, { "epoch": 0.07023533715400561, "grad_norm": 1.0847615883655475, "learning_rate": 4.994789335065245e-06, "loss": 0.2044, "step": 72 }, { "epoch": 0.07121082794781124, "grad_norm": 1.168873122167307, "learning_rate": 4.9942554465409e-06, "loss": 0.1934, "step": 73 }, { "epoch": 0.07218631874161688, "grad_norm": 1.019104243672844, "learning_rate": 4.993695555545789e-06, "loss": 0.1905, "step": 74 }, { "epoch": 0.0731618095354225, "grad_norm": 1.1308032239714556, "learning_rate": 4.993109667916742e-06, "loss": 0.193, "step": 75 }, { "epoch": 0.07413730032922815, "grad_norm": 1.1401766992269902, "learning_rate": 4.992497789761606e-06, "loss": 0.1793, "step": 76 }, { "epoch": 0.07511279112303378, "grad_norm": 1.1144999371666817, "learning_rate": 4.991859927459174e-06, "loss": 0.2023, "step": 77 }, { "epoch": 0.0760882819168394, "grad_norm": 1.191890901059724, "learning_rate": 4.991196087659125e-06, "loss": 0.1858, "step": 78 }, { "epoch": 0.07706377271064505, "grad_norm": 1.11445392411597, "learning_rate": 4.990506277281952e-06, "loss": 0.1843, "step": 79 }, { "epoch": 0.07803926350445067, "grad_norm": 1.1630978983896232, "learning_rate": 4.989790503518888e-06, "loss": 0.1775, "step": 80 }, { "epoch": 0.07901475429825632, "grad_norm": 1.3514121202700597, "learning_rate": 4.989048773831834e-06, "loss": 0.1879, "step": 81 }, { "epoch": 0.07999024509206194, "grad_norm": 1.0138569119140257, "learning_rate": 4.988281095953279e-06, "loss": 0.1712, "step": 82 }, { "epoch": 0.08096573588586757, "grad_norm": 1.2876601226086022, "learning_rate": 4.987487477886221e-06, "loss": 0.1806, "step": 83 }, { "epoch": 0.08194122667967321, "grad_norm": 1.0842905018940538, "learning_rate": 4.986667927904084e-06, "loss": 0.172, "step": 84 }, { "epoch": 0.08291671747347884, "grad_norm": 1.0011346482797712, "learning_rate": 4.985822454550629e-06, "loss": 0.1775, "step": 85 }, { "epoch": 0.08389220826728448, "grad_norm": 1.322374107432227, "learning_rate": 4.984951066639869e-06, "loss": 0.1872, "step": 86 }, { "epoch": 0.08486769906109011, "grad_norm": 1.0531176300373897, "learning_rate": 4.984053773255971e-06, "loss": 0.1848, "step": 87 }, { "epoch": 0.08584318985489574, "grad_norm": 1.0696103709835127, "learning_rate": 4.9831305837531684e-06, "loss": 0.1789, "step": 88 }, { "epoch": 0.08681868064870138, "grad_norm": 1.048271406336639, "learning_rate": 4.98218150775566e-06, "loss": 0.1763, "step": 89 }, { "epoch": 0.08779417144250701, "grad_norm": 0.9981980018547214, "learning_rate": 4.9812065551575075e-06, "loss": 0.1746, "step": 90 }, { "epoch": 0.08876966223631265, "grad_norm": 1.0986911182229246, "learning_rate": 4.9802057361225375e-06, "loss": 0.2019, "step": 91 }, { "epoch": 0.08974515303011828, "grad_norm": 1.0482728855361627, "learning_rate": 4.97917906108423e-06, "loss": 0.1725, "step": 92 }, { "epoch": 0.0907206438239239, "grad_norm": 1.0599473411074543, "learning_rate": 4.978126540745615e-06, "loss": 0.1804, "step": 93 }, { "epoch": 0.09169613461772955, "grad_norm": 1.078553827467801, "learning_rate": 4.977048186079155e-06, "loss": 0.2011, "step": 94 }, { "epoch": 0.09267162541153517, "grad_norm": 1.0444452777467914, "learning_rate": 4.975944008326638e-06, "loss": 0.1835, "step": 95 }, { "epoch": 0.09364711620534082, "grad_norm": 1.0791056951072164, "learning_rate": 4.974814018999053e-06, "loss": 0.1824, "step": 96 }, { "epoch": 0.09462260699914644, "grad_norm": 1.078717583871865, "learning_rate": 4.973658229876476e-06, "loss": 0.189, "step": 97 }, { "epoch": 0.09559809779295209, "grad_norm": 1.124139067655016, "learning_rate": 4.97247665300794e-06, "loss": 0.1832, "step": 98 }, { "epoch": 0.09657358858675771, "grad_norm": 1.0195889033855914, "learning_rate": 4.971269300711318e-06, "loss": 0.1766, "step": 99 }, { "epoch": 0.09754907938056334, "grad_norm": 1.1122862478289546, "learning_rate": 4.970036185573189e-06, "loss": 0.1805, "step": 100 }, { "epoch": 0.09852457017436898, "grad_norm": 1.1044026960421833, "learning_rate": 4.968777320448707e-06, "loss": 0.1836, "step": 101 }, { "epoch": 0.09950006096817461, "grad_norm": 1.204423675193318, "learning_rate": 4.96749271846147e-06, "loss": 0.1767, "step": 102 }, { "epoch": 0.10047555176198025, "grad_norm": 1.0248442266116176, "learning_rate": 4.96618239300338e-06, "loss": 0.1912, "step": 103 }, { "epoch": 0.10145104255578588, "grad_norm": 1.1740907364465507, "learning_rate": 4.964846357734505e-06, "loss": 0.1726, "step": 104 }, { "epoch": 0.10242653334959151, "grad_norm": 0.9983984379050335, "learning_rate": 4.963484626582937e-06, "loss": 0.1787, "step": 105 }, { "epoch": 0.10340202414339715, "grad_norm": 1.0692395971458715, "learning_rate": 4.9620972137446456e-06, "loss": 0.179, "step": 106 }, { "epoch": 0.10437751493720278, "grad_norm": 1.0955109652052737, "learning_rate": 4.96068413368333e-06, "loss": 0.1765, "step": 107 }, { "epoch": 0.10535300573100842, "grad_norm": 1.0684942634997572, "learning_rate": 4.959245401130269e-06, "loss": 0.172, "step": 108 }, { "epoch": 0.10632849652481405, "grad_norm": 0.9995901623406969, "learning_rate": 4.957781031084169e-06, "loss": 0.1805, "step": 109 }, { "epoch": 0.10730398731861968, "grad_norm": 0.9323445364396948, "learning_rate": 4.956291038811003e-06, "loss": 0.1853, "step": 110 }, { "epoch": 0.10827947811242532, "grad_norm": 1.1033013476735185, "learning_rate": 4.954775439843855e-06, "loss": 0.1873, "step": 111 }, { "epoch": 0.10925496890623095, "grad_norm": 1.1691417741452548, "learning_rate": 4.9532342499827584e-06, "loss": 0.185, "step": 112 }, { "epoch": 0.11023045970003659, "grad_norm": 1.1918610910255243, "learning_rate": 4.95166748529453e-06, "loss": 0.1892, "step": 113 }, { "epoch": 0.11120595049384221, "grad_norm": 1.1014111858940245, "learning_rate": 4.950075162112602e-06, "loss": 0.1735, "step": 114 }, { "epoch": 0.11218144128764784, "grad_norm": 1.216567039997431, "learning_rate": 4.9484572970368516e-06, "loss": 0.1838, "step": 115 }, { "epoch": 0.11315693208145348, "grad_norm": 1.1097171959187067, "learning_rate": 4.946813906933432e-06, "loss": 0.1874, "step": 116 }, { "epoch": 0.11413242287525911, "grad_norm": 0.967977115534241, "learning_rate": 4.94514500893459e-06, "loss": 0.1738, "step": 117 }, { "epoch": 0.11510791366906475, "grad_norm": 0.9657457430010942, "learning_rate": 4.943450620438491e-06, "loss": 0.1752, "step": 118 }, { "epoch": 0.11608340446287038, "grad_norm": 0.9166596111467156, "learning_rate": 4.941730759109041e-06, "loss": 0.1818, "step": 119 }, { "epoch": 0.11705889525667601, "grad_norm": 0.9450152131596894, "learning_rate": 4.939985442875695e-06, "loss": 0.1799, "step": 120 }, { "epoch": 0.11803438605048165, "grad_norm": 0.938177849072337, "learning_rate": 4.938214689933276e-06, "loss": 0.1735, "step": 121 }, { "epoch": 0.11900987684428728, "grad_norm": 1.0415880049786728, "learning_rate": 4.93641851874178e-06, "loss": 0.1868, "step": 122 }, { "epoch": 0.11998536763809292, "grad_norm": 0.9746213204544258, "learning_rate": 4.9345969480261925e-06, "loss": 0.1829, "step": 123 }, { "epoch": 0.12096085843189855, "grad_norm": 0.9444433221181193, "learning_rate": 4.932749996776282e-06, "loss": 0.1731, "step": 124 }, { "epoch": 0.12193634922570418, "grad_norm": 1.0120487885636702, "learning_rate": 4.93087768424641e-06, "loss": 0.1886, "step": 125 }, { "epoch": 0.12291184001950982, "grad_norm": 0.9837069724067731, "learning_rate": 4.928980029955329e-06, "loss": 0.1821, "step": 126 }, { "epoch": 0.12388733081331545, "grad_norm": 0.9425235884710698, "learning_rate": 4.927057053685975e-06, "loss": 0.1801, "step": 127 }, { "epoch": 0.12486282160712109, "grad_norm": 1.0113092843870408, "learning_rate": 4.925108775485269e-06, "loss": 0.1796, "step": 128 }, { "epoch": 0.12583831240092672, "grad_norm": 1.0163088829406406, "learning_rate": 4.923135215663897e-06, "loss": 0.1654, "step": 129 }, { "epoch": 0.12681380319473234, "grad_norm": 1.164966417682963, "learning_rate": 4.921136394796109e-06, "loss": 0.1803, "step": 130 }, { "epoch": 0.12778929398853797, "grad_norm": 1.1687221264442198, "learning_rate": 4.919112333719498e-06, "loss": 0.1878, "step": 131 }, { "epoch": 0.12876478478234363, "grad_norm": 0.9888510085952801, "learning_rate": 4.9170630535347866e-06, "loss": 0.1771, "step": 132 }, { "epoch": 0.12974027557614926, "grad_norm": 1.2032250564065257, "learning_rate": 4.914988575605602e-06, "loss": 0.1893, "step": 133 }, { "epoch": 0.13071576636995488, "grad_norm": 0.9809911156410793, "learning_rate": 4.91288892155826e-06, "loss": 0.1728, "step": 134 }, { "epoch": 0.1316912571637605, "grad_norm": 1.2533075523741326, "learning_rate": 4.910764113281533e-06, "loss": 0.1836, "step": 135 }, { "epoch": 0.13266674795756614, "grad_norm": 1.0029237484325686, "learning_rate": 4.908614172926426e-06, "loss": 0.1757, "step": 136 }, { "epoch": 0.1336422387513718, "grad_norm": 1.0323575219354129, "learning_rate": 4.906439122905946e-06, "loss": 0.161, "step": 137 }, { "epoch": 0.13461772954517742, "grad_norm": 0.9974877118091979, "learning_rate": 4.904238985894864e-06, "loss": 0.1644, "step": 138 }, { "epoch": 0.13559322033898305, "grad_norm": 0.8968544196822601, "learning_rate": 4.902013784829483e-06, "loss": 0.1541, "step": 139 }, { "epoch": 0.13656871113278868, "grad_norm": 1.067306106970932, "learning_rate": 4.899763542907399e-06, "loss": 0.1689, "step": 140 }, { "epoch": 0.1375442019265943, "grad_norm": 0.9831550732650544, "learning_rate": 4.897488283587253e-06, "loss": 0.1764, "step": 141 }, { "epoch": 0.13851969272039996, "grad_norm": 1.098512924207099, "learning_rate": 4.895188030588495e-06, "loss": 0.1671, "step": 142 }, { "epoch": 0.1394951835142056, "grad_norm": 0.9813597480687253, "learning_rate": 4.892862807891131e-06, "loss": 0.1723, "step": 143 }, { "epoch": 0.14047067430801122, "grad_norm": 0.9059688793766486, "learning_rate": 4.890512639735475e-06, "loss": 0.1585, "step": 144 }, { "epoch": 0.14144616510181685, "grad_norm": 0.9421090522610057, "learning_rate": 4.888137550621897e-06, "loss": 0.1667, "step": 145 }, { "epoch": 0.14242165589562247, "grad_norm": 0.9984203476569828, "learning_rate": 4.885737565310565e-06, "loss": 0.1748, "step": 146 }, { "epoch": 0.14339714668942813, "grad_norm": 0.9344709716653912, "learning_rate": 4.883312708821188e-06, "loss": 0.172, "step": 147 }, { "epoch": 0.14437263748323376, "grad_norm": 0.995510093417273, "learning_rate": 4.880863006432758e-06, "loss": 0.1704, "step": 148 }, { "epoch": 0.14534812827703938, "grad_norm": 0.9658034107481427, "learning_rate": 4.878388483683281e-06, "loss": 0.1779, "step": 149 }, { "epoch": 0.146323619070845, "grad_norm": 1.0283682187427396, "learning_rate": 4.8758891663695165e-06, "loss": 0.1733, "step": 150 }, { "epoch": 0.14729910986465064, "grad_norm": 1.0848455109349466, "learning_rate": 4.873365080546706e-06, "loss": 0.1883, "step": 151 }, { "epoch": 0.1482746006584563, "grad_norm": 1.0034493473399255, "learning_rate": 4.870816252528298e-06, "loss": 0.1725, "step": 152 }, { "epoch": 0.14925009145226192, "grad_norm": 1.0714206307709666, "learning_rate": 4.868242708885681e-06, "loss": 0.1855, "step": 153 }, { "epoch": 0.15022558224606755, "grad_norm": 1.0894809817408557, "learning_rate": 4.8656444764479005e-06, "loss": 0.1815, "step": 154 }, { "epoch": 0.15120107303987318, "grad_norm": 1.0188695855109409, "learning_rate": 4.863021582301381e-06, "loss": 0.171, "step": 155 }, { "epoch": 0.1521765638336788, "grad_norm": 1.0015265493016945, "learning_rate": 4.860374053789643e-06, "loss": 0.1703, "step": 156 }, { "epoch": 0.15315205462748446, "grad_norm": 0.9798681079088354, "learning_rate": 4.857701918513023e-06, "loss": 0.1762, "step": 157 }, { "epoch": 0.1541275454212901, "grad_norm": 0.9770978700089492, "learning_rate": 4.855005204328378e-06, "loss": 0.159, "step": 158 }, { "epoch": 0.15510303621509572, "grad_norm": 1.0074537647270105, "learning_rate": 4.8522839393487976e-06, "loss": 0.1701, "step": 159 }, { "epoch": 0.15607852700890135, "grad_norm": 0.983344185894382, "learning_rate": 4.849538151943316e-06, "loss": 0.1742, "step": 160 }, { "epoch": 0.15705401780270697, "grad_norm": 0.9614400815443347, "learning_rate": 4.84676787073661e-06, "loss": 0.173, "step": 161 }, { "epoch": 0.15802950859651263, "grad_norm": 0.9526818664956649, "learning_rate": 4.843973124608703e-06, "loss": 0.1732, "step": 162 }, { "epoch": 0.15900499939031826, "grad_norm": 0.8985619765911587, "learning_rate": 4.841153942694664e-06, "loss": 0.1665, "step": 163 }, { "epoch": 0.15998049018412389, "grad_norm": 0.9257494791929499, "learning_rate": 4.838310354384304e-06, "loss": 0.1684, "step": 164 }, { "epoch": 0.1609559809779295, "grad_norm": 0.9783281983993328, "learning_rate": 4.835442389321867e-06, "loss": 0.1692, "step": 165 }, { "epoch": 0.16193147177173514, "grad_norm": 0.8866263340546827, "learning_rate": 4.832550077405727e-06, "loss": 0.1563, "step": 166 }, { "epoch": 0.1629069625655408, "grad_norm": 1.0759800513612392, "learning_rate": 4.829633448788072e-06, "loss": 0.1796, "step": 167 }, { "epoch": 0.16388245335934642, "grad_norm": 1.0415739389017458, "learning_rate": 4.826692533874586e-06, "loss": 0.1736, "step": 168 }, { "epoch": 0.16485794415315205, "grad_norm": 1.0321014340070922, "learning_rate": 4.823727363324142e-06, "loss": 0.1763, "step": 169 }, { "epoch": 0.16583343494695768, "grad_norm": 1.149401130099103, "learning_rate": 4.820737968048471e-06, "loss": 0.1735, "step": 170 }, { "epoch": 0.1668089257407633, "grad_norm": 0.9668062539942579, "learning_rate": 4.8177243792118515e-06, "loss": 0.1843, "step": 171 }, { "epoch": 0.16778441653456896, "grad_norm": 1.0145067381113042, "learning_rate": 4.8146866282307725e-06, "loss": 0.1719, "step": 172 }, { "epoch": 0.1687599073283746, "grad_norm": 0.9146830744890082, "learning_rate": 4.811624746773616e-06, "loss": 0.1597, "step": 173 }, { "epoch": 0.16973539812218022, "grad_norm": 0.9826930086582164, "learning_rate": 4.808538766760321e-06, "loss": 0.1539, "step": 174 }, { "epoch": 0.17071088891598585, "grad_norm": 0.9622129610913185, "learning_rate": 4.805428720362054e-06, "loss": 0.1682, "step": 175 }, { "epoch": 0.17168637970979148, "grad_norm": 1.200349435662473, "learning_rate": 4.8022946400008705e-06, "loss": 0.1687, "step": 176 }, { "epoch": 0.17266187050359713, "grad_norm": 1.0757087822872222, "learning_rate": 4.79913655834938e-06, "loss": 0.1812, "step": 177 }, { "epoch": 0.17363736129740276, "grad_norm": 1.0368299327714814, "learning_rate": 4.795954508330403e-06, "loss": 0.1721, "step": 178 }, { "epoch": 0.1746128520912084, "grad_norm": 1.1790584459327953, "learning_rate": 4.79274852311663e-06, "loss": 0.1708, "step": 179 }, { "epoch": 0.17558834288501401, "grad_norm": 0.9846847616375829, "learning_rate": 4.7895186361302736e-06, "loss": 0.1639, "step": 180 }, { "epoch": 0.17656383367881964, "grad_norm": 0.9524671440100206, "learning_rate": 4.786264881042722e-06, "loss": 0.1544, "step": 181 }, { "epoch": 0.1775393244726253, "grad_norm": 1.0337867693599667, "learning_rate": 4.782987291774186e-06, "loss": 0.1586, "step": 182 }, { "epoch": 0.17851481526643093, "grad_norm": 0.955523189772594, "learning_rate": 4.779685902493346e-06, "loss": 0.1703, "step": 183 }, { "epoch": 0.17949030606023655, "grad_norm": 0.8949175563809436, "learning_rate": 4.776360747616999e-06, "loss": 0.1536, "step": 184 }, { "epoch": 0.18046579685404218, "grad_norm": 1.0812607035187014, "learning_rate": 4.773011861809694e-06, "loss": 0.1729, "step": 185 }, { "epoch": 0.1814412876478478, "grad_norm": 0.9776277191751103, "learning_rate": 4.769639279983372e-06, "loss": 0.1745, "step": 186 }, { "epoch": 0.18241677844165347, "grad_norm": 1.0274757082533237, "learning_rate": 4.76624303729701e-06, "loss": 0.1651, "step": 187 }, { "epoch": 0.1833922692354591, "grad_norm": 1.0244957983233014, "learning_rate": 4.762823169156242e-06, "loss": 0.1642, "step": 188 }, { "epoch": 0.18436776002926472, "grad_norm": 0.9535466834456596, "learning_rate": 4.7593797112129995e-06, "loss": 0.1701, "step": 189 }, { "epoch": 0.18534325082307035, "grad_norm": 1.0223048067541962, "learning_rate": 4.755912699365135e-06, "loss": 0.1619, "step": 190 }, { "epoch": 0.18631874161687598, "grad_norm": 0.9985286606007027, "learning_rate": 4.752422169756048e-06, "loss": 0.1541, "step": 191 }, { "epoch": 0.18729423241068163, "grad_norm": 0.9550222906892099, "learning_rate": 4.748908158774312e-06, "loss": 0.1583, "step": 192 }, { "epoch": 0.18826972320448726, "grad_norm": 1.0220426764043464, "learning_rate": 4.745370703053291e-06, "loss": 0.1685, "step": 193 }, { "epoch": 0.1892452139982929, "grad_norm": 0.9903476130292974, "learning_rate": 4.741809839470758e-06, "loss": 0.1624, "step": 194 }, { "epoch": 0.19022070479209852, "grad_norm": 0.9340244675396043, "learning_rate": 4.738225605148514e-06, "loss": 0.171, "step": 195 }, { "epoch": 0.19119619558590417, "grad_norm": 0.9933142717509476, "learning_rate": 4.734618037451997e-06, "loss": 0.1677, "step": 196 }, { "epoch": 0.1921716863797098, "grad_norm": 1.1481162666912414, "learning_rate": 4.730987173989894e-06, "loss": 0.1654, "step": 197 }, { "epoch": 0.19314717717351543, "grad_norm": 1.0010562488790584, "learning_rate": 4.727333052613749e-06, "loss": 0.1689, "step": 198 }, { "epoch": 0.19412266796732106, "grad_norm": 0.943204986199299, "learning_rate": 4.7236557114175705e-06, "loss": 0.153, "step": 199 }, { "epoch": 0.19509815876112668, "grad_norm": 1.004212156065797, "learning_rate": 4.71995518873743e-06, "loss": 0.1738, "step": 200 }, { "epoch": 0.19607364955493234, "grad_norm": 0.9943567195079659, "learning_rate": 4.716231523151065e-06, "loss": 0.1621, "step": 201 }, { "epoch": 0.19704914034873797, "grad_norm": 0.9438877027311198, "learning_rate": 4.712484753477478e-06, "loss": 0.1608, "step": 202 }, { "epoch": 0.1980246311425436, "grad_norm": 1.0575615752386678, "learning_rate": 4.708714918776527e-06, "loss": 0.1712, "step": 203 }, { "epoch": 0.19900012193634922, "grad_norm": 0.9584987238953627, "learning_rate": 4.704922058348526e-06, "loss": 0.1586, "step": 204 }, { "epoch": 0.19997561273015485, "grad_norm": 0.9822024229285666, "learning_rate": 4.701106211733827e-06, "loss": 0.1633, "step": 205 }, { "epoch": 0.2009511035239605, "grad_norm": 1.0567476044736286, "learning_rate": 4.697267418712415e-06, "loss": 0.1682, "step": 206 }, { "epoch": 0.20192659431776613, "grad_norm": 0.9248879785588529, "learning_rate": 4.693405719303487e-06, "loss": 0.158, "step": 207 }, { "epoch": 0.20290208511157176, "grad_norm": 1.0157347860608708, "learning_rate": 4.689521153765039e-06, "loss": 0.1585, "step": 208 }, { "epoch": 0.2038775759053774, "grad_norm": 1.0718965104259908, "learning_rate": 4.685613762593446e-06, "loss": 0.1664, "step": 209 }, { "epoch": 0.20485306669918302, "grad_norm": 1.1039750845125622, "learning_rate": 4.681683586523037e-06, "loss": 0.1678, "step": 210 }, { "epoch": 0.20582855749298867, "grad_norm": 1.0549278846097074, "learning_rate": 4.677730666525675e-06, "loss": 0.1661, "step": 211 }, { "epoch": 0.2068040482867943, "grad_norm": 1.0350210106238715, "learning_rate": 4.673755043810324e-06, "loss": 0.1629, "step": 212 }, { "epoch": 0.20777953908059993, "grad_norm": 1.0928587210817202, "learning_rate": 4.669756759822625e-06, "loss": 0.1641, "step": 213 }, { "epoch": 0.20875502987440556, "grad_norm": 1.0139520854082134, "learning_rate": 4.66573585624446e-06, "loss": 0.1689, "step": 214 }, { "epoch": 0.20973052066821118, "grad_norm": 1.1017739109840923, "learning_rate": 4.661692374993519e-06, "loss": 0.1624, "step": 215 }, { "epoch": 0.21070601146201684, "grad_norm": 1.2340168628656984, "learning_rate": 4.657626358222864e-06, "loss": 0.1756, "step": 216 }, { "epoch": 0.21168150225582247, "grad_norm": 0.9230324919898438, "learning_rate": 4.653537848320488e-06, "loss": 0.1602, "step": 217 }, { "epoch": 0.2126569930496281, "grad_norm": 1.0516263439718125, "learning_rate": 4.6494268879088745e-06, "loss": 0.161, "step": 218 }, { "epoch": 0.21363248384343372, "grad_norm": 1.0294336700433153, "learning_rate": 4.6452935198445496e-06, "loss": 0.169, "step": 219 }, { "epoch": 0.21460797463723935, "grad_norm": 0.950242561108817, "learning_rate": 4.64113778721764e-06, "loss": 0.1686, "step": 220 }, { "epoch": 0.215583465431045, "grad_norm": 1.0972582723890745, "learning_rate": 4.636959733351422e-06, "loss": 0.1581, "step": 221 }, { "epoch": 0.21655895622485064, "grad_norm": 0.9521794002072761, "learning_rate": 4.632759401801869e-06, "loss": 0.1592, "step": 222 }, { "epoch": 0.21753444701865626, "grad_norm": 1.1612379311952095, "learning_rate": 4.628536836357196e-06, "loss": 0.166, "step": 223 }, { "epoch": 0.2185099378124619, "grad_norm": 0.9211219216002743, "learning_rate": 4.6242920810374095e-06, "loss": 0.1542, "step": 224 }, { "epoch": 0.21948542860626752, "grad_norm": 0.937788934386043, "learning_rate": 4.62002518009384e-06, "loss": 0.1725, "step": 225 }, { "epoch": 0.22046091940007317, "grad_norm": 0.9773446856643697, "learning_rate": 4.615736178008687e-06, "loss": 0.1566, "step": 226 }, { "epoch": 0.2214364101938788, "grad_norm": 0.9516125327713063, "learning_rate": 4.611425119494552e-06, "loss": 0.158, "step": 227 }, { "epoch": 0.22241190098768443, "grad_norm": 1.0157719146604811, "learning_rate": 4.607092049493973e-06, "loss": 0.1633, "step": 228 }, { "epoch": 0.22338739178149006, "grad_norm": 0.9692342820358121, "learning_rate": 4.6027370131789614e-06, "loss": 0.1622, "step": 229 }, { "epoch": 0.22436288257529569, "grad_norm": 0.9487979994417061, "learning_rate": 4.5983600559505196e-06, "loss": 0.1596, "step": 230 }, { "epoch": 0.22533837336910134, "grad_norm": 1.0470848627284035, "learning_rate": 4.593961223438177e-06, "loss": 0.1707, "step": 231 }, { "epoch": 0.22631386416290697, "grad_norm": 0.9284433450109106, "learning_rate": 4.589540561499512e-06, "loss": 0.149, "step": 232 }, { "epoch": 0.2272893549567126, "grad_norm": 1.0411956514134435, "learning_rate": 4.585098116219674e-06, "loss": 0.1666, "step": 233 }, { "epoch": 0.22826484575051822, "grad_norm": 0.9982980604258467, "learning_rate": 4.580633933910901e-06, "loss": 0.1582, "step": 234 }, { "epoch": 0.22924033654432385, "grad_norm": 1.0353148821593277, "learning_rate": 4.576148061112039e-06, "loss": 0.1798, "step": 235 }, { "epoch": 0.2302158273381295, "grad_norm": 0.8623535833361177, "learning_rate": 4.571640544588056e-06, "loss": 0.1568, "step": 236 }, { "epoch": 0.23119131813193514, "grad_norm": 1.141933497627487, "learning_rate": 4.567111431329555e-06, "loss": 0.1556, "step": 237 }, { "epoch": 0.23216680892574076, "grad_norm": 0.9999053398774652, "learning_rate": 4.562560768552283e-06, "loss": 0.1603, "step": 238 }, { "epoch": 0.2331422997195464, "grad_norm": 0.9284632452687369, "learning_rate": 4.55798860369664e-06, "loss": 0.1458, "step": 239 }, { "epoch": 0.23411779051335202, "grad_norm": 0.9872863479156412, "learning_rate": 4.553394984427184e-06, "loss": 0.1621, "step": 240 }, { "epoch": 0.23509328130715768, "grad_norm": 0.9100468729196777, "learning_rate": 4.548779958632134e-06, "loss": 0.1568, "step": 241 }, { "epoch": 0.2360687721009633, "grad_norm": 0.8444568550070792, "learning_rate": 4.54414357442287e-06, "loss": 0.1408, "step": 242 }, { "epoch": 0.23704426289476893, "grad_norm": 1.0223092529639413, "learning_rate": 4.539485880133433e-06, "loss": 0.1609, "step": 243 }, { "epoch": 0.23801975368857456, "grad_norm": 0.9414329883963978, "learning_rate": 4.534806924320021e-06, "loss": 0.1528, "step": 244 }, { "epoch": 0.2389952444823802, "grad_norm": 1.0175078568206852, "learning_rate": 4.5301067557604815e-06, "loss": 0.1607, "step": 245 }, { "epoch": 0.23997073527618584, "grad_norm": 0.9503464111258508, "learning_rate": 4.525385423453803e-06, "loss": 0.1621, "step": 246 }, { "epoch": 0.24094622606999147, "grad_norm": 1.0317509367040238, "learning_rate": 4.520642976619607e-06, "loss": 0.1562, "step": 247 }, { "epoch": 0.2419217168637971, "grad_norm": 0.9406630214637611, "learning_rate": 4.515879464697629e-06, "loss": 0.1524, "step": 248 }, { "epoch": 0.24289720765760273, "grad_norm": 0.9116204738820602, "learning_rate": 4.51109493734721e-06, "loss": 0.1589, "step": 249 }, { "epoch": 0.24387269845140835, "grad_norm": 0.9686707124955454, "learning_rate": 4.506289444446775e-06, "loss": 0.1661, "step": 250 }, { "epoch": 0.244848189245214, "grad_norm": 0.8855697170450265, "learning_rate": 4.5014630360933136e-06, "loss": 0.1525, "step": 251 }, { "epoch": 0.24582368003901964, "grad_norm": 0.8408850368006234, "learning_rate": 4.496615762601857e-06, "loss": 0.1553, "step": 252 }, { "epoch": 0.24679917083282527, "grad_norm": 0.9791331885043342, "learning_rate": 4.491747674504956e-06, "loss": 0.1613, "step": 253 }, { "epoch": 0.2477746616266309, "grad_norm": 0.8536640710320617, "learning_rate": 4.48685882255215e-06, "loss": 0.1561, "step": 254 }, { "epoch": 0.24875015242043652, "grad_norm": 0.9108673457445514, "learning_rate": 4.481949257709442e-06, "loss": 0.1581, "step": 255 }, { "epoch": 0.24972564321424218, "grad_norm": 0.9929310498573025, "learning_rate": 4.477019031158767e-06, "loss": 0.1635, "step": 256 }, { "epoch": 0.2507011340080478, "grad_norm": 0.9319735021715781, "learning_rate": 4.472068194297453e-06, "loss": 0.1535, "step": 257 }, { "epoch": 0.25167662480185343, "grad_norm": 0.8234410520072005, "learning_rate": 4.467096798737694e-06, "loss": 0.1419, "step": 258 }, { "epoch": 0.2526521155956591, "grad_norm": 0.8293157655059741, "learning_rate": 4.462104896306004e-06, "loss": 0.1509, "step": 259 }, { "epoch": 0.2536276063894647, "grad_norm": 0.981516914054671, "learning_rate": 4.457092539042682e-06, "loss": 0.1503, "step": 260 }, { "epoch": 0.25460309718327034, "grad_norm": 0.9280314009979089, "learning_rate": 4.452059779201267e-06, "loss": 0.1524, "step": 261 }, { "epoch": 0.25557858797707594, "grad_norm": 0.9107843402176423, "learning_rate": 4.44700666924799e-06, "loss": 0.1599, "step": 262 }, { "epoch": 0.2565540787708816, "grad_norm": 0.9504200684566895, "learning_rate": 4.441933261861239e-06, "loss": 0.1603, "step": 263 }, { "epoch": 0.25752956956468726, "grad_norm": 0.8932484399394331, "learning_rate": 4.436839609930992e-06, "loss": 0.166, "step": 264 }, { "epoch": 0.25850506035849286, "grad_norm": 0.9657424201285992, "learning_rate": 4.431725766558284e-06, "loss": 0.1592, "step": 265 }, { "epoch": 0.2594805511522985, "grad_norm": 0.9280800047846618, "learning_rate": 4.426591785054637e-06, "loss": 0.1692, "step": 266 }, { "epoch": 0.2604560419461041, "grad_norm": 0.909753774216958, "learning_rate": 4.421437718941517e-06, "loss": 0.1546, "step": 267 }, { "epoch": 0.26143153273990977, "grad_norm": 0.9249375767292966, "learning_rate": 4.416263621949769e-06, "loss": 0.158, "step": 268 }, { "epoch": 0.2624070235337154, "grad_norm": 0.9638546484727502, "learning_rate": 4.41106954801906e-06, "loss": 0.1549, "step": 269 }, { "epoch": 0.263382514327521, "grad_norm": 0.9676726584506395, "learning_rate": 4.4058555512973135e-06, "loss": 0.1526, "step": 270 }, { "epoch": 0.2643580051213267, "grad_norm": 0.9643850490217997, "learning_rate": 4.4006216861401475e-06, "loss": 0.1507, "step": 271 }, { "epoch": 0.2653334959151323, "grad_norm": 0.9455841602811957, "learning_rate": 4.395368007110307e-06, "loss": 0.1609, "step": 272 }, { "epoch": 0.26630898670893793, "grad_norm": 1.0032153539367255, "learning_rate": 4.390094568977099e-06, "loss": 0.1517, "step": 273 }, { "epoch": 0.2672844775027436, "grad_norm": 0.9286676590827331, "learning_rate": 4.384801426715814e-06, "loss": 0.1509, "step": 274 }, { "epoch": 0.2682599682965492, "grad_norm": 0.8635512942768435, "learning_rate": 4.379488635507157e-06, "loss": 0.1479, "step": 275 }, { "epoch": 0.26923545909035485, "grad_norm": 0.9730864421049288, "learning_rate": 4.3741562507366754e-06, "loss": 0.1532, "step": 276 }, { "epoch": 0.27021094988416045, "grad_norm": 0.8738885156811014, "learning_rate": 4.368804327994174e-06, "loss": 0.1579, "step": 277 }, { "epoch": 0.2711864406779661, "grad_norm": 0.9907365851233481, "learning_rate": 4.363432923073144e-06, "loss": 0.1543, "step": 278 }, { "epoch": 0.27216193147177176, "grad_norm": 0.9242295443836225, "learning_rate": 4.3580420919701745e-06, "loss": 0.1584, "step": 279 }, { "epoch": 0.27313742226557736, "grad_norm": 0.9088209928537309, "learning_rate": 4.352631890884373e-06, "loss": 0.1547, "step": 280 }, { "epoch": 0.274112913059383, "grad_norm": 0.9101013673566347, "learning_rate": 4.347202376216775e-06, "loss": 0.1648, "step": 281 }, { "epoch": 0.2750884038531886, "grad_norm": 0.9053322800776012, "learning_rate": 4.341753604569764e-06, "loss": 0.1488, "step": 282 }, { "epoch": 0.27606389464699427, "grad_norm": 0.9622509070577334, "learning_rate": 4.336285632746472e-06, "loss": 0.155, "step": 283 }, { "epoch": 0.2770393854407999, "grad_norm": 0.9097255445557655, "learning_rate": 4.330798517750194e-06, "loss": 0.138, "step": 284 }, { "epoch": 0.2780148762346055, "grad_norm": 0.915630210467682, "learning_rate": 4.3252923167837905e-06, "loss": 0.1533, "step": 285 }, { "epoch": 0.2789903670284112, "grad_norm": 0.9410519479526743, "learning_rate": 4.319767087249094e-06, "loss": 0.1588, "step": 286 }, { "epoch": 0.2799658578222168, "grad_norm": 0.9225807794360538, "learning_rate": 4.314222886746304e-06, "loss": 0.1546, "step": 287 }, { "epoch": 0.28094134861602243, "grad_norm": 0.8732635665891079, "learning_rate": 4.308659773073398e-06, "loss": 0.1677, "step": 288 }, { "epoch": 0.2819168394098281, "grad_norm": 0.979628005682833, "learning_rate": 4.303077804225517e-06, "loss": 0.1626, "step": 289 }, { "epoch": 0.2828923302036337, "grad_norm": 0.9055375259430347, "learning_rate": 4.297477038394368e-06, "loss": 0.1675, "step": 290 }, { "epoch": 0.28386782099743935, "grad_norm": 0.9480989580020738, "learning_rate": 4.291857533967616e-06, "loss": 0.1574, "step": 291 }, { "epoch": 0.28484331179124495, "grad_norm": 0.8985804614382488, "learning_rate": 4.286219349528274e-06, "loss": 0.1557, "step": 292 }, { "epoch": 0.2858188025850506, "grad_norm": 1.0146034373434893, "learning_rate": 4.280562543854091e-06, "loss": 0.1645, "step": 293 }, { "epoch": 0.28679429337885626, "grad_norm": 0.890713161023523, "learning_rate": 4.274887175916946e-06, "loss": 0.1567, "step": 294 }, { "epoch": 0.28776978417266186, "grad_norm": 0.9267996009148974, "learning_rate": 4.269193304882226e-06, "loss": 0.16, "step": 295 }, { "epoch": 0.2887452749664675, "grad_norm": 0.8938807739528849, "learning_rate": 4.263480990108212e-06, "loss": 0.1578, "step": 296 }, { "epoch": 0.2897207657602731, "grad_norm": 0.9382622618633856, "learning_rate": 4.257750291145457e-06, "loss": 0.1594, "step": 297 }, { "epoch": 0.29069625655407877, "grad_norm": 0.9825847935980099, "learning_rate": 4.252001267736174e-06, "loss": 0.1555, "step": 298 }, { "epoch": 0.2916717473478844, "grad_norm": 0.8990364135002812, "learning_rate": 4.246233979813602e-06, "loss": 0.1603, "step": 299 }, { "epoch": 0.29264723814169, "grad_norm": 0.8656451974006956, "learning_rate": 4.24044848750139e-06, "loss": 0.1491, "step": 300 }, { "epoch": 0.2936227289354957, "grad_norm": 0.9253142530395264, "learning_rate": 4.234644851112965e-06, "loss": 0.1607, "step": 301 }, { "epoch": 0.2945982197293013, "grad_norm": 0.8969951216981602, "learning_rate": 4.228823131150904e-06, "loss": 0.1521, "step": 302 }, { "epoch": 0.29557371052310694, "grad_norm": 0.9403876224279274, "learning_rate": 4.222983388306308e-06, "loss": 0.1616, "step": 303 }, { "epoch": 0.2965492013169126, "grad_norm": 0.8925996666326684, "learning_rate": 4.217125683458162e-06, "loss": 0.1581, "step": 304 }, { "epoch": 0.2975246921107182, "grad_norm": 0.8585766705268447, "learning_rate": 4.211250077672704e-06, "loss": 0.1492, "step": 305 }, { "epoch": 0.29850018290452385, "grad_norm": 0.9350703216606489, "learning_rate": 4.20535663220279e-06, "loss": 0.1587, "step": 306 }, { "epoch": 0.29947567369832945, "grad_norm": 0.9188676922420881, "learning_rate": 4.199445408487253e-06, "loss": 0.1593, "step": 307 }, { "epoch": 0.3004511644921351, "grad_norm": 0.9547113904108987, "learning_rate": 4.1935164681502626e-06, "loss": 0.151, "step": 308 }, { "epoch": 0.30142665528594076, "grad_norm": 1.0033584349157778, "learning_rate": 4.187569873000684e-06, "loss": 0.1629, "step": 309 }, { "epoch": 0.30240214607974636, "grad_norm": 0.9859571353144448, "learning_rate": 4.181605685031433e-06, "loss": 0.1734, "step": 310 }, { "epoch": 0.303377636873552, "grad_norm": 0.8228370416849764, "learning_rate": 4.1756239664188275e-06, "loss": 0.1509, "step": 311 }, { "epoch": 0.3043531276673576, "grad_norm": 1.0367963117038492, "learning_rate": 4.169624779521944e-06, "loss": 0.1571, "step": 312 }, { "epoch": 0.30532861846116327, "grad_norm": 0.9669617210813681, "learning_rate": 4.163608186881964e-06, "loss": 0.1628, "step": 313 }, { "epoch": 0.3063041092549689, "grad_norm": 1.0103654228797525, "learning_rate": 4.157574251221522e-06, "loss": 0.1595, "step": 314 }, { "epoch": 0.3072796000487745, "grad_norm": 0.9081265589361568, "learning_rate": 4.1515230354440526e-06, "loss": 0.1524, "step": 315 }, { "epoch": 0.3082550908425802, "grad_norm": 0.924809859907981, "learning_rate": 4.145454602633137e-06, "loss": 0.1619, "step": 316 }, { "epoch": 0.3092305816363858, "grad_norm": 0.8914505960487598, "learning_rate": 4.139369016051838e-06, "loss": 0.1603, "step": 317 }, { "epoch": 0.31020607243019144, "grad_norm": 0.8961623985939995, "learning_rate": 4.1332663391420515e-06, "loss": 0.1604, "step": 318 }, { "epoch": 0.3111815632239971, "grad_norm": 0.8168107498311156, "learning_rate": 4.127146635523835e-06, "loss": 0.1437, "step": 319 }, { "epoch": 0.3121570540178027, "grad_norm": 0.8664110012815645, "learning_rate": 4.121009968994751e-06, "loss": 0.1506, "step": 320 }, { "epoch": 0.31313254481160835, "grad_norm": 0.9602948973971756, "learning_rate": 4.114856403529196e-06, "loss": 0.1463, "step": 321 }, { "epoch": 0.31410803560541395, "grad_norm": 0.9778713250873332, "learning_rate": 4.10868600327774e-06, "loss": 0.1615, "step": 322 }, { "epoch": 0.3150835263992196, "grad_norm": 1.0278280133622393, "learning_rate": 4.102498832566454e-06, "loss": 0.1537, "step": 323 }, { "epoch": 0.31605901719302526, "grad_norm": 0.9256019026472626, "learning_rate": 4.096294955896239e-06, "loss": 0.1611, "step": 324 }, { "epoch": 0.31703450798683086, "grad_norm": 0.8287537359495541, "learning_rate": 4.090074437942155e-06, "loss": 0.1533, "step": 325 }, { "epoch": 0.3180099987806365, "grad_norm": 0.8835278034235962, "learning_rate": 4.083837343552749e-06, "loss": 0.1493, "step": 326 }, { "epoch": 0.3189854895744421, "grad_norm": 0.8259069804781106, "learning_rate": 4.077583737749373e-06, "loss": 0.1424, "step": 327 }, { "epoch": 0.31996098036824777, "grad_norm": 0.941189764723523, "learning_rate": 4.07131368572551e-06, "loss": 0.1523, "step": 328 }, { "epoch": 0.3209364711620534, "grad_norm": 0.9164961397020092, "learning_rate": 4.0650272528460955e-06, "loss": 0.1539, "step": 329 }, { "epoch": 0.321911961955859, "grad_norm": 0.8541305146982892, "learning_rate": 4.058724504646834e-06, "loss": 0.1377, "step": 330 }, { "epoch": 0.3228874527496647, "grad_norm": 0.9562169457452313, "learning_rate": 4.052405506833516e-06, "loss": 0.1534, "step": 331 }, { "epoch": 0.3238629435434703, "grad_norm": 0.9199148267026572, "learning_rate": 4.046070325281333e-06, "loss": 0.1471, "step": 332 }, { "epoch": 0.32483843433727594, "grad_norm": 0.9205784248133079, "learning_rate": 4.039719026034191e-06, "loss": 0.1403, "step": 333 }, { "epoch": 0.3258139251310816, "grad_norm": 0.8621247018342253, "learning_rate": 4.0333516753040225e-06, "loss": 0.1365, "step": 334 }, { "epoch": 0.3267894159248872, "grad_norm": 0.9318993173542444, "learning_rate": 4.026968339470097e-06, "loss": 0.1446, "step": 335 }, { "epoch": 0.32776490671869285, "grad_norm": 0.9970504184721904, "learning_rate": 4.020569085078324e-06, "loss": 0.1553, "step": 336 }, { "epoch": 0.32874039751249845, "grad_norm": 0.9288139122169886, "learning_rate": 4.014153978840568e-06, "loss": 0.1457, "step": 337 }, { "epoch": 0.3297158883063041, "grad_norm": 0.947326368514792, "learning_rate": 4.007723087633943e-06, "loss": 0.1567, "step": 338 }, { "epoch": 0.33069137910010976, "grad_norm": 0.933169285653779, "learning_rate": 4.001276478500127e-06, "loss": 0.1564, "step": 339 }, { "epoch": 0.33166686989391536, "grad_norm": 0.8619195635693533, "learning_rate": 3.994814218644649e-06, "loss": 0.1409, "step": 340 }, { "epoch": 0.332642360687721, "grad_norm": 0.9533171506313436, "learning_rate": 3.988336375436201e-06, "loss": 0.1472, "step": 341 }, { "epoch": 0.3336178514815266, "grad_norm": 0.9085852674744761, "learning_rate": 3.981843016405928e-06, "loss": 0.1518, "step": 342 }, { "epoch": 0.3345933422753323, "grad_norm": 0.8015484531793845, "learning_rate": 3.975334209246727e-06, "loss": 0.1448, "step": 343 }, { "epoch": 0.33556883306913793, "grad_norm": 0.9409823279258879, "learning_rate": 3.968810021812544e-06, "loss": 0.1555, "step": 344 }, { "epoch": 0.33654432386294353, "grad_norm": 0.8142592274829353, "learning_rate": 3.962270522117659e-06, "loss": 0.1457, "step": 345 }, { "epoch": 0.3375198146567492, "grad_norm": 0.9076632142318111, "learning_rate": 3.955715778335984e-06, "loss": 0.1538, "step": 346 }, { "epoch": 0.3384953054505548, "grad_norm": 0.8974350187857667, "learning_rate": 3.949145858800348e-06, "loss": 0.1481, "step": 347 }, { "epoch": 0.33947079624436044, "grad_norm": 0.8901297893092709, "learning_rate": 3.942560832001789e-06, "loss": 0.1503, "step": 348 }, { "epoch": 0.3404462870381661, "grad_norm": 0.8560837002365613, "learning_rate": 3.935960766588835e-06, "loss": 0.1528, "step": 349 }, { "epoch": 0.3414217778319717, "grad_norm": 0.8443274184309183, "learning_rate": 3.9293457313667905e-06, "loss": 0.1436, "step": 350 }, { "epoch": 0.34239726862577735, "grad_norm": 1.0282756436782237, "learning_rate": 3.922715795297022e-06, "loss": 0.1482, "step": 351 }, { "epoch": 0.34337275941958295, "grad_norm": 0.9065229091333961, "learning_rate": 3.916071027496234e-06, "loss": 0.1445, "step": 352 }, { "epoch": 0.3443482502133886, "grad_norm": 0.9714447847850254, "learning_rate": 3.909411497235752e-06, "loss": 0.1473, "step": 353 }, { "epoch": 0.34532374100719426, "grad_norm": 0.9561585485417536, "learning_rate": 3.902737273940799e-06, "loss": 0.1471, "step": 354 }, { "epoch": 0.34629923180099986, "grad_norm": 0.8900543390831095, "learning_rate": 3.8960484271897736e-06, "loss": 0.1527, "step": 355 }, { "epoch": 0.3472747225948055, "grad_norm": 0.9827409295726128, "learning_rate": 3.889345026713519e-06, "loss": 0.1534, "step": 356 }, { "epoch": 0.3482502133886111, "grad_norm": 0.8940002734667856, "learning_rate": 3.882627142394605e-06, "loss": 0.1486, "step": 357 }, { "epoch": 0.3492257041824168, "grad_norm": 0.9944296349312072, "learning_rate": 3.875894844266592e-06, "loss": 0.1531, "step": 358 }, { "epoch": 0.35020119497622243, "grad_norm": 0.9120571509586781, "learning_rate": 3.869148202513303e-06, "loss": 0.1606, "step": 359 }, { "epoch": 0.35117668577002803, "grad_norm": 0.9095948987953094, "learning_rate": 3.862387287468095e-06, "loss": 0.137, "step": 360 }, { "epoch": 0.3521521765638337, "grad_norm": 0.877930473526069, "learning_rate": 3.85561216961312e-06, "loss": 0.1408, "step": 361 }, { "epoch": 0.3531276673576393, "grad_norm": 0.9699360794407749, "learning_rate": 3.8488229195785965e-06, "loss": 0.147, "step": 362 }, { "epoch": 0.35410315815144494, "grad_norm": 0.8874366855608934, "learning_rate": 3.842019608142068e-06, "loss": 0.1446, "step": 363 }, { "epoch": 0.3550786489452506, "grad_norm": 0.8644114874928793, "learning_rate": 3.83520230622767e-06, "loss": 0.1484, "step": 364 }, { "epoch": 0.3560541397390562, "grad_norm": 0.9393508952046424, "learning_rate": 3.8283710849053835e-06, "loss": 0.1415, "step": 365 }, { "epoch": 0.35702963053286185, "grad_norm": 0.915863410960202, "learning_rate": 3.821526015390302e-06, "loss": 0.1386, "step": 366 }, { "epoch": 0.35800512132666745, "grad_norm": 0.9939022499746282, "learning_rate": 3.814667169041887e-06, "loss": 0.156, "step": 367 }, { "epoch": 0.3589806121204731, "grad_norm": 0.9698635070361811, "learning_rate": 3.8077946173632175e-06, "loss": 0.1351, "step": 368 }, { "epoch": 0.35995610291427876, "grad_norm": 0.9476385232414736, "learning_rate": 3.800908432000254e-06, "loss": 0.1427, "step": 369 }, { "epoch": 0.36093159370808436, "grad_norm": 0.856130584200129, "learning_rate": 3.7940086847410875e-06, "loss": 0.1516, "step": 370 }, { "epoch": 0.36190708450189, "grad_norm": 0.8579511362373262, "learning_rate": 3.7870954475151873e-06, "loss": 0.1453, "step": 371 }, { "epoch": 0.3628825752956956, "grad_norm": 0.9244210071492551, "learning_rate": 3.780168792392658e-06, "loss": 0.1403, "step": 372 }, { "epoch": 0.3638580660895013, "grad_norm": 0.8916214840343722, "learning_rate": 3.7732287915834842e-06, "loss": 0.1344, "step": 373 }, { "epoch": 0.36483355688330693, "grad_norm": 0.9285987965737049, "learning_rate": 3.766275517436779e-06, "loss": 0.1473, "step": 374 }, { "epoch": 0.36580904767711253, "grad_norm": 1.0202253515003419, "learning_rate": 3.759309042440028e-06, "loss": 0.1626, "step": 375 }, { "epoch": 0.3667845384709182, "grad_norm": 0.9202482904809027, "learning_rate": 3.752329439218337e-06, "loss": 0.1437, "step": 376 }, { "epoch": 0.3677600292647238, "grad_norm": 0.868921559794134, "learning_rate": 3.7453367805336697e-06, "loss": 0.1411, "step": 377 }, { "epoch": 0.36873552005852944, "grad_norm": 0.9278377375229888, "learning_rate": 3.738331139284096e-06, "loss": 0.1462, "step": 378 }, { "epoch": 0.3697110108523351, "grad_norm": 0.9555741240850121, "learning_rate": 3.731312588503024e-06, "loss": 0.1525, "step": 379 }, { "epoch": 0.3706865016461407, "grad_norm": 0.9746070816089245, "learning_rate": 3.7242812013584474e-06, "loss": 0.1509, "step": 380 }, { "epoch": 0.37166199243994635, "grad_norm": 0.8847369605779092, "learning_rate": 3.717237051152175e-06, "loss": 0.1433, "step": 381 }, { "epoch": 0.37263748323375195, "grad_norm": 1.0235063474249009, "learning_rate": 3.710180211319071e-06, "loss": 0.1486, "step": 382 }, { "epoch": 0.3736129740275576, "grad_norm": 0.8950016936249444, "learning_rate": 3.703110755426289e-06, "loss": 0.139, "step": 383 }, { "epoch": 0.37458846482136327, "grad_norm": 0.8847042144920154, "learning_rate": 3.696028757172503e-06, "loss": 0.1521, "step": 384 }, { "epoch": 0.37556395561516887, "grad_norm": 0.9106515631528226, "learning_rate": 3.68893429038714e-06, "loss": 0.146, "step": 385 }, { "epoch": 0.3765394464089745, "grad_norm": 0.9546055014669876, "learning_rate": 3.681827429029613e-06, "loss": 0.1431, "step": 386 }, { "epoch": 0.3775149372027802, "grad_norm": 0.8713617926689637, "learning_rate": 3.6747082471885454e-06, "loss": 0.1484, "step": 387 }, { "epoch": 0.3784904279965858, "grad_norm": 0.8981560581623195, "learning_rate": 3.6675768190810023e-06, "loss": 0.1405, "step": 388 }, { "epoch": 0.37946591879039143, "grad_norm": 0.9576903681690667, "learning_rate": 3.6604332190517144e-06, "loss": 0.1461, "step": 389 }, { "epoch": 0.38044140958419703, "grad_norm": 0.9982163299550169, "learning_rate": 3.653277521572305e-06, "loss": 0.15, "step": 390 }, { "epoch": 0.3814169003780027, "grad_norm": 0.8524151991534643, "learning_rate": 3.6461098012405116e-06, "loss": 0.1446, "step": 391 }, { "epoch": 0.38239239117180834, "grad_norm": 0.9831471719181231, "learning_rate": 3.6389301327794114e-06, "loss": 0.1671, "step": 392 }, { "epoch": 0.38336788196561394, "grad_norm": 0.9182177457148154, "learning_rate": 3.6317385910366365e-06, "loss": 0.1476, "step": 393 }, { "epoch": 0.3843433727594196, "grad_norm": 0.9198423171501263, "learning_rate": 3.624535250983601e-06, "loss": 0.153, "step": 394 }, { "epoch": 0.3853188635532252, "grad_norm": 0.8808600354881101, "learning_rate": 3.6173201877147134e-06, "loss": 0.1456, "step": 395 }, { "epoch": 0.38629435434703085, "grad_norm": 0.8829324038702485, "learning_rate": 3.6100934764465973e-06, "loss": 0.1413, "step": 396 }, { "epoch": 0.3872698451408365, "grad_norm": 0.8876955529556401, "learning_rate": 3.6028551925173046e-06, "loss": 0.1512, "step": 397 }, { "epoch": 0.3882453359346421, "grad_norm": 0.8555692280523547, "learning_rate": 3.595605411385533e-06, "loss": 0.1428, "step": 398 }, { "epoch": 0.38922082672844777, "grad_norm": 0.8764418318857389, "learning_rate": 3.5883442086298375e-06, "loss": 0.1481, "step": 399 }, { "epoch": 0.39019631752225337, "grad_norm": 0.926181501103571, "learning_rate": 3.581071659947842e-06, "loss": 0.1394, "step": 400 }, { "epoch": 0.391171808316059, "grad_norm": 0.9087710728115378, "learning_rate": 3.5737878411554515e-06, "loss": 0.1396, "step": 401 }, { "epoch": 0.3921472991098647, "grad_norm": 0.9084832908918349, "learning_rate": 3.566492828186063e-06, "loss": 0.1457, "step": 402 }, { "epoch": 0.3931227899036703, "grad_norm": 0.9198265733188117, "learning_rate": 3.5591866970897695e-06, "loss": 0.1428, "step": 403 }, { "epoch": 0.39409828069747593, "grad_norm": 0.9028014238279914, "learning_rate": 3.55186952403257e-06, "loss": 0.1459, "step": 404 }, { "epoch": 0.39507377149128153, "grad_norm": 0.7803451616200923, "learning_rate": 3.5445413852955783e-06, "loss": 0.1405, "step": 405 }, { "epoch": 0.3960492622850872, "grad_norm": 0.8807693627140049, "learning_rate": 3.5372023572742224e-06, "loss": 0.144, "step": 406 }, { "epoch": 0.39702475307889284, "grad_norm": 0.9065679616664823, "learning_rate": 3.52985251647745e-06, "loss": 0.152, "step": 407 }, { "epoch": 0.39800024387269844, "grad_norm": 0.9126785313651608, "learning_rate": 3.522491939526932e-06, "loss": 0.1527, "step": 408 }, { "epoch": 0.3989757346665041, "grad_norm": 0.9401530910172149, "learning_rate": 3.515120703156264e-06, "loss": 0.1546, "step": 409 }, { "epoch": 0.3999512254603097, "grad_norm": 0.919541385453652, "learning_rate": 3.507738884210164e-06, "loss": 0.1397, "step": 410 }, { "epoch": 0.40092671625411536, "grad_norm": 0.9257840949340401, "learning_rate": 3.500346559643675e-06, "loss": 0.1435, "step": 411 }, { "epoch": 0.401902207047921, "grad_norm": 0.9038785967347172, "learning_rate": 3.4929438065213568e-06, "loss": 0.147, "step": 412 }, { "epoch": 0.4028776978417266, "grad_norm": 1.0676896384715988, "learning_rate": 3.48553070201649e-06, "loss": 0.1558, "step": 413 }, { "epoch": 0.40385318863553227, "grad_norm": 0.9754468401487543, "learning_rate": 3.4781073234102665e-06, "loss": 0.1449, "step": 414 }, { "epoch": 0.40482867942933787, "grad_norm": 0.8939601187260615, "learning_rate": 3.470673748090984e-06, "loss": 0.1396, "step": 415 }, { "epoch": 0.4058041702231435, "grad_norm": 0.9427428788545429, "learning_rate": 3.4632300535532415e-06, "loss": 0.1427, "step": 416 }, { "epoch": 0.4067796610169492, "grad_norm": 0.887782687868998, "learning_rate": 3.4557763173971293e-06, "loss": 0.1449, "step": 417 }, { "epoch": 0.4077551518107548, "grad_norm": 1.0018593992113438, "learning_rate": 3.4483126173274227e-06, "loss": 0.1473, "step": 418 }, { "epoch": 0.40873064260456043, "grad_norm": 0.9539144339332059, "learning_rate": 3.440839031152769e-06, "loss": 0.1432, "step": 419 }, { "epoch": 0.40970613339836603, "grad_norm": 0.8796520440270823, "learning_rate": 3.4333556367848773e-06, "loss": 0.1356, "step": 420 }, { "epoch": 0.4106816241921717, "grad_norm": 0.8930556857980667, "learning_rate": 3.425862512237708e-06, "loss": 0.1414, "step": 421 }, { "epoch": 0.41165711498597735, "grad_norm": 0.9130386801180627, "learning_rate": 3.418359735626656e-06, "loss": 0.144, "step": 422 }, { "epoch": 0.41263260577978295, "grad_norm": 0.8591156628779346, "learning_rate": 3.4108473851677408e-06, "loss": 0.1413, "step": 423 }, { "epoch": 0.4136080965735886, "grad_norm": 0.9654725301304397, "learning_rate": 3.4033255391767865e-06, "loss": 0.1492, "step": 424 }, { "epoch": 0.4145835873673942, "grad_norm": 0.9083019018961785, "learning_rate": 3.395794276068609e-06, "loss": 0.1452, "step": 425 }, { "epoch": 0.41555907816119986, "grad_norm": 0.8891185711032482, "learning_rate": 3.388253674356196e-06, "loss": 0.1415, "step": 426 }, { "epoch": 0.4165345689550055, "grad_norm": 0.9375975455278247, "learning_rate": 3.380703812649891e-06, "loss": 0.1403, "step": 427 }, { "epoch": 0.4175100597488111, "grad_norm": 0.9020491738743014, "learning_rate": 3.3731447696565713e-06, "loss": 0.1455, "step": 428 }, { "epoch": 0.41848555054261677, "grad_norm": 0.8574973373936586, "learning_rate": 3.3655766241788285e-06, "loss": 0.1368, "step": 429 }, { "epoch": 0.41946104133642237, "grad_norm": 0.8757023776531063, "learning_rate": 3.357999455114148e-06, "loss": 0.1446, "step": 430 }, { "epoch": 0.420436532130228, "grad_norm": 0.9837347529941712, "learning_rate": 3.350413341454086e-06, "loss": 0.1506, "step": 431 }, { "epoch": 0.4214120229240337, "grad_norm": 0.8472380588550584, "learning_rate": 3.3428183622834442e-06, "loss": 0.1513, "step": 432 }, { "epoch": 0.4223875137178393, "grad_norm": 0.8387818280951195, "learning_rate": 3.3352145967794463e-06, "loss": 0.1403, "step": 433 }, { "epoch": 0.42336300451164494, "grad_norm": 0.8972955470900738, "learning_rate": 3.3276021242109152e-06, "loss": 0.1411, "step": 434 }, { "epoch": 0.42433849530545054, "grad_norm": 0.9676494460221445, "learning_rate": 3.319981023937442e-06, "loss": 0.1508, "step": 435 }, { "epoch": 0.4253139860992562, "grad_norm": 0.8603463763707745, "learning_rate": 3.3123513754085635e-06, "loss": 0.1412, "step": 436 }, { "epoch": 0.42628947689306185, "grad_norm": 0.8735162753932564, "learning_rate": 3.3047132581629297e-06, "loss": 0.1356, "step": 437 }, { "epoch": 0.42726496768686745, "grad_norm": 0.946772891936297, "learning_rate": 3.297066751827478e-06, "loss": 0.1364, "step": 438 }, { "epoch": 0.4282404584806731, "grad_norm": 0.9312293341123083, "learning_rate": 3.2894119361166e-06, "loss": 0.1384, "step": 439 }, { "epoch": 0.4292159492744787, "grad_norm": 0.8668697057456637, "learning_rate": 3.2817488908313132e-06, "loss": 0.1407, "step": 440 }, { "epoch": 0.43019144006828436, "grad_norm": 0.9062795443546382, "learning_rate": 3.274077695858428e-06, "loss": 0.1511, "step": 441 }, { "epoch": 0.43116693086209, "grad_norm": 0.9197029171414413, "learning_rate": 3.2663984311697137e-06, "loss": 0.1443, "step": 442 }, { "epoch": 0.4321424216558956, "grad_norm": 0.9874318993389217, "learning_rate": 3.2587111768210677e-06, "loss": 0.1522, "step": 443 }, { "epoch": 0.43311791244970127, "grad_norm": 0.8997910308707128, "learning_rate": 3.251016012951678e-06, "loss": 0.1376, "step": 444 }, { "epoch": 0.43409340324350687, "grad_norm": 0.8780597656452647, "learning_rate": 3.2433130197831877e-06, "loss": 0.141, "step": 445 }, { "epoch": 0.4350688940373125, "grad_norm": 0.8943843664972011, "learning_rate": 3.2356022776188623e-06, "loss": 0.1394, "step": 446 }, { "epoch": 0.4360443848311182, "grad_norm": 0.9140727392888497, "learning_rate": 3.227883866842749e-06, "loss": 0.1384, "step": 447 }, { "epoch": 0.4370198756249238, "grad_norm": 0.8988648983999619, "learning_rate": 3.2201578679188396e-06, "loss": 0.1383, "step": 448 }, { "epoch": 0.43799536641872944, "grad_norm": 0.809856506892555, "learning_rate": 3.2124243613902316e-06, "loss": 0.1404, "step": 449 }, { "epoch": 0.43897085721253504, "grad_norm": 0.9050899458550025, "learning_rate": 3.204683427878291e-06, "loss": 0.1467, "step": 450 }, { "epoch": 0.4399463480063407, "grad_norm": 0.8840121082189991, "learning_rate": 3.196935148081808e-06, "loss": 0.1446, "step": 451 }, { "epoch": 0.44092183880014635, "grad_norm": 0.855404279656821, "learning_rate": 3.189179602776157e-06, "loss": 0.1362, "step": 452 }, { "epoch": 0.44189732959395195, "grad_norm": 0.8416485853165431, "learning_rate": 3.181416872812455e-06, "loss": 0.1381, "step": 453 }, { "epoch": 0.4428728203877576, "grad_norm": 0.826136304893414, "learning_rate": 3.1736470391167195e-06, "loss": 0.1439, "step": 454 }, { "epoch": 0.4438483111815632, "grad_norm": 0.888073097598988, "learning_rate": 3.1658701826890237e-06, "loss": 0.1472, "step": 455 }, { "epoch": 0.44482380197536886, "grad_norm": 0.8935863712022185, "learning_rate": 3.158086384602652e-06, "loss": 0.1401, "step": 456 }, { "epoch": 0.4457992927691745, "grad_norm": 0.8809290645701081, "learning_rate": 3.150295726003256e-06, "loss": 0.1527, "step": 457 }, { "epoch": 0.4467747835629801, "grad_norm": 0.8388623216005323, "learning_rate": 3.142498288108007e-06, "loss": 0.1365, "step": 458 }, { "epoch": 0.44775027435678577, "grad_norm": 0.9162474538739859, "learning_rate": 3.13469415220475e-06, "loss": 0.1405, "step": 459 }, { "epoch": 0.44872576515059137, "grad_norm": 0.9225875521669346, "learning_rate": 3.1268833996511584e-06, "loss": 0.1479, "step": 460 }, { "epoch": 0.449701255944397, "grad_norm": 0.8772214033934059, "learning_rate": 3.119066111873879e-06, "loss": 0.1312, "step": 461 }, { "epoch": 0.4506767467382027, "grad_norm": 0.8397821561643128, "learning_rate": 3.1112423703676937e-06, "loss": 0.1344, "step": 462 }, { "epoch": 0.4516522375320083, "grad_norm": 0.9056702372440215, "learning_rate": 3.1034122566946596e-06, "loss": 0.1329, "step": 463 }, { "epoch": 0.45262772832581394, "grad_norm": 0.9828835323972128, "learning_rate": 3.0955758524832663e-06, "loss": 0.1454, "step": 464 }, { "epoch": 0.45360321911961954, "grad_norm": 0.949595108992025, "learning_rate": 3.0877332394275806e-06, "loss": 0.1393, "step": 465 }, { "epoch": 0.4545787099134252, "grad_norm": 0.8820964745475348, "learning_rate": 3.079884499286396e-06, "loss": 0.1303, "step": 466 }, { "epoch": 0.45555420070723085, "grad_norm": 0.8539218038263852, "learning_rate": 3.0720297138823814e-06, "loss": 0.1369, "step": 467 }, { "epoch": 0.45652969150103645, "grad_norm": 0.9065920605308372, "learning_rate": 3.0641689651012253e-06, "loss": 0.1418, "step": 468 }, { "epoch": 0.4575051822948421, "grad_norm": 0.8602281531380704, "learning_rate": 3.056302334890786e-06, "loss": 0.1377, "step": 469 }, { "epoch": 0.4584806730886477, "grad_norm": 0.8667507800649346, "learning_rate": 3.0484299052602355e-06, "loss": 0.1399, "step": 470 }, { "epoch": 0.45945616388245336, "grad_norm": 0.913545132678621, "learning_rate": 3.040551758279204e-06, "loss": 0.1369, "step": 471 }, { "epoch": 0.460431654676259, "grad_norm": 0.8738158833673688, "learning_rate": 3.032667976076923e-06, "loss": 0.1473, "step": 472 }, { "epoch": 0.4614071454700646, "grad_norm": 0.8801680622206783, "learning_rate": 3.0247786408413725e-06, "loss": 0.1492, "step": 473 }, { "epoch": 0.4623826362638703, "grad_norm": 0.876084094436277, "learning_rate": 3.0168838348184226e-06, "loss": 0.1429, "step": 474 }, { "epoch": 0.4633581270576759, "grad_norm": 0.8429300809717286, "learning_rate": 3.008983640310976e-06, "loss": 0.1239, "step": 475 }, { "epoch": 0.46433361785148153, "grad_norm": 0.8564086119743741, "learning_rate": 3.0010781396781104e-06, "loss": 0.1416, "step": 476 }, { "epoch": 0.4653091086452872, "grad_norm": 0.9263549241409399, "learning_rate": 2.993167415334218e-06, "loss": 0.147, "step": 477 }, { "epoch": 0.4662845994390928, "grad_norm": 0.9103066887514243, "learning_rate": 2.9852515497481494e-06, "loss": 0.149, "step": 478 }, { "epoch": 0.46726009023289844, "grad_norm": 0.8633007227845262, "learning_rate": 2.977330625442352e-06, "loss": 0.1331, "step": 479 }, { "epoch": 0.46823558102670404, "grad_norm": 0.9029118321491253, "learning_rate": 2.9694047249920095e-06, "loss": 0.1508, "step": 480 }, { "epoch": 0.4692110718205097, "grad_norm": 0.8011901107322851, "learning_rate": 2.961473931024182e-06, "loss": 0.1347, "step": 481 }, { "epoch": 0.47018656261431535, "grad_norm": 0.953456056583228, "learning_rate": 2.953538326216944e-06, "loss": 0.1386, "step": 482 }, { "epoch": 0.47116205340812095, "grad_norm": 0.9087484393028001, "learning_rate": 2.9455979932985237e-06, "loss": 0.1378, "step": 483 }, { "epoch": 0.4721375442019266, "grad_norm": 0.8453741920558421, "learning_rate": 2.9376530150464388e-06, "loss": 0.1366, "step": 484 }, { "epoch": 0.4731130349957322, "grad_norm": 0.9941873485158667, "learning_rate": 2.9297034742866336e-06, "loss": 0.1436, "step": 485 }, { "epoch": 0.47408852578953786, "grad_norm": 0.9362727469002547, "learning_rate": 2.921749453892618e-06, "loss": 0.1335, "step": 486 }, { "epoch": 0.4750640165833435, "grad_norm": 0.8486907157670049, "learning_rate": 2.913791036784601e-06, "loss": 0.1311, "step": 487 }, { "epoch": 0.4760395073771491, "grad_norm": 0.941827373025834, "learning_rate": 2.9058283059286275e-06, "loss": 0.1384, "step": 488 }, { "epoch": 0.4770149981709548, "grad_norm": 1.0578890093377278, "learning_rate": 2.8978613443357107e-06, "loss": 0.1532, "step": 489 }, { "epoch": 0.4779904889647604, "grad_norm": 0.9196862748457214, "learning_rate": 2.889890235060972e-06, "loss": 0.1451, "step": 490 }, { "epoch": 0.47896597975856603, "grad_norm": 0.9252094112063749, "learning_rate": 2.88191506120277e-06, "loss": 0.1478, "step": 491 }, { "epoch": 0.4799414705523717, "grad_norm": 0.9936636709844183, "learning_rate": 2.873935905901839e-06, "loss": 0.1383, "step": 492 }, { "epoch": 0.4809169613461773, "grad_norm": 0.9220512507450012, "learning_rate": 2.865952852340417e-06, "loss": 0.1414, "step": 493 }, { "epoch": 0.48189245213998294, "grad_norm": 0.8881328759002474, "learning_rate": 2.8579659837413816e-06, "loss": 0.1381, "step": 494 }, { "epoch": 0.48286794293378854, "grad_norm": 0.9493132086314534, "learning_rate": 2.8499753833673827e-06, "loss": 0.1356, "step": 495 }, { "epoch": 0.4838434337275942, "grad_norm": 0.8586849465773233, "learning_rate": 2.8419811345199736e-06, "loss": 0.1288, "step": 496 }, { "epoch": 0.48481892452139985, "grad_norm": 0.8051743928108915, "learning_rate": 2.8339833205387434e-06, "loss": 0.1272, "step": 497 }, { "epoch": 0.48579441531520545, "grad_norm": 0.8245960012411253, "learning_rate": 2.825982024800445e-06, "loss": 0.136, "step": 498 }, { "epoch": 0.4867699061090111, "grad_norm": 0.8849438079565617, "learning_rate": 2.8179773307181307e-06, "loss": 0.1401, "step": 499 }, { "epoch": 0.4877453969028167, "grad_norm": 1.0134005395077472, "learning_rate": 2.8099693217402807e-06, "loss": 0.1455, "step": 500 }, { "epoch": 0.4877453969028167, "eval_loss": 0.13670527935028076, "eval_runtime": 121.3231, "eval_samples_per_second": 5.465, "eval_steps_per_second": 0.684, "step": 500 }, { "epoch": 0.48872088769662236, "grad_norm": 0.9319226416659817, "learning_rate": 2.8019580813499304e-06, "loss": 0.1443, "step": 501 }, { "epoch": 0.489696378490428, "grad_norm": 0.9042087056844109, "learning_rate": 2.793943693063805e-06, "loss": 0.1382, "step": 502 }, { "epoch": 0.4906718692842336, "grad_norm": 0.8443370914686048, "learning_rate": 2.7859262404314453e-06, "loss": 0.1226, "step": 503 }, { "epoch": 0.4916473600780393, "grad_norm": 1.0430324211655913, "learning_rate": 2.7779058070343367e-06, "loss": 0.147, "step": 504 }, { "epoch": 0.4926228508718449, "grad_norm": 0.9189006680486591, "learning_rate": 2.7698824764850406e-06, "loss": 0.1328, "step": 505 }, { "epoch": 0.49359834166565053, "grad_norm": 0.8384229297869797, "learning_rate": 2.7618563324263208e-06, "loss": 0.1357, "step": 506 }, { "epoch": 0.4945738324594562, "grad_norm": 0.9260402086024966, "learning_rate": 2.7538274585302707e-06, "loss": 0.1328, "step": 507 }, { "epoch": 0.4955493232532618, "grad_norm": 0.8698081640274972, "learning_rate": 2.745795938497443e-06, "loss": 0.1334, "step": 508 }, { "epoch": 0.49652481404706744, "grad_norm": 0.9155791950153026, "learning_rate": 2.737761856055975e-06, "loss": 0.1355, "step": 509 }, { "epoch": 0.49750030484087304, "grad_norm": 0.9389141674328999, "learning_rate": 2.7297252949607195e-06, "loss": 0.1338, "step": 510 }, { "epoch": 0.4984757956346787, "grad_norm": 0.9095624175043169, "learning_rate": 2.721686338992366e-06, "loss": 0.1371, "step": 511 }, { "epoch": 0.49945128642848435, "grad_norm": 0.8431669151453126, "learning_rate": 2.7136450719565732e-06, "loss": 0.1246, "step": 512 }, { "epoch": 0.50042677722229, "grad_norm": 0.8523982451840252, "learning_rate": 2.7056015776830907e-06, "loss": 0.1227, "step": 513 }, { "epoch": 0.5014022680160956, "grad_norm": 0.8800355441963557, "learning_rate": 2.6975559400248876e-06, "loss": 0.1376, "step": 514 }, { "epoch": 0.5023777588099012, "grad_norm": 0.9058981894611112, "learning_rate": 2.6895082428572777e-06, "loss": 0.1417, "step": 515 }, { "epoch": 0.5033532496037069, "grad_norm": 0.8780804727101097, "learning_rate": 2.681458570077045e-06, "loss": 0.1346, "step": 516 }, { "epoch": 0.5043287403975125, "grad_norm": 0.9298727825050748, "learning_rate": 2.673407005601569e-06, "loss": 0.1457, "step": 517 }, { "epoch": 0.5053042311913182, "grad_norm": 0.8489401556484758, "learning_rate": 2.6653536333679504e-06, "loss": 0.1331, "step": 518 }, { "epoch": 0.5062797219851237, "grad_norm": 0.8898533627090344, "learning_rate": 2.6572985373321344e-06, "loss": 0.1434, "step": 519 }, { "epoch": 0.5072552127789294, "grad_norm": 0.9081368712381108, "learning_rate": 2.649241801468039e-06, "loss": 0.1403, "step": 520 }, { "epoch": 0.508230703572735, "grad_norm": 0.8906355358423613, "learning_rate": 2.641183509766675e-06, "loss": 0.1309, "step": 521 }, { "epoch": 0.5092061943665407, "grad_norm": 0.9492976151605389, "learning_rate": 2.633123746235274e-06, "loss": 0.1434, "step": 522 }, { "epoch": 0.5101816851603463, "grad_norm": 0.9210546856991715, "learning_rate": 2.6250625948964125e-06, "loss": 0.1362, "step": 523 }, { "epoch": 0.5111571759541519, "grad_norm": 0.884606453071947, "learning_rate": 2.617000139787132e-06, "loss": 0.1412, "step": 524 }, { "epoch": 0.5121326667479575, "grad_norm": 0.9149743385923425, "learning_rate": 2.608936464958068e-06, "loss": 0.1471, "step": 525 }, { "epoch": 0.5131081575417632, "grad_norm": 0.9748860734433248, "learning_rate": 2.6008716544725692e-06, "loss": 0.1419, "step": 526 }, { "epoch": 0.5140836483355689, "grad_norm": 0.8273165692055457, "learning_rate": 2.5928057924058264e-06, "loss": 0.1289, "step": 527 }, { "epoch": 0.5150591391293745, "grad_norm": 0.9005080399868409, "learning_rate": 2.5847389628439905e-06, "loss": 0.1295, "step": 528 }, { "epoch": 0.51603462992318, "grad_norm": 0.8085104883382979, "learning_rate": 2.576671249883301e-06, "loss": 0.1341, "step": 529 }, { "epoch": 0.5170101207169857, "grad_norm": 0.9357055967861888, "learning_rate": 2.568602737629204e-06, "loss": 0.1383, "step": 530 }, { "epoch": 0.5179856115107914, "grad_norm": 0.9860801651928193, "learning_rate": 2.5605335101954796e-06, "loss": 0.1395, "step": 531 }, { "epoch": 0.518961102304597, "grad_norm": 1.0023841873957609, "learning_rate": 2.552463651703365e-06, "loss": 0.1391, "step": 532 }, { "epoch": 0.5199365930984027, "grad_norm": 0.9970746647895117, "learning_rate": 2.5443932462806733e-06, "loss": 0.1345, "step": 533 }, { "epoch": 0.5209120838922082, "grad_norm": 0.8823424259442537, "learning_rate": 2.5363223780609214e-06, "loss": 0.1303, "step": 534 }, { "epoch": 0.5218875746860139, "grad_norm": 0.8990272295179849, "learning_rate": 2.52825113118245e-06, "loss": 0.138, "step": 535 }, { "epoch": 0.5228630654798195, "grad_norm": 0.964490190997079, "learning_rate": 2.520179589787547e-06, "loss": 0.1419, "step": 536 }, { "epoch": 0.5238385562736252, "grad_norm": 0.9850971200909243, "learning_rate": 2.5121078380215713e-06, "loss": 0.1378, "step": 537 }, { "epoch": 0.5248140470674308, "grad_norm": 0.8473816563308804, "learning_rate": 2.5040359600320747e-06, "loss": 0.1331, "step": 538 }, { "epoch": 0.5257895378612364, "grad_norm": 0.9063261263061652, "learning_rate": 2.4959640399679253e-06, "loss": 0.143, "step": 539 }, { "epoch": 0.526765028655042, "grad_norm": 0.8833983728636542, "learning_rate": 2.4878921619784295e-06, "loss": 0.1337, "step": 540 }, { "epoch": 0.5277405194488477, "grad_norm": 0.8657377624673289, "learning_rate": 2.4798204102124533e-06, "loss": 0.1344, "step": 541 }, { "epoch": 0.5287160102426534, "grad_norm": 0.9202779408760426, "learning_rate": 2.4717488688175513e-06, "loss": 0.1315, "step": 542 }, { "epoch": 0.529691501036459, "grad_norm": 0.9088907887564982, "learning_rate": 2.4636776219390794e-06, "loss": 0.1329, "step": 543 }, { "epoch": 0.5306669918302646, "grad_norm": 0.9310598535059849, "learning_rate": 2.4556067537193276e-06, "loss": 0.1338, "step": 544 }, { "epoch": 0.5316424826240702, "grad_norm": 0.9003405042179257, "learning_rate": 2.4475363482966356e-06, "loss": 0.1362, "step": 545 }, { "epoch": 0.5326179734178759, "grad_norm": 0.8206989405119627, "learning_rate": 2.4394664898045208e-06, "loss": 0.133, "step": 546 }, { "epoch": 0.5335934642116815, "grad_norm": 0.8742977014572649, "learning_rate": 2.4313972623707964e-06, "loss": 0.1343, "step": 547 }, { "epoch": 0.5345689550054872, "grad_norm": 0.8727361447556689, "learning_rate": 2.4233287501167e-06, "loss": 0.1318, "step": 548 }, { "epoch": 0.5355444457992927, "grad_norm": 0.9233330686910666, "learning_rate": 2.4152610371560095e-06, "loss": 0.141, "step": 549 }, { "epoch": 0.5365199365930984, "grad_norm": 0.8918087087553142, "learning_rate": 2.4071942075941744e-06, "loss": 0.1329, "step": 550 }, { "epoch": 0.537495427386904, "grad_norm": 0.849661983110438, "learning_rate": 2.3991283455274316e-06, "loss": 0.1288, "step": 551 }, { "epoch": 0.5384709181807097, "grad_norm": 0.8538891952062707, "learning_rate": 2.391063535041933e-06, "loss": 0.1322, "step": 552 }, { "epoch": 0.5394464089745153, "grad_norm": 0.8801508104924974, "learning_rate": 2.3829998602128685e-06, "loss": 0.1239, "step": 553 }, { "epoch": 0.5404218997683209, "grad_norm": 0.8524417866635733, "learning_rate": 2.3749374051035883e-06, "loss": 0.1308, "step": 554 }, { "epoch": 0.5413973905621265, "grad_norm": 0.8827154465529061, "learning_rate": 2.3668762537647254e-06, "loss": 0.1377, "step": 555 }, { "epoch": 0.5423728813559322, "grad_norm": 0.8050527885767014, "learning_rate": 2.358816490233326e-06, "loss": 0.127, "step": 556 }, { "epoch": 0.5433483721497379, "grad_norm": 0.8172596997607479, "learning_rate": 2.3507581985319623e-06, "loss": 0.1212, "step": 557 }, { "epoch": 0.5443238629435435, "grad_norm": 0.8963200330399226, "learning_rate": 2.342701462667866e-06, "loss": 0.1401, "step": 558 }, { "epoch": 0.5452993537373491, "grad_norm": 0.887052377675435, "learning_rate": 2.3346463666320512e-06, "loss": 0.1323, "step": 559 }, { "epoch": 0.5462748445311547, "grad_norm": 0.9127806776302525, "learning_rate": 2.3265929943984317e-06, "loss": 0.1361, "step": 560 }, { "epoch": 0.5472503353249604, "grad_norm": 0.9147589526822345, "learning_rate": 2.318541429922956e-06, "loss": 0.1363, "step": 561 }, { "epoch": 0.548225826118766, "grad_norm": 0.8944082387887866, "learning_rate": 2.310491757142723e-06, "loss": 0.1414, "step": 562 }, { "epoch": 0.5492013169125717, "grad_norm": 0.8666818426784915, "learning_rate": 2.3024440599751132e-06, "loss": 0.1338, "step": 563 }, { "epoch": 0.5501768077063772, "grad_norm": 0.8384112838731508, "learning_rate": 2.29439842231691e-06, "loss": 0.128, "step": 564 }, { "epoch": 0.5511522985001829, "grad_norm": 0.8598917633955132, "learning_rate": 2.2863549280434285e-06, "loss": 0.1378, "step": 565 }, { "epoch": 0.5521277892939885, "grad_norm": 0.8490812429791822, "learning_rate": 2.2783136610076345e-06, "loss": 0.1302, "step": 566 }, { "epoch": 0.5531032800877942, "grad_norm": 0.9671910225128955, "learning_rate": 2.270274705039282e-06, "loss": 0.1371, "step": 567 }, { "epoch": 0.5540787708815998, "grad_norm": 0.9181284807350794, "learning_rate": 2.2622381439440255e-06, "loss": 0.1435, "step": 568 }, { "epoch": 0.5550542616754054, "grad_norm": 0.7950987192294108, "learning_rate": 2.2542040615025584e-06, "loss": 0.1219, "step": 569 }, { "epoch": 0.556029752469211, "grad_norm": 1.0089717587920861, "learning_rate": 2.24617254146973e-06, "loss": 0.1292, "step": 570 }, { "epoch": 0.5570052432630167, "grad_norm": 0.9858255149161776, "learning_rate": 2.23814366757368e-06, "loss": 0.1393, "step": 571 }, { "epoch": 0.5579807340568224, "grad_norm": 0.8129514529071972, "learning_rate": 2.23011752351496e-06, "loss": 0.1235, "step": 572 }, { "epoch": 0.558956224850628, "grad_norm": 0.9188253431247291, "learning_rate": 2.222094192965664e-06, "loss": 0.1275, "step": 573 }, { "epoch": 0.5599317156444336, "grad_norm": 0.9256786564401078, "learning_rate": 2.214073759568555e-06, "loss": 0.1383, "step": 574 }, { "epoch": 0.5609072064382392, "grad_norm": 0.810478233219589, "learning_rate": 2.2060563069361955e-06, "loss": 0.1221, "step": 575 }, { "epoch": 0.5618826972320449, "grad_norm": 0.8139352487642532, "learning_rate": 2.19804191865007e-06, "loss": 0.1203, "step": 576 }, { "epoch": 0.5628581880258505, "grad_norm": 0.8148607289634421, "learning_rate": 2.19003067825972e-06, "loss": 0.127, "step": 577 }, { "epoch": 0.5638336788196562, "grad_norm": 0.8730605763594029, "learning_rate": 2.1820226692818693e-06, "loss": 0.1294, "step": 578 }, { "epoch": 0.5648091696134617, "grad_norm": 0.9081540528860818, "learning_rate": 2.174017975199556e-06, "loss": 0.1418, "step": 579 }, { "epoch": 0.5657846604072674, "grad_norm": 0.964077636816459, "learning_rate": 2.1660166794612574e-06, "loss": 0.1549, "step": 580 }, { "epoch": 0.566760151201073, "grad_norm": 0.8682612244819073, "learning_rate": 2.158018865480027e-06, "loss": 0.1339, "step": 581 }, { "epoch": 0.5677356419948787, "grad_norm": 0.9087722584693885, "learning_rate": 2.1500246166326177e-06, "loss": 0.138, "step": 582 }, { "epoch": 0.5687111327886843, "grad_norm": 0.8284282518619499, "learning_rate": 2.1420340162586196e-06, "loss": 0.1273, "step": 583 }, { "epoch": 0.5696866235824899, "grad_norm": 0.899367551533117, "learning_rate": 2.1340471476595836e-06, "loss": 0.1325, "step": 584 }, { "epoch": 0.5706621143762955, "grad_norm": 0.8721503402061037, "learning_rate": 2.1260640940981616e-06, "loss": 0.1373, "step": 585 }, { "epoch": 0.5716376051701012, "grad_norm": 0.8007443556277751, "learning_rate": 2.11808493879723e-06, "loss": 0.1314, "step": 586 }, { "epoch": 0.5726130959639069, "grad_norm": 0.843399287875525, "learning_rate": 2.110109764939029e-06, "loss": 0.1273, "step": 587 }, { "epoch": 0.5735885867577125, "grad_norm": 0.8694236616188882, "learning_rate": 2.10213865566429e-06, "loss": 0.1292, "step": 588 }, { "epoch": 0.5745640775515181, "grad_norm": 0.944807048930237, "learning_rate": 2.0941716940713733e-06, "loss": 0.136, "step": 589 }, { "epoch": 0.5755395683453237, "grad_norm": 0.8012154070123548, "learning_rate": 2.086208963215399e-06, "loss": 0.1219, "step": 590 }, { "epoch": 0.5765150591391294, "grad_norm": 0.9713169032385847, "learning_rate": 2.0782505461073822e-06, "loss": 0.1272, "step": 591 }, { "epoch": 0.577490549932935, "grad_norm": 0.833986482831455, "learning_rate": 2.0702965257133664e-06, "loss": 0.1211, "step": 592 }, { "epoch": 0.5784660407267407, "grad_norm": 0.9223111755392438, "learning_rate": 2.062346984953562e-06, "loss": 0.1229, "step": 593 }, { "epoch": 0.5794415315205462, "grad_norm": 0.9302281584720075, "learning_rate": 2.0544020067014776e-06, "loss": 0.1383, "step": 594 }, { "epoch": 0.5804170223143519, "grad_norm": 0.9426149459659149, "learning_rate": 2.0464616737830566e-06, "loss": 0.1337, "step": 595 }, { "epoch": 0.5813925131081575, "grad_norm": 0.8882577504858508, "learning_rate": 2.03852606897582e-06, "loss": 0.1244, "step": 596 }, { "epoch": 0.5823680039019632, "grad_norm": 0.8456021234816907, "learning_rate": 2.0305952750079918e-06, "loss": 0.1324, "step": 597 }, { "epoch": 0.5833434946957688, "grad_norm": 0.933740394118853, "learning_rate": 2.0226693745576494e-06, "loss": 0.136, "step": 598 }, { "epoch": 0.5843189854895744, "grad_norm": 0.9462976911157949, "learning_rate": 2.0147484502518514e-06, "loss": 0.1286, "step": 599 }, { "epoch": 0.58529447628338, "grad_norm": 0.9098208372409959, "learning_rate": 2.006832584665783e-06, "loss": 0.132, "step": 600 }, { "epoch": 0.5862699670771857, "grad_norm": 0.795331526345968, "learning_rate": 1.99892186032189e-06, "loss": 0.1151, "step": 601 }, { "epoch": 0.5872454578709914, "grad_norm": 0.8664446011660321, "learning_rate": 1.9910163596890247e-06, "loss": 0.1288, "step": 602 }, { "epoch": 0.588220948664797, "grad_norm": 0.8610758137851285, "learning_rate": 1.983116165181578e-06, "loss": 0.1239, "step": 603 }, { "epoch": 0.5891964394586026, "grad_norm": 0.8560179576447229, "learning_rate": 1.9752213591586288e-06, "loss": 0.1276, "step": 604 }, { "epoch": 0.5901719302524082, "grad_norm": 0.8603930368509891, "learning_rate": 1.9673320239230783e-06, "loss": 0.1305, "step": 605 }, { "epoch": 0.5911474210462139, "grad_norm": 0.8443701642736344, "learning_rate": 1.9594482417207973e-06, "loss": 0.1261, "step": 606 }, { "epoch": 0.5921229118400195, "grad_norm": 0.7909939129569269, "learning_rate": 1.951570094739765e-06, "loss": 0.1211, "step": 607 }, { "epoch": 0.5930984026338252, "grad_norm": 0.931299792718923, "learning_rate": 1.9436976651092143e-06, "loss": 0.1382, "step": 608 }, { "epoch": 0.5940738934276307, "grad_norm": 0.8781253875372456, "learning_rate": 1.9358310348987755e-06, "loss": 0.1273, "step": 609 }, { "epoch": 0.5950493842214364, "grad_norm": 0.8570421027558416, "learning_rate": 1.9279702861176203e-06, "loss": 0.1342, "step": 610 }, { "epoch": 0.596024875015242, "grad_norm": 0.8451871775967923, "learning_rate": 1.9201155007136045e-06, "loss": 0.1365, "step": 611 }, { "epoch": 0.5970003658090477, "grad_norm": 0.9384520823084104, "learning_rate": 1.9122667605724202e-06, "loss": 0.1348, "step": 612 }, { "epoch": 0.5979758566028534, "grad_norm": 0.8922350956011054, "learning_rate": 1.9044241475167339e-06, "loss": 0.1348, "step": 613 }, { "epoch": 0.5989513473966589, "grad_norm": 0.8385907676608247, "learning_rate": 1.8965877433053409e-06, "loss": 0.1287, "step": 614 }, { "epoch": 0.5999268381904646, "grad_norm": 0.8438875643530197, "learning_rate": 1.8887576296323069e-06, "loss": 0.1234, "step": 615 }, { "epoch": 0.6009023289842702, "grad_norm": 0.8782492476271099, "learning_rate": 1.8809338881261213e-06, "loss": 0.1312, "step": 616 }, { "epoch": 0.6018778197780759, "grad_norm": 0.8210821361205313, "learning_rate": 1.8731166003488427e-06, "loss": 0.1256, "step": 617 }, { "epoch": 0.6028533105718815, "grad_norm": 0.8313931745166832, "learning_rate": 1.86530584779525e-06, "loss": 0.1269, "step": 618 }, { "epoch": 0.6038288013656871, "grad_norm": 0.8248814372114766, "learning_rate": 1.857501711891993e-06, "loss": 0.1277, "step": 619 }, { "epoch": 0.6048042921594927, "grad_norm": 0.8300683641876326, "learning_rate": 1.8497042739967447e-06, "loss": 0.1321, "step": 620 }, { "epoch": 0.6057797829532984, "grad_norm": 0.8967700470622615, "learning_rate": 1.841913615397348e-06, "loss": 0.1273, "step": 621 }, { "epoch": 0.606755273747104, "grad_norm": 0.937831066737166, "learning_rate": 1.834129817310977e-06, "loss": 0.1293, "step": 622 }, { "epoch": 0.6077307645409097, "grad_norm": 0.8608866113112006, "learning_rate": 1.826352960883281e-06, "loss": 0.1276, "step": 623 }, { "epoch": 0.6087062553347152, "grad_norm": 0.8562897857077051, "learning_rate": 1.8185831271875457e-06, "loss": 0.1204, "step": 624 }, { "epoch": 0.6096817461285209, "grad_norm": 0.8326388599515545, "learning_rate": 1.8108203972238436e-06, "loss": 0.124, "step": 625 }, { "epoch": 0.6106572369223265, "grad_norm": 0.9146045731524001, "learning_rate": 1.8030648519181926e-06, "loss": 0.1327, "step": 626 }, { "epoch": 0.6116327277161322, "grad_norm": 0.8125344568360147, "learning_rate": 1.7953165721217086e-06, "loss": 0.1206, "step": 627 }, { "epoch": 0.6126082185099379, "grad_norm": 0.8788372999011178, "learning_rate": 1.7875756386097686e-06, "loss": 0.1376, "step": 628 }, { "epoch": 0.6135837093037434, "grad_norm": 0.8647255719378499, "learning_rate": 1.779842132081162e-06, "loss": 0.1243, "step": 629 }, { "epoch": 0.614559200097549, "grad_norm": 0.8756476987196234, "learning_rate": 1.7721161331572522e-06, "loss": 0.132, "step": 630 }, { "epoch": 0.6155346908913547, "grad_norm": 0.9563185708098122, "learning_rate": 1.7643977223811392e-06, "loss": 0.1364, "step": 631 }, { "epoch": 0.6165101816851604, "grad_norm": 0.8320678225276503, "learning_rate": 1.7566869802168132e-06, "loss": 0.1189, "step": 632 }, { "epoch": 0.617485672478966, "grad_norm": 0.9093829860333068, "learning_rate": 1.7489839870483236e-06, "loss": 0.1277, "step": 633 }, { "epoch": 0.6184611632727716, "grad_norm": 0.9011561073082258, "learning_rate": 1.7412888231789327e-06, "loss": 0.1358, "step": 634 }, { "epoch": 0.6194366540665772, "grad_norm": 1.0084984996858963, "learning_rate": 1.7336015688302869e-06, "loss": 0.1268, "step": 635 }, { "epoch": 0.6204121448603829, "grad_norm": 0.9152283340289553, "learning_rate": 1.725922304141573e-06, "loss": 0.1295, "step": 636 }, { "epoch": 0.6213876356541885, "grad_norm": 0.9238351863571306, "learning_rate": 1.718251109168688e-06, "loss": 0.1318, "step": 637 }, { "epoch": 0.6223631264479942, "grad_norm": 0.866304353620812, "learning_rate": 1.7105880638834007e-06, "loss": 0.1212, "step": 638 }, { "epoch": 0.6233386172417997, "grad_norm": 0.95501299054228, "learning_rate": 1.7029332481725234e-06, "loss": 0.1364, "step": 639 }, { "epoch": 0.6243141080356054, "grad_norm": 0.8839661191980624, "learning_rate": 1.6952867418370707e-06, "loss": 0.127, "step": 640 }, { "epoch": 0.625289598829411, "grad_norm": 0.8559388882214561, "learning_rate": 1.6876486245914375e-06, "loss": 0.1249, "step": 641 }, { "epoch": 0.6262650896232167, "grad_norm": 0.9364435799205211, "learning_rate": 1.6800189760625585e-06, "loss": 0.1329, "step": 642 }, { "epoch": 0.6272405804170224, "grad_norm": 0.8651014555168739, "learning_rate": 1.672397875789086e-06, "loss": 0.125, "step": 643 }, { "epoch": 0.6282160712108279, "grad_norm": 0.8892108367186486, "learning_rate": 1.6647854032205547e-06, "loss": 0.1304, "step": 644 }, { "epoch": 0.6291915620046336, "grad_norm": 0.8503679132174746, "learning_rate": 1.6571816377165568e-06, "loss": 0.1249, "step": 645 }, { "epoch": 0.6301670527984392, "grad_norm": 0.9314057258730163, "learning_rate": 1.6495866585459142e-06, "loss": 0.131, "step": 646 }, { "epoch": 0.6311425435922449, "grad_norm": 0.9002310278187038, "learning_rate": 1.6420005448858522e-06, "loss": 0.1329, "step": 647 }, { "epoch": 0.6321180343860505, "grad_norm": 0.8263610327236112, "learning_rate": 1.6344233758211717e-06, "loss": 0.1217, "step": 648 }, { "epoch": 0.6330935251798561, "grad_norm": 0.8099996895450952, "learning_rate": 1.6268552303434298e-06, "loss": 0.1211, "step": 649 }, { "epoch": 0.6340690159736617, "grad_norm": 0.8528126491211017, "learning_rate": 1.6192961873501096e-06, "loss": 0.1297, "step": 650 }, { "epoch": 0.6350445067674674, "grad_norm": 0.8585959373381494, "learning_rate": 1.611746325643805e-06, "loss": 0.1295, "step": 651 }, { "epoch": 0.636019997561273, "grad_norm": 0.8655236759185483, "learning_rate": 1.6042057239313919e-06, "loss": 0.1246, "step": 652 }, { "epoch": 0.6369954883550787, "grad_norm": 0.8207113182539624, "learning_rate": 1.5966744608232137e-06, "loss": 0.12, "step": 653 }, { "epoch": 0.6379709791488842, "grad_norm": 0.8614663092183132, "learning_rate": 1.5891526148322594e-06, "loss": 0.1187, "step": 654 }, { "epoch": 0.6389464699426899, "grad_norm": 0.9001233935199348, "learning_rate": 1.5816402643733441e-06, "loss": 0.1276, "step": 655 }, { "epoch": 0.6399219607364955, "grad_norm": 0.8508200436128184, "learning_rate": 1.5741374877622922e-06, "loss": 0.1267, "step": 656 }, { "epoch": 0.6408974515303012, "grad_norm": 0.9227739705381223, "learning_rate": 1.566644363215123e-06, "loss": 0.1344, "step": 657 }, { "epoch": 0.6418729423241069, "grad_norm": 0.853646215929862, "learning_rate": 1.5591609688472313e-06, "loss": 0.1363, "step": 658 }, { "epoch": 0.6428484331179124, "grad_norm": 0.9878682768114257, "learning_rate": 1.551687382672578e-06, "loss": 0.1354, "step": 659 }, { "epoch": 0.643823923911718, "grad_norm": 0.8622608892200061, "learning_rate": 1.5442236826028705e-06, "loss": 0.1281, "step": 660 }, { "epoch": 0.6447994147055237, "grad_norm": 0.8868098887146023, "learning_rate": 1.5367699464467596e-06, "loss": 0.1316, "step": 661 }, { "epoch": 0.6457749054993294, "grad_norm": 0.8568792305689042, "learning_rate": 1.5293262519090169e-06, "loss": 0.1289, "step": 662 }, { "epoch": 0.646750396293135, "grad_norm": 0.8345103045814822, "learning_rate": 1.5218926765897345e-06, "loss": 0.1238, "step": 663 }, { "epoch": 0.6477258870869406, "grad_norm": 0.8296168710839235, "learning_rate": 1.5144692979835103e-06, "loss": 0.1279, "step": 664 }, { "epoch": 0.6487013778807462, "grad_norm": 0.8651995871451914, "learning_rate": 1.5070561934786437e-06, "loss": 0.131, "step": 665 }, { "epoch": 0.6496768686745519, "grad_norm": 0.9169910575714909, "learning_rate": 1.4996534403563267e-06, "loss": 0.1334, "step": 666 }, { "epoch": 0.6506523594683575, "grad_norm": 0.7693306954167114, "learning_rate": 1.4922611157898364e-06, "loss": 0.1168, "step": 667 }, { "epoch": 0.6516278502621632, "grad_norm": 0.863479055519848, "learning_rate": 1.4848792968437376e-06, "loss": 0.1248, "step": 668 }, { "epoch": 0.6526033410559687, "grad_norm": 0.8668901068891359, "learning_rate": 1.477508060473069e-06, "loss": 0.1321, "step": 669 }, { "epoch": 0.6535788318497744, "grad_norm": 0.8378480519694856, "learning_rate": 1.4701474835225515e-06, "loss": 0.1312, "step": 670 }, { "epoch": 0.65455432264358, "grad_norm": 0.811001196208897, "learning_rate": 1.4627976427257784e-06, "loss": 0.1201, "step": 671 }, { "epoch": 0.6555298134373857, "grad_norm": 0.8636188619772824, "learning_rate": 1.4554586147044225e-06, "loss": 0.1365, "step": 672 }, { "epoch": 0.6565053042311914, "grad_norm": 0.85744996390658, "learning_rate": 1.4481304759674303e-06, "loss": 0.1249, "step": 673 }, { "epoch": 0.6574807950249969, "grad_norm": 0.8435636506247961, "learning_rate": 1.4408133029102322e-06, "loss": 0.1195, "step": 674 }, { "epoch": 0.6584562858188026, "grad_norm": 0.8509255085231769, "learning_rate": 1.4335071718139379e-06, "loss": 0.1183, "step": 675 }, { "epoch": 0.6594317766126082, "grad_norm": 0.8793960273956329, "learning_rate": 1.4262121588445493e-06, "loss": 0.1245, "step": 676 }, { "epoch": 0.6604072674064139, "grad_norm": 0.8970156523355322, "learning_rate": 1.418928340052159e-06, "loss": 0.1286, "step": 677 }, { "epoch": 0.6613827582002195, "grad_norm": 0.8644162321708901, "learning_rate": 1.411655791370164e-06, "loss": 0.1252, "step": 678 }, { "epoch": 0.6623582489940251, "grad_norm": 0.8396437394115862, "learning_rate": 1.4043945886144673e-06, "loss": 0.1179, "step": 679 }, { "epoch": 0.6633337397878307, "grad_norm": 0.8448586465328236, "learning_rate": 1.397144807482696e-06, "loss": 0.1252, "step": 680 }, { "epoch": 0.6643092305816364, "grad_norm": 0.8718372351836149, "learning_rate": 1.3899065235534031e-06, "loss": 0.1216, "step": 681 }, { "epoch": 0.665284721375442, "grad_norm": 0.8407139122291305, "learning_rate": 1.382679812285287e-06, "loss": 0.1283, "step": 682 }, { "epoch": 0.6662602121692477, "grad_norm": 0.8404593100461625, "learning_rate": 1.375464749016399e-06, "loss": 0.1188, "step": 683 }, { "epoch": 0.6672357029630532, "grad_norm": 0.940858079210062, "learning_rate": 1.3682614089633637e-06, "loss": 0.129, "step": 684 }, { "epoch": 0.6682111937568589, "grad_norm": 0.8569084229583452, "learning_rate": 1.3610698672205897e-06, "loss": 0.1281, "step": 685 }, { "epoch": 0.6691866845506645, "grad_norm": 0.8797960805331662, "learning_rate": 1.3538901987594882e-06, "loss": 0.1292, "step": 686 }, { "epoch": 0.6701621753444702, "grad_norm": 0.8374942493553367, "learning_rate": 1.3467224784276961e-06, "loss": 0.1309, "step": 687 }, { "epoch": 0.6711376661382759, "grad_norm": 0.7710018857731183, "learning_rate": 1.339566780948287e-06, "loss": 0.1139, "step": 688 }, { "epoch": 0.6721131569320814, "grad_norm": 0.8657391409702617, "learning_rate": 1.3324231809189985e-06, "loss": 0.1309, "step": 689 }, { "epoch": 0.6730886477258871, "grad_norm": 0.8127948325482396, "learning_rate": 1.3252917528114556e-06, "loss": 0.1267, "step": 690 }, { "epoch": 0.6740641385196927, "grad_norm": 0.8122952852781401, "learning_rate": 1.3181725709703874e-06, "loss": 0.1213, "step": 691 }, { "epoch": 0.6750396293134984, "grad_norm": 0.851528828559218, "learning_rate": 1.3110657096128606e-06, "loss": 0.12, "step": 692 }, { "epoch": 0.676015120107304, "grad_norm": 0.86862592870714, "learning_rate": 1.3039712428274976e-06, "loss": 0.1231, "step": 693 }, { "epoch": 0.6769906109011096, "grad_norm": 0.8696639730293406, "learning_rate": 1.2968892445737117e-06, "loss": 0.1294, "step": 694 }, { "epoch": 0.6779661016949152, "grad_norm": 0.9002780776036385, "learning_rate": 1.2898197886809289e-06, "loss": 0.1256, "step": 695 }, { "epoch": 0.6789415924887209, "grad_norm": 0.8530790957831813, "learning_rate": 1.2827629488478254e-06, "loss": 0.13, "step": 696 }, { "epoch": 0.6799170832825265, "grad_norm": 0.8450717054989292, "learning_rate": 1.2757187986415528e-06, "loss": 0.1306, "step": 697 }, { "epoch": 0.6808925740763322, "grad_norm": 0.8714024729066125, "learning_rate": 1.2686874114969767e-06, "loss": 0.1323, "step": 698 }, { "epoch": 0.6818680648701377, "grad_norm": 0.8591419822215062, "learning_rate": 1.2616688607159045e-06, "loss": 0.1396, "step": 699 }, { "epoch": 0.6828435556639434, "grad_norm": 0.8361114781431095, "learning_rate": 1.2546632194663305e-06, "loss": 0.1254, "step": 700 }, { "epoch": 0.683819046457749, "grad_norm": 0.8242742327238427, "learning_rate": 1.2476705607816639e-06, "loss": 0.1234, "step": 701 }, { "epoch": 0.6847945372515547, "grad_norm": 0.8637116393284514, "learning_rate": 1.2406909575599717e-06, "loss": 0.125, "step": 702 }, { "epoch": 0.6857700280453604, "grad_norm": 0.8884418622966431, "learning_rate": 1.2337244825632217e-06, "loss": 0.135, "step": 703 }, { "epoch": 0.6867455188391659, "grad_norm": 0.8275460859037888, "learning_rate": 1.2267712084165168e-06, "loss": 0.1158, "step": 704 }, { "epoch": 0.6877210096329716, "grad_norm": 0.8553642751195779, "learning_rate": 1.2198312076073427e-06, "loss": 0.128, "step": 705 }, { "epoch": 0.6886965004267772, "grad_norm": 0.8837324796776882, "learning_rate": 1.2129045524848138e-06, "loss": 0.128, "step": 706 }, { "epoch": 0.6896719912205829, "grad_norm": 0.8270144975513911, "learning_rate": 1.205991315258914e-06, "loss": 0.1185, "step": 707 }, { "epoch": 0.6906474820143885, "grad_norm": 0.8738272053373212, "learning_rate": 1.199091567999746e-06, "loss": 0.1264, "step": 708 }, { "epoch": 0.6916229728081941, "grad_norm": 0.8344314699090399, "learning_rate": 1.1922053826367833e-06, "loss": 0.1152, "step": 709 }, { "epoch": 0.6925984636019997, "grad_norm": 0.8650165946821683, "learning_rate": 1.1853328309581139e-06, "loss": 0.1283, "step": 710 }, { "epoch": 0.6935739543958054, "grad_norm": 0.8374700294304306, "learning_rate": 1.1784739846096982e-06, "loss": 0.1238, "step": 711 }, { "epoch": 0.694549445189611, "grad_norm": 0.8380502969131793, "learning_rate": 1.1716289150946173e-06, "loss": 0.1234, "step": 712 }, { "epoch": 0.6955249359834167, "grad_norm": 0.9064794381405805, "learning_rate": 1.1647976937723315e-06, "loss": 0.1306, "step": 713 }, { "epoch": 0.6965004267772222, "grad_norm": 0.832398941447385, "learning_rate": 1.1579803918579322e-06, "loss": 0.125, "step": 714 }, { "epoch": 0.6974759175710279, "grad_norm": 0.8174727802602623, "learning_rate": 1.1511770804214046e-06, "loss": 0.1248, "step": 715 }, { "epoch": 0.6984514083648335, "grad_norm": 0.8474018969337508, "learning_rate": 1.1443878303868805e-06, "loss": 0.1338, "step": 716 }, { "epoch": 0.6994268991586392, "grad_norm": 0.8466582401794491, "learning_rate": 1.1376127125319065e-06, "loss": 0.1148, "step": 717 }, { "epoch": 0.7004023899524449, "grad_norm": 0.8912296637806607, "learning_rate": 1.1308517974866973e-06, "loss": 0.1274, "step": 718 }, { "epoch": 0.7013778807462504, "grad_norm": 0.9491089244563283, "learning_rate": 1.1241051557334087e-06, "loss": 0.124, "step": 719 }, { "epoch": 0.7023533715400561, "grad_norm": 0.9380181290695212, "learning_rate": 1.117372857605395e-06, "loss": 0.1354, "step": 720 }, { "epoch": 0.7033288623338617, "grad_norm": 0.8177994093605291, "learning_rate": 1.110654973286481e-06, "loss": 0.1164, "step": 721 }, { "epoch": 0.7043043531276674, "grad_norm": 0.8347927318132022, "learning_rate": 1.1039515728102273e-06, "loss": 0.1275, "step": 722 }, { "epoch": 0.705279843921473, "grad_norm": 0.8970474242898054, "learning_rate": 1.0972627260592014e-06, "loss": 0.1231, "step": 723 }, { "epoch": 0.7062553347152786, "grad_norm": 0.795847982137367, "learning_rate": 1.0905885027642484e-06, "loss": 0.1177, "step": 724 }, { "epoch": 0.7072308255090842, "grad_norm": 0.8814347768470174, "learning_rate": 1.0839289725037669e-06, "loss": 0.1334, "step": 725 }, { "epoch": 0.7082063163028899, "grad_norm": 0.828253666509173, "learning_rate": 1.0772842047029786e-06, "loss": 0.1218, "step": 726 }, { "epoch": 0.7091818070966955, "grad_norm": 0.812866815136845, "learning_rate": 1.0706542686332102e-06, "loss": 0.1143, "step": 727 }, { "epoch": 0.7101572978905012, "grad_norm": 0.8556204829905426, "learning_rate": 1.064039233411166e-06, "loss": 0.1264, "step": 728 }, { "epoch": 0.7111327886843067, "grad_norm": 0.8841697516339617, "learning_rate": 1.057439167998212e-06, "loss": 0.1278, "step": 729 }, { "epoch": 0.7121082794781124, "grad_norm": 0.8636362569900907, "learning_rate": 1.050854141199652e-06, "loss": 0.1235, "step": 730 }, { "epoch": 0.713083770271918, "grad_norm": 0.8283746245515586, "learning_rate": 1.0442842216640168e-06, "loss": 0.1208, "step": 731 }, { "epoch": 0.7140592610657237, "grad_norm": 0.8728491657133771, "learning_rate": 1.037729477882341e-06, "loss": 0.117, "step": 732 }, { "epoch": 0.7150347518595294, "grad_norm": 0.8915555008915115, "learning_rate": 1.0311899781874563e-06, "loss": 0.1293, "step": 733 }, { "epoch": 0.7160102426533349, "grad_norm": 0.8380362900075755, "learning_rate": 1.0246657907532727e-06, "loss": 0.1216, "step": 734 }, { "epoch": 0.7169857334471406, "grad_norm": 0.8565497720933896, "learning_rate": 1.018156983594073e-06, "loss": 0.1149, "step": 735 }, { "epoch": 0.7179612242409462, "grad_norm": 0.8358087384760822, "learning_rate": 1.0116636245637997e-06, "loss": 0.115, "step": 736 }, { "epoch": 0.7189367150347519, "grad_norm": 0.8847500849421209, "learning_rate": 1.0051857813553516e-06, "loss": 0.123, "step": 737 }, { "epoch": 0.7199122058285575, "grad_norm": 0.8794173909505707, "learning_rate": 9.987235214998741e-07, "loss": 0.1277, "step": 738 }, { "epoch": 0.7208876966223631, "grad_norm": 0.8665043928636729, "learning_rate": 9.922769123660564e-07, "loss": 0.1197, "step": 739 }, { "epoch": 0.7218631874161687, "grad_norm": 0.8361820544782747, "learning_rate": 9.858460211594331e-07, "loss": 0.124, "step": 740 }, { "epoch": 0.7228386782099744, "grad_norm": 0.846259877698448, "learning_rate": 9.79430914921677e-07, "loss": 0.1207, "step": 741 }, { "epoch": 0.72381416900378, "grad_norm": 0.8468390031221599, "learning_rate": 9.730316605299041e-07, "loss": 0.1226, "step": 742 }, { "epoch": 0.7247896597975857, "grad_norm": 0.9014801026605848, "learning_rate": 9.666483246959785e-07, "loss": 0.119, "step": 743 }, { "epoch": 0.7257651505913912, "grad_norm": 0.9115482145040746, "learning_rate": 9.602809739658105e-07, "loss": 0.1245, "step": 744 }, { "epoch": 0.7267406413851969, "grad_norm": 0.8643571047269124, "learning_rate": 9.53929674718668e-07, "loss": 0.1216, "step": 745 }, { "epoch": 0.7277161321790026, "grad_norm": 0.8382233183752549, "learning_rate": 9.475944931664851e-07, "loss": 0.125, "step": 746 }, { "epoch": 0.7286916229728082, "grad_norm": 0.8536460797398011, "learning_rate": 9.412754953531664e-07, "loss": 0.1223, "step": 747 }, { "epoch": 0.7296671137666139, "grad_norm": 0.8756760541580888, "learning_rate": 9.349727471539052e-07, "loss": 0.1215, "step": 748 }, { "epoch": 0.7306426045604194, "grad_norm": 0.884349162839785, "learning_rate": 9.286863142744907e-07, "loss": 0.1237, "step": 749 }, { "epoch": 0.7316180953542251, "grad_norm": 0.8232259080657391, "learning_rate": 9.224162622506283e-07, "loss": 0.118, "step": 750 }, { "epoch": 0.7325935861480307, "grad_norm": 0.8865909364907526, "learning_rate": 9.161626564472511e-07, "loss": 0.1271, "step": 751 }, { "epoch": 0.7335690769418364, "grad_norm": 0.8299409590937745, "learning_rate": 9.099255620578451e-07, "loss": 0.1212, "step": 752 }, { "epoch": 0.734544567735642, "grad_norm": 0.841743660255562, "learning_rate": 9.037050441037614e-07, "loss": 0.1209, "step": 753 }, { "epoch": 0.7355200585294476, "grad_norm": 0.9030416307912014, "learning_rate": 8.975011674335468e-07, "loss": 0.1271, "step": 754 }, { "epoch": 0.7364955493232532, "grad_norm": 0.8090547787848019, "learning_rate": 8.9131399672226e-07, "loss": 0.1174, "step": 755 }, { "epoch": 0.7374710401170589, "grad_norm": 0.8525347836022709, "learning_rate": 8.851435964708044e-07, "loss": 0.1138, "step": 756 }, { "epoch": 0.7384465309108645, "grad_norm": 0.8419910445268589, "learning_rate": 8.789900310052491e-07, "loss": 0.1215, "step": 757 }, { "epoch": 0.7394220217046702, "grad_norm": 0.8490121207415494, "learning_rate": 8.728533644761647e-07, "loss": 0.1235, "step": 758 }, { "epoch": 0.7403975124984757, "grad_norm": 0.8982804136005714, "learning_rate": 8.667336608579488e-07, "loss": 0.1226, "step": 759 }, { "epoch": 0.7413730032922814, "grad_norm": 0.8467362174197229, "learning_rate": 8.606309839481628e-07, "loss": 0.123, "step": 760 }, { "epoch": 0.742348494086087, "grad_norm": 0.8380377962583168, "learning_rate": 8.545453973668643e-07, "loss": 0.1182, "step": 761 }, { "epoch": 0.7433239848798927, "grad_norm": 0.8630957150176094, "learning_rate": 8.484769645559482e-07, "loss": 0.1195, "step": 762 }, { "epoch": 0.7442994756736984, "grad_norm": 0.8473968774046894, "learning_rate": 8.424257487784787e-07, "loss": 0.1221, "step": 763 }, { "epoch": 0.7452749664675039, "grad_norm": 0.9113778016570447, "learning_rate": 8.363918131180371e-07, "loss": 0.1256, "step": 764 }, { "epoch": 0.7462504572613096, "grad_norm": 0.9276069699799473, "learning_rate": 8.303752204780563e-07, "loss": 0.1248, "step": 765 }, { "epoch": 0.7472259480551152, "grad_norm": 0.8167762715238328, "learning_rate": 8.243760335811734e-07, "loss": 0.1246, "step": 766 }, { "epoch": 0.7482014388489209, "grad_norm": 0.8357245883738734, "learning_rate": 8.183943149685678e-07, "loss": 0.1173, "step": 767 }, { "epoch": 0.7491769296427265, "grad_norm": 0.8026503286373335, "learning_rate": 8.124301269993168e-07, "loss": 0.1151, "step": 768 }, { "epoch": 0.7501524204365322, "grad_norm": 0.9315071378112674, "learning_rate": 8.06483531849738e-07, "loss": 0.1235, "step": 769 }, { "epoch": 0.7511279112303377, "grad_norm": 0.9034958615476297, "learning_rate": 8.00554591512748e-07, "loss": 0.1262, "step": 770 }, { "epoch": 0.7521034020241434, "grad_norm": 0.8669626589106916, "learning_rate": 7.946433677972104e-07, "loss": 0.1243, "step": 771 }, { "epoch": 0.753078892817949, "grad_norm": 0.8864921293794893, "learning_rate": 7.887499223272968e-07, "loss": 0.1298, "step": 772 }, { "epoch": 0.7540543836117547, "grad_norm": 0.8392771141328131, "learning_rate": 7.828743165418393e-07, "loss": 0.111, "step": 773 }, { "epoch": 0.7550298744055604, "grad_norm": 0.821355233876034, "learning_rate": 7.770166116936923e-07, "loss": 0.1139, "step": 774 }, { "epoch": 0.7560053651993659, "grad_norm": 0.8688766276433155, "learning_rate": 7.711768688490962e-07, "loss": 0.1219, "step": 775 }, { "epoch": 0.7569808559931716, "grad_norm": 0.9962941039117293, "learning_rate": 7.653551488870356e-07, "loss": 0.1283, "step": 776 }, { "epoch": 0.7579563467869772, "grad_norm": 0.8627221635187925, "learning_rate": 7.595515124986106e-07, "loss": 0.1237, "step": 777 }, { "epoch": 0.7589318375807829, "grad_norm": 0.8592174071340959, "learning_rate": 7.537660201863989e-07, "loss": 0.1236, "step": 778 }, { "epoch": 0.7599073283745885, "grad_norm": 0.8914444059496847, "learning_rate": 7.479987322638274e-07, "loss": 0.126, "step": 779 }, { "epoch": 0.7608828191683941, "grad_norm": 0.838972249342774, "learning_rate": 7.422497088545436e-07, "loss": 0.1225, "step": 780 }, { "epoch": 0.7618583099621997, "grad_norm": 0.8350154394486518, "learning_rate": 7.365190098917896e-07, "loss": 0.1172, "step": 781 }, { "epoch": 0.7628338007560054, "grad_norm": 0.8527764567297885, "learning_rate": 7.308066951177742e-07, "loss": 0.1194, "step": 782 }, { "epoch": 0.763809291549811, "grad_norm": 0.7880927395631528, "learning_rate": 7.251128240830543e-07, "loss": 0.1122, "step": 783 }, { "epoch": 0.7647847823436167, "grad_norm": 0.8570625979073652, "learning_rate": 7.194374561459094e-07, "loss": 0.1275, "step": 784 }, { "epoch": 0.7657602731374222, "grad_norm": 0.9193780613269857, "learning_rate": 7.137806504717276e-07, "loss": 0.136, "step": 785 }, { "epoch": 0.7667357639312279, "grad_norm": 0.8788105843538093, "learning_rate": 7.081424660323846e-07, "loss": 0.1213, "step": 786 }, { "epoch": 0.7677112547250335, "grad_norm": 0.8629850615115177, "learning_rate": 7.025229616056326e-07, "loss": 0.1223, "step": 787 }, { "epoch": 0.7686867455188392, "grad_norm": 0.8509382827108314, "learning_rate": 6.969221957744832e-07, "loss": 0.1325, "step": 788 }, { "epoch": 0.7696622363126449, "grad_norm": 0.8104302202106697, "learning_rate": 6.913402269266026e-07, "loss": 0.1193, "step": 789 }, { "epoch": 0.7706377271064504, "grad_norm": 0.7876537221330931, "learning_rate": 6.857771132536958e-07, "loss": 0.1177, "step": 790 }, { "epoch": 0.771613217900256, "grad_norm": 0.8452359406579485, "learning_rate": 6.802329127509072e-07, "loss": 0.1236, "step": 791 }, { "epoch": 0.7725887086940617, "grad_norm": 0.9041467869529476, "learning_rate": 6.747076832162094e-07, "loss": 0.1331, "step": 792 }, { "epoch": 0.7735641994878674, "grad_norm": 0.8750663033303995, "learning_rate": 6.692014822498066e-07, "loss": 0.1179, "step": 793 }, { "epoch": 0.774539690281673, "grad_norm": 0.7709557191713534, "learning_rate": 6.637143672535282e-07, "loss": 0.1123, "step": 794 }, { "epoch": 0.7755151810754786, "grad_norm": 0.8489536750773737, "learning_rate": 6.582463954302368e-07, "loss": 0.1189, "step": 795 }, { "epoch": 0.7764906718692842, "grad_norm": 0.7986353115358458, "learning_rate": 6.527976237832256e-07, "loss": 0.1145, "step": 796 }, { "epoch": 0.7774661626630899, "grad_norm": 0.8636955041864243, "learning_rate": 6.473681091156289e-07, "loss": 0.1235, "step": 797 }, { "epoch": 0.7784416534568955, "grad_norm": 0.8485394380202599, "learning_rate": 6.419579080298263e-07, "loss": 0.1218, "step": 798 }, { "epoch": 0.7794171442507012, "grad_norm": 0.8626133292687657, "learning_rate": 6.36567076926857e-07, "loss": 0.1227, "step": 799 }, { "epoch": 0.7803926350445067, "grad_norm": 0.8270229140897313, "learning_rate": 6.311956720058263e-07, "loss": 0.1278, "step": 800 }, { "epoch": 0.7813681258383124, "grad_norm": 0.8821105926200152, "learning_rate": 6.258437492633254e-07, "loss": 0.1251, "step": 801 }, { "epoch": 0.782343616632118, "grad_norm": 0.7683315982295926, "learning_rate": 6.205113644928429e-07, "loss": 0.1139, "step": 802 }, { "epoch": 0.7833191074259237, "grad_norm": 0.8270618740008568, "learning_rate": 6.151985732841867e-07, "loss": 0.1117, "step": 803 }, { "epoch": 0.7842945982197294, "grad_norm": 0.8155271658159645, "learning_rate": 6.099054310229008e-07, "loss": 0.1308, "step": 804 }, { "epoch": 0.7852700890135349, "grad_norm": 0.7877985298756666, "learning_rate": 6.046319928896926e-07, "loss": 0.1157, "step": 805 }, { "epoch": 0.7862455798073406, "grad_norm": 0.8527729286111828, "learning_rate": 5.993783138598532e-07, "loss": 0.1254, "step": 806 }, { "epoch": 0.7872210706011462, "grad_norm": 0.8162522251396951, "learning_rate": 5.941444487026877e-07, "loss": 0.1193, "step": 807 }, { "epoch": 0.7881965613949519, "grad_norm": 0.8198785573921169, "learning_rate": 5.889304519809402e-07, "loss": 0.1225, "step": 808 }, { "epoch": 0.7891720521887575, "grad_norm": 0.8649744547672364, "learning_rate": 5.837363780502309e-07, "loss": 0.1301, "step": 809 }, { "epoch": 0.7901475429825631, "grad_norm": 0.8253170535592403, "learning_rate": 5.785622810584834e-07, "loss": 0.1173, "step": 810 }, { "epoch": 0.7911230337763687, "grad_norm": 0.7949976111986673, "learning_rate": 5.734082149453634e-07, "loss": 0.1134, "step": 811 }, { "epoch": 0.7920985245701744, "grad_norm": 0.8803222417671256, "learning_rate": 5.682742334417171e-07, "loss": 0.125, "step": 812 }, { "epoch": 0.79307401536398, "grad_norm": 0.8262641815619376, "learning_rate": 5.631603900690077e-07, "loss": 0.1214, "step": 813 }, { "epoch": 0.7940495061577857, "grad_norm": 0.849372433500513, "learning_rate": 5.580667381387623e-07, "loss": 0.1244, "step": 814 }, { "epoch": 0.7950249969515912, "grad_norm": 0.7908963105671865, "learning_rate": 5.529933307520102e-07, "loss": 0.1087, "step": 815 }, { "epoch": 0.7960004877453969, "grad_norm": 0.8908242498302418, "learning_rate": 5.47940220798735e-07, "loss": 0.1171, "step": 816 }, { "epoch": 0.7969759785392025, "grad_norm": 0.8393797963498955, "learning_rate": 5.429074609573184e-07, "loss": 0.1226, "step": 817 }, { "epoch": 0.7979514693330082, "grad_norm": 0.8535887142626548, "learning_rate": 5.378951036939966e-07, "loss": 0.1238, "step": 818 }, { "epoch": 0.7989269601268139, "grad_norm": 0.8500240894184807, "learning_rate": 5.329032012623064e-07, "loss": 0.1244, "step": 819 }, { "epoch": 0.7999024509206194, "grad_norm": 0.8632462798593744, "learning_rate": 5.279318057025476e-07, "loss": 0.1131, "step": 820 }, { "epoch": 0.8008779417144251, "grad_norm": 0.8502544027437935, "learning_rate": 5.229809688412337e-07, "loss": 0.125, "step": 821 }, { "epoch": 0.8018534325082307, "grad_norm": 0.8275247834448751, "learning_rate": 5.180507422905585e-07, "loss": 0.1182, "step": 822 }, { "epoch": 0.8028289233020364, "grad_norm": 0.8976360407929419, "learning_rate": 5.131411774478503e-07, "loss": 0.1308, "step": 823 }, { "epoch": 0.803804414095842, "grad_norm": 0.8925880179518361, "learning_rate": 5.082523254950447e-07, "loss": 0.1194, "step": 824 }, { "epoch": 0.8047799048896476, "grad_norm": 0.7964366205604348, "learning_rate": 5.033842373981429e-07, "loss": 0.1157, "step": 825 }, { "epoch": 0.8057553956834532, "grad_norm": 0.8629927042044921, "learning_rate": 4.985369639066867e-07, "loss": 0.1231, "step": 826 }, { "epoch": 0.8067308864772589, "grad_norm": 0.8130144409152532, "learning_rate": 4.937105555532252e-07, "loss": 0.1205, "step": 827 }, { "epoch": 0.8077063772710645, "grad_norm": 0.7968686512185147, "learning_rate": 4.889050626527905e-07, "loss": 0.1222, "step": 828 }, { "epoch": 0.8086818680648702, "grad_norm": 0.8084128367948819, "learning_rate": 4.841205353023715e-07, "loss": 0.1176, "step": 829 }, { "epoch": 0.8096573588586757, "grad_norm": 0.8260965745640515, "learning_rate": 4.79357023380394e-07, "loss": 0.1168, "step": 830 }, { "epoch": 0.8106328496524814, "grad_norm": 0.8136591394426701, "learning_rate": 4.746145765461965e-07, "loss": 0.1246, "step": 831 }, { "epoch": 0.811608340446287, "grad_norm": 0.84880591266123, "learning_rate": 4.6989324423951847e-07, "loss": 0.1207, "step": 832 }, { "epoch": 0.8125838312400927, "grad_norm": 0.8381317559160933, "learning_rate": 4.651930756799794e-07, "loss": 0.1214, "step": 833 }, { "epoch": 0.8135593220338984, "grad_norm": 0.8225053300538132, "learning_rate": 4.6051411986656775e-07, "loss": 0.1222, "step": 834 }, { "epoch": 0.8145348128277039, "grad_norm": 0.7758201411310363, "learning_rate": 4.5585642557713083e-07, "loss": 0.1108, "step": 835 }, { "epoch": 0.8155103036215096, "grad_norm": 0.7841422014242995, "learning_rate": 4.512200413678672e-07, "loss": 0.1149, "step": 836 }, { "epoch": 0.8164857944153152, "grad_norm": 0.8433260125398965, "learning_rate": 4.466050155728166e-07, "loss": 0.1246, "step": 837 }, { "epoch": 0.8174612852091209, "grad_norm": 0.8004993062409861, "learning_rate": 4.420113963033607e-07, "loss": 0.1139, "step": 838 }, { "epoch": 0.8184367760029265, "grad_norm": 0.8544261665582314, "learning_rate": 4.374392314477174e-07, "loss": 0.1442, "step": 839 }, { "epoch": 0.8194122667967321, "grad_norm": 0.8269193033610706, "learning_rate": 4.3288856867044564e-07, "loss": 0.117, "step": 840 }, { "epoch": 0.8203877575905377, "grad_norm": 0.8056654304417562, "learning_rate": 4.2835945541194414e-07, "loss": 0.1106, "step": 841 }, { "epoch": 0.8213632483843434, "grad_norm": 0.8225949604333195, "learning_rate": 4.238519388879614e-07, "loss": 0.118, "step": 842 }, { "epoch": 0.822338739178149, "grad_norm": 0.8057418784933338, "learning_rate": 4.1936606608909887e-07, "loss": 0.1155, "step": 843 }, { "epoch": 0.8233142299719547, "grad_norm": 0.8424817751997794, "learning_rate": 4.149018837803262e-07, "loss": 0.124, "step": 844 }, { "epoch": 0.8242897207657602, "grad_norm": 0.7785385373841229, "learning_rate": 4.1045943850048843e-07, "loss": 0.1229, "step": 845 }, { "epoch": 0.8252652115595659, "grad_norm": 0.8139230651458013, "learning_rate": 4.0603877656182373e-07, "loss": 0.1201, "step": 846 }, { "epoch": 0.8262407023533715, "grad_norm": 0.8136291299903247, "learning_rate": 4.0163994404948144e-07, "loss": 0.1147, "step": 847 }, { "epoch": 0.8272161931471772, "grad_norm": 0.7604644453749246, "learning_rate": 3.9726298682103923e-07, "loss": 0.1136, "step": 848 }, { "epoch": 0.8281916839409829, "grad_norm": 0.8003636956695743, "learning_rate": 3.929079505060271e-07, "loss": 0.1104, "step": 849 }, { "epoch": 0.8291671747347884, "grad_norm": 0.8051074779072083, "learning_rate": 3.8857488050544903e-07, "loss": 0.1196, "step": 850 }, { "epoch": 0.8301426655285941, "grad_norm": 0.8417299086430947, "learning_rate": 3.842638219913142e-07, "loss": 0.1187, "step": 851 }, { "epoch": 0.8311181563223997, "grad_norm": 0.8473564595322601, "learning_rate": 3.799748199061612e-07, "loss": 0.1238, "step": 852 }, { "epoch": 0.8320936471162054, "grad_norm": 0.8530209378089045, "learning_rate": 3.7570791896259147e-07, "loss": 0.1195, "step": 853 }, { "epoch": 0.833069137910011, "grad_norm": 0.8683334198893611, "learning_rate": 3.7146316364280426e-07, "loss": 0.1202, "step": 854 }, { "epoch": 0.8340446287038166, "grad_norm": 0.8185200067165666, "learning_rate": 3.6724059819813223e-07, "loss": 0.1244, "step": 855 }, { "epoch": 0.8350201194976222, "grad_norm": 0.8871001344843794, "learning_rate": 3.630402666485783e-07, "loss": 0.1257, "step": 856 }, { "epoch": 0.8359956102914279, "grad_norm": 0.8335712456953084, "learning_rate": 3.5886221278236045e-07, "loss": 0.117, "step": 857 }, { "epoch": 0.8369711010852335, "grad_norm": 0.8724817716079213, "learning_rate": 3.547064801554509e-07, "loss": 0.1223, "step": 858 }, { "epoch": 0.8379465918790392, "grad_norm": 0.8327646410251778, "learning_rate": 3.5057311209112625e-07, "loss": 0.1273, "step": 859 }, { "epoch": 0.8389220826728447, "grad_norm": 0.8553723242208955, "learning_rate": 3.4646215167951166e-07, "loss": 0.1237, "step": 860 }, { "epoch": 0.8398975734666504, "grad_norm": 0.8667258133361339, "learning_rate": 3.4237364177713615e-07, "loss": 0.1194, "step": 861 }, { "epoch": 0.840873064260456, "grad_norm": 0.775526945884146, "learning_rate": 3.383076250064815e-07, "loss": 0.111, "step": 862 }, { "epoch": 0.8418485550542617, "grad_norm": 0.8350772764316703, "learning_rate": 3.342641437555411e-07, "loss": 0.1227, "step": 863 }, { "epoch": 0.8428240458480674, "grad_norm": 0.8737189518152014, "learning_rate": 3.3024324017737555e-07, "loss": 0.1219, "step": 864 }, { "epoch": 0.8437995366418729, "grad_norm": 0.8525338624818053, "learning_rate": 3.262449561896766e-07, "loss": 0.1208, "step": 865 }, { "epoch": 0.8447750274356786, "grad_norm": 0.7808415133904315, "learning_rate": 3.2226933347432516e-07, "loss": 0.1115, "step": 866 }, { "epoch": 0.8457505182294842, "grad_norm": 0.8212690915534371, "learning_rate": 3.183164134769631e-07, "loss": 0.1192, "step": 867 }, { "epoch": 0.8467260090232899, "grad_norm": 0.8432223045748855, "learning_rate": 3.143862374065548e-07, "loss": 0.1271, "step": 868 }, { "epoch": 0.8477014998170955, "grad_norm": 0.809073364351975, "learning_rate": 3.104788462349612e-07, "loss": 0.1164, "step": 869 }, { "epoch": 0.8486769906109011, "grad_norm": 0.8122295362266347, "learning_rate": 3.065942806965139e-07, "loss": 0.1166, "step": 870 }, { "epoch": 0.8496524814047067, "grad_norm": 0.8628135926038828, "learning_rate": 3.0273258128758585e-07, "loss": 0.1231, "step": 871 }, { "epoch": 0.8506279721985124, "grad_norm": 0.9038737225256949, "learning_rate": 2.98893788266173e-07, "loss": 0.1269, "step": 872 }, { "epoch": 0.851603462992318, "grad_norm": 0.8415109669107367, "learning_rate": 2.9507794165147487e-07, "loss": 0.1185, "step": 873 }, { "epoch": 0.8525789537861237, "grad_norm": 0.8260696094654174, "learning_rate": 2.9128508122347324e-07, "loss": 0.1209, "step": 874 }, { "epoch": 0.8535544445799292, "grad_norm": 0.7950123942247705, "learning_rate": 2.875152465225234e-07, "loss": 0.1107, "step": 875 }, { "epoch": 0.8545299353737349, "grad_norm": 0.8036333920809307, "learning_rate": 2.837684768489354e-07, "loss": 0.1159, "step": 876 }, { "epoch": 0.8555054261675406, "grad_norm": 0.8543568873961171, "learning_rate": 2.800448112625709e-07, "loss": 0.1278, "step": 877 }, { "epoch": 0.8564809169613462, "grad_norm": 0.8705052724557507, "learning_rate": 2.7634428858242995e-07, "loss": 0.1244, "step": 878 }, { "epoch": 0.8574564077551519, "grad_norm": 0.8648954628473264, "learning_rate": 2.7266694738625143e-07, "loss": 0.1134, "step": 879 }, { "epoch": 0.8584318985489574, "grad_norm": 0.8569598558609475, "learning_rate": 2.690128260101069e-07, "loss": 0.1216, "step": 880 }, { "epoch": 0.8594073893427631, "grad_norm": 0.8615706510023747, "learning_rate": 2.6538196254800393e-07, "loss": 0.1282, "step": 881 }, { "epoch": 0.8603828801365687, "grad_norm": 0.8654838887587664, "learning_rate": 2.617743948514867e-07, "loss": 0.1279, "step": 882 }, { "epoch": 0.8613583709303744, "grad_norm": 0.8150117399404916, "learning_rate": 2.5819016052924224e-07, "loss": 0.1186, "step": 883 }, { "epoch": 0.86233386172418, "grad_norm": 0.8042159227280485, "learning_rate": 2.5462929694670986e-07, "loss": 0.1162, "step": 884 }, { "epoch": 0.8633093525179856, "grad_norm": 0.8098200477108216, "learning_rate": 2.5109184122568797e-07, "loss": 0.1079, "step": 885 }, { "epoch": 0.8642848433117912, "grad_norm": 0.8037754437911059, "learning_rate": 2.4757783024395244e-07, "loss": 0.1168, "step": 886 }, { "epoch": 0.8652603341055969, "grad_norm": 0.8174214190622696, "learning_rate": 2.44087300634866e-07, "loss": 0.1143, "step": 887 }, { "epoch": 0.8662358248994025, "grad_norm": 0.9250267238278224, "learning_rate": 2.4062028878700074e-07, "loss": 0.1242, "step": 888 }, { "epoch": 0.8672113156932082, "grad_norm": 0.843085805747338, "learning_rate": 2.3717683084375832e-07, "loss": 0.1179, "step": 889 }, { "epoch": 0.8681868064870137, "grad_norm": 0.8349662619877875, "learning_rate": 2.3375696270299093e-07, "loss": 0.1193, "step": 890 }, { "epoch": 0.8691622972808194, "grad_norm": 0.7931357279276846, "learning_rate": 2.3036072001662829e-07, "loss": 0.1112, "step": 891 }, { "epoch": 0.870137788074625, "grad_norm": 0.8058217864023394, "learning_rate": 2.2698813819030802e-07, "loss": 0.1128, "step": 892 }, { "epoch": 0.8711132788684307, "grad_norm": 0.8105881503397407, "learning_rate": 2.2363925238300167e-07, "loss": 0.1186, "step": 893 }, { "epoch": 0.8720887696622364, "grad_norm": 0.793047905194702, "learning_rate": 2.2031409750665422e-07, "loss": 0.1136, "step": 894 }, { "epoch": 0.8730642604560419, "grad_norm": 0.887128059993288, "learning_rate": 2.170127082258147e-07, "loss": 0.1179, "step": 895 }, { "epoch": 0.8740397512498476, "grad_norm": 0.7729380382439517, "learning_rate": 2.1373511895727866e-07, "loss": 0.1093, "step": 896 }, { "epoch": 0.8750152420436532, "grad_norm": 0.8334653630393726, "learning_rate": 2.1048136386972645e-07, "loss": 0.1201, "step": 897 }, { "epoch": 0.8759907328374589, "grad_norm": 0.8109427974591241, "learning_rate": 2.0725147688337054e-07, "loss": 0.1209, "step": 898 }, { "epoch": 0.8769662236312645, "grad_norm": 0.8075287607010287, "learning_rate": 2.040454916695972e-07, "loss": 0.1104, "step": 899 }, { "epoch": 0.8779417144250701, "grad_norm": 0.7907021456077222, "learning_rate": 2.0086344165062054e-07, "loss": 0.1186, "step": 900 }, { "epoch": 0.8789172052188757, "grad_norm": 0.8321531164794602, "learning_rate": 1.9770535999912967e-07, "loss": 0.1242, "step": 901 }, { "epoch": 0.8798926960126814, "grad_norm": 0.836504402336581, "learning_rate": 1.945712796379462e-07, "loss": 0.1195, "step": 902 }, { "epoch": 0.880868186806487, "grad_norm": 0.8414921499203574, "learning_rate": 1.914612332396787e-07, "loss": 0.1191, "step": 903 }, { "epoch": 0.8818436776002927, "grad_norm": 0.8268050582099995, "learning_rate": 1.883752532263844e-07, "loss": 0.1212, "step": 904 }, { "epoch": 0.8828191683940982, "grad_norm": 0.7804225339880899, "learning_rate": 1.8531337176922792e-07, "loss": 0.109, "step": 905 }, { "epoch": 0.8837946591879039, "grad_norm": 0.8089563734380512, "learning_rate": 1.8227562078814903e-07, "loss": 0.1185, "step": 906 }, { "epoch": 0.8847701499817096, "grad_norm": 0.8524343318848043, "learning_rate": 1.7926203195152898e-07, "loss": 0.1236, "step": 907 }, { "epoch": 0.8857456407755152, "grad_norm": 0.8146203036325108, "learning_rate": 1.7627263667585882e-07, "loss": 0.1118, "step": 908 }, { "epoch": 0.8867211315693209, "grad_norm": 0.8438170384778132, "learning_rate": 1.7330746612541387e-07, "loss": 0.1153, "step": 909 }, { "epoch": 0.8876966223631264, "grad_norm": 0.816511650231362, "learning_rate": 1.7036655121192875e-07, "loss": 0.1262, "step": 910 }, { "epoch": 0.8886721131569321, "grad_norm": 0.8384874737504223, "learning_rate": 1.6744992259427272e-07, "loss": 0.1241, "step": 911 }, { "epoch": 0.8896476039507377, "grad_norm": 0.8111140618017998, "learning_rate": 1.6455761067813352e-07, "loss": 0.1137, "step": 912 }, { "epoch": 0.8906230947445434, "grad_norm": 0.8137411483354814, "learning_rate": 1.6168964561569716e-07, "loss": 0.1177, "step": 913 }, { "epoch": 0.891598585538349, "grad_norm": 0.7826271884075486, "learning_rate": 1.5884605730533686e-07, "loss": 0.1173, "step": 914 }, { "epoch": 0.8925740763321546, "grad_norm": 0.8815949613672592, "learning_rate": 1.5602687539129745e-07, "loss": 0.1224, "step": 915 }, { "epoch": 0.8935495671259602, "grad_norm": 0.8365843017681764, "learning_rate": 1.5323212926339037e-07, "loss": 0.119, "step": 916 }, { "epoch": 0.8945250579197659, "grad_norm": 0.7838035275941735, "learning_rate": 1.504618480566844e-07, "loss": 0.1137, "step": 917 }, { "epoch": 0.8955005487135715, "grad_norm": 0.8571652908859096, "learning_rate": 1.4771606065120293e-07, "loss": 0.1176, "step": 918 }, { "epoch": 0.8964760395073772, "grad_norm": 0.8042632494131905, "learning_rate": 1.4499479567162328e-07, "loss": 0.1193, "step": 919 }, { "epoch": 0.8974515303011827, "grad_norm": 0.8620451657826478, "learning_rate": 1.4229808148697732e-07, "loss": 0.1245, "step": 920 }, { "epoch": 0.8984270210949884, "grad_norm": 0.8376113680024069, "learning_rate": 1.396259462103572e-07, "loss": 0.1101, "step": 921 }, { "epoch": 0.899402511888794, "grad_norm": 0.8311869479243955, "learning_rate": 1.3697841769861996e-07, "loss": 0.1198, "step": 922 }, { "epoch": 0.9003780026825997, "grad_norm": 0.869133372824549, "learning_rate": 1.3435552355210018e-07, "loss": 0.1172, "step": 923 }, { "epoch": 0.9013534934764054, "grad_norm": 0.8549023606786766, "learning_rate": 1.3175729111431946e-07, "loss": 0.1248, "step": 924 }, { "epoch": 0.9023289842702109, "grad_norm": 0.8266104807259295, "learning_rate": 1.2918374747170225e-07, "loss": 0.1157, "step": 925 }, { "epoch": 0.9033044750640166, "grad_norm": 0.8725974407693242, "learning_rate": 1.266349194532951e-07, "loss": 0.1212, "step": 926 }, { "epoch": 0.9042799658578222, "grad_norm": 0.8869251827136297, "learning_rate": 1.2411083363048386e-07, "loss": 0.1168, "step": 927 }, { "epoch": 0.9052554566516279, "grad_norm": 0.838978233063295, "learning_rate": 1.2161151631671974e-07, "loss": 0.1193, "step": 928 }, { "epoch": 0.9062309474454335, "grad_norm": 0.8158933020010286, "learning_rate": 1.1913699356724317e-07, "loss": 0.1139, "step": 929 }, { "epoch": 0.9072064382392391, "grad_norm": 0.8302262653149984, "learning_rate": 1.166872911788125e-07, "loss": 0.1193, "step": 930 }, { "epoch": 0.9081819290330447, "grad_norm": 0.8281059144615461, "learning_rate": 1.1426243468943582e-07, "loss": 0.1218, "step": 931 }, { "epoch": 0.9091574198268504, "grad_norm": 0.8150387396832428, "learning_rate": 1.1186244937810315e-07, "loss": 0.1133, "step": 932 }, { "epoch": 0.910132910620656, "grad_norm": 0.8174814241523217, "learning_rate": 1.0948736026452495e-07, "loss": 0.1111, "step": 933 }, { "epoch": 0.9111084014144617, "grad_norm": 0.8072685450600614, "learning_rate": 1.071371921088693e-07, "loss": 0.1219, "step": 934 }, { "epoch": 0.9120838922082672, "grad_norm": 0.8462154845847925, "learning_rate": 1.0481196941150574e-07, "loss": 0.1131, "step": 935 }, { "epoch": 0.9130593830020729, "grad_norm": 0.8194674338836323, "learning_rate": 1.0251171641274788e-07, "loss": 0.1208, "step": 936 }, { "epoch": 0.9140348737958786, "grad_norm": 0.8051460380945417, "learning_rate": 1.0023645709260233e-07, "loss": 0.111, "step": 937 }, { "epoch": 0.9150103645896842, "grad_norm": 0.7953737847156309, "learning_rate": 9.798621517051716e-08, "loss": 0.1134, "step": 938 }, { "epoch": 0.9159858553834899, "grad_norm": 0.8137082279917823, "learning_rate": 9.576101410513655e-08, "loss": 0.1184, "step": 939 }, { "epoch": 0.9169613461772954, "grad_norm": 0.81376889042028, "learning_rate": 9.356087709405465e-08, "loss": 0.1141, "step": 940 }, { "epoch": 0.9179368369711011, "grad_norm": 0.8493873934604731, "learning_rate": 9.138582707357429e-08, "loss": 0.1178, "step": 941 }, { "epoch": 0.9189123277649067, "grad_norm": 0.8480580946913866, "learning_rate": 8.923588671846784e-08, "loss": 0.119, "step": 942 }, { "epoch": 0.9198878185587124, "grad_norm": 0.8259767599447636, "learning_rate": 8.711107844174089e-08, "loss": 0.1224, "step": 943 }, { "epoch": 0.920863309352518, "grad_norm": 0.8414744578683574, "learning_rate": 8.501142439439808e-08, "loss": 0.1184, "step": 944 }, { "epoch": 0.9218388001463236, "grad_norm": 0.7500528197265454, "learning_rate": 8.29369464652141e-08, "loss": 0.0985, "step": 945 }, { "epoch": 0.9228142909401292, "grad_norm": 0.8122867841326354, "learning_rate": 8.088766628050193e-08, "loss": 0.1102, "step": 946 }, { "epoch": 0.9237897817339349, "grad_norm": 0.8378630528612887, "learning_rate": 7.886360520389158e-08, "loss": 0.1187, "step": 947 }, { "epoch": 0.9247652725277405, "grad_norm": 0.7993318052605702, "learning_rate": 7.686478433610339e-08, "loss": 0.113, "step": 948 }, { "epoch": 0.9257407633215462, "grad_norm": 0.8282335985098225, "learning_rate": 7.48912245147318e-08, "loss": 0.1183, "step": 949 }, { "epoch": 0.9267162541153517, "grad_norm": 0.8238348823155386, "learning_rate": 7.294294631402493e-08, "loss": 0.1202, "step": 950 }, { "epoch": 0.9276917449091574, "grad_norm": 0.8024852399758585, "learning_rate": 7.101997004467203e-08, "loss": 0.122, "step": 951 }, { "epoch": 0.9286672357029631, "grad_norm": 0.7847553156886314, "learning_rate": 6.912231575359057e-08, "loss": 0.1104, "step": 952 }, { "epoch": 0.9296427264967687, "grad_norm": 0.8975888719583766, "learning_rate": 6.725000322371916e-08, "loss": 0.1186, "step": 953 }, { "epoch": 0.9306182172905744, "grad_norm": 0.8212833223497472, "learning_rate": 6.540305197380859e-08, "loss": 0.122, "step": 954 }, { "epoch": 0.9315937080843799, "grad_norm": 0.7982479941876005, "learning_rate": 6.358148125822e-08, "loss": 0.1136, "step": 955 }, { "epoch": 0.9325691988781856, "grad_norm": 0.8592832767848003, "learning_rate": 6.178531006672484e-08, "loss": 0.1276, "step": 956 }, { "epoch": 0.9335446896719912, "grad_norm": 0.8873811445927061, "learning_rate": 6.001455712430492e-08, "loss": 0.126, "step": 957 }, { "epoch": 0.9345201804657969, "grad_norm": 0.8556443661624066, "learning_rate": 5.826924089095881e-08, "loss": 0.1176, "step": 958 }, { "epoch": 0.9354956712596025, "grad_norm": 0.7924528652097824, "learning_rate": 5.654937956150852e-08, "loss": 0.1208, "step": 959 }, { "epoch": 0.9364711620534081, "grad_norm": 0.8207355407913433, "learning_rate": 5.4854991065410866e-08, "loss": 0.1236, "step": 960 }, { "epoch": 0.9374466528472137, "grad_norm": 0.823166152063427, "learning_rate": 5.3186093066568965e-08, "loss": 0.1195, "step": 961 }, { "epoch": 0.9384221436410194, "grad_norm": 0.9014815011677116, "learning_rate": 5.154270296314878e-08, "loss": 0.1317, "step": 962 }, { "epoch": 0.939397634434825, "grad_norm": 0.8362981567362682, "learning_rate": 4.992483788739927e-08, "loss": 0.1208, "step": 963 }, { "epoch": 0.9403731252286307, "grad_norm": 0.840511469516904, "learning_rate": 4.833251470547084e-08, "loss": 0.1293, "step": 964 }, { "epoch": 0.9413486160224362, "grad_norm": 0.831812913196967, "learning_rate": 4.6765750017242206e-08, "loss": 0.1217, "step": 965 }, { "epoch": 0.9423241068162419, "grad_norm": 0.8367591325240412, "learning_rate": 4.5224560156145734e-08, "loss": 0.118, "step": 966 }, { "epoch": 0.9432995976100476, "grad_norm": 0.879426458908269, "learning_rate": 4.370896118899792e-08, "loss": 0.1145, "step": 967 }, { "epoch": 0.9442750884038532, "grad_norm": 0.8098437070862734, "learning_rate": 4.221896891583144e-08, "loss": 0.1144, "step": 968 }, { "epoch": 0.9452505791976589, "grad_norm": 0.7667200836569177, "learning_rate": 4.0754598869730824e-08, "loss": 0.11, "step": 969 }, { "epoch": 0.9462260699914644, "grad_norm": 0.8556949255820121, "learning_rate": 3.931586631667039e-08, "loss": 0.1246, "step": 970 }, { "epoch": 0.9472015607852701, "grad_norm": 0.8278433518221362, "learning_rate": 3.7902786255354927e-08, "loss": 0.1244, "step": 971 }, { "epoch": 0.9481770515790757, "grad_norm": 0.8279334840526833, "learning_rate": 3.65153734170634e-08, "loss": 0.1167, "step": 972 }, { "epoch": 0.9491525423728814, "grad_norm": 0.8237905685355518, "learning_rate": 3.515364226549523e-08, "loss": 0.1205, "step": 973 }, { "epoch": 0.950128033166687, "grad_norm": 0.8493136167563948, "learning_rate": 3.381760699662062e-08, "loss": 0.1223, "step": 974 }, { "epoch": 0.9511035239604926, "grad_norm": 0.8361911343136549, "learning_rate": 3.25072815385305e-08, "loss": 0.1161, "step": 975 }, { "epoch": 0.9520790147542982, "grad_norm": 0.80954482668594, "learning_rate": 3.1222679551293486e-08, "loss": 0.1129, "step": 976 }, { "epoch": 0.9530545055481039, "grad_norm": 0.82063025704283, "learning_rate": 2.996381442681162e-08, "loss": 0.1191, "step": 977 }, { "epoch": 0.9540299963419095, "grad_norm": 0.828098268541136, "learning_rate": 2.8730699288682107e-08, "loss": 0.1262, "step": 978 }, { "epoch": 0.9550054871357152, "grad_norm": 0.8093239279254736, "learning_rate": 2.7523346992060217e-08, "loss": 0.1159, "step": 979 }, { "epoch": 0.9559809779295207, "grad_norm": 0.8448131471914473, "learning_rate": 2.634177012352468e-08, "loss": 0.1181, "step": 980 }, { "epoch": 0.9569564687233264, "grad_norm": 0.850456037528909, "learning_rate": 2.5185981000946657e-08, "loss": 0.1232, "step": 981 }, { "epoch": 0.9579319595171321, "grad_norm": 0.7998292649044384, "learning_rate": 2.405599167336209e-08, "loss": 0.1152, "step": 982 }, { "epoch": 0.9589074503109377, "grad_norm": 0.8711285830989083, "learning_rate": 2.295181392084511e-08, "loss": 0.1192, "step": 983 }, { "epoch": 0.9598829411047434, "grad_norm": 0.831799017458, "learning_rate": 2.187345925438594e-08, "loss": 0.1143, "step": 984 }, { "epoch": 0.9608584318985489, "grad_norm": 0.8032074365365065, "learning_rate": 2.0820938915770417e-08, "loss": 0.1105, "step": 985 }, { "epoch": 0.9618339226923546, "grad_norm": 0.7978427999769494, "learning_rate": 1.9794263877463138e-08, "loss": 0.1061, "step": 986 }, { "epoch": 0.9628094134861602, "grad_norm": 0.8519627713839724, "learning_rate": 1.879344484249257e-08, "loss": 0.1229, "step": 987 }, { "epoch": 0.9637849042799659, "grad_norm": 0.8131083664763187, "learning_rate": 1.781849224434029e-08, "loss": 0.1178, "step": 988 }, { "epoch": 0.9647603950737715, "grad_norm": 0.8711093922420086, "learning_rate": 1.686941624683164e-08, "loss": 0.1268, "step": 989 }, { "epoch": 0.9657358858675771, "grad_norm": 0.8153485749863514, "learning_rate": 1.5946226744029402e-08, "loss": 0.1139, "step": 990 }, { "epoch": 0.9667113766613827, "grad_norm": 0.8470338609149345, "learning_rate": 1.504893336013169e-08, "loss": 0.117, "step": 991 }, { "epoch": 0.9676868674551884, "grad_norm": 0.8952280575519709, "learning_rate": 1.4177545449370889e-08, "loss": 0.1193, "step": 992 }, { "epoch": 0.968662358248994, "grad_norm": 0.8393629890478395, "learning_rate": 1.3332072095916527e-08, "loss": 0.1148, "step": 993 }, { "epoch": 0.9696378490427997, "grad_norm": 0.8484999365301013, "learning_rate": 1.2512522113779235e-08, "loss": 0.1212, "step": 994 }, { "epoch": 0.9706133398366052, "grad_norm": 0.8047573717380992, "learning_rate": 1.1718904046721657e-08, "loss": 0.1169, "step": 995 }, { "epoch": 0.9715888306304109, "grad_norm": 0.7801955720110736, "learning_rate": 1.0951226168166574e-08, "loss": 0.1158, "step": 996 }, { "epoch": 0.9725643214242166, "grad_norm": 0.8469472102608405, "learning_rate": 1.0209496481112247e-08, "loss": 0.119, "step": 997 }, { "epoch": 0.9735398122180222, "grad_norm": 0.7899772186862825, "learning_rate": 9.493722718048326e-09, "loss": 0.108, "step": 998 }, { "epoch": 0.9745153030118279, "grad_norm": 0.8122633623112819, "learning_rate": 8.803912340875076e-09, "loss": 0.1149, "step": 999 }, { "epoch": 0.9754907938056334, "grad_norm": 0.861388140402043, "learning_rate": 8.140072540826772e-09, "loss": 0.1229, "step": 1000 }, { "epoch": 0.9754907938056334, "eval_loss": 0.11711173504590988, "eval_runtime": 127.3978, "eval_samples_per_second": 5.204, "eval_steps_per_second": 0.652, "step": 1000 }, { "epoch": 0.9764662845994391, "grad_norm": 0.795759380098638, "learning_rate": 7.502210238395091e-09, "loss": 0.1202, "step": 1001 }, { "epoch": 0.9774417753932447, "grad_norm": 0.7911131054538534, "learning_rate": 6.890332083258622e-09, "loss": 0.1206, "step": 1002 }, { "epoch": 0.9784172661870504, "grad_norm": 0.8225435058710419, "learning_rate": 6.304444454212077e-09, "loss": 0.1253, "step": 1003 }, { "epoch": 0.979392756980856, "grad_norm": 0.7897526541638892, "learning_rate": 5.7445534591002435e-09, "loss": 0.115, "step": 1004 }, { "epoch": 0.9803682477746616, "grad_norm": 0.8241723887349869, "learning_rate": 5.210664934754972e-09, "loss": 0.1248, "step": 1005 }, { "epoch": 0.9813437385684672, "grad_norm": 0.8404643550865318, "learning_rate": 4.702784446934116e-09, "loss": 0.1185, "step": 1006 }, { "epoch": 0.9823192293622729, "grad_norm": 0.8329025507816016, "learning_rate": 4.220917290262139e-09, "loss": 0.1221, "step": 1007 }, { "epoch": 0.9832947201560785, "grad_norm": 0.8272251709257826, "learning_rate": 3.765068488177093e-09, "loss": 0.1243, "step": 1008 }, { "epoch": 0.9842702109498842, "grad_norm": 0.8289663001334425, "learning_rate": 3.335242792876503e-09, "loss": 0.1121, "step": 1009 }, { "epoch": 0.9852457017436898, "grad_norm": 0.8935773554975892, "learning_rate": 2.931444685269069e-09, "loss": 0.1296, "step": 1010 }, { "epoch": 0.9862211925374954, "grad_norm": 0.8147353591204097, "learning_rate": 2.553678374926649e-09, "loss": 0.1197, "step": 1011 }, { "epoch": 0.9871966833313011, "grad_norm": 0.8356734189663587, "learning_rate": 2.201947800041515e-09, "loss": 0.1233, "step": 1012 }, { "epoch": 0.9881721741251067, "grad_norm": 0.8277826045315564, "learning_rate": 1.8762566273852754e-09, "loss": 0.1102, "step": 1013 }, { "epoch": 0.9891476649189124, "grad_norm": 0.837735922536016, "learning_rate": 1.576608252269185e-09, "loss": 0.1282, "step": 1014 }, { "epoch": 0.9901231557127179, "grad_norm": 0.8185965403465947, "learning_rate": 1.3030057985108368e-09, "loss": 0.1184, "step": 1015 }, { "epoch": 0.9910986465065236, "grad_norm": 0.7801036446264883, "learning_rate": 1.055452118400302e-09, "loss": 0.111, "step": 1016 }, { "epoch": 0.9920741373003292, "grad_norm": 0.7975320236568699, "learning_rate": 8.339497926704299e-10, "loss": 0.114, "step": 1017 }, { "epoch": 0.9930496280941349, "grad_norm": 0.8525955644349905, "learning_rate": 6.385011304704814e-10, "loss": 0.1257, "step": 1018 }, { "epoch": 0.9940251188879405, "grad_norm": 0.8091050888964465, "learning_rate": 4.691081693411481e-10, "loss": 0.1147, "step": 1019 }, { "epoch": 0.9950006096817461, "grad_norm": 0.9205192964480091, "learning_rate": 3.2577267519484645e-10, "loss": 0.1257, "step": 1020 }, { "epoch": 0.9959761004755517, "grad_norm": 0.816877602174351, "learning_rate": 2.0849614229601078e-10, "loss": 0.1145, "step": 1021 }, { "epoch": 0.9969515912693574, "grad_norm": 0.8091963243679661, "learning_rate": 1.1727979324527294e-10, "loss": 0.1185, "step": 1022 }, { "epoch": 0.997927082063163, "grad_norm": 0.8922096606338227, "learning_rate": 5.212457896835998e-11, "loss": 0.118, "step": 1023 }, { "epoch": 0.9989025728569687, "grad_norm": 0.7963658344295353, "learning_rate": 1.303117870443682e-11, "loss": 0.118, "step": 1024 }, { "epoch": 0.9998780636507743, "grad_norm": 0.832964772750828, "learning_rate": 0.0, "loss": 0.1208, "step": 1025 }, { "epoch": 0.9998780636507743, "step": 1025, "total_flos": 241738000957440.0, "train_loss": 0.1467944400630346, "train_runtime": 47673.2991, "train_samples_per_second": 1.376, "train_steps_per_second": 0.022 } ], "logging_steps": 1, "max_steps": 1025, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 241738000957440.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }