{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3069132202869639, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 8.53342435652647e-05, "grad_norm": 0.82421875, "learning_rate": 1.0000000000000002e-06, "loss": 1.8, "step": 1 }, { "epoch": 0.0001706684871305294, "grad_norm": 0.9140625, "learning_rate": 2.0000000000000003e-06, "loss": 1.8859, "step": 2 }, { "epoch": 0.0002560027306957941, "grad_norm": 0.90625, "learning_rate": 3e-06, "loss": 1.7003, "step": 3 }, { "epoch": 0.0003413369742610588, "grad_norm": 0.76171875, "learning_rate": 4.000000000000001e-06, "loss": 1.6892, "step": 4 }, { "epoch": 0.0004266712178263235, "grad_norm": 0.77734375, "learning_rate": 5e-06, "loss": 1.6493, "step": 5 }, { "epoch": 0.0005120054613915882, "grad_norm": 0.80859375, "learning_rate": 6e-06, "loss": 1.7291, "step": 6 }, { "epoch": 0.0005973397049568529, "grad_norm": 0.8203125, "learning_rate": 7.000000000000001e-06, "loss": 1.7985, "step": 7 }, { "epoch": 0.0006826739485221176, "grad_norm": 0.875, "learning_rate": 8.000000000000001e-06, "loss": 1.6985, "step": 8 }, { "epoch": 0.0007680081920873823, "grad_norm": 0.8671875, "learning_rate": 9e-06, "loss": 1.9519, "step": 9 }, { "epoch": 0.000853342435652647, "grad_norm": 0.83984375, "learning_rate": 1e-05, "loss": 1.8502, "step": 10 }, { "epoch": 0.0009386766792179117, "grad_norm": 0.921875, "learning_rate": 1.1000000000000001e-05, "loss": 1.6864, "step": 11 }, { "epoch": 0.0010240109227831764, "grad_norm": 0.90625, "learning_rate": 1.2e-05, "loss": 2.0213, "step": 12 }, { "epoch": 0.001109345166348441, "grad_norm": 0.8046875, "learning_rate": 1.3000000000000001e-05, "loss": 1.7538, "step": 13 }, { "epoch": 0.0011946794099137058, "grad_norm": 0.921875, "learning_rate": 1.4000000000000001e-05, "loss": 1.8914, "step": 14 }, { "epoch": 0.0012800136534789704, "grad_norm": 0.890625, "learning_rate": 1.5e-05, "loss": 1.7576, "step": 15 }, { "epoch": 0.0013653478970442352, "grad_norm": 0.84375, "learning_rate": 1.6000000000000003e-05, "loss": 1.6817, "step": 16 }, { "epoch": 0.0014506821406094998, "grad_norm": 0.91015625, "learning_rate": 1.7000000000000003e-05, "loss": 1.9613, "step": 17 }, { "epoch": 0.0015360163841747646, "grad_norm": 0.86328125, "learning_rate": 1.8e-05, "loss": 1.692, "step": 18 }, { "epoch": 0.0016213506277400291, "grad_norm": 0.7578125, "learning_rate": 1.9e-05, "loss": 1.5385, "step": 19 }, { "epoch": 0.001706684871305294, "grad_norm": 0.8515625, "learning_rate": 2e-05, "loss": 1.8615, "step": 20 }, { "epoch": 0.0017920191148705585, "grad_norm": 0.83984375, "learning_rate": 2.1e-05, "loss": 1.7127, "step": 21 }, { "epoch": 0.0018773533584358233, "grad_norm": 0.80859375, "learning_rate": 2.2000000000000003e-05, "loss": 1.8537, "step": 22 }, { "epoch": 0.001962687602001088, "grad_norm": 0.80078125, "learning_rate": 2.3000000000000003e-05, "loss": 2.0063, "step": 23 }, { "epoch": 0.0020480218455663527, "grad_norm": 0.79296875, "learning_rate": 2.4e-05, "loss": 1.636, "step": 24 }, { "epoch": 0.0021333560891316173, "grad_norm": 0.98046875, "learning_rate": 2.5e-05, "loss": 1.7672, "step": 25 }, { "epoch": 0.002218690332696882, "grad_norm": 0.81640625, "learning_rate": 2.6000000000000002e-05, "loss": 2.0052, "step": 26 }, { "epoch": 0.002304024576262147, "grad_norm": 0.9453125, "learning_rate": 2.7000000000000002e-05, "loss": 1.9785, "step": 27 }, { "epoch": 0.0023893588198274115, "grad_norm": 0.85546875, "learning_rate": 2.8000000000000003e-05, "loss": 1.7909, "step": 28 }, { "epoch": 0.002474693063392676, "grad_norm": 0.765625, "learning_rate": 2.9e-05, "loss": 1.8871, "step": 29 }, { "epoch": 0.0025600273069579407, "grad_norm": 0.8203125, "learning_rate": 3e-05, "loss": 1.707, "step": 30 }, { "epoch": 0.0026453615505232057, "grad_norm": 0.796875, "learning_rate": 3.1e-05, "loss": 1.9111, "step": 31 }, { "epoch": 0.0027306957940884703, "grad_norm": 0.73828125, "learning_rate": 3.2000000000000005e-05, "loss": 1.6338, "step": 32 }, { "epoch": 0.002816030037653735, "grad_norm": 0.80078125, "learning_rate": 3.3e-05, "loss": 1.8687, "step": 33 }, { "epoch": 0.0029013642812189995, "grad_norm": 0.72265625, "learning_rate": 3.4000000000000007e-05, "loss": 1.7552, "step": 34 }, { "epoch": 0.0029866985247842645, "grad_norm": 0.71875, "learning_rate": 3.5e-05, "loss": 1.6459, "step": 35 }, { "epoch": 0.003072032768349529, "grad_norm": 0.89453125, "learning_rate": 3.6e-05, "loss": 1.8945, "step": 36 }, { "epoch": 0.0031573670119147937, "grad_norm": 0.86328125, "learning_rate": 3.7e-05, "loss": 1.8111, "step": 37 }, { "epoch": 0.0032427012554800583, "grad_norm": 0.84765625, "learning_rate": 3.8e-05, "loss": 1.8449, "step": 38 }, { "epoch": 0.0033280354990453233, "grad_norm": 0.73828125, "learning_rate": 3.9000000000000006e-05, "loss": 1.7649, "step": 39 }, { "epoch": 0.003413369742610588, "grad_norm": 0.82421875, "learning_rate": 4e-05, "loss": 1.8078, "step": 40 }, { "epoch": 0.0034987039861758525, "grad_norm": 0.74609375, "learning_rate": 4.1e-05, "loss": 1.6198, "step": 41 }, { "epoch": 0.003584038229741117, "grad_norm": 0.90625, "learning_rate": 4.2e-05, "loss": 1.7853, "step": 42 }, { "epoch": 0.003669372473306382, "grad_norm": 0.828125, "learning_rate": 4.3e-05, "loss": 1.8162, "step": 43 }, { "epoch": 0.0037547067168716467, "grad_norm": 0.7265625, "learning_rate": 4.4000000000000006e-05, "loss": 1.8106, "step": 44 }, { "epoch": 0.0038400409604369113, "grad_norm": 0.69921875, "learning_rate": 4.5e-05, "loss": 1.6102, "step": 45 }, { "epoch": 0.003925375204002176, "grad_norm": 0.6875, "learning_rate": 4.600000000000001e-05, "loss": 1.6463, "step": 46 }, { "epoch": 0.004010709447567441, "grad_norm": 0.66015625, "learning_rate": 4.7e-05, "loss": 1.546, "step": 47 }, { "epoch": 0.0040960436911327055, "grad_norm": 0.703125, "learning_rate": 4.8e-05, "loss": 1.7966, "step": 48 }, { "epoch": 0.00418137793469797, "grad_norm": 0.796875, "learning_rate": 4.9e-05, "loss": 1.9185, "step": 49 }, { "epoch": 0.004266712178263235, "grad_norm": 0.70703125, "learning_rate": 5e-05, "loss": 1.6677, "step": 50 }, { "epoch": 0.004352046421828499, "grad_norm": 0.765625, "learning_rate": 5.1000000000000006e-05, "loss": 1.7566, "step": 51 }, { "epoch": 0.004437380665393764, "grad_norm": 0.7578125, "learning_rate": 5.2000000000000004e-05, "loss": 1.6925, "step": 52 }, { "epoch": 0.004522714908959029, "grad_norm": 0.8515625, "learning_rate": 5.300000000000001e-05, "loss": 1.8328, "step": 53 }, { "epoch": 0.004608049152524294, "grad_norm": 0.73046875, "learning_rate": 5.4000000000000005e-05, "loss": 1.906, "step": 54 }, { "epoch": 0.0046933833960895585, "grad_norm": 0.68359375, "learning_rate": 5.500000000000001e-05, "loss": 1.6741, "step": 55 }, { "epoch": 0.004778717639654823, "grad_norm": 0.67578125, "learning_rate": 5.6000000000000006e-05, "loss": 1.8366, "step": 56 }, { "epoch": 0.004864051883220088, "grad_norm": 0.69921875, "learning_rate": 5.6999999999999996e-05, "loss": 1.733, "step": 57 }, { "epoch": 0.004949386126785352, "grad_norm": 0.6171875, "learning_rate": 5.8e-05, "loss": 1.8005, "step": 58 }, { "epoch": 0.005034720370350617, "grad_norm": 0.734375, "learning_rate": 5.9e-05, "loss": 1.7329, "step": 59 }, { "epoch": 0.005120054613915881, "grad_norm": 0.69140625, "learning_rate": 6e-05, "loss": 1.7772, "step": 60 }, { "epoch": 0.005205388857481147, "grad_norm": 0.671875, "learning_rate": 6.1e-05, "loss": 1.6311, "step": 61 }, { "epoch": 0.0052907231010464115, "grad_norm": 0.59765625, "learning_rate": 6.2e-05, "loss": 1.506, "step": 62 }, { "epoch": 0.005376057344611676, "grad_norm": 0.6796875, "learning_rate": 6.3e-05, "loss": 1.6361, "step": 63 }, { "epoch": 0.005461391588176941, "grad_norm": 0.65234375, "learning_rate": 6.400000000000001e-05, "loss": 1.6185, "step": 64 }, { "epoch": 0.005546725831742205, "grad_norm": 0.6328125, "learning_rate": 6.500000000000001e-05, "loss": 1.7215, "step": 65 }, { "epoch": 0.00563206007530747, "grad_norm": 0.671875, "learning_rate": 6.6e-05, "loss": 1.8679, "step": 66 }, { "epoch": 0.005717394318872734, "grad_norm": 0.5859375, "learning_rate": 6.7e-05, "loss": 1.7063, "step": 67 }, { "epoch": 0.005802728562437999, "grad_norm": 0.58984375, "learning_rate": 6.800000000000001e-05, "loss": 1.6245, "step": 68 }, { "epoch": 0.0058880628060032645, "grad_norm": 0.578125, "learning_rate": 6.9e-05, "loss": 1.8109, "step": 69 }, { "epoch": 0.005973397049568529, "grad_norm": 0.67578125, "learning_rate": 7e-05, "loss": 1.9045, "step": 70 }, { "epoch": 0.006058731293133794, "grad_norm": 0.70703125, "learning_rate": 7.1e-05, "loss": 1.6802, "step": 71 }, { "epoch": 0.006144065536699058, "grad_norm": 0.66015625, "learning_rate": 7.2e-05, "loss": 1.757, "step": 72 }, { "epoch": 0.006229399780264323, "grad_norm": 0.6171875, "learning_rate": 7.3e-05, "loss": 1.7513, "step": 73 }, { "epoch": 0.006314734023829587, "grad_norm": 0.60546875, "learning_rate": 7.4e-05, "loss": 1.8557, "step": 74 }, { "epoch": 0.006400068267394852, "grad_norm": 0.5625, "learning_rate": 7.500000000000001e-05, "loss": 1.6747, "step": 75 }, { "epoch": 0.006485402510960117, "grad_norm": 0.6015625, "learning_rate": 7.6e-05, "loss": 1.6629, "step": 76 }, { "epoch": 0.006570736754525381, "grad_norm": 0.78515625, "learning_rate": 7.7e-05, "loss": 1.8731, "step": 77 }, { "epoch": 0.006656070998090647, "grad_norm": 0.59765625, "learning_rate": 7.800000000000001e-05, "loss": 1.6409, "step": 78 }, { "epoch": 0.006741405241655911, "grad_norm": 0.63671875, "learning_rate": 7.900000000000001e-05, "loss": 1.6955, "step": 79 }, { "epoch": 0.006826739485221176, "grad_norm": 0.58203125, "learning_rate": 8e-05, "loss": 1.6168, "step": 80 }, { "epoch": 0.00691207372878644, "grad_norm": 0.60546875, "learning_rate": 8.1e-05, "loss": 1.8731, "step": 81 }, { "epoch": 0.006997407972351705, "grad_norm": 0.5625, "learning_rate": 8.2e-05, "loss": 1.8115, "step": 82 }, { "epoch": 0.00708274221591697, "grad_norm": 0.6328125, "learning_rate": 8.3e-05, "loss": 1.8266, "step": 83 }, { "epoch": 0.007168076459482234, "grad_norm": 0.5390625, "learning_rate": 8.4e-05, "loss": 1.6428, "step": 84 }, { "epoch": 0.007253410703047499, "grad_norm": 0.53515625, "learning_rate": 8.5e-05, "loss": 1.6117, "step": 85 }, { "epoch": 0.007338744946612764, "grad_norm": 0.5, "learning_rate": 8.6e-05, "loss": 1.5626, "step": 86 }, { "epoch": 0.007424079190178029, "grad_norm": 0.51953125, "learning_rate": 8.7e-05, "loss": 1.686, "step": 87 }, { "epoch": 0.007509413433743293, "grad_norm": 0.50390625, "learning_rate": 8.800000000000001e-05, "loss": 1.6377, "step": 88 }, { "epoch": 0.007594747677308558, "grad_norm": 0.48046875, "learning_rate": 8.900000000000001e-05, "loss": 1.5294, "step": 89 }, { "epoch": 0.007680081920873823, "grad_norm": 0.50390625, "learning_rate": 9e-05, "loss": 1.6331, "step": 90 }, { "epoch": 0.007765416164439087, "grad_norm": 0.5078125, "learning_rate": 9.1e-05, "loss": 1.725, "step": 91 }, { "epoch": 0.007850750408004353, "grad_norm": 0.5, "learning_rate": 9.200000000000001e-05, "loss": 1.6189, "step": 92 }, { "epoch": 0.007936084651569617, "grad_norm": 0.546875, "learning_rate": 9.300000000000001e-05, "loss": 1.5317, "step": 93 }, { "epoch": 0.008021418895134882, "grad_norm": 0.48828125, "learning_rate": 9.4e-05, "loss": 1.4674, "step": 94 }, { "epoch": 0.008106753138700146, "grad_norm": 0.55078125, "learning_rate": 9.5e-05, "loss": 1.8128, "step": 95 }, { "epoch": 0.008192087382265411, "grad_norm": 0.54296875, "learning_rate": 9.6e-05, "loss": 1.5884, "step": 96 }, { "epoch": 0.008277421625830676, "grad_norm": 0.46875, "learning_rate": 9.7e-05, "loss": 1.6662, "step": 97 }, { "epoch": 0.00836275586939594, "grad_norm": 0.5078125, "learning_rate": 9.8e-05, "loss": 1.5444, "step": 98 }, { "epoch": 0.008448090112961205, "grad_norm": 0.57421875, "learning_rate": 9.900000000000001e-05, "loss": 1.8447, "step": 99 }, { "epoch": 0.00853342435652647, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5976, "step": 100 }, { "epoch": 0.008618758600091734, "grad_norm": 0.6640625, "learning_rate": 9.999497487437187e-05, "loss": 1.4566, "step": 101 }, { "epoch": 0.008704092843656999, "grad_norm": 0.5078125, "learning_rate": 9.998994974874373e-05, "loss": 1.9021, "step": 102 }, { "epoch": 0.008789427087222263, "grad_norm": 0.435546875, "learning_rate": 9.998492462311558e-05, "loss": 1.4717, "step": 103 }, { "epoch": 0.008874761330787528, "grad_norm": 0.482421875, "learning_rate": 9.997989949748744e-05, "loss": 1.5066, "step": 104 }, { "epoch": 0.008960095574352792, "grad_norm": 0.578125, "learning_rate": 9.99748743718593e-05, "loss": 1.553, "step": 105 }, { "epoch": 0.009045429817918059, "grad_norm": 0.4609375, "learning_rate": 9.996984924623116e-05, "loss": 1.4151, "step": 106 }, { "epoch": 0.009130764061483323, "grad_norm": 0.4921875, "learning_rate": 9.996482412060301e-05, "loss": 1.7529, "step": 107 }, { "epoch": 0.009216098305048588, "grad_norm": 0.486328125, "learning_rate": 9.995979899497487e-05, "loss": 1.6218, "step": 108 }, { "epoch": 0.009301432548613852, "grad_norm": 0.4765625, "learning_rate": 9.995477386934674e-05, "loss": 1.6131, "step": 109 }, { "epoch": 0.009386766792179117, "grad_norm": 0.55078125, "learning_rate": 9.99497487437186e-05, "loss": 1.7024, "step": 110 }, { "epoch": 0.009472101035744382, "grad_norm": 0.4453125, "learning_rate": 9.994472361809045e-05, "loss": 1.5234, "step": 111 }, { "epoch": 0.009557435279309646, "grad_norm": 0.482421875, "learning_rate": 9.993969849246232e-05, "loss": 1.4346, "step": 112 }, { "epoch": 0.00964276952287491, "grad_norm": 0.427734375, "learning_rate": 9.993467336683417e-05, "loss": 1.5256, "step": 113 }, { "epoch": 0.009728103766440175, "grad_norm": 0.486328125, "learning_rate": 9.992964824120603e-05, "loss": 1.7505, "step": 114 }, { "epoch": 0.00981343801000544, "grad_norm": 0.462890625, "learning_rate": 9.99246231155779e-05, "loss": 1.534, "step": 115 }, { "epoch": 0.009898772253570704, "grad_norm": 0.5546875, "learning_rate": 9.991959798994976e-05, "loss": 1.7212, "step": 116 }, { "epoch": 0.009984106497135969, "grad_norm": 0.4609375, "learning_rate": 9.99145728643216e-05, "loss": 1.4882, "step": 117 }, { "epoch": 0.010069440740701234, "grad_norm": 0.443359375, "learning_rate": 9.990954773869348e-05, "loss": 1.6496, "step": 118 }, { "epoch": 0.010154774984266498, "grad_norm": 0.47265625, "learning_rate": 9.990452261306533e-05, "loss": 1.547, "step": 119 }, { "epoch": 0.010240109227831763, "grad_norm": 0.47265625, "learning_rate": 9.989949748743719e-05, "loss": 1.548, "step": 120 }, { "epoch": 0.010325443471397027, "grad_norm": 0.41015625, "learning_rate": 9.989447236180905e-05, "loss": 1.5738, "step": 121 }, { "epoch": 0.010410777714962294, "grad_norm": 0.4296875, "learning_rate": 9.988944723618092e-05, "loss": 1.4623, "step": 122 }, { "epoch": 0.010496111958527558, "grad_norm": 0.482421875, "learning_rate": 9.988442211055276e-05, "loss": 1.5758, "step": 123 }, { "epoch": 0.010581446202092823, "grad_norm": 0.396484375, "learning_rate": 9.987939698492463e-05, "loss": 1.4766, "step": 124 }, { "epoch": 0.010666780445658088, "grad_norm": 0.45703125, "learning_rate": 9.987437185929649e-05, "loss": 1.5325, "step": 125 }, { "epoch": 0.010752114689223352, "grad_norm": 0.52734375, "learning_rate": 9.986934673366835e-05, "loss": 1.5734, "step": 126 }, { "epoch": 0.010837448932788617, "grad_norm": 0.43359375, "learning_rate": 9.98643216080402e-05, "loss": 1.5174, "step": 127 }, { "epoch": 0.010922783176353881, "grad_norm": 0.384765625, "learning_rate": 9.985929648241207e-05, "loss": 1.4156, "step": 128 }, { "epoch": 0.011008117419919146, "grad_norm": 0.392578125, "learning_rate": 9.985427135678392e-05, "loss": 1.5875, "step": 129 }, { "epoch": 0.01109345166348441, "grad_norm": 0.37109375, "learning_rate": 9.984924623115578e-05, "loss": 1.4329, "step": 130 }, { "epoch": 0.011178785907049675, "grad_norm": 0.4453125, "learning_rate": 9.984422110552765e-05, "loss": 1.6382, "step": 131 }, { "epoch": 0.01126412015061494, "grad_norm": 0.435546875, "learning_rate": 9.983919597989951e-05, "loss": 1.5618, "step": 132 }, { "epoch": 0.011349454394180204, "grad_norm": 0.41796875, "learning_rate": 9.983417085427136e-05, "loss": 1.4175, "step": 133 }, { "epoch": 0.011434788637745469, "grad_norm": 0.44921875, "learning_rate": 9.982914572864322e-05, "loss": 1.5475, "step": 134 }, { "epoch": 0.011520122881310733, "grad_norm": 0.4765625, "learning_rate": 9.982412060301508e-05, "loss": 1.5458, "step": 135 }, { "epoch": 0.011605457124875998, "grad_norm": 0.46875, "learning_rate": 9.981909547738694e-05, "loss": 1.5643, "step": 136 }, { "epoch": 0.011690791368441263, "grad_norm": 0.44140625, "learning_rate": 9.98140703517588e-05, "loss": 1.7032, "step": 137 }, { "epoch": 0.011776125612006529, "grad_norm": 0.384765625, "learning_rate": 9.980904522613065e-05, "loss": 1.4243, "step": 138 }, { "epoch": 0.011861459855571794, "grad_norm": 0.392578125, "learning_rate": 9.980402010050252e-05, "loss": 1.4475, "step": 139 }, { "epoch": 0.011946794099137058, "grad_norm": 0.427734375, "learning_rate": 9.979899497487438e-05, "loss": 1.5821, "step": 140 }, { "epoch": 0.012032128342702323, "grad_norm": 0.435546875, "learning_rate": 9.979396984924624e-05, "loss": 1.5276, "step": 141 }, { "epoch": 0.012117462586267587, "grad_norm": 0.4296875, "learning_rate": 9.978894472361809e-05, "loss": 1.5591, "step": 142 }, { "epoch": 0.012202796829832852, "grad_norm": 0.392578125, "learning_rate": 9.978391959798995e-05, "loss": 1.3586, "step": 143 }, { "epoch": 0.012288131073398116, "grad_norm": 0.45703125, "learning_rate": 9.977889447236181e-05, "loss": 1.506, "step": 144 }, { "epoch": 0.012373465316963381, "grad_norm": 0.451171875, "learning_rate": 9.977386934673367e-05, "loss": 1.5435, "step": 145 }, { "epoch": 0.012458799560528646, "grad_norm": 0.431640625, "learning_rate": 9.976884422110552e-05, "loss": 1.5242, "step": 146 }, { "epoch": 0.01254413380409391, "grad_norm": 0.439453125, "learning_rate": 9.97638190954774e-05, "loss": 1.5617, "step": 147 }, { "epoch": 0.012629468047659175, "grad_norm": 0.392578125, "learning_rate": 9.975879396984925e-05, "loss": 1.6439, "step": 148 }, { "epoch": 0.01271480229122444, "grad_norm": 0.40625, "learning_rate": 9.975376884422111e-05, "loss": 1.4132, "step": 149 }, { "epoch": 0.012800136534789704, "grad_norm": 0.40234375, "learning_rate": 9.974874371859297e-05, "loss": 1.5351, "step": 150 }, { "epoch": 0.012885470778354969, "grad_norm": 0.396484375, "learning_rate": 9.974371859296483e-05, "loss": 1.5083, "step": 151 }, { "epoch": 0.012970805021920233, "grad_norm": 0.41796875, "learning_rate": 9.973869346733668e-05, "loss": 1.3979, "step": 152 }, { "epoch": 0.013056139265485498, "grad_norm": 0.43359375, "learning_rate": 9.973366834170856e-05, "loss": 1.4763, "step": 153 }, { "epoch": 0.013141473509050762, "grad_norm": 0.427734375, "learning_rate": 9.97286432160804e-05, "loss": 1.3848, "step": 154 }, { "epoch": 0.013226807752616029, "grad_norm": 0.37109375, "learning_rate": 9.972361809045227e-05, "loss": 1.4794, "step": 155 }, { "epoch": 0.013312141996181293, "grad_norm": 0.392578125, "learning_rate": 9.971859296482412e-05, "loss": 1.6615, "step": 156 }, { "epoch": 0.013397476239746558, "grad_norm": 0.404296875, "learning_rate": 9.971356783919599e-05, "loss": 1.459, "step": 157 }, { "epoch": 0.013482810483311822, "grad_norm": 0.458984375, "learning_rate": 9.970854271356784e-05, "loss": 1.4856, "step": 158 }, { "epoch": 0.013568144726877087, "grad_norm": 0.390625, "learning_rate": 9.97035175879397e-05, "loss": 1.5139, "step": 159 }, { "epoch": 0.013653478970442352, "grad_norm": 0.455078125, "learning_rate": 9.969849246231156e-05, "loss": 1.4959, "step": 160 }, { "epoch": 0.013738813214007616, "grad_norm": 0.400390625, "learning_rate": 9.969346733668343e-05, "loss": 1.5225, "step": 161 }, { "epoch": 0.01382414745757288, "grad_norm": 0.5078125, "learning_rate": 9.968844221105527e-05, "loss": 1.4846, "step": 162 }, { "epoch": 0.013909481701138145, "grad_norm": 0.369140625, "learning_rate": 9.968341708542715e-05, "loss": 1.461, "step": 163 }, { "epoch": 0.01399481594470341, "grad_norm": 0.388671875, "learning_rate": 9.9678391959799e-05, "loss": 1.6441, "step": 164 }, { "epoch": 0.014080150188268675, "grad_norm": 0.421875, "learning_rate": 9.967336683417086e-05, "loss": 1.5616, "step": 165 }, { "epoch": 0.01416548443183394, "grad_norm": 0.37890625, "learning_rate": 9.966834170854272e-05, "loss": 1.4335, "step": 166 }, { "epoch": 0.014250818675399204, "grad_norm": 0.3828125, "learning_rate": 9.966331658291458e-05, "loss": 1.3676, "step": 167 }, { "epoch": 0.014336152918964468, "grad_norm": 0.4140625, "learning_rate": 9.965829145728643e-05, "loss": 1.6303, "step": 168 }, { "epoch": 0.014421487162529733, "grad_norm": 0.380859375, "learning_rate": 9.96532663316583e-05, "loss": 1.5304, "step": 169 }, { "epoch": 0.014506821406094998, "grad_norm": 0.400390625, "learning_rate": 9.964824120603016e-05, "loss": 1.5157, "step": 170 }, { "epoch": 0.014592155649660264, "grad_norm": 0.455078125, "learning_rate": 9.964321608040202e-05, "loss": 1.5339, "step": 171 }, { "epoch": 0.014677489893225528, "grad_norm": 0.375, "learning_rate": 9.963819095477387e-05, "loss": 1.4739, "step": 172 }, { "epoch": 0.014762824136790793, "grad_norm": 0.4375, "learning_rate": 9.963316582914573e-05, "loss": 1.3935, "step": 173 }, { "epoch": 0.014848158380356058, "grad_norm": 0.37890625, "learning_rate": 9.962814070351759e-05, "loss": 1.4655, "step": 174 }, { "epoch": 0.014933492623921322, "grad_norm": 0.41796875, "learning_rate": 9.962311557788945e-05, "loss": 1.4473, "step": 175 }, { "epoch": 0.015018826867486587, "grad_norm": 0.435546875, "learning_rate": 9.961809045226132e-05, "loss": 1.4698, "step": 176 }, { "epoch": 0.015104161111051851, "grad_norm": 0.37109375, "learning_rate": 9.961306532663316e-05, "loss": 1.5137, "step": 177 }, { "epoch": 0.015189495354617116, "grad_norm": 0.3671875, "learning_rate": 9.960804020100503e-05, "loss": 1.5024, "step": 178 }, { "epoch": 0.01527482959818238, "grad_norm": 0.416015625, "learning_rate": 9.960301507537689e-05, "loss": 1.4906, "step": 179 }, { "epoch": 0.015360163841747645, "grad_norm": 0.396484375, "learning_rate": 9.959798994974875e-05, "loss": 1.4285, "step": 180 }, { "epoch": 0.01544549808531291, "grad_norm": 0.373046875, "learning_rate": 9.95929648241206e-05, "loss": 1.3331, "step": 181 }, { "epoch": 0.015530832328878174, "grad_norm": 0.421875, "learning_rate": 9.958793969849247e-05, "loss": 1.5696, "step": 182 }, { "epoch": 0.015616166572443439, "grad_norm": 0.4453125, "learning_rate": 9.958291457286432e-05, "loss": 1.7032, "step": 183 }, { "epoch": 0.015701500816008705, "grad_norm": 0.396484375, "learning_rate": 9.957788944723619e-05, "loss": 1.4965, "step": 184 }, { "epoch": 0.01578683505957397, "grad_norm": 0.3984375, "learning_rate": 9.957286432160805e-05, "loss": 1.4278, "step": 185 }, { "epoch": 0.015872169303139234, "grad_norm": 0.353515625, "learning_rate": 9.956783919597991e-05, "loss": 1.3238, "step": 186 }, { "epoch": 0.0159575035467045, "grad_norm": 0.4296875, "learning_rate": 9.956281407035176e-05, "loss": 1.5803, "step": 187 }, { "epoch": 0.016042837790269764, "grad_norm": 0.396484375, "learning_rate": 9.955778894472362e-05, "loss": 1.5097, "step": 188 }, { "epoch": 0.016128172033835028, "grad_norm": 0.376953125, "learning_rate": 9.955276381909548e-05, "loss": 1.5302, "step": 189 }, { "epoch": 0.016213506277400293, "grad_norm": 0.421875, "learning_rate": 9.954773869346734e-05, "loss": 1.4605, "step": 190 }, { "epoch": 0.016298840520965557, "grad_norm": 0.431640625, "learning_rate": 9.954271356783919e-05, "loss": 1.4818, "step": 191 }, { "epoch": 0.016384174764530822, "grad_norm": 0.44921875, "learning_rate": 9.953768844221107e-05, "loss": 1.5929, "step": 192 }, { "epoch": 0.016469509008096087, "grad_norm": 0.373046875, "learning_rate": 9.953266331658292e-05, "loss": 1.6972, "step": 193 }, { "epoch": 0.01655484325166135, "grad_norm": 0.423828125, "learning_rate": 9.952763819095478e-05, "loss": 1.7035, "step": 194 }, { "epoch": 0.016640177495226616, "grad_norm": 0.4140625, "learning_rate": 9.952261306532664e-05, "loss": 1.522, "step": 195 }, { "epoch": 0.01672551173879188, "grad_norm": 0.40234375, "learning_rate": 9.95175879396985e-05, "loss": 1.5499, "step": 196 }, { "epoch": 0.016810845982357145, "grad_norm": 0.396484375, "learning_rate": 9.951256281407035e-05, "loss": 1.3498, "step": 197 }, { "epoch": 0.01689618022592241, "grad_norm": 0.423828125, "learning_rate": 9.950753768844223e-05, "loss": 1.4946, "step": 198 }, { "epoch": 0.016981514469487674, "grad_norm": 0.412109375, "learning_rate": 9.950251256281408e-05, "loss": 1.4106, "step": 199 }, { "epoch": 0.01706684871305294, "grad_norm": 0.34375, "learning_rate": 9.949748743718594e-05, "loss": 1.3061, "step": 200 }, { "epoch": 0.017152182956618203, "grad_norm": 0.3671875, "learning_rate": 9.94924623115578e-05, "loss": 1.6192, "step": 201 }, { "epoch": 0.017237517200183468, "grad_norm": 0.376953125, "learning_rate": 9.948743718592966e-05, "loss": 1.6002, "step": 202 }, { "epoch": 0.017322851443748732, "grad_norm": 0.44921875, "learning_rate": 9.948241206030151e-05, "loss": 1.4208, "step": 203 }, { "epoch": 0.017408185687313997, "grad_norm": 0.359375, "learning_rate": 9.947738693467337e-05, "loss": 1.4471, "step": 204 }, { "epoch": 0.01749351993087926, "grad_norm": 0.3828125, "learning_rate": 9.947236180904523e-05, "loss": 1.5423, "step": 205 }, { "epoch": 0.017578854174444526, "grad_norm": 0.390625, "learning_rate": 9.94673366834171e-05, "loss": 1.4858, "step": 206 }, { "epoch": 0.01766418841800979, "grad_norm": 0.369140625, "learning_rate": 9.946231155778894e-05, "loss": 1.492, "step": 207 }, { "epoch": 0.017749522661575055, "grad_norm": 0.35546875, "learning_rate": 9.94572864321608e-05, "loss": 1.4228, "step": 208 }, { "epoch": 0.01783485690514032, "grad_norm": 0.359375, "learning_rate": 9.945226130653267e-05, "loss": 1.3717, "step": 209 }, { "epoch": 0.017920191148705585, "grad_norm": 0.36328125, "learning_rate": 9.944723618090453e-05, "loss": 1.4904, "step": 210 }, { "epoch": 0.018005525392270853, "grad_norm": 0.34375, "learning_rate": 9.944221105527639e-05, "loss": 1.3978, "step": 211 }, { "epoch": 0.018090859635836117, "grad_norm": 0.41015625, "learning_rate": 9.943718592964824e-05, "loss": 1.4583, "step": 212 }, { "epoch": 0.018176193879401382, "grad_norm": 0.36328125, "learning_rate": 9.94321608040201e-05, "loss": 1.5892, "step": 213 }, { "epoch": 0.018261528122966646, "grad_norm": 0.37109375, "learning_rate": 9.942713567839197e-05, "loss": 1.5942, "step": 214 }, { "epoch": 0.01834686236653191, "grad_norm": 0.35546875, "learning_rate": 9.942211055276383e-05, "loss": 1.5318, "step": 215 }, { "epoch": 0.018432196610097176, "grad_norm": 0.359375, "learning_rate": 9.941708542713568e-05, "loss": 1.4243, "step": 216 }, { "epoch": 0.01851753085366244, "grad_norm": 0.392578125, "learning_rate": 9.941206030150754e-05, "loss": 1.3628, "step": 217 }, { "epoch": 0.018602865097227705, "grad_norm": 0.3359375, "learning_rate": 9.94070351758794e-05, "loss": 1.4799, "step": 218 }, { "epoch": 0.01868819934079297, "grad_norm": 0.34765625, "learning_rate": 9.940201005025126e-05, "loss": 1.5227, "step": 219 }, { "epoch": 0.018773533584358234, "grad_norm": 0.341796875, "learning_rate": 9.939698492462311e-05, "loss": 1.436, "step": 220 }, { "epoch": 0.0188588678279235, "grad_norm": 0.3125, "learning_rate": 9.939195979899499e-05, "loss": 1.4322, "step": 221 }, { "epoch": 0.018944202071488763, "grad_norm": 0.373046875, "learning_rate": 9.938693467336683e-05, "loss": 1.5913, "step": 222 }, { "epoch": 0.019029536315054028, "grad_norm": 0.373046875, "learning_rate": 9.93819095477387e-05, "loss": 1.5052, "step": 223 }, { "epoch": 0.019114870558619292, "grad_norm": 0.390625, "learning_rate": 9.937688442211056e-05, "loss": 1.6388, "step": 224 }, { "epoch": 0.019200204802184557, "grad_norm": 0.416015625, "learning_rate": 9.937185929648242e-05, "loss": 1.5977, "step": 225 }, { "epoch": 0.01928553904574982, "grad_norm": 0.400390625, "learning_rate": 9.936683417085427e-05, "loss": 1.4548, "step": 226 }, { "epoch": 0.019370873289315086, "grad_norm": 0.369140625, "learning_rate": 9.936180904522614e-05, "loss": 1.4918, "step": 227 }, { "epoch": 0.01945620753288035, "grad_norm": 0.4296875, "learning_rate": 9.935678391959799e-05, "loss": 1.5212, "step": 228 }, { "epoch": 0.019541541776445615, "grad_norm": 0.5234375, "learning_rate": 9.935175879396985e-05, "loss": 1.5645, "step": 229 }, { "epoch": 0.01962687602001088, "grad_norm": 0.37109375, "learning_rate": 9.934673366834172e-05, "loss": 1.4794, "step": 230 }, { "epoch": 0.019712210263576144, "grad_norm": 0.4296875, "learning_rate": 9.934170854271358e-05, "loss": 1.4586, "step": 231 }, { "epoch": 0.01979754450714141, "grad_norm": 0.369140625, "learning_rate": 9.933668341708543e-05, "loss": 1.5006, "step": 232 }, { "epoch": 0.019882878750706674, "grad_norm": 0.388671875, "learning_rate": 9.933165829145729e-05, "loss": 1.5719, "step": 233 }, { "epoch": 0.019968212994271938, "grad_norm": 0.37109375, "learning_rate": 9.932663316582915e-05, "loss": 1.3318, "step": 234 }, { "epoch": 0.020053547237837203, "grad_norm": 0.388671875, "learning_rate": 9.932160804020101e-05, "loss": 1.5506, "step": 235 }, { "epoch": 0.020138881481402467, "grad_norm": 0.357421875, "learning_rate": 9.931658291457286e-05, "loss": 1.4031, "step": 236 }, { "epoch": 0.020224215724967732, "grad_norm": 0.3828125, "learning_rate": 9.931155778894474e-05, "loss": 1.5384, "step": 237 }, { "epoch": 0.020309549968532997, "grad_norm": 0.3515625, "learning_rate": 9.930653266331659e-05, "loss": 1.3319, "step": 238 }, { "epoch": 0.02039488421209826, "grad_norm": 0.39453125, "learning_rate": 9.930150753768845e-05, "loss": 1.441, "step": 239 }, { "epoch": 0.020480218455663526, "grad_norm": 0.353515625, "learning_rate": 9.929648241206031e-05, "loss": 1.4311, "step": 240 }, { "epoch": 0.02056555269922879, "grad_norm": 0.380859375, "learning_rate": 9.929145728643217e-05, "loss": 1.4405, "step": 241 }, { "epoch": 0.020650886942794055, "grad_norm": 0.322265625, "learning_rate": 9.928643216080402e-05, "loss": 1.3873, "step": 242 }, { "epoch": 0.02073622118635932, "grad_norm": 0.380859375, "learning_rate": 9.928140703517588e-05, "loss": 1.6589, "step": 243 }, { "epoch": 0.020821555429924588, "grad_norm": 0.396484375, "learning_rate": 9.927638190954774e-05, "loss": 1.4386, "step": 244 }, { "epoch": 0.020906889673489852, "grad_norm": 0.384765625, "learning_rate": 9.927135678391961e-05, "loss": 1.4542, "step": 245 }, { "epoch": 0.020992223917055117, "grad_norm": 0.36328125, "learning_rate": 9.926633165829147e-05, "loss": 1.4885, "step": 246 }, { "epoch": 0.02107755816062038, "grad_norm": 0.39453125, "learning_rate": 9.926130653266332e-05, "loss": 1.5343, "step": 247 }, { "epoch": 0.021162892404185646, "grad_norm": 0.38671875, "learning_rate": 9.925628140703518e-05, "loss": 1.4289, "step": 248 }, { "epoch": 0.02124822664775091, "grad_norm": 0.3515625, "learning_rate": 9.925125628140703e-05, "loss": 1.5846, "step": 249 }, { "epoch": 0.021333560891316175, "grad_norm": 0.40625, "learning_rate": 9.92462311557789e-05, "loss": 1.308, "step": 250 }, { "epoch": 0.02141889513488144, "grad_norm": 0.34375, "learning_rate": 9.924120603015075e-05, "loss": 1.6103, "step": 251 }, { "epoch": 0.021504229378446704, "grad_norm": 0.365234375, "learning_rate": 9.923618090452261e-05, "loss": 1.5454, "step": 252 }, { "epoch": 0.02158956362201197, "grad_norm": 0.380859375, "learning_rate": 9.923115577889448e-05, "loss": 1.5327, "step": 253 }, { "epoch": 0.021674897865577233, "grad_norm": 0.3828125, "learning_rate": 9.922613065326634e-05, "loss": 1.4348, "step": 254 }, { "epoch": 0.021760232109142498, "grad_norm": 0.341796875, "learning_rate": 9.922110552763819e-05, "loss": 1.5543, "step": 255 }, { "epoch": 0.021845566352707763, "grad_norm": 0.39453125, "learning_rate": 9.921608040201006e-05, "loss": 1.4348, "step": 256 }, { "epoch": 0.021930900596273027, "grad_norm": 0.34375, "learning_rate": 9.921105527638191e-05, "loss": 1.3951, "step": 257 }, { "epoch": 0.022016234839838292, "grad_norm": 0.3671875, "learning_rate": 9.920603015075377e-05, "loss": 1.3103, "step": 258 }, { "epoch": 0.022101569083403556, "grad_norm": 0.353515625, "learning_rate": 9.920100502512563e-05, "loss": 1.3765, "step": 259 }, { "epoch": 0.02218690332696882, "grad_norm": 0.341796875, "learning_rate": 9.91959798994975e-05, "loss": 1.3495, "step": 260 }, { "epoch": 0.022272237570534086, "grad_norm": 0.3515625, "learning_rate": 9.919095477386935e-05, "loss": 1.3266, "step": 261 }, { "epoch": 0.02235757181409935, "grad_norm": 0.30078125, "learning_rate": 9.918592964824122e-05, "loss": 1.2198, "step": 262 }, { "epoch": 0.022442906057664615, "grad_norm": 0.32421875, "learning_rate": 9.918090452261307e-05, "loss": 1.419, "step": 263 }, { "epoch": 0.02252824030122988, "grad_norm": 0.34765625, "learning_rate": 9.917587939698493e-05, "loss": 1.341, "step": 264 }, { "epoch": 0.022613574544795144, "grad_norm": 0.30859375, "learning_rate": 9.917085427135678e-05, "loss": 1.323, "step": 265 }, { "epoch": 0.02269890878836041, "grad_norm": 0.37890625, "learning_rate": 9.916582914572866e-05, "loss": 1.3533, "step": 266 }, { "epoch": 0.022784243031925673, "grad_norm": 0.3515625, "learning_rate": 9.91608040201005e-05, "loss": 1.3956, "step": 267 }, { "epoch": 0.022869577275490938, "grad_norm": 0.41015625, "learning_rate": 9.915577889447237e-05, "loss": 1.5848, "step": 268 }, { "epoch": 0.022954911519056202, "grad_norm": 0.40625, "learning_rate": 9.915075376884423e-05, "loss": 1.499, "step": 269 }, { "epoch": 0.023040245762621467, "grad_norm": 0.3671875, "learning_rate": 9.914572864321609e-05, "loss": 1.4106, "step": 270 }, { "epoch": 0.02312558000618673, "grad_norm": 0.37109375, "learning_rate": 9.914070351758794e-05, "loss": 1.4622, "step": 271 }, { "epoch": 0.023210914249751996, "grad_norm": 0.359375, "learning_rate": 9.913567839195981e-05, "loss": 1.3194, "step": 272 }, { "epoch": 0.02329624849331726, "grad_norm": 0.37890625, "learning_rate": 9.913065326633166e-05, "loss": 1.432, "step": 273 }, { "epoch": 0.023381582736882525, "grad_norm": 0.376953125, "learning_rate": 9.912562814070352e-05, "loss": 1.4197, "step": 274 }, { "epoch": 0.02346691698044779, "grad_norm": 0.345703125, "learning_rate": 9.912060301507539e-05, "loss": 1.3807, "step": 275 }, { "epoch": 0.023552251224013058, "grad_norm": 0.380859375, "learning_rate": 9.911557788944725e-05, "loss": 1.4024, "step": 276 }, { "epoch": 0.023637585467578322, "grad_norm": 0.380859375, "learning_rate": 9.91105527638191e-05, "loss": 1.4863, "step": 277 }, { "epoch": 0.023722919711143587, "grad_norm": 0.375, "learning_rate": 9.910552763819096e-05, "loss": 1.314, "step": 278 }, { "epoch": 0.02380825395470885, "grad_norm": 0.375, "learning_rate": 9.910050251256282e-05, "loss": 1.4456, "step": 279 }, { "epoch": 0.023893588198274116, "grad_norm": 0.333984375, "learning_rate": 9.909547738693468e-05, "loss": 1.5877, "step": 280 }, { "epoch": 0.02397892244183938, "grad_norm": 0.400390625, "learning_rate": 9.909045226130653e-05, "loss": 1.5869, "step": 281 }, { "epoch": 0.024064256685404645, "grad_norm": 0.357421875, "learning_rate": 9.90854271356784e-05, "loss": 1.5391, "step": 282 }, { "epoch": 0.02414959092896991, "grad_norm": 0.322265625, "learning_rate": 9.908040201005026e-05, "loss": 1.3901, "step": 283 }, { "epoch": 0.024234925172535175, "grad_norm": 0.345703125, "learning_rate": 9.90753768844221e-05, "loss": 1.4891, "step": 284 }, { "epoch": 0.02432025941610044, "grad_norm": 0.306640625, "learning_rate": 9.907035175879398e-05, "loss": 1.3531, "step": 285 }, { "epoch": 0.024405593659665704, "grad_norm": 0.421875, "learning_rate": 9.906532663316583e-05, "loss": 1.3986, "step": 286 }, { "epoch": 0.02449092790323097, "grad_norm": 0.34765625, "learning_rate": 9.906030150753769e-05, "loss": 1.5053, "step": 287 }, { "epoch": 0.024576262146796233, "grad_norm": 0.361328125, "learning_rate": 9.905527638190955e-05, "loss": 1.4861, "step": 288 }, { "epoch": 0.024661596390361498, "grad_norm": 0.361328125, "learning_rate": 9.905025125628141e-05, "loss": 1.3849, "step": 289 }, { "epoch": 0.024746930633926762, "grad_norm": 0.357421875, "learning_rate": 9.904522613065326e-05, "loss": 1.4285, "step": 290 }, { "epoch": 0.024832264877492027, "grad_norm": 0.341796875, "learning_rate": 9.904020100502514e-05, "loss": 1.3654, "step": 291 }, { "epoch": 0.02491759912105729, "grad_norm": 0.357421875, "learning_rate": 9.903517587939699e-05, "loss": 1.453, "step": 292 }, { "epoch": 0.025002933364622556, "grad_norm": 0.3515625, "learning_rate": 9.903015075376885e-05, "loss": 1.4705, "step": 293 }, { "epoch": 0.02508826760818782, "grad_norm": 0.353515625, "learning_rate": 9.902512562814071e-05, "loss": 1.4251, "step": 294 }, { "epoch": 0.025173601851753085, "grad_norm": 0.310546875, "learning_rate": 9.902010050251257e-05, "loss": 1.2653, "step": 295 }, { "epoch": 0.02525893609531835, "grad_norm": 0.34765625, "learning_rate": 9.901507537688442e-05, "loss": 1.4037, "step": 296 }, { "epoch": 0.025344270338883614, "grad_norm": 0.345703125, "learning_rate": 9.901005025125628e-05, "loss": 1.522, "step": 297 }, { "epoch": 0.02542960458244888, "grad_norm": 0.369140625, "learning_rate": 9.900502512562815e-05, "loss": 1.2837, "step": 298 }, { "epoch": 0.025514938826014143, "grad_norm": 0.33984375, "learning_rate": 9.900000000000001e-05, "loss": 1.3896, "step": 299 }, { "epoch": 0.025600273069579408, "grad_norm": 0.3984375, "learning_rate": 9.899497487437186e-05, "loss": 1.5719, "step": 300 }, { "epoch": 0.025685607313144673, "grad_norm": 0.4296875, "learning_rate": 9.898994974874373e-05, "loss": 1.6433, "step": 301 }, { "epoch": 0.025770941556709937, "grad_norm": 0.3984375, "learning_rate": 9.898492462311558e-05, "loss": 1.4654, "step": 302 }, { "epoch": 0.025856275800275202, "grad_norm": 0.34375, "learning_rate": 9.897989949748744e-05, "loss": 1.5617, "step": 303 }, { "epoch": 0.025941610043840466, "grad_norm": 0.3671875, "learning_rate": 9.89748743718593e-05, "loss": 1.3666, "step": 304 }, { "epoch": 0.02602694428740573, "grad_norm": 0.423828125, "learning_rate": 9.896984924623117e-05, "loss": 1.3791, "step": 305 }, { "epoch": 0.026112278530970996, "grad_norm": 0.326171875, "learning_rate": 9.896482412060301e-05, "loss": 1.9473, "step": 306 }, { "epoch": 0.02619761277453626, "grad_norm": 0.400390625, "learning_rate": 9.895979899497489e-05, "loss": 1.3326, "step": 307 }, { "epoch": 0.026282947018101525, "grad_norm": 0.341796875, "learning_rate": 9.895477386934674e-05, "loss": 1.4549, "step": 308 }, { "epoch": 0.026368281261666793, "grad_norm": 0.357421875, "learning_rate": 9.89497487437186e-05, "loss": 1.3408, "step": 309 }, { "epoch": 0.026453615505232057, "grad_norm": 0.375, "learning_rate": 9.894472361809046e-05, "loss": 1.4272, "step": 310 }, { "epoch": 0.026538949748797322, "grad_norm": 0.322265625, "learning_rate": 9.893969849246232e-05, "loss": 1.3771, "step": 311 }, { "epoch": 0.026624283992362587, "grad_norm": 0.384765625, "learning_rate": 9.893467336683417e-05, "loss": 1.406, "step": 312 }, { "epoch": 0.02670961823592785, "grad_norm": 0.32421875, "learning_rate": 9.892964824120604e-05, "loss": 1.2111, "step": 313 }, { "epoch": 0.026794952479493116, "grad_norm": 0.50390625, "learning_rate": 9.89246231155779e-05, "loss": 1.44, "step": 314 }, { "epoch": 0.02688028672305838, "grad_norm": 0.34765625, "learning_rate": 9.891959798994975e-05, "loss": 1.3375, "step": 315 }, { "epoch": 0.026965620966623645, "grad_norm": 0.4453125, "learning_rate": 9.891457286432161e-05, "loss": 1.3964, "step": 316 }, { "epoch": 0.02705095521018891, "grad_norm": 0.40234375, "learning_rate": 9.890954773869347e-05, "loss": 1.2975, "step": 317 }, { "epoch": 0.027136289453754174, "grad_norm": 0.43359375, "learning_rate": 9.890452261306533e-05, "loss": 1.5118, "step": 318 }, { "epoch": 0.02722162369731944, "grad_norm": 0.322265625, "learning_rate": 9.889949748743718e-05, "loss": 1.4032, "step": 319 }, { "epoch": 0.027306957940884703, "grad_norm": 0.330078125, "learning_rate": 9.889447236180906e-05, "loss": 1.355, "step": 320 }, { "epoch": 0.027392292184449968, "grad_norm": 0.328125, "learning_rate": 9.88894472361809e-05, "loss": 1.4382, "step": 321 }, { "epoch": 0.027477626428015232, "grad_norm": 0.322265625, "learning_rate": 9.888442211055277e-05, "loss": 1.3795, "step": 322 }, { "epoch": 0.027562960671580497, "grad_norm": 0.37109375, "learning_rate": 9.887939698492463e-05, "loss": 1.4173, "step": 323 }, { "epoch": 0.02764829491514576, "grad_norm": 0.37109375, "learning_rate": 9.887437185929649e-05, "loss": 1.6702, "step": 324 }, { "epoch": 0.027733629158711026, "grad_norm": 0.341796875, "learning_rate": 9.886934673366834e-05, "loss": 1.5437, "step": 325 }, { "epoch": 0.02781896340227629, "grad_norm": 0.357421875, "learning_rate": 9.886432160804021e-05, "loss": 1.5578, "step": 326 }, { "epoch": 0.027904297645841555, "grad_norm": 0.45703125, "learning_rate": 9.885929648241206e-05, "loss": 1.5042, "step": 327 }, { "epoch": 0.02798963188940682, "grad_norm": 0.349609375, "learning_rate": 9.885427135678393e-05, "loss": 1.5055, "step": 328 }, { "epoch": 0.028074966132972085, "grad_norm": 0.34765625, "learning_rate": 9.884924623115577e-05, "loss": 1.4351, "step": 329 }, { "epoch": 0.02816030037653735, "grad_norm": 0.326171875, "learning_rate": 9.884422110552765e-05, "loss": 1.2851, "step": 330 }, { "epoch": 0.028245634620102614, "grad_norm": 0.39453125, "learning_rate": 9.88391959798995e-05, "loss": 1.5372, "step": 331 }, { "epoch": 0.02833096886366788, "grad_norm": 0.35546875, "learning_rate": 9.883417085427136e-05, "loss": 1.4504, "step": 332 }, { "epoch": 0.028416303107233143, "grad_norm": 0.361328125, "learning_rate": 9.882914572864322e-05, "loss": 1.4873, "step": 333 }, { "epoch": 0.028501637350798407, "grad_norm": 0.34765625, "learning_rate": 9.882412060301508e-05, "loss": 1.5475, "step": 334 }, { "epoch": 0.028586971594363672, "grad_norm": 0.349609375, "learning_rate": 9.881909547738693e-05, "loss": 1.3558, "step": 335 }, { "epoch": 0.028672305837928937, "grad_norm": 0.33203125, "learning_rate": 9.881407035175881e-05, "loss": 1.441, "step": 336 }, { "epoch": 0.0287576400814942, "grad_norm": 0.337890625, "learning_rate": 9.880904522613066e-05, "loss": 1.4192, "step": 337 }, { "epoch": 0.028842974325059466, "grad_norm": 0.37109375, "learning_rate": 9.880402010050252e-05, "loss": 1.4601, "step": 338 }, { "epoch": 0.02892830856862473, "grad_norm": 0.37109375, "learning_rate": 9.879899497487438e-05, "loss": 1.4178, "step": 339 }, { "epoch": 0.029013642812189995, "grad_norm": 0.318359375, "learning_rate": 9.879396984924624e-05, "loss": 1.416, "step": 340 }, { "epoch": 0.029098977055755263, "grad_norm": 0.50390625, "learning_rate": 9.878894472361809e-05, "loss": 1.3266, "step": 341 }, { "epoch": 0.029184311299320528, "grad_norm": 0.373046875, "learning_rate": 9.878391959798995e-05, "loss": 1.4582, "step": 342 }, { "epoch": 0.029269645542885792, "grad_norm": 0.34375, "learning_rate": 9.877889447236182e-05, "loss": 1.461, "step": 343 }, { "epoch": 0.029354979786451057, "grad_norm": 0.3359375, "learning_rate": 9.877386934673368e-05, "loss": 1.4722, "step": 344 }, { "epoch": 0.02944031403001632, "grad_norm": 0.33203125, "learning_rate": 9.876884422110553e-05, "loss": 1.2707, "step": 345 }, { "epoch": 0.029525648273581586, "grad_norm": 0.32421875, "learning_rate": 9.87638190954774e-05, "loss": 1.2089, "step": 346 }, { "epoch": 0.02961098251714685, "grad_norm": 0.384765625, "learning_rate": 9.875879396984925e-05, "loss": 1.5013, "step": 347 }, { "epoch": 0.029696316760712115, "grad_norm": 0.34375, "learning_rate": 9.875376884422111e-05, "loss": 1.46, "step": 348 }, { "epoch": 0.02978165100427738, "grad_norm": 0.361328125, "learning_rate": 9.874874371859297e-05, "loss": 1.527, "step": 349 }, { "epoch": 0.029866985247842644, "grad_norm": 0.35546875, "learning_rate": 9.874371859296482e-05, "loss": 1.4721, "step": 350 }, { "epoch": 0.02995231949140791, "grad_norm": 0.33984375, "learning_rate": 9.873869346733668e-05, "loss": 1.4308, "step": 351 }, { "epoch": 0.030037653734973174, "grad_norm": 0.34375, "learning_rate": 9.873366834170855e-05, "loss": 1.423, "step": 352 }, { "epoch": 0.030122987978538438, "grad_norm": 0.361328125, "learning_rate": 9.872864321608041e-05, "loss": 1.3206, "step": 353 }, { "epoch": 0.030208322222103703, "grad_norm": 0.3671875, "learning_rate": 9.872361809045226e-05, "loss": 1.5606, "step": 354 }, { "epoch": 0.030293656465668967, "grad_norm": 0.373046875, "learning_rate": 9.871859296482413e-05, "loss": 1.4185, "step": 355 }, { "epoch": 0.030378990709234232, "grad_norm": 0.357421875, "learning_rate": 9.871356783919598e-05, "loss": 1.3768, "step": 356 }, { "epoch": 0.030464324952799497, "grad_norm": 0.341796875, "learning_rate": 9.870854271356784e-05, "loss": 1.4831, "step": 357 }, { "epoch": 0.03054965919636476, "grad_norm": 0.318359375, "learning_rate": 9.870351758793969e-05, "loss": 1.4651, "step": 358 }, { "epoch": 0.030634993439930026, "grad_norm": 0.328125, "learning_rate": 9.869849246231157e-05, "loss": 1.3874, "step": 359 }, { "epoch": 0.03072032768349529, "grad_norm": 0.322265625, "learning_rate": 9.869346733668342e-05, "loss": 1.5274, "step": 360 }, { "epoch": 0.030805661927060555, "grad_norm": 0.337890625, "learning_rate": 9.868844221105528e-05, "loss": 1.462, "step": 361 }, { "epoch": 0.03089099617062582, "grad_norm": 0.37890625, "learning_rate": 9.868341708542714e-05, "loss": 1.5312, "step": 362 }, { "epoch": 0.030976330414191084, "grad_norm": 0.375, "learning_rate": 9.8678391959799e-05, "loss": 1.44, "step": 363 }, { "epoch": 0.03106166465775635, "grad_norm": 0.33203125, "learning_rate": 9.867336683417085e-05, "loss": 1.4294, "step": 364 }, { "epoch": 0.031146998901321613, "grad_norm": 0.390625, "learning_rate": 9.866834170854273e-05, "loss": 1.4829, "step": 365 }, { "epoch": 0.031232333144886878, "grad_norm": 0.37109375, "learning_rate": 9.866331658291457e-05, "loss": 1.5956, "step": 366 }, { "epoch": 0.031317667388452146, "grad_norm": 0.333984375, "learning_rate": 9.865829145728644e-05, "loss": 1.4825, "step": 367 }, { "epoch": 0.03140300163201741, "grad_norm": 0.369140625, "learning_rate": 9.86532663316583e-05, "loss": 1.4988, "step": 368 }, { "epoch": 0.031488335875582675, "grad_norm": 0.3046875, "learning_rate": 9.864824120603016e-05, "loss": 1.3063, "step": 369 }, { "epoch": 0.03157367011914794, "grad_norm": 0.3125, "learning_rate": 9.864321608040201e-05, "loss": 1.267, "step": 370 }, { "epoch": 0.031659004362713204, "grad_norm": 0.333984375, "learning_rate": 9.863819095477388e-05, "loss": 1.2518, "step": 371 }, { "epoch": 0.03174433860627847, "grad_norm": 0.337890625, "learning_rate": 9.863316582914573e-05, "loss": 1.3486, "step": 372 }, { "epoch": 0.03182967284984373, "grad_norm": 0.35546875, "learning_rate": 9.86281407035176e-05, "loss": 1.4999, "step": 373 }, { "epoch": 0.031915007093409, "grad_norm": 0.3671875, "learning_rate": 9.862311557788944e-05, "loss": 1.4815, "step": 374 }, { "epoch": 0.03200034133697426, "grad_norm": 0.35546875, "learning_rate": 9.861809045226132e-05, "loss": 1.5226, "step": 375 }, { "epoch": 0.03208567558053953, "grad_norm": 0.359375, "learning_rate": 9.861306532663317e-05, "loss": 1.4169, "step": 376 }, { "epoch": 0.03217100982410479, "grad_norm": 0.3359375, "learning_rate": 9.860804020100503e-05, "loss": 1.2659, "step": 377 }, { "epoch": 0.032256344067670056, "grad_norm": 0.328125, "learning_rate": 9.860301507537689e-05, "loss": 1.3325, "step": 378 }, { "epoch": 0.03234167831123532, "grad_norm": 0.3359375, "learning_rate": 9.859798994974875e-05, "loss": 1.2241, "step": 379 }, { "epoch": 0.032427012554800586, "grad_norm": 0.390625, "learning_rate": 9.85929648241206e-05, "loss": 1.6017, "step": 380 }, { "epoch": 0.03251234679836585, "grad_norm": 0.359375, "learning_rate": 9.858793969849246e-05, "loss": 1.436, "step": 381 }, { "epoch": 0.032597681041931115, "grad_norm": 0.337890625, "learning_rate": 9.858291457286433e-05, "loss": 1.4302, "step": 382 }, { "epoch": 0.03268301528549638, "grad_norm": 0.353515625, "learning_rate": 9.857788944723619e-05, "loss": 1.3834, "step": 383 }, { "epoch": 0.032768349529061644, "grad_norm": 0.3359375, "learning_rate": 9.857286432160805e-05, "loss": 1.3019, "step": 384 }, { "epoch": 0.03285368377262691, "grad_norm": 0.380859375, "learning_rate": 9.85678391959799e-05, "loss": 1.487, "step": 385 }, { "epoch": 0.03293901801619217, "grad_norm": 0.318359375, "learning_rate": 9.856281407035176e-05, "loss": 1.3938, "step": 386 }, { "epoch": 0.03302435225975744, "grad_norm": 0.43359375, "learning_rate": 9.855778894472362e-05, "loss": 1.488, "step": 387 }, { "epoch": 0.0331096865033227, "grad_norm": 0.337890625, "learning_rate": 9.855276381909548e-05, "loss": 1.3579, "step": 388 }, { "epoch": 0.03319502074688797, "grad_norm": 0.326171875, "learning_rate": 9.854773869346733e-05, "loss": 1.3319, "step": 389 }, { "epoch": 0.03328035499045323, "grad_norm": 0.345703125, "learning_rate": 9.85427135678392e-05, "loss": 1.3677, "step": 390 }, { "epoch": 0.033365689234018496, "grad_norm": 0.380859375, "learning_rate": 9.853768844221106e-05, "loss": 1.4472, "step": 391 }, { "epoch": 0.03345102347758376, "grad_norm": 0.326171875, "learning_rate": 9.853266331658292e-05, "loss": 1.388, "step": 392 }, { "epoch": 0.033536357721149025, "grad_norm": 0.333984375, "learning_rate": 9.852763819095477e-05, "loss": 1.2981, "step": 393 }, { "epoch": 0.03362169196471429, "grad_norm": 0.3203125, "learning_rate": 9.852261306532664e-05, "loss": 1.5465, "step": 394 }, { "epoch": 0.033707026208279554, "grad_norm": 0.306640625, "learning_rate": 9.851758793969849e-05, "loss": 1.3837, "step": 395 }, { "epoch": 0.03379236045184482, "grad_norm": 0.37109375, "learning_rate": 9.851256281407035e-05, "loss": 1.5756, "step": 396 }, { "epoch": 0.033877694695410084, "grad_norm": 0.33203125, "learning_rate": 9.850753768844222e-05, "loss": 1.3774, "step": 397 }, { "epoch": 0.03396302893897535, "grad_norm": 0.34375, "learning_rate": 9.850251256281408e-05, "loss": 1.3807, "step": 398 }, { "epoch": 0.03404836318254061, "grad_norm": 0.35546875, "learning_rate": 9.849748743718593e-05, "loss": 1.4118, "step": 399 }, { "epoch": 0.03413369742610588, "grad_norm": 0.384765625, "learning_rate": 9.84924623115578e-05, "loss": 1.5277, "step": 400 }, { "epoch": 0.03421903166967114, "grad_norm": 0.380859375, "learning_rate": 9.848743718592965e-05, "loss": 1.5215, "step": 401 }, { "epoch": 0.034304365913236406, "grad_norm": 0.326171875, "learning_rate": 9.848241206030151e-05, "loss": 1.5074, "step": 402 }, { "epoch": 0.03438970015680167, "grad_norm": 0.3671875, "learning_rate": 9.847738693467337e-05, "loss": 1.3304, "step": 403 }, { "epoch": 0.034475034400366936, "grad_norm": 0.333984375, "learning_rate": 9.847236180904524e-05, "loss": 1.3198, "step": 404 }, { "epoch": 0.0345603686439322, "grad_norm": 0.34765625, "learning_rate": 9.846733668341709e-05, "loss": 1.3872, "step": 405 }, { "epoch": 0.034645702887497465, "grad_norm": 0.353515625, "learning_rate": 9.846231155778895e-05, "loss": 1.4763, "step": 406 }, { "epoch": 0.03473103713106273, "grad_norm": 0.3203125, "learning_rate": 9.845728643216081e-05, "loss": 1.366, "step": 407 }, { "epoch": 0.034816371374627994, "grad_norm": 0.400390625, "learning_rate": 9.845226130653267e-05, "loss": 1.3394, "step": 408 }, { "epoch": 0.03490170561819326, "grad_norm": 0.33984375, "learning_rate": 9.844723618090452e-05, "loss": 1.4859, "step": 409 }, { "epoch": 0.03498703986175852, "grad_norm": 0.37109375, "learning_rate": 9.84422110552764e-05, "loss": 1.4893, "step": 410 }, { "epoch": 0.03507237410532379, "grad_norm": 0.349609375, "learning_rate": 9.843718592964824e-05, "loss": 1.4897, "step": 411 }, { "epoch": 0.03515770834888905, "grad_norm": 0.328125, "learning_rate": 9.84321608040201e-05, "loss": 1.4898, "step": 412 }, { "epoch": 0.03524304259245432, "grad_norm": 0.3203125, "learning_rate": 9.842713567839197e-05, "loss": 1.2922, "step": 413 }, { "epoch": 0.03532837683601958, "grad_norm": 0.3359375, "learning_rate": 9.842211055276383e-05, "loss": 1.4774, "step": 414 }, { "epoch": 0.035413711079584846, "grad_norm": 0.298828125, "learning_rate": 9.841708542713568e-05, "loss": 1.3156, "step": 415 }, { "epoch": 0.03549904532315011, "grad_norm": 0.294921875, "learning_rate": 9.841206030150754e-05, "loss": 1.219, "step": 416 }, { "epoch": 0.035584379566715375, "grad_norm": 0.31640625, "learning_rate": 9.84070351758794e-05, "loss": 1.346, "step": 417 }, { "epoch": 0.03566971381028064, "grad_norm": 0.369140625, "learning_rate": 9.840201005025126e-05, "loss": 1.3559, "step": 418 }, { "epoch": 0.035755048053845905, "grad_norm": 0.345703125, "learning_rate": 9.839698492462313e-05, "loss": 1.3243, "step": 419 }, { "epoch": 0.03584038229741117, "grad_norm": 0.33984375, "learning_rate": 9.839195979899497e-05, "loss": 1.492, "step": 420 }, { "epoch": 0.035925716540976434, "grad_norm": 0.326171875, "learning_rate": 9.838693467336684e-05, "loss": 1.324, "step": 421 }, { "epoch": 0.036011050784541705, "grad_norm": 0.353515625, "learning_rate": 9.83819095477387e-05, "loss": 1.4559, "step": 422 }, { "epoch": 0.03609638502810697, "grad_norm": 0.3359375, "learning_rate": 9.837688442211056e-05, "loss": 1.2916, "step": 423 }, { "epoch": 0.036181719271672234, "grad_norm": 0.4609375, "learning_rate": 9.837185929648241e-05, "loss": 1.5289, "step": 424 }, { "epoch": 0.0362670535152375, "grad_norm": 0.33984375, "learning_rate": 9.836683417085427e-05, "loss": 1.4793, "step": 425 }, { "epoch": 0.036352387758802764, "grad_norm": 0.3203125, "learning_rate": 9.836180904522613e-05, "loss": 1.3065, "step": 426 }, { "epoch": 0.03643772200236803, "grad_norm": 0.3984375, "learning_rate": 9.8356783919598e-05, "loss": 1.5644, "step": 427 }, { "epoch": 0.03652305624593329, "grad_norm": 0.328125, "learning_rate": 9.835175879396984e-05, "loss": 1.4515, "step": 428 }, { "epoch": 0.03660839048949856, "grad_norm": 0.341796875, "learning_rate": 9.834673366834172e-05, "loss": 1.3688, "step": 429 }, { "epoch": 0.03669372473306382, "grad_norm": 0.3125, "learning_rate": 9.834170854271357e-05, "loss": 1.4453, "step": 430 }, { "epoch": 0.036779058976629087, "grad_norm": 0.302734375, "learning_rate": 9.833668341708543e-05, "loss": 1.4015, "step": 431 }, { "epoch": 0.03686439322019435, "grad_norm": 0.306640625, "learning_rate": 9.833165829145729e-05, "loss": 1.3695, "step": 432 }, { "epoch": 0.036949727463759616, "grad_norm": 0.33203125, "learning_rate": 9.832663316582915e-05, "loss": 1.2991, "step": 433 }, { "epoch": 0.03703506170732488, "grad_norm": 0.365234375, "learning_rate": 9.8321608040201e-05, "loss": 1.4742, "step": 434 }, { "epoch": 0.037120395950890145, "grad_norm": 0.318359375, "learning_rate": 9.831658291457288e-05, "loss": 1.3426, "step": 435 }, { "epoch": 0.03720573019445541, "grad_norm": 0.3515625, "learning_rate": 9.831155778894473e-05, "loss": 1.3282, "step": 436 }, { "epoch": 0.037291064438020674, "grad_norm": 0.322265625, "learning_rate": 9.830653266331659e-05, "loss": 1.4286, "step": 437 }, { "epoch": 0.03737639868158594, "grad_norm": 0.333984375, "learning_rate": 9.830150753768844e-05, "loss": 1.5547, "step": 438 }, { "epoch": 0.0374617329251512, "grad_norm": 0.384765625, "learning_rate": 9.829648241206031e-05, "loss": 1.3978, "step": 439 }, { "epoch": 0.03754706716871647, "grad_norm": 0.34375, "learning_rate": 9.829145728643216e-05, "loss": 1.4236, "step": 440 }, { "epoch": 0.03763240141228173, "grad_norm": 0.306640625, "learning_rate": 9.828643216080402e-05, "loss": 1.3891, "step": 441 }, { "epoch": 0.037717735655847, "grad_norm": 0.369140625, "learning_rate": 9.828140703517589e-05, "loss": 1.5679, "step": 442 }, { "epoch": 0.03780306989941226, "grad_norm": 0.3515625, "learning_rate": 9.827638190954775e-05, "loss": 1.377, "step": 443 }, { "epoch": 0.037888404142977526, "grad_norm": 0.34375, "learning_rate": 9.82713567839196e-05, "loss": 1.4544, "step": 444 }, { "epoch": 0.03797373838654279, "grad_norm": 0.333984375, "learning_rate": 9.826633165829147e-05, "loss": 1.4464, "step": 445 }, { "epoch": 0.038059072630108055, "grad_norm": 0.314453125, "learning_rate": 9.826130653266332e-05, "loss": 1.3857, "step": 446 }, { "epoch": 0.03814440687367332, "grad_norm": 0.33984375, "learning_rate": 9.825628140703518e-05, "loss": 1.4554, "step": 447 }, { "epoch": 0.038229741117238585, "grad_norm": 0.41796875, "learning_rate": 9.825125628140704e-05, "loss": 1.4405, "step": 448 }, { "epoch": 0.03831507536080385, "grad_norm": 0.326171875, "learning_rate": 9.82462311557789e-05, "loss": 1.4376, "step": 449 }, { "epoch": 0.038400409604369114, "grad_norm": 0.484375, "learning_rate": 9.824120603015075e-05, "loss": 1.5418, "step": 450 }, { "epoch": 0.03848574384793438, "grad_norm": 0.318359375, "learning_rate": 9.823618090452262e-05, "loss": 1.3599, "step": 451 }, { "epoch": 0.03857107809149964, "grad_norm": 0.341796875, "learning_rate": 9.823115577889448e-05, "loss": 1.2787, "step": 452 }, { "epoch": 0.03865641233506491, "grad_norm": 0.34375, "learning_rate": 9.822613065326634e-05, "loss": 1.3217, "step": 453 }, { "epoch": 0.03874174657863017, "grad_norm": 0.330078125, "learning_rate": 9.822110552763819e-05, "loss": 1.2746, "step": 454 }, { "epoch": 0.03882708082219544, "grad_norm": 0.404296875, "learning_rate": 9.821608040201005e-05, "loss": 1.3369, "step": 455 }, { "epoch": 0.0389124150657607, "grad_norm": 0.333984375, "learning_rate": 9.821105527638191e-05, "loss": 1.4238, "step": 456 }, { "epoch": 0.038997749309325966, "grad_norm": 0.34765625, "learning_rate": 9.820603015075378e-05, "loss": 1.4062, "step": 457 }, { "epoch": 0.03908308355289123, "grad_norm": 0.42578125, "learning_rate": 9.820100502512564e-05, "loss": 1.5235, "step": 458 }, { "epoch": 0.039168417796456495, "grad_norm": 0.326171875, "learning_rate": 9.819597989949749e-05, "loss": 1.411, "step": 459 }, { "epoch": 0.03925375204002176, "grad_norm": 0.390625, "learning_rate": 9.819095477386935e-05, "loss": 1.3977, "step": 460 }, { "epoch": 0.039339086283587024, "grad_norm": 0.322265625, "learning_rate": 9.818592964824121e-05, "loss": 1.5666, "step": 461 }, { "epoch": 0.03942442052715229, "grad_norm": 0.33984375, "learning_rate": 9.818090452261307e-05, "loss": 1.4605, "step": 462 }, { "epoch": 0.03950975477071755, "grad_norm": 0.373046875, "learning_rate": 9.817587939698492e-05, "loss": 1.4669, "step": 463 }, { "epoch": 0.03959508901428282, "grad_norm": 0.2578125, "learning_rate": 9.81708542713568e-05, "loss": 2.2431, "step": 464 }, { "epoch": 0.03968042325784808, "grad_norm": 0.34765625, "learning_rate": 9.816582914572864e-05, "loss": 1.3993, "step": 465 }, { "epoch": 0.03976575750141335, "grad_norm": 0.3359375, "learning_rate": 9.81608040201005e-05, "loss": 1.3912, "step": 466 }, { "epoch": 0.03985109174497861, "grad_norm": 0.38671875, "learning_rate": 9.815577889447236e-05, "loss": 1.3504, "step": 467 }, { "epoch": 0.039936425988543876, "grad_norm": 0.318359375, "learning_rate": 9.815075376884423e-05, "loss": 1.2783, "step": 468 }, { "epoch": 0.04002176023210914, "grad_norm": 0.29296875, "learning_rate": 9.814572864321608e-05, "loss": 1.4368, "step": 469 }, { "epoch": 0.040107094475674405, "grad_norm": 0.33203125, "learning_rate": 9.814070351758794e-05, "loss": 1.2483, "step": 470 }, { "epoch": 0.04019242871923967, "grad_norm": 0.3984375, "learning_rate": 9.81356783919598e-05, "loss": 1.2745, "step": 471 }, { "epoch": 0.040277762962804935, "grad_norm": 0.35546875, "learning_rate": 9.813065326633167e-05, "loss": 1.1939, "step": 472 }, { "epoch": 0.0403630972063702, "grad_norm": 0.34375, "learning_rate": 9.812562814070351e-05, "loss": 1.3728, "step": 473 }, { "epoch": 0.040448431449935464, "grad_norm": 0.3359375, "learning_rate": 9.812060301507539e-05, "loss": 1.4527, "step": 474 }, { "epoch": 0.04053376569350073, "grad_norm": 0.326171875, "learning_rate": 9.811557788944724e-05, "loss": 1.5409, "step": 475 }, { "epoch": 0.04061909993706599, "grad_norm": 0.33984375, "learning_rate": 9.81105527638191e-05, "loss": 1.4426, "step": 476 }, { "epoch": 0.04070443418063126, "grad_norm": 0.333984375, "learning_rate": 9.810552763819096e-05, "loss": 1.3423, "step": 477 }, { "epoch": 0.04078976842419652, "grad_norm": 0.34765625, "learning_rate": 9.810050251256282e-05, "loss": 1.4133, "step": 478 }, { "epoch": 0.04087510266776179, "grad_norm": 0.34765625, "learning_rate": 9.809547738693467e-05, "loss": 1.3434, "step": 479 }, { "epoch": 0.04096043691132705, "grad_norm": 0.302734375, "learning_rate": 9.809045226130655e-05, "loss": 1.5089, "step": 480 }, { "epoch": 0.041045771154892316, "grad_norm": 0.33203125, "learning_rate": 9.80854271356784e-05, "loss": 1.185, "step": 481 }, { "epoch": 0.04113110539845758, "grad_norm": 0.419921875, "learning_rate": 9.808040201005026e-05, "loss": 1.4114, "step": 482 }, { "epoch": 0.041216439642022845, "grad_norm": 0.35546875, "learning_rate": 9.807537688442211e-05, "loss": 1.2592, "step": 483 }, { "epoch": 0.04130177388558811, "grad_norm": 0.35546875, "learning_rate": 9.807035175879398e-05, "loss": 1.3393, "step": 484 }, { "epoch": 0.041387108129153374, "grad_norm": 0.322265625, "learning_rate": 9.806532663316583e-05, "loss": 1.5252, "step": 485 }, { "epoch": 0.04147244237271864, "grad_norm": 0.3671875, "learning_rate": 9.806030150753769e-05, "loss": 1.5749, "step": 486 }, { "epoch": 0.04155777661628391, "grad_norm": 0.3046875, "learning_rate": 9.805527638190956e-05, "loss": 1.3505, "step": 487 }, { "epoch": 0.041643110859849175, "grad_norm": 0.431640625, "learning_rate": 9.805025125628142e-05, "loss": 1.3747, "step": 488 }, { "epoch": 0.04172844510341444, "grad_norm": 0.30859375, "learning_rate": 9.804522613065327e-05, "loss": 1.393, "step": 489 }, { "epoch": 0.041813779346979704, "grad_norm": 0.34375, "learning_rate": 9.804020100502513e-05, "loss": 1.3142, "step": 490 }, { "epoch": 0.04189911359054497, "grad_norm": 0.33203125, "learning_rate": 9.803517587939699e-05, "loss": 1.2547, "step": 491 }, { "epoch": 0.04198444783411023, "grad_norm": 0.31640625, "learning_rate": 9.803015075376885e-05, "loss": 1.3805, "step": 492 }, { "epoch": 0.0420697820776755, "grad_norm": 0.412109375, "learning_rate": 9.802512562814071e-05, "loss": 1.4887, "step": 493 }, { "epoch": 0.04215511632124076, "grad_norm": 0.322265625, "learning_rate": 9.802010050251256e-05, "loss": 1.5492, "step": 494 }, { "epoch": 0.04224045056480603, "grad_norm": 0.310546875, "learning_rate": 9.801507537688442e-05, "loss": 1.4367, "step": 495 }, { "epoch": 0.04232578480837129, "grad_norm": 0.31640625, "learning_rate": 9.801005025125629e-05, "loss": 1.3598, "step": 496 }, { "epoch": 0.042411119051936556, "grad_norm": 0.341796875, "learning_rate": 9.800502512562815e-05, "loss": 1.5329, "step": 497 }, { "epoch": 0.04249645329550182, "grad_norm": 0.33203125, "learning_rate": 9.8e-05, "loss": 1.3892, "step": 498 }, { "epoch": 0.042581787539067086, "grad_norm": 0.345703125, "learning_rate": 9.799497487437186e-05, "loss": 1.4771, "step": 499 }, { "epoch": 0.04266712178263235, "grad_norm": 0.318359375, "learning_rate": 9.798994974874372e-05, "loss": 1.4531, "step": 500 }, { "epoch": 0.042752456026197615, "grad_norm": 0.361328125, "learning_rate": 9.798492462311558e-05, "loss": 1.4213, "step": 501 }, { "epoch": 0.04283779026976288, "grad_norm": 0.333984375, "learning_rate": 9.797989949748743e-05, "loss": 1.445, "step": 502 }, { "epoch": 0.042923124513328144, "grad_norm": 0.36328125, "learning_rate": 9.797487437185931e-05, "loss": 1.4574, "step": 503 }, { "epoch": 0.04300845875689341, "grad_norm": 0.466796875, "learning_rate": 9.796984924623116e-05, "loss": 1.6703, "step": 504 }, { "epoch": 0.04309379300045867, "grad_norm": 0.39453125, "learning_rate": 9.796482412060302e-05, "loss": 1.3828, "step": 505 }, { "epoch": 0.04317912724402394, "grad_norm": 0.337890625, "learning_rate": 9.795979899497488e-05, "loss": 1.4373, "step": 506 }, { "epoch": 0.0432644614875892, "grad_norm": 0.376953125, "learning_rate": 9.795477386934674e-05, "loss": 1.5437, "step": 507 }, { "epoch": 0.04334979573115447, "grad_norm": 0.330078125, "learning_rate": 9.794974874371859e-05, "loss": 1.4256, "step": 508 }, { "epoch": 0.04343512997471973, "grad_norm": 0.365234375, "learning_rate": 9.794472361809047e-05, "loss": 1.3799, "step": 509 }, { "epoch": 0.043520464218284996, "grad_norm": 0.33984375, "learning_rate": 9.793969849246231e-05, "loss": 1.3046, "step": 510 }, { "epoch": 0.04360579846185026, "grad_norm": 0.35546875, "learning_rate": 9.793467336683418e-05, "loss": 1.4718, "step": 511 }, { "epoch": 0.043691132705415525, "grad_norm": 0.34765625, "learning_rate": 9.792964824120604e-05, "loss": 1.3132, "step": 512 }, { "epoch": 0.04377646694898079, "grad_norm": 0.3359375, "learning_rate": 9.79246231155779e-05, "loss": 1.4986, "step": 513 }, { "epoch": 0.043861801192546054, "grad_norm": 0.451171875, "learning_rate": 9.791959798994975e-05, "loss": 1.3882, "step": 514 }, { "epoch": 0.04394713543611132, "grad_norm": 0.3359375, "learning_rate": 9.791457286432161e-05, "loss": 1.4305, "step": 515 }, { "epoch": 0.044032469679676584, "grad_norm": 0.359375, "learning_rate": 9.790954773869347e-05, "loss": 1.2547, "step": 516 }, { "epoch": 0.04411780392324185, "grad_norm": 0.291015625, "learning_rate": 9.790452261306533e-05, "loss": 1.4184, "step": 517 }, { "epoch": 0.04420313816680711, "grad_norm": 0.318359375, "learning_rate": 9.789949748743718e-05, "loss": 1.3546, "step": 518 }, { "epoch": 0.04428847241037238, "grad_norm": 0.3359375, "learning_rate": 9.789447236180906e-05, "loss": 1.5497, "step": 519 }, { "epoch": 0.04437380665393764, "grad_norm": 0.326171875, "learning_rate": 9.788944723618091e-05, "loss": 1.3478, "step": 520 }, { "epoch": 0.044459140897502906, "grad_norm": 0.345703125, "learning_rate": 9.788442211055277e-05, "loss": 1.3972, "step": 521 }, { "epoch": 0.04454447514106817, "grad_norm": 0.314453125, "learning_rate": 9.787939698492463e-05, "loss": 1.2856, "step": 522 }, { "epoch": 0.044629809384633436, "grad_norm": 0.31640625, "learning_rate": 9.78743718592965e-05, "loss": 1.2654, "step": 523 }, { "epoch": 0.0447151436281987, "grad_norm": 0.34375, "learning_rate": 9.786934673366834e-05, "loss": 1.38, "step": 524 }, { "epoch": 0.044800477871763965, "grad_norm": 0.287109375, "learning_rate": 9.78643216080402e-05, "loss": 1.1772, "step": 525 }, { "epoch": 0.04488581211532923, "grad_norm": 0.330078125, "learning_rate": 9.785929648241207e-05, "loss": 1.368, "step": 526 }, { "epoch": 0.044971146358894494, "grad_norm": 0.306640625, "learning_rate": 9.785427135678393e-05, "loss": 1.3993, "step": 527 }, { "epoch": 0.04505648060245976, "grad_norm": 0.349609375, "learning_rate": 9.784924623115579e-05, "loss": 1.3716, "step": 528 }, { "epoch": 0.04514181484602502, "grad_norm": 0.345703125, "learning_rate": 9.784422110552764e-05, "loss": 1.4416, "step": 529 }, { "epoch": 0.04522714908959029, "grad_norm": 0.3515625, "learning_rate": 9.78391959798995e-05, "loss": 1.3523, "step": 530 }, { "epoch": 0.04531248333315555, "grad_norm": 0.33984375, "learning_rate": 9.783417085427136e-05, "loss": 1.373, "step": 531 }, { "epoch": 0.04539781757672082, "grad_norm": 0.30859375, "learning_rate": 9.782914572864322e-05, "loss": 1.3263, "step": 532 }, { "epoch": 0.04548315182028608, "grad_norm": 0.3515625, "learning_rate": 9.782412060301507e-05, "loss": 1.3304, "step": 533 }, { "epoch": 0.045568486063851346, "grad_norm": 0.33203125, "learning_rate": 9.781909547738694e-05, "loss": 1.4301, "step": 534 }, { "epoch": 0.04565382030741661, "grad_norm": 0.396484375, "learning_rate": 9.78140703517588e-05, "loss": 1.3432, "step": 535 }, { "epoch": 0.045739154550981875, "grad_norm": 0.318359375, "learning_rate": 9.780904522613066e-05, "loss": 1.3278, "step": 536 }, { "epoch": 0.04582448879454714, "grad_norm": 0.296875, "learning_rate": 9.780402010050251e-05, "loss": 1.3933, "step": 537 }, { "epoch": 0.045909823038112405, "grad_norm": 0.30859375, "learning_rate": 9.779899497487438e-05, "loss": 1.3747, "step": 538 }, { "epoch": 0.04599515728167767, "grad_norm": 0.349609375, "learning_rate": 9.779396984924623e-05, "loss": 1.4705, "step": 539 }, { "epoch": 0.046080491525242934, "grad_norm": 0.294921875, "learning_rate": 9.77889447236181e-05, "loss": 1.3608, "step": 540 }, { "epoch": 0.0461658257688082, "grad_norm": 0.2890625, "learning_rate": 9.778391959798996e-05, "loss": 1.3604, "step": 541 }, { "epoch": 0.04625116001237346, "grad_norm": 0.357421875, "learning_rate": 9.777889447236182e-05, "loss": 1.366, "step": 542 }, { "epoch": 0.04633649425593873, "grad_norm": 0.330078125, "learning_rate": 9.777386934673367e-05, "loss": 1.3359, "step": 543 }, { "epoch": 0.04642182849950399, "grad_norm": 0.330078125, "learning_rate": 9.776884422110554e-05, "loss": 1.3241, "step": 544 }, { "epoch": 0.04650716274306926, "grad_norm": 0.36328125, "learning_rate": 9.776381909547739e-05, "loss": 1.3774, "step": 545 }, { "epoch": 0.04659249698663452, "grad_norm": 0.330078125, "learning_rate": 9.775879396984925e-05, "loss": 1.4116, "step": 546 }, { "epoch": 0.046677831230199786, "grad_norm": 0.30859375, "learning_rate": 9.77537688442211e-05, "loss": 1.2552, "step": 547 }, { "epoch": 0.04676316547376505, "grad_norm": 0.3203125, "learning_rate": 9.774874371859298e-05, "loss": 1.3401, "step": 548 }, { "epoch": 0.046848499717330315, "grad_norm": 0.34765625, "learning_rate": 9.774371859296483e-05, "loss": 1.31, "step": 549 }, { "epoch": 0.04693383396089558, "grad_norm": 0.361328125, "learning_rate": 9.773869346733669e-05, "loss": 1.4162, "step": 550 }, { "epoch": 0.047019168204460844, "grad_norm": 0.279296875, "learning_rate": 9.773366834170855e-05, "loss": 1.3854, "step": 551 }, { "epoch": 0.047104502448026116, "grad_norm": 0.333984375, "learning_rate": 9.772864321608041e-05, "loss": 1.3167, "step": 552 }, { "epoch": 0.04718983669159138, "grad_norm": 0.310546875, "learning_rate": 9.772361809045226e-05, "loss": 1.4153, "step": 553 }, { "epoch": 0.047275170935156645, "grad_norm": 0.306640625, "learning_rate": 9.771859296482414e-05, "loss": 1.3265, "step": 554 }, { "epoch": 0.04736050517872191, "grad_norm": 0.333984375, "learning_rate": 9.771356783919598e-05, "loss": 1.4612, "step": 555 }, { "epoch": 0.047445839422287174, "grad_norm": 0.333984375, "learning_rate": 9.770854271356785e-05, "loss": 1.3782, "step": 556 }, { "epoch": 0.04753117366585244, "grad_norm": 0.326171875, "learning_rate": 9.770351758793971e-05, "loss": 1.3727, "step": 557 }, { "epoch": 0.0476165079094177, "grad_norm": 0.375, "learning_rate": 9.769849246231157e-05, "loss": 1.4461, "step": 558 }, { "epoch": 0.04770184215298297, "grad_norm": 0.32421875, "learning_rate": 9.769346733668342e-05, "loss": 1.1829, "step": 559 }, { "epoch": 0.04778717639654823, "grad_norm": 0.302734375, "learning_rate": 9.768844221105528e-05, "loss": 1.3462, "step": 560 }, { "epoch": 0.0478725106401135, "grad_norm": 0.330078125, "learning_rate": 9.768341708542714e-05, "loss": 1.4409, "step": 561 }, { "epoch": 0.04795784488367876, "grad_norm": 0.328125, "learning_rate": 9.7678391959799e-05, "loss": 1.4291, "step": 562 }, { "epoch": 0.048043179127244026, "grad_norm": 0.345703125, "learning_rate": 9.767336683417085e-05, "loss": 1.2478, "step": 563 }, { "epoch": 0.04812851337080929, "grad_norm": 0.328125, "learning_rate": 9.766834170854271e-05, "loss": 1.5019, "step": 564 }, { "epoch": 0.048213847614374555, "grad_norm": 0.314453125, "learning_rate": 9.766331658291458e-05, "loss": 1.3371, "step": 565 }, { "epoch": 0.04829918185793982, "grad_norm": 0.42578125, "learning_rate": 9.765829145728644e-05, "loss": 1.7739, "step": 566 }, { "epoch": 0.048384516101505085, "grad_norm": 0.3359375, "learning_rate": 9.76532663316583e-05, "loss": 1.431, "step": 567 }, { "epoch": 0.04846985034507035, "grad_norm": 0.298828125, "learning_rate": 9.764824120603015e-05, "loss": 1.3027, "step": 568 }, { "epoch": 0.048555184588635614, "grad_norm": 0.345703125, "learning_rate": 9.764321608040201e-05, "loss": 1.4325, "step": 569 }, { "epoch": 0.04864051883220088, "grad_norm": 0.3515625, "learning_rate": 9.763819095477387e-05, "loss": 1.4314, "step": 570 }, { "epoch": 0.04872585307576614, "grad_norm": 0.34375, "learning_rate": 9.763316582914574e-05, "loss": 1.438, "step": 571 }, { "epoch": 0.04881118731933141, "grad_norm": 0.3125, "learning_rate": 9.762814070351758e-05, "loss": 1.2522, "step": 572 }, { "epoch": 0.04889652156289667, "grad_norm": 0.36328125, "learning_rate": 9.762311557788946e-05, "loss": 1.4434, "step": 573 }, { "epoch": 0.04898185580646194, "grad_norm": 0.373046875, "learning_rate": 9.761809045226131e-05, "loss": 1.4041, "step": 574 }, { "epoch": 0.0490671900500272, "grad_norm": 0.388671875, "learning_rate": 9.761306532663317e-05, "loss": 1.3432, "step": 575 }, { "epoch": 0.049152524293592466, "grad_norm": 0.302734375, "learning_rate": 9.760804020100502e-05, "loss": 1.2897, "step": 576 }, { "epoch": 0.04923785853715773, "grad_norm": 0.34765625, "learning_rate": 9.76030150753769e-05, "loss": 1.4813, "step": 577 }, { "epoch": 0.049323192780722995, "grad_norm": 0.369140625, "learning_rate": 9.759798994974874e-05, "loss": 1.6552, "step": 578 }, { "epoch": 0.04940852702428826, "grad_norm": 0.30859375, "learning_rate": 9.75929648241206e-05, "loss": 1.3383, "step": 579 }, { "epoch": 0.049493861267853524, "grad_norm": 0.328125, "learning_rate": 9.758793969849247e-05, "loss": 1.4323, "step": 580 }, { "epoch": 0.04957919551141879, "grad_norm": 0.328125, "learning_rate": 9.758291457286433e-05, "loss": 1.3754, "step": 581 }, { "epoch": 0.04966452975498405, "grad_norm": 0.30859375, "learning_rate": 9.757788944723618e-05, "loss": 1.2542, "step": 582 }, { "epoch": 0.04974986399854932, "grad_norm": 0.333984375, "learning_rate": 9.757286432160805e-05, "loss": 1.3337, "step": 583 }, { "epoch": 0.04983519824211458, "grad_norm": 0.34375, "learning_rate": 9.75678391959799e-05, "loss": 1.5184, "step": 584 }, { "epoch": 0.04992053248567985, "grad_norm": 0.357421875, "learning_rate": 9.756281407035176e-05, "loss": 1.3454, "step": 585 }, { "epoch": 0.05000586672924511, "grad_norm": 0.271484375, "learning_rate": 9.755778894472363e-05, "loss": 1.5811, "step": 586 }, { "epoch": 0.050091200972810376, "grad_norm": 0.310546875, "learning_rate": 9.755276381909549e-05, "loss": 1.3913, "step": 587 }, { "epoch": 0.05017653521637564, "grad_norm": 0.30078125, "learning_rate": 9.754773869346734e-05, "loss": 1.2978, "step": 588 }, { "epoch": 0.050261869459940905, "grad_norm": 0.3359375, "learning_rate": 9.754271356783921e-05, "loss": 1.4808, "step": 589 }, { "epoch": 0.05034720370350617, "grad_norm": 0.388671875, "learning_rate": 9.753768844221106e-05, "loss": 1.3697, "step": 590 }, { "epoch": 0.050432537947071435, "grad_norm": 0.310546875, "learning_rate": 9.753266331658292e-05, "loss": 1.428, "step": 591 }, { "epoch": 0.0505178721906367, "grad_norm": 0.341796875, "learning_rate": 9.752763819095477e-05, "loss": 1.3562, "step": 592 }, { "epoch": 0.050603206434201964, "grad_norm": 0.314453125, "learning_rate": 9.752261306532665e-05, "loss": 1.1277, "step": 593 }, { "epoch": 0.05068854067776723, "grad_norm": 0.32421875, "learning_rate": 9.75175879396985e-05, "loss": 1.2116, "step": 594 }, { "epoch": 0.05077387492133249, "grad_norm": 0.35546875, "learning_rate": 9.751256281407036e-05, "loss": 1.4622, "step": 595 }, { "epoch": 0.05085920916489776, "grad_norm": 0.28515625, "learning_rate": 9.750753768844222e-05, "loss": 1.3907, "step": 596 }, { "epoch": 0.05094454340846302, "grad_norm": 0.35546875, "learning_rate": 9.750251256281408e-05, "loss": 1.5221, "step": 597 }, { "epoch": 0.05102987765202829, "grad_norm": 0.306640625, "learning_rate": 9.749748743718593e-05, "loss": 1.318, "step": 598 }, { "epoch": 0.05111521189559355, "grad_norm": 0.31640625, "learning_rate": 9.749246231155779e-05, "loss": 1.4771, "step": 599 }, { "epoch": 0.051200546139158816, "grad_norm": 0.322265625, "learning_rate": 9.748743718592965e-05, "loss": 1.2633, "step": 600 }, { "epoch": 0.05128588038272408, "grad_norm": 0.337890625, "learning_rate": 9.748241206030152e-05, "loss": 1.4679, "step": 601 }, { "epoch": 0.051371214626289345, "grad_norm": 0.328125, "learning_rate": 9.747738693467338e-05, "loss": 1.4968, "step": 602 }, { "epoch": 0.05145654886985461, "grad_norm": 0.32421875, "learning_rate": 9.747236180904523e-05, "loss": 1.3159, "step": 603 }, { "epoch": 0.051541883113419874, "grad_norm": 0.294921875, "learning_rate": 9.746733668341709e-05, "loss": 1.3781, "step": 604 }, { "epoch": 0.05162721735698514, "grad_norm": 0.314453125, "learning_rate": 9.746231155778895e-05, "loss": 1.4195, "step": 605 }, { "epoch": 0.051712551600550404, "grad_norm": 0.314453125, "learning_rate": 9.745728643216081e-05, "loss": 1.4356, "step": 606 }, { "epoch": 0.05179788584411567, "grad_norm": 0.357421875, "learning_rate": 9.745226130653266e-05, "loss": 1.3743, "step": 607 }, { "epoch": 0.05188322008768093, "grad_norm": 0.310546875, "learning_rate": 9.744723618090452e-05, "loss": 1.3155, "step": 608 }, { "epoch": 0.0519685543312462, "grad_norm": 0.337890625, "learning_rate": 9.744221105527638e-05, "loss": 1.4083, "step": 609 }, { "epoch": 0.05205388857481146, "grad_norm": 0.30859375, "learning_rate": 9.743718592964825e-05, "loss": 1.3564, "step": 610 }, { "epoch": 0.052139222818376726, "grad_norm": 0.333984375, "learning_rate": 9.74321608040201e-05, "loss": 1.5243, "step": 611 }, { "epoch": 0.05222455706194199, "grad_norm": 0.298828125, "learning_rate": 9.742713567839197e-05, "loss": 1.3934, "step": 612 }, { "epoch": 0.052309891305507256, "grad_norm": 0.330078125, "learning_rate": 9.742211055276382e-05, "loss": 1.3353, "step": 613 }, { "epoch": 0.05239522554907252, "grad_norm": 0.32421875, "learning_rate": 9.741708542713568e-05, "loss": 1.353, "step": 614 }, { "epoch": 0.052480559792637785, "grad_norm": 0.29296875, "learning_rate": 9.741206030150754e-05, "loss": 1.3131, "step": 615 }, { "epoch": 0.05256589403620305, "grad_norm": 0.333984375, "learning_rate": 9.74070351758794e-05, "loss": 1.3904, "step": 616 }, { "epoch": 0.05265122827976832, "grad_norm": 0.3359375, "learning_rate": 9.740201005025125e-05, "loss": 1.3883, "step": 617 }, { "epoch": 0.052736562523333586, "grad_norm": 0.3046875, "learning_rate": 9.739698492462313e-05, "loss": 1.3321, "step": 618 }, { "epoch": 0.05282189676689885, "grad_norm": 0.30859375, "learning_rate": 9.739195979899498e-05, "loss": 1.4628, "step": 619 }, { "epoch": 0.052907231010464115, "grad_norm": 0.30078125, "learning_rate": 9.738693467336684e-05, "loss": 1.381, "step": 620 }, { "epoch": 0.05299256525402938, "grad_norm": 0.38671875, "learning_rate": 9.73819095477387e-05, "loss": 1.3673, "step": 621 }, { "epoch": 0.053077899497594644, "grad_norm": 0.341796875, "learning_rate": 9.737688442211056e-05, "loss": 1.3133, "step": 622 }, { "epoch": 0.05316323374115991, "grad_norm": 0.3046875, "learning_rate": 9.737185929648241e-05, "loss": 1.5619, "step": 623 }, { "epoch": 0.05324856798472517, "grad_norm": 0.3046875, "learning_rate": 9.736683417085427e-05, "loss": 1.3775, "step": 624 }, { "epoch": 0.05333390222829044, "grad_norm": 0.330078125, "learning_rate": 9.736180904522614e-05, "loss": 1.2801, "step": 625 }, { "epoch": 0.0534192364718557, "grad_norm": 0.3359375, "learning_rate": 9.7356783919598e-05, "loss": 1.2728, "step": 626 }, { "epoch": 0.05350457071542097, "grad_norm": 0.32421875, "learning_rate": 9.735175879396985e-05, "loss": 1.3673, "step": 627 }, { "epoch": 0.05358990495898623, "grad_norm": 0.353515625, "learning_rate": 9.734673366834172e-05, "loss": 1.5111, "step": 628 }, { "epoch": 0.053675239202551496, "grad_norm": 0.34375, "learning_rate": 9.734170854271357e-05, "loss": 1.3328, "step": 629 }, { "epoch": 0.05376057344611676, "grad_norm": 0.28125, "learning_rate": 9.733668341708543e-05, "loss": 1.3164, "step": 630 }, { "epoch": 0.053845907689682025, "grad_norm": 0.314453125, "learning_rate": 9.73316582914573e-05, "loss": 1.4929, "step": 631 }, { "epoch": 0.05393124193324729, "grad_norm": 0.359375, "learning_rate": 9.732663316582916e-05, "loss": 1.1437, "step": 632 }, { "epoch": 0.054016576176812554, "grad_norm": 0.3515625, "learning_rate": 9.7321608040201e-05, "loss": 1.4573, "step": 633 }, { "epoch": 0.05410191042037782, "grad_norm": 0.345703125, "learning_rate": 9.731658291457287e-05, "loss": 1.5108, "step": 634 }, { "epoch": 0.054187244663943084, "grad_norm": 0.349609375, "learning_rate": 9.731155778894473e-05, "loss": 1.3038, "step": 635 }, { "epoch": 0.05427257890750835, "grad_norm": 0.318359375, "learning_rate": 9.730653266331659e-05, "loss": 1.2705, "step": 636 }, { "epoch": 0.05435791315107361, "grad_norm": 0.361328125, "learning_rate": 9.730150753768845e-05, "loss": 1.3002, "step": 637 }, { "epoch": 0.05444324739463888, "grad_norm": 0.34765625, "learning_rate": 9.72964824120603e-05, "loss": 1.2631, "step": 638 }, { "epoch": 0.05452858163820414, "grad_norm": 0.3359375, "learning_rate": 9.729145728643216e-05, "loss": 1.4492, "step": 639 }, { "epoch": 0.054613915881769406, "grad_norm": 0.3515625, "learning_rate": 9.728643216080403e-05, "loss": 1.2772, "step": 640 }, { "epoch": 0.05469925012533467, "grad_norm": 0.31640625, "learning_rate": 9.728140703517589e-05, "loss": 1.2737, "step": 641 }, { "epoch": 0.054784584368899936, "grad_norm": 0.3125, "learning_rate": 9.727638190954774e-05, "loss": 1.3252, "step": 642 }, { "epoch": 0.0548699186124652, "grad_norm": 0.326171875, "learning_rate": 9.72713567839196e-05, "loss": 1.5284, "step": 643 }, { "epoch": 0.054955252856030465, "grad_norm": 0.345703125, "learning_rate": 9.726633165829146e-05, "loss": 1.2824, "step": 644 }, { "epoch": 0.05504058709959573, "grad_norm": 0.41015625, "learning_rate": 9.726130653266332e-05, "loss": 1.519, "step": 645 }, { "epoch": 0.055125921343160994, "grad_norm": 0.333984375, "learning_rate": 9.725628140703517e-05, "loss": 1.4557, "step": 646 }, { "epoch": 0.05521125558672626, "grad_norm": 0.376953125, "learning_rate": 9.725125628140705e-05, "loss": 1.2523, "step": 647 }, { "epoch": 0.05529658983029152, "grad_norm": 0.32421875, "learning_rate": 9.72462311557789e-05, "loss": 1.2909, "step": 648 }, { "epoch": 0.05538192407385679, "grad_norm": 0.314453125, "learning_rate": 9.724120603015076e-05, "loss": 1.4421, "step": 649 }, { "epoch": 0.05546725831742205, "grad_norm": 0.322265625, "learning_rate": 9.723618090452262e-05, "loss": 1.4081, "step": 650 }, { "epoch": 0.05555259256098732, "grad_norm": 0.353515625, "learning_rate": 9.723115577889448e-05, "loss": 1.3829, "step": 651 }, { "epoch": 0.05563792680455258, "grad_norm": 0.345703125, "learning_rate": 9.722613065326633e-05, "loss": 1.4267, "step": 652 }, { "epoch": 0.055723261048117846, "grad_norm": 0.337890625, "learning_rate": 9.72211055276382e-05, "loss": 1.4874, "step": 653 }, { "epoch": 0.05580859529168311, "grad_norm": 0.326171875, "learning_rate": 9.721608040201005e-05, "loss": 1.31, "step": 654 }, { "epoch": 0.055893929535248375, "grad_norm": 0.322265625, "learning_rate": 9.721105527638192e-05, "loss": 1.3142, "step": 655 }, { "epoch": 0.05597926377881364, "grad_norm": 0.3515625, "learning_rate": 9.720603015075376e-05, "loss": 1.2545, "step": 656 }, { "epoch": 0.056064598022378904, "grad_norm": 0.376953125, "learning_rate": 9.720100502512564e-05, "loss": 1.3988, "step": 657 }, { "epoch": 0.05614993226594417, "grad_norm": 0.38671875, "learning_rate": 9.719597989949749e-05, "loss": 1.4185, "step": 658 }, { "epoch": 0.056235266509509434, "grad_norm": 0.31640625, "learning_rate": 9.719095477386935e-05, "loss": 2.0381, "step": 659 }, { "epoch": 0.0563206007530747, "grad_norm": 0.373046875, "learning_rate": 9.718592964824121e-05, "loss": 1.5739, "step": 660 }, { "epoch": 0.05640593499663996, "grad_norm": 0.29296875, "learning_rate": 9.718090452261307e-05, "loss": 1.3885, "step": 661 }, { "epoch": 0.05649126924020523, "grad_norm": 0.3046875, "learning_rate": 9.717587939698492e-05, "loss": 1.3189, "step": 662 }, { "epoch": 0.05657660348377049, "grad_norm": 0.326171875, "learning_rate": 9.71708542713568e-05, "loss": 1.42, "step": 663 }, { "epoch": 0.05666193772733576, "grad_norm": 0.306640625, "learning_rate": 9.716582914572865e-05, "loss": 1.2109, "step": 664 }, { "epoch": 0.05674727197090102, "grad_norm": 0.30859375, "learning_rate": 9.716080402010051e-05, "loss": 1.3912, "step": 665 }, { "epoch": 0.056832606214466286, "grad_norm": 0.29296875, "learning_rate": 9.715577889447237e-05, "loss": 1.3246, "step": 666 }, { "epoch": 0.05691794045803155, "grad_norm": 0.318359375, "learning_rate": 9.715075376884423e-05, "loss": 1.4655, "step": 667 }, { "epoch": 0.057003274701596815, "grad_norm": 0.345703125, "learning_rate": 9.714572864321608e-05, "loss": 1.3685, "step": 668 }, { "epoch": 0.05708860894516208, "grad_norm": 0.330078125, "learning_rate": 9.714070351758794e-05, "loss": 1.4323, "step": 669 }, { "epoch": 0.057173943188727344, "grad_norm": 0.412109375, "learning_rate": 9.71356783919598e-05, "loss": 1.3601, "step": 670 }, { "epoch": 0.05725927743229261, "grad_norm": 0.3359375, "learning_rate": 9.713065326633167e-05, "loss": 1.4362, "step": 671 }, { "epoch": 0.05734461167585787, "grad_norm": 0.330078125, "learning_rate": 9.712562814070352e-05, "loss": 2.746, "step": 672 }, { "epoch": 0.05742994591942314, "grad_norm": 0.314453125, "learning_rate": 9.712060301507538e-05, "loss": 1.3721, "step": 673 }, { "epoch": 0.0575152801629884, "grad_norm": 0.2890625, "learning_rate": 9.711557788944724e-05, "loss": 1.2796, "step": 674 }, { "epoch": 0.05760061440655367, "grad_norm": 0.322265625, "learning_rate": 9.71105527638191e-05, "loss": 1.5044, "step": 675 }, { "epoch": 0.05768594865011893, "grad_norm": 0.328125, "learning_rate": 9.710552763819096e-05, "loss": 1.2025, "step": 676 }, { "epoch": 0.057771282893684196, "grad_norm": 0.46484375, "learning_rate": 9.710050251256281e-05, "loss": 1.5289, "step": 677 }, { "epoch": 0.05785661713724946, "grad_norm": 0.3203125, "learning_rate": 9.709547738693468e-05, "loss": 1.413, "step": 678 }, { "epoch": 0.057941951380814725, "grad_norm": 0.33203125, "learning_rate": 9.709045226130654e-05, "loss": 1.4112, "step": 679 }, { "epoch": 0.05802728562437999, "grad_norm": 0.28515625, "learning_rate": 9.70854271356784e-05, "loss": 1.3151, "step": 680 }, { "epoch": 0.058112619867945255, "grad_norm": 0.37109375, "learning_rate": 9.708040201005025e-05, "loss": 1.3537, "step": 681 }, { "epoch": 0.058197954111510526, "grad_norm": 0.34765625, "learning_rate": 9.707537688442212e-05, "loss": 1.3003, "step": 682 }, { "epoch": 0.05828328835507579, "grad_norm": 0.333984375, "learning_rate": 9.707035175879397e-05, "loss": 1.3926, "step": 683 }, { "epoch": 0.058368622598641055, "grad_norm": 0.310546875, "learning_rate": 9.706532663316583e-05, "loss": 1.3556, "step": 684 }, { "epoch": 0.05845395684220632, "grad_norm": 0.306640625, "learning_rate": 9.706030150753768e-05, "loss": 1.2722, "step": 685 }, { "epoch": 0.058539291085771585, "grad_norm": 0.34765625, "learning_rate": 9.705527638190956e-05, "loss": 1.3896, "step": 686 }, { "epoch": 0.05862462532933685, "grad_norm": 0.333984375, "learning_rate": 9.70502512562814e-05, "loss": 1.2862, "step": 687 }, { "epoch": 0.058709959572902114, "grad_norm": 0.322265625, "learning_rate": 9.704522613065327e-05, "loss": 1.46, "step": 688 }, { "epoch": 0.05879529381646738, "grad_norm": 0.31640625, "learning_rate": 9.704020100502513e-05, "loss": 1.4346, "step": 689 }, { "epoch": 0.05888062806003264, "grad_norm": 0.3671875, "learning_rate": 9.703517587939699e-05, "loss": 1.3685, "step": 690 }, { "epoch": 0.05896596230359791, "grad_norm": 0.33984375, "learning_rate": 9.703015075376884e-05, "loss": 1.3749, "step": 691 }, { "epoch": 0.05905129654716317, "grad_norm": 0.3046875, "learning_rate": 9.702512562814072e-05, "loss": 1.1969, "step": 692 }, { "epoch": 0.05913663079072844, "grad_norm": 0.35546875, "learning_rate": 9.702010050251257e-05, "loss": 1.3639, "step": 693 }, { "epoch": 0.0592219650342937, "grad_norm": 0.34765625, "learning_rate": 9.701507537688443e-05, "loss": 1.5146, "step": 694 }, { "epoch": 0.059307299277858966, "grad_norm": 0.314453125, "learning_rate": 9.701005025125629e-05, "loss": 1.3232, "step": 695 }, { "epoch": 0.05939263352142423, "grad_norm": 0.322265625, "learning_rate": 9.700502512562815e-05, "loss": 1.2522, "step": 696 }, { "epoch": 0.059477967764989495, "grad_norm": 0.28515625, "learning_rate": 9.7e-05, "loss": 1.4553, "step": 697 }, { "epoch": 0.05956330200855476, "grad_norm": 0.310546875, "learning_rate": 9.699497487437188e-05, "loss": 1.0855, "step": 698 }, { "epoch": 0.059648636252120024, "grad_norm": 0.28125, "learning_rate": 9.698994974874372e-05, "loss": 1.3087, "step": 699 }, { "epoch": 0.05973397049568529, "grad_norm": 0.369140625, "learning_rate": 9.698492462311559e-05, "loss": 1.4624, "step": 700 }, { "epoch": 0.05981930473925055, "grad_norm": 0.35546875, "learning_rate": 9.697989949748743e-05, "loss": 1.4732, "step": 701 }, { "epoch": 0.05990463898281582, "grad_norm": 0.3125, "learning_rate": 9.697487437185931e-05, "loss": 1.3887, "step": 702 }, { "epoch": 0.05998997322638108, "grad_norm": 0.310546875, "learning_rate": 9.696984924623116e-05, "loss": 1.4674, "step": 703 }, { "epoch": 0.06007530746994635, "grad_norm": 0.322265625, "learning_rate": 9.696482412060302e-05, "loss": 1.3927, "step": 704 }, { "epoch": 0.06016064171351161, "grad_norm": 0.298828125, "learning_rate": 9.695979899497488e-05, "loss": 1.2565, "step": 705 }, { "epoch": 0.060245975957076876, "grad_norm": 0.302734375, "learning_rate": 9.695477386934674e-05, "loss": 1.2928, "step": 706 }, { "epoch": 0.06033131020064214, "grad_norm": 0.322265625, "learning_rate": 9.694974874371859e-05, "loss": 1.3397, "step": 707 }, { "epoch": 0.060416644444207405, "grad_norm": 0.322265625, "learning_rate": 9.694472361809045e-05, "loss": 1.3858, "step": 708 }, { "epoch": 0.06050197868777267, "grad_norm": 0.318359375, "learning_rate": 9.693969849246232e-05, "loss": 1.4811, "step": 709 }, { "epoch": 0.060587312931337935, "grad_norm": 0.322265625, "learning_rate": 9.693467336683418e-05, "loss": 1.4033, "step": 710 }, { "epoch": 0.0606726471749032, "grad_norm": 0.330078125, "learning_rate": 9.692964824120604e-05, "loss": 1.46, "step": 711 }, { "epoch": 0.060757981418468464, "grad_norm": 0.34765625, "learning_rate": 9.692462311557789e-05, "loss": 1.3448, "step": 712 }, { "epoch": 0.06084331566203373, "grad_norm": 0.326171875, "learning_rate": 9.691959798994975e-05, "loss": 1.4918, "step": 713 }, { "epoch": 0.06092864990559899, "grad_norm": 0.30078125, "learning_rate": 9.691457286432161e-05, "loss": 1.2975, "step": 714 }, { "epoch": 0.06101398414916426, "grad_norm": 0.34765625, "learning_rate": 9.690954773869348e-05, "loss": 1.3874, "step": 715 }, { "epoch": 0.06109931839272952, "grad_norm": 0.365234375, "learning_rate": 9.690452261306532e-05, "loss": 1.4599, "step": 716 }, { "epoch": 0.06118465263629479, "grad_norm": 0.302734375, "learning_rate": 9.689949748743719e-05, "loss": 1.3242, "step": 717 }, { "epoch": 0.06126998687986005, "grad_norm": 0.32421875, "learning_rate": 9.689447236180905e-05, "loss": 1.4184, "step": 718 }, { "epoch": 0.061355321123425316, "grad_norm": 0.337890625, "learning_rate": 9.688944723618091e-05, "loss": 1.5936, "step": 719 }, { "epoch": 0.06144065536699058, "grad_norm": 0.2734375, "learning_rate": 9.688442211055276e-05, "loss": 1.1906, "step": 720 }, { "epoch": 0.061525989610555845, "grad_norm": 0.294921875, "learning_rate": 9.687939698492463e-05, "loss": 1.2857, "step": 721 }, { "epoch": 0.06161132385412111, "grad_norm": 0.298828125, "learning_rate": 9.687437185929648e-05, "loss": 1.2152, "step": 722 }, { "epoch": 0.061696658097686374, "grad_norm": 0.3046875, "learning_rate": 9.686934673366834e-05, "loss": 1.3036, "step": 723 }, { "epoch": 0.06178199234125164, "grad_norm": 0.294921875, "learning_rate": 9.686432160804021e-05, "loss": 1.3545, "step": 724 }, { "epoch": 0.061867326584816904, "grad_norm": 0.353515625, "learning_rate": 9.685929648241207e-05, "loss": 1.6225, "step": 725 }, { "epoch": 0.06195266082838217, "grad_norm": 0.30859375, "learning_rate": 9.685427135678392e-05, "loss": 1.1705, "step": 726 }, { "epoch": 0.06203799507194743, "grad_norm": 0.298828125, "learning_rate": 9.684924623115579e-05, "loss": 1.2082, "step": 727 }, { "epoch": 0.0621233293155127, "grad_norm": 0.318359375, "learning_rate": 9.684422110552764e-05, "loss": 1.2985, "step": 728 }, { "epoch": 0.06220866355907796, "grad_norm": 0.306640625, "learning_rate": 9.68391959798995e-05, "loss": 1.2335, "step": 729 }, { "epoch": 0.062293997802643226, "grad_norm": 0.3125, "learning_rate": 9.683417085427137e-05, "loss": 1.2785, "step": 730 }, { "epoch": 0.06237933204620849, "grad_norm": 0.380859375, "learning_rate": 9.682914572864323e-05, "loss": 1.4843, "step": 731 }, { "epoch": 0.062464666289773756, "grad_norm": 0.333984375, "learning_rate": 9.682412060301508e-05, "loss": 1.2533, "step": 732 }, { "epoch": 0.06255000053333902, "grad_norm": 0.337890625, "learning_rate": 9.681909547738694e-05, "loss": 1.4767, "step": 733 }, { "epoch": 0.06263533477690429, "grad_norm": 0.373046875, "learning_rate": 9.68140703517588e-05, "loss": 1.2843, "step": 734 }, { "epoch": 0.06272066902046955, "grad_norm": 0.32421875, "learning_rate": 9.680904522613066e-05, "loss": 1.514, "step": 735 }, { "epoch": 0.06280600326403482, "grad_norm": 0.32421875, "learning_rate": 9.680402010050251e-05, "loss": 1.4528, "step": 736 }, { "epoch": 0.06289133750760008, "grad_norm": 0.3046875, "learning_rate": 9.679899497487439e-05, "loss": 1.3571, "step": 737 }, { "epoch": 0.06297667175116535, "grad_norm": 0.35546875, "learning_rate": 9.679396984924623e-05, "loss": 1.2832, "step": 738 }, { "epoch": 0.06306200599473061, "grad_norm": 0.3203125, "learning_rate": 9.67889447236181e-05, "loss": 1.4314, "step": 739 }, { "epoch": 0.06314734023829588, "grad_norm": 0.38671875, "learning_rate": 9.678391959798996e-05, "loss": 1.5038, "step": 740 }, { "epoch": 0.06323267448186114, "grad_norm": 0.314453125, "learning_rate": 9.677889447236182e-05, "loss": 1.2716, "step": 741 }, { "epoch": 0.06331800872542641, "grad_norm": 0.328125, "learning_rate": 9.677386934673367e-05, "loss": 1.2287, "step": 742 }, { "epoch": 0.06340334296899167, "grad_norm": 0.326171875, "learning_rate": 9.676884422110553e-05, "loss": 1.3997, "step": 743 }, { "epoch": 0.06348867721255694, "grad_norm": 0.34765625, "learning_rate": 9.67638190954774e-05, "loss": 1.5631, "step": 744 }, { "epoch": 0.0635740114561222, "grad_norm": 0.33984375, "learning_rate": 9.675879396984924e-05, "loss": 1.4241, "step": 745 }, { "epoch": 0.06365934569968747, "grad_norm": 0.345703125, "learning_rate": 9.675376884422112e-05, "loss": 1.2748, "step": 746 }, { "epoch": 0.06374467994325272, "grad_norm": 0.361328125, "learning_rate": 9.674874371859297e-05, "loss": 1.3819, "step": 747 }, { "epoch": 0.063830014186818, "grad_norm": 0.35546875, "learning_rate": 9.674371859296483e-05, "loss": 1.3393, "step": 748 }, { "epoch": 0.06391534843038325, "grad_norm": 0.333984375, "learning_rate": 9.673869346733668e-05, "loss": 1.3685, "step": 749 }, { "epoch": 0.06400068267394853, "grad_norm": 0.34375, "learning_rate": 9.673366834170855e-05, "loss": 1.3799, "step": 750 }, { "epoch": 0.06408601691751378, "grad_norm": 0.314453125, "learning_rate": 9.67286432160804e-05, "loss": 1.265, "step": 751 }, { "epoch": 0.06417135116107905, "grad_norm": 0.37109375, "learning_rate": 9.672361809045226e-05, "loss": 1.5527, "step": 752 }, { "epoch": 0.06425668540464431, "grad_norm": 0.32421875, "learning_rate": 9.671859296482412e-05, "loss": 1.3506, "step": 753 }, { "epoch": 0.06434201964820958, "grad_norm": 0.33984375, "learning_rate": 9.671356783919599e-05, "loss": 1.4867, "step": 754 }, { "epoch": 0.06442735389177484, "grad_norm": 0.3671875, "learning_rate": 9.670854271356784e-05, "loss": 1.3884, "step": 755 }, { "epoch": 0.06451268813534011, "grad_norm": 0.353515625, "learning_rate": 9.670351758793971e-05, "loss": 1.3691, "step": 756 }, { "epoch": 0.06459802237890537, "grad_norm": 0.34375, "learning_rate": 9.669849246231156e-05, "loss": 1.3279, "step": 757 }, { "epoch": 0.06468335662247064, "grad_norm": 0.34375, "learning_rate": 9.669346733668342e-05, "loss": 1.3145, "step": 758 }, { "epoch": 0.0647686908660359, "grad_norm": 0.494140625, "learning_rate": 9.668844221105528e-05, "loss": 1.4105, "step": 759 }, { "epoch": 0.06485402510960117, "grad_norm": 0.412109375, "learning_rate": 9.668341708542715e-05, "loss": 1.3765, "step": 760 }, { "epoch": 0.06493935935316643, "grad_norm": 0.349609375, "learning_rate": 9.6678391959799e-05, "loss": 1.3581, "step": 761 }, { "epoch": 0.0650246935967317, "grad_norm": 0.330078125, "learning_rate": 9.667336683417087e-05, "loss": 1.3577, "step": 762 }, { "epoch": 0.06511002784029696, "grad_norm": 0.31640625, "learning_rate": 9.666834170854272e-05, "loss": 1.3019, "step": 763 }, { "epoch": 0.06519536208386223, "grad_norm": 0.32421875, "learning_rate": 9.666331658291458e-05, "loss": 1.4107, "step": 764 }, { "epoch": 0.06528069632742749, "grad_norm": 0.30078125, "learning_rate": 9.665829145728643e-05, "loss": 1.2924, "step": 765 }, { "epoch": 0.06536603057099276, "grad_norm": 0.33203125, "learning_rate": 9.66532663316583e-05, "loss": 1.3826, "step": 766 }, { "epoch": 0.06545136481455802, "grad_norm": 0.318359375, "learning_rate": 9.664824120603015e-05, "loss": 1.2879, "step": 767 }, { "epoch": 0.06553669905812329, "grad_norm": 0.33984375, "learning_rate": 9.664321608040201e-05, "loss": 1.4199, "step": 768 }, { "epoch": 0.06562203330168855, "grad_norm": 0.298828125, "learning_rate": 9.663819095477388e-05, "loss": 1.342, "step": 769 }, { "epoch": 0.06570736754525382, "grad_norm": 0.314453125, "learning_rate": 9.663316582914574e-05, "loss": 1.3338, "step": 770 }, { "epoch": 0.06579270178881907, "grad_norm": 0.306640625, "learning_rate": 9.662814070351759e-05, "loss": 1.3405, "step": 771 }, { "epoch": 0.06587803603238435, "grad_norm": 0.337890625, "learning_rate": 9.662311557788946e-05, "loss": 1.4719, "step": 772 }, { "epoch": 0.0659633702759496, "grad_norm": 0.31640625, "learning_rate": 9.661809045226131e-05, "loss": 1.3346, "step": 773 }, { "epoch": 0.06604870451951488, "grad_norm": 0.3046875, "learning_rate": 9.661306532663317e-05, "loss": 1.3901, "step": 774 }, { "epoch": 0.06613403876308013, "grad_norm": 0.3125, "learning_rate": 9.660804020100504e-05, "loss": 1.3265, "step": 775 }, { "epoch": 0.0662193730066454, "grad_norm": 0.3046875, "learning_rate": 9.66030150753769e-05, "loss": 1.2461, "step": 776 }, { "epoch": 0.06630470725021066, "grad_norm": 0.31640625, "learning_rate": 9.659798994974875e-05, "loss": 1.3661, "step": 777 }, { "epoch": 0.06639004149377593, "grad_norm": 0.35546875, "learning_rate": 9.659296482412061e-05, "loss": 1.4615, "step": 778 }, { "epoch": 0.0664753757373412, "grad_norm": 0.296875, "learning_rate": 9.658793969849247e-05, "loss": 1.3211, "step": 779 }, { "epoch": 0.06656070998090646, "grad_norm": 0.310546875, "learning_rate": 9.658291457286432e-05, "loss": 1.3174, "step": 780 }, { "epoch": 0.06664604422447173, "grad_norm": 0.333984375, "learning_rate": 9.657788944723618e-05, "loss": 1.3867, "step": 781 }, { "epoch": 0.06673137846803699, "grad_norm": 0.310546875, "learning_rate": 9.657286432160804e-05, "loss": 1.318, "step": 782 }, { "epoch": 0.06681671271160226, "grad_norm": 0.349609375, "learning_rate": 9.65678391959799e-05, "loss": 1.449, "step": 783 }, { "epoch": 0.06690204695516752, "grad_norm": 0.3828125, "learning_rate": 9.656281407035175e-05, "loss": 1.5028, "step": 784 }, { "epoch": 0.06698738119873279, "grad_norm": 0.28125, "learning_rate": 9.655778894472363e-05, "loss": 1.2669, "step": 785 }, { "epoch": 0.06707271544229805, "grad_norm": 0.32421875, "learning_rate": 9.655276381909548e-05, "loss": 1.3624, "step": 786 }, { "epoch": 0.06715804968586332, "grad_norm": 0.271484375, "learning_rate": 9.654773869346734e-05, "loss": 1.1874, "step": 787 }, { "epoch": 0.06724338392942858, "grad_norm": 0.306640625, "learning_rate": 9.65427135678392e-05, "loss": 1.2312, "step": 788 }, { "epoch": 0.06732871817299385, "grad_norm": 0.353515625, "learning_rate": 9.653768844221106e-05, "loss": 1.3271, "step": 789 }, { "epoch": 0.06741405241655911, "grad_norm": 0.333984375, "learning_rate": 9.653266331658291e-05, "loss": 1.3787, "step": 790 }, { "epoch": 0.06749938666012438, "grad_norm": 0.291015625, "learning_rate": 9.652763819095479e-05, "loss": 1.2081, "step": 791 }, { "epoch": 0.06758472090368964, "grad_norm": 0.32421875, "learning_rate": 9.652261306532664e-05, "loss": 1.2338, "step": 792 }, { "epoch": 0.06767005514725491, "grad_norm": 0.302734375, "learning_rate": 9.65175879396985e-05, "loss": 1.3317, "step": 793 }, { "epoch": 0.06775538939082017, "grad_norm": 0.330078125, "learning_rate": 9.651256281407036e-05, "loss": 1.0677, "step": 794 }, { "epoch": 0.06784072363438544, "grad_norm": 0.3125, "learning_rate": 9.650753768844222e-05, "loss": 1.3388, "step": 795 }, { "epoch": 0.0679260578779507, "grad_norm": 0.328125, "learning_rate": 9.650251256281407e-05, "loss": 1.2781, "step": 796 }, { "epoch": 0.06801139212151597, "grad_norm": 0.31640625, "learning_rate": 9.649748743718593e-05, "loss": 1.3382, "step": 797 }, { "epoch": 0.06809672636508123, "grad_norm": 0.30859375, "learning_rate": 9.64924623115578e-05, "loss": 1.3792, "step": 798 }, { "epoch": 0.0681820606086465, "grad_norm": 0.330078125, "learning_rate": 9.648743718592966e-05, "loss": 1.2831, "step": 799 }, { "epoch": 0.06826739485221175, "grad_norm": 0.31640625, "learning_rate": 9.64824120603015e-05, "loss": 1.3086, "step": 800 }, { "epoch": 0.06835272909577703, "grad_norm": 0.333984375, "learning_rate": 9.647738693467338e-05, "loss": 1.5027, "step": 801 }, { "epoch": 0.06843806333934228, "grad_norm": 0.330078125, "learning_rate": 9.647236180904523e-05, "loss": 1.2856, "step": 802 }, { "epoch": 0.06852339758290756, "grad_norm": 0.3359375, "learning_rate": 9.646733668341709e-05, "loss": 1.4071, "step": 803 }, { "epoch": 0.06860873182647281, "grad_norm": 0.318359375, "learning_rate": 9.646231155778895e-05, "loss": 1.3371, "step": 804 }, { "epoch": 0.06869406607003808, "grad_norm": 0.310546875, "learning_rate": 9.645728643216081e-05, "loss": 1.3227, "step": 805 }, { "epoch": 0.06877940031360334, "grad_norm": 0.306640625, "learning_rate": 9.645226130653266e-05, "loss": 1.5008, "step": 806 }, { "epoch": 0.06886473455716861, "grad_norm": 0.310546875, "learning_rate": 9.644723618090454e-05, "loss": 1.2905, "step": 807 }, { "epoch": 0.06895006880073387, "grad_norm": 0.3125, "learning_rate": 9.644221105527639e-05, "loss": 1.3807, "step": 808 }, { "epoch": 0.06903540304429914, "grad_norm": 0.3125, "learning_rate": 9.643718592964825e-05, "loss": 1.4242, "step": 809 }, { "epoch": 0.0691207372878644, "grad_norm": 0.3046875, "learning_rate": 9.64321608040201e-05, "loss": 1.4082, "step": 810 }, { "epoch": 0.06920607153142967, "grad_norm": 0.35546875, "learning_rate": 9.642713567839196e-05, "loss": 1.4116, "step": 811 }, { "epoch": 0.06929140577499493, "grad_norm": 0.32421875, "learning_rate": 9.642211055276382e-05, "loss": 1.4158, "step": 812 }, { "epoch": 0.0693767400185602, "grad_norm": 0.291015625, "learning_rate": 9.641708542713568e-05, "loss": 1.2911, "step": 813 }, { "epoch": 0.06946207426212546, "grad_norm": 0.34765625, "learning_rate": 9.641206030150755e-05, "loss": 1.3035, "step": 814 }, { "epoch": 0.06954740850569073, "grad_norm": 0.375, "learning_rate": 9.64070351758794e-05, "loss": 1.5684, "step": 815 }, { "epoch": 0.06963274274925599, "grad_norm": 0.294921875, "learning_rate": 9.640201005025126e-05, "loss": 1.387, "step": 816 }, { "epoch": 0.06971807699282126, "grad_norm": 0.314453125, "learning_rate": 9.639698492462312e-05, "loss": 1.3563, "step": 817 }, { "epoch": 0.06980341123638652, "grad_norm": 0.3203125, "learning_rate": 9.639195979899498e-05, "loss": 1.2745, "step": 818 }, { "epoch": 0.06988874547995179, "grad_norm": 0.3203125, "learning_rate": 9.638693467336683e-05, "loss": 1.4439, "step": 819 }, { "epoch": 0.06997407972351705, "grad_norm": 0.345703125, "learning_rate": 9.63819095477387e-05, "loss": 1.4016, "step": 820 }, { "epoch": 0.07005941396708232, "grad_norm": 0.314453125, "learning_rate": 9.637688442211055e-05, "loss": 1.2745, "step": 821 }, { "epoch": 0.07014474821064758, "grad_norm": 0.294921875, "learning_rate": 9.637185929648242e-05, "loss": 1.2946, "step": 822 }, { "epoch": 0.07023008245421285, "grad_norm": 0.302734375, "learning_rate": 9.636683417085428e-05, "loss": 1.2733, "step": 823 }, { "epoch": 0.0703154166977781, "grad_norm": 0.310546875, "learning_rate": 9.636180904522614e-05, "loss": 1.3107, "step": 824 }, { "epoch": 0.07040075094134338, "grad_norm": 0.3359375, "learning_rate": 9.635678391959799e-05, "loss": 1.3516, "step": 825 }, { "epoch": 0.07048608518490863, "grad_norm": 0.3359375, "learning_rate": 9.635175879396985e-05, "loss": 1.3862, "step": 826 }, { "epoch": 0.0705714194284739, "grad_norm": 0.28125, "learning_rate": 9.634673366834171e-05, "loss": 1.3199, "step": 827 }, { "epoch": 0.07065675367203916, "grad_norm": 0.28515625, "learning_rate": 9.634170854271357e-05, "loss": 1.3849, "step": 828 }, { "epoch": 0.07074208791560443, "grad_norm": 0.3515625, "learning_rate": 9.633668341708542e-05, "loss": 1.4096, "step": 829 }, { "epoch": 0.07082742215916969, "grad_norm": 0.333984375, "learning_rate": 9.63316582914573e-05, "loss": 1.3158, "step": 830 }, { "epoch": 0.07091275640273496, "grad_norm": 0.302734375, "learning_rate": 9.632663316582915e-05, "loss": 1.1482, "step": 831 }, { "epoch": 0.07099809064630022, "grad_norm": 0.32421875, "learning_rate": 9.632160804020101e-05, "loss": 1.2551, "step": 832 }, { "epoch": 0.07108342488986549, "grad_norm": 0.296875, "learning_rate": 9.631658291457287e-05, "loss": 1.3848, "step": 833 }, { "epoch": 0.07116875913343075, "grad_norm": 0.345703125, "learning_rate": 9.631155778894473e-05, "loss": 1.4597, "step": 834 }, { "epoch": 0.07125409337699602, "grad_norm": 0.302734375, "learning_rate": 9.630653266331658e-05, "loss": 1.2314, "step": 835 }, { "epoch": 0.07133942762056128, "grad_norm": 0.3125, "learning_rate": 9.630150753768846e-05, "loss": 1.3928, "step": 836 }, { "epoch": 0.07142476186412655, "grad_norm": 0.318359375, "learning_rate": 9.62964824120603e-05, "loss": 1.4368, "step": 837 }, { "epoch": 0.07151009610769181, "grad_norm": 0.34375, "learning_rate": 9.629145728643217e-05, "loss": 1.5076, "step": 838 }, { "epoch": 0.07159543035125708, "grad_norm": 0.337890625, "learning_rate": 9.628643216080403e-05, "loss": 1.3296, "step": 839 }, { "epoch": 0.07168076459482234, "grad_norm": 0.34765625, "learning_rate": 9.628140703517589e-05, "loss": 1.302, "step": 840 }, { "epoch": 0.07176609883838761, "grad_norm": 0.314453125, "learning_rate": 9.627638190954774e-05, "loss": 1.5028, "step": 841 }, { "epoch": 0.07185143308195287, "grad_norm": 0.30859375, "learning_rate": 9.62713567839196e-05, "loss": 1.3537, "step": 842 }, { "epoch": 0.07193676732551814, "grad_norm": 0.3359375, "learning_rate": 9.626633165829146e-05, "loss": 1.415, "step": 843 }, { "epoch": 0.07202210156908341, "grad_norm": 0.29296875, "learning_rate": 9.626130653266333e-05, "loss": 1.3223, "step": 844 }, { "epoch": 0.07210743581264867, "grad_norm": 0.32421875, "learning_rate": 9.625628140703517e-05, "loss": 1.396, "step": 845 }, { "epoch": 0.07219277005621394, "grad_norm": 0.3203125, "learning_rate": 9.625125628140704e-05, "loss": 1.2808, "step": 846 }, { "epoch": 0.0722781042997792, "grad_norm": 0.30859375, "learning_rate": 9.62462311557789e-05, "loss": 1.2221, "step": 847 }, { "epoch": 0.07236343854334447, "grad_norm": 0.287109375, "learning_rate": 9.624120603015076e-05, "loss": 1.4332, "step": 848 }, { "epoch": 0.07244877278690973, "grad_norm": 0.318359375, "learning_rate": 9.623618090452262e-05, "loss": 1.3363, "step": 849 }, { "epoch": 0.072534107030475, "grad_norm": 0.30078125, "learning_rate": 9.623115577889447e-05, "loss": 1.3362, "step": 850 }, { "epoch": 0.07261944127404026, "grad_norm": 0.353515625, "learning_rate": 9.622613065326633e-05, "loss": 1.4657, "step": 851 }, { "epoch": 0.07270477551760553, "grad_norm": 0.333984375, "learning_rate": 9.62211055276382e-05, "loss": 1.5472, "step": 852 }, { "epoch": 0.07279010976117078, "grad_norm": 0.31640625, "learning_rate": 9.621608040201006e-05, "loss": 1.3259, "step": 853 }, { "epoch": 0.07287544400473606, "grad_norm": 0.296875, "learning_rate": 9.62110552763819e-05, "loss": 1.2589, "step": 854 }, { "epoch": 0.07296077824830131, "grad_norm": 0.296875, "learning_rate": 9.620603015075378e-05, "loss": 1.2032, "step": 855 }, { "epoch": 0.07304611249186659, "grad_norm": 0.3359375, "learning_rate": 9.620100502512563e-05, "loss": 1.4066, "step": 856 }, { "epoch": 0.07313144673543184, "grad_norm": 0.30859375, "learning_rate": 9.619597989949749e-05, "loss": 1.3098, "step": 857 }, { "epoch": 0.07321678097899711, "grad_norm": 0.400390625, "learning_rate": 9.619095477386934e-05, "loss": 1.4806, "step": 858 }, { "epoch": 0.07330211522256237, "grad_norm": 0.302734375, "learning_rate": 9.618592964824122e-05, "loss": 1.3792, "step": 859 }, { "epoch": 0.07338744946612764, "grad_norm": 0.32421875, "learning_rate": 9.618090452261306e-05, "loss": 1.4728, "step": 860 }, { "epoch": 0.0734727837096929, "grad_norm": 0.3046875, "learning_rate": 9.617587939698493e-05, "loss": 1.4013, "step": 861 }, { "epoch": 0.07355811795325817, "grad_norm": 0.3046875, "learning_rate": 9.617085427135679e-05, "loss": 1.2793, "step": 862 }, { "epoch": 0.07364345219682343, "grad_norm": 0.337890625, "learning_rate": 9.616582914572865e-05, "loss": 1.3481, "step": 863 }, { "epoch": 0.0737287864403887, "grad_norm": 0.314453125, "learning_rate": 9.61608040201005e-05, "loss": 1.4183, "step": 864 }, { "epoch": 0.07381412068395396, "grad_norm": 0.33203125, "learning_rate": 9.615577889447237e-05, "loss": 1.3814, "step": 865 }, { "epoch": 0.07389945492751923, "grad_norm": 0.32421875, "learning_rate": 9.615075376884422e-05, "loss": 1.354, "step": 866 }, { "epoch": 0.07398478917108449, "grad_norm": 0.359375, "learning_rate": 9.614572864321608e-05, "loss": 1.3322, "step": 867 }, { "epoch": 0.07407012341464976, "grad_norm": 0.306640625, "learning_rate": 9.614070351758795e-05, "loss": 1.301, "step": 868 }, { "epoch": 0.07415545765821502, "grad_norm": 0.29296875, "learning_rate": 9.613567839195981e-05, "loss": 1.4269, "step": 869 }, { "epoch": 0.07424079190178029, "grad_norm": 0.3125, "learning_rate": 9.613065326633166e-05, "loss": 1.4436, "step": 870 }, { "epoch": 0.07432612614534555, "grad_norm": 0.32421875, "learning_rate": 9.612562814070353e-05, "loss": 1.376, "step": 871 }, { "epoch": 0.07441146038891082, "grad_norm": 0.306640625, "learning_rate": 9.612060301507538e-05, "loss": 1.2913, "step": 872 }, { "epoch": 0.07449679463247608, "grad_norm": 0.294921875, "learning_rate": 9.611557788944724e-05, "loss": 1.1825, "step": 873 }, { "epoch": 0.07458212887604135, "grad_norm": 0.35546875, "learning_rate": 9.611055276381909e-05, "loss": 1.4007, "step": 874 }, { "epoch": 0.0746674631196066, "grad_norm": 0.359375, "learning_rate": 9.610552763819097e-05, "loss": 1.3464, "step": 875 }, { "epoch": 0.07475279736317188, "grad_norm": 0.32421875, "learning_rate": 9.610050251256282e-05, "loss": 1.4501, "step": 876 }, { "epoch": 0.07483813160673713, "grad_norm": 0.345703125, "learning_rate": 9.609547738693468e-05, "loss": 1.3211, "step": 877 }, { "epoch": 0.0749234658503024, "grad_norm": 0.2890625, "learning_rate": 9.609045226130654e-05, "loss": 1.2537, "step": 878 }, { "epoch": 0.07500880009386766, "grad_norm": 0.298828125, "learning_rate": 9.60854271356784e-05, "loss": 1.3352, "step": 879 }, { "epoch": 0.07509413433743294, "grad_norm": 0.3046875, "learning_rate": 9.608040201005025e-05, "loss": 1.2167, "step": 880 }, { "epoch": 0.0751794685809982, "grad_norm": 0.30859375, "learning_rate": 9.607537688442211e-05, "loss": 1.2767, "step": 881 }, { "epoch": 0.07526480282456346, "grad_norm": 0.330078125, "learning_rate": 9.607035175879397e-05, "loss": 1.4463, "step": 882 }, { "epoch": 0.07535013706812872, "grad_norm": 0.306640625, "learning_rate": 9.606532663316584e-05, "loss": 1.2891, "step": 883 }, { "epoch": 0.075435471311694, "grad_norm": 0.27734375, "learning_rate": 9.60603015075377e-05, "loss": 1.3248, "step": 884 }, { "epoch": 0.07552080555525925, "grad_norm": 0.341796875, "learning_rate": 9.605527638190955e-05, "loss": 1.4177, "step": 885 }, { "epoch": 0.07560613979882452, "grad_norm": 0.3203125, "learning_rate": 9.605025125628141e-05, "loss": 1.2561, "step": 886 }, { "epoch": 0.07569147404238978, "grad_norm": 0.310546875, "learning_rate": 9.604522613065327e-05, "loss": 1.7487, "step": 887 }, { "epoch": 0.07577680828595505, "grad_norm": 0.291015625, "learning_rate": 9.604020100502513e-05, "loss": 1.3968, "step": 888 }, { "epoch": 0.07586214252952031, "grad_norm": 0.5703125, "learning_rate": 9.603517587939698e-05, "loss": 1.5181, "step": 889 }, { "epoch": 0.07594747677308558, "grad_norm": 0.310546875, "learning_rate": 9.603015075376884e-05, "loss": 1.3616, "step": 890 }, { "epoch": 0.07603281101665084, "grad_norm": 0.349609375, "learning_rate": 9.60251256281407e-05, "loss": 1.451, "step": 891 }, { "epoch": 0.07611814526021611, "grad_norm": 0.35546875, "learning_rate": 9.602010050251257e-05, "loss": 1.3771, "step": 892 }, { "epoch": 0.07620347950378137, "grad_norm": 0.3515625, "learning_rate": 9.601507537688442e-05, "loss": 1.3326, "step": 893 }, { "epoch": 0.07628881374734664, "grad_norm": 0.3203125, "learning_rate": 9.601005025125629e-05, "loss": 1.359, "step": 894 }, { "epoch": 0.0763741479909119, "grad_norm": 0.3125, "learning_rate": 9.600502512562814e-05, "loss": 1.4046, "step": 895 }, { "epoch": 0.07645948223447717, "grad_norm": 0.31640625, "learning_rate": 9.6e-05, "loss": 1.4966, "step": 896 }, { "epoch": 0.07654481647804243, "grad_norm": 0.3046875, "learning_rate": 9.599497487437186e-05, "loss": 1.4438, "step": 897 }, { "epoch": 0.0766301507216077, "grad_norm": 0.341796875, "learning_rate": 9.598994974874373e-05, "loss": 1.3693, "step": 898 }, { "epoch": 0.07671548496517296, "grad_norm": 0.3046875, "learning_rate": 9.598492462311557e-05, "loss": 1.3065, "step": 899 }, { "epoch": 0.07680081920873823, "grad_norm": 0.322265625, "learning_rate": 9.597989949748745e-05, "loss": 1.3884, "step": 900 }, { "epoch": 0.07688615345230349, "grad_norm": 0.318359375, "learning_rate": 9.59748743718593e-05, "loss": 1.297, "step": 901 }, { "epoch": 0.07697148769586876, "grad_norm": 0.279296875, "learning_rate": 9.596984924623116e-05, "loss": 1.3187, "step": 902 }, { "epoch": 0.07705682193943401, "grad_norm": 0.349609375, "learning_rate": 9.596482412060302e-05, "loss": 1.6206, "step": 903 }, { "epoch": 0.07714215618299929, "grad_norm": 0.30078125, "learning_rate": 9.595979899497489e-05, "loss": 1.3043, "step": 904 }, { "epoch": 0.07722749042656454, "grad_norm": 0.359375, "learning_rate": 9.595477386934673e-05, "loss": 1.5438, "step": 905 }, { "epoch": 0.07731282467012981, "grad_norm": 0.3125, "learning_rate": 9.59497487437186e-05, "loss": 1.2039, "step": 906 }, { "epoch": 0.07739815891369507, "grad_norm": 0.31640625, "learning_rate": 9.594472361809046e-05, "loss": 1.5157, "step": 907 }, { "epoch": 0.07748349315726034, "grad_norm": 0.314453125, "learning_rate": 9.593969849246232e-05, "loss": 1.1498, "step": 908 }, { "epoch": 0.07756882740082562, "grad_norm": 0.34765625, "learning_rate": 9.593467336683417e-05, "loss": 1.5209, "step": 909 }, { "epoch": 0.07765416164439087, "grad_norm": 0.322265625, "learning_rate": 9.592964824120604e-05, "loss": 1.4398, "step": 910 }, { "epoch": 0.07773949588795614, "grad_norm": 0.341796875, "learning_rate": 9.592462311557789e-05, "loss": 1.3306, "step": 911 }, { "epoch": 0.0778248301315214, "grad_norm": 0.353515625, "learning_rate": 9.591959798994975e-05, "loss": 1.3401, "step": 912 }, { "epoch": 0.07791016437508667, "grad_norm": 0.3125, "learning_rate": 9.591457286432162e-05, "loss": 1.3406, "step": 913 }, { "epoch": 0.07799549861865193, "grad_norm": 0.33984375, "learning_rate": 9.590954773869348e-05, "loss": 1.3676, "step": 914 }, { "epoch": 0.0780808328622172, "grad_norm": 0.31640625, "learning_rate": 9.590452261306533e-05, "loss": 1.2864, "step": 915 }, { "epoch": 0.07816616710578246, "grad_norm": 0.359375, "learning_rate": 9.589949748743719e-05, "loss": 1.4584, "step": 916 }, { "epoch": 0.07825150134934773, "grad_norm": 0.3203125, "learning_rate": 9.589447236180905e-05, "loss": 1.3652, "step": 917 }, { "epoch": 0.07833683559291299, "grad_norm": 0.3046875, "learning_rate": 9.588944723618091e-05, "loss": 1.3157, "step": 918 }, { "epoch": 0.07842216983647826, "grad_norm": 0.296875, "learning_rate": 9.588442211055276e-05, "loss": 1.3047, "step": 919 }, { "epoch": 0.07850750408004352, "grad_norm": 0.31640625, "learning_rate": 9.587939698492462e-05, "loss": 1.3499, "step": 920 }, { "epoch": 0.07859283832360879, "grad_norm": 0.314453125, "learning_rate": 9.587437185929649e-05, "loss": 1.3093, "step": 921 }, { "epoch": 0.07867817256717405, "grad_norm": 0.318359375, "learning_rate": 9.586934673366835e-05, "loss": 1.5205, "step": 922 }, { "epoch": 0.07876350681073932, "grad_norm": 0.33203125, "learning_rate": 9.586432160804021e-05, "loss": 1.3924, "step": 923 }, { "epoch": 0.07884884105430458, "grad_norm": 0.341796875, "learning_rate": 9.585929648241206e-05, "loss": 1.1967, "step": 924 }, { "epoch": 0.07893417529786985, "grad_norm": 0.29296875, "learning_rate": 9.585427135678392e-05, "loss": 1.3746, "step": 925 }, { "epoch": 0.0790195095414351, "grad_norm": 0.298828125, "learning_rate": 9.584924623115578e-05, "loss": 1.274, "step": 926 }, { "epoch": 0.07910484378500038, "grad_norm": 0.330078125, "learning_rate": 9.584422110552764e-05, "loss": 1.3245, "step": 927 }, { "epoch": 0.07919017802856564, "grad_norm": 0.32421875, "learning_rate": 9.583919597989949e-05, "loss": 1.4353, "step": 928 }, { "epoch": 0.07927551227213091, "grad_norm": 0.287109375, "learning_rate": 9.583417085427137e-05, "loss": 1.3904, "step": 929 }, { "epoch": 0.07936084651569617, "grad_norm": 0.3359375, "learning_rate": 9.582914572864322e-05, "loss": 1.3832, "step": 930 }, { "epoch": 0.07944618075926144, "grad_norm": 0.2890625, "learning_rate": 9.582412060301508e-05, "loss": 1.2578, "step": 931 }, { "epoch": 0.0795315150028267, "grad_norm": 0.341796875, "learning_rate": 9.581909547738694e-05, "loss": 1.248, "step": 932 }, { "epoch": 0.07961684924639197, "grad_norm": 0.333984375, "learning_rate": 9.58140703517588e-05, "loss": 1.3457, "step": 933 }, { "epoch": 0.07970218348995722, "grad_norm": 0.302734375, "learning_rate": 9.580904522613065e-05, "loss": 1.3833, "step": 934 }, { "epoch": 0.0797875177335225, "grad_norm": 0.318359375, "learning_rate": 9.580402010050251e-05, "loss": 1.3946, "step": 935 }, { "epoch": 0.07987285197708775, "grad_norm": 0.30078125, "learning_rate": 9.579899497487438e-05, "loss": 1.237, "step": 936 }, { "epoch": 0.07995818622065302, "grad_norm": 0.3359375, "learning_rate": 9.579396984924624e-05, "loss": 1.4489, "step": 937 }, { "epoch": 0.08004352046421828, "grad_norm": 0.37109375, "learning_rate": 9.578894472361809e-05, "loss": 1.6768, "step": 938 }, { "epoch": 0.08012885470778355, "grad_norm": 0.298828125, "learning_rate": 9.578391959798996e-05, "loss": 1.3023, "step": 939 }, { "epoch": 0.08021418895134881, "grad_norm": 0.322265625, "learning_rate": 9.577889447236181e-05, "loss": 1.4101, "step": 940 }, { "epoch": 0.08029952319491408, "grad_norm": 0.384765625, "learning_rate": 9.577386934673367e-05, "loss": 1.5069, "step": 941 }, { "epoch": 0.08038485743847934, "grad_norm": 0.306640625, "learning_rate": 9.576884422110553e-05, "loss": 1.414, "step": 942 }, { "epoch": 0.08047019168204461, "grad_norm": 0.2734375, "learning_rate": 9.57638190954774e-05, "loss": 1.4306, "step": 943 }, { "epoch": 0.08055552592560987, "grad_norm": 0.318359375, "learning_rate": 9.575879396984924e-05, "loss": 1.2921, "step": 944 }, { "epoch": 0.08064086016917514, "grad_norm": 0.28515625, "learning_rate": 9.575376884422112e-05, "loss": 1.2511, "step": 945 }, { "epoch": 0.0807261944127404, "grad_norm": 0.2734375, "learning_rate": 9.574874371859297e-05, "loss": 1.2058, "step": 946 }, { "epoch": 0.08081152865630567, "grad_norm": 0.30859375, "learning_rate": 9.574371859296483e-05, "loss": 1.2351, "step": 947 }, { "epoch": 0.08089686289987093, "grad_norm": 0.345703125, "learning_rate": 9.573869346733669e-05, "loss": 1.4025, "step": 948 }, { "epoch": 0.0809821971434362, "grad_norm": 0.32421875, "learning_rate": 9.573366834170855e-05, "loss": 1.3862, "step": 949 }, { "epoch": 0.08106753138700146, "grad_norm": 0.318359375, "learning_rate": 9.57286432160804e-05, "loss": 1.5174, "step": 950 }, { "epoch": 0.08115286563056673, "grad_norm": 0.294921875, "learning_rate": 9.572361809045227e-05, "loss": 1.3707, "step": 951 }, { "epoch": 0.08123819987413199, "grad_norm": 0.3203125, "learning_rate": 9.571859296482413e-05, "loss": 1.2754, "step": 952 }, { "epoch": 0.08132353411769726, "grad_norm": 0.37109375, "learning_rate": 9.571356783919599e-05, "loss": 1.2164, "step": 953 }, { "epoch": 0.08140886836126252, "grad_norm": 0.298828125, "learning_rate": 9.570854271356784e-05, "loss": 1.3342, "step": 954 }, { "epoch": 0.08149420260482779, "grad_norm": 0.296875, "learning_rate": 9.57035175879397e-05, "loss": 1.2289, "step": 955 }, { "epoch": 0.08157953684839304, "grad_norm": 0.34375, "learning_rate": 9.569849246231156e-05, "loss": 1.4531, "step": 956 }, { "epoch": 0.08166487109195832, "grad_norm": 0.306640625, "learning_rate": 9.569346733668342e-05, "loss": 1.2437, "step": 957 }, { "epoch": 0.08175020533552357, "grad_norm": 0.3125, "learning_rate": 9.568844221105529e-05, "loss": 1.2783, "step": 958 }, { "epoch": 0.08183553957908885, "grad_norm": 0.32421875, "learning_rate": 9.568341708542713e-05, "loss": 1.3499, "step": 959 }, { "epoch": 0.0819208738226541, "grad_norm": 0.31640625, "learning_rate": 9.5678391959799e-05, "loss": 1.4048, "step": 960 }, { "epoch": 0.08200620806621937, "grad_norm": 0.341796875, "learning_rate": 9.567336683417086e-05, "loss": 1.397, "step": 961 }, { "epoch": 0.08209154230978463, "grad_norm": 0.3359375, "learning_rate": 9.566834170854272e-05, "loss": 1.3072, "step": 962 }, { "epoch": 0.0821768765533499, "grad_norm": 0.29296875, "learning_rate": 9.566331658291457e-05, "loss": 1.2147, "step": 963 }, { "epoch": 0.08226221079691516, "grad_norm": 0.296875, "learning_rate": 9.565829145728644e-05, "loss": 1.3292, "step": 964 }, { "epoch": 0.08234754504048043, "grad_norm": 0.365234375, "learning_rate": 9.565326633165829e-05, "loss": 1.348, "step": 965 }, { "epoch": 0.08243287928404569, "grad_norm": 0.32421875, "learning_rate": 9.564824120603016e-05, "loss": 1.4118, "step": 966 }, { "epoch": 0.08251821352761096, "grad_norm": 0.31640625, "learning_rate": 9.5643216080402e-05, "loss": 1.372, "step": 967 }, { "epoch": 0.08260354777117622, "grad_norm": 0.357421875, "learning_rate": 9.563819095477388e-05, "loss": 1.3445, "step": 968 }, { "epoch": 0.08268888201474149, "grad_norm": 0.330078125, "learning_rate": 9.563316582914573e-05, "loss": 1.4447, "step": 969 }, { "epoch": 0.08277421625830675, "grad_norm": 0.322265625, "learning_rate": 9.562814070351759e-05, "loss": 1.3272, "step": 970 }, { "epoch": 0.08285955050187202, "grad_norm": 0.31640625, "learning_rate": 9.562311557788945e-05, "loss": 1.2451, "step": 971 }, { "epoch": 0.08294488474543728, "grad_norm": 0.322265625, "learning_rate": 9.561809045226131e-05, "loss": 1.3698, "step": 972 }, { "epoch": 0.08303021898900255, "grad_norm": 0.34375, "learning_rate": 9.561306532663316e-05, "loss": 1.323, "step": 973 }, { "epoch": 0.08311555323256782, "grad_norm": 0.28125, "learning_rate": 9.560804020100504e-05, "loss": 1.2683, "step": 974 }, { "epoch": 0.08320088747613308, "grad_norm": 0.31640625, "learning_rate": 9.560301507537689e-05, "loss": 1.2856, "step": 975 }, { "epoch": 0.08328622171969835, "grad_norm": 0.306640625, "learning_rate": 9.559798994974875e-05, "loss": 1.3442, "step": 976 }, { "epoch": 0.08337155596326361, "grad_norm": 0.314453125, "learning_rate": 9.559296482412061e-05, "loss": 1.3305, "step": 977 }, { "epoch": 0.08345689020682888, "grad_norm": 0.333984375, "learning_rate": 9.558793969849247e-05, "loss": 1.3871, "step": 978 }, { "epoch": 0.08354222445039414, "grad_norm": 0.3046875, "learning_rate": 9.558291457286432e-05, "loss": 1.1914, "step": 979 }, { "epoch": 0.08362755869395941, "grad_norm": 0.359375, "learning_rate": 9.55778894472362e-05, "loss": 1.5542, "step": 980 }, { "epoch": 0.08371289293752467, "grad_norm": 0.35546875, "learning_rate": 9.557286432160804e-05, "loss": 1.3395, "step": 981 }, { "epoch": 0.08379822718108994, "grad_norm": 0.31640625, "learning_rate": 9.556783919597991e-05, "loss": 1.2492, "step": 982 }, { "epoch": 0.0838835614246552, "grad_norm": 0.318359375, "learning_rate": 9.556281407035176e-05, "loss": 1.3243, "step": 983 }, { "epoch": 0.08396889566822047, "grad_norm": 0.330078125, "learning_rate": 9.555778894472363e-05, "loss": 1.4897, "step": 984 }, { "epoch": 0.08405422991178572, "grad_norm": 0.296875, "learning_rate": 9.555276381909548e-05, "loss": 1.2071, "step": 985 }, { "epoch": 0.084139564155351, "grad_norm": 0.306640625, "learning_rate": 9.554773869346734e-05, "loss": 1.4489, "step": 986 }, { "epoch": 0.08422489839891625, "grad_norm": 0.3515625, "learning_rate": 9.55427135678392e-05, "loss": 1.2155, "step": 987 }, { "epoch": 0.08431023264248153, "grad_norm": 0.294921875, "learning_rate": 9.553768844221107e-05, "loss": 1.3922, "step": 988 }, { "epoch": 0.08439556688604678, "grad_norm": 0.359375, "learning_rate": 9.553266331658291e-05, "loss": 1.3206, "step": 989 }, { "epoch": 0.08448090112961205, "grad_norm": 0.357421875, "learning_rate": 9.552763819095478e-05, "loss": 1.7447, "step": 990 }, { "epoch": 0.08456623537317731, "grad_norm": 0.298828125, "learning_rate": 9.552261306532664e-05, "loss": 1.349, "step": 991 }, { "epoch": 0.08465156961674258, "grad_norm": 0.314453125, "learning_rate": 9.55175879396985e-05, "loss": 1.4593, "step": 992 }, { "epoch": 0.08473690386030784, "grad_norm": 0.390625, "learning_rate": 9.551256281407036e-05, "loss": 1.2754, "step": 993 }, { "epoch": 0.08482223810387311, "grad_norm": 0.34375, "learning_rate": 9.550753768844221e-05, "loss": 1.3154, "step": 994 }, { "epoch": 0.08490757234743837, "grad_norm": 0.291015625, "learning_rate": 9.550251256281407e-05, "loss": 1.3023, "step": 995 }, { "epoch": 0.08499290659100364, "grad_norm": 0.3671875, "learning_rate": 9.549748743718593e-05, "loss": 1.3258, "step": 996 }, { "epoch": 0.0850782408345689, "grad_norm": 0.318359375, "learning_rate": 9.54924623115578e-05, "loss": 1.4054, "step": 997 }, { "epoch": 0.08516357507813417, "grad_norm": 0.314453125, "learning_rate": 9.548743718592965e-05, "loss": 1.4067, "step": 998 }, { "epoch": 0.08524890932169943, "grad_norm": 0.302734375, "learning_rate": 9.548241206030151e-05, "loss": 1.2185, "step": 999 }, { "epoch": 0.0853342435652647, "grad_norm": 0.34765625, "learning_rate": 9.547738693467337e-05, "loss": 1.4602, "step": 1000 }, { "epoch": 0.08541957780882996, "grad_norm": 0.265625, "learning_rate": 9.547236180904523e-05, "loss": 1.1682, "step": 1001 }, { "epoch": 0.08550491205239523, "grad_norm": 0.3125, "learning_rate": 9.546733668341708e-05, "loss": 1.5237, "step": 1002 }, { "epoch": 0.08559024629596049, "grad_norm": 0.330078125, "learning_rate": 9.546231155778896e-05, "loss": 1.4155, "step": 1003 }, { "epoch": 0.08567558053952576, "grad_norm": 0.32421875, "learning_rate": 9.54572864321608e-05, "loss": 1.3246, "step": 1004 }, { "epoch": 0.08576091478309102, "grad_norm": 0.29296875, "learning_rate": 9.545226130653267e-05, "loss": 1.4683, "step": 1005 }, { "epoch": 0.08584624902665629, "grad_norm": 0.326171875, "learning_rate": 9.544723618090453e-05, "loss": 1.3286, "step": 1006 }, { "epoch": 0.08593158327022155, "grad_norm": 0.306640625, "learning_rate": 9.544221105527639e-05, "loss": 1.3636, "step": 1007 }, { "epoch": 0.08601691751378682, "grad_norm": 0.33984375, "learning_rate": 9.543718592964824e-05, "loss": 1.4854, "step": 1008 }, { "epoch": 0.08610225175735207, "grad_norm": 0.328125, "learning_rate": 9.543216080402011e-05, "loss": 1.3744, "step": 1009 }, { "epoch": 0.08618758600091735, "grad_norm": 0.310546875, "learning_rate": 9.542713567839196e-05, "loss": 1.2739, "step": 1010 }, { "epoch": 0.0862729202444826, "grad_norm": 0.373046875, "learning_rate": 9.542211055276382e-05, "loss": 1.2112, "step": 1011 }, { "epoch": 0.08635825448804788, "grad_norm": 0.3203125, "learning_rate": 9.541708542713569e-05, "loss": 1.3066, "step": 1012 }, { "epoch": 0.08644358873161313, "grad_norm": 0.32421875, "learning_rate": 9.541206030150755e-05, "loss": 1.6257, "step": 1013 }, { "epoch": 0.0865289229751784, "grad_norm": 0.369140625, "learning_rate": 9.54070351758794e-05, "loss": 1.315, "step": 1014 }, { "epoch": 0.08661425721874366, "grad_norm": 0.318359375, "learning_rate": 9.540201005025126e-05, "loss": 1.4225, "step": 1015 }, { "epoch": 0.08669959146230893, "grad_norm": 0.279296875, "learning_rate": 9.539698492462312e-05, "loss": 1.3597, "step": 1016 }, { "epoch": 0.08678492570587419, "grad_norm": 0.30859375, "learning_rate": 9.539195979899498e-05, "loss": 1.4484, "step": 1017 }, { "epoch": 0.08687025994943946, "grad_norm": 0.380859375, "learning_rate": 9.538693467336683e-05, "loss": 1.4764, "step": 1018 }, { "epoch": 0.08695559419300472, "grad_norm": 0.333984375, "learning_rate": 9.538190954773871e-05, "loss": 1.4303, "step": 1019 }, { "epoch": 0.08704092843656999, "grad_norm": 0.38671875, "learning_rate": 9.537688442211056e-05, "loss": 1.2781, "step": 1020 }, { "epoch": 0.08712626268013525, "grad_norm": 0.294921875, "learning_rate": 9.537185929648242e-05, "loss": 1.1953, "step": 1021 }, { "epoch": 0.08721159692370052, "grad_norm": 0.306640625, "learning_rate": 9.536683417085428e-05, "loss": 1.1877, "step": 1022 }, { "epoch": 0.08729693116726578, "grad_norm": 0.328125, "learning_rate": 9.536180904522614e-05, "loss": 1.2809, "step": 1023 }, { "epoch": 0.08738226541083105, "grad_norm": 0.3125, "learning_rate": 9.535678391959799e-05, "loss": 1.3766, "step": 1024 }, { "epoch": 0.08746759965439631, "grad_norm": 0.314453125, "learning_rate": 9.535175879396985e-05, "loss": 1.3612, "step": 1025 }, { "epoch": 0.08755293389796158, "grad_norm": 0.359375, "learning_rate": 9.534673366834171e-05, "loss": 1.4097, "step": 1026 }, { "epoch": 0.08763826814152684, "grad_norm": 0.359375, "learning_rate": 9.534170854271358e-05, "loss": 1.4027, "step": 1027 }, { "epoch": 0.08772360238509211, "grad_norm": 0.283203125, "learning_rate": 9.533668341708543e-05, "loss": 1.2262, "step": 1028 }, { "epoch": 0.08780893662865737, "grad_norm": 0.31640625, "learning_rate": 9.533165829145729e-05, "loss": 1.4883, "step": 1029 }, { "epoch": 0.08789427087222264, "grad_norm": 0.28515625, "learning_rate": 9.532663316582915e-05, "loss": 1.2614, "step": 1030 }, { "epoch": 0.0879796051157879, "grad_norm": 0.359375, "learning_rate": 9.532160804020101e-05, "loss": 1.2919, "step": 1031 }, { "epoch": 0.08806493935935317, "grad_norm": 0.345703125, "learning_rate": 9.531658291457287e-05, "loss": 1.5493, "step": 1032 }, { "epoch": 0.08815027360291842, "grad_norm": 0.31640625, "learning_rate": 9.531155778894472e-05, "loss": 1.2417, "step": 1033 }, { "epoch": 0.0882356078464837, "grad_norm": 0.3203125, "learning_rate": 9.530653266331658e-05, "loss": 1.2649, "step": 1034 }, { "epoch": 0.08832094209004895, "grad_norm": 0.30859375, "learning_rate": 9.530150753768845e-05, "loss": 1.4258, "step": 1035 }, { "epoch": 0.08840627633361423, "grad_norm": 0.310546875, "learning_rate": 9.529648241206031e-05, "loss": 1.3556, "step": 1036 }, { "epoch": 0.08849161057717948, "grad_norm": 0.310546875, "learning_rate": 9.529145728643216e-05, "loss": 1.448, "step": 1037 }, { "epoch": 0.08857694482074475, "grad_norm": 0.291015625, "learning_rate": 9.528643216080403e-05, "loss": 1.2203, "step": 1038 }, { "epoch": 0.08866227906431003, "grad_norm": 0.330078125, "learning_rate": 9.528140703517588e-05, "loss": 1.3738, "step": 1039 }, { "epoch": 0.08874761330787528, "grad_norm": 0.3671875, "learning_rate": 9.527638190954774e-05, "loss": 1.2509, "step": 1040 }, { "epoch": 0.08883294755144056, "grad_norm": 0.353515625, "learning_rate": 9.52713567839196e-05, "loss": 1.4614, "step": 1041 }, { "epoch": 0.08891828179500581, "grad_norm": 0.287109375, "learning_rate": 9.526633165829147e-05, "loss": 1.1727, "step": 1042 }, { "epoch": 0.08900361603857108, "grad_norm": 0.298828125, "learning_rate": 9.526130653266331e-05, "loss": 1.4499, "step": 1043 }, { "epoch": 0.08908895028213634, "grad_norm": 0.3046875, "learning_rate": 9.525628140703518e-05, "loss": 1.2172, "step": 1044 }, { "epoch": 0.08917428452570161, "grad_norm": 0.296875, "learning_rate": 9.525125628140704e-05, "loss": 1.2474, "step": 1045 }, { "epoch": 0.08925961876926687, "grad_norm": 0.27734375, "learning_rate": 9.52462311557789e-05, "loss": 1.3159, "step": 1046 }, { "epoch": 0.08934495301283214, "grad_norm": 0.3203125, "learning_rate": 9.524120603015075e-05, "loss": 1.3855, "step": 1047 }, { "epoch": 0.0894302872563974, "grad_norm": 0.33203125, "learning_rate": 9.523618090452263e-05, "loss": 1.1638, "step": 1048 }, { "epoch": 0.08951562149996267, "grad_norm": 0.337890625, "learning_rate": 9.523115577889447e-05, "loss": 1.442, "step": 1049 }, { "epoch": 0.08960095574352793, "grad_norm": 0.306640625, "learning_rate": 9.522613065326634e-05, "loss": 1.1956, "step": 1050 }, { "epoch": 0.0896862899870932, "grad_norm": 0.310546875, "learning_rate": 9.52211055276382e-05, "loss": 1.2994, "step": 1051 }, { "epoch": 0.08977162423065846, "grad_norm": 0.365234375, "learning_rate": 9.521608040201006e-05, "loss": 1.3061, "step": 1052 }, { "epoch": 0.08985695847422373, "grad_norm": 0.3203125, "learning_rate": 9.521105527638191e-05, "loss": 1.435, "step": 1053 }, { "epoch": 0.08994229271778899, "grad_norm": 0.3203125, "learning_rate": 9.520603015075378e-05, "loss": 1.4175, "step": 1054 }, { "epoch": 0.09002762696135426, "grad_norm": 0.326171875, "learning_rate": 9.520100502512563e-05, "loss": 1.2454, "step": 1055 }, { "epoch": 0.09011296120491952, "grad_norm": 0.306640625, "learning_rate": 9.51959798994975e-05, "loss": 1.2274, "step": 1056 }, { "epoch": 0.09019829544848479, "grad_norm": 0.3046875, "learning_rate": 9.519095477386936e-05, "loss": 1.3122, "step": 1057 }, { "epoch": 0.09028362969205005, "grad_norm": 0.275390625, "learning_rate": 9.518592964824122e-05, "loss": 1.2216, "step": 1058 }, { "epoch": 0.09036896393561532, "grad_norm": 0.296875, "learning_rate": 9.518090452261307e-05, "loss": 1.3435, "step": 1059 }, { "epoch": 0.09045429817918058, "grad_norm": 0.30859375, "learning_rate": 9.517587939698493e-05, "loss": 1.1502, "step": 1060 }, { "epoch": 0.09053963242274585, "grad_norm": 0.31640625, "learning_rate": 9.517085427135679e-05, "loss": 1.1575, "step": 1061 }, { "epoch": 0.0906249666663111, "grad_norm": 0.298828125, "learning_rate": 9.516582914572865e-05, "loss": 1.209, "step": 1062 }, { "epoch": 0.09071030090987638, "grad_norm": 0.28515625, "learning_rate": 9.51608040201005e-05, "loss": 1.2504, "step": 1063 }, { "epoch": 0.09079563515344163, "grad_norm": 0.30859375, "learning_rate": 9.515577889447236e-05, "loss": 1.3055, "step": 1064 }, { "epoch": 0.0908809693970069, "grad_norm": 0.306640625, "learning_rate": 9.515075376884423e-05, "loss": 1.2386, "step": 1065 }, { "epoch": 0.09096630364057216, "grad_norm": 0.3359375, "learning_rate": 9.514572864321609e-05, "loss": 1.2577, "step": 1066 }, { "epoch": 0.09105163788413743, "grad_norm": 0.349609375, "learning_rate": 9.514070351758795e-05, "loss": 1.3365, "step": 1067 }, { "epoch": 0.09113697212770269, "grad_norm": 0.306640625, "learning_rate": 9.51356783919598e-05, "loss": 1.2264, "step": 1068 }, { "epoch": 0.09122230637126796, "grad_norm": 0.30078125, "learning_rate": 9.513065326633166e-05, "loss": 1.4203, "step": 1069 }, { "epoch": 0.09130764061483322, "grad_norm": 0.306640625, "learning_rate": 9.512562814070352e-05, "loss": 1.4671, "step": 1070 }, { "epoch": 0.09139297485839849, "grad_norm": 0.34765625, "learning_rate": 9.512060301507538e-05, "loss": 1.2146, "step": 1071 }, { "epoch": 0.09147830910196375, "grad_norm": 0.314453125, "learning_rate": 9.511557788944723e-05, "loss": 1.4928, "step": 1072 }, { "epoch": 0.09156364334552902, "grad_norm": 0.33203125, "learning_rate": 9.511055276381911e-05, "loss": 1.2909, "step": 1073 }, { "epoch": 0.09164897758909428, "grad_norm": 0.298828125, "learning_rate": 9.510552763819096e-05, "loss": 1.4021, "step": 1074 }, { "epoch": 0.09173431183265955, "grad_norm": 0.283203125, "learning_rate": 9.510050251256282e-05, "loss": 1.2152, "step": 1075 }, { "epoch": 0.09181964607622481, "grad_norm": 0.46875, "learning_rate": 9.509547738693467e-05, "loss": 1.499, "step": 1076 }, { "epoch": 0.09190498031979008, "grad_norm": 0.298828125, "learning_rate": 9.509045226130654e-05, "loss": 1.3289, "step": 1077 }, { "epoch": 0.09199031456335534, "grad_norm": 0.34375, "learning_rate": 9.508542713567839e-05, "loss": 1.3843, "step": 1078 }, { "epoch": 0.09207564880692061, "grad_norm": 0.2890625, "learning_rate": 9.508040201005025e-05, "loss": 1.3072, "step": 1079 }, { "epoch": 0.09216098305048587, "grad_norm": 0.30859375, "learning_rate": 9.507537688442212e-05, "loss": 1.3309, "step": 1080 }, { "epoch": 0.09224631729405114, "grad_norm": 0.3203125, "learning_rate": 9.507035175879398e-05, "loss": 1.3491, "step": 1081 }, { "epoch": 0.0923316515376164, "grad_norm": 0.310546875, "learning_rate": 9.506532663316583e-05, "loss": 1.4421, "step": 1082 }, { "epoch": 0.09241698578118167, "grad_norm": 0.34375, "learning_rate": 9.50603015075377e-05, "loss": 1.279, "step": 1083 }, { "epoch": 0.09250232002474693, "grad_norm": 0.27734375, "learning_rate": 9.505527638190955e-05, "loss": 1.1988, "step": 1084 }, { "epoch": 0.0925876542683122, "grad_norm": 0.265625, "learning_rate": 9.505025125628141e-05, "loss": 1.1984, "step": 1085 }, { "epoch": 0.09267298851187745, "grad_norm": 0.341796875, "learning_rate": 9.504522613065327e-05, "loss": 1.2976, "step": 1086 }, { "epoch": 0.09275832275544273, "grad_norm": 0.306640625, "learning_rate": 9.504020100502514e-05, "loss": 1.3143, "step": 1087 }, { "epoch": 0.09284365699900798, "grad_norm": 0.291015625, "learning_rate": 9.503517587939698e-05, "loss": 1.222, "step": 1088 }, { "epoch": 0.09292899124257326, "grad_norm": 0.384765625, "learning_rate": 9.503015075376886e-05, "loss": 1.4583, "step": 1089 }, { "epoch": 0.09301432548613851, "grad_norm": 0.3046875, "learning_rate": 9.502512562814071e-05, "loss": 1.3027, "step": 1090 }, { "epoch": 0.09309965972970378, "grad_norm": 0.326171875, "learning_rate": 9.502010050251257e-05, "loss": 1.2766, "step": 1091 }, { "epoch": 0.09318499397326904, "grad_norm": 0.333984375, "learning_rate": 9.501507537688442e-05, "loss": 1.367, "step": 1092 }, { "epoch": 0.09327032821683431, "grad_norm": 0.330078125, "learning_rate": 9.50100502512563e-05, "loss": 1.318, "step": 1093 }, { "epoch": 0.09335566246039957, "grad_norm": 0.322265625, "learning_rate": 9.500502512562814e-05, "loss": 1.3332, "step": 1094 }, { "epoch": 0.09344099670396484, "grad_norm": 0.287109375, "learning_rate": 9.5e-05, "loss": 1.156, "step": 1095 }, { "epoch": 0.0935263309475301, "grad_norm": 0.291015625, "learning_rate": 9.499497487437187e-05, "loss": 1.3936, "step": 1096 }, { "epoch": 0.09361166519109537, "grad_norm": 0.28125, "learning_rate": 9.498994974874373e-05, "loss": 1.4093, "step": 1097 }, { "epoch": 0.09369699943466063, "grad_norm": 0.298828125, "learning_rate": 9.498492462311558e-05, "loss": 1.1657, "step": 1098 }, { "epoch": 0.0937823336782259, "grad_norm": 0.310546875, "learning_rate": 9.497989949748744e-05, "loss": 1.1372, "step": 1099 }, { "epoch": 0.09386766792179116, "grad_norm": 0.353515625, "learning_rate": 9.49748743718593e-05, "loss": 1.4971, "step": 1100 }, { "epoch": 0.09395300216535643, "grad_norm": 0.306640625, "learning_rate": 9.496984924623116e-05, "loss": 1.4222, "step": 1101 }, { "epoch": 0.09403833640892169, "grad_norm": 0.408203125, "learning_rate": 9.496482412060303e-05, "loss": 1.3926, "step": 1102 }, { "epoch": 0.09412367065248696, "grad_norm": 0.32421875, "learning_rate": 9.495979899497487e-05, "loss": 1.3004, "step": 1103 }, { "epoch": 0.09420900489605223, "grad_norm": 0.298828125, "learning_rate": 9.495477386934674e-05, "loss": 1.2803, "step": 1104 }, { "epoch": 0.09429433913961749, "grad_norm": 0.32421875, "learning_rate": 9.49497487437186e-05, "loss": 1.3128, "step": 1105 }, { "epoch": 0.09437967338318276, "grad_norm": 0.357421875, "learning_rate": 9.494472361809046e-05, "loss": 1.3343, "step": 1106 }, { "epoch": 0.09446500762674802, "grad_norm": 0.306640625, "learning_rate": 9.493969849246231e-05, "loss": 1.2732, "step": 1107 }, { "epoch": 0.09455034187031329, "grad_norm": 0.287109375, "learning_rate": 9.493467336683417e-05, "loss": 1.3553, "step": 1108 }, { "epoch": 0.09463567611387855, "grad_norm": 0.328125, "learning_rate": 9.492964824120603e-05, "loss": 1.2705, "step": 1109 }, { "epoch": 0.09472101035744382, "grad_norm": 0.34375, "learning_rate": 9.49246231155779e-05, "loss": 1.3513, "step": 1110 }, { "epoch": 0.09480634460100908, "grad_norm": 0.310546875, "learning_rate": 9.491959798994974e-05, "loss": 1.3373, "step": 1111 }, { "epoch": 0.09489167884457435, "grad_norm": 0.29296875, "learning_rate": 9.491457286432162e-05, "loss": 1.2673, "step": 1112 }, { "epoch": 0.0949770130881396, "grad_norm": 0.306640625, "learning_rate": 9.490954773869347e-05, "loss": 1.3147, "step": 1113 }, { "epoch": 0.09506234733170488, "grad_norm": 0.3203125, "learning_rate": 9.490452261306533e-05, "loss": 1.2911, "step": 1114 }, { "epoch": 0.09514768157527013, "grad_norm": 0.302734375, "learning_rate": 9.489949748743719e-05, "loss": 1.2077, "step": 1115 }, { "epoch": 0.0952330158188354, "grad_norm": 0.326171875, "learning_rate": 9.489447236180905e-05, "loss": 1.3118, "step": 1116 }, { "epoch": 0.09531835006240066, "grad_norm": 0.326171875, "learning_rate": 9.48894472361809e-05, "loss": 1.2258, "step": 1117 }, { "epoch": 0.09540368430596594, "grad_norm": 0.412109375, "learning_rate": 9.488442211055278e-05, "loss": 1.5091, "step": 1118 }, { "epoch": 0.0954890185495312, "grad_norm": 0.337890625, "learning_rate": 9.487939698492463e-05, "loss": 1.3728, "step": 1119 }, { "epoch": 0.09557435279309646, "grad_norm": 0.30859375, "learning_rate": 9.487437185929649e-05, "loss": 1.2788, "step": 1120 }, { "epoch": 0.09565968703666172, "grad_norm": 0.287109375, "learning_rate": 9.486934673366835e-05, "loss": 1.1832, "step": 1121 }, { "epoch": 0.095745021280227, "grad_norm": 0.310546875, "learning_rate": 9.486432160804021e-05, "loss": 1.3591, "step": 1122 }, { "epoch": 0.09583035552379225, "grad_norm": 0.283203125, "learning_rate": 9.485929648241206e-05, "loss": 1.2637, "step": 1123 }, { "epoch": 0.09591568976735752, "grad_norm": 0.314453125, "learning_rate": 9.485427135678392e-05, "loss": 1.2691, "step": 1124 }, { "epoch": 0.09600102401092278, "grad_norm": 0.294921875, "learning_rate": 9.484924623115578e-05, "loss": 1.3089, "step": 1125 }, { "epoch": 0.09608635825448805, "grad_norm": 0.33984375, "learning_rate": 9.484422110552765e-05, "loss": 1.2502, "step": 1126 }, { "epoch": 0.09617169249805331, "grad_norm": 0.42578125, "learning_rate": 9.48391959798995e-05, "loss": 1.2362, "step": 1127 }, { "epoch": 0.09625702674161858, "grad_norm": 0.326171875, "learning_rate": 9.483417085427137e-05, "loss": 1.4225, "step": 1128 }, { "epoch": 0.09634236098518384, "grad_norm": 0.30078125, "learning_rate": 9.482914572864322e-05, "loss": 1.2085, "step": 1129 }, { "epoch": 0.09642769522874911, "grad_norm": 0.318359375, "learning_rate": 9.482412060301508e-05, "loss": 1.2116, "step": 1130 }, { "epoch": 0.09651302947231437, "grad_norm": 0.302734375, "learning_rate": 9.481909547738694e-05, "loss": 1.3269, "step": 1131 }, { "epoch": 0.09659836371587964, "grad_norm": 0.296875, "learning_rate": 9.48140703517588e-05, "loss": 1.3203, "step": 1132 }, { "epoch": 0.0966836979594449, "grad_norm": 0.302734375, "learning_rate": 9.480904522613065e-05, "loss": 1.513, "step": 1133 }, { "epoch": 0.09676903220301017, "grad_norm": 0.3046875, "learning_rate": 9.480402010050252e-05, "loss": 1.3283, "step": 1134 }, { "epoch": 0.09685436644657543, "grad_norm": 0.322265625, "learning_rate": 9.479899497487438e-05, "loss": 1.4475, "step": 1135 }, { "epoch": 0.0969397006901407, "grad_norm": 0.314453125, "learning_rate": 9.479396984924624e-05, "loss": 1.3202, "step": 1136 }, { "epoch": 0.09702503493370596, "grad_norm": 0.302734375, "learning_rate": 9.47889447236181e-05, "loss": 1.2201, "step": 1137 }, { "epoch": 0.09711036917727123, "grad_norm": 0.314453125, "learning_rate": 9.478391959798995e-05, "loss": 1.2863, "step": 1138 }, { "epoch": 0.09719570342083649, "grad_norm": 0.314453125, "learning_rate": 9.477889447236181e-05, "loss": 1.3521, "step": 1139 }, { "epoch": 0.09728103766440176, "grad_norm": 0.283203125, "learning_rate": 9.477386934673366e-05, "loss": 1.2675, "step": 1140 }, { "epoch": 0.09736637190796701, "grad_norm": 0.33984375, "learning_rate": 9.476884422110554e-05, "loss": 1.3831, "step": 1141 }, { "epoch": 0.09745170615153229, "grad_norm": 0.31640625, "learning_rate": 9.476381909547739e-05, "loss": 1.2899, "step": 1142 }, { "epoch": 0.09753704039509754, "grad_norm": 0.28515625, "learning_rate": 9.475879396984925e-05, "loss": 1.3978, "step": 1143 }, { "epoch": 0.09762237463866281, "grad_norm": 0.294921875, "learning_rate": 9.475376884422111e-05, "loss": 1.2152, "step": 1144 }, { "epoch": 0.09770770888222807, "grad_norm": 0.3125, "learning_rate": 9.474874371859297e-05, "loss": 1.41, "step": 1145 }, { "epoch": 0.09779304312579334, "grad_norm": 0.3671875, "learning_rate": 9.474371859296482e-05, "loss": 1.1995, "step": 1146 }, { "epoch": 0.0978783773693586, "grad_norm": 0.333984375, "learning_rate": 9.47386934673367e-05, "loss": 1.3036, "step": 1147 }, { "epoch": 0.09796371161292387, "grad_norm": 0.296875, "learning_rate": 9.473366834170854e-05, "loss": 1.3866, "step": 1148 }, { "epoch": 0.09804904585648913, "grad_norm": 0.306640625, "learning_rate": 9.47286432160804e-05, "loss": 1.4042, "step": 1149 }, { "epoch": 0.0981343801000544, "grad_norm": 0.3125, "learning_rate": 9.472361809045227e-05, "loss": 1.3863, "step": 1150 }, { "epoch": 0.09821971434361966, "grad_norm": 0.31640625, "learning_rate": 9.471859296482413e-05, "loss": 1.5364, "step": 1151 }, { "epoch": 0.09830504858718493, "grad_norm": 0.296875, "learning_rate": 9.471356783919598e-05, "loss": 1.1391, "step": 1152 }, { "epoch": 0.09839038283075019, "grad_norm": 0.330078125, "learning_rate": 9.470854271356784e-05, "loss": 1.3903, "step": 1153 }, { "epoch": 0.09847571707431546, "grad_norm": 0.306640625, "learning_rate": 9.47035175879397e-05, "loss": 1.3542, "step": 1154 }, { "epoch": 0.09856105131788072, "grad_norm": 0.318359375, "learning_rate": 9.469849246231156e-05, "loss": 1.2436, "step": 1155 }, { "epoch": 0.09864638556144599, "grad_norm": 0.302734375, "learning_rate": 9.469346733668341e-05, "loss": 1.313, "step": 1156 }, { "epoch": 0.09873171980501125, "grad_norm": 0.291015625, "learning_rate": 9.468844221105529e-05, "loss": 1.3168, "step": 1157 }, { "epoch": 0.09881705404857652, "grad_norm": 0.322265625, "learning_rate": 9.468341708542714e-05, "loss": 1.1924, "step": 1158 }, { "epoch": 0.09890238829214178, "grad_norm": 0.326171875, "learning_rate": 9.4678391959799e-05, "loss": 1.4191, "step": 1159 }, { "epoch": 0.09898772253570705, "grad_norm": 0.29296875, "learning_rate": 9.467336683417086e-05, "loss": 1.3188, "step": 1160 }, { "epoch": 0.0990730567792723, "grad_norm": 0.28515625, "learning_rate": 9.466834170854272e-05, "loss": 1.2368, "step": 1161 }, { "epoch": 0.09915839102283758, "grad_norm": 0.3359375, "learning_rate": 9.466331658291457e-05, "loss": 1.606, "step": 1162 }, { "epoch": 0.09924372526640284, "grad_norm": 0.330078125, "learning_rate": 9.465829145728645e-05, "loss": 1.4095, "step": 1163 }, { "epoch": 0.0993290595099681, "grad_norm": 0.302734375, "learning_rate": 9.46532663316583e-05, "loss": 1.2883, "step": 1164 }, { "epoch": 0.09941439375353336, "grad_norm": 0.3125, "learning_rate": 9.464824120603016e-05, "loss": 1.3407, "step": 1165 }, { "epoch": 0.09949972799709864, "grad_norm": 0.298828125, "learning_rate": 9.464321608040202e-05, "loss": 1.4751, "step": 1166 }, { "epoch": 0.0995850622406639, "grad_norm": 0.3515625, "learning_rate": 9.463819095477388e-05, "loss": 1.2236, "step": 1167 }, { "epoch": 0.09967039648422917, "grad_norm": 0.404296875, "learning_rate": 9.463316582914573e-05, "loss": 1.3881, "step": 1168 }, { "epoch": 0.09975573072779444, "grad_norm": 0.31640625, "learning_rate": 9.462814070351759e-05, "loss": 1.3522, "step": 1169 }, { "epoch": 0.0998410649713597, "grad_norm": 0.337890625, "learning_rate": 9.462311557788945e-05, "loss": 1.0582, "step": 1170 }, { "epoch": 0.09992639921492497, "grad_norm": 0.328125, "learning_rate": 9.461809045226132e-05, "loss": 1.3409, "step": 1171 }, { "epoch": 0.10001173345849022, "grad_norm": 0.357421875, "learning_rate": 9.461306532663317e-05, "loss": 1.2444, "step": 1172 }, { "epoch": 0.1000970677020555, "grad_norm": 0.322265625, "learning_rate": 9.460804020100503e-05, "loss": 1.3634, "step": 1173 }, { "epoch": 0.10018240194562075, "grad_norm": 0.34375, "learning_rate": 9.460301507537689e-05, "loss": 1.3582, "step": 1174 }, { "epoch": 0.10026773618918602, "grad_norm": 0.275390625, "learning_rate": 9.459798994974874e-05, "loss": 1.303, "step": 1175 }, { "epoch": 0.10035307043275128, "grad_norm": 0.2890625, "learning_rate": 9.459296482412061e-05, "loss": 1.3325, "step": 1176 }, { "epoch": 0.10043840467631655, "grad_norm": 0.328125, "learning_rate": 9.458793969849246e-05, "loss": 1.3358, "step": 1177 }, { "epoch": 0.10052373891988181, "grad_norm": 0.296875, "learning_rate": 9.458291457286432e-05, "loss": 1.2335, "step": 1178 }, { "epoch": 0.10060907316344708, "grad_norm": 0.32421875, "learning_rate": 9.457788944723619e-05, "loss": 1.1906, "step": 1179 }, { "epoch": 0.10069440740701234, "grad_norm": 0.2890625, "learning_rate": 9.457286432160805e-05, "loss": 1.4378, "step": 1180 }, { "epoch": 0.10077974165057761, "grad_norm": 0.32421875, "learning_rate": 9.45678391959799e-05, "loss": 1.3206, "step": 1181 }, { "epoch": 0.10086507589414287, "grad_norm": 0.349609375, "learning_rate": 9.456281407035177e-05, "loss": 1.3138, "step": 1182 }, { "epoch": 0.10095041013770814, "grad_norm": 0.330078125, "learning_rate": 9.455778894472362e-05, "loss": 1.2635, "step": 1183 }, { "epoch": 0.1010357443812734, "grad_norm": 0.30859375, "learning_rate": 9.455276381909548e-05, "loss": 1.3609, "step": 1184 }, { "epoch": 0.10112107862483867, "grad_norm": 0.328125, "learning_rate": 9.454773869346733e-05, "loss": 1.494, "step": 1185 }, { "epoch": 0.10120641286840393, "grad_norm": 0.3125, "learning_rate": 9.45427135678392e-05, "loss": 1.5018, "step": 1186 }, { "epoch": 0.1012917471119692, "grad_norm": 0.310546875, "learning_rate": 9.453768844221105e-05, "loss": 1.2777, "step": 1187 }, { "epoch": 0.10137708135553446, "grad_norm": 0.294921875, "learning_rate": 9.453266331658292e-05, "loss": 1.3416, "step": 1188 }, { "epoch": 0.10146241559909973, "grad_norm": 0.333984375, "learning_rate": 9.452763819095478e-05, "loss": 1.5682, "step": 1189 }, { "epoch": 0.10154774984266499, "grad_norm": 0.34765625, "learning_rate": 9.452261306532664e-05, "loss": 1.4004, "step": 1190 }, { "epoch": 0.10163308408623026, "grad_norm": 0.361328125, "learning_rate": 9.451758793969849e-05, "loss": 1.3492, "step": 1191 }, { "epoch": 0.10171841832979552, "grad_norm": 0.298828125, "learning_rate": 9.451256281407037e-05, "loss": 1.3, "step": 1192 }, { "epoch": 0.10180375257336079, "grad_norm": 0.3203125, "learning_rate": 9.450753768844221e-05, "loss": 1.1861, "step": 1193 }, { "epoch": 0.10188908681692604, "grad_norm": 0.3359375, "learning_rate": 9.450251256281408e-05, "loss": 1.342, "step": 1194 }, { "epoch": 0.10197442106049132, "grad_norm": 0.318359375, "learning_rate": 9.449748743718594e-05, "loss": 1.3087, "step": 1195 }, { "epoch": 0.10205975530405657, "grad_norm": 0.318359375, "learning_rate": 9.44924623115578e-05, "loss": 1.1589, "step": 1196 }, { "epoch": 0.10214508954762185, "grad_norm": 0.306640625, "learning_rate": 9.448743718592965e-05, "loss": 1.2609, "step": 1197 }, { "epoch": 0.1022304237911871, "grad_norm": 0.291015625, "learning_rate": 9.448241206030152e-05, "loss": 1.334, "step": 1198 }, { "epoch": 0.10231575803475237, "grad_norm": 0.31640625, "learning_rate": 9.447738693467337e-05, "loss": 1.4618, "step": 1199 }, { "epoch": 0.10240109227831763, "grad_norm": 0.326171875, "learning_rate": 9.447236180904523e-05, "loss": 1.2337, "step": 1200 }, { "epoch": 0.1024864265218829, "grad_norm": 0.369140625, "learning_rate": 9.446733668341708e-05, "loss": 1.4728, "step": 1201 }, { "epoch": 0.10257176076544816, "grad_norm": 0.30859375, "learning_rate": 9.446231155778896e-05, "loss": 1.3977, "step": 1202 }, { "epoch": 0.10265709500901343, "grad_norm": 0.30078125, "learning_rate": 9.445728643216081e-05, "loss": 1.4819, "step": 1203 }, { "epoch": 0.10274242925257869, "grad_norm": 0.3203125, "learning_rate": 9.445226130653267e-05, "loss": 1.24, "step": 1204 }, { "epoch": 0.10282776349614396, "grad_norm": 0.33203125, "learning_rate": 9.444723618090453e-05, "loss": 1.4156, "step": 1205 }, { "epoch": 0.10291309773970922, "grad_norm": 0.283203125, "learning_rate": 9.444221105527638e-05, "loss": 1.2449, "step": 1206 }, { "epoch": 0.10299843198327449, "grad_norm": 0.318359375, "learning_rate": 9.443718592964824e-05, "loss": 1.4098, "step": 1207 }, { "epoch": 0.10308376622683975, "grad_norm": 0.318359375, "learning_rate": 9.44321608040201e-05, "loss": 1.477, "step": 1208 }, { "epoch": 0.10316910047040502, "grad_norm": 0.306640625, "learning_rate": 9.442713567839197e-05, "loss": 1.4231, "step": 1209 }, { "epoch": 0.10325443471397028, "grad_norm": 0.294921875, "learning_rate": 9.442211055276381e-05, "loss": 1.1363, "step": 1210 }, { "epoch": 0.10333976895753555, "grad_norm": 0.37109375, "learning_rate": 9.441708542713569e-05, "loss": 1.2383, "step": 1211 }, { "epoch": 0.10342510320110081, "grad_norm": 0.353515625, "learning_rate": 9.441206030150754e-05, "loss": 1.4936, "step": 1212 }, { "epoch": 0.10351043744466608, "grad_norm": 4.3125, "learning_rate": 9.44070351758794e-05, "loss": 1.5887, "step": 1213 }, { "epoch": 0.10359577168823134, "grad_norm": 0.318359375, "learning_rate": 9.440201005025126e-05, "loss": 1.2671, "step": 1214 }, { "epoch": 0.10368110593179661, "grad_norm": 0.306640625, "learning_rate": 9.439698492462312e-05, "loss": 1.3602, "step": 1215 }, { "epoch": 0.10376644017536187, "grad_norm": 0.279296875, "learning_rate": 9.439195979899497e-05, "loss": 1.2588, "step": 1216 }, { "epoch": 0.10385177441892714, "grad_norm": 0.310546875, "learning_rate": 9.438693467336683e-05, "loss": 1.3548, "step": 1217 }, { "epoch": 0.1039371086624924, "grad_norm": 0.353515625, "learning_rate": 9.43819095477387e-05, "loss": 1.4364, "step": 1218 }, { "epoch": 0.10402244290605767, "grad_norm": 0.283203125, "learning_rate": 9.437688442211056e-05, "loss": 1.2901, "step": 1219 }, { "epoch": 0.10410777714962292, "grad_norm": 0.30859375, "learning_rate": 9.437185929648241e-05, "loss": 1.3073, "step": 1220 }, { "epoch": 0.1041931113931882, "grad_norm": 0.3125, "learning_rate": 9.436683417085428e-05, "loss": 1.3083, "step": 1221 }, { "epoch": 0.10427844563675345, "grad_norm": 0.3203125, "learning_rate": 9.436180904522613e-05, "loss": 1.3715, "step": 1222 }, { "epoch": 0.10436377988031872, "grad_norm": 0.33984375, "learning_rate": 9.4356783919598e-05, "loss": 1.4082, "step": 1223 }, { "epoch": 0.10444911412388398, "grad_norm": 0.3046875, "learning_rate": 9.435175879396986e-05, "loss": 1.1787, "step": 1224 }, { "epoch": 0.10453444836744925, "grad_norm": 0.29296875, "learning_rate": 9.434673366834172e-05, "loss": 1.2793, "step": 1225 }, { "epoch": 0.10461978261101451, "grad_norm": 0.3671875, "learning_rate": 9.434170854271357e-05, "loss": 1.3337, "step": 1226 }, { "epoch": 0.10470511685457978, "grad_norm": 0.30859375, "learning_rate": 9.433668341708544e-05, "loss": 1.4945, "step": 1227 }, { "epoch": 0.10479045109814504, "grad_norm": 0.3984375, "learning_rate": 9.433165829145729e-05, "loss": 1.462, "step": 1228 }, { "epoch": 0.10487578534171031, "grad_norm": 0.30859375, "learning_rate": 9.432663316582915e-05, "loss": 1.2387, "step": 1229 }, { "epoch": 0.10496111958527557, "grad_norm": 0.3125, "learning_rate": 9.432160804020101e-05, "loss": 1.3557, "step": 1230 }, { "epoch": 0.10504645382884084, "grad_norm": 0.302734375, "learning_rate": 9.431658291457288e-05, "loss": 1.3068, "step": 1231 }, { "epoch": 0.1051317880724061, "grad_norm": 0.32421875, "learning_rate": 9.431155778894472e-05, "loss": 1.3712, "step": 1232 }, { "epoch": 0.10521712231597137, "grad_norm": 0.296875, "learning_rate": 9.430653266331659e-05, "loss": 1.2886, "step": 1233 }, { "epoch": 0.10530245655953664, "grad_norm": 0.28515625, "learning_rate": 9.430150753768845e-05, "loss": 1.1449, "step": 1234 }, { "epoch": 0.1053877908031019, "grad_norm": 0.314453125, "learning_rate": 9.429648241206031e-05, "loss": 1.3099, "step": 1235 }, { "epoch": 0.10547312504666717, "grad_norm": 0.3046875, "learning_rate": 9.429145728643216e-05, "loss": 1.3578, "step": 1236 }, { "epoch": 0.10555845929023243, "grad_norm": 0.3515625, "learning_rate": 9.428643216080403e-05, "loss": 1.2041, "step": 1237 }, { "epoch": 0.1056437935337977, "grad_norm": 0.310546875, "learning_rate": 9.428140703517588e-05, "loss": 1.3481, "step": 1238 }, { "epoch": 0.10572912777736296, "grad_norm": 0.259765625, "learning_rate": 9.427638190954775e-05, "loss": 1.1176, "step": 1239 }, { "epoch": 0.10581446202092823, "grad_norm": 0.310546875, "learning_rate": 9.427135678391961e-05, "loss": 1.2424, "step": 1240 }, { "epoch": 0.10589979626449349, "grad_norm": 0.29296875, "learning_rate": 9.426633165829146e-05, "loss": 1.1172, "step": 1241 }, { "epoch": 0.10598513050805876, "grad_norm": 0.330078125, "learning_rate": 9.426130653266332e-05, "loss": 1.383, "step": 1242 }, { "epoch": 0.10607046475162402, "grad_norm": 0.30859375, "learning_rate": 9.425628140703518e-05, "loss": 1.3154, "step": 1243 }, { "epoch": 0.10615579899518929, "grad_norm": 0.29296875, "learning_rate": 9.425125628140704e-05, "loss": 1.1934, "step": 1244 }, { "epoch": 0.10624113323875455, "grad_norm": 0.287109375, "learning_rate": 9.424623115577889e-05, "loss": 1.3165, "step": 1245 }, { "epoch": 0.10632646748231982, "grad_norm": 0.302734375, "learning_rate": 9.424120603015077e-05, "loss": 1.1671, "step": 1246 }, { "epoch": 0.10641180172588507, "grad_norm": 0.283203125, "learning_rate": 9.423618090452261e-05, "loss": 1.2845, "step": 1247 }, { "epoch": 0.10649713596945035, "grad_norm": 0.333984375, "learning_rate": 9.423115577889448e-05, "loss": 1.3658, "step": 1248 }, { "epoch": 0.1065824702130156, "grad_norm": 0.32421875, "learning_rate": 9.422613065326632e-05, "loss": 1.3176, "step": 1249 }, { "epoch": 0.10666780445658088, "grad_norm": 0.3359375, "learning_rate": 9.42211055276382e-05, "loss": 1.3618, "step": 1250 }, { "epoch": 0.10675313870014613, "grad_norm": 0.326171875, "learning_rate": 9.421608040201005e-05, "loss": 1.2907, "step": 1251 }, { "epoch": 0.1068384729437114, "grad_norm": 0.32421875, "learning_rate": 9.421105527638191e-05, "loss": 1.4733, "step": 1252 }, { "epoch": 0.10692380718727666, "grad_norm": 0.318359375, "learning_rate": 9.420603015075377e-05, "loss": 1.4793, "step": 1253 }, { "epoch": 0.10700914143084193, "grad_norm": 0.3203125, "learning_rate": 9.420100502512564e-05, "loss": 1.2664, "step": 1254 }, { "epoch": 0.10709447567440719, "grad_norm": 0.328125, "learning_rate": 9.419597989949748e-05, "loss": 1.4601, "step": 1255 }, { "epoch": 0.10717980991797246, "grad_norm": 0.30078125, "learning_rate": 9.419095477386936e-05, "loss": 1.2501, "step": 1256 }, { "epoch": 0.10726514416153772, "grad_norm": 0.30078125, "learning_rate": 9.418592964824121e-05, "loss": 1.3505, "step": 1257 }, { "epoch": 0.10735047840510299, "grad_norm": 0.33203125, "learning_rate": 9.418090452261307e-05, "loss": 1.2536, "step": 1258 }, { "epoch": 0.10743581264866825, "grad_norm": 0.353515625, "learning_rate": 9.417587939698493e-05, "loss": 1.4409, "step": 1259 }, { "epoch": 0.10752114689223352, "grad_norm": 0.32421875, "learning_rate": 9.41708542713568e-05, "loss": 1.2701, "step": 1260 }, { "epoch": 0.10760648113579878, "grad_norm": 0.2890625, "learning_rate": 9.416582914572864e-05, "loss": 1.3374, "step": 1261 }, { "epoch": 0.10769181537936405, "grad_norm": 0.3359375, "learning_rate": 9.41608040201005e-05, "loss": 1.3722, "step": 1262 }, { "epoch": 0.10777714962292931, "grad_norm": 0.30859375, "learning_rate": 9.415577889447237e-05, "loss": 1.3045, "step": 1263 }, { "epoch": 0.10786248386649458, "grad_norm": 0.294921875, "learning_rate": 9.415075376884423e-05, "loss": 1.2823, "step": 1264 }, { "epoch": 0.10794781811005984, "grad_norm": 0.318359375, "learning_rate": 9.414572864321608e-05, "loss": 1.4643, "step": 1265 }, { "epoch": 0.10803315235362511, "grad_norm": 0.283203125, "learning_rate": 9.414070351758795e-05, "loss": 1.3461, "step": 1266 }, { "epoch": 0.10811848659719037, "grad_norm": 0.3515625, "learning_rate": 9.41356783919598e-05, "loss": 1.4715, "step": 1267 }, { "epoch": 0.10820382084075564, "grad_norm": 0.365234375, "learning_rate": 9.413065326633166e-05, "loss": 1.2455, "step": 1268 }, { "epoch": 0.1082891550843209, "grad_norm": 0.310546875, "learning_rate": 9.412562814070352e-05, "loss": 1.2344, "step": 1269 }, { "epoch": 0.10837448932788617, "grad_norm": 0.28125, "learning_rate": 9.412060301507539e-05, "loss": 1.2255, "step": 1270 }, { "epoch": 0.10845982357145142, "grad_norm": 0.298828125, "learning_rate": 9.411557788944724e-05, "loss": 1.3467, "step": 1271 }, { "epoch": 0.1085451578150167, "grad_norm": 0.328125, "learning_rate": 9.41105527638191e-05, "loss": 1.4181, "step": 1272 }, { "epoch": 0.10863049205858195, "grad_norm": 0.28515625, "learning_rate": 9.410552763819096e-05, "loss": 1.1255, "step": 1273 }, { "epoch": 0.10871582630214723, "grad_norm": 0.333984375, "learning_rate": 9.410050251256282e-05, "loss": 1.3963, "step": 1274 }, { "epoch": 0.10880116054571248, "grad_norm": 0.3125, "learning_rate": 9.409547738693468e-05, "loss": 1.3, "step": 1275 }, { "epoch": 0.10888649478927775, "grad_norm": 0.283203125, "learning_rate": 9.409045226130653e-05, "loss": 1.3346, "step": 1276 }, { "epoch": 0.10897182903284301, "grad_norm": 0.345703125, "learning_rate": 9.40854271356784e-05, "loss": 1.4137, "step": 1277 }, { "epoch": 0.10905716327640828, "grad_norm": 0.357421875, "learning_rate": 9.408040201005026e-05, "loss": 1.4214, "step": 1278 }, { "epoch": 0.10914249751997354, "grad_norm": 0.28125, "learning_rate": 9.407537688442212e-05, "loss": 1.387, "step": 1279 }, { "epoch": 0.10922783176353881, "grad_norm": 0.330078125, "learning_rate": 9.407035175879397e-05, "loss": 1.4503, "step": 1280 }, { "epoch": 0.10931316600710407, "grad_norm": 0.3046875, "learning_rate": 9.406532663316583e-05, "loss": 1.3951, "step": 1281 }, { "epoch": 0.10939850025066934, "grad_norm": 0.365234375, "learning_rate": 9.406030150753769e-05, "loss": 1.6485, "step": 1282 }, { "epoch": 0.1094838344942346, "grad_norm": 0.3125, "learning_rate": 9.405527638190955e-05, "loss": 1.3528, "step": 1283 }, { "epoch": 0.10956916873779987, "grad_norm": 0.31640625, "learning_rate": 9.40502512562814e-05, "loss": 1.3463, "step": 1284 }, { "epoch": 0.10965450298136513, "grad_norm": 0.337890625, "learning_rate": 9.404522613065328e-05, "loss": 1.4435, "step": 1285 }, { "epoch": 0.1097398372249304, "grad_norm": 0.3046875, "learning_rate": 9.404020100502513e-05, "loss": 1.2741, "step": 1286 }, { "epoch": 0.10982517146849566, "grad_norm": 0.296875, "learning_rate": 9.403517587939699e-05, "loss": 1.2217, "step": 1287 }, { "epoch": 0.10991050571206093, "grad_norm": 0.333984375, "learning_rate": 9.403015075376885e-05, "loss": 1.4001, "step": 1288 }, { "epoch": 0.10999583995562619, "grad_norm": 0.333984375, "learning_rate": 9.402512562814071e-05, "loss": 1.5025, "step": 1289 }, { "epoch": 0.11008117419919146, "grad_norm": 0.30859375, "learning_rate": 9.402010050251256e-05, "loss": 1.3302, "step": 1290 }, { "epoch": 0.11016650844275672, "grad_norm": 0.29296875, "learning_rate": 9.401507537688444e-05, "loss": 1.3329, "step": 1291 }, { "epoch": 0.11025184268632199, "grad_norm": 0.29296875, "learning_rate": 9.401005025125628e-05, "loss": 1.2649, "step": 1292 }, { "epoch": 0.11033717692988725, "grad_norm": 0.322265625, "learning_rate": 9.400502512562815e-05, "loss": 1.4107, "step": 1293 }, { "epoch": 0.11042251117345252, "grad_norm": 0.279296875, "learning_rate": 9.4e-05, "loss": 1.3736, "step": 1294 }, { "epoch": 0.11050784541701777, "grad_norm": 0.322265625, "learning_rate": 9.399497487437187e-05, "loss": 1.2525, "step": 1295 }, { "epoch": 0.11059317966058305, "grad_norm": 0.341796875, "learning_rate": 9.398994974874372e-05, "loss": 1.4613, "step": 1296 }, { "epoch": 0.1106785139041483, "grad_norm": 0.326171875, "learning_rate": 9.398492462311558e-05, "loss": 1.4314, "step": 1297 }, { "epoch": 0.11076384814771358, "grad_norm": 0.33203125, "learning_rate": 9.397989949748744e-05, "loss": 1.1376, "step": 1298 }, { "epoch": 0.11084918239127885, "grad_norm": 0.3515625, "learning_rate": 9.39748743718593e-05, "loss": 1.2772, "step": 1299 }, { "epoch": 0.1109345166348441, "grad_norm": 0.287109375, "learning_rate": 9.396984924623115e-05, "loss": 1.3063, "step": 1300 }, { "epoch": 0.11101985087840938, "grad_norm": 0.30078125, "learning_rate": 9.396482412060303e-05, "loss": 1.3122, "step": 1301 }, { "epoch": 0.11110518512197463, "grad_norm": 0.3203125, "learning_rate": 9.395979899497488e-05, "loss": 1.4451, "step": 1302 }, { "epoch": 0.1111905193655399, "grad_norm": 0.326171875, "learning_rate": 9.395477386934674e-05, "loss": 1.327, "step": 1303 }, { "epoch": 0.11127585360910516, "grad_norm": 0.306640625, "learning_rate": 9.39497487437186e-05, "loss": 1.3376, "step": 1304 }, { "epoch": 0.11136118785267043, "grad_norm": 0.267578125, "learning_rate": 9.394472361809046e-05, "loss": 1.2783, "step": 1305 }, { "epoch": 0.11144652209623569, "grad_norm": 0.34375, "learning_rate": 9.393969849246231e-05, "loss": 1.4416, "step": 1306 }, { "epoch": 0.11153185633980096, "grad_norm": 0.333984375, "learning_rate": 9.393467336683417e-05, "loss": 1.434, "step": 1307 }, { "epoch": 0.11161719058336622, "grad_norm": 0.287109375, "learning_rate": 9.392964824120604e-05, "loss": 1.3551, "step": 1308 }, { "epoch": 0.11170252482693149, "grad_norm": 0.341796875, "learning_rate": 9.39246231155779e-05, "loss": 1.3086, "step": 1309 }, { "epoch": 0.11178785907049675, "grad_norm": 0.298828125, "learning_rate": 9.391959798994975e-05, "loss": 1.2479, "step": 1310 }, { "epoch": 0.11187319331406202, "grad_norm": 0.296875, "learning_rate": 9.391457286432161e-05, "loss": 1.2318, "step": 1311 }, { "epoch": 0.11195852755762728, "grad_norm": 0.318359375, "learning_rate": 9.390954773869347e-05, "loss": 1.4304, "step": 1312 }, { "epoch": 0.11204386180119255, "grad_norm": 0.31640625, "learning_rate": 9.390452261306533e-05, "loss": 1.3709, "step": 1313 }, { "epoch": 0.11212919604475781, "grad_norm": 0.322265625, "learning_rate": 9.38994974874372e-05, "loss": 1.2007, "step": 1314 }, { "epoch": 0.11221453028832308, "grad_norm": 0.29296875, "learning_rate": 9.389447236180904e-05, "loss": 1.3801, "step": 1315 }, { "epoch": 0.11229986453188834, "grad_norm": 0.337890625, "learning_rate": 9.38894472361809e-05, "loss": 1.4037, "step": 1316 }, { "epoch": 0.11238519877545361, "grad_norm": 0.322265625, "learning_rate": 9.388442211055277e-05, "loss": 1.307, "step": 1317 }, { "epoch": 0.11247053301901887, "grad_norm": 0.31640625, "learning_rate": 9.387939698492463e-05, "loss": 1.2305, "step": 1318 }, { "epoch": 0.11255586726258414, "grad_norm": 0.416015625, "learning_rate": 9.387437185929648e-05, "loss": 1.5274, "step": 1319 }, { "epoch": 0.1126412015061494, "grad_norm": 0.322265625, "learning_rate": 9.386934673366835e-05, "loss": 1.357, "step": 1320 }, { "epoch": 0.11272653574971467, "grad_norm": 0.31640625, "learning_rate": 9.38643216080402e-05, "loss": 1.1687, "step": 1321 }, { "epoch": 0.11281186999327993, "grad_norm": 0.3671875, "learning_rate": 9.385929648241206e-05, "loss": 1.4293, "step": 1322 }, { "epoch": 0.1128972042368452, "grad_norm": 0.333984375, "learning_rate": 9.385427135678393e-05, "loss": 1.4857, "step": 1323 }, { "epoch": 0.11298253848041045, "grad_norm": 0.37109375, "learning_rate": 9.384924623115579e-05, "loss": 1.4498, "step": 1324 }, { "epoch": 0.11306787272397573, "grad_norm": 0.33203125, "learning_rate": 9.384422110552764e-05, "loss": 1.3203, "step": 1325 }, { "epoch": 0.11315320696754098, "grad_norm": 0.31640625, "learning_rate": 9.38391959798995e-05, "loss": 1.3093, "step": 1326 }, { "epoch": 0.11323854121110626, "grad_norm": 0.3046875, "learning_rate": 9.383417085427136e-05, "loss": 1.3309, "step": 1327 }, { "epoch": 0.11332387545467151, "grad_norm": 0.359375, "learning_rate": 9.382914572864322e-05, "loss": 1.6233, "step": 1328 }, { "epoch": 0.11340920969823678, "grad_norm": 0.298828125, "learning_rate": 9.382412060301507e-05, "loss": 1.4111, "step": 1329 }, { "epoch": 0.11349454394180204, "grad_norm": 0.306640625, "learning_rate": 9.381909547738695e-05, "loss": 1.4237, "step": 1330 }, { "epoch": 0.11357987818536731, "grad_norm": 0.302734375, "learning_rate": 9.38140703517588e-05, "loss": 1.4405, "step": 1331 }, { "epoch": 0.11366521242893257, "grad_norm": 0.287109375, "learning_rate": 9.380904522613066e-05, "loss": 1.3108, "step": 1332 }, { "epoch": 0.11375054667249784, "grad_norm": 0.314453125, "learning_rate": 9.380402010050252e-05, "loss": 1.2497, "step": 1333 }, { "epoch": 0.1138358809160631, "grad_norm": 0.326171875, "learning_rate": 9.379899497487438e-05, "loss": 1.3772, "step": 1334 }, { "epoch": 0.11392121515962837, "grad_norm": 0.29296875, "learning_rate": 9.379396984924623e-05, "loss": 1.3153, "step": 1335 }, { "epoch": 0.11400654940319363, "grad_norm": 0.3203125, "learning_rate": 9.37889447236181e-05, "loss": 1.3436, "step": 1336 }, { "epoch": 0.1140918836467589, "grad_norm": 0.3203125, "learning_rate": 9.378391959798995e-05, "loss": 1.4253, "step": 1337 }, { "epoch": 0.11417721789032416, "grad_norm": 0.369140625, "learning_rate": 9.377889447236182e-05, "loss": 1.1631, "step": 1338 }, { "epoch": 0.11426255213388943, "grad_norm": 0.2734375, "learning_rate": 9.377386934673368e-05, "loss": 1.1159, "step": 1339 }, { "epoch": 0.11434788637745469, "grad_norm": 0.318359375, "learning_rate": 9.376884422110554e-05, "loss": 1.1696, "step": 1340 }, { "epoch": 0.11443322062101996, "grad_norm": 0.328125, "learning_rate": 9.376381909547739e-05, "loss": 1.2474, "step": 1341 }, { "epoch": 0.11451855486458522, "grad_norm": 0.283203125, "learning_rate": 9.375879396984925e-05, "loss": 1.3009, "step": 1342 }, { "epoch": 0.11460388910815049, "grad_norm": 0.28515625, "learning_rate": 9.375376884422111e-05, "loss": 1.1385, "step": 1343 }, { "epoch": 0.11468922335171575, "grad_norm": 0.3125, "learning_rate": 9.374874371859297e-05, "loss": 1.2862, "step": 1344 }, { "epoch": 0.11477455759528102, "grad_norm": 0.306640625, "learning_rate": 9.374371859296482e-05, "loss": 1.2026, "step": 1345 }, { "epoch": 0.11485989183884628, "grad_norm": 0.361328125, "learning_rate": 9.373869346733668e-05, "loss": 1.3274, "step": 1346 }, { "epoch": 0.11494522608241155, "grad_norm": 0.27734375, "learning_rate": 9.373366834170855e-05, "loss": 1.3037, "step": 1347 }, { "epoch": 0.1150305603259768, "grad_norm": 0.326171875, "learning_rate": 9.372864321608041e-05, "loss": 1.4523, "step": 1348 }, { "epoch": 0.11511589456954208, "grad_norm": 0.306640625, "learning_rate": 9.372361809045227e-05, "loss": 1.2325, "step": 1349 }, { "epoch": 0.11520122881310733, "grad_norm": 0.396484375, "learning_rate": 9.371859296482412e-05, "loss": 1.23, "step": 1350 }, { "epoch": 0.1152865630566726, "grad_norm": 0.28515625, "learning_rate": 9.371356783919598e-05, "loss": 1.3123, "step": 1351 }, { "epoch": 0.11537189730023786, "grad_norm": 0.27734375, "learning_rate": 9.370854271356784e-05, "loss": 1.108, "step": 1352 }, { "epoch": 0.11545723154380313, "grad_norm": 0.30859375, "learning_rate": 9.37035175879397e-05, "loss": 1.3374, "step": 1353 }, { "epoch": 0.11554256578736839, "grad_norm": 0.328125, "learning_rate": 9.369849246231155e-05, "loss": 1.4087, "step": 1354 }, { "epoch": 0.11562790003093366, "grad_norm": 0.330078125, "learning_rate": 9.369346733668343e-05, "loss": 1.2411, "step": 1355 }, { "epoch": 0.11571323427449892, "grad_norm": 0.30859375, "learning_rate": 9.368844221105528e-05, "loss": 1.2342, "step": 1356 }, { "epoch": 0.1157985685180642, "grad_norm": 0.361328125, "learning_rate": 9.368341708542714e-05, "loss": 1.3929, "step": 1357 }, { "epoch": 0.11588390276162945, "grad_norm": 0.322265625, "learning_rate": 9.367839195979899e-05, "loss": 1.3551, "step": 1358 }, { "epoch": 0.11596923700519472, "grad_norm": 0.31640625, "learning_rate": 9.367336683417086e-05, "loss": 1.2237, "step": 1359 }, { "epoch": 0.11605457124875998, "grad_norm": 0.302734375, "learning_rate": 9.366834170854271e-05, "loss": 1.2868, "step": 1360 }, { "epoch": 0.11613990549232525, "grad_norm": 0.326171875, "learning_rate": 9.366331658291457e-05, "loss": 1.2755, "step": 1361 }, { "epoch": 0.11622523973589051, "grad_norm": 0.322265625, "learning_rate": 9.365829145728644e-05, "loss": 1.4521, "step": 1362 }, { "epoch": 0.11631057397945578, "grad_norm": 0.283203125, "learning_rate": 9.36532663316583e-05, "loss": 1.2697, "step": 1363 }, { "epoch": 0.11639590822302105, "grad_norm": 0.296875, "learning_rate": 9.364824120603015e-05, "loss": 1.3259, "step": 1364 }, { "epoch": 0.11648124246658631, "grad_norm": 0.3125, "learning_rate": 9.364321608040202e-05, "loss": 1.3184, "step": 1365 }, { "epoch": 0.11656657671015158, "grad_norm": 0.322265625, "learning_rate": 9.363819095477387e-05, "loss": 1.2799, "step": 1366 }, { "epoch": 0.11665191095371684, "grad_norm": 0.310546875, "learning_rate": 9.363316582914573e-05, "loss": 1.3662, "step": 1367 }, { "epoch": 0.11673724519728211, "grad_norm": 0.279296875, "learning_rate": 9.36281407035176e-05, "loss": 1.2788, "step": 1368 }, { "epoch": 0.11682257944084737, "grad_norm": 0.287109375, "learning_rate": 9.362311557788946e-05, "loss": 1.1637, "step": 1369 }, { "epoch": 0.11690791368441264, "grad_norm": 0.306640625, "learning_rate": 9.36180904522613e-05, "loss": 1.2888, "step": 1370 }, { "epoch": 0.1169932479279779, "grad_norm": 0.369140625, "learning_rate": 9.361306532663318e-05, "loss": 1.4161, "step": 1371 }, { "epoch": 0.11707858217154317, "grad_norm": 0.3515625, "learning_rate": 9.360804020100503e-05, "loss": 1.4456, "step": 1372 }, { "epoch": 0.11716391641510843, "grad_norm": 0.30859375, "learning_rate": 9.360301507537689e-05, "loss": 1.5161, "step": 1373 }, { "epoch": 0.1172492506586737, "grad_norm": 0.310546875, "learning_rate": 9.359798994974874e-05, "loss": 1.3524, "step": 1374 }, { "epoch": 0.11733458490223896, "grad_norm": 0.4453125, "learning_rate": 9.359296482412062e-05, "loss": 1.47, "step": 1375 }, { "epoch": 0.11741991914580423, "grad_norm": 0.314453125, "learning_rate": 9.358793969849246e-05, "loss": 1.2959, "step": 1376 }, { "epoch": 0.11750525338936949, "grad_norm": 0.369140625, "learning_rate": 9.358291457286433e-05, "loss": 1.5436, "step": 1377 }, { "epoch": 0.11759058763293476, "grad_norm": 0.3046875, "learning_rate": 9.357788944723619e-05, "loss": 1.3796, "step": 1378 }, { "epoch": 0.11767592187650001, "grad_norm": 0.322265625, "learning_rate": 9.357286432160805e-05, "loss": 1.082, "step": 1379 }, { "epoch": 0.11776125612006529, "grad_norm": 0.298828125, "learning_rate": 9.35678391959799e-05, "loss": 1.3027, "step": 1380 }, { "epoch": 0.11784659036363054, "grad_norm": 0.3125, "learning_rate": 9.356281407035176e-05, "loss": 1.3783, "step": 1381 }, { "epoch": 0.11793192460719581, "grad_norm": 0.361328125, "learning_rate": 9.355778894472362e-05, "loss": 1.2626, "step": 1382 }, { "epoch": 0.11801725885076107, "grad_norm": 0.291015625, "learning_rate": 9.355276381909549e-05, "loss": 1.4151, "step": 1383 }, { "epoch": 0.11810259309432634, "grad_norm": 0.3046875, "learning_rate": 9.354773869346735e-05, "loss": 1.2773, "step": 1384 }, { "epoch": 0.1181879273378916, "grad_norm": 0.310546875, "learning_rate": 9.35427135678392e-05, "loss": 1.3266, "step": 1385 }, { "epoch": 0.11827326158145687, "grad_norm": 0.3203125, "learning_rate": 9.353768844221106e-05, "loss": 1.2559, "step": 1386 }, { "epoch": 0.11835859582502213, "grad_norm": 0.3359375, "learning_rate": 9.353266331658292e-05, "loss": 1.3789, "step": 1387 }, { "epoch": 0.1184439300685874, "grad_norm": 0.31640625, "learning_rate": 9.352763819095478e-05, "loss": 1.4742, "step": 1388 }, { "epoch": 0.11852926431215266, "grad_norm": 0.2890625, "learning_rate": 9.352261306532663e-05, "loss": 1.4176, "step": 1389 }, { "epoch": 0.11861459855571793, "grad_norm": 0.2890625, "learning_rate": 9.351758793969849e-05, "loss": 1.3704, "step": 1390 }, { "epoch": 0.11869993279928319, "grad_norm": 0.26171875, "learning_rate": 9.351256281407035e-05, "loss": 1.22, "step": 1391 }, { "epoch": 0.11878526704284846, "grad_norm": 0.294921875, "learning_rate": 9.350753768844222e-05, "loss": 1.2012, "step": 1392 }, { "epoch": 0.11887060128641372, "grad_norm": 0.322265625, "learning_rate": 9.350251256281406e-05, "loss": 1.4565, "step": 1393 }, { "epoch": 0.11895593552997899, "grad_norm": 0.3828125, "learning_rate": 9.349748743718594e-05, "loss": 1.5512, "step": 1394 }, { "epoch": 0.11904126977354425, "grad_norm": 0.310546875, "learning_rate": 9.349246231155779e-05, "loss": 1.2607, "step": 1395 }, { "epoch": 0.11912660401710952, "grad_norm": 0.322265625, "learning_rate": 9.348743718592965e-05, "loss": 1.3142, "step": 1396 }, { "epoch": 0.11921193826067478, "grad_norm": 0.3203125, "learning_rate": 9.348241206030151e-05, "loss": 1.3598, "step": 1397 }, { "epoch": 0.11929727250424005, "grad_norm": 0.294921875, "learning_rate": 9.347738693467338e-05, "loss": 1.3794, "step": 1398 }, { "epoch": 0.1193826067478053, "grad_norm": 0.3359375, "learning_rate": 9.347236180904522e-05, "loss": 1.1818, "step": 1399 }, { "epoch": 0.11946794099137058, "grad_norm": 0.36328125, "learning_rate": 9.34673366834171e-05, "loss": 1.4556, "step": 1400 }, { "epoch": 0.11955327523493584, "grad_norm": 0.30078125, "learning_rate": 9.346231155778895e-05, "loss": 1.2601, "step": 1401 }, { "epoch": 0.1196386094785011, "grad_norm": 0.275390625, "learning_rate": 9.345728643216081e-05, "loss": 1.2135, "step": 1402 }, { "epoch": 0.11972394372206636, "grad_norm": 0.3359375, "learning_rate": 9.345226130653266e-05, "loss": 1.4675, "step": 1403 }, { "epoch": 0.11980927796563164, "grad_norm": 0.298828125, "learning_rate": 9.344723618090453e-05, "loss": 1.4075, "step": 1404 }, { "epoch": 0.1198946122091969, "grad_norm": 0.318359375, "learning_rate": 9.344221105527638e-05, "loss": 1.3385, "step": 1405 }, { "epoch": 0.11997994645276217, "grad_norm": 0.287109375, "learning_rate": 9.343718592964824e-05, "loss": 1.3084, "step": 1406 }, { "epoch": 0.12006528069632742, "grad_norm": 0.30859375, "learning_rate": 9.34321608040201e-05, "loss": 1.2108, "step": 1407 }, { "epoch": 0.1201506149398927, "grad_norm": 0.3046875, "learning_rate": 9.342713567839197e-05, "loss": 1.4003, "step": 1408 }, { "epoch": 0.12023594918345795, "grad_norm": 0.294921875, "learning_rate": 9.342211055276382e-05, "loss": 1.2786, "step": 1409 }, { "epoch": 0.12032128342702322, "grad_norm": 0.326171875, "learning_rate": 9.341708542713569e-05, "loss": 1.3677, "step": 1410 }, { "epoch": 0.12040661767058848, "grad_norm": 0.330078125, "learning_rate": 9.341206030150754e-05, "loss": 1.2692, "step": 1411 }, { "epoch": 0.12049195191415375, "grad_norm": 0.310546875, "learning_rate": 9.34070351758794e-05, "loss": 1.2877, "step": 1412 }, { "epoch": 0.12057728615771901, "grad_norm": 0.333984375, "learning_rate": 9.340201005025126e-05, "loss": 1.3741, "step": 1413 }, { "epoch": 0.12066262040128428, "grad_norm": 0.328125, "learning_rate": 9.339698492462313e-05, "loss": 1.228, "step": 1414 }, { "epoch": 0.12074795464484954, "grad_norm": 0.2734375, "learning_rate": 9.339195979899498e-05, "loss": 1.2013, "step": 1415 }, { "epoch": 0.12083328888841481, "grad_norm": 0.5, "learning_rate": 9.338693467336684e-05, "loss": 1.3256, "step": 1416 }, { "epoch": 0.12091862313198007, "grad_norm": 0.3203125, "learning_rate": 9.33819095477387e-05, "loss": 1.3276, "step": 1417 }, { "epoch": 0.12100395737554534, "grad_norm": 0.33203125, "learning_rate": 9.337688442211056e-05, "loss": 1.3572, "step": 1418 }, { "epoch": 0.1210892916191106, "grad_norm": 0.341796875, "learning_rate": 9.337185929648241e-05, "loss": 1.514, "step": 1419 }, { "epoch": 0.12117462586267587, "grad_norm": 0.28125, "learning_rate": 9.336683417085427e-05, "loss": 1.2285, "step": 1420 }, { "epoch": 0.12125996010624113, "grad_norm": 0.306640625, "learning_rate": 9.336180904522613e-05, "loss": 1.2514, "step": 1421 }, { "epoch": 0.1213452943498064, "grad_norm": 0.326171875, "learning_rate": 9.3356783919598e-05, "loss": 1.3792, "step": 1422 }, { "epoch": 0.12143062859337166, "grad_norm": 0.408203125, "learning_rate": 9.335175879396986e-05, "loss": 1.2483, "step": 1423 }, { "epoch": 0.12151596283693693, "grad_norm": 0.345703125, "learning_rate": 9.33467336683417e-05, "loss": 1.4372, "step": 1424 }, { "epoch": 0.12160129708050219, "grad_norm": 0.328125, "learning_rate": 9.334170854271357e-05, "loss": 1.1626, "step": 1425 }, { "epoch": 0.12168663132406746, "grad_norm": 0.31640625, "learning_rate": 9.333668341708543e-05, "loss": 1.2953, "step": 1426 }, { "epoch": 0.12177196556763271, "grad_norm": 0.322265625, "learning_rate": 9.333165829145729e-05, "loss": 1.3465, "step": 1427 }, { "epoch": 0.12185729981119799, "grad_norm": 0.291015625, "learning_rate": 9.332663316582914e-05, "loss": 1.1558, "step": 1428 }, { "epoch": 0.12194263405476326, "grad_norm": 0.29296875, "learning_rate": 9.332160804020102e-05, "loss": 1.1799, "step": 1429 }, { "epoch": 0.12202796829832852, "grad_norm": 0.390625, "learning_rate": 9.331658291457287e-05, "loss": 1.3799, "step": 1430 }, { "epoch": 0.12211330254189379, "grad_norm": 0.31640625, "learning_rate": 9.331155778894473e-05, "loss": 1.4263, "step": 1431 }, { "epoch": 0.12219863678545904, "grad_norm": 0.302734375, "learning_rate": 9.330653266331659e-05, "loss": 1.3859, "step": 1432 }, { "epoch": 0.12228397102902432, "grad_norm": 0.3671875, "learning_rate": 9.330150753768845e-05, "loss": 1.438, "step": 1433 }, { "epoch": 0.12236930527258957, "grad_norm": 0.318359375, "learning_rate": 9.32964824120603e-05, "loss": 1.4246, "step": 1434 }, { "epoch": 0.12245463951615485, "grad_norm": 0.3125, "learning_rate": 9.329145728643216e-05, "loss": 1.393, "step": 1435 }, { "epoch": 0.1225399737597201, "grad_norm": 0.30859375, "learning_rate": 9.328643216080402e-05, "loss": 1.2203, "step": 1436 }, { "epoch": 0.12262530800328537, "grad_norm": 0.310546875, "learning_rate": 9.328140703517589e-05, "loss": 1.2853, "step": 1437 }, { "epoch": 0.12271064224685063, "grad_norm": 0.296875, "learning_rate": 9.327638190954773e-05, "loss": 1.2891, "step": 1438 }, { "epoch": 0.1227959764904159, "grad_norm": 0.333984375, "learning_rate": 9.327135678391961e-05, "loss": 1.4089, "step": 1439 }, { "epoch": 0.12288131073398116, "grad_norm": 0.35546875, "learning_rate": 9.326633165829146e-05, "loss": 1.3873, "step": 1440 }, { "epoch": 0.12296664497754643, "grad_norm": 0.27734375, "learning_rate": 9.326130653266332e-05, "loss": 1.243, "step": 1441 }, { "epoch": 0.12305197922111169, "grad_norm": 0.283203125, "learning_rate": 9.325628140703518e-05, "loss": 1.1859, "step": 1442 }, { "epoch": 0.12313731346467696, "grad_norm": 0.28515625, "learning_rate": 9.325125628140704e-05, "loss": 1.309, "step": 1443 }, { "epoch": 0.12322264770824222, "grad_norm": 0.291015625, "learning_rate": 9.324623115577889e-05, "loss": 1.1615, "step": 1444 }, { "epoch": 0.12330798195180749, "grad_norm": 0.326171875, "learning_rate": 9.324120603015077e-05, "loss": 1.6309, "step": 1445 }, { "epoch": 0.12339331619537275, "grad_norm": 0.30859375, "learning_rate": 9.323618090452262e-05, "loss": 1.3351, "step": 1446 }, { "epoch": 0.12347865043893802, "grad_norm": 1.1171875, "learning_rate": 9.323115577889448e-05, "loss": 1.4337, "step": 1447 }, { "epoch": 0.12356398468250328, "grad_norm": 0.3203125, "learning_rate": 9.322613065326634e-05, "loss": 1.4513, "step": 1448 }, { "epoch": 0.12364931892606855, "grad_norm": 0.33203125, "learning_rate": 9.32211055276382e-05, "loss": 1.4264, "step": 1449 }, { "epoch": 0.12373465316963381, "grad_norm": 0.298828125, "learning_rate": 9.321608040201005e-05, "loss": 1.1431, "step": 1450 }, { "epoch": 0.12381998741319908, "grad_norm": 0.298828125, "learning_rate": 9.321105527638191e-05, "loss": 1.287, "step": 1451 }, { "epoch": 0.12390532165676434, "grad_norm": 0.3046875, "learning_rate": 9.320603015075378e-05, "loss": 1.3962, "step": 1452 }, { "epoch": 0.12399065590032961, "grad_norm": 0.306640625, "learning_rate": 9.320100502512564e-05, "loss": 1.2267, "step": 1453 }, { "epoch": 0.12407599014389487, "grad_norm": 0.2890625, "learning_rate": 9.319597989949749e-05, "loss": 1.2399, "step": 1454 }, { "epoch": 0.12416132438746014, "grad_norm": 0.35546875, "learning_rate": 9.319095477386935e-05, "loss": 1.3872, "step": 1455 }, { "epoch": 0.1242466586310254, "grad_norm": 0.31640625, "learning_rate": 9.318592964824121e-05, "loss": 1.276, "step": 1456 }, { "epoch": 0.12433199287459067, "grad_norm": 0.328125, "learning_rate": 9.318090452261307e-05, "loss": 1.227, "step": 1457 }, { "epoch": 0.12441732711815592, "grad_norm": 0.37109375, "learning_rate": 9.317587939698493e-05, "loss": 1.3223, "step": 1458 }, { "epoch": 0.1245026613617212, "grad_norm": 0.3125, "learning_rate": 9.317085427135678e-05, "loss": 1.2242, "step": 1459 }, { "epoch": 0.12458799560528645, "grad_norm": 0.318359375, "learning_rate": 9.316582914572864e-05, "loss": 1.4762, "step": 1460 }, { "epoch": 0.12467332984885172, "grad_norm": 0.34375, "learning_rate": 9.316080402010051e-05, "loss": 1.3564, "step": 1461 }, { "epoch": 0.12475866409241698, "grad_norm": 0.3203125, "learning_rate": 9.315577889447237e-05, "loss": 1.4305, "step": 1462 }, { "epoch": 0.12484399833598225, "grad_norm": 0.28515625, "learning_rate": 9.315075376884422e-05, "loss": 1.2938, "step": 1463 }, { "epoch": 0.12492933257954751, "grad_norm": 0.294921875, "learning_rate": 9.314572864321609e-05, "loss": 1.2876, "step": 1464 }, { "epoch": 0.12501466682311277, "grad_norm": 0.314453125, "learning_rate": 9.314070351758794e-05, "loss": 1.3643, "step": 1465 }, { "epoch": 0.12510000106667804, "grad_norm": 0.28515625, "learning_rate": 9.31356783919598e-05, "loss": 1.2878, "step": 1466 }, { "epoch": 0.1251853353102433, "grad_norm": 0.3125, "learning_rate": 9.313065326633165e-05, "loss": 1.3228, "step": 1467 }, { "epoch": 0.12527066955380858, "grad_norm": 0.318359375, "learning_rate": 9.312562814070353e-05, "loss": 1.2678, "step": 1468 }, { "epoch": 0.12535600379737383, "grad_norm": 0.3203125, "learning_rate": 9.312060301507538e-05, "loss": 1.3039, "step": 1469 }, { "epoch": 0.1254413380409391, "grad_norm": 0.3125, "learning_rate": 9.311557788944724e-05, "loss": 1.3492, "step": 1470 }, { "epoch": 0.12552667228450437, "grad_norm": 0.31640625, "learning_rate": 9.31105527638191e-05, "loss": 1.4216, "step": 1471 }, { "epoch": 0.12561200652806964, "grad_norm": 0.337890625, "learning_rate": 9.310552763819096e-05, "loss": 1.2161, "step": 1472 }, { "epoch": 0.12569734077163489, "grad_norm": 0.27734375, "learning_rate": 9.310050251256281e-05, "loss": 1.2469, "step": 1473 }, { "epoch": 0.12578267501520016, "grad_norm": 0.314453125, "learning_rate": 9.309547738693469e-05, "loss": 1.4488, "step": 1474 }, { "epoch": 0.12586800925876543, "grad_norm": 0.28515625, "learning_rate": 9.309045226130653e-05, "loss": 1.4823, "step": 1475 }, { "epoch": 0.1259533435023307, "grad_norm": 0.3359375, "learning_rate": 9.30854271356784e-05, "loss": 1.3294, "step": 1476 }, { "epoch": 0.12603867774589594, "grad_norm": 0.353515625, "learning_rate": 9.308040201005026e-05, "loss": 1.4159, "step": 1477 }, { "epoch": 0.12612401198946122, "grad_norm": 0.306640625, "learning_rate": 9.307537688442212e-05, "loss": 1.2869, "step": 1478 }, { "epoch": 0.1262093462330265, "grad_norm": 0.32421875, "learning_rate": 9.307035175879397e-05, "loss": 1.3068, "step": 1479 }, { "epoch": 0.12629468047659176, "grad_norm": 0.2734375, "learning_rate": 9.306532663316585e-05, "loss": 1.2561, "step": 1480 }, { "epoch": 0.126380014720157, "grad_norm": 0.345703125, "learning_rate": 9.30603015075377e-05, "loss": 1.3812, "step": 1481 }, { "epoch": 0.12646534896372227, "grad_norm": 0.328125, "learning_rate": 9.305527638190956e-05, "loss": 1.4297, "step": 1482 }, { "epoch": 0.12655068320728755, "grad_norm": 0.345703125, "learning_rate": 9.30502512562814e-05, "loss": 1.5409, "step": 1483 }, { "epoch": 0.12663601745085282, "grad_norm": 0.314453125, "learning_rate": 9.304522613065328e-05, "loss": 1.2521, "step": 1484 }, { "epoch": 0.12672135169441806, "grad_norm": 0.30078125, "learning_rate": 9.304020100502513e-05, "loss": 1.1897, "step": 1485 }, { "epoch": 0.12680668593798333, "grad_norm": 0.310546875, "learning_rate": 9.303517587939699e-05, "loss": 1.2765, "step": 1486 }, { "epoch": 0.1268920201815486, "grad_norm": 0.291015625, "learning_rate": 9.303015075376885e-05, "loss": 1.1291, "step": 1487 }, { "epoch": 0.12697735442511388, "grad_norm": 0.341796875, "learning_rate": 9.302512562814071e-05, "loss": 1.373, "step": 1488 }, { "epoch": 0.12706268866867912, "grad_norm": 0.357421875, "learning_rate": 9.302010050251256e-05, "loss": 1.1695, "step": 1489 }, { "epoch": 0.1271480229122444, "grad_norm": 0.3671875, "learning_rate": 9.301507537688442e-05, "loss": 1.2835, "step": 1490 }, { "epoch": 0.12723335715580966, "grad_norm": 0.3359375, "learning_rate": 9.301005025125629e-05, "loss": 1.351, "step": 1491 }, { "epoch": 0.12731869139937493, "grad_norm": 0.32421875, "learning_rate": 9.300502512562815e-05, "loss": 1.3164, "step": 1492 }, { "epoch": 0.1274040256429402, "grad_norm": 0.314453125, "learning_rate": 9.300000000000001e-05, "loss": 1.3673, "step": 1493 }, { "epoch": 0.12748935988650545, "grad_norm": 0.337890625, "learning_rate": 9.299497487437186e-05, "loss": 1.449, "step": 1494 }, { "epoch": 0.12757469413007072, "grad_norm": 0.35546875, "learning_rate": 9.298994974874372e-05, "loss": 1.2609, "step": 1495 }, { "epoch": 0.127660028373636, "grad_norm": 0.302734375, "learning_rate": 9.298492462311558e-05, "loss": 1.3044, "step": 1496 }, { "epoch": 0.12774536261720126, "grad_norm": 0.310546875, "learning_rate": 9.297989949748745e-05, "loss": 1.2145, "step": 1497 }, { "epoch": 0.1278306968607665, "grad_norm": 0.28125, "learning_rate": 9.29748743718593e-05, "loss": 1.3209, "step": 1498 }, { "epoch": 0.12791603110433178, "grad_norm": 0.3046875, "learning_rate": 9.296984924623116e-05, "loss": 1.2787, "step": 1499 }, { "epoch": 0.12800136534789705, "grad_norm": 0.283203125, "learning_rate": 9.296482412060302e-05, "loss": 1.15, "step": 1500 }, { "epoch": 0.15355891455024426, "grad_norm": 0.330078125, "learning_rate": 9.295979899497488e-05, "loss": 1.2346, "step": 1501 }, { "epoch": 0.15366121895700657, "grad_norm": 0.30078125, "learning_rate": 9.295477386934673e-05, "loss": 1.2548, "step": 1502 }, { "epoch": 0.1537635233637689, "grad_norm": 0.3515625, "learning_rate": 9.29497487437186e-05, "loss": 1.3803, "step": 1503 }, { "epoch": 0.1538658277705312, "grad_norm": 0.357421875, "learning_rate": 9.294472361809045e-05, "loss": 1.3935, "step": 1504 }, { "epoch": 0.15396813217729355, "grad_norm": 0.33984375, "learning_rate": 9.293969849246231e-05, "loss": 1.4877, "step": 1505 }, { "epoch": 0.15407043658405586, "grad_norm": 0.298828125, "learning_rate": 9.293467336683418e-05, "loss": 1.366, "step": 1506 }, { "epoch": 0.15417274099081818, "grad_norm": 0.349609375, "learning_rate": 9.292964824120604e-05, "loss": 1.2657, "step": 1507 }, { "epoch": 0.1542750453975805, "grad_norm": 0.318359375, "learning_rate": 9.292462311557789e-05, "loss": 1.4584, "step": 1508 }, { "epoch": 0.1543773498043428, "grad_norm": 0.314453125, "learning_rate": 9.291959798994976e-05, "loss": 1.3361, "step": 1509 }, { "epoch": 0.15447965421110516, "grad_norm": 0.37890625, "learning_rate": 9.291457286432161e-05, "loss": 1.4785, "step": 1510 }, { "epoch": 0.15458195861786747, "grad_norm": 0.3203125, "learning_rate": 9.290954773869347e-05, "loss": 1.2981, "step": 1511 }, { "epoch": 0.1546842630246298, "grad_norm": 0.283203125, "learning_rate": 9.290452261306532e-05, "loss": 1.2296, "step": 1512 }, { "epoch": 0.1547865674313921, "grad_norm": 0.318359375, "learning_rate": 9.28994974874372e-05, "loss": 1.3675, "step": 1513 }, { "epoch": 0.15488887183815442, "grad_norm": 0.298828125, "learning_rate": 9.289447236180905e-05, "loss": 1.2596, "step": 1514 }, { "epoch": 0.15499117624491676, "grad_norm": 0.291015625, "learning_rate": 9.288944723618091e-05, "loss": 1.3923, "step": 1515 }, { "epoch": 0.15509348065167908, "grad_norm": 0.337890625, "learning_rate": 9.288442211055277e-05, "loss": 1.3485, "step": 1516 }, { "epoch": 0.1551957850584414, "grad_norm": 0.298828125, "learning_rate": 9.287939698492463e-05, "loss": 1.2968, "step": 1517 }, { "epoch": 0.1552980894652037, "grad_norm": 0.296875, "learning_rate": 9.287437185929648e-05, "loss": 1.1658, "step": 1518 }, { "epoch": 0.15540039387196602, "grad_norm": 0.3125, "learning_rate": 9.286934673366836e-05, "loss": 1.3286, "step": 1519 }, { "epoch": 0.15550269827872837, "grad_norm": 0.349609375, "learning_rate": 9.28643216080402e-05, "loss": 1.3154, "step": 1520 }, { "epoch": 0.15560500268549068, "grad_norm": 0.353515625, "learning_rate": 9.285929648241207e-05, "loss": 1.3204, "step": 1521 }, { "epoch": 0.155707307092253, "grad_norm": 0.34375, "learning_rate": 9.285427135678393e-05, "loss": 1.5133, "step": 1522 }, { "epoch": 0.15580961149901532, "grad_norm": 0.306640625, "learning_rate": 9.284924623115579e-05, "loss": 1.2492, "step": 1523 }, { "epoch": 0.15591191590577763, "grad_norm": 0.328125, "learning_rate": 9.284422110552764e-05, "loss": 1.3145, "step": 1524 }, { "epoch": 0.15601422031253998, "grad_norm": 0.29296875, "learning_rate": 9.28391959798995e-05, "loss": 1.2699, "step": 1525 }, { "epoch": 0.1561165247193023, "grad_norm": 0.294921875, "learning_rate": 9.283417085427136e-05, "loss": 1.2763, "step": 1526 }, { "epoch": 0.1562188291260646, "grad_norm": 0.330078125, "learning_rate": 9.282914572864323e-05, "loss": 1.4155, "step": 1527 }, { "epoch": 0.15632113353282692, "grad_norm": 0.302734375, "learning_rate": 9.282412060301507e-05, "loss": 1.2837, "step": 1528 }, { "epoch": 0.15642343793958924, "grad_norm": 0.302734375, "learning_rate": 9.281909547738694e-05, "loss": 1.3232, "step": 1529 }, { "epoch": 0.15652574234635158, "grad_norm": 0.287109375, "learning_rate": 9.28140703517588e-05, "loss": 1.2722, "step": 1530 }, { "epoch": 0.1566280467531139, "grad_norm": 0.29296875, "learning_rate": 9.280904522613066e-05, "loss": 1.4178, "step": 1531 }, { "epoch": 0.1567303511598762, "grad_norm": 0.296875, "learning_rate": 9.280402010050252e-05, "loss": 1.306, "step": 1532 }, { "epoch": 0.15683265556663853, "grad_norm": 0.294921875, "learning_rate": 9.279899497487437e-05, "loss": 1.4261, "step": 1533 }, { "epoch": 0.15693495997340084, "grad_norm": 0.349609375, "learning_rate": 9.279396984924623e-05, "loss": 1.2852, "step": 1534 }, { "epoch": 0.1570372643801632, "grad_norm": 0.333984375, "learning_rate": 9.27889447236181e-05, "loss": 1.3846, "step": 1535 }, { "epoch": 0.1571395687869255, "grad_norm": 0.3828125, "learning_rate": 9.278391959798996e-05, "loss": 1.4102, "step": 1536 }, { "epoch": 0.15724187319368782, "grad_norm": 0.330078125, "learning_rate": 9.27788944723618e-05, "loss": 1.1593, "step": 1537 }, { "epoch": 0.15734417760045014, "grad_norm": 0.337890625, "learning_rate": 9.277386934673368e-05, "loss": 1.2609, "step": 1538 }, { "epoch": 0.15744648200721245, "grad_norm": 0.2734375, "learning_rate": 9.276884422110553e-05, "loss": 1.2475, "step": 1539 }, { "epoch": 0.1575487864139748, "grad_norm": 0.3203125, "learning_rate": 9.276381909547739e-05, "loss": 1.468, "step": 1540 }, { "epoch": 0.1576510908207371, "grad_norm": 0.353515625, "learning_rate": 9.275879396984925e-05, "loss": 1.4137, "step": 1541 }, { "epoch": 0.15775339522749943, "grad_norm": 0.3203125, "learning_rate": 9.275376884422111e-05, "loss": 1.4562, "step": 1542 }, { "epoch": 0.15785569963426174, "grad_norm": 0.318359375, "learning_rate": 9.274874371859296e-05, "loss": 1.2902, "step": 1543 }, { "epoch": 0.15795800404102406, "grad_norm": 0.359375, "learning_rate": 9.274371859296483e-05, "loss": 1.4781, "step": 1544 }, { "epoch": 0.1580603084477864, "grad_norm": 0.333984375, "learning_rate": 9.273869346733669e-05, "loss": 1.2979, "step": 1545 }, { "epoch": 0.15816261285454872, "grad_norm": 0.345703125, "learning_rate": 9.273366834170855e-05, "loss": 1.3402, "step": 1546 }, { "epoch": 0.15826491726131103, "grad_norm": 0.345703125, "learning_rate": 9.27286432160804e-05, "loss": 1.2811, "step": 1547 }, { "epoch": 0.15836722166807335, "grad_norm": 0.30078125, "learning_rate": 9.272361809045227e-05, "loss": 1.2658, "step": 1548 }, { "epoch": 0.15846952607483566, "grad_norm": 0.333984375, "learning_rate": 9.271859296482412e-05, "loss": 1.3889, "step": 1549 }, { "epoch": 0.158571830481598, "grad_norm": 0.298828125, "learning_rate": 9.271356783919598e-05, "loss": 1.2575, "step": 1550 }, { "epoch": 0.15867413488836032, "grad_norm": 0.33984375, "learning_rate": 9.270854271356785e-05, "loss": 1.449, "step": 1551 }, { "epoch": 0.15877643929512264, "grad_norm": 0.283203125, "learning_rate": 9.270351758793971e-05, "loss": 1.3009, "step": 1552 }, { "epoch": 0.15887874370188496, "grad_norm": 0.34765625, "learning_rate": 9.269849246231156e-05, "loss": 1.3837, "step": 1553 }, { "epoch": 0.15898104810864727, "grad_norm": 0.314453125, "learning_rate": 9.269346733668343e-05, "loss": 1.5433, "step": 1554 }, { "epoch": 0.15908335251540962, "grad_norm": 0.310546875, "learning_rate": 9.268844221105528e-05, "loss": 1.4125, "step": 1555 }, { "epoch": 0.15918565692217193, "grad_norm": 0.30859375, "learning_rate": 9.268341708542714e-05, "loss": 1.3466, "step": 1556 }, { "epoch": 0.15928796132893425, "grad_norm": 0.34765625, "learning_rate": 9.2678391959799e-05, "loss": 1.2347, "step": 1557 }, { "epoch": 0.15939026573569656, "grad_norm": 0.302734375, "learning_rate": 9.267336683417087e-05, "loss": 1.323, "step": 1558 }, { "epoch": 0.15949257014245888, "grad_norm": 0.37890625, "learning_rate": 9.266834170854272e-05, "loss": 1.2902, "step": 1559 }, { "epoch": 0.1595948745492212, "grad_norm": 0.302734375, "learning_rate": 9.266331658291458e-05, "loss": 1.2616, "step": 1560 }, { "epoch": 0.15969717895598354, "grad_norm": 0.333984375, "learning_rate": 9.265829145728644e-05, "loss": 1.2072, "step": 1561 }, { "epoch": 0.15979948336274585, "grad_norm": 0.31640625, "learning_rate": 9.26532663316583e-05, "loss": 1.5675, "step": 1562 }, { "epoch": 0.15990178776950817, "grad_norm": 0.357421875, "learning_rate": 9.264824120603015e-05, "loss": 1.4115, "step": 1563 }, { "epoch": 0.16000409217627048, "grad_norm": 0.294921875, "learning_rate": 9.264321608040201e-05, "loss": 1.2056, "step": 1564 }, { "epoch": 0.1601063965830328, "grad_norm": 0.27734375, "learning_rate": 9.263819095477387e-05, "loss": 1.138, "step": 1565 }, { "epoch": 0.16020870098979514, "grad_norm": 0.333984375, "learning_rate": 9.263316582914574e-05, "loss": 1.3898, "step": 1566 }, { "epoch": 0.16031100539655746, "grad_norm": 0.287109375, "learning_rate": 9.26281407035176e-05, "loss": 1.2141, "step": 1567 }, { "epoch": 0.16041330980331978, "grad_norm": 0.3046875, "learning_rate": 9.262311557788945e-05, "loss": 1.311, "step": 1568 }, { "epoch": 0.1605156142100821, "grad_norm": 0.287109375, "learning_rate": 9.261809045226131e-05, "loss": 1.199, "step": 1569 }, { "epoch": 0.1606179186168444, "grad_norm": 0.333984375, "learning_rate": 9.261306532663317e-05, "loss": 1.4812, "step": 1570 }, { "epoch": 0.16072022302360675, "grad_norm": 0.28125, "learning_rate": 9.260804020100503e-05, "loss": 1.2285, "step": 1571 }, { "epoch": 0.16082252743036907, "grad_norm": 0.32421875, "learning_rate": 9.260301507537688e-05, "loss": 1.3063, "step": 1572 }, { "epoch": 0.16092483183713138, "grad_norm": 0.306640625, "learning_rate": 9.259798994974876e-05, "loss": 1.2574, "step": 1573 }, { "epoch": 0.1610271362438937, "grad_norm": 0.302734375, "learning_rate": 9.25929648241206e-05, "loss": 1.3088, "step": 1574 }, { "epoch": 0.16112944065065601, "grad_norm": 0.310546875, "learning_rate": 9.258793969849247e-05, "loss": 1.2179, "step": 1575 }, { "epoch": 0.16123174505741836, "grad_norm": 0.466796875, "learning_rate": 9.258291457286432e-05, "loss": 1.3441, "step": 1576 }, { "epoch": 0.16133404946418067, "grad_norm": 0.34375, "learning_rate": 9.257788944723619e-05, "loss": 1.3764, "step": 1577 }, { "epoch": 0.161436353870943, "grad_norm": 0.318359375, "learning_rate": 9.257286432160804e-05, "loss": 1.3811, "step": 1578 }, { "epoch": 0.1615386582777053, "grad_norm": 0.27734375, "learning_rate": 9.25678391959799e-05, "loss": 1.3129, "step": 1579 }, { "epoch": 0.16164096268446762, "grad_norm": 0.28515625, "learning_rate": 9.256281407035176e-05, "loss": 1.2735, "step": 1580 }, { "epoch": 0.16174326709122996, "grad_norm": 0.3046875, "learning_rate": 9.255778894472363e-05, "loss": 1.3448, "step": 1581 }, { "epoch": 0.16184557149799228, "grad_norm": 0.294921875, "learning_rate": 9.255276381909547e-05, "loss": 1.2142, "step": 1582 }, { "epoch": 0.1619478759047546, "grad_norm": 0.326171875, "learning_rate": 9.254773869346735e-05, "loss": 1.2651, "step": 1583 }, { "epoch": 0.1620501803115169, "grad_norm": 0.29296875, "learning_rate": 9.25427135678392e-05, "loss": 1.1892, "step": 1584 }, { "epoch": 0.16215248471827923, "grad_norm": 0.30078125, "learning_rate": 9.253768844221106e-05, "loss": 1.2552, "step": 1585 }, { "epoch": 0.16225478912504157, "grad_norm": 0.345703125, "learning_rate": 9.253266331658292e-05, "loss": 1.3523, "step": 1586 }, { "epoch": 0.1623570935318039, "grad_norm": 0.357421875, "learning_rate": 9.252763819095478e-05, "loss": 1.4275, "step": 1587 }, { "epoch": 0.1624593979385662, "grad_norm": 0.3515625, "learning_rate": 9.252261306532663e-05, "loss": 1.3137, "step": 1588 }, { "epoch": 0.16256170234532852, "grad_norm": 0.279296875, "learning_rate": 9.251758793969851e-05, "loss": 1.2555, "step": 1589 }, { "epoch": 0.16266400675209083, "grad_norm": 0.291015625, "learning_rate": 9.251256281407036e-05, "loss": 1.178, "step": 1590 }, { "epoch": 0.16276631115885318, "grad_norm": 0.32421875, "learning_rate": 9.250753768844222e-05, "loss": 1.2815, "step": 1591 }, { "epoch": 0.1628686155656155, "grad_norm": 0.35546875, "learning_rate": 9.250251256281407e-05, "loss": 1.1241, "step": 1592 }, { "epoch": 0.1629709199723778, "grad_norm": 0.3203125, "learning_rate": 9.249748743718594e-05, "loss": 1.287, "step": 1593 }, { "epoch": 0.16307322437914012, "grad_norm": 0.318359375, "learning_rate": 9.249246231155779e-05, "loss": 1.3928, "step": 1594 }, { "epoch": 0.16317552878590244, "grad_norm": 0.298828125, "learning_rate": 9.248743718592965e-05, "loss": 1.3587, "step": 1595 }, { "epoch": 0.16327783319266478, "grad_norm": 0.30859375, "learning_rate": 9.248241206030152e-05, "loss": 1.4456, "step": 1596 }, { "epoch": 0.1633801375994271, "grad_norm": 0.310546875, "learning_rate": 9.247738693467338e-05, "loss": 1.2184, "step": 1597 }, { "epoch": 0.16348244200618942, "grad_norm": 0.359375, "learning_rate": 9.247236180904523e-05, "loss": 1.4365, "step": 1598 }, { "epoch": 0.16358474641295173, "grad_norm": 0.275390625, "learning_rate": 9.246733668341709e-05, "loss": 1.2976, "step": 1599 }, { "epoch": 0.16368705081971405, "grad_norm": 0.384765625, "learning_rate": 9.246231155778895e-05, "loss": 1.3051, "step": 1600 }, { "epoch": 0.1637893552264764, "grad_norm": 0.34765625, "learning_rate": 9.245728643216081e-05, "loss": 1.4675, "step": 1601 }, { "epoch": 0.1638916596332387, "grad_norm": 0.314453125, "learning_rate": 9.245226130653267e-05, "loss": 1.4059, "step": 1602 }, { "epoch": 0.16399396404000102, "grad_norm": 0.296875, "learning_rate": 9.244723618090452e-05, "loss": 1.3653, "step": 1603 }, { "epoch": 0.16409626844676334, "grad_norm": 0.33984375, "learning_rate": 9.244221105527638e-05, "loss": 1.2357, "step": 1604 }, { "epoch": 0.16419857285352565, "grad_norm": 0.326171875, "learning_rate": 9.243718592964823e-05, "loss": 1.2528, "step": 1605 }, { "epoch": 0.164300877260288, "grad_norm": 0.306640625, "learning_rate": 9.243216080402011e-05, "loss": 1.2042, "step": 1606 }, { "epoch": 0.1644031816670503, "grad_norm": 0.34765625, "learning_rate": 9.242713567839196e-05, "loss": 1.2687, "step": 1607 }, { "epoch": 0.16450548607381263, "grad_norm": 0.29296875, "learning_rate": 9.242211055276382e-05, "loss": 1.4098, "step": 1608 }, { "epoch": 0.16460779048057494, "grad_norm": 0.322265625, "learning_rate": 9.241708542713568e-05, "loss": 1.3566, "step": 1609 }, { "epoch": 0.16471009488733726, "grad_norm": 0.30859375, "learning_rate": 9.241206030150754e-05, "loss": 1.1662, "step": 1610 }, { "epoch": 0.1648123992940996, "grad_norm": 0.2890625, "learning_rate": 9.240703517587939e-05, "loss": 1.3317, "step": 1611 }, { "epoch": 0.16491470370086192, "grad_norm": 0.326171875, "learning_rate": 9.240201005025127e-05, "loss": 1.4334, "step": 1612 }, { "epoch": 0.16501700810762424, "grad_norm": 0.33203125, "learning_rate": 9.239698492462312e-05, "loss": 1.2236, "step": 1613 }, { "epoch": 0.16511931251438655, "grad_norm": 0.28515625, "learning_rate": 9.239195979899498e-05, "loss": 1.3276, "step": 1614 }, { "epoch": 0.16522161692114887, "grad_norm": 0.31640625, "learning_rate": 9.238693467336684e-05, "loss": 1.358, "step": 1615 }, { "epoch": 0.1653239213279112, "grad_norm": 0.310546875, "learning_rate": 9.23819095477387e-05, "loss": 1.3055, "step": 1616 }, { "epoch": 0.16542622573467353, "grad_norm": 0.310546875, "learning_rate": 9.237688442211055e-05, "loss": 1.2512, "step": 1617 }, { "epoch": 0.16552853014143584, "grad_norm": 0.306640625, "learning_rate": 9.237185929648243e-05, "loss": 1.2963, "step": 1618 }, { "epoch": 0.16563083454819816, "grad_norm": 0.34765625, "learning_rate": 9.236683417085427e-05, "loss": 1.38, "step": 1619 }, { "epoch": 0.16573313895496047, "grad_norm": 0.31640625, "learning_rate": 9.236180904522614e-05, "loss": 1.4313, "step": 1620 }, { "epoch": 0.16583544336172282, "grad_norm": 0.3046875, "learning_rate": 9.235678391959799e-05, "loss": 1.2958, "step": 1621 }, { "epoch": 0.16593774776848513, "grad_norm": 0.32421875, "learning_rate": 9.235175879396986e-05, "loss": 1.3825, "step": 1622 }, { "epoch": 0.16604005217524745, "grad_norm": 0.30078125, "learning_rate": 9.234673366834171e-05, "loss": 1.353, "step": 1623 }, { "epoch": 0.16614235658200976, "grad_norm": 0.330078125, "learning_rate": 9.234170854271357e-05, "loss": 1.3458, "step": 1624 }, { "epoch": 0.16624466098877208, "grad_norm": 0.314453125, "learning_rate": 9.233668341708543e-05, "loss": 1.26, "step": 1625 }, { "epoch": 0.16634696539553442, "grad_norm": 0.330078125, "learning_rate": 9.23316582914573e-05, "loss": 1.2797, "step": 1626 }, { "epoch": 0.16644926980229674, "grad_norm": 0.326171875, "learning_rate": 9.232663316582914e-05, "loss": 1.3798, "step": 1627 }, { "epoch": 0.16655157420905906, "grad_norm": 0.306640625, "learning_rate": 9.232160804020102e-05, "loss": 1.2309, "step": 1628 }, { "epoch": 0.16665387861582137, "grad_norm": 0.291015625, "learning_rate": 9.231658291457287e-05, "loss": 1.3051, "step": 1629 }, { "epoch": 0.1667561830225837, "grad_norm": 0.28515625, "learning_rate": 9.231155778894473e-05, "loss": 1.1063, "step": 1630 }, { "epoch": 0.16685848742934603, "grad_norm": 0.27734375, "learning_rate": 9.230653266331659e-05, "loss": 1.277, "step": 1631 }, { "epoch": 0.16696079183610835, "grad_norm": 0.294921875, "learning_rate": 9.230150753768845e-05, "loss": 1.2432, "step": 1632 }, { "epoch": 0.16706309624287066, "grad_norm": 0.34765625, "learning_rate": 9.22964824120603e-05, "loss": 1.5938, "step": 1633 }, { "epoch": 0.16716540064963298, "grad_norm": 0.3125, "learning_rate": 9.229145728643216e-05, "loss": 1.3362, "step": 1634 }, { "epoch": 0.1672677050563953, "grad_norm": 0.28515625, "learning_rate": 9.228643216080403e-05, "loss": 1.2522, "step": 1635 }, { "epoch": 0.16737000946315764, "grad_norm": 0.3125, "learning_rate": 9.228140703517588e-05, "loss": 1.3154, "step": 1636 }, { "epoch": 0.16747231386991995, "grad_norm": 0.271484375, "learning_rate": 9.227638190954774e-05, "loss": 1.3019, "step": 1637 }, { "epoch": 0.16757461827668227, "grad_norm": 0.314453125, "learning_rate": 9.22713567839196e-05, "loss": 1.4128, "step": 1638 }, { "epoch": 0.16767692268344458, "grad_norm": 0.291015625, "learning_rate": 9.226633165829146e-05, "loss": 1.3354, "step": 1639 }, { "epoch": 0.1677792270902069, "grad_norm": 0.326171875, "learning_rate": 9.226130653266331e-05, "loss": 1.3858, "step": 1640 }, { "epoch": 0.16788153149696924, "grad_norm": 0.41015625, "learning_rate": 9.225628140703519e-05, "loss": 1.4336, "step": 1641 }, { "epoch": 0.16798383590373156, "grad_norm": 0.32421875, "learning_rate": 9.225125628140703e-05, "loss": 1.3268, "step": 1642 }, { "epoch": 0.16808614031049388, "grad_norm": 0.29296875, "learning_rate": 9.22462311557789e-05, "loss": 1.3165, "step": 1643 }, { "epoch": 0.1681884447172562, "grad_norm": 0.33203125, "learning_rate": 9.224120603015076e-05, "loss": 1.4154, "step": 1644 }, { "epoch": 0.1682907491240185, "grad_norm": 0.306640625, "learning_rate": 9.223618090452262e-05, "loss": 1.2417, "step": 1645 }, { "epoch": 0.16839305353078085, "grad_norm": 0.328125, "learning_rate": 9.223115577889447e-05, "loss": 1.2462, "step": 1646 }, { "epoch": 0.16849535793754317, "grad_norm": 0.3203125, "learning_rate": 9.222613065326634e-05, "loss": 1.2048, "step": 1647 }, { "epoch": 0.16859766234430548, "grad_norm": 0.349609375, "learning_rate": 9.222110552763819e-05, "loss": 1.3316, "step": 1648 }, { "epoch": 0.1686999667510678, "grad_norm": 0.30078125, "learning_rate": 9.221608040201005e-05, "loss": 1.3575, "step": 1649 }, { "epoch": 0.1688022711578301, "grad_norm": 0.318359375, "learning_rate": 9.221105527638192e-05, "loss": 1.4781, "step": 1650 }, { "epoch": 0.16890457556459246, "grad_norm": 0.35546875, "learning_rate": 9.220603015075378e-05, "loss": 1.2977, "step": 1651 }, { "epoch": 0.16900687997135477, "grad_norm": 0.33984375, "learning_rate": 9.220100502512563e-05, "loss": 1.2581, "step": 1652 }, { "epoch": 0.1691091843781171, "grad_norm": 0.326171875, "learning_rate": 9.219597989949749e-05, "loss": 1.3487, "step": 1653 }, { "epoch": 0.1692114887848794, "grad_norm": 0.283203125, "learning_rate": 9.219095477386935e-05, "loss": 1.2185, "step": 1654 }, { "epoch": 0.16931379319164172, "grad_norm": 0.314453125, "learning_rate": 9.218592964824121e-05, "loss": 1.4111, "step": 1655 }, { "epoch": 0.16941609759840406, "grad_norm": 0.35546875, "learning_rate": 9.218090452261306e-05, "loss": 1.3675, "step": 1656 }, { "epoch": 0.16951840200516638, "grad_norm": 0.32421875, "learning_rate": 9.217587939698494e-05, "loss": 1.3928, "step": 1657 }, { "epoch": 0.1696207064119287, "grad_norm": 0.287109375, "learning_rate": 9.217085427135679e-05, "loss": 1.2325, "step": 1658 }, { "epoch": 0.169723010818691, "grad_norm": 0.34375, "learning_rate": 9.216582914572865e-05, "loss": 1.2944, "step": 1659 }, { "epoch": 0.16982531522545333, "grad_norm": 0.28515625, "learning_rate": 9.216080402010051e-05, "loss": 1.2395, "step": 1660 }, { "epoch": 0.16992761963221567, "grad_norm": 0.2734375, "learning_rate": 9.215577889447237e-05, "loss": 1.3592, "step": 1661 }, { "epoch": 0.170029924038978, "grad_norm": 0.291015625, "learning_rate": 9.215075376884422e-05, "loss": 1.3152, "step": 1662 }, { "epoch": 0.1701322284457403, "grad_norm": 0.318359375, "learning_rate": 9.21457286432161e-05, "loss": 1.3014, "step": 1663 }, { "epoch": 0.17023453285250262, "grad_norm": 0.306640625, "learning_rate": 9.214070351758794e-05, "loss": 1.2549, "step": 1664 }, { "epoch": 0.17033683725926493, "grad_norm": 0.396484375, "learning_rate": 9.21356783919598e-05, "loss": 1.3299, "step": 1665 }, { "epoch": 0.17043914166602728, "grad_norm": 0.30859375, "learning_rate": 9.213065326633167e-05, "loss": 1.3312, "step": 1666 }, { "epoch": 0.1705414460727896, "grad_norm": 0.302734375, "learning_rate": 9.212562814070353e-05, "loss": 1.1033, "step": 1667 }, { "epoch": 0.1706437504795519, "grad_norm": 0.38671875, "learning_rate": 9.212060301507538e-05, "loss": 1.3803, "step": 1668 }, { "epoch": 0.17074605488631422, "grad_norm": 0.36328125, "learning_rate": 9.211557788944724e-05, "loss": 1.3409, "step": 1669 }, { "epoch": 0.17084835929307654, "grad_norm": 0.33203125, "learning_rate": 9.21105527638191e-05, "loss": 1.206, "step": 1670 }, { "epoch": 0.17095066369983888, "grad_norm": 0.28515625, "learning_rate": 9.210552763819095e-05, "loss": 1.287, "step": 1671 }, { "epoch": 0.1710529681066012, "grad_norm": 0.36328125, "learning_rate": 9.210050251256281e-05, "loss": 1.3579, "step": 1672 }, { "epoch": 0.17115527251336352, "grad_norm": 0.287109375, "learning_rate": 9.209547738693468e-05, "loss": 1.3364, "step": 1673 }, { "epoch": 0.17125757692012583, "grad_norm": 0.359375, "learning_rate": 9.209045226130654e-05, "loss": 1.4099, "step": 1674 }, { "epoch": 0.17135988132688815, "grad_norm": 0.333984375, "learning_rate": 9.208542713567839e-05, "loss": 1.3591, "step": 1675 }, { "epoch": 0.1714621857336505, "grad_norm": 0.32421875, "learning_rate": 9.208040201005026e-05, "loss": 1.481, "step": 1676 }, { "epoch": 0.1715644901404128, "grad_norm": 0.2890625, "learning_rate": 9.207537688442211e-05, "loss": 1.3703, "step": 1677 }, { "epoch": 0.17166679454717512, "grad_norm": 0.31640625, "learning_rate": 9.207035175879397e-05, "loss": 1.2713, "step": 1678 }, { "epoch": 0.17176909895393744, "grad_norm": 0.30078125, "learning_rate": 9.206532663316583e-05, "loss": 1.2143, "step": 1679 }, { "epoch": 0.17187140336069975, "grad_norm": 0.302734375, "learning_rate": 9.20603015075377e-05, "loss": 1.4203, "step": 1680 }, { "epoch": 0.1719737077674621, "grad_norm": 0.37109375, "learning_rate": 9.205527638190954e-05, "loss": 1.4291, "step": 1681 }, { "epoch": 0.1720760121742244, "grad_norm": 0.32421875, "learning_rate": 9.205025125628142e-05, "loss": 1.2199, "step": 1682 }, { "epoch": 0.17217831658098673, "grad_norm": 0.35546875, "learning_rate": 9.204522613065327e-05, "loss": 1.3086, "step": 1683 }, { "epoch": 0.17228062098774904, "grad_norm": 0.314453125, "learning_rate": 9.204020100502513e-05, "loss": 1.3145, "step": 1684 }, { "epoch": 0.17238292539451136, "grad_norm": 0.41015625, "learning_rate": 9.203517587939698e-05, "loss": 1.3096, "step": 1685 }, { "epoch": 0.17248522980127368, "grad_norm": 0.291015625, "learning_rate": 9.203015075376885e-05, "loss": 1.3241, "step": 1686 }, { "epoch": 0.17258753420803602, "grad_norm": 0.296875, "learning_rate": 9.20251256281407e-05, "loss": 1.3976, "step": 1687 }, { "epoch": 0.17268983861479834, "grad_norm": 0.33203125, "learning_rate": 9.202010050251257e-05, "loss": 1.3742, "step": 1688 }, { "epoch": 0.17279214302156065, "grad_norm": 0.3515625, "learning_rate": 9.201507537688443e-05, "loss": 1.3982, "step": 1689 }, { "epoch": 0.17289444742832297, "grad_norm": 0.26953125, "learning_rate": 9.201005025125629e-05, "loss": 1.243, "step": 1690 }, { "epoch": 0.17299675183508528, "grad_norm": 0.328125, "learning_rate": 9.200502512562814e-05, "loss": 1.3548, "step": 1691 }, { "epoch": 0.17309905624184763, "grad_norm": 0.279296875, "learning_rate": 9.200000000000001e-05, "loss": 1.2694, "step": 1692 }, { "epoch": 0.17320136064860994, "grad_norm": 0.337890625, "learning_rate": 9.199497487437186e-05, "loss": 1.1923, "step": 1693 }, { "epoch": 0.17330366505537226, "grad_norm": 0.3359375, "learning_rate": 9.198994974874372e-05, "loss": 1.3961, "step": 1694 }, { "epoch": 0.17340596946213457, "grad_norm": 0.326171875, "learning_rate": 9.198492462311559e-05, "loss": 1.4562, "step": 1695 }, { "epoch": 0.1735082738688969, "grad_norm": 0.271484375, "learning_rate": 9.197989949748745e-05, "loss": 1.2344, "step": 1696 }, { "epoch": 0.17361057827565923, "grad_norm": 0.3125, "learning_rate": 9.19748743718593e-05, "loss": 1.3808, "step": 1697 }, { "epoch": 0.17371288268242155, "grad_norm": 0.28125, "learning_rate": 9.196984924623117e-05, "loss": 1.2461, "step": 1698 }, { "epoch": 0.17381518708918386, "grad_norm": 0.33203125, "learning_rate": 9.196482412060302e-05, "loss": 1.3063, "step": 1699 }, { "epoch": 0.17391749149594618, "grad_norm": 0.345703125, "learning_rate": 9.195979899497488e-05, "loss": 1.4944, "step": 1700 }, { "epoch": 0.1740197959027085, "grad_norm": 0.3046875, "learning_rate": 9.195477386934673e-05, "loss": 1.2582, "step": 1701 }, { "epoch": 0.17412210030947084, "grad_norm": 0.33203125, "learning_rate": 9.19497487437186e-05, "loss": 1.2295, "step": 1702 }, { "epoch": 0.17422440471623316, "grad_norm": 0.333984375, "learning_rate": 9.194472361809046e-05, "loss": 1.4492, "step": 1703 }, { "epoch": 0.17432670912299547, "grad_norm": 0.30078125, "learning_rate": 9.193969849246232e-05, "loss": 1.2569, "step": 1704 }, { "epoch": 0.1744290135297578, "grad_norm": 0.337890625, "learning_rate": 9.193467336683418e-05, "loss": 1.3971, "step": 1705 }, { "epoch": 0.1745313179365201, "grad_norm": 0.330078125, "learning_rate": 9.192964824120603e-05, "loss": 1.2995, "step": 1706 }, { "epoch": 0.17463362234328245, "grad_norm": 0.26953125, "learning_rate": 9.192462311557789e-05, "loss": 1.2128, "step": 1707 }, { "epoch": 0.17473592675004476, "grad_norm": 0.421875, "learning_rate": 9.191959798994975e-05, "loss": 1.3353, "step": 1708 }, { "epoch": 0.17483823115680708, "grad_norm": 0.3515625, "learning_rate": 9.191457286432161e-05, "loss": 1.3012, "step": 1709 }, { "epoch": 0.1749405355635694, "grad_norm": 0.3203125, "learning_rate": 9.190954773869346e-05, "loss": 1.2806, "step": 1710 }, { "epoch": 0.1750428399703317, "grad_norm": 0.3515625, "learning_rate": 9.190452261306534e-05, "loss": 1.3339, "step": 1711 }, { "epoch": 0.17514514437709405, "grad_norm": 0.279296875, "learning_rate": 9.189949748743719e-05, "loss": 1.2492, "step": 1712 }, { "epoch": 0.17524744878385637, "grad_norm": 0.380859375, "learning_rate": 9.189447236180905e-05, "loss": 1.3901, "step": 1713 }, { "epoch": 0.17534975319061868, "grad_norm": 0.31640625, "learning_rate": 9.188944723618091e-05, "loss": 1.2139, "step": 1714 }, { "epoch": 0.175452057597381, "grad_norm": 0.357421875, "learning_rate": 9.188442211055277e-05, "loss": 1.3047, "step": 1715 }, { "epoch": 0.17555436200414332, "grad_norm": 0.30859375, "learning_rate": 9.187939698492462e-05, "loss": 1.3216, "step": 1716 }, { "epoch": 0.17565666641090566, "grad_norm": 0.2890625, "learning_rate": 9.187437185929648e-05, "loss": 1.3469, "step": 1717 }, { "epoch": 0.17575897081766798, "grad_norm": 0.318359375, "learning_rate": 9.186934673366835e-05, "loss": 1.2826, "step": 1718 }, { "epoch": 0.1758612752244303, "grad_norm": 0.287109375, "learning_rate": 9.186432160804021e-05, "loss": 1.201, "step": 1719 }, { "epoch": 0.1759635796311926, "grad_norm": 0.306640625, "learning_rate": 9.185929648241206e-05, "loss": 1.3041, "step": 1720 }, { "epoch": 0.17606588403795492, "grad_norm": 0.287109375, "learning_rate": 9.185427135678393e-05, "loss": 1.194, "step": 1721 }, { "epoch": 0.17616818844471727, "grad_norm": 0.34375, "learning_rate": 9.184924623115578e-05, "loss": 1.2185, "step": 1722 }, { "epoch": 0.17627049285147958, "grad_norm": 0.296875, "learning_rate": 9.184422110552764e-05, "loss": 1.2872, "step": 1723 }, { "epoch": 0.1763727972582419, "grad_norm": 0.287109375, "learning_rate": 9.18391959798995e-05, "loss": 1.3425, "step": 1724 }, { "epoch": 0.1764751016650042, "grad_norm": 0.2890625, "learning_rate": 9.183417085427137e-05, "loss": 1.2125, "step": 1725 }, { "epoch": 0.17657740607176653, "grad_norm": 0.302734375, "learning_rate": 9.182914572864321e-05, "loss": 1.2667, "step": 1726 }, { "epoch": 0.17667971047852887, "grad_norm": 0.29296875, "learning_rate": 9.182412060301509e-05, "loss": 1.1883, "step": 1727 }, { "epoch": 0.1767820148852912, "grad_norm": 0.30859375, "learning_rate": 9.181909547738694e-05, "loss": 1.3468, "step": 1728 }, { "epoch": 0.1768843192920535, "grad_norm": 0.294921875, "learning_rate": 9.18140703517588e-05, "loss": 1.419, "step": 1729 }, { "epoch": 0.17698662369881582, "grad_norm": 0.32421875, "learning_rate": 9.180904522613065e-05, "loss": 1.1533, "step": 1730 }, { "epoch": 0.17708892810557814, "grad_norm": 0.30078125, "learning_rate": 9.180402010050252e-05, "loss": 1.3651, "step": 1731 }, { "epoch": 0.17719123251234048, "grad_norm": 0.333984375, "learning_rate": 9.179899497487437e-05, "loss": 1.3139, "step": 1732 }, { "epoch": 0.1772935369191028, "grad_norm": 0.30078125, "learning_rate": 9.179396984924624e-05, "loss": 1.271, "step": 1733 }, { "epoch": 0.1773958413258651, "grad_norm": 0.2890625, "learning_rate": 9.17889447236181e-05, "loss": 1.2729, "step": 1734 }, { "epoch": 0.17749814573262743, "grad_norm": 0.298828125, "learning_rate": 9.178391959798996e-05, "loss": 1.4803, "step": 1735 }, { "epoch": 0.17760045013938974, "grad_norm": 0.314453125, "learning_rate": 9.177889447236181e-05, "loss": 1.2183, "step": 1736 }, { "epoch": 0.1777027545461521, "grad_norm": 0.29296875, "learning_rate": 9.177386934673367e-05, "loss": 1.3386, "step": 1737 }, { "epoch": 0.1778050589529144, "grad_norm": 0.294921875, "learning_rate": 9.176884422110553e-05, "loss": 1.2796, "step": 1738 }, { "epoch": 0.17790736335967672, "grad_norm": 0.3125, "learning_rate": 9.17638190954774e-05, "loss": 1.233, "step": 1739 }, { "epoch": 0.17800966776643903, "grad_norm": 0.2890625, "learning_rate": 9.175879396984926e-05, "loss": 1.3442, "step": 1740 }, { "epoch": 0.17811197217320135, "grad_norm": 0.337890625, "learning_rate": 9.17537688442211e-05, "loss": 1.3043, "step": 1741 }, { "epoch": 0.1782142765799637, "grad_norm": 0.322265625, "learning_rate": 9.174874371859297e-05, "loss": 1.4444, "step": 1742 }, { "epoch": 0.178316580986726, "grad_norm": 0.33203125, "learning_rate": 9.174371859296483e-05, "loss": 1.2418, "step": 1743 }, { "epoch": 0.17841888539348832, "grad_norm": 0.30859375, "learning_rate": 9.173869346733669e-05, "loss": 1.4208, "step": 1744 }, { "epoch": 0.17852118980025064, "grad_norm": 0.267578125, "learning_rate": 9.173366834170854e-05, "loss": 1.1904, "step": 1745 }, { "epoch": 0.17862349420701296, "grad_norm": 0.357421875, "learning_rate": 9.17286432160804e-05, "loss": 1.3898, "step": 1746 }, { "epoch": 0.1787257986137753, "grad_norm": 0.322265625, "learning_rate": 9.172361809045226e-05, "loss": 1.3413, "step": 1747 }, { "epoch": 0.17882810302053762, "grad_norm": 0.330078125, "learning_rate": 9.171859296482412e-05, "loss": 1.275, "step": 1748 }, { "epoch": 0.17893040742729993, "grad_norm": 0.298828125, "learning_rate": 9.171356783919597e-05, "loss": 1.297, "step": 1749 }, { "epoch": 0.17903271183406225, "grad_norm": 0.310546875, "learning_rate": 9.170854271356785e-05, "loss": 1.248, "step": 1750 }, { "epoch": 0.17913501624082456, "grad_norm": 0.3203125, "learning_rate": 9.17035175879397e-05, "loss": 1.3892, "step": 1751 }, { "epoch": 0.1792373206475869, "grad_norm": 0.3203125, "learning_rate": 9.169849246231156e-05, "loss": 1.2857, "step": 1752 }, { "epoch": 0.17933962505434922, "grad_norm": 0.296875, "learning_rate": 9.169346733668342e-05, "loss": 1.3273, "step": 1753 }, { "epoch": 0.17944192946111154, "grad_norm": 0.330078125, "learning_rate": 9.168844221105528e-05, "loss": 1.4495, "step": 1754 }, { "epoch": 0.17954423386787385, "grad_norm": 0.337890625, "learning_rate": 9.168341708542713e-05, "loss": 1.2872, "step": 1755 }, { "epoch": 0.17964653827463617, "grad_norm": 0.298828125, "learning_rate": 9.167839195979901e-05, "loss": 1.2667, "step": 1756 }, { "epoch": 0.1797488426813985, "grad_norm": 0.314453125, "learning_rate": 9.167336683417086e-05, "loss": 1.2202, "step": 1757 }, { "epoch": 0.17985114708816083, "grad_norm": 0.3203125, "learning_rate": 9.166834170854272e-05, "loss": 1.437, "step": 1758 }, { "epoch": 0.17995345149492314, "grad_norm": 0.294921875, "learning_rate": 9.166331658291458e-05, "loss": 1.3129, "step": 1759 }, { "epoch": 0.18005575590168546, "grad_norm": 0.296875, "learning_rate": 9.165829145728644e-05, "loss": 1.2678, "step": 1760 }, { "epoch": 0.18015806030844778, "grad_norm": 0.326171875, "learning_rate": 9.165326633165829e-05, "loss": 1.129, "step": 1761 }, { "epoch": 0.18026036471521012, "grad_norm": 0.30859375, "learning_rate": 9.164824120603015e-05, "loss": 1.3449, "step": 1762 }, { "epoch": 0.18036266912197244, "grad_norm": 0.318359375, "learning_rate": 9.164321608040201e-05, "loss": 1.4864, "step": 1763 }, { "epoch": 0.18046497352873475, "grad_norm": 0.31640625, "learning_rate": 9.163819095477388e-05, "loss": 1.2433, "step": 1764 }, { "epoch": 0.18056727793549707, "grad_norm": 0.28125, "learning_rate": 9.163316582914573e-05, "loss": 1.3965, "step": 1765 }, { "epoch": 0.18066958234225938, "grad_norm": 0.3515625, "learning_rate": 9.16281407035176e-05, "loss": 1.3814, "step": 1766 }, { "epoch": 0.18077188674902173, "grad_norm": 0.279296875, "learning_rate": 9.162311557788945e-05, "loss": 1.3571, "step": 1767 }, { "epoch": 0.18087419115578404, "grad_norm": 0.302734375, "learning_rate": 9.161809045226131e-05, "loss": 1.2272, "step": 1768 }, { "epoch": 0.18097649556254636, "grad_norm": 0.28515625, "learning_rate": 9.161306532663317e-05, "loss": 1.221, "step": 1769 }, { "epoch": 0.18107879996930867, "grad_norm": 0.271484375, "learning_rate": 9.160804020100504e-05, "loss": 1.3506, "step": 1770 }, { "epoch": 0.181181104376071, "grad_norm": 0.287109375, "learning_rate": 9.160301507537688e-05, "loss": 1.3588, "step": 1771 }, { "epoch": 0.18128340878283333, "grad_norm": 0.302734375, "learning_rate": 9.159798994974875e-05, "loss": 1.3389, "step": 1772 }, { "epoch": 0.18138571318959565, "grad_norm": 0.2734375, "learning_rate": 9.159296482412061e-05, "loss": 1.204, "step": 1773 }, { "epoch": 0.18148801759635796, "grad_norm": 0.310546875, "learning_rate": 9.158793969849247e-05, "loss": 1.3818, "step": 1774 }, { "epoch": 0.18159032200312028, "grad_norm": 0.298828125, "learning_rate": 9.158291457286433e-05, "loss": 1.2955, "step": 1775 }, { "epoch": 0.1816926264098826, "grad_norm": 0.302734375, "learning_rate": 9.157788944723618e-05, "loss": 1.2536, "step": 1776 }, { "epoch": 0.18179493081664494, "grad_norm": 0.56640625, "learning_rate": 9.157286432160804e-05, "loss": 1.4514, "step": 1777 }, { "epoch": 0.18189723522340726, "grad_norm": 0.287109375, "learning_rate": 9.15678391959799e-05, "loss": 1.2602, "step": 1778 }, { "epoch": 0.18199953963016957, "grad_norm": 0.34375, "learning_rate": 9.156281407035177e-05, "loss": 1.3923, "step": 1779 }, { "epoch": 0.1821018440369319, "grad_norm": 0.33203125, "learning_rate": 9.155778894472362e-05, "loss": 1.3078, "step": 1780 }, { "epoch": 0.1822041484436942, "grad_norm": 0.3125, "learning_rate": 9.155276381909548e-05, "loss": 1.4303, "step": 1781 }, { "epoch": 0.18230645285045655, "grad_norm": 0.326171875, "learning_rate": 9.154773869346734e-05, "loss": 1.2688, "step": 1782 }, { "epoch": 0.18240875725721886, "grad_norm": 0.29296875, "learning_rate": 9.15427135678392e-05, "loss": 2.0682, "step": 1783 }, { "epoch": 0.18251106166398118, "grad_norm": 0.322265625, "learning_rate": 9.153768844221105e-05, "loss": 1.2682, "step": 1784 }, { "epoch": 0.1826133660707435, "grad_norm": 0.36328125, "learning_rate": 9.153266331658293e-05, "loss": 1.6212, "step": 1785 }, { "epoch": 0.1827156704775058, "grad_norm": 0.35546875, "learning_rate": 9.152763819095477e-05, "loss": 1.3894, "step": 1786 }, { "epoch": 0.18281797488426815, "grad_norm": 0.3203125, "learning_rate": 9.152261306532664e-05, "loss": 1.3493, "step": 1787 }, { "epoch": 0.18292027929103047, "grad_norm": 0.314453125, "learning_rate": 9.15175879396985e-05, "loss": 1.3689, "step": 1788 }, { "epoch": 0.18302258369779278, "grad_norm": 0.310546875, "learning_rate": 9.151256281407036e-05, "loss": 1.3308, "step": 1789 }, { "epoch": 0.1831248881045551, "grad_norm": 0.361328125, "learning_rate": 9.150753768844221e-05, "loss": 1.4499, "step": 1790 }, { "epoch": 0.18322719251131742, "grad_norm": 0.291015625, "learning_rate": 9.150251256281408e-05, "loss": 1.3601, "step": 1791 }, { "epoch": 0.18332949691807976, "grad_norm": 0.296875, "learning_rate": 9.149748743718593e-05, "loss": 1.3126, "step": 1792 }, { "epoch": 0.18343180132484208, "grad_norm": 0.376953125, "learning_rate": 9.14924623115578e-05, "loss": 1.2775, "step": 1793 }, { "epoch": 0.1835341057316044, "grad_norm": 0.29296875, "learning_rate": 9.148743718592964e-05, "loss": 1.2917, "step": 1794 }, { "epoch": 0.1836364101383667, "grad_norm": 0.328125, "learning_rate": 9.148241206030152e-05, "loss": 1.3955, "step": 1795 }, { "epoch": 0.18373871454512902, "grad_norm": 0.32421875, "learning_rate": 9.147738693467337e-05, "loss": 1.2771, "step": 1796 }, { "epoch": 0.18384101895189137, "grad_norm": 0.2890625, "learning_rate": 9.147236180904523e-05, "loss": 1.3374, "step": 1797 }, { "epoch": 0.18394332335865368, "grad_norm": 0.3203125, "learning_rate": 9.146733668341709e-05, "loss": 1.4208, "step": 1798 }, { "epoch": 0.184045627765416, "grad_norm": 0.306640625, "learning_rate": 9.146231155778895e-05, "loss": 1.3981, "step": 1799 }, { "epoch": 0.1841479321721783, "grad_norm": 0.298828125, "learning_rate": 9.14572864321608e-05, "loss": 1.3801, "step": 1800 }, { "epoch": 0.18425023657894063, "grad_norm": 0.298828125, "learning_rate": 9.145226130653268e-05, "loss": 1.2722, "step": 1801 }, { "epoch": 0.18435254098570297, "grad_norm": 0.33203125, "learning_rate": 9.144723618090453e-05, "loss": 1.2895, "step": 1802 }, { "epoch": 0.1844548453924653, "grad_norm": 0.31640625, "learning_rate": 9.144221105527639e-05, "loss": 1.2018, "step": 1803 }, { "epoch": 0.1845571497992276, "grad_norm": 0.31640625, "learning_rate": 9.143718592964825e-05, "loss": 1.2419, "step": 1804 }, { "epoch": 0.18465945420598992, "grad_norm": 0.30859375, "learning_rate": 9.143216080402011e-05, "loss": 1.4157, "step": 1805 }, { "epoch": 0.18476175861275224, "grad_norm": 0.3203125, "learning_rate": 9.142713567839196e-05, "loss": 1.2267, "step": 1806 }, { "epoch": 0.18486406301951455, "grad_norm": 0.337890625, "learning_rate": 9.142211055276382e-05, "loss": 1.3648, "step": 1807 }, { "epoch": 0.1849663674262769, "grad_norm": 0.2734375, "learning_rate": 9.141708542713568e-05, "loss": 1.1896, "step": 1808 }, { "epoch": 0.1850686718330392, "grad_norm": 0.3671875, "learning_rate": 9.141206030150755e-05, "loss": 1.4298, "step": 1809 }, { "epoch": 0.18517097623980153, "grad_norm": 0.369140625, "learning_rate": 9.14070351758794e-05, "loss": 1.262, "step": 1810 }, { "epoch": 0.18527328064656384, "grad_norm": 0.2890625, "learning_rate": 9.140201005025126e-05, "loss": 1.4046, "step": 1811 }, { "epoch": 0.18537558505332616, "grad_norm": 0.3359375, "learning_rate": 9.139698492462312e-05, "loss": 1.3531, "step": 1812 }, { "epoch": 0.1854778894600885, "grad_norm": 0.3203125, "learning_rate": 9.139195979899498e-05, "loss": 1.3751, "step": 1813 }, { "epoch": 0.18558019386685082, "grad_norm": 0.34375, "learning_rate": 9.138693467336684e-05, "loss": 1.4116, "step": 1814 }, { "epoch": 0.18568249827361313, "grad_norm": 0.341796875, "learning_rate": 9.138190954773869e-05, "loss": 1.2862, "step": 1815 }, { "epoch": 0.18578480268037545, "grad_norm": 0.302734375, "learning_rate": 9.137688442211055e-05, "loss": 1.2613, "step": 1816 }, { "epoch": 0.18588710708713777, "grad_norm": 0.29296875, "learning_rate": 9.137185929648242e-05, "loss": 1.3423, "step": 1817 }, { "epoch": 0.1859894114939001, "grad_norm": 0.31640625, "learning_rate": 9.136683417085428e-05, "loss": 1.2318, "step": 1818 }, { "epoch": 0.18609171590066242, "grad_norm": 0.294921875, "learning_rate": 9.136180904522613e-05, "loss": 1.1601, "step": 1819 }, { "epoch": 0.18619402030742474, "grad_norm": 0.2734375, "learning_rate": 9.1356783919598e-05, "loss": 1.1958, "step": 1820 }, { "epoch": 0.18629632471418706, "grad_norm": 0.306640625, "learning_rate": 9.135175879396985e-05, "loss": 1.4096, "step": 1821 }, { "epoch": 0.18639862912094937, "grad_norm": 0.275390625, "learning_rate": 9.134673366834171e-05, "loss": 1.2596, "step": 1822 }, { "epoch": 0.18650093352771172, "grad_norm": 0.359375, "learning_rate": 9.134170854271357e-05, "loss": 1.324, "step": 1823 }, { "epoch": 0.18660323793447403, "grad_norm": 0.310546875, "learning_rate": 9.133668341708544e-05, "loss": 1.3728, "step": 1824 }, { "epoch": 0.18670554234123635, "grad_norm": 0.296875, "learning_rate": 9.133165829145728e-05, "loss": 1.3299, "step": 1825 }, { "epoch": 0.18680784674799866, "grad_norm": 0.29296875, "learning_rate": 9.132663316582915e-05, "loss": 1.2416, "step": 1826 }, { "epoch": 0.18691015115476098, "grad_norm": 0.29296875, "learning_rate": 9.132160804020101e-05, "loss": 1.3128, "step": 1827 }, { "epoch": 0.18701245556152332, "grad_norm": 0.328125, "learning_rate": 9.131658291457287e-05, "loss": 1.5047, "step": 1828 }, { "epoch": 0.18711475996828564, "grad_norm": 0.275390625, "learning_rate": 9.131155778894472e-05, "loss": 1.2577, "step": 1829 }, { "epoch": 0.18721706437504795, "grad_norm": 0.31640625, "learning_rate": 9.13065326633166e-05, "loss": 1.3849, "step": 1830 }, { "epoch": 0.18731936878181027, "grad_norm": 0.28125, "learning_rate": 9.130150753768844e-05, "loss": 1.2895, "step": 1831 }, { "epoch": 0.18742167318857259, "grad_norm": 0.283203125, "learning_rate": 9.12964824120603e-05, "loss": 1.3109, "step": 1832 }, { "epoch": 0.18752397759533493, "grad_norm": 0.302734375, "learning_rate": 9.129145728643217e-05, "loss": 1.2323, "step": 1833 }, { "epoch": 0.18762628200209724, "grad_norm": 0.3203125, "learning_rate": 9.128643216080403e-05, "loss": 1.3735, "step": 1834 }, { "epoch": 0.18772858640885956, "grad_norm": 0.2578125, "learning_rate": 9.128140703517588e-05, "loss": 1.133, "step": 1835 }, { "epoch": 0.18783089081562188, "grad_norm": 0.322265625, "learning_rate": 9.127638190954775e-05, "loss": 1.2754, "step": 1836 }, { "epoch": 0.1879331952223842, "grad_norm": 0.328125, "learning_rate": 9.12713567839196e-05, "loss": 1.2892, "step": 1837 }, { "epoch": 0.18803549962914654, "grad_norm": 0.302734375, "learning_rate": 9.126633165829146e-05, "loss": 1.2518, "step": 1838 }, { "epoch": 0.18813780403590885, "grad_norm": 0.3125, "learning_rate": 9.126130653266331e-05, "loss": 1.2133, "step": 1839 }, { "epoch": 0.18824010844267117, "grad_norm": 0.294921875, "learning_rate": 9.125628140703519e-05, "loss": 1.3176, "step": 1840 }, { "epoch": 0.18834241284943348, "grad_norm": 0.33203125, "learning_rate": 9.125125628140704e-05, "loss": 1.1976, "step": 1841 }, { "epoch": 0.1884447172561958, "grad_norm": 0.337890625, "learning_rate": 9.12462311557789e-05, "loss": 1.3181, "step": 1842 }, { "epoch": 0.18854702166295814, "grad_norm": 0.37109375, "learning_rate": 9.124120603015076e-05, "loss": 1.5471, "step": 1843 }, { "epoch": 0.18864932606972046, "grad_norm": 0.345703125, "learning_rate": 9.123618090452262e-05, "loss": 1.4038, "step": 1844 }, { "epoch": 0.18875163047648277, "grad_norm": 0.3203125, "learning_rate": 9.123115577889447e-05, "loss": 1.3686, "step": 1845 }, { "epoch": 0.1888539348832451, "grad_norm": 0.345703125, "learning_rate": 9.122613065326633e-05, "loss": 1.2725, "step": 1846 }, { "epoch": 0.1889562392900074, "grad_norm": 0.310546875, "learning_rate": 9.12211055276382e-05, "loss": 1.3187, "step": 1847 }, { "epoch": 0.18905854369676975, "grad_norm": 0.318359375, "learning_rate": 9.121608040201006e-05, "loss": 1.1635, "step": 1848 }, { "epoch": 0.18916084810353206, "grad_norm": 0.30078125, "learning_rate": 9.121105527638192e-05, "loss": 1.2876, "step": 1849 }, { "epoch": 0.18926315251029438, "grad_norm": 0.294921875, "learning_rate": 9.120603015075377e-05, "loss": 1.2888, "step": 1850 }, { "epoch": 0.1893654569170567, "grad_norm": 0.33984375, "learning_rate": 9.120100502512563e-05, "loss": 1.3443, "step": 1851 }, { "epoch": 0.189467761323819, "grad_norm": 0.353515625, "learning_rate": 9.119597989949749e-05, "loss": 1.3101, "step": 1852 }, { "epoch": 0.18957006573058136, "grad_norm": 0.298828125, "learning_rate": 9.119095477386935e-05, "loss": 1.2101, "step": 1853 }, { "epoch": 0.18967237013734367, "grad_norm": 0.30078125, "learning_rate": 9.11859296482412e-05, "loss": 1.3859, "step": 1854 }, { "epoch": 0.189774674544106, "grad_norm": 0.306640625, "learning_rate": 9.118090452261306e-05, "loss": 1.4062, "step": 1855 }, { "epoch": 0.1898769789508683, "grad_norm": 0.328125, "learning_rate": 9.117587939698493e-05, "loss": 1.2257, "step": 1856 }, { "epoch": 0.18997928335763062, "grad_norm": 0.29296875, "learning_rate": 9.117085427135679e-05, "loss": 1.1512, "step": 1857 }, { "epoch": 0.19008158776439296, "grad_norm": 0.296875, "learning_rate": 9.116582914572864e-05, "loss": 1.259, "step": 1858 }, { "epoch": 0.19018389217115528, "grad_norm": 0.294921875, "learning_rate": 9.116080402010051e-05, "loss": 1.3598, "step": 1859 }, { "epoch": 0.1902861965779176, "grad_norm": 0.359375, "learning_rate": 9.115577889447236e-05, "loss": 1.346, "step": 1860 }, { "epoch": 0.1903885009846799, "grad_norm": 0.302734375, "learning_rate": 9.115075376884422e-05, "loss": 1.3072, "step": 1861 }, { "epoch": 0.19049080539144223, "grad_norm": 0.2890625, "learning_rate": 9.114572864321609e-05, "loss": 1.2539, "step": 1862 }, { "epoch": 0.19059310979820457, "grad_norm": 0.31640625, "learning_rate": 9.114070351758795e-05, "loss": 1.5602, "step": 1863 }, { "epoch": 0.19069541420496688, "grad_norm": 0.34765625, "learning_rate": 9.11356783919598e-05, "loss": 1.5689, "step": 1864 }, { "epoch": 0.1907977186117292, "grad_norm": 0.291015625, "learning_rate": 9.113065326633167e-05, "loss": 1.2763, "step": 1865 }, { "epoch": 0.19090002301849152, "grad_norm": 0.310546875, "learning_rate": 9.112562814070352e-05, "loss": 1.3337, "step": 1866 }, { "epoch": 0.19100232742525383, "grad_norm": 0.318359375, "learning_rate": 9.112060301507538e-05, "loss": 1.4017, "step": 1867 }, { "epoch": 0.19110463183201618, "grad_norm": 0.31640625, "learning_rate": 9.111557788944724e-05, "loss": 1.1626, "step": 1868 }, { "epoch": 0.1912069362387785, "grad_norm": 0.287109375, "learning_rate": 9.11105527638191e-05, "loss": 1.2432, "step": 1869 }, { "epoch": 0.1913092406455408, "grad_norm": 0.294921875, "learning_rate": 9.110552763819095e-05, "loss": 1.3988, "step": 1870 }, { "epoch": 0.19141154505230312, "grad_norm": 0.310546875, "learning_rate": 9.110050251256282e-05, "loss": 1.2806, "step": 1871 }, { "epoch": 0.19151384945906544, "grad_norm": 0.32421875, "learning_rate": 9.109547738693468e-05, "loss": 1.3549, "step": 1872 }, { "epoch": 0.19161615386582778, "grad_norm": 0.29296875, "learning_rate": 9.109045226130654e-05, "loss": 1.4051, "step": 1873 }, { "epoch": 0.1917184582725901, "grad_norm": 0.3359375, "learning_rate": 9.108542713567839e-05, "loss": 1.4142, "step": 1874 }, { "epoch": 0.1918207626793524, "grad_norm": 0.314453125, "learning_rate": 9.108040201005026e-05, "loss": 1.358, "step": 1875 }, { "epoch": 0.19192306708611473, "grad_norm": 0.38671875, "learning_rate": 9.107537688442211e-05, "loss": 1.3958, "step": 1876 }, { "epoch": 0.19202537149287704, "grad_norm": 0.337890625, "learning_rate": 9.107035175879398e-05, "loss": 1.3837, "step": 1877 }, { "epoch": 0.1921276758996394, "grad_norm": 0.31640625, "learning_rate": 9.106532663316584e-05, "loss": 1.2804, "step": 1878 }, { "epoch": 0.1922299803064017, "grad_norm": 0.29296875, "learning_rate": 9.10603015075377e-05, "loss": 1.3235, "step": 1879 }, { "epoch": 0.19233228471316402, "grad_norm": 0.28515625, "learning_rate": 9.105527638190955e-05, "loss": 1.3627, "step": 1880 }, { "epoch": 0.19243458911992634, "grad_norm": 0.337890625, "learning_rate": 9.105025125628141e-05, "loss": 1.3471, "step": 1881 }, { "epoch": 0.19253689352668865, "grad_norm": 0.3359375, "learning_rate": 9.104522613065327e-05, "loss": 1.3054, "step": 1882 }, { "epoch": 0.192639197933451, "grad_norm": 0.337890625, "learning_rate": 9.104020100502513e-05, "loss": 1.3257, "step": 1883 }, { "epoch": 0.1927415023402133, "grad_norm": 0.3125, "learning_rate": 9.1035175879397e-05, "loss": 1.1562, "step": 1884 }, { "epoch": 0.19284380674697563, "grad_norm": 0.353515625, "learning_rate": 9.103015075376884e-05, "loss": 1.1333, "step": 1885 }, { "epoch": 0.19294611115373794, "grad_norm": 0.337890625, "learning_rate": 9.10251256281407e-05, "loss": 1.3051, "step": 1886 }, { "epoch": 0.19304841556050026, "grad_norm": 0.328125, "learning_rate": 9.102010050251257e-05, "loss": 1.2238, "step": 1887 }, { "epoch": 0.1931507199672626, "grad_norm": 0.32421875, "learning_rate": 9.101507537688443e-05, "loss": 1.3721, "step": 1888 }, { "epoch": 0.19325302437402492, "grad_norm": 0.29296875, "learning_rate": 9.101005025125628e-05, "loss": 1.3025, "step": 1889 }, { "epoch": 0.19335532878078723, "grad_norm": 0.3125, "learning_rate": 9.100502512562814e-05, "loss": 1.1809, "step": 1890 }, { "epoch": 0.19345763318754955, "grad_norm": 0.328125, "learning_rate": 9.1e-05, "loss": 1.3444, "step": 1891 }, { "epoch": 0.19355993759431186, "grad_norm": 0.2890625, "learning_rate": 9.099497487437186e-05, "loss": 1.2677, "step": 1892 }, { "epoch": 0.1936622420010742, "grad_norm": 0.2890625, "learning_rate": 9.098994974874371e-05, "loss": 1.2641, "step": 1893 }, { "epoch": 0.19376454640783652, "grad_norm": 0.30078125, "learning_rate": 9.098492462311559e-05, "loss": 1.3316, "step": 1894 }, { "epoch": 0.19386685081459884, "grad_norm": 0.28125, "learning_rate": 9.097989949748744e-05, "loss": 1.3412, "step": 1895 }, { "epoch": 0.19396915522136116, "grad_norm": 0.298828125, "learning_rate": 9.09748743718593e-05, "loss": 1.3454, "step": 1896 }, { "epoch": 0.19407145962812347, "grad_norm": 0.349609375, "learning_rate": 9.096984924623116e-05, "loss": 1.4278, "step": 1897 }, { "epoch": 0.19417376403488582, "grad_norm": 0.306640625, "learning_rate": 9.096482412060302e-05, "loss": 1.2006, "step": 1898 }, { "epoch": 0.19427606844164813, "grad_norm": 0.3125, "learning_rate": 9.095979899497487e-05, "loss": 1.3616, "step": 1899 }, { "epoch": 0.19437837284841045, "grad_norm": 0.337890625, "learning_rate": 9.095477386934675e-05, "loss": 1.4358, "step": 1900 }, { "epoch": 0.19448067725517276, "grad_norm": 0.326171875, "learning_rate": 9.09497487437186e-05, "loss": 1.3655, "step": 1901 }, { "epoch": 0.19458298166193508, "grad_norm": 0.296875, "learning_rate": 9.094472361809046e-05, "loss": 1.1665, "step": 1902 }, { "epoch": 0.19468528606869742, "grad_norm": 0.314453125, "learning_rate": 9.09396984924623e-05, "loss": 1.3715, "step": 1903 }, { "epoch": 0.19478759047545974, "grad_norm": 0.28125, "learning_rate": 9.093467336683418e-05, "loss": 1.4138, "step": 1904 }, { "epoch": 0.19488989488222205, "grad_norm": 0.302734375, "learning_rate": 9.092964824120603e-05, "loss": 1.1934, "step": 1905 }, { "epoch": 0.19499219928898437, "grad_norm": 0.32421875, "learning_rate": 9.092462311557789e-05, "loss": 1.3299, "step": 1906 }, { "epoch": 0.19509450369574668, "grad_norm": 0.318359375, "learning_rate": 9.091959798994975e-05, "loss": 1.2497, "step": 1907 }, { "epoch": 0.19519680810250903, "grad_norm": 0.294921875, "learning_rate": 9.091457286432162e-05, "loss": 1.2256, "step": 1908 }, { "epoch": 0.19529911250927134, "grad_norm": 0.30078125, "learning_rate": 9.090954773869347e-05, "loss": 1.182, "step": 1909 }, { "epoch": 0.19540141691603366, "grad_norm": 0.33203125, "learning_rate": 9.090452261306534e-05, "loss": 1.3239, "step": 1910 }, { "epoch": 0.19550372132279598, "grad_norm": 0.306640625, "learning_rate": 9.089949748743719e-05, "loss": 1.2864, "step": 1911 }, { "epoch": 0.1956060257295583, "grad_norm": 0.28125, "learning_rate": 9.089447236180905e-05, "loss": 1.1711, "step": 1912 }, { "epoch": 0.19570833013632064, "grad_norm": 0.306640625, "learning_rate": 9.088944723618091e-05, "loss": 1.3749, "step": 1913 }, { "epoch": 0.19581063454308295, "grad_norm": 0.310546875, "learning_rate": 9.088442211055278e-05, "loss": 1.4663, "step": 1914 }, { "epoch": 0.19591293894984527, "grad_norm": 0.294921875, "learning_rate": 9.087939698492462e-05, "loss": 1.2761, "step": 1915 }, { "epoch": 0.19601524335660758, "grad_norm": 0.2890625, "learning_rate": 9.087437185929649e-05, "loss": 1.2449, "step": 1916 }, { "epoch": 0.1961175477633699, "grad_norm": 0.318359375, "learning_rate": 9.086934673366835e-05, "loss": 1.3676, "step": 1917 }, { "epoch": 0.19621985217013224, "grad_norm": 0.30078125, "learning_rate": 9.086432160804021e-05, "loss": 1.335, "step": 1918 }, { "epoch": 0.19632215657689456, "grad_norm": 0.333984375, "learning_rate": 9.085929648241206e-05, "loss": 1.4627, "step": 1919 }, { "epoch": 0.19642446098365687, "grad_norm": 0.291015625, "learning_rate": 9.085427135678392e-05, "loss": 1.2172, "step": 1920 }, { "epoch": 0.1965267653904192, "grad_norm": 0.337890625, "learning_rate": 9.084924623115578e-05, "loss": 1.3467, "step": 1921 }, { "epoch": 0.1966290697971815, "grad_norm": 0.341796875, "learning_rate": 9.084422110552764e-05, "loss": 1.3214, "step": 1922 }, { "epoch": 0.19673137420394385, "grad_norm": 0.318359375, "learning_rate": 9.08391959798995e-05, "loss": 1.4099, "step": 1923 }, { "epoch": 0.19683367861070616, "grad_norm": 0.287109375, "learning_rate": 9.083417085427136e-05, "loss": 1.3046, "step": 1924 }, { "epoch": 0.19693598301746848, "grad_norm": 0.3125, "learning_rate": 9.082914572864322e-05, "loss": 1.4134, "step": 1925 }, { "epoch": 0.1970382874242308, "grad_norm": 0.306640625, "learning_rate": 9.082412060301508e-05, "loss": 1.2746, "step": 1926 }, { "epoch": 0.1971405918309931, "grad_norm": 0.31640625, "learning_rate": 9.081909547738694e-05, "loss": 1.2667, "step": 1927 }, { "epoch": 0.19724289623775546, "grad_norm": 0.384765625, "learning_rate": 9.081407035175879e-05, "loss": 1.1924, "step": 1928 }, { "epoch": 0.19734520064451777, "grad_norm": 0.3203125, "learning_rate": 9.080904522613067e-05, "loss": 1.3751, "step": 1929 }, { "epoch": 0.1974475050512801, "grad_norm": 0.28125, "learning_rate": 9.080402010050251e-05, "loss": 1.3946, "step": 1930 }, { "epoch": 0.1975498094580424, "grad_norm": 0.3046875, "learning_rate": 9.079899497487438e-05, "loss": 1.2784, "step": 1931 }, { "epoch": 0.19765211386480472, "grad_norm": 0.326171875, "learning_rate": 9.079396984924624e-05, "loss": 1.4042, "step": 1932 }, { "epoch": 0.19775441827156703, "grad_norm": 0.318359375, "learning_rate": 9.07889447236181e-05, "loss": 1.3185, "step": 1933 }, { "epoch": 0.19785672267832938, "grad_norm": 0.333984375, "learning_rate": 9.078391959798995e-05, "loss": 1.3514, "step": 1934 }, { "epoch": 0.1979590270850917, "grad_norm": 0.30859375, "learning_rate": 9.077889447236181e-05, "loss": 1.2139, "step": 1935 }, { "epoch": 0.198061331491854, "grad_norm": 0.31640625, "learning_rate": 9.077386934673367e-05, "loss": 1.4079, "step": 1936 }, { "epoch": 0.19816363589861632, "grad_norm": 0.3203125, "learning_rate": 9.076884422110553e-05, "loss": 1.2877, "step": 1937 }, { "epoch": 0.19826594030537864, "grad_norm": 0.29296875, "learning_rate": 9.076381909547738e-05, "loss": 1.2036, "step": 1938 }, { "epoch": 0.19836824471214098, "grad_norm": 0.30859375, "learning_rate": 9.075879396984926e-05, "loss": 1.1893, "step": 1939 }, { "epoch": 0.1984705491189033, "grad_norm": 0.365234375, "learning_rate": 9.075376884422111e-05, "loss": 1.343, "step": 1940 }, { "epoch": 0.19857285352566562, "grad_norm": 0.287109375, "learning_rate": 9.074874371859297e-05, "loss": 1.1982, "step": 1941 }, { "epoch": 0.19867515793242793, "grad_norm": 0.291015625, "learning_rate": 9.074371859296483e-05, "loss": 1.2391, "step": 1942 }, { "epoch": 0.19877746233919025, "grad_norm": 0.287109375, "learning_rate": 9.073869346733669e-05, "loss": 1.3099, "step": 1943 }, { "epoch": 0.1988797667459526, "grad_norm": 0.28515625, "learning_rate": 9.073366834170854e-05, "loss": 1.1498, "step": 1944 }, { "epoch": 0.1989820711527149, "grad_norm": 0.31640625, "learning_rate": 9.072864321608042e-05, "loss": 1.3714, "step": 1945 }, { "epoch": 0.19908437555947722, "grad_norm": 0.322265625, "learning_rate": 9.072361809045227e-05, "loss": 1.3595, "step": 1946 }, { "epoch": 0.19918667996623954, "grad_norm": 0.369140625, "learning_rate": 9.071859296482413e-05, "loss": 1.3612, "step": 1947 }, { "epoch": 0.19928898437300185, "grad_norm": 0.419921875, "learning_rate": 9.071356783919599e-05, "loss": 1.5744, "step": 1948 }, { "epoch": 0.1993912887797642, "grad_norm": 0.322265625, "learning_rate": 9.070854271356785e-05, "loss": 1.2913, "step": 1949 }, { "epoch": 0.1994935931865265, "grad_norm": 0.341796875, "learning_rate": 9.07035175879397e-05, "loss": 1.2631, "step": 1950 }, { "epoch": 0.19959589759328883, "grad_norm": 0.349609375, "learning_rate": 9.069849246231156e-05, "loss": 1.4542, "step": 1951 }, { "epoch": 0.19969820200005114, "grad_norm": 0.3046875, "learning_rate": 9.069346733668342e-05, "loss": 1.3186, "step": 1952 }, { "epoch": 0.19980050640681346, "grad_norm": 0.31640625, "learning_rate": 9.068844221105529e-05, "loss": 1.3571, "step": 1953 }, { "epoch": 0.1999028108135758, "grad_norm": 0.376953125, "learning_rate": 9.068341708542713e-05, "loss": 1.2792, "step": 1954 }, { "epoch": 0.20000511522033812, "grad_norm": 0.27734375, "learning_rate": 9.0678391959799e-05, "loss": 1.2678, "step": 1955 }, { "epoch": 0.20010741962710044, "grad_norm": 0.283203125, "learning_rate": 9.067336683417086e-05, "loss": 1.3041, "step": 1956 }, { "epoch": 0.20020972403386275, "grad_norm": 0.375, "learning_rate": 9.066834170854272e-05, "loss": 1.3326, "step": 1957 }, { "epoch": 0.20031202844062507, "grad_norm": 0.2890625, "learning_rate": 9.066331658291458e-05, "loss": 1.2095, "step": 1958 }, { "epoch": 0.2004143328473874, "grad_norm": 0.322265625, "learning_rate": 9.065829145728643e-05, "loss": 1.3881, "step": 1959 }, { "epoch": 0.20051663725414973, "grad_norm": 0.28515625, "learning_rate": 9.06532663316583e-05, "loss": 1.2477, "step": 1960 }, { "epoch": 0.20061894166091204, "grad_norm": 0.35546875, "learning_rate": 9.064824120603016e-05, "loss": 1.2735, "step": 1961 }, { "epoch": 0.20072124606767436, "grad_norm": 0.3046875, "learning_rate": 9.064321608040202e-05, "loss": 1.3607, "step": 1962 }, { "epoch": 0.20082355047443667, "grad_norm": 0.3203125, "learning_rate": 9.063819095477387e-05, "loss": 1.3117, "step": 1963 }, { "epoch": 0.20092585488119902, "grad_norm": 0.296875, "learning_rate": 9.063316582914573e-05, "loss": 1.2575, "step": 1964 }, { "epoch": 0.20102815928796133, "grad_norm": 0.33203125, "learning_rate": 9.062814070351759e-05, "loss": 1.2619, "step": 1965 }, { "epoch": 0.20113046369472365, "grad_norm": 0.318359375, "learning_rate": 9.062311557788945e-05, "loss": 1.4094, "step": 1966 }, { "epoch": 0.20123276810148596, "grad_norm": 0.31640625, "learning_rate": 9.06180904522613e-05, "loss": 1.3687, "step": 1967 }, { "epoch": 0.20133507250824828, "grad_norm": 0.283203125, "learning_rate": 9.061306532663318e-05, "loss": 1.3135, "step": 1968 }, { "epoch": 0.20143737691501062, "grad_norm": 0.357421875, "learning_rate": 9.060804020100502e-05, "loss": 1.3346, "step": 1969 }, { "epoch": 0.20153968132177294, "grad_norm": 0.29296875, "learning_rate": 9.060301507537689e-05, "loss": 1.2596, "step": 1970 }, { "epoch": 0.20164198572853526, "grad_norm": 0.32421875, "learning_rate": 9.059798994974875e-05, "loss": 1.2723, "step": 1971 }, { "epoch": 0.20174429013529757, "grad_norm": 0.310546875, "learning_rate": 9.059296482412061e-05, "loss": 1.2645, "step": 1972 }, { "epoch": 0.2018465945420599, "grad_norm": 0.31640625, "learning_rate": 9.058793969849246e-05, "loss": 1.3082, "step": 1973 }, { "epoch": 0.20194889894882223, "grad_norm": 0.30859375, "learning_rate": 9.058291457286433e-05, "loss": 1.234, "step": 1974 }, { "epoch": 0.20205120335558455, "grad_norm": 0.3203125, "learning_rate": 9.057788944723618e-05, "loss": 1.3674, "step": 1975 }, { "epoch": 0.20215350776234686, "grad_norm": 0.298828125, "learning_rate": 9.057286432160805e-05, "loss": 1.2796, "step": 1976 }, { "epoch": 0.20225581216910918, "grad_norm": 0.3046875, "learning_rate": 9.056783919597991e-05, "loss": 1.3276, "step": 1977 }, { "epoch": 0.2023581165758715, "grad_norm": 0.328125, "learning_rate": 9.056281407035177e-05, "loss": 1.3207, "step": 1978 }, { "epoch": 0.20246042098263384, "grad_norm": 0.314453125, "learning_rate": 9.055778894472362e-05, "loss": 1.4606, "step": 1979 }, { "epoch": 0.20256272538939615, "grad_norm": 0.296875, "learning_rate": 9.055276381909548e-05, "loss": 1.327, "step": 1980 }, { "epoch": 0.20266502979615847, "grad_norm": 0.318359375, "learning_rate": 9.054773869346734e-05, "loss": 1.2809, "step": 1981 }, { "epoch": 0.20276733420292078, "grad_norm": 0.328125, "learning_rate": 9.05427135678392e-05, "loss": 1.4314, "step": 1982 }, { "epoch": 0.2028696386096831, "grad_norm": 0.314453125, "learning_rate": 9.053768844221105e-05, "loss": 1.297, "step": 1983 }, { "epoch": 0.20297194301644544, "grad_norm": 0.2734375, "learning_rate": 9.053266331658293e-05, "loss": 1.2353, "step": 1984 }, { "epoch": 0.20307424742320776, "grad_norm": 0.302734375, "learning_rate": 9.052763819095478e-05, "loss": 1.3004, "step": 1985 }, { "epoch": 0.20317655182997008, "grad_norm": 0.30078125, "learning_rate": 9.052261306532664e-05, "loss": 1.2385, "step": 1986 }, { "epoch": 0.2032788562367324, "grad_norm": 0.32421875, "learning_rate": 9.05175879396985e-05, "loss": 1.4001, "step": 1987 }, { "epoch": 0.2033811606434947, "grad_norm": 0.283203125, "learning_rate": 9.051256281407036e-05, "loss": 1.2766, "step": 1988 }, { "epoch": 0.20348346505025705, "grad_norm": 0.45703125, "learning_rate": 9.050753768844221e-05, "loss": 1.4681, "step": 1989 }, { "epoch": 0.20358576945701937, "grad_norm": 0.30859375, "learning_rate": 9.050251256281407e-05, "loss": 1.2539, "step": 1990 }, { "epoch": 0.20368807386378168, "grad_norm": 0.302734375, "learning_rate": 9.049748743718594e-05, "loss": 1.3183, "step": 1991 }, { "epoch": 0.203790378270544, "grad_norm": 0.27734375, "learning_rate": 9.04924623115578e-05, "loss": 1.293, "step": 1992 }, { "epoch": 0.20389268267730631, "grad_norm": 0.296875, "learning_rate": 9.048743718592966e-05, "loss": 1.1698, "step": 1993 }, { "epoch": 0.20399498708406866, "grad_norm": 0.298828125, "learning_rate": 9.048241206030151e-05, "loss": 1.1652, "step": 1994 }, { "epoch": 0.20409729149083097, "grad_norm": 0.34375, "learning_rate": 9.047738693467337e-05, "loss": 1.1658, "step": 1995 }, { "epoch": 0.2041995958975933, "grad_norm": 0.310546875, "learning_rate": 9.047236180904523e-05, "loss": 1.4806, "step": 1996 }, { "epoch": 0.2043019003043556, "grad_norm": 0.310546875, "learning_rate": 9.04673366834171e-05, "loss": 1.3443, "step": 1997 }, { "epoch": 0.20440420471111792, "grad_norm": 0.306640625, "learning_rate": 9.046231155778894e-05, "loss": 1.3218, "step": 1998 }, { "epoch": 0.20450650911788026, "grad_norm": 0.3359375, "learning_rate": 9.04572864321608e-05, "loss": 1.1825, "step": 1999 }, { "epoch": 0.20460881352464258, "grad_norm": 0.298828125, "learning_rate": 9.045226130653267e-05, "loss": 1.3207, "step": 2000 }, { "epoch": 0.2047111179314049, "grad_norm": 0.3125, "learning_rate": 9.044723618090453e-05, "loss": 1.3915, "step": 2001 }, { "epoch": 0.2048134223381672, "grad_norm": 0.34375, "learning_rate": 9.044221105527638e-05, "loss": 1.2614, "step": 2002 }, { "epoch": 0.20491572674492953, "grad_norm": 0.318359375, "learning_rate": 9.043718592964825e-05, "loss": 1.1976, "step": 2003 }, { "epoch": 0.20501803115169187, "grad_norm": 0.310546875, "learning_rate": 9.04321608040201e-05, "loss": 1.3486, "step": 2004 }, { "epoch": 0.2051203355584542, "grad_norm": 0.306640625, "learning_rate": 9.042713567839196e-05, "loss": 1.4098, "step": 2005 }, { "epoch": 0.2052226399652165, "grad_norm": 0.328125, "learning_rate": 9.042211055276383e-05, "loss": 1.3424, "step": 2006 }, { "epoch": 0.20532494437197882, "grad_norm": 0.30859375, "learning_rate": 9.041708542713569e-05, "loss": 1.4767, "step": 2007 }, { "epoch": 0.20542724877874113, "grad_norm": 0.31640625, "learning_rate": 9.041206030150754e-05, "loss": 1.2861, "step": 2008 }, { "epoch": 0.20552955318550348, "grad_norm": 0.279296875, "learning_rate": 9.040703517587941e-05, "loss": 1.1353, "step": 2009 }, { "epoch": 0.2056318575922658, "grad_norm": 0.3125, "learning_rate": 9.040201005025126e-05, "loss": 1.2924, "step": 2010 }, { "epoch": 0.2057341619990281, "grad_norm": 0.28125, "learning_rate": 9.039698492462312e-05, "loss": 1.2415, "step": 2011 }, { "epoch": 0.20583646640579042, "grad_norm": 0.296875, "learning_rate": 9.039195979899497e-05, "loss": 1.2404, "step": 2012 }, { "epoch": 0.20593877081255274, "grad_norm": 0.29296875, "learning_rate": 9.038693467336685e-05, "loss": 1.2177, "step": 2013 }, { "epoch": 0.20604107521931508, "grad_norm": 0.333984375, "learning_rate": 9.03819095477387e-05, "loss": 1.4137, "step": 2014 }, { "epoch": 0.2061433796260774, "grad_norm": 0.296875, "learning_rate": 9.037688442211056e-05, "loss": 1.2525, "step": 2015 }, { "epoch": 0.20624568403283972, "grad_norm": 0.35546875, "learning_rate": 9.037185929648242e-05, "loss": 1.361, "step": 2016 }, { "epoch": 0.20634798843960203, "grad_norm": 0.314453125, "learning_rate": 9.036683417085428e-05, "loss": 1.2103, "step": 2017 }, { "epoch": 0.20645029284636435, "grad_norm": 0.294921875, "learning_rate": 9.036180904522613e-05, "loss": 1.2673, "step": 2018 }, { "epoch": 0.2065525972531267, "grad_norm": 0.3046875, "learning_rate": 9.0356783919598e-05, "loss": 1.4846, "step": 2019 }, { "epoch": 0.206654901659889, "grad_norm": 0.294921875, "learning_rate": 9.035175879396985e-05, "loss": 1.204, "step": 2020 }, { "epoch": 0.20675720606665132, "grad_norm": 0.294921875, "learning_rate": 9.034673366834172e-05, "loss": 1.4021, "step": 2021 }, { "epoch": 0.20685951047341364, "grad_norm": 0.32421875, "learning_rate": 9.034170854271358e-05, "loss": 1.2186, "step": 2022 }, { "epoch": 0.20696181488017595, "grad_norm": 0.287109375, "learning_rate": 9.033668341708544e-05, "loss": 1.2232, "step": 2023 }, { "epoch": 0.2070641192869383, "grad_norm": 0.296875, "learning_rate": 9.033165829145729e-05, "loss": 1.2634, "step": 2024 }, { "epoch": 0.2071664236937006, "grad_norm": 0.33203125, "learning_rate": 9.032663316582915e-05, "loss": 1.4685, "step": 2025 }, { "epoch": 0.20726872810046293, "grad_norm": 0.3203125, "learning_rate": 9.032160804020101e-05, "loss": 1.3344, "step": 2026 }, { "epoch": 0.20737103250722524, "grad_norm": 0.302734375, "learning_rate": 9.031658291457287e-05, "loss": 1.2402, "step": 2027 }, { "epoch": 0.20747333691398756, "grad_norm": 0.3125, "learning_rate": 9.031155778894472e-05, "loss": 1.2688, "step": 2028 }, { "epoch": 0.2075756413207499, "grad_norm": 0.333984375, "learning_rate": 9.030653266331658e-05, "loss": 1.273, "step": 2029 }, { "epoch": 0.20767794572751222, "grad_norm": 0.296875, "learning_rate": 9.030150753768845e-05, "loss": 1.1764, "step": 2030 }, { "epoch": 0.20778025013427454, "grad_norm": 0.341796875, "learning_rate": 9.02964824120603e-05, "loss": 1.2588, "step": 2031 }, { "epoch": 0.20788255454103685, "grad_norm": 0.34765625, "learning_rate": 9.029145728643217e-05, "loss": 1.4664, "step": 2032 }, { "epoch": 0.20798485894779917, "grad_norm": 0.26953125, "learning_rate": 9.028643216080402e-05, "loss": 1.3256, "step": 2033 }, { "epoch": 0.2080871633545615, "grad_norm": 0.302734375, "learning_rate": 9.028140703517588e-05, "loss": 1.3394, "step": 2034 }, { "epoch": 0.20818946776132383, "grad_norm": 0.32421875, "learning_rate": 9.027638190954774e-05, "loss": 1.2647, "step": 2035 }, { "epoch": 0.20829177216808614, "grad_norm": 0.3203125, "learning_rate": 9.02713567839196e-05, "loss": 1.3064, "step": 2036 }, { "epoch": 0.20839407657484846, "grad_norm": 0.3046875, "learning_rate": 9.026633165829145e-05, "loss": 1.2726, "step": 2037 }, { "epoch": 0.20849638098161077, "grad_norm": 0.8984375, "learning_rate": 9.026130653266333e-05, "loss": 1.33, "step": 2038 }, { "epoch": 0.20859868538837312, "grad_norm": 0.291015625, "learning_rate": 9.025628140703518e-05, "loss": 1.3325, "step": 2039 }, { "epoch": 0.20870098979513543, "grad_norm": 0.28515625, "learning_rate": 9.025125628140704e-05, "loss": 1.4136, "step": 2040 }, { "epoch": 0.20880329420189775, "grad_norm": 0.2490234375, "learning_rate": 9.02462311557789e-05, "loss": 1.1753, "step": 2041 }, { "epoch": 0.20890559860866006, "grad_norm": 0.298828125, "learning_rate": 9.024120603015076e-05, "loss": 1.1866, "step": 2042 }, { "epoch": 0.20900790301542238, "grad_norm": 0.330078125, "learning_rate": 9.023618090452261e-05, "loss": 1.2339, "step": 2043 }, { "epoch": 0.20911020742218472, "grad_norm": 0.302734375, "learning_rate": 9.023115577889447e-05, "loss": 1.2965, "step": 2044 }, { "epoch": 0.20921251182894704, "grad_norm": 0.322265625, "learning_rate": 9.022613065326634e-05, "loss": 1.3755, "step": 2045 }, { "epoch": 0.20931481623570936, "grad_norm": 0.32421875, "learning_rate": 9.02211055276382e-05, "loss": 1.4173, "step": 2046 }, { "epoch": 0.20941712064247167, "grad_norm": 0.3203125, "learning_rate": 9.021608040201005e-05, "loss": 1.4359, "step": 2047 }, { "epoch": 0.209519425049234, "grad_norm": 0.306640625, "learning_rate": 9.021105527638192e-05, "loss": 1.2875, "step": 2048 }, { "epoch": 0.20962172945599633, "grad_norm": 0.322265625, "learning_rate": 9.020603015075377e-05, "loss": 1.2629, "step": 2049 }, { "epoch": 0.20972403386275865, "grad_norm": 0.310546875, "learning_rate": 9.020100502512563e-05, "loss": 1.2109, "step": 2050 }, { "epoch": 0.20982633826952096, "grad_norm": 0.314453125, "learning_rate": 9.01959798994975e-05, "loss": 1.2753, "step": 2051 }, { "epoch": 0.20992864267628328, "grad_norm": 0.30859375, "learning_rate": 9.019095477386936e-05, "loss": 1.1911, "step": 2052 }, { "epoch": 0.2100309470830456, "grad_norm": 0.3125, "learning_rate": 9.01859296482412e-05, "loss": 1.2505, "step": 2053 }, { "epoch": 0.21013325148980794, "grad_norm": 0.51171875, "learning_rate": 9.018090452261308e-05, "loss": 1.5321, "step": 2054 }, { "epoch": 0.21023555589657025, "grad_norm": 0.341796875, "learning_rate": 9.017587939698493e-05, "loss": 1.4055, "step": 2055 }, { "epoch": 0.21033786030333257, "grad_norm": 0.341796875, "learning_rate": 9.017085427135679e-05, "loss": 1.3572, "step": 2056 }, { "epoch": 0.21044016471009488, "grad_norm": 0.318359375, "learning_rate": 9.016582914572865e-05, "loss": 1.5077, "step": 2057 }, { "epoch": 0.2105424691168572, "grad_norm": 0.3515625, "learning_rate": 9.016080402010052e-05, "loss": 1.4844, "step": 2058 }, { "epoch": 0.21064477352361952, "grad_norm": 0.31640625, "learning_rate": 9.015577889447236e-05, "loss": 1.3549, "step": 2059 }, { "epoch": 0.21074707793038186, "grad_norm": 0.3515625, "learning_rate": 9.015075376884423e-05, "loss": 1.2443, "step": 2060 }, { "epoch": 0.21084938233714418, "grad_norm": 0.32421875, "learning_rate": 9.014572864321609e-05, "loss": 1.2995, "step": 2061 }, { "epoch": 0.2109516867439065, "grad_norm": 0.3359375, "learning_rate": 9.014070351758795e-05, "loss": 1.1773, "step": 2062 }, { "epoch": 0.2110539911506688, "grad_norm": 0.287109375, "learning_rate": 9.01356783919598e-05, "loss": 1.352, "step": 2063 }, { "epoch": 0.21115629555743112, "grad_norm": 0.384765625, "learning_rate": 9.013065326633166e-05, "loss": 1.294, "step": 2064 }, { "epoch": 0.21125859996419347, "grad_norm": 0.283203125, "learning_rate": 9.012562814070352e-05, "loss": 1.3368, "step": 2065 }, { "epoch": 0.21136090437095578, "grad_norm": 0.32421875, "learning_rate": 9.012060301507537e-05, "loss": 1.2738, "step": 2066 }, { "epoch": 0.2114632087777181, "grad_norm": 0.326171875, "learning_rate": 9.011557788944725e-05, "loss": 1.4362, "step": 2067 }, { "epoch": 0.2115655131844804, "grad_norm": 0.2734375, "learning_rate": 9.01105527638191e-05, "loss": 1.3641, "step": 2068 }, { "epoch": 0.21166781759124273, "grad_norm": 0.3125, "learning_rate": 9.010552763819096e-05, "loss": 1.3745, "step": 2069 }, { "epoch": 0.21177012199800507, "grad_norm": 0.279296875, "learning_rate": 9.010050251256282e-05, "loss": 1.2297, "step": 2070 }, { "epoch": 0.2118724264047674, "grad_norm": 0.298828125, "learning_rate": 9.009547738693468e-05, "loss": 1.2202, "step": 2071 }, { "epoch": 0.2119747308115297, "grad_norm": 0.279296875, "learning_rate": 9.009045226130653e-05, "loss": 1.2316, "step": 2072 }, { "epoch": 0.21207703521829202, "grad_norm": 0.357421875, "learning_rate": 9.008542713567839e-05, "loss": 1.3687, "step": 2073 }, { "epoch": 0.21217933962505434, "grad_norm": 0.33203125, "learning_rate": 9.008040201005025e-05, "loss": 1.2848, "step": 2074 }, { "epoch": 0.21228164403181668, "grad_norm": 0.291015625, "learning_rate": 9.007537688442212e-05, "loss": 1.2104, "step": 2075 }, { "epoch": 0.212383948438579, "grad_norm": 0.306640625, "learning_rate": 9.007035175879396e-05, "loss": 1.3671, "step": 2076 }, { "epoch": 0.2124862528453413, "grad_norm": 0.283203125, "learning_rate": 9.006532663316584e-05, "loss": 1.2061, "step": 2077 }, { "epoch": 0.21258855725210363, "grad_norm": 0.328125, "learning_rate": 9.006030150753769e-05, "loss": 1.0694, "step": 2078 }, { "epoch": 0.21269086165886594, "grad_norm": 0.32421875, "learning_rate": 9.005527638190955e-05, "loss": 1.3455, "step": 2079 }, { "epoch": 0.2127931660656283, "grad_norm": 0.3046875, "learning_rate": 9.005025125628141e-05, "loss": 1.3924, "step": 2080 }, { "epoch": 0.2128954704723906, "grad_norm": 0.296875, "learning_rate": 9.004522613065327e-05, "loss": 1.1912, "step": 2081 }, { "epoch": 0.21299777487915292, "grad_norm": 0.30078125, "learning_rate": 9.004020100502512e-05, "loss": 1.3938, "step": 2082 }, { "epoch": 0.21310007928591523, "grad_norm": 0.349609375, "learning_rate": 9.0035175879397e-05, "loss": 1.4127, "step": 2083 }, { "epoch": 0.21320238369267755, "grad_norm": 0.267578125, "learning_rate": 9.003015075376885e-05, "loss": 1.3153, "step": 2084 }, { "epoch": 0.2133046880994399, "grad_norm": 0.318359375, "learning_rate": 9.002512562814071e-05, "loss": 1.3857, "step": 2085 }, { "epoch": 0.2134069925062022, "grad_norm": 0.296875, "learning_rate": 9.002010050251257e-05, "loss": 1.1757, "step": 2086 }, { "epoch": 0.21350929691296452, "grad_norm": 0.30078125, "learning_rate": 9.001507537688443e-05, "loss": 1.2431, "step": 2087 }, { "epoch": 0.21361160131972684, "grad_norm": 0.330078125, "learning_rate": 9.001005025125628e-05, "loss": 1.4681, "step": 2088 }, { "epoch": 0.21371390572648916, "grad_norm": 0.298828125, "learning_rate": 9.000502512562814e-05, "loss": 1.3761, "step": 2089 }, { "epoch": 0.2138162101332515, "grad_norm": 0.3125, "learning_rate": 9e-05, "loss": 1.2978, "step": 2090 }, { "epoch": 0.21391851454001382, "grad_norm": 0.33203125, "learning_rate": 8.999497487437187e-05, "loss": 1.3183, "step": 2091 }, { "epoch": 0.21402081894677613, "grad_norm": 0.369140625, "learning_rate": 8.998994974874372e-05, "loss": 1.488, "step": 2092 }, { "epoch": 0.21412312335353845, "grad_norm": 0.3046875, "learning_rate": 8.998492462311559e-05, "loss": 1.436, "step": 2093 }, { "epoch": 0.21422542776030076, "grad_norm": 0.283203125, "learning_rate": 8.997989949748744e-05, "loss": 1.1459, "step": 2094 }, { "epoch": 0.2143277321670631, "grad_norm": 0.349609375, "learning_rate": 8.99748743718593e-05, "loss": 1.2978, "step": 2095 }, { "epoch": 0.21443003657382542, "grad_norm": 0.2734375, "learning_rate": 8.996984924623116e-05, "loss": 1.2275, "step": 2096 }, { "epoch": 0.21453234098058774, "grad_norm": 0.35546875, "learning_rate": 8.996482412060301e-05, "loss": 1.345, "step": 2097 }, { "epoch": 0.21463464538735005, "grad_norm": 0.287109375, "learning_rate": 8.995979899497487e-05, "loss": 1.2951, "step": 2098 }, { "epoch": 0.21473694979411237, "grad_norm": 0.326171875, "learning_rate": 8.995477386934674e-05, "loss": 1.2303, "step": 2099 }, { "epoch": 0.2148392542008747, "grad_norm": 0.3203125, "learning_rate": 8.99497487437186e-05, "loss": 1.4191, "step": 2100 }, { "epoch": 0.21494155860763703, "grad_norm": 0.3046875, "learning_rate": 8.994472361809045e-05, "loss": 1.2987, "step": 2101 }, { "epoch": 0.21504386301439934, "grad_norm": 0.322265625, "learning_rate": 8.993969849246232e-05, "loss": 1.3087, "step": 2102 }, { "epoch": 0.21514616742116166, "grad_norm": 0.291015625, "learning_rate": 8.993467336683417e-05, "loss": 1.3447, "step": 2103 }, { "epoch": 0.21524847182792398, "grad_norm": 0.30859375, "learning_rate": 8.992964824120603e-05, "loss": 1.371, "step": 2104 }, { "epoch": 0.21535077623468632, "grad_norm": 0.3359375, "learning_rate": 8.992462311557788e-05, "loss": 1.4364, "step": 2105 }, { "epoch": 0.21545308064144864, "grad_norm": 0.306640625, "learning_rate": 8.991959798994976e-05, "loss": 1.3914, "step": 2106 }, { "epoch": 0.21555538504821095, "grad_norm": 0.322265625, "learning_rate": 8.99145728643216e-05, "loss": 1.333, "step": 2107 }, { "epoch": 0.21565768945497327, "grad_norm": 0.291015625, "learning_rate": 8.990954773869347e-05, "loss": 1.3417, "step": 2108 }, { "epoch": 0.21575999386173558, "grad_norm": 0.3125, "learning_rate": 8.990452261306533e-05, "loss": 1.3477, "step": 2109 }, { "epoch": 0.21586229826849793, "grad_norm": 0.275390625, "learning_rate": 8.989949748743719e-05, "loss": 1.1834, "step": 2110 }, { "epoch": 0.21596460267526024, "grad_norm": 0.318359375, "learning_rate": 8.989447236180904e-05, "loss": 1.4264, "step": 2111 }, { "epoch": 0.21606690708202256, "grad_norm": 0.27734375, "learning_rate": 8.988944723618092e-05, "loss": 1.3688, "step": 2112 }, { "epoch": 0.21616921148878487, "grad_norm": 0.328125, "learning_rate": 8.988442211055276e-05, "loss": 1.0705, "step": 2113 }, { "epoch": 0.2162715158955472, "grad_norm": 0.326171875, "learning_rate": 8.987939698492463e-05, "loss": 1.2983, "step": 2114 }, { "epoch": 0.21637382030230953, "grad_norm": 0.50390625, "learning_rate": 8.987437185929649e-05, "loss": 1.2808, "step": 2115 }, { "epoch": 0.21647612470907185, "grad_norm": 0.318359375, "learning_rate": 8.986934673366835e-05, "loss": 1.2684, "step": 2116 }, { "epoch": 0.21657842911583416, "grad_norm": 0.357421875, "learning_rate": 8.98643216080402e-05, "loss": 1.5127, "step": 2117 }, { "epoch": 0.21668073352259648, "grad_norm": 0.53125, "learning_rate": 8.985929648241207e-05, "loss": 1.4531, "step": 2118 }, { "epoch": 0.2167830379293588, "grad_norm": 0.32421875, "learning_rate": 8.985427135678392e-05, "loss": 1.4417, "step": 2119 }, { "epoch": 0.21688534233612114, "grad_norm": 0.35546875, "learning_rate": 8.984924623115579e-05, "loss": 1.2386, "step": 2120 }, { "epoch": 0.21698764674288346, "grad_norm": 0.283203125, "learning_rate": 8.984422110552763e-05, "loss": 1.2102, "step": 2121 }, { "epoch": 0.21708995114964577, "grad_norm": 0.3046875, "learning_rate": 8.983919597989951e-05, "loss": 1.3421, "step": 2122 }, { "epoch": 0.2171922555564081, "grad_norm": 0.3203125, "learning_rate": 8.983417085427136e-05, "loss": 1.2552, "step": 2123 }, { "epoch": 0.2172945599631704, "grad_norm": 0.30859375, "learning_rate": 8.982914572864322e-05, "loss": 1.3547, "step": 2124 }, { "epoch": 0.21739686436993275, "grad_norm": 0.294921875, "learning_rate": 8.982412060301508e-05, "loss": 1.2123, "step": 2125 }, { "epoch": 0.21749916877669506, "grad_norm": 0.255859375, "learning_rate": 8.981909547738694e-05, "loss": 1.2857, "step": 2126 }, { "epoch": 0.21760147318345738, "grad_norm": 0.322265625, "learning_rate": 8.981407035175879e-05, "loss": 1.2973, "step": 2127 }, { "epoch": 0.2177037775902197, "grad_norm": 0.337890625, "learning_rate": 8.980904522613067e-05, "loss": 1.3425, "step": 2128 }, { "epoch": 0.217806081996982, "grad_norm": 0.330078125, "learning_rate": 8.980402010050252e-05, "loss": 1.4658, "step": 2129 }, { "epoch": 0.21790838640374435, "grad_norm": 0.279296875, "learning_rate": 8.979899497487438e-05, "loss": 1.1998, "step": 2130 }, { "epoch": 0.21801069081050667, "grad_norm": 0.296875, "learning_rate": 8.979396984924624e-05, "loss": 1.3988, "step": 2131 }, { "epoch": 0.21811299521726898, "grad_norm": 0.314453125, "learning_rate": 8.978894472361809e-05, "loss": 1.3859, "step": 2132 }, { "epoch": 0.2182152996240313, "grad_norm": 0.314453125, "learning_rate": 8.978391959798995e-05, "loss": 1.4262, "step": 2133 }, { "epoch": 0.21831760403079362, "grad_norm": 0.330078125, "learning_rate": 8.977889447236181e-05, "loss": 1.378, "step": 2134 }, { "epoch": 0.21841990843755596, "grad_norm": 0.30859375, "learning_rate": 8.977386934673368e-05, "loss": 1.2466, "step": 2135 }, { "epoch": 0.21852221284431828, "grad_norm": 0.298828125, "learning_rate": 8.976884422110552e-05, "loss": 1.194, "step": 2136 }, { "epoch": 0.2186245172510806, "grad_norm": 0.294921875, "learning_rate": 8.976381909547739e-05, "loss": 1.3254, "step": 2137 }, { "epoch": 0.2187268216578429, "grad_norm": 0.302734375, "learning_rate": 8.975879396984925e-05, "loss": 1.3857, "step": 2138 }, { "epoch": 0.21882912606460522, "grad_norm": 0.3359375, "learning_rate": 8.975376884422111e-05, "loss": 1.3741, "step": 2139 }, { "epoch": 0.21893143047136757, "grad_norm": 0.349609375, "learning_rate": 8.974874371859296e-05, "loss": 1.1945, "step": 2140 }, { "epoch": 0.21903373487812988, "grad_norm": 0.294921875, "learning_rate": 8.974371859296483e-05, "loss": 1.3371, "step": 2141 }, { "epoch": 0.2191360392848922, "grad_norm": 0.275390625, "learning_rate": 8.973869346733668e-05, "loss": 1.2369, "step": 2142 }, { "epoch": 0.2192383436916545, "grad_norm": 0.3125, "learning_rate": 8.973366834170854e-05, "loss": 1.4174, "step": 2143 }, { "epoch": 0.21934064809841683, "grad_norm": 0.296875, "learning_rate": 8.97286432160804e-05, "loss": 1.4221, "step": 2144 }, { "epoch": 0.21944295250517917, "grad_norm": 0.3203125, "learning_rate": 8.972361809045227e-05, "loss": 1.2968, "step": 2145 }, { "epoch": 0.2195452569119415, "grad_norm": 0.32421875, "learning_rate": 8.971859296482412e-05, "loss": 1.187, "step": 2146 }, { "epoch": 0.2196475613187038, "grad_norm": 0.322265625, "learning_rate": 8.971356783919599e-05, "loss": 1.3433, "step": 2147 }, { "epoch": 0.21974986572546612, "grad_norm": 0.333984375, "learning_rate": 8.970854271356784e-05, "loss": 1.396, "step": 2148 }, { "epoch": 0.21985217013222844, "grad_norm": 0.333984375, "learning_rate": 8.97035175879397e-05, "loss": 1.3328, "step": 2149 }, { "epoch": 0.21995447453899078, "grad_norm": 0.306640625, "learning_rate": 8.969849246231157e-05, "loss": 1.2382, "step": 2150 }, { "epoch": 0.2200567789457531, "grad_norm": 0.271484375, "learning_rate": 8.969346733668343e-05, "loss": 1.2312, "step": 2151 }, { "epoch": 0.2201590833525154, "grad_norm": 0.384765625, "learning_rate": 8.968844221105528e-05, "loss": 1.3846, "step": 2152 }, { "epoch": 0.22026138775927773, "grad_norm": 0.341796875, "learning_rate": 8.968341708542714e-05, "loss": 1.49, "step": 2153 }, { "epoch": 0.22036369216604004, "grad_norm": 0.314453125, "learning_rate": 8.9678391959799e-05, "loss": 1.2359, "step": 2154 }, { "epoch": 0.2204659965728024, "grad_norm": 0.291015625, "learning_rate": 8.967336683417086e-05, "loss": 1.264, "step": 2155 }, { "epoch": 0.2205683009795647, "grad_norm": 0.328125, "learning_rate": 8.966834170854271e-05, "loss": 1.4617, "step": 2156 }, { "epoch": 0.22067060538632702, "grad_norm": 0.404296875, "learning_rate": 8.966331658291459e-05, "loss": 1.3678, "step": 2157 }, { "epoch": 0.22077290979308933, "grad_norm": 0.314453125, "learning_rate": 8.965829145728643e-05, "loss": 1.4072, "step": 2158 }, { "epoch": 0.22087521419985165, "grad_norm": 0.3359375, "learning_rate": 8.96532663316583e-05, "loss": 1.4649, "step": 2159 }, { "epoch": 0.220977518606614, "grad_norm": 0.357421875, "learning_rate": 8.964824120603016e-05, "loss": 1.3982, "step": 2160 }, { "epoch": 0.2210798230133763, "grad_norm": 0.291015625, "learning_rate": 8.964321608040202e-05, "loss": 1.3732, "step": 2161 }, { "epoch": 0.22118212742013862, "grad_norm": 0.318359375, "learning_rate": 8.963819095477387e-05, "loss": 1.533, "step": 2162 }, { "epoch": 0.22128443182690094, "grad_norm": 0.3125, "learning_rate": 8.963316582914573e-05, "loss": 1.23, "step": 2163 }, { "epoch": 0.22138673623366326, "grad_norm": 0.2890625, "learning_rate": 8.962814070351759e-05, "loss": 1.1676, "step": 2164 }, { "epoch": 0.2214890406404256, "grad_norm": 0.3125, "learning_rate": 8.962311557788945e-05, "loss": 1.5745, "step": 2165 }, { "epoch": 0.22159134504718792, "grad_norm": 0.291015625, "learning_rate": 8.961809045226132e-05, "loss": 1.3242, "step": 2166 }, { "epoch": 0.22169364945395023, "grad_norm": 0.30859375, "learning_rate": 8.961306532663317e-05, "loss": 1.244, "step": 2167 }, { "epoch": 0.22179595386071255, "grad_norm": 0.333984375, "learning_rate": 8.960804020100503e-05, "loss": 1.444, "step": 2168 }, { "epoch": 0.22189825826747486, "grad_norm": 0.357421875, "learning_rate": 8.960301507537689e-05, "loss": 1.3133, "step": 2169 }, { "epoch": 0.2220005626742372, "grad_norm": 0.369140625, "learning_rate": 8.959798994974875e-05, "loss": 1.3671, "step": 2170 }, { "epoch": 0.22210286708099952, "grad_norm": 0.326171875, "learning_rate": 8.95929648241206e-05, "loss": 1.2656, "step": 2171 }, { "epoch": 0.22220517148776184, "grad_norm": 0.349609375, "learning_rate": 8.958793969849246e-05, "loss": 1.3748, "step": 2172 }, { "epoch": 0.22230747589452415, "grad_norm": 0.34765625, "learning_rate": 8.958291457286432e-05, "loss": 1.2292, "step": 2173 }, { "epoch": 0.22240978030128647, "grad_norm": 0.302734375, "learning_rate": 8.957788944723619e-05, "loss": 1.4189, "step": 2174 }, { "epoch": 0.2225120847080488, "grad_norm": 0.302734375, "learning_rate": 8.957286432160803e-05, "loss": 1.2883, "step": 2175 }, { "epoch": 0.22261438911481113, "grad_norm": 0.27734375, "learning_rate": 8.956783919597991e-05, "loss": 1.3188, "step": 2176 }, { "epoch": 0.22271669352157344, "grad_norm": 0.306640625, "learning_rate": 8.956281407035176e-05, "loss": 1.275, "step": 2177 }, { "epoch": 0.22281899792833576, "grad_norm": 0.30078125, "learning_rate": 8.955778894472362e-05, "loss": 1.4773, "step": 2178 }, { "epoch": 0.22292130233509808, "grad_norm": 0.33984375, "learning_rate": 8.955276381909548e-05, "loss": 1.3933, "step": 2179 }, { "epoch": 0.2230236067418604, "grad_norm": 0.298828125, "learning_rate": 8.954773869346734e-05, "loss": 1.1767, "step": 2180 }, { "epoch": 0.22312591114862274, "grad_norm": 0.34375, "learning_rate": 8.95427135678392e-05, "loss": 1.4305, "step": 2181 }, { "epoch": 0.22322821555538505, "grad_norm": 0.306640625, "learning_rate": 8.953768844221107e-05, "loss": 1.166, "step": 2182 }, { "epoch": 0.22333051996214737, "grad_norm": 0.326171875, "learning_rate": 8.953266331658292e-05, "loss": 1.5651, "step": 2183 }, { "epoch": 0.22343282436890968, "grad_norm": 0.2890625, "learning_rate": 8.952763819095478e-05, "loss": 1.1262, "step": 2184 }, { "epoch": 0.223535128775672, "grad_norm": 0.3046875, "learning_rate": 8.952261306532663e-05, "loss": 1.237, "step": 2185 }, { "epoch": 0.22363743318243434, "grad_norm": 0.302734375, "learning_rate": 8.95175879396985e-05, "loss": 1.4067, "step": 2186 }, { "epoch": 0.22373973758919666, "grad_norm": 0.318359375, "learning_rate": 8.951256281407035e-05, "loss": 1.3759, "step": 2187 }, { "epoch": 0.22384204199595897, "grad_norm": 0.26953125, "learning_rate": 8.950753768844221e-05, "loss": 1.2976, "step": 2188 }, { "epoch": 0.2239443464027213, "grad_norm": 0.3046875, "learning_rate": 8.950251256281408e-05, "loss": 1.3606, "step": 2189 }, { "epoch": 0.2240466508094836, "grad_norm": 0.2890625, "learning_rate": 8.949748743718594e-05, "loss": 1.1233, "step": 2190 }, { "epoch": 0.22414895521624595, "grad_norm": 0.279296875, "learning_rate": 8.949246231155779e-05, "loss": 1.2726, "step": 2191 }, { "epoch": 0.22425125962300826, "grad_norm": 0.298828125, "learning_rate": 8.948743718592966e-05, "loss": 1.2684, "step": 2192 }, { "epoch": 0.22435356402977058, "grad_norm": 0.3046875, "learning_rate": 8.948241206030151e-05, "loss": 1.4078, "step": 2193 }, { "epoch": 0.2244558684365329, "grad_norm": 0.283203125, "learning_rate": 8.947738693467337e-05, "loss": 1.321, "step": 2194 }, { "epoch": 0.2245581728432952, "grad_norm": 0.298828125, "learning_rate": 8.947236180904523e-05, "loss": 1.3039, "step": 2195 }, { "epoch": 0.22466047725005756, "grad_norm": 0.337890625, "learning_rate": 8.94673366834171e-05, "loss": 1.3312, "step": 2196 }, { "epoch": 0.22476278165681987, "grad_norm": 0.330078125, "learning_rate": 8.946231155778895e-05, "loss": 1.3054, "step": 2197 }, { "epoch": 0.2248650860635822, "grad_norm": 0.271484375, "learning_rate": 8.945728643216081e-05, "loss": 1.184, "step": 2198 }, { "epoch": 0.2249673904703445, "grad_norm": 0.349609375, "learning_rate": 8.945226130653267e-05, "loss": 1.4471, "step": 2199 }, { "epoch": 0.22506969487710682, "grad_norm": 0.2890625, "learning_rate": 8.944723618090453e-05, "loss": 1.2415, "step": 2200 }, { "epoch": 0.22517199928386916, "grad_norm": 0.302734375, "learning_rate": 8.944221105527638e-05, "loss": 1.3041, "step": 2201 }, { "epoch": 0.22527430369063148, "grad_norm": 0.310546875, "learning_rate": 8.943718592964824e-05, "loss": 1.2962, "step": 2202 }, { "epoch": 0.2253766080973938, "grad_norm": 0.28515625, "learning_rate": 8.94321608040201e-05, "loss": 1.2353, "step": 2203 }, { "epoch": 0.2254789125041561, "grad_norm": 0.306640625, "learning_rate": 8.942713567839197e-05, "loss": 1.168, "step": 2204 }, { "epoch": 0.22558121691091843, "grad_norm": 0.359375, "learning_rate": 8.942211055276383e-05, "loss": 1.3574, "step": 2205 }, { "epoch": 0.22568352131768077, "grad_norm": 0.31640625, "learning_rate": 8.941708542713568e-05, "loss": 1.2858, "step": 2206 }, { "epoch": 0.22578582572444308, "grad_norm": 0.38671875, "learning_rate": 8.941206030150754e-05, "loss": 1.3231, "step": 2207 }, { "epoch": 0.2258881301312054, "grad_norm": 0.287109375, "learning_rate": 8.94070351758794e-05, "loss": 1.2293, "step": 2208 }, { "epoch": 0.22599043453796772, "grad_norm": 0.328125, "learning_rate": 8.940201005025126e-05, "loss": 1.354, "step": 2209 }, { "epoch": 0.22609273894473003, "grad_norm": 0.28515625, "learning_rate": 8.939698492462311e-05, "loss": 1.2387, "step": 2210 }, { "epoch": 0.22619504335149238, "grad_norm": 0.3046875, "learning_rate": 8.939195979899499e-05, "loss": 1.2641, "step": 2211 }, { "epoch": 0.2262973477582547, "grad_norm": 0.357421875, "learning_rate": 8.938693467336684e-05, "loss": 1.341, "step": 2212 }, { "epoch": 0.226399652165017, "grad_norm": 0.27734375, "learning_rate": 8.93819095477387e-05, "loss": 1.1686, "step": 2213 }, { "epoch": 0.22650195657177932, "grad_norm": 0.2890625, "learning_rate": 8.937688442211055e-05, "loss": 1.3455, "step": 2214 }, { "epoch": 0.22660426097854164, "grad_norm": 0.30859375, "learning_rate": 8.937185929648242e-05, "loss": 1.4182, "step": 2215 }, { "epoch": 0.22670656538530398, "grad_norm": 0.3046875, "learning_rate": 8.936683417085427e-05, "loss": 1.3055, "step": 2216 }, { "epoch": 0.2268088697920663, "grad_norm": 0.3984375, "learning_rate": 8.936180904522613e-05, "loss": 1.2774, "step": 2217 }, { "epoch": 0.2269111741988286, "grad_norm": 0.310546875, "learning_rate": 8.9356783919598e-05, "loss": 1.2455, "step": 2218 }, { "epoch": 0.22701347860559093, "grad_norm": 0.302734375, "learning_rate": 8.935175879396986e-05, "loss": 1.1941, "step": 2219 }, { "epoch": 0.22711578301235325, "grad_norm": 0.28125, "learning_rate": 8.93467336683417e-05, "loss": 1.2661, "step": 2220 }, { "epoch": 0.2272180874191156, "grad_norm": 0.3203125, "learning_rate": 8.934170854271358e-05, "loss": 1.3698, "step": 2221 }, { "epoch": 0.2273203918258779, "grad_norm": 0.267578125, "learning_rate": 8.933668341708543e-05, "loss": 1.2613, "step": 2222 }, { "epoch": 0.22742269623264022, "grad_norm": 0.3359375, "learning_rate": 8.933165829145729e-05, "loss": 1.2739, "step": 2223 }, { "epoch": 0.22752500063940254, "grad_norm": 0.30859375, "learning_rate": 8.932663316582915e-05, "loss": 1.2198, "step": 2224 }, { "epoch": 0.22762730504616485, "grad_norm": 0.34375, "learning_rate": 8.932160804020101e-05, "loss": 1.4704, "step": 2225 }, { "epoch": 0.2277296094529272, "grad_norm": 0.27734375, "learning_rate": 8.931658291457286e-05, "loss": 1.3201, "step": 2226 }, { "epoch": 0.2278319138596895, "grad_norm": 0.306640625, "learning_rate": 8.931155778894474e-05, "loss": 1.3618, "step": 2227 }, { "epoch": 0.22793421826645183, "grad_norm": 0.306640625, "learning_rate": 8.930653266331659e-05, "loss": 1.2951, "step": 2228 }, { "epoch": 0.22803652267321414, "grad_norm": 0.326171875, "learning_rate": 8.930150753768845e-05, "loss": 1.4667, "step": 2229 }, { "epoch": 0.22813882707997646, "grad_norm": 0.302734375, "learning_rate": 8.92964824120603e-05, "loss": 1.3523, "step": 2230 }, { "epoch": 0.2282411314867388, "grad_norm": 0.37109375, "learning_rate": 8.929145728643217e-05, "loss": 1.5443, "step": 2231 }, { "epoch": 0.22834343589350112, "grad_norm": 0.337890625, "learning_rate": 8.928643216080402e-05, "loss": 1.1923, "step": 2232 }, { "epoch": 0.22844574030026343, "grad_norm": 0.291015625, "learning_rate": 8.928140703517588e-05, "loss": 1.271, "step": 2233 }, { "epoch": 0.22854804470702575, "grad_norm": 0.283203125, "learning_rate": 8.927638190954775e-05, "loss": 1.2412, "step": 2234 }, { "epoch": 0.22865034911378807, "grad_norm": 0.2734375, "learning_rate": 8.927135678391961e-05, "loss": 1.1517, "step": 2235 }, { "epoch": 0.2287526535205504, "grad_norm": 0.27734375, "learning_rate": 8.926633165829146e-05, "loss": 1.2549, "step": 2236 }, { "epoch": 0.22885495792731272, "grad_norm": 0.28125, "learning_rate": 8.926130653266332e-05, "loss": 1.1289, "step": 2237 }, { "epoch": 0.22895726233407504, "grad_norm": 0.32421875, "learning_rate": 8.925628140703518e-05, "loss": 1.3875, "step": 2238 }, { "epoch": 0.22905956674083736, "grad_norm": 0.287109375, "learning_rate": 8.925125628140704e-05, "loss": 1.2833, "step": 2239 }, { "epoch": 0.22916187114759967, "grad_norm": 0.3046875, "learning_rate": 8.92462311557789e-05, "loss": 1.2983, "step": 2240 }, { "epoch": 0.22926417555436202, "grad_norm": 0.322265625, "learning_rate": 8.924120603015075e-05, "loss": 1.2805, "step": 2241 }, { "epoch": 0.22936647996112433, "grad_norm": 0.28125, "learning_rate": 8.923618090452261e-05, "loss": 1.2393, "step": 2242 }, { "epoch": 0.22946878436788665, "grad_norm": 0.337890625, "learning_rate": 8.923115577889448e-05, "loss": 1.5522, "step": 2243 }, { "epoch": 0.22957108877464896, "grad_norm": 0.30078125, "learning_rate": 8.922613065326634e-05, "loss": 1.3631, "step": 2244 }, { "epoch": 0.22967339318141128, "grad_norm": 0.30078125, "learning_rate": 8.922110552763819e-05, "loss": 1.2893, "step": 2245 }, { "epoch": 0.22977569758817362, "grad_norm": 0.357421875, "learning_rate": 8.921608040201005e-05, "loss": 1.4064, "step": 2246 }, { "epoch": 0.22987800199493594, "grad_norm": 0.306640625, "learning_rate": 8.921105527638191e-05, "loss": 1.2209, "step": 2247 }, { "epoch": 0.22998030640169825, "grad_norm": 0.349609375, "learning_rate": 8.920603015075377e-05, "loss": 1.4656, "step": 2248 }, { "epoch": 0.23008261080846057, "grad_norm": 0.330078125, "learning_rate": 8.920100502512562e-05, "loss": 1.5254, "step": 2249 }, { "epoch": 0.23018491521522289, "grad_norm": 0.345703125, "learning_rate": 8.91959798994975e-05, "loss": 1.3181, "step": 2250 }, { "epoch": 0.23028721962198523, "grad_norm": 0.310546875, "learning_rate": 8.919095477386935e-05, "loss": 1.222, "step": 2251 }, { "epoch": 0.23038952402874754, "grad_norm": 0.34765625, "learning_rate": 8.918592964824121e-05, "loss": 1.1647, "step": 2252 }, { "epoch": 0.23049182843550986, "grad_norm": 0.322265625, "learning_rate": 8.918090452261307e-05, "loss": 1.2895, "step": 2253 }, { "epoch": 0.23059413284227218, "grad_norm": 0.3203125, "learning_rate": 8.917587939698493e-05, "loss": 1.3574, "step": 2254 }, { "epoch": 0.2306964372490345, "grad_norm": 0.322265625, "learning_rate": 8.917085427135678e-05, "loss": 1.2063, "step": 2255 }, { "epoch": 0.23079874165579684, "grad_norm": 0.294921875, "learning_rate": 8.916582914572866e-05, "loss": 1.3906, "step": 2256 }, { "epoch": 0.23090104606255915, "grad_norm": 0.328125, "learning_rate": 8.91608040201005e-05, "loss": 1.2297, "step": 2257 }, { "epoch": 0.23100335046932147, "grad_norm": 0.322265625, "learning_rate": 8.915577889447237e-05, "loss": 1.34, "step": 2258 }, { "epoch": 0.23110565487608378, "grad_norm": 0.310546875, "learning_rate": 8.915075376884423e-05, "loss": 1.2502, "step": 2259 }, { "epoch": 0.2312079592828461, "grad_norm": 0.3125, "learning_rate": 8.914572864321609e-05, "loss": 1.3726, "step": 2260 }, { "epoch": 0.23131026368960844, "grad_norm": 0.326171875, "learning_rate": 8.914070351758794e-05, "loss": 1.2808, "step": 2261 }, { "epoch": 0.23141256809637076, "grad_norm": 0.326171875, "learning_rate": 8.91356783919598e-05, "loss": 1.2937, "step": 2262 }, { "epoch": 0.23151487250313307, "grad_norm": 0.66015625, "learning_rate": 8.913065326633166e-05, "loss": 1.3673, "step": 2263 }, { "epoch": 0.2316171769098954, "grad_norm": 0.3203125, "learning_rate": 8.912562814070353e-05, "loss": 1.3771, "step": 2264 }, { "epoch": 0.2317194813166577, "grad_norm": 0.33984375, "learning_rate": 8.912060301507537e-05, "loss": 1.3323, "step": 2265 }, { "epoch": 0.23182178572342005, "grad_norm": 0.3359375, "learning_rate": 8.911557788944725e-05, "loss": 1.4286, "step": 2266 }, { "epoch": 0.23192409013018236, "grad_norm": 0.318359375, "learning_rate": 8.91105527638191e-05, "loss": 1.4318, "step": 2267 }, { "epoch": 0.23202639453694468, "grad_norm": 0.296875, "learning_rate": 8.910552763819096e-05, "loss": 1.2079, "step": 2268 }, { "epoch": 0.232128698943707, "grad_norm": 0.3046875, "learning_rate": 8.910050251256282e-05, "loss": 1.0874, "step": 2269 }, { "epoch": 0.2322310033504693, "grad_norm": 0.328125, "learning_rate": 8.909547738693468e-05, "loss": 1.2283, "step": 2270 }, { "epoch": 0.23233330775723166, "grad_norm": 0.359375, "learning_rate": 8.909045226130653e-05, "loss": 1.4397, "step": 2271 }, { "epoch": 0.23243561216399397, "grad_norm": 0.271484375, "learning_rate": 8.90854271356784e-05, "loss": 1.1857, "step": 2272 }, { "epoch": 0.2325379165707563, "grad_norm": 0.2734375, "learning_rate": 8.908040201005026e-05, "loss": 1.18, "step": 2273 }, { "epoch": 0.2326402209775186, "grad_norm": 0.302734375, "learning_rate": 8.907537688442212e-05, "loss": 1.1835, "step": 2274 }, { "epoch": 0.23274252538428092, "grad_norm": 0.291015625, "learning_rate": 8.907035175879398e-05, "loss": 1.1926, "step": 2275 }, { "epoch": 0.23284482979104326, "grad_norm": 0.306640625, "learning_rate": 8.906532663316583e-05, "loss": 1.2627, "step": 2276 }, { "epoch": 0.23294713419780558, "grad_norm": 0.341796875, "learning_rate": 8.906030150753769e-05, "loss": 1.2658, "step": 2277 }, { "epoch": 0.2330494386045679, "grad_norm": 0.3359375, "learning_rate": 8.905527638190955e-05, "loss": 1.2217, "step": 2278 }, { "epoch": 0.2331517430113302, "grad_norm": 0.3125, "learning_rate": 8.905025125628142e-05, "loss": 1.3111, "step": 2279 }, { "epoch": 0.23325404741809252, "grad_norm": 0.357421875, "learning_rate": 8.904522613065326e-05, "loss": 1.4406, "step": 2280 }, { "epoch": 0.23335635182485487, "grad_norm": 0.314453125, "learning_rate": 8.904020100502513e-05, "loss": 1.4352, "step": 2281 }, { "epoch": 0.23345865623161718, "grad_norm": 0.33203125, "learning_rate": 8.903517587939699e-05, "loss": 1.3062, "step": 2282 }, { "epoch": 0.2335609606383795, "grad_norm": 0.29296875, "learning_rate": 8.903015075376885e-05, "loss": 1.3182, "step": 2283 }, { "epoch": 0.23366326504514182, "grad_norm": 0.27734375, "learning_rate": 8.90251256281407e-05, "loss": 1.1784, "step": 2284 }, { "epoch": 0.23376556945190413, "grad_norm": 0.29296875, "learning_rate": 8.902010050251257e-05, "loss": 1.3282, "step": 2285 }, { "epoch": 0.23386787385866648, "grad_norm": 0.349609375, "learning_rate": 8.901507537688442e-05, "loss": 1.4107, "step": 2286 }, { "epoch": 0.2339701782654288, "grad_norm": 0.302734375, "learning_rate": 8.901005025125628e-05, "loss": 1.4278, "step": 2287 }, { "epoch": 0.2340724826721911, "grad_norm": 0.294921875, "learning_rate": 8.900502512562815e-05, "loss": 1.4059, "step": 2288 }, { "epoch": 0.23417478707895342, "grad_norm": 0.322265625, "learning_rate": 8.900000000000001e-05, "loss": 1.4541, "step": 2289 }, { "epoch": 0.23427709148571574, "grad_norm": 0.2890625, "learning_rate": 8.899497487437186e-05, "loss": 1.2032, "step": 2290 }, { "epoch": 0.23437939589247808, "grad_norm": 0.345703125, "learning_rate": 8.898994974874373e-05, "loss": 1.3368, "step": 2291 }, { "epoch": 0.2344817002992404, "grad_norm": 0.296875, "learning_rate": 8.898492462311558e-05, "loss": 1.2588, "step": 2292 }, { "epoch": 0.2345840047060027, "grad_norm": 0.30859375, "learning_rate": 8.897989949748744e-05, "loss": 1.5248, "step": 2293 }, { "epoch": 0.23468630911276503, "grad_norm": 0.322265625, "learning_rate": 8.897487437185929e-05, "loss": 1.3509, "step": 2294 }, { "epoch": 0.23478861351952734, "grad_norm": 0.3828125, "learning_rate": 8.896984924623117e-05, "loss": 1.2052, "step": 2295 }, { "epoch": 0.2348909179262897, "grad_norm": 0.318359375, "learning_rate": 8.896482412060302e-05, "loss": 1.382, "step": 2296 }, { "epoch": 0.234993222333052, "grad_norm": 0.283203125, "learning_rate": 8.895979899497488e-05, "loss": 1.4202, "step": 2297 }, { "epoch": 0.23509552673981432, "grad_norm": 0.306640625, "learning_rate": 8.895477386934674e-05, "loss": 1.2853, "step": 2298 }, { "epoch": 0.23519783114657664, "grad_norm": 0.296875, "learning_rate": 8.89497487437186e-05, "loss": 1.2985, "step": 2299 }, { "epoch": 0.23530013555333895, "grad_norm": 0.33203125, "learning_rate": 8.894472361809045e-05, "loss": 1.398, "step": 2300 }, { "epoch": 0.2354024399601013, "grad_norm": 0.298828125, "learning_rate": 8.893969849246233e-05, "loss": 1.2192, "step": 2301 }, { "epoch": 0.2355047443668636, "grad_norm": 0.341796875, "learning_rate": 8.893467336683417e-05, "loss": 1.2473, "step": 2302 }, { "epoch": 0.23560704877362593, "grad_norm": 0.30859375, "learning_rate": 8.892964824120604e-05, "loss": 1.3319, "step": 2303 }, { "epoch": 0.23570935318038824, "grad_norm": 0.30859375, "learning_rate": 8.89246231155779e-05, "loss": 1.2781, "step": 2304 }, { "epoch": 0.23581165758715056, "grad_norm": 0.30859375, "learning_rate": 8.891959798994976e-05, "loss": 1.3943, "step": 2305 }, { "epoch": 0.23591396199391287, "grad_norm": 0.271484375, "learning_rate": 8.891457286432161e-05, "loss": 1.2693, "step": 2306 }, { "epoch": 0.23601626640067522, "grad_norm": 0.322265625, "learning_rate": 8.890954773869347e-05, "loss": 1.3676, "step": 2307 }, { "epoch": 0.23611857080743753, "grad_norm": 0.296875, "learning_rate": 8.890452261306533e-05, "loss": 1.3028, "step": 2308 }, { "epoch": 0.23622087521419985, "grad_norm": 0.294921875, "learning_rate": 8.88994974874372e-05, "loss": 1.2217, "step": 2309 }, { "epoch": 0.23632317962096216, "grad_norm": 0.296875, "learning_rate": 8.889447236180904e-05, "loss": 1.271, "step": 2310 }, { "epoch": 0.23642548402772448, "grad_norm": 0.302734375, "learning_rate": 8.88894472361809e-05, "loss": 1.2208, "step": 2311 }, { "epoch": 0.23652778843448682, "grad_norm": 0.30859375, "learning_rate": 8.888442211055277e-05, "loss": 1.493, "step": 2312 }, { "epoch": 0.23663009284124914, "grad_norm": 0.28515625, "learning_rate": 8.887939698492463e-05, "loss": 1.3257, "step": 2313 }, { "epoch": 0.23673239724801146, "grad_norm": 0.328125, "learning_rate": 8.887437185929649e-05, "loss": 1.2654, "step": 2314 }, { "epoch": 0.23683470165477377, "grad_norm": 0.265625, "learning_rate": 8.886934673366834e-05, "loss": 1.2585, "step": 2315 }, { "epoch": 0.2369370060615361, "grad_norm": 0.349609375, "learning_rate": 8.88643216080402e-05, "loss": 1.4337, "step": 2316 }, { "epoch": 0.23703931046829843, "grad_norm": 0.279296875, "learning_rate": 8.885929648241206e-05, "loss": 1.1464, "step": 2317 }, { "epoch": 0.23714161487506075, "grad_norm": 0.306640625, "learning_rate": 8.885427135678393e-05, "loss": 1.2744, "step": 2318 }, { "epoch": 0.23724391928182306, "grad_norm": 0.291015625, "learning_rate": 8.884924623115577e-05, "loss": 1.4869, "step": 2319 }, { "epoch": 0.23734622368858538, "grad_norm": 0.3125, "learning_rate": 8.884422110552765e-05, "loss": 1.3421, "step": 2320 }, { "epoch": 0.2374485280953477, "grad_norm": 0.34375, "learning_rate": 8.88391959798995e-05, "loss": 1.4379, "step": 2321 }, { "epoch": 0.23755083250211004, "grad_norm": 0.296875, "learning_rate": 8.883417085427136e-05, "loss": 1.3244, "step": 2322 }, { "epoch": 0.23765313690887235, "grad_norm": 0.2890625, "learning_rate": 8.882914572864321e-05, "loss": 1.3385, "step": 2323 }, { "epoch": 0.23775544131563467, "grad_norm": 0.3671875, "learning_rate": 8.882412060301508e-05, "loss": 1.192, "step": 2324 }, { "epoch": 0.23785774572239698, "grad_norm": 0.31640625, "learning_rate": 8.881909547738693e-05, "loss": 1.2999, "step": 2325 }, { "epoch": 0.2379600501291593, "grad_norm": 0.337890625, "learning_rate": 8.88140703517588e-05, "loss": 1.379, "step": 2326 }, { "epoch": 0.23806235453592164, "grad_norm": 0.298828125, "learning_rate": 8.880904522613066e-05, "loss": 1.2447, "step": 2327 }, { "epoch": 0.23816465894268396, "grad_norm": 0.330078125, "learning_rate": 8.880402010050252e-05, "loss": 1.2053, "step": 2328 }, { "epoch": 0.23826696334944628, "grad_norm": 0.322265625, "learning_rate": 8.879899497487437e-05, "loss": 1.3617, "step": 2329 }, { "epoch": 0.2383692677562086, "grad_norm": 0.353515625, "learning_rate": 8.879396984924624e-05, "loss": 1.3339, "step": 2330 }, { "epoch": 0.2384715721629709, "grad_norm": 0.318359375, "learning_rate": 8.878894472361809e-05, "loss": 1.3141, "step": 2331 }, { "epoch": 0.23857387656973325, "grad_norm": 0.337890625, "learning_rate": 8.878391959798995e-05, "loss": 1.2668, "step": 2332 }, { "epoch": 0.23867618097649557, "grad_norm": 0.34765625, "learning_rate": 8.877889447236182e-05, "loss": 1.4639, "step": 2333 }, { "epoch": 0.23877848538325788, "grad_norm": 0.267578125, "learning_rate": 8.877386934673368e-05, "loss": 1.2663, "step": 2334 }, { "epoch": 0.2388807897900202, "grad_norm": 0.34375, "learning_rate": 8.876884422110553e-05, "loss": 1.3883, "step": 2335 }, { "epoch": 0.23898309419678251, "grad_norm": 0.29296875, "learning_rate": 8.87638190954774e-05, "loss": 1.1962, "step": 2336 }, { "epoch": 0.23908539860354486, "grad_norm": 0.291015625, "learning_rate": 8.875879396984925e-05, "loss": 1.2374, "step": 2337 }, { "epoch": 0.23918770301030717, "grad_norm": 0.3359375, "learning_rate": 8.875376884422111e-05, "loss": 1.4339, "step": 2338 }, { "epoch": 0.2392900074170695, "grad_norm": 0.3046875, "learning_rate": 8.874874371859296e-05, "loss": 1.2313, "step": 2339 }, { "epoch": 0.2393923118238318, "grad_norm": 0.30859375, "learning_rate": 8.874371859296484e-05, "loss": 1.2587, "step": 2340 }, { "epoch": 0.23949461623059412, "grad_norm": 0.359375, "learning_rate": 8.873869346733669e-05, "loss": 1.2651, "step": 2341 }, { "epoch": 0.23959692063735646, "grad_norm": 0.30859375, "learning_rate": 8.873366834170855e-05, "loss": 1.3698, "step": 2342 }, { "epoch": 0.23969922504411878, "grad_norm": 0.318359375, "learning_rate": 8.872864321608041e-05, "loss": 1.2214, "step": 2343 }, { "epoch": 0.2398015294508811, "grad_norm": 0.380859375, "learning_rate": 8.872361809045227e-05, "loss": 1.4404, "step": 2344 }, { "epoch": 0.2399038338576434, "grad_norm": 0.330078125, "learning_rate": 8.871859296482412e-05, "loss": 1.2665, "step": 2345 }, { "epoch": 0.24000613826440573, "grad_norm": 0.404296875, "learning_rate": 8.871356783919598e-05, "loss": 1.4166, "step": 2346 }, { "epoch": 0.24010844267116807, "grad_norm": 0.37890625, "learning_rate": 8.870854271356784e-05, "loss": 1.3424, "step": 2347 }, { "epoch": 0.2402107470779304, "grad_norm": 0.31640625, "learning_rate": 8.87035175879397e-05, "loss": 1.3546, "step": 2348 }, { "epoch": 0.2403130514846927, "grad_norm": 0.314453125, "learning_rate": 8.869849246231157e-05, "loss": 1.2761, "step": 2349 }, { "epoch": 0.24041535589145502, "grad_norm": 0.279296875, "learning_rate": 8.869346733668342e-05, "loss": 1.3518, "step": 2350 }, { "epoch": 0.24051766029821733, "grad_norm": 0.337890625, "learning_rate": 8.868844221105528e-05, "loss": 1.3681, "step": 2351 }, { "epoch": 0.24061996470497968, "grad_norm": 0.314453125, "learning_rate": 8.868341708542714e-05, "loss": 1.2492, "step": 2352 }, { "epoch": 0.240722269111742, "grad_norm": 0.310546875, "learning_rate": 8.8678391959799e-05, "loss": 1.2725, "step": 2353 }, { "epoch": 0.2408245735185043, "grad_norm": 0.3125, "learning_rate": 8.867336683417085e-05, "loss": 1.2389, "step": 2354 }, { "epoch": 0.24092687792526662, "grad_norm": 0.283203125, "learning_rate": 8.866834170854271e-05, "loss": 1.3316, "step": 2355 }, { "epoch": 0.24102918233202894, "grad_norm": 0.294921875, "learning_rate": 8.866331658291458e-05, "loss": 1.2479, "step": 2356 }, { "epoch": 0.24113148673879128, "grad_norm": 0.3203125, "learning_rate": 8.865829145728644e-05, "loss": 1.2899, "step": 2357 }, { "epoch": 0.2412337911455536, "grad_norm": 0.296875, "learning_rate": 8.865326633165829e-05, "loss": 1.37, "step": 2358 }, { "epoch": 0.24133609555231592, "grad_norm": 0.294921875, "learning_rate": 8.864824120603016e-05, "loss": 1.1622, "step": 2359 }, { "epoch": 0.24143839995907823, "grad_norm": 0.30078125, "learning_rate": 8.864321608040201e-05, "loss": 1.283, "step": 2360 }, { "epoch": 0.24154070436584055, "grad_norm": 0.3125, "learning_rate": 8.863819095477387e-05, "loss": 1.236, "step": 2361 }, { "epoch": 0.2416430087726029, "grad_norm": 0.337890625, "learning_rate": 8.863316582914573e-05, "loss": 1.4388, "step": 2362 }, { "epoch": 0.2417453131793652, "grad_norm": 0.28515625, "learning_rate": 8.86281407035176e-05, "loss": 1.2878, "step": 2363 }, { "epoch": 0.24184761758612752, "grad_norm": 0.306640625, "learning_rate": 8.862311557788944e-05, "loss": 1.2116, "step": 2364 }, { "epoch": 0.24194992199288984, "grad_norm": 0.2890625, "learning_rate": 8.861809045226132e-05, "loss": 1.0578, "step": 2365 }, { "epoch": 0.24205222639965215, "grad_norm": 0.326171875, "learning_rate": 8.861306532663317e-05, "loss": 1.4497, "step": 2366 }, { "epoch": 0.2421545308064145, "grad_norm": 0.33984375, "learning_rate": 8.860804020100503e-05, "loss": 1.3916, "step": 2367 }, { "epoch": 0.2422568352131768, "grad_norm": 0.34765625, "learning_rate": 8.860301507537689e-05, "loss": 1.4761, "step": 2368 }, { "epoch": 0.24235913961993913, "grad_norm": 0.306640625, "learning_rate": 8.859798994974875e-05, "loss": 1.3576, "step": 2369 }, { "epoch": 0.24246144402670144, "grad_norm": 0.28515625, "learning_rate": 8.85929648241206e-05, "loss": 1.1123, "step": 2370 }, { "epoch": 0.24256374843346376, "grad_norm": 0.298828125, "learning_rate": 8.858793969849246e-05, "loss": 1.336, "step": 2371 }, { "epoch": 0.2426660528402261, "grad_norm": 0.3125, "learning_rate": 8.858291457286433e-05, "loss": 1.2926, "step": 2372 }, { "epoch": 0.24276835724698842, "grad_norm": 0.30859375, "learning_rate": 8.857788944723619e-05, "loss": 1.2395, "step": 2373 }, { "epoch": 0.24287066165375074, "grad_norm": 0.345703125, "learning_rate": 8.857286432160804e-05, "loss": 1.2775, "step": 2374 }, { "epoch": 0.24297296606051305, "grad_norm": 0.3046875, "learning_rate": 8.856783919597991e-05, "loss": 1.2714, "step": 2375 }, { "epoch": 0.24307527046727537, "grad_norm": 0.29296875, "learning_rate": 8.856281407035176e-05, "loss": 1.2552, "step": 2376 }, { "epoch": 0.2431775748740377, "grad_norm": 0.30078125, "learning_rate": 8.855778894472362e-05, "loss": 1.2561, "step": 2377 }, { "epoch": 0.24327987928080003, "grad_norm": 0.33203125, "learning_rate": 8.855276381909549e-05, "loss": 1.3577, "step": 2378 }, { "epoch": 0.24338218368756234, "grad_norm": 0.32421875, "learning_rate": 8.854773869346735e-05, "loss": 1.1852, "step": 2379 }, { "epoch": 0.24348448809432466, "grad_norm": 0.291015625, "learning_rate": 8.85427135678392e-05, "loss": 1.2295, "step": 2380 }, { "epoch": 0.24358679250108697, "grad_norm": 0.26953125, "learning_rate": 8.853768844221106e-05, "loss": 1.1819, "step": 2381 }, { "epoch": 0.24368909690784932, "grad_norm": 0.306640625, "learning_rate": 8.853266331658292e-05, "loss": 1.327, "step": 2382 }, { "epoch": 0.24379140131461163, "grad_norm": 0.353515625, "learning_rate": 8.852763819095478e-05, "loss": 1.4168, "step": 2383 }, { "epoch": 0.24389370572137395, "grad_norm": 0.3359375, "learning_rate": 8.852261306532664e-05, "loss": 1.3866, "step": 2384 }, { "epoch": 0.24399601012813626, "grad_norm": 0.291015625, "learning_rate": 8.851758793969849e-05, "loss": 1.2381, "step": 2385 }, { "epoch": 0.24409831453489858, "grad_norm": 0.310546875, "learning_rate": 8.851256281407035e-05, "loss": 1.3144, "step": 2386 }, { "epoch": 0.24420061894166092, "grad_norm": 0.296875, "learning_rate": 8.850753768844222e-05, "loss": 1.2372, "step": 2387 }, { "epoch": 0.24430292334842324, "grad_norm": 0.318359375, "learning_rate": 8.850251256281408e-05, "loss": 1.2327, "step": 2388 }, { "epoch": 0.24440522775518556, "grad_norm": 0.294921875, "learning_rate": 8.849748743718593e-05, "loss": 1.4287, "step": 2389 }, { "epoch": 0.24450753216194787, "grad_norm": 0.3046875, "learning_rate": 8.849246231155779e-05, "loss": 1.3101, "step": 2390 }, { "epoch": 0.2446098365687102, "grad_norm": 0.314453125, "learning_rate": 8.848743718592965e-05, "loss": 1.247, "step": 2391 }, { "epoch": 0.24471214097547253, "grad_norm": 0.32421875, "learning_rate": 8.848241206030151e-05, "loss": 1.2632, "step": 2392 }, { "epoch": 0.24481444538223485, "grad_norm": 0.3125, "learning_rate": 8.847738693467336e-05, "loss": 1.2876, "step": 2393 }, { "epoch": 0.24491674978899716, "grad_norm": 0.298828125, "learning_rate": 8.847236180904524e-05, "loss": 1.3558, "step": 2394 }, { "epoch": 0.24501905419575948, "grad_norm": 0.318359375, "learning_rate": 8.846733668341709e-05, "loss": 1.4837, "step": 2395 }, { "epoch": 0.2451213586025218, "grad_norm": 0.306640625, "learning_rate": 8.846231155778895e-05, "loss": 1.2343, "step": 2396 }, { "epoch": 0.24522366300928414, "grad_norm": 0.33984375, "learning_rate": 8.845728643216081e-05, "loss": 1.4627, "step": 2397 }, { "epoch": 0.24532596741604645, "grad_norm": 0.326171875, "learning_rate": 8.845226130653267e-05, "loss": 1.2122, "step": 2398 }, { "epoch": 0.24542827182280877, "grad_norm": 0.306640625, "learning_rate": 8.844723618090452e-05, "loss": 1.2573, "step": 2399 }, { "epoch": 0.24553057622957108, "grad_norm": 0.28515625, "learning_rate": 8.84422110552764e-05, "loss": 1.3786, "step": 2400 }, { "epoch": 0.2456328806363334, "grad_norm": 0.32421875, "learning_rate": 8.843718592964824e-05, "loss": 1.3175, "step": 2401 }, { "epoch": 0.24573518504309574, "grad_norm": 0.349609375, "learning_rate": 8.84321608040201e-05, "loss": 1.2314, "step": 2402 }, { "epoch": 0.24583748944985806, "grad_norm": 0.330078125, "learning_rate": 8.842713567839196e-05, "loss": 1.3994, "step": 2403 }, { "epoch": 0.24593979385662038, "grad_norm": 0.328125, "learning_rate": 8.842211055276383e-05, "loss": 1.3168, "step": 2404 }, { "epoch": 0.2460420982633827, "grad_norm": 0.26171875, "learning_rate": 8.841708542713568e-05, "loss": 1.0834, "step": 2405 }, { "epoch": 0.246144402670145, "grad_norm": 0.27734375, "learning_rate": 8.841206030150754e-05, "loss": 1.2019, "step": 2406 }, { "epoch": 0.24624670707690735, "grad_norm": 0.396484375, "learning_rate": 8.84070351758794e-05, "loss": 1.3916, "step": 2407 }, { "epoch": 0.24634901148366967, "grad_norm": 0.3515625, "learning_rate": 8.840201005025127e-05, "loss": 1.232, "step": 2408 }, { "epoch": 0.24645131589043198, "grad_norm": 0.28125, "learning_rate": 8.839698492462311e-05, "loss": 1.2637, "step": 2409 }, { "epoch": 0.2465536202971943, "grad_norm": 0.32421875, "learning_rate": 8.839195979899499e-05, "loss": 1.5129, "step": 2410 }, { "epoch": 0.2466559247039566, "grad_norm": 0.279296875, "learning_rate": 8.838693467336684e-05, "loss": 1.761, "step": 2411 }, { "epoch": 0.24675822911071896, "grad_norm": 0.2890625, "learning_rate": 8.83819095477387e-05, "loss": 1.2509, "step": 2412 }, { "epoch": 0.24686053351748127, "grad_norm": 0.31640625, "learning_rate": 8.837688442211056e-05, "loss": 1.2109, "step": 2413 }, { "epoch": 0.2469628379242436, "grad_norm": 0.275390625, "learning_rate": 8.837185929648242e-05, "loss": 1.2738, "step": 2414 }, { "epoch": 0.2470651423310059, "grad_norm": 0.298828125, "learning_rate": 8.836683417085427e-05, "loss": 1.2109, "step": 2415 }, { "epoch": 0.24716744673776822, "grad_norm": 0.298828125, "learning_rate": 8.836180904522613e-05, "loss": 1.276, "step": 2416 }, { "epoch": 0.24726975114453056, "grad_norm": 0.33203125, "learning_rate": 8.8356783919598e-05, "loss": 1.4437, "step": 2417 }, { "epoch": 0.24737205555129288, "grad_norm": 0.291015625, "learning_rate": 8.835175879396986e-05, "loss": 1.2568, "step": 2418 }, { "epoch": 0.2474743599580552, "grad_norm": 0.35546875, "learning_rate": 8.834673366834171e-05, "loss": 1.5726, "step": 2419 }, { "epoch": 0.2475766643648175, "grad_norm": 0.412109375, "learning_rate": 8.834170854271357e-05, "loss": 1.2507, "step": 2420 }, { "epoch": 0.24767896877157983, "grad_norm": 0.29296875, "learning_rate": 8.833668341708543e-05, "loss": 1.1842, "step": 2421 }, { "epoch": 0.24778127317834217, "grad_norm": 0.30078125, "learning_rate": 8.833165829145729e-05, "loss": 1.2968, "step": 2422 }, { "epoch": 0.2478835775851045, "grad_norm": 0.31640625, "learning_rate": 8.832663316582916e-05, "loss": 1.3644, "step": 2423 }, { "epoch": 0.2479858819918668, "grad_norm": 0.2734375, "learning_rate": 8.8321608040201e-05, "loss": 1.249, "step": 2424 }, { "epoch": 0.24808818639862912, "grad_norm": 0.328125, "learning_rate": 8.831658291457287e-05, "loss": 1.3359, "step": 2425 }, { "epoch": 0.24819049080539143, "grad_norm": 0.2890625, "learning_rate": 8.831155778894473e-05, "loss": 1.3231, "step": 2426 }, { "epoch": 0.24829279521215378, "grad_norm": 0.2890625, "learning_rate": 8.830653266331659e-05, "loss": 1.182, "step": 2427 }, { "epoch": 0.2483950996189161, "grad_norm": 0.275390625, "learning_rate": 8.830150753768844e-05, "loss": 1.2308, "step": 2428 }, { "epoch": 0.2484974040256784, "grad_norm": 0.30859375, "learning_rate": 8.829648241206031e-05, "loss": 1.3883, "step": 2429 }, { "epoch": 0.24859970843244072, "grad_norm": 0.373046875, "learning_rate": 8.829145728643216e-05, "loss": 1.5232, "step": 2430 }, { "epoch": 0.24870201283920304, "grad_norm": 0.322265625, "learning_rate": 8.828643216080402e-05, "loss": 1.4264, "step": 2431 }, { "epoch": 0.24880431724596536, "grad_norm": 0.2890625, "learning_rate": 8.828140703517587e-05, "loss": 1.2733, "step": 2432 }, { "epoch": 0.2489066216527277, "grad_norm": 0.31640625, "learning_rate": 8.827638190954775e-05, "loss": 1.4233, "step": 2433 }, { "epoch": 0.24900892605949002, "grad_norm": 0.33203125, "learning_rate": 8.82713567839196e-05, "loss": 1.2491, "step": 2434 }, { "epoch": 0.24911123046625233, "grad_norm": 0.298828125, "learning_rate": 8.826633165829146e-05, "loss": 1.2952, "step": 2435 }, { "epoch": 0.24921353487301465, "grad_norm": 0.30078125, "learning_rate": 8.826130653266332e-05, "loss": 1.1876, "step": 2436 }, { "epoch": 0.24931583927977696, "grad_norm": 0.4140625, "learning_rate": 8.825628140703518e-05, "loss": 1.1698, "step": 2437 }, { "epoch": 0.2494181436865393, "grad_norm": 0.28125, "learning_rate": 8.825125628140703e-05, "loss": 1.3104, "step": 2438 }, { "epoch": 0.24952044809330162, "grad_norm": 0.30078125, "learning_rate": 8.824623115577891e-05, "loss": 1.313, "step": 2439 }, { "epoch": 0.24962275250006394, "grad_norm": 0.37890625, "learning_rate": 8.824120603015076e-05, "loss": 1.2309, "step": 2440 }, { "epoch": 0.24972505690682625, "grad_norm": 0.306640625, "learning_rate": 8.823618090452262e-05, "loss": 1.3572, "step": 2441 }, { "epoch": 0.24982736131358857, "grad_norm": 0.2890625, "learning_rate": 8.823115577889448e-05, "loss": 1.173, "step": 2442 }, { "epoch": 0.2499296657203509, "grad_norm": 0.3046875, "learning_rate": 8.822613065326634e-05, "loss": 1.2637, "step": 2443 }, { "epoch": 0.2500319701271132, "grad_norm": 0.369140625, "learning_rate": 8.822110552763819e-05, "loss": 1.1946, "step": 2444 }, { "epoch": 0.25013427453387554, "grad_norm": 0.34765625, "learning_rate": 8.821608040201007e-05, "loss": 1.1003, "step": 2445 }, { "epoch": 0.2502365789406379, "grad_norm": 0.341796875, "learning_rate": 8.821105527638191e-05, "loss": 1.3119, "step": 2446 }, { "epoch": 0.2503388833474002, "grad_norm": 0.287109375, "learning_rate": 8.820603015075378e-05, "loss": 1.1347, "step": 2447 }, { "epoch": 0.2504411877541625, "grad_norm": 0.3359375, "learning_rate": 8.820100502512562e-05, "loss": 1.3459, "step": 2448 }, { "epoch": 0.2505434921609248, "grad_norm": 0.353515625, "learning_rate": 8.81959798994975e-05, "loss": 1.1383, "step": 2449 }, { "epoch": 0.25064579656768715, "grad_norm": 0.30078125, "learning_rate": 8.819095477386935e-05, "loss": 1.204, "step": 2450 }, { "epoch": 0.2507481009744495, "grad_norm": 0.275390625, "learning_rate": 8.818592964824121e-05, "loss": 1.271, "step": 2451 }, { "epoch": 0.2508504053812118, "grad_norm": 0.271484375, "learning_rate": 8.818090452261307e-05, "loss": 1.281, "step": 2452 }, { "epoch": 0.2509527097879741, "grad_norm": 0.337890625, "learning_rate": 8.817587939698493e-05, "loss": 1.4707, "step": 2453 }, { "epoch": 0.2510550141947364, "grad_norm": 0.302734375, "learning_rate": 8.817085427135678e-05, "loss": 1.1374, "step": 2454 }, { "epoch": 0.25115731860149876, "grad_norm": 0.302734375, "learning_rate": 8.816582914572865e-05, "loss": 1.2328, "step": 2455 }, { "epoch": 0.2512596230082611, "grad_norm": 0.314453125, "learning_rate": 8.816080402010051e-05, "loss": 1.3623, "step": 2456 }, { "epoch": 0.2513619274150234, "grad_norm": 0.306640625, "learning_rate": 8.815577889447237e-05, "loss": 1.3461, "step": 2457 }, { "epoch": 0.25146423182178573, "grad_norm": 0.3359375, "learning_rate": 8.815075376884423e-05, "loss": 1.1101, "step": 2458 }, { "epoch": 0.251566536228548, "grad_norm": 0.326171875, "learning_rate": 8.814572864321608e-05, "loss": 1.341, "step": 2459 }, { "epoch": 0.25166884063531036, "grad_norm": 0.310546875, "learning_rate": 8.814070351758794e-05, "loss": 1.2733, "step": 2460 }, { "epoch": 0.2517711450420727, "grad_norm": 0.3125, "learning_rate": 8.81356783919598e-05, "loss": 1.3274, "step": 2461 }, { "epoch": 0.251873449448835, "grad_norm": 0.291015625, "learning_rate": 8.813065326633167e-05, "loss": 1.1975, "step": 2462 }, { "epoch": 0.25197575385559734, "grad_norm": 0.3203125, "learning_rate": 8.812562814070351e-05, "loss": 1.3258, "step": 2463 }, { "epoch": 0.25207805826235963, "grad_norm": 0.357421875, "learning_rate": 8.812060301507538e-05, "loss": 1.3287, "step": 2464 }, { "epoch": 0.25218036266912197, "grad_norm": 0.28515625, "learning_rate": 8.811557788944724e-05, "loss": 1.2669, "step": 2465 }, { "epoch": 0.2522826670758843, "grad_norm": 0.314453125, "learning_rate": 8.81105527638191e-05, "loss": 1.4663, "step": 2466 }, { "epoch": 0.2523849714826466, "grad_norm": 0.330078125, "learning_rate": 8.810552763819095e-05, "loss": 1.2789, "step": 2467 }, { "epoch": 0.25248727588940895, "grad_norm": 0.30078125, "learning_rate": 8.810050251256282e-05, "loss": 1.2982, "step": 2468 }, { "epoch": 0.25258958029617123, "grad_norm": 0.318359375, "learning_rate": 8.809547738693467e-05, "loss": 1.3498, "step": 2469 }, { "epoch": 0.2526918847029336, "grad_norm": 0.298828125, "learning_rate": 8.809045226130654e-05, "loss": 1.2849, "step": 2470 }, { "epoch": 0.2527941891096959, "grad_norm": 0.310546875, "learning_rate": 8.80854271356784e-05, "loss": 1.4032, "step": 2471 }, { "epoch": 0.2528964935164582, "grad_norm": 0.337890625, "learning_rate": 8.808040201005026e-05, "loss": 1.3192, "step": 2472 }, { "epoch": 0.25299879792322055, "grad_norm": 0.279296875, "learning_rate": 8.807537688442211e-05, "loss": 1.2856, "step": 2473 }, { "epoch": 0.25310110232998284, "grad_norm": 0.32421875, "learning_rate": 8.807035175879398e-05, "loss": 1.3788, "step": 2474 }, { "epoch": 0.2532034067367452, "grad_norm": 0.30078125, "learning_rate": 8.806532663316583e-05, "loss": 1.237, "step": 2475 }, { "epoch": 0.25330571114350753, "grad_norm": 0.3046875, "learning_rate": 8.80603015075377e-05, "loss": 1.255, "step": 2476 }, { "epoch": 0.2534080155502698, "grad_norm": 0.287109375, "learning_rate": 8.805527638190956e-05, "loss": 1.1906, "step": 2477 }, { "epoch": 0.25351031995703216, "grad_norm": 0.28125, "learning_rate": 8.805025125628142e-05, "loss": 1.346, "step": 2478 }, { "epoch": 0.25361262436379445, "grad_norm": 0.28515625, "learning_rate": 8.804522613065327e-05, "loss": 1.2363, "step": 2479 }, { "epoch": 0.2537149287705568, "grad_norm": 0.30859375, "learning_rate": 8.804020100502513e-05, "loss": 1.2531, "step": 2480 }, { "epoch": 0.25381723317731913, "grad_norm": 0.337890625, "learning_rate": 8.803517587939699e-05, "loss": 1.3067, "step": 2481 }, { "epoch": 0.2539195375840814, "grad_norm": 0.30859375, "learning_rate": 8.803015075376885e-05, "loss": 1.3658, "step": 2482 }, { "epoch": 0.25402184199084377, "grad_norm": 0.291015625, "learning_rate": 8.80251256281407e-05, "loss": 1.2461, "step": 2483 }, { "epoch": 0.25412414639760605, "grad_norm": 0.314453125, "learning_rate": 8.802010050251258e-05, "loss": 1.3223, "step": 2484 }, { "epoch": 0.2542264508043684, "grad_norm": 0.314453125, "learning_rate": 8.801507537688443e-05, "loss": 1.3924, "step": 2485 }, { "epoch": 0.25432875521113074, "grad_norm": 0.34765625, "learning_rate": 8.801005025125629e-05, "loss": 1.4305, "step": 2486 }, { "epoch": 0.25443105961789303, "grad_norm": 0.294921875, "learning_rate": 8.800502512562815e-05, "loss": 1.3333, "step": 2487 }, { "epoch": 0.2545333640246554, "grad_norm": 0.33203125, "learning_rate": 8.800000000000001e-05, "loss": 1.4884, "step": 2488 }, { "epoch": 0.25463566843141766, "grad_norm": 0.5234375, "learning_rate": 8.799497487437186e-05, "loss": 1.2143, "step": 2489 }, { "epoch": 0.25473797283818, "grad_norm": 0.287109375, "learning_rate": 8.798994974874372e-05, "loss": 1.3136, "step": 2490 }, { "epoch": 0.25484027724494235, "grad_norm": 0.294921875, "learning_rate": 8.798492462311558e-05, "loss": 1.2748, "step": 2491 }, { "epoch": 0.25494258165170464, "grad_norm": 0.361328125, "learning_rate": 8.797989949748745e-05, "loss": 1.405, "step": 2492 }, { "epoch": 0.255044886058467, "grad_norm": 0.306640625, "learning_rate": 8.797487437185931e-05, "loss": 1.2446, "step": 2493 }, { "epoch": 0.25514719046522927, "grad_norm": 0.322265625, "learning_rate": 8.796984924623116e-05, "loss": 1.4839, "step": 2494 }, { "epoch": 0.2552494948719916, "grad_norm": 0.27734375, "learning_rate": 8.796482412060302e-05, "loss": 1.2331, "step": 2495 }, { "epoch": 0.25535179927875395, "grad_norm": 0.2734375, "learning_rate": 8.795979899497487e-05, "loss": 1.2983, "step": 2496 }, { "epoch": 0.25545410368551624, "grad_norm": 0.328125, "learning_rate": 8.795477386934674e-05, "loss": 1.3523, "step": 2497 }, { "epoch": 0.2555564080922786, "grad_norm": 0.30078125, "learning_rate": 8.794974874371859e-05, "loss": 1.2694, "step": 2498 }, { "epoch": 0.2556587124990409, "grad_norm": 0.31640625, "learning_rate": 8.794472361809045e-05, "loss": 1.3928, "step": 2499 }, { "epoch": 0.2557610169058032, "grad_norm": 0.296875, "learning_rate": 8.793969849246232e-05, "loss": 1.3093, "step": 2500 }, { "epoch": 0.25586332131256556, "grad_norm": 0.328125, "learning_rate": 8.793467336683418e-05, "loss": 1.3712, "step": 2501 }, { "epoch": 0.25596562571932785, "grad_norm": 0.27734375, "learning_rate": 8.792964824120603e-05, "loss": 1.2625, "step": 2502 }, { "epoch": 0.2560679301260902, "grad_norm": 0.35546875, "learning_rate": 8.79246231155779e-05, "loss": 1.2562, "step": 2503 }, { "epoch": 0.2561702345328525, "grad_norm": 0.3203125, "learning_rate": 8.791959798994975e-05, "loss": 1.4446, "step": 2504 }, { "epoch": 0.2562725389396148, "grad_norm": 0.33203125, "learning_rate": 8.791457286432161e-05, "loss": 1.4237, "step": 2505 }, { "epoch": 0.25637484334637717, "grad_norm": 0.333984375, "learning_rate": 8.790954773869347e-05, "loss": 1.2571, "step": 2506 }, { "epoch": 0.25647714775313946, "grad_norm": 0.291015625, "learning_rate": 8.790452261306534e-05, "loss": 1.2226, "step": 2507 }, { "epoch": 0.2565794521599018, "grad_norm": 0.30859375, "learning_rate": 8.789949748743718e-05, "loss": 1.2939, "step": 2508 }, { "epoch": 0.2566817565666641, "grad_norm": 0.283203125, "learning_rate": 8.789447236180906e-05, "loss": 1.2856, "step": 2509 }, { "epoch": 0.25678406097342643, "grad_norm": 0.30859375, "learning_rate": 8.788944723618091e-05, "loss": 1.3568, "step": 2510 }, { "epoch": 0.2568863653801888, "grad_norm": 0.3359375, "learning_rate": 8.788442211055277e-05, "loss": 1.3614, "step": 2511 }, { "epoch": 0.25698866978695106, "grad_norm": 0.3046875, "learning_rate": 8.787939698492462e-05, "loss": 1.3224, "step": 2512 }, { "epoch": 0.2570909741937134, "grad_norm": 0.296875, "learning_rate": 8.78743718592965e-05, "loss": 1.3664, "step": 2513 }, { "epoch": 0.2571932786004757, "grad_norm": 0.57421875, "learning_rate": 8.786934673366834e-05, "loss": 1.2402, "step": 2514 }, { "epoch": 0.25729558300723804, "grad_norm": 0.36328125, "learning_rate": 8.78643216080402e-05, "loss": 1.2686, "step": 2515 }, { "epoch": 0.2573978874140004, "grad_norm": 0.365234375, "learning_rate": 8.785929648241207e-05, "loss": 1.4722, "step": 2516 }, { "epoch": 0.25750019182076267, "grad_norm": 0.310546875, "learning_rate": 8.785427135678393e-05, "loss": 1.4582, "step": 2517 }, { "epoch": 0.257602496227525, "grad_norm": 0.326171875, "learning_rate": 8.784924623115578e-05, "loss": 1.318, "step": 2518 }, { "epoch": 0.2577048006342873, "grad_norm": 0.302734375, "learning_rate": 8.784422110552765e-05, "loss": 1.2485, "step": 2519 }, { "epoch": 0.25780710504104964, "grad_norm": 0.2890625, "learning_rate": 8.78391959798995e-05, "loss": 1.1807, "step": 2520 }, { "epoch": 0.257909409447812, "grad_norm": 0.306640625, "learning_rate": 8.783417085427136e-05, "loss": 1.2946, "step": 2521 }, { "epoch": 0.2580117138545743, "grad_norm": 0.353515625, "learning_rate": 8.782914572864323e-05, "loss": 1.4654, "step": 2522 }, { "epoch": 0.2581140182613366, "grad_norm": 0.35546875, "learning_rate": 8.782412060301509e-05, "loss": 1.3093, "step": 2523 }, { "epoch": 0.2582163226680989, "grad_norm": 0.298828125, "learning_rate": 8.781909547738694e-05, "loss": 1.383, "step": 2524 }, { "epoch": 0.25831862707486125, "grad_norm": 0.306640625, "learning_rate": 8.78140703517588e-05, "loss": 1.4855, "step": 2525 }, { "epoch": 0.2584209314816236, "grad_norm": 0.32421875, "learning_rate": 8.780904522613066e-05, "loss": 1.2943, "step": 2526 }, { "epoch": 0.2585232358883859, "grad_norm": 0.291015625, "learning_rate": 8.780402010050251e-05, "loss": 1.2308, "step": 2527 }, { "epoch": 0.2586255402951482, "grad_norm": 0.33203125, "learning_rate": 8.779899497487437e-05, "loss": 1.4374, "step": 2528 }, { "epoch": 0.2587278447019105, "grad_norm": 0.3203125, "learning_rate": 8.779396984924623e-05, "loss": 1.256, "step": 2529 }, { "epoch": 0.25883014910867286, "grad_norm": 0.3046875, "learning_rate": 8.77889447236181e-05, "loss": 1.304, "step": 2530 }, { "epoch": 0.2589324535154352, "grad_norm": 0.3046875, "learning_rate": 8.778391959798994e-05, "loss": 1.302, "step": 2531 }, { "epoch": 0.2590347579221975, "grad_norm": 0.32421875, "learning_rate": 8.777889447236182e-05, "loss": 1.4424, "step": 2532 }, { "epoch": 0.25913706232895983, "grad_norm": 0.29296875, "learning_rate": 8.777386934673367e-05, "loss": 1.2931, "step": 2533 }, { "epoch": 0.2592393667357221, "grad_norm": 0.31640625, "learning_rate": 8.776884422110553e-05, "loss": 1.1826, "step": 2534 }, { "epoch": 0.25934167114248446, "grad_norm": 0.3203125, "learning_rate": 8.776381909547739e-05, "loss": 1.3807, "step": 2535 }, { "epoch": 0.2594439755492468, "grad_norm": 0.298828125, "learning_rate": 8.775879396984925e-05, "loss": 1.1657, "step": 2536 }, { "epoch": 0.2595462799560091, "grad_norm": 0.28125, "learning_rate": 8.77537688442211e-05, "loss": 1.2216, "step": 2537 }, { "epoch": 0.25964858436277144, "grad_norm": 0.29296875, "learning_rate": 8.774874371859298e-05, "loss": 1.421, "step": 2538 }, { "epoch": 0.2597508887695337, "grad_norm": 0.314453125, "learning_rate": 8.774371859296483e-05, "loss": 1.3412, "step": 2539 }, { "epoch": 0.25985319317629607, "grad_norm": 0.326171875, "learning_rate": 8.773869346733669e-05, "loss": 1.2944, "step": 2540 }, { "epoch": 0.2599554975830584, "grad_norm": 0.326171875, "learning_rate": 8.773366834170854e-05, "loss": 1.2215, "step": 2541 }, { "epoch": 0.2600578019898207, "grad_norm": 0.310546875, "learning_rate": 8.772864321608041e-05, "loss": 1.1204, "step": 2542 }, { "epoch": 0.26016010639658305, "grad_norm": 0.28515625, "learning_rate": 8.772361809045226e-05, "loss": 1.2282, "step": 2543 }, { "epoch": 0.26026241080334533, "grad_norm": 0.333984375, "learning_rate": 8.771859296482412e-05, "loss": 1.2974, "step": 2544 }, { "epoch": 0.2603647152101077, "grad_norm": 0.294921875, "learning_rate": 8.771356783919598e-05, "loss": 1.3554, "step": 2545 }, { "epoch": 0.26046701961687, "grad_norm": 0.287109375, "learning_rate": 8.770854271356785e-05, "loss": 1.329, "step": 2546 }, { "epoch": 0.2605693240236323, "grad_norm": 0.326171875, "learning_rate": 8.77035175879397e-05, "loss": 1.28, "step": 2547 }, { "epoch": 0.26067162843039465, "grad_norm": 0.28125, "learning_rate": 8.769849246231157e-05, "loss": 1.1819, "step": 2548 }, { "epoch": 0.26077393283715694, "grad_norm": 0.361328125, "learning_rate": 8.769346733668342e-05, "loss": 1.422, "step": 2549 }, { "epoch": 0.2608762372439193, "grad_norm": 0.27734375, "learning_rate": 8.768844221105528e-05, "loss": 1.2408, "step": 2550 }, { "epoch": 0.26097854165068163, "grad_norm": 0.2890625, "learning_rate": 8.768341708542714e-05, "loss": 1.1474, "step": 2551 }, { "epoch": 0.2610808460574439, "grad_norm": 0.322265625, "learning_rate": 8.7678391959799e-05, "loss": 1.3153, "step": 2552 }, { "epoch": 0.26118315046420626, "grad_norm": 0.30859375, "learning_rate": 8.767336683417085e-05, "loss": 1.3141, "step": 2553 }, { "epoch": 0.26128545487096855, "grad_norm": 0.3203125, "learning_rate": 8.766834170854273e-05, "loss": 1.2633, "step": 2554 }, { "epoch": 0.2613877592777309, "grad_norm": 0.283203125, "learning_rate": 8.766331658291458e-05, "loss": 1.3578, "step": 2555 }, { "epoch": 0.26149006368449323, "grad_norm": 0.310546875, "learning_rate": 8.765829145728644e-05, "loss": 1.3082, "step": 2556 }, { "epoch": 0.2615923680912555, "grad_norm": 0.296875, "learning_rate": 8.765326633165829e-05, "loss": 1.2846, "step": 2557 }, { "epoch": 0.26169467249801787, "grad_norm": 0.302734375, "learning_rate": 8.764824120603016e-05, "loss": 1.3471, "step": 2558 }, { "epoch": 0.26179697690478015, "grad_norm": 0.3046875, "learning_rate": 8.764321608040201e-05, "loss": 1.308, "step": 2559 }, { "epoch": 0.2618992813115425, "grad_norm": 0.28515625, "learning_rate": 8.763819095477387e-05, "loss": 1.1872, "step": 2560 }, { "epoch": 0.26200158571830484, "grad_norm": 0.2890625, "learning_rate": 8.763316582914574e-05, "loss": 1.4082, "step": 2561 }, { "epoch": 0.26210389012506713, "grad_norm": 0.302734375, "learning_rate": 8.762814070351758e-05, "loss": 1.3155, "step": 2562 }, { "epoch": 0.2622061945318295, "grad_norm": 0.28515625, "learning_rate": 8.762311557788945e-05, "loss": 1.2965, "step": 2563 }, { "epoch": 0.26230849893859176, "grad_norm": 0.3046875, "learning_rate": 8.761809045226131e-05, "loss": 1.245, "step": 2564 }, { "epoch": 0.2624108033453541, "grad_norm": 0.302734375, "learning_rate": 8.761306532663317e-05, "loss": 1.3927, "step": 2565 }, { "epoch": 0.26251310775211645, "grad_norm": 0.28515625, "learning_rate": 8.760804020100502e-05, "loss": 1.1191, "step": 2566 }, { "epoch": 0.26261541215887874, "grad_norm": 0.388671875, "learning_rate": 8.76030150753769e-05, "loss": 1.3636, "step": 2567 }, { "epoch": 0.2627177165656411, "grad_norm": 0.35546875, "learning_rate": 8.759798994974874e-05, "loss": 1.3622, "step": 2568 }, { "epoch": 0.26282002097240337, "grad_norm": 0.326171875, "learning_rate": 8.75929648241206e-05, "loss": 1.4425, "step": 2569 }, { "epoch": 0.2629223253791657, "grad_norm": 0.341796875, "learning_rate": 8.758793969849247e-05, "loss": 1.3842, "step": 2570 }, { "epoch": 0.26302462978592805, "grad_norm": 0.296875, "learning_rate": 8.758291457286433e-05, "loss": 1.2743, "step": 2571 }, { "epoch": 0.26312693419269034, "grad_norm": 0.296875, "learning_rate": 8.757788944723618e-05, "loss": 1.2812, "step": 2572 }, { "epoch": 0.2632292385994527, "grad_norm": 0.296875, "learning_rate": 8.757286432160804e-05, "loss": 1.2752, "step": 2573 }, { "epoch": 0.263331543006215, "grad_norm": 0.349609375, "learning_rate": 8.75678391959799e-05, "loss": 1.2779, "step": 2574 }, { "epoch": 0.2634338474129773, "grad_norm": 0.34765625, "learning_rate": 8.756281407035176e-05, "loss": 1.3704, "step": 2575 }, { "epoch": 0.26353615181973966, "grad_norm": 0.263671875, "learning_rate": 8.755778894472361e-05, "loss": 1.1308, "step": 2576 }, { "epoch": 0.26363845622650195, "grad_norm": 0.27734375, "learning_rate": 8.755276381909549e-05, "loss": 1.1902, "step": 2577 }, { "epoch": 0.2637407606332643, "grad_norm": 0.29296875, "learning_rate": 8.754773869346734e-05, "loss": 1.1601, "step": 2578 }, { "epoch": 0.2638430650400266, "grad_norm": 0.2734375, "learning_rate": 8.75427135678392e-05, "loss": 1.1896, "step": 2579 }, { "epoch": 0.2639453694467889, "grad_norm": 0.287109375, "learning_rate": 8.753768844221106e-05, "loss": 1.4207, "step": 2580 }, { "epoch": 0.26404767385355127, "grad_norm": 0.30078125, "learning_rate": 8.753266331658292e-05, "loss": 1.2276, "step": 2581 }, { "epoch": 0.26414997826031356, "grad_norm": 0.2890625, "learning_rate": 8.752763819095477e-05, "loss": 1.3212, "step": 2582 }, { "epoch": 0.2642522826670759, "grad_norm": 0.3046875, "learning_rate": 8.752261306532665e-05, "loss": 1.2532, "step": 2583 }, { "epoch": 0.2643545870738382, "grad_norm": 0.359375, "learning_rate": 8.75175879396985e-05, "loss": 1.4928, "step": 2584 }, { "epoch": 0.26445689148060053, "grad_norm": 0.30078125, "learning_rate": 8.751256281407036e-05, "loss": 1.3882, "step": 2585 }, { "epoch": 0.2645591958873629, "grad_norm": 0.296875, "learning_rate": 8.750753768844222e-05, "loss": 1.3669, "step": 2586 }, { "epoch": 0.26466150029412516, "grad_norm": 0.30859375, "learning_rate": 8.750251256281408e-05, "loss": 1.1608, "step": 2587 }, { "epoch": 0.2647638047008875, "grad_norm": 0.2890625, "learning_rate": 8.749748743718593e-05, "loss": 1.1918, "step": 2588 }, { "epoch": 0.2648661091076498, "grad_norm": 0.32421875, "learning_rate": 8.749246231155779e-05, "loss": 1.3735, "step": 2589 }, { "epoch": 0.26496841351441214, "grad_norm": 0.333984375, "learning_rate": 8.748743718592965e-05, "loss": 1.4091, "step": 2590 }, { "epoch": 0.2650707179211745, "grad_norm": 0.326171875, "learning_rate": 8.748241206030152e-05, "loss": 1.2506, "step": 2591 }, { "epoch": 0.26517302232793677, "grad_norm": 0.333984375, "learning_rate": 8.747738693467336e-05, "loss": 1.2955, "step": 2592 }, { "epoch": 0.2652753267346991, "grad_norm": 0.3125, "learning_rate": 8.747236180904523e-05, "loss": 1.3568, "step": 2593 }, { "epoch": 0.2653776311414614, "grad_norm": 0.3203125, "learning_rate": 8.746733668341709e-05, "loss": 1.3322, "step": 2594 }, { "epoch": 0.26547993554822374, "grad_norm": 0.3046875, "learning_rate": 8.746231155778895e-05, "loss": 1.2753, "step": 2595 }, { "epoch": 0.2655822399549861, "grad_norm": 0.32421875, "learning_rate": 8.745728643216081e-05, "loss": 1.227, "step": 2596 }, { "epoch": 0.2656845443617484, "grad_norm": 0.341796875, "learning_rate": 8.745226130653266e-05, "loss": 1.2689, "step": 2597 }, { "epoch": 0.2657868487685107, "grad_norm": 0.32421875, "learning_rate": 8.744723618090452e-05, "loss": 1.3139, "step": 2598 }, { "epoch": 0.265889153175273, "grad_norm": 0.294921875, "learning_rate": 8.744221105527639e-05, "loss": 1.2866, "step": 2599 }, { "epoch": 0.26599145758203535, "grad_norm": 0.330078125, "learning_rate": 8.743718592964825e-05, "loss": 1.2891, "step": 2600 }, { "epoch": 0.2660937619887977, "grad_norm": 0.341796875, "learning_rate": 8.74321608040201e-05, "loss": 1.3092, "step": 2601 }, { "epoch": 0.26619606639556, "grad_norm": 0.376953125, "learning_rate": 8.742713567839197e-05, "loss": 1.3697, "step": 2602 }, { "epoch": 0.2662983708023223, "grad_norm": 0.30078125, "learning_rate": 8.742211055276382e-05, "loss": 1.3384, "step": 2603 }, { "epoch": 0.2664006752090846, "grad_norm": 0.310546875, "learning_rate": 8.741708542713568e-05, "loss": 1.1937, "step": 2604 }, { "epoch": 0.26650297961584696, "grad_norm": 0.3125, "learning_rate": 8.741206030150753e-05, "loss": 1.2435, "step": 2605 }, { "epoch": 0.2666052840226093, "grad_norm": 0.333984375, "learning_rate": 8.74070351758794e-05, "loss": 1.1976, "step": 2606 }, { "epoch": 0.2667075884293716, "grad_norm": 0.310546875, "learning_rate": 8.740201005025125e-05, "loss": 1.3179, "step": 2607 }, { "epoch": 0.26680989283613393, "grad_norm": 0.322265625, "learning_rate": 8.739698492462312e-05, "loss": 1.2266, "step": 2608 }, { "epoch": 0.2669121972428962, "grad_norm": 0.310546875, "learning_rate": 8.739195979899498e-05, "loss": 1.3869, "step": 2609 }, { "epoch": 0.26701450164965856, "grad_norm": 0.318359375, "learning_rate": 8.738693467336684e-05, "loss": 1.2248, "step": 2610 }, { "epoch": 0.2671168060564209, "grad_norm": 0.33984375, "learning_rate": 8.738190954773869e-05, "loss": 1.3424, "step": 2611 }, { "epoch": 0.2672191104631832, "grad_norm": 0.306640625, "learning_rate": 8.737688442211056e-05, "loss": 1.3217, "step": 2612 }, { "epoch": 0.26732141486994554, "grad_norm": 0.287109375, "learning_rate": 8.737185929648241e-05, "loss": 1.2382, "step": 2613 }, { "epoch": 0.2674237192767078, "grad_norm": 0.34765625, "learning_rate": 8.736683417085428e-05, "loss": 1.3698, "step": 2614 }, { "epoch": 0.26752602368347017, "grad_norm": 0.27734375, "learning_rate": 8.736180904522614e-05, "loss": 1.1441, "step": 2615 }, { "epoch": 0.26762832809023246, "grad_norm": 0.333984375, "learning_rate": 8.7356783919598e-05, "loss": 1.2693, "step": 2616 }, { "epoch": 0.2677306324969948, "grad_norm": 0.328125, "learning_rate": 8.735175879396985e-05, "loss": 1.4, "step": 2617 }, { "epoch": 0.26783293690375715, "grad_norm": 0.30859375, "learning_rate": 8.734673366834172e-05, "loss": 1.2597, "step": 2618 }, { "epoch": 0.26793524131051943, "grad_norm": 0.294921875, "learning_rate": 8.734170854271357e-05, "loss": 1.1935, "step": 2619 }, { "epoch": 0.2680375457172818, "grad_norm": 0.3203125, "learning_rate": 8.733668341708543e-05, "loss": 1.3349, "step": 2620 }, { "epoch": 0.26813985012404407, "grad_norm": 0.318359375, "learning_rate": 8.733165829145728e-05, "loss": 1.2376, "step": 2621 }, { "epoch": 0.2682421545308064, "grad_norm": 0.318359375, "learning_rate": 8.732663316582916e-05, "loss": 1.2902, "step": 2622 }, { "epoch": 0.26834445893756875, "grad_norm": 0.296875, "learning_rate": 8.7321608040201e-05, "loss": 1.2709, "step": 2623 }, { "epoch": 0.26844676334433104, "grad_norm": 0.337890625, "learning_rate": 8.731658291457287e-05, "loss": 1.3944, "step": 2624 }, { "epoch": 0.2685490677510934, "grad_norm": 0.38671875, "learning_rate": 8.731155778894473e-05, "loss": 1.3606, "step": 2625 }, { "epoch": 0.2686513721578557, "grad_norm": 0.283203125, "learning_rate": 8.730653266331659e-05, "loss": 1.2543, "step": 2626 }, { "epoch": 0.268753676564618, "grad_norm": 0.31640625, "learning_rate": 8.730150753768844e-05, "loss": 1.2533, "step": 2627 }, { "epoch": 0.26885598097138036, "grad_norm": 0.302734375, "learning_rate": 8.72964824120603e-05, "loss": 1.2278, "step": 2628 }, { "epoch": 0.26895828537814265, "grad_norm": 0.32421875, "learning_rate": 8.729145728643217e-05, "loss": 1.3452, "step": 2629 }, { "epoch": 0.269060589784905, "grad_norm": 0.291015625, "learning_rate": 8.728643216080403e-05, "loss": 1.2531, "step": 2630 }, { "epoch": 0.2691628941916673, "grad_norm": 0.31640625, "learning_rate": 8.728140703517589e-05, "loss": 1.3713, "step": 2631 }, { "epoch": 0.2692651985984296, "grad_norm": 0.3046875, "learning_rate": 8.727638190954774e-05, "loss": 1.1919, "step": 2632 }, { "epoch": 0.26936750300519197, "grad_norm": 0.287109375, "learning_rate": 8.72713567839196e-05, "loss": 1.2771, "step": 2633 }, { "epoch": 0.26946980741195425, "grad_norm": 0.333984375, "learning_rate": 8.726633165829146e-05, "loss": 1.2883, "step": 2634 }, { "epoch": 0.2695721118187166, "grad_norm": 0.287109375, "learning_rate": 8.726130653266332e-05, "loss": 1.1977, "step": 2635 }, { "epoch": 0.2696744162254789, "grad_norm": 0.314453125, "learning_rate": 8.725628140703517e-05, "loss": 1.2361, "step": 2636 }, { "epoch": 0.26977672063224123, "grad_norm": 0.27734375, "learning_rate": 8.725125628140703e-05, "loss": 1.2424, "step": 2637 }, { "epoch": 0.2698790250390036, "grad_norm": 0.302734375, "learning_rate": 8.72462311557789e-05, "loss": 1.3551, "step": 2638 }, { "epoch": 0.26998132944576586, "grad_norm": 0.310546875, "learning_rate": 8.724120603015076e-05, "loss": 1.1335, "step": 2639 }, { "epoch": 0.2700836338525282, "grad_norm": 0.33984375, "learning_rate": 8.723618090452261e-05, "loss": 1.4061, "step": 2640 }, { "epoch": 0.2701859382592905, "grad_norm": 0.2890625, "learning_rate": 8.723115577889448e-05, "loss": 1.1638, "step": 2641 }, { "epoch": 0.27028824266605284, "grad_norm": 0.3046875, "learning_rate": 8.722613065326633e-05, "loss": 1.2136, "step": 2642 }, { "epoch": 0.2703905470728152, "grad_norm": 0.357421875, "learning_rate": 8.722110552763819e-05, "loss": 1.3762, "step": 2643 }, { "epoch": 0.27049285147957747, "grad_norm": 0.30859375, "learning_rate": 8.721608040201005e-05, "loss": 1.2217, "step": 2644 }, { "epoch": 0.2705951558863398, "grad_norm": 0.302734375, "learning_rate": 8.721105527638192e-05, "loss": 1.2642, "step": 2645 }, { "epoch": 0.2706974602931021, "grad_norm": 0.365234375, "learning_rate": 8.720603015075377e-05, "loss": 1.3418, "step": 2646 }, { "epoch": 0.27079976469986444, "grad_norm": 0.310546875, "learning_rate": 8.720100502512564e-05, "loss": 1.3, "step": 2647 }, { "epoch": 0.2709020691066268, "grad_norm": 0.306640625, "learning_rate": 8.719597989949749e-05, "loss": 1.3531, "step": 2648 }, { "epoch": 0.2710043735133891, "grad_norm": 0.283203125, "learning_rate": 8.719095477386935e-05, "loss": 1.2158, "step": 2649 }, { "epoch": 0.2711066779201514, "grad_norm": 0.3203125, "learning_rate": 8.71859296482412e-05, "loss": 1.2892, "step": 2650 }, { "epoch": 0.2712089823269137, "grad_norm": 0.326171875, "learning_rate": 8.718090452261308e-05, "loss": 1.3833, "step": 2651 }, { "epoch": 0.27131128673367605, "grad_norm": 0.34375, "learning_rate": 8.717587939698492e-05, "loss": 1.225, "step": 2652 }, { "epoch": 0.2714135911404384, "grad_norm": 0.267578125, "learning_rate": 8.717085427135679e-05, "loss": 1.17, "step": 2653 }, { "epoch": 0.2715158955472007, "grad_norm": 0.353515625, "learning_rate": 8.716582914572865e-05, "loss": 1.2544, "step": 2654 }, { "epoch": 0.271618199953963, "grad_norm": 0.55078125, "learning_rate": 8.716080402010051e-05, "loss": 1.2882, "step": 2655 }, { "epoch": 0.2717205043607253, "grad_norm": 0.314453125, "learning_rate": 8.715577889447236e-05, "loss": 2.1961, "step": 2656 }, { "epoch": 0.27182280876748766, "grad_norm": 0.322265625, "learning_rate": 8.715075376884423e-05, "loss": 1.3558, "step": 2657 }, { "epoch": 0.27192511317425, "grad_norm": 0.3125, "learning_rate": 8.714572864321608e-05, "loss": 1.2967, "step": 2658 }, { "epoch": 0.2720274175810123, "grad_norm": 0.310546875, "learning_rate": 8.714070351758794e-05, "loss": 1.2884, "step": 2659 }, { "epoch": 0.27212972198777463, "grad_norm": 0.302734375, "learning_rate": 8.713567839195981e-05, "loss": 1.3628, "step": 2660 }, { "epoch": 0.2722320263945369, "grad_norm": 0.291015625, "learning_rate": 8.713065326633167e-05, "loss": 1.2816, "step": 2661 }, { "epoch": 0.27233433080129926, "grad_norm": 0.328125, "learning_rate": 8.712562814070352e-05, "loss": 1.3329, "step": 2662 }, { "epoch": 0.2724366352080616, "grad_norm": 0.302734375, "learning_rate": 8.712060301507538e-05, "loss": 1.4082, "step": 2663 }, { "epoch": 0.2725389396148239, "grad_norm": 0.314453125, "learning_rate": 8.711557788944724e-05, "loss": 1.2249, "step": 2664 }, { "epoch": 0.27264124402158624, "grad_norm": 0.287109375, "learning_rate": 8.71105527638191e-05, "loss": 1.2578, "step": 2665 }, { "epoch": 0.2727435484283485, "grad_norm": 0.349609375, "learning_rate": 8.710552763819095e-05, "loss": 1.4183, "step": 2666 }, { "epoch": 0.27284585283511087, "grad_norm": 0.294921875, "learning_rate": 8.710050251256281e-05, "loss": 1.3319, "step": 2667 }, { "epoch": 0.2729481572418732, "grad_norm": 0.326171875, "learning_rate": 8.709547738693468e-05, "loss": 1.3479, "step": 2668 }, { "epoch": 0.2730504616486355, "grad_norm": 0.283203125, "learning_rate": 8.709045226130654e-05, "loss": 1.2902, "step": 2669 }, { "epoch": 0.27315276605539784, "grad_norm": 0.275390625, "learning_rate": 8.70854271356784e-05, "loss": 1.2325, "step": 2670 }, { "epoch": 0.27325507046216013, "grad_norm": 0.3203125, "learning_rate": 8.708040201005025e-05, "loss": 1.3145, "step": 2671 }, { "epoch": 0.2733573748689225, "grad_norm": 0.416015625, "learning_rate": 8.707537688442211e-05, "loss": 1.5189, "step": 2672 }, { "epoch": 0.2734596792756848, "grad_norm": 0.3046875, "learning_rate": 8.707035175879397e-05, "loss": 1.1732, "step": 2673 }, { "epoch": 0.2735619836824471, "grad_norm": 0.283203125, "learning_rate": 8.706532663316583e-05, "loss": 1.2507, "step": 2674 }, { "epoch": 0.27366428808920945, "grad_norm": 0.298828125, "learning_rate": 8.706030150753768e-05, "loss": 1.2585, "step": 2675 }, { "epoch": 0.27376659249597174, "grad_norm": 0.3046875, "learning_rate": 8.705527638190956e-05, "loss": 1.3931, "step": 2676 }, { "epoch": 0.2738688969027341, "grad_norm": 0.296875, "learning_rate": 8.705025125628141e-05, "loss": 1.1675, "step": 2677 }, { "epoch": 0.2739712013094964, "grad_norm": 0.291015625, "learning_rate": 8.704522613065327e-05, "loss": 1.2555, "step": 2678 }, { "epoch": 0.2740735057162587, "grad_norm": 0.294921875, "learning_rate": 8.704020100502513e-05, "loss": 1.3707, "step": 2679 }, { "epoch": 0.27417581012302106, "grad_norm": 0.30078125, "learning_rate": 8.7035175879397e-05, "loss": 1.3261, "step": 2680 }, { "epoch": 0.27427811452978335, "grad_norm": 0.27734375, "learning_rate": 8.703015075376884e-05, "loss": 1.2162, "step": 2681 }, { "epoch": 0.2743804189365457, "grad_norm": 0.322265625, "learning_rate": 8.70251256281407e-05, "loss": 1.2336, "step": 2682 }, { "epoch": 0.27448272334330803, "grad_norm": 0.30859375, "learning_rate": 8.702010050251257e-05, "loss": 1.4929, "step": 2683 }, { "epoch": 0.2745850277500703, "grad_norm": 0.34765625, "learning_rate": 8.701507537688443e-05, "loss": 1.4326, "step": 2684 }, { "epoch": 0.27468733215683266, "grad_norm": 0.267578125, "learning_rate": 8.701005025125628e-05, "loss": 1.316, "step": 2685 }, { "epoch": 0.27478963656359495, "grad_norm": 0.275390625, "learning_rate": 8.700502512562815e-05, "loss": 1.229, "step": 2686 }, { "epoch": 0.2748919409703573, "grad_norm": 0.3125, "learning_rate": 8.7e-05, "loss": 1.3003, "step": 2687 }, { "epoch": 0.27499424537711964, "grad_norm": 0.3046875, "learning_rate": 8.699497487437186e-05, "loss": 1.5353, "step": 2688 }, { "epoch": 0.2750965497838819, "grad_norm": 0.333984375, "learning_rate": 8.698994974874372e-05, "loss": 1.3388, "step": 2689 }, { "epoch": 0.27519885419064427, "grad_norm": 0.365234375, "learning_rate": 8.698492462311559e-05, "loss": 1.2782, "step": 2690 }, { "epoch": 0.27530115859740656, "grad_norm": 0.326171875, "learning_rate": 8.697989949748744e-05, "loss": 1.2682, "step": 2691 }, { "epoch": 0.2754034630041689, "grad_norm": 0.337890625, "learning_rate": 8.697487437185931e-05, "loss": 1.4232, "step": 2692 }, { "epoch": 0.27550576741093125, "grad_norm": 0.345703125, "learning_rate": 8.696984924623116e-05, "loss": 1.3877, "step": 2693 }, { "epoch": 0.27560807181769353, "grad_norm": 0.33203125, "learning_rate": 8.696482412060302e-05, "loss": 1.1227, "step": 2694 }, { "epoch": 0.2757103762244559, "grad_norm": 0.3125, "learning_rate": 8.695979899497488e-05, "loss": 1.1016, "step": 2695 }, { "epoch": 0.27581268063121817, "grad_norm": 0.318359375, "learning_rate": 8.695477386934675e-05, "loss": 1.3558, "step": 2696 }, { "epoch": 0.2759149850379805, "grad_norm": 0.2890625, "learning_rate": 8.69497487437186e-05, "loss": 1.2154, "step": 2697 }, { "epoch": 0.27601728944474285, "grad_norm": 0.28515625, "learning_rate": 8.694472361809046e-05, "loss": 1.2455, "step": 2698 }, { "epoch": 0.27611959385150514, "grad_norm": 0.28125, "learning_rate": 8.693969849246232e-05, "loss": 1.3463, "step": 2699 }, { "epoch": 0.2762218982582675, "grad_norm": 0.3046875, "learning_rate": 8.693467336683418e-05, "loss": 1.2462, "step": 2700 }, { "epoch": 0.27632420266502977, "grad_norm": 0.32421875, "learning_rate": 8.692964824120603e-05, "loss": 1.2499, "step": 2701 }, { "epoch": 0.2764265070717921, "grad_norm": 0.326171875, "learning_rate": 8.692462311557789e-05, "loss": 1.1878, "step": 2702 }, { "epoch": 0.27652881147855446, "grad_norm": 0.279296875, "learning_rate": 8.691959798994975e-05, "loss": 1.1283, "step": 2703 }, { "epoch": 0.27663111588531675, "grad_norm": 0.3203125, "learning_rate": 8.691457286432161e-05, "loss": 1.2597, "step": 2704 }, { "epoch": 0.2767334202920791, "grad_norm": 0.29296875, "learning_rate": 8.690954773869348e-05, "loss": 1.2644, "step": 2705 }, { "epoch": 0.2768357246988414, "grad_norm": 0.310546875, "learning_rate": 8.690452261306532e-05, "loss": 1.2131, "step": 2706 }, { "epoch": 0.2769380291056037, "grad_norm": 0.3046875, "learning_rate": 8.689949748743719e-05, "loss": 1.325, "step": 2707 }, { "epoch": 0.27704033351236607, "grad_norm": 0.28515625, "learning_rate": 8.689447236180905e-05, "loss": 1.3073, "step": 2708 }, { "epoch": 0.27714263791912835, "grad_norm": 0.2734375, "learning_rate": 8.688944723618091e-05, "loss": 1.269, "step": 2709 }, { "epoch": 0.2772449423258907, "grad_norm": 0.3671875, "learning_rate": 8.688442211055276e-05, "loss": 1.3373, "step": 2710 }, { "epoch": 0.277347246732653, "grad_norm": 0.33984375, "learning_rate": 8.687939698492464e-05, "loss": 1.3192, "step": 2711 }, { "epoch": 0.27744955113941533, "grad_norm": 0.330078125, "learning_rate": 8.687437185929648e-05, "loss": 1.3356, "step": 2712 }, { "epoch": 0.2775518555461777, "grad_norm": 0.318359375, "learning_rate": 8.686934673366835e-05, "loss": 1.2267, "step": 2713 }, { "epoch": 0.27765415995293996, "grad_norm": 0.30078125, "learning_rate": 8.68643216080402e-05, "loss": 1.2895, "step": 2714 }, { "epoch": 0.2777564643597023, "grad_norm": 0.353515625, "learning_rate": 8.685929648241207e-05, "loss": 1.3493, "step": 2715 }, { "epoch": 0.2778587687664646, "grad_norm": 0.36328125, "learning_rate": 8.685427135678392e-05, "loss": 1.3142, "step": 2716 }, { "epoch": 0.27796107317322694, "grad_norm": 0.373046875, "learning_rate": 8.684924623115578e-05, "loss": 1.3491, "step": 2717 }, { "epoch": 0.2780633775799893, "grad_norm": 0.291015625, "learning_rate": 8.684422110552764e-05, "loss": 1.2482, "step": 2718 }, { "epoch": 0.27816568198675157, "grad_norm": 0.3515625, "learning_rate": 8.68391959798995e-05, "loss": 1.3565, "step": 2719 }, { "epoch": 0.2782679863935139, "grad_norm": 0.271484375, "learning_rate": 8.683417085427135e-05, "loss": 1.2082, "step": 2720 }, { "epoch": 0.2783702908002762, "grad_norm": 0.314453125, "learning_rate": 8.682914572864323e-05, "loss": 1.3252, "step": 2721 }, { "epoch": 0.27847259520703854, "grad_norm": 0.26171875, "learning_rate": 8.682412060301508e-05, "loss": 1.2134, "step": 2722 }, { "epoch": 0.2785748996138009, "grad_norm": 0.310546875, "learning_rate": 8.681909547738694e-05, "loss": 1.3013, "step": 2723 }, { "epoch": 0.2786772040205632, "grad_norm": 0.31640625, "learning_rate": 8.68140703517588e-05, "loss": 1.2969, "step": 2724 }, { "epoch": 0.2787795084273255, "grad_norm": 0.32421875, "learning_rate": 8.680904522613066e-05, "loss": 1.3255, "step": 2725 }, { "epoch": 0.2788818128340878, "grad_norm": 0.296875, "learning_rate": 8.680402010050251e-05, "loss": 1.2339, "step": 2726 }, { "epoch": 0.27898411724085015, "grad_norm": 0.345703125, "learning_rate": 8.679899497487439e-05, "loss": 1.3045, "step": 2727 }, { "epoch": 0.2790864216476125, "grad_norm": 0.32421875, "learning_rate": 8.679396984924624e-05, "loss": 1.3166, "step": 2728 }, { "epoch": 0.2791887260543748, "grad_norm": 0.294921875, "learning_rate": 8.67889447236181e-05, "loss": 1.19, "step": 2729 }, { "epoch": 0.2792910304611371, "grad_norm": 0.328125, "learning_rate": 8.678391959798995e-05, "loss": 1.1467, "step": 2730 }, { "epoch": 0.2793933348678994, "grad_norm": 0.302734375, "learning_rate": 8.677889447236182e-05, "loss": 1.3073, "step": 2731 }, { "epoch": 0.27949563927466176, "grad_norm": 0.287109375, "learning_rate": 8.677386934673367e-05, "loss": 1.3157, "step": 2732 }, { "epoch": 0.2795979436814241, "grad_norm": 0.306640625, "learning_rate": 8.676884422110553e-05, "loss": 1.2355, "step": 2733 }, { "epoch": 0.2797002480881864, "grad_norm": 0.31640625, "learning_rate": 8.67638190954774e-05, "loss": 1.3147, "step": 2734 }, { "epoch": 0.27980255249494873, "grad_norm": 0.28515625, "learning_rate": 8.675879396984926e-05, "loss": 1.2172, "step": 2735 }, { "epoch": 0.279904856901711, "grad_norm": 0.287109375, "learning_rate": 8.67537688442211e-05, "loss": 1.2741, "step": 2736 }, { "epoch": 0.28000716130847336, "grad_norm": 0.33203125, "learning_rate": 8.674874371859297e-05, "loss": 1.3384, "step": 2737 }, { "epoch": 0.2801094657152357, "grad_norm": 0.28125, "learning_rate": 8.674371859296483e-05, "loss": 1.338, "step": 2738 }, { "epoch": 0.280211770121998, "grad_norm": 0.3046875, "learning_rate": 8.673869346733669e-05, "loss": 1.4115, "step": 2739 }, { "epoch": 0.28031407452876034, "grad_norm": 0.314453125, "learning_rate": 8.673366834170855e-05, "loss": 1.2106, "step": 2740 }, { "epoch": 0.2804163789355226, "grad_norm": 0.31640625, "learning_rate": 8.67286432160804e-05, "loss": 1.3035, "step": 2741 }, { "epoch": 0.28051868334228497, "grad_norm": 0.283203125, "learning_rate": 8.672361809045226e-05, "loss": 1.2221, "step": 2742 }, { "epoch": 0.2806209877490473, "grad_norm": 0.447265625, "learning_rate": 8.671859296482413e-05, "loss": 1.4653, "step": 2743 }, { "epoch": 0.2807232921558096, "grad_norm": 0.33984375, "learning_rate": 8.671356783919599e-05, "loss": 1.2618, "step": 2744 }, { "epoch": 0.28082559656257194, "grad_norm": 0.29296875, "learning_rate": 8.670854271356784e-05, "loss": 1.2783, "step": 2745 }, { "epoch": 0.28092790096933423, "grad_norm": 0.306640625, "learning_rate": 8.67035175879397e-05, "loss": 1.3686, "step": 2746 }, { "epoch": 0.2810302053760966, "grad_norm": 0.283203125, "learning_rate": 8.669849246231156e-05, "loss": 1.1746, "step": 2747 }, { "epoch": 0.2811325097828589, "grad_norm": 0.291015625, "learning_rate": 8.669346733668342e-05, "loss": 1.116, "step": 2748 }, { "epoch": 0.2812348141896212, "grad_norm": 0.2890625, "learning_rate": 8.668844221105527e-05, "loss": 1.2711, "step": 2749 }, { "epoch": 0.28133711859638355, "grad_norm": 0.328125, "learning_rate": 8.668341708542715e-05, "loss": 1.193, "step": 2750 }, { "epoch": 0.28143942300314584, "grad_norm": 0.314453125, "learning_rate": 8.6678391959799e-05, "loss": 1.2475, "step": 2751 }, { "epoch": 0.2815417274099082, "grad_norm": 0.294921875, "learning_rate": 8.667336683417086e-05, "loss": 1.3928, "step": 2752 }, { "epoch": 0.2816440318166705, "grad_norm": 0.26953125, "learning_rate": 8.666834170854272e-05, "loss": 1.3675, "step": 2753 }, { "epoch": 0.2817463362234328, "grad_norm": 0.26953125, "learning_rate": 8.666331658291458e-05, "loss": 1.2208, "step": 2754 }, { "epoch": 0.28184864063019516, "grad_norm": 0.28125, "learning_rate": 8.665829145728643e-05, "loss": 1.3302, "step": 2755 }, { "epoch": 0.28195094503695745, "grad_norm": 0.333984375, "learning_rate": 8.66532663316583e-05, "loss": 1.34, "step": 2756 }, { "epoch": 0.2820532494437198, "grad_norm": 0.302734375, "learning_rate": 8.664824120603015e-05, "loss": 1.1633, "step": 2757 }, { "epoch": 0.28215555385048213, "grad_norm": 0.35546875, "learning_rate": 8.664321608040202e-05, "loss": 1.0938, "step": 2758 }, { "epoch": 0.2822578582572444, "grad_norm": 0.27734375, "learning_rate": 8.663819095477388e-05, "loss": 1.1774, "step": 2759 }, { "epoch": 0.28236016266400676, "grad_norm": 0.322265625, "learning_rate": 8.663316582914574e-05, "loss": 1.4109, "step": 2760 }, { "epoch": 0.28246246707076905, "grad_norm": 0.29296875, "learning_rate": 8.662814070351759e-05, "loss": 1.1813, "step": 2761 }, { "epoch": 0.2825647714775314, "grad_norm": 0.37109375, "learning_rate": 8.662311557788945e-05, "loss": 1.3598, "step": 2762 }, { "epoch": 0.28266707588429374, "grad_norm": 0.341796875, "learning_rate": 8.661809045226131e-05, "loss": 1.3125, "step": 2763 }, { "epoch": 0.282769380291056, "grad_norm": 0.33984375, "learning_rate": 8.661306532663317e-05, "loss": 1.2261, "step": 2764 }, { "epoch": 0.28287168469781837, "grad_norm": 0.314453125, "learning_rate": 8.660804020100502e-05, "loss": 1.3012, "step": 2765 }, { "epoch": 0.28297398910458066, "grad_norm": 0.294921875, "learning_rate": 8.66030150753769e-05, "loss": 1.4463, "step": 2766 }, { "epoch": 0.283076293511343, "grad_norm": 0.4375, "learning_rate": 8.659798994974875e-05, "loss": 1.2871, "step": 2767 }, { "epoch": 0.28317859791810535, "grad_norm": 0.28515625, "learning_rate": 8.659296482412061e-05, "loss": 1.3626, "step": 2768 }, { "epoch": 0.28328090232486763, "grad_norm": 0.328125, "learning_rate": 8.658793969849247e-05, "loss": 1.1349, "step": 2769 }, { "epoch": 0.28338320673163, "grad_norm": 0.306640625, "learning_rate": 8.658291457286433e-05, "loss": 1.4026, "step": 2770 }, { "epoch": 0.28348551113839227, "grad_norm": 0.2890625, "learning_rate": 8.657788944723618e-05, "loss": 1.2518, "step": 2771 }, { "epoch": 0.2835878155451546, "grad_norm": 0.294921875, "learning_rate": 8.657286432160804e-05, "loss": 1.4393, "step": 2772 }, { "epoch": 0.28369011995191695, "grad_norm": 0.326171875, "learning_rate": 8.65678391959799e-05, "loss": 1.3177, "step": 2773 }, { "epoch": 0.28379242435867924, "grad_norm": 0.322265625, "learning_rate": 8.656281407035177e-05, "loss": 1.27, "step": 2774 }, { "epoch": 0.2838947287654416, "grad_norm": 0.283203125, "learning_rate": 8.655778894472362e-05, "loss": 1.2309, "step": 2775 }, { "epoch": 0.28399703317220387, "grad_norm": 0.294921875, "learning_rate": 8.655276381909548e-05, "loss": 1.1821, "step": 2776 }, { "epoch": 0.2840993375789662, "grad_norm": 0.29296875, "learning_rate": 8.654773869346734e-05, "loss": 1.2228, "step": 2777 }, { "epoch": 0.28420164198572856, "grad_norm": 0.33984375, "learning_rate": 8.65427135678392e-05, "loss": 1.4383, "step": 2778 }, { "epoch": 0.28430394639249085, "grad_norm": 0.30859375, "learning_rate": 8.653768844221106e-05, "loss": 1.2895, "step": 2779 }, { "epoch": 0.2844062507992532, "grad_norm": 0.283203125, "learning_rate": 8.653266331658291e-05, "loss": 1.4215, "step": 2780 }, { "epoch": 0.2845085552060155, "grad_norm": 0.33203125, "learning_rate": 8.652763819095477e-05, "loss": 1.2833, "step": 2781 }, { "epoch": 0.2846108596127778, "grad_norm": 0.302734375, "learning_rate": 8.652261306532664e-05, "loss": 1.2301, "step": 2782 }, { "epoch": 0.28471316401954017, "grad_norm": 0.3359375, "learning_rate": 8.65175879396985e-05, "loss": 1.2644, "step": 2783 }, { "epoch": 0.28481546842630245, "grad_norm": 0.31640625, "learning_rate": 8.651256281407035e-05, "loss": 1.268, "step": 2784 }, { "epoch": 0.2849177728330648, "grad_norm": 0.30078125, "learning_rate": 8.650753768844222e-05, "loss": 1.2828, "step": 2785 }, { "epoch": 0.2850200772398271, "grad_norm": 0.296875, "learning_rate": 8.650251256281407e-05, "loss": 1.3515, "step": 2786 }, { "epoch": 0.28512238164658943, "grad_norm": 0.392578125, "learning_rate": 8.649748743718593e-05, "loss": 1.3456, "step": 2787 }, { "epoch": 0.2852246860533518, "grad_norm": 0.3359375, "learning_rate": 8.64924623115578e-05, "loss": 1.2792, "step": 2788 }, { "epoch": 0.28532699046011406, "grad_norm": 0.296875, "learning_rate": 8.648743718592966e-05, "loss": 1.3036, "step": 2789 }, { "epoch": 0.2854292948668764, "grad_norm": 0.314453125, "learning_rate": 8.64824120603015e-05, "loss": 1.3678, "step": 2790 }, { "epoch": 0.2855315992736387, "grad_norm": 0.35546875, "learning_rate": 8.647738693467337e-05, "loss": 1.3589, "step": 2791 }, { "epoch": 0.28563390368040104, "grad_norm": 0.3125, "learning_rate": 8.647236180904523e-05, "loss": 1.3689, "step": 2792 }, { "epoch": 0.2857362080871634, "grad_norm": 0.318359375, "learning_rate": 8.646733668341709e-05, "loss": 1.2581, "step": 2793 }, { "epoch": 0.28583851249392567, "grad_norm": 0.337890625, "learning_rate": 8.646231155778894e-05, "loss": 1.3456, "step": 2794 }, { "epoch": 0.285940816900688, "grad_norm": 0.3671875, "learning_rate": 8.645728643216082e-05, "loss": 1.2595, "step": 2795 }, { "epoch": 0.2860431213074503, "grad_norm": 0.375, "learning_rate": 8.645226130653266e-05, "loss": 1.1435, "step": 2796 }, { "epoch": 0.28614542571421264, "grad_norm": 0.29296875, "learning_rate": 8.644723618090453e-05, "loss": 1.3483, "step": 2797 }, { "epoch": 0.286247730120975, "grad_norm": 0.318359375, "learning_rate": 8.644221105527639e-05, "loss": 1.3046, "step": 2798 }, { "epoch": 0.2863500345277373, "grad_norm": 0.3125, "learning_rate": 8.643718592964825e-05, "loss": 1.3652, "step": 2799 }, { "epoch": 0.2864523389344996, "grad_norm": 0.30859375, "learning_rate": 8.64321608040201e-05, "loss": 1.1933, "step": 2800 }, { "epoch": 0.2865546433412619, "grad_norm": 0.310546875, "learning_rate": 8.642713567839197e-05, "loss": 1.2848, "step": 2801 }, { "epoch": 0.28665694774802425, "grad_norm": 0.33984375, "learning_rate": 8.642211055276382e-05, "loss": 1.2001, "step": 2802 }, { "epoch": 0.2867592521547866, "grad_norm": 0.33984375, "learning_rate": 8.641708542713568e-05, "loss": 1.3415, "step": 2803 }, { "epoch": 0.2868615565615489, "grad_norm": 0.330078125, "learning_rate": 8.641206030150755e-05, "loss": 1.3523, "step": 2804 }, { "epoch": 0.2869638609683112, "grad_norm": 0.28515625, "learning_rate": 8.640703517587941e-05, "loss": 1.286, "step": 2805 }, { "epoch": 0.2870661653750735, "grad_norm": 0.32421875, "learning_rate": 8.640201005025126e-05, "loss": 1.2798, "step": 2806 }, { "epoch": 0.28716846978183586, "grad_norm": 0.31640625, "learning_rate": 8.639698492462312e-05, "loss": 1.3141, "step": 2807 }, { "epoch": 0.2872707741885982, "grad_norm": 0.29296875, "learning_rate": 8.639195979899498e-05, "loss": 1.3701, "step": 2808 }, { "epoch": 0.2873730785953605, "grad_norm": 0.31640625, "learning_rate": 8.638693467336684e-05, "loss": 1.31, "step": 2809 }, { "epoch": 0.28747538300212283, "grad_norm": 0.283203125, "learning_rate": 8.638190954773869e-05, "loss": 1.2792, "step": 2810 }, { "epoch": 0.2875776874088851, "grad_norm": 0.302734375, "learning_rate": 8.637688442211055e-05, "loss": 1.1935, "step": 2811 }, { "epoch": 0.28767999181564746, "grad_norm": 0.353515625, "learning_rate": 8.637185929648242e-05, "loss": 1.3588, "step": 2812 }, { "epoch": 0.2877822962224098, "grad_norm": 0.302734375, "learning_rate": 8.636683417085428e-05, "loss": 1.3472, "step": 2813 }, { "epoch": 0.2878846006291721, "grad_norm": 0.302734375, "learning_rate": 8.636180904522614e-05, "loss": 1.2856, "step": 2814 }, { "epoch": 0.28798690503593444, "grad_norm": 0.279296875, "learning_rate": 8.635678391959799e-05, "loss": 1.3072, "step": 2815 }, { "epoch": 0.2880892094426967, "grad_norm": 0.3125, "learning_rate": 8.635175879396985e-05, "loss": 1.2753, "step": 2816 }, { "epoch": 0.28819151384945907, "grad_norm": 0.30859375, "learning_rate": 8.634673366834171e-05, "loss": 1.2794, "step": 2817 }, { "epoch": 0.2882938182562214, "grad_norm": 0.32421875, "learning_rate": 8.634170854271357e-05, "loss": 1.5109, "step": 2818 }, { "epoch": 0.2883961226629837, "grad_norm": 0.2890625, "learning_rate": 8.633668341708542e-05, "loss": 1.2535, "step": 2819 }, { "epoch": 0.28849842706974604, "grad_norm": 0.29296875, "learning_rate": 8.63316582914573e-05, "loss": 1.1817, "step": 2820 }, { "epoch": 0.28860073147650833, "grad_norm": 0.275390625, "learning_rate": 8.632663316582915e-05, "loss": 1.2252, "step": 2821 }, { "epoch": 0.2887030358832707, "grad_norm": 0.32421875, "learning_rate": 8.632160804020101e-05, "loss": 1.411, "step": 2822 }, { "epoch": 0.288805340290033, "grad_norm": 0.359375, "learning_rate": 8.631658291457286e-05, "loss": 1.3788, "step": 2823 }, { "epoch": 0.2889076446967953, "grad_norm": 0.337890625, "learning_rate": 8.631155778894473e-05, "loss": 1.1647, "step": 2824 }, { "epoch": 0.28900994910355765, "grad_norm": 0.283203125, "learning_rate": 8.630653266331658e-05, "loss": 1.2299, "step": 2825 }, { "epoch": 0.28911225351031994, "grad_norm": 0.27734375, "learning_rate": 8.630150753768844e-05, "loss": 1.1826, "step": 2826 }, { "epoch": 0.2892145579170823, "grad_norm": 0.314453125, "learning_rate": 8.62964824120603e-05, "loss": 1.0549, "step": 2827 }, { "epoch": 0.2893168623238446, "grad_norm": 0.310546875, "learning_rate": 8.629145728643217e-05, "loss": 1.3753, "step": 2828 }, { "epoch": 0.2894191667306069, "grad_norm": 0.3125, "learning_rate": 8.628643216080402e-05, "loss": 1.2292, "step": 2829 }, { "epoch": 0.28952147113736926, "grad_norm": 0.28515625, "learning_rate": 8.628140703517589e-05, "loss": 1.1877, "step": 2830 }, { "epoch": 0.28962377554413155, "grad_norm": 0.302734375, "learning_rate": 8.627638190954774e-05, "loss": 1.3454, "step": 2831 }, { "epoch": 0.2897260799508939, "grad_norm": 0.2734375, "learning_rate": 8.62713567839196e-05, "loss": 1.2519, "step": 2832 }, { "epoch": 0.28982838435765623, "grad_norm": 0.32421875, "learning_rate": 8.626633165829146e-05, "loss": 1.2465, "step": 2833 }, { "epoch": 0.2899306887644185, "grad_norm": 0.271484375, "learning_rate": 8.626130653266333e-05, "loss": 1.1173, "step": 2834 }, { "epoch": 0.29003299317118086, "grad_norm": 0.3046875, "learning_rate": 8.625628140703518e-05, "loss": 1.2512, "step": 2835 }, { "epoch": 0.29013529757794315, "grad_norm": 0.26953125, "learning_rate": 8.625125628140705e-05, "loss": 1.1709, "step": 2836 }, { "epoch": 0.2902376019847055, "grad_norm": 0.302734375, "learning_rate": 8.62462311557789e-05, "loss": 1.3282, "step": 2837 }, { "epoch": 0.29033990639146784, "grad_norm": 0.3046875, "learning_rate": 8.624120603015076e-05, "loss": 1.2196, "step": 2838 }, { "epoch": 0.2904422107982301, "grad_norm": 0.32421875, "learning_rate": 8.623618090452261e-05, "loss": 1.2177, "step": 2839 }, { "epoch": 0.29054451520499247, "grad_norm": 0.318359375, "learning_rate": 8.623115577889449e-05, "loss": 1.1915, "step": 2840 }, { "epoch": 0.29064681961175476, "grad_norm": 0.34765625, "learning_rate": 8.622613065326633e-05, "loss": 1.4107, "step": 2841 }, { "epoch": 0.2907491240185171, "grad_norm": 0.328125, "learning_rate": 8.62211055276382e-05, "loss": 1.3164, "step": 2842 }, { "epoch": 0.29085142842527945, "grad_norm": 0.318359375, "learning_rate": 8.621608040201006e-05, "loss": 1.264, "step": 2843 }, { "epoch": 0.29095373283204173, "grad_norm": 0.27734375, "learning_rate": 8.621105527638192e-05, "loss": 1.2058, "step": 2844 }, { "epoch": 0.2910560372388041, "grad_norm": 0.302734375, "learning_rate": 8.620603015075377e-05, "loss": 1.393, "step": 2845 }, { "epoch": 0.29115834164556637, "grad_norm": 0.3125, "learning_rate": 8.620100502512563e-05, "loss": 1.2132, "step": 2846 }, { "epoch": 0.2912606460523287, "grad_norm": 0.322265625, "learning_rate": 8.619597989949749e-05, "loss": 1.4702, "step": 2847 }, { "epoch": 0.29136295045909105, "grad_norm": 0.2734375, "learning_rate": 8.619095477386935e-05, "loss": 1.2564, "step": 2848 }, { "epoch": 0.29146525486585334, "grad_norm": 0.341796875, "learning_rate": 8.618592964824122e-05, "loss": 1.3934, "step": 2849 }, { "epoch": 0.2915675592726157, "grad_norm": 0.279296875, "learning_rate": 8.618090452261306e-05, "loss": 1.2953, "step": 2850 }, { "epoch": 0.29166986367937797, "grad_norm": 0.333984375, "learning_rate": 8.617587939698493e-05, "loss": 1.1858, "step": 2851 }, { "epoch": 0.2917721680861403, "grad_norm": 0.357421875, "learning_rate": 8.617085427135679e-05, "loss": 1.4151, "step": 2852 }, { "epoch": 0.29187447249290266, "grad_norm": 0.28125, "learning_rate": 8.616582914572865e-05, "loss": 1.2433, "step": 2853 }, { "epoch": 0.29197677689966495, "grad_norm": 0.3203125, "learning_rate": 8.61608040201005e-05, "loss": 1.4271, "step": 2854 }, { "epoch": 0.2920790813064273, "grad_norm": 0.302734375, "learning_rate": 8.615577889447236e-05, "loss": 1.0983, "step": 2855 }, { "epoch": 0.2921813857131896, "grad_norm": 0.326171875, "learning_rate": 8.615075376884422e-05, "loss": 1.2965, "step": 2856 }, { "epoch": 0.2922836901199519, "grad_norm": 0.337890625, "learning_rate": 8.614572864321609e-05, "loss": 1.2515, "step": 2857 }, { "epoch": 0.29238599452671427, "grad_norm": 0.3046875, "learning_rate": 8.614070351758793e-05, "loss": 1.2414, "step": 2858 }, { "epoch": 0.29248829893347655, "grad_norm": 0.275390625, "learning_rate": 8.613567839195981e-05, "loss": 1.2809, "step": 2859 }, { "epoch": 0.2925906033402389, "grad_norm": 0.3125, "learning_rate": 8.613065326633166e-05, "loss": 1.2884, "step": 2860 }, { "epoch": 0.2926929077470012, "grad_norm": 0.310546875, "learning_rate": 8.612562814070352e-05, "loss": 1.3647, "step": 2861 }, { "epoch": 0.29279521215376353, "grad_norm": 0.3515625, "learning_rate": 8.612060301507538e-05, "loss": 1.2993, "step": 2862 }, { "epoch": 0.2928975165605259, "grad_norm": 0.28515625, "learning_rate": 8.611557788944724e-05, "loss": 1.2742, "step": 2863 }, { "epoch": 0.29299982096728816, "grad_norm": 0.357421875, "learning_rate": 8.611055276381909e-05, "loss": 1.462, "step": 2864 }, { "epoch": 0.2931021253740505, "grad_norm": 0.318359375, "learning_rate": 8.610552763819097e-05, "loss": 1.4236, "step": 2865 }, { "epoch": 0.2932044297808128, "grad_norm": 0.29296875, "learning_rate": 8.610050251256282e-05, "loss": 1.1439, "step": 2866 }, { "epoch": 0.29330673418757514, "grad_norm": 0.298828125, "learning_rate": 8.609547738693468e-05, "loss": 1.3208, "step": 2867 }, { "epoch": 0.2934090385943374, "grad_norm": 0.3203125, "learning_rate": 8.609045226130654e-05, "loss": 1.2454, "step": 2868 }, { "epoch": 0.29351134300109977, "grad_norm": 0.31640625, "learning_rate": 8.60854271356784e-05, "loss": 1.3906, "step": 2869 }, { "epoch": 0.2936136474078621, "grad_norm": 0.326171875, "learning_rate": 8.608040201005025e-05, "loss": 1.4226, "step": 2870 }, { "epoch": 0.2937159518146244, "grad_norm": 0.263671875, "learning_rate": 8.607537688442211e-05, "loss": 1.1754, "step": 2871 }, { "epoch": 0.29381825622138674, "grad_norm": 0.298828125, "learning_rate": 8.607035175879398e-05, "loss": 1.313, "step": 2872 }, { "epoch": 0.29392056062814903, "grad_norm": 0.2890625, "learning_rate": 8.606532663316584e-05, "loss": 1.1734, "step": 2873 }, { "epoch": 0.2940228650349114, "grad_norm": 0.271484375, "learning_rate": 8.606030150753769e-05, "loss": 1.2436, "step": 2874 }, { "epoch": 0.2941251694416737, "grad_norm": 0.322265625, "learning_rate": 8.605527638190956e-05, "loss": 1.3063, "step": 2875 }, { "epoch": 0.294227473848436, "grad_norm": 0.294921875, "learning_rate": 8.605025125628141e-05, "loss": 1.4807, "step": 2876 }, { "epoch": 0.29432977825519835, "grad_norm": 0.314453125, "learning_rate": 8.604522613065327e-05, "loss": 1.28, "step": 2877 }, { "epoch": 0.29443208266196064, "grad_norm": 0.294921875, "learning_rate": 8.604020100502513e-05, "loss": 1.2602, "step": 2878 }, { "epoch": 0.294534387068723, "grad_norm": 0.9375, "learning_rate": 8.6035175879397e-05, "loss": 1.2516, "step": 2879 }, { "epoch": 0.2946366914754853, "grad_norm": 0.30078125, "learning_rate": 8.603015075376884e-05, "loss": 1.2672, "step": 2880 }, { "epoch": 0.2947389958822476, "grad_norm": 0.287109375, "learning_rate": 8.60251256281407e-05, "loss": 1.1709, "step": 2881 }, { "epoch": 0.29484130028900996, "grad_norm": 0.33984375, "learning_rate": 8.602010050251257e-05, "loss": 1.3053, "step": 2882 }, { "epoch": 0.29494360469577224, "grad_norm": 0.3046875, "learning_rate": 8.601507537688443e-05, "loss": 1.3093, "step": 2883 }, { "epoch": 0.2950459091025346, "grad_norm": 0.275390625, "learning_rate": 8.601005025125628e-05, "loss": 1.2323, "step": 2884 }, { "epoch": 0.29514821350929693, "grad_norm": 0.341796875, "learning_rate": 8.600502512562814e-05, "loss": 1.5389, "step": 2885 }, { "epoch": 0.2952505179160592, "grad_norm": 0.287109375, "learning_rate": 8.6e-05, "loss": 1.4941, "step": 2886 }, { "epoch": 0.29535282232282156, "grad_norm": 0.3671875, "learning_rate": 8.599497487437187e-05, "loss": 1.306, "step": 2887 }, { "epoch": 0.29545512672958385, "grad_norm": 0.294921875, "learning_rate": 8.598994974874373e-05, "loss": 1.2073, "step": 2888 }, { "epoch": 0.2955574311363462, "grad_norm": 0.29296875, "learning_rate": 8.598492462311558e-05, "loss": 1.3055, "step": 2889 }, { "epoch": 0.29565973554310854, "grad_norm": 0.294921875, "learning_rate": 8.597989949748744e-05, "loss": 1.3324, "step": 2890 }, { "epoch": 0.2957620399498708, "grad_norm": 0.328125, "learning_rate": 8.59748743718593e-05, "loss": 1.2904, "step": 2891 }, { "epoch": 0.29586434435663317, "grad_norm": 0.2890625, "learning_rate": 8.596984924623116e-05, "loss": 1.2032, "step": 2892 }, { "epoch": 0.29596664876339546, "grad_norm": 0.291015625, "learning_rate": 8.596482412060301e-05, "loss": 1.3338, "step": 2893 }, { "epoch": 0.2960689531701578, "grad_norm": 0.283203125, "learning_rate": 8.595979899497489e-05, "loss": 1.2141, "step": 2894 }, { "epoch": 0.29617125757692014, "grad_norm": 0.29296875, "learning_rate": 8.595477386934673e-05, "loss": 1.2024, "step": 2895 }, { "epoch": 0.29627356198368243, "grad_norm": 0.294921875, "learning_rate": 8.59497487437186e-05, "loss": 1.2922, "step": 2896 }, { "epoch": 0.2963758663904448, "grad_norm": 0.302734375, "learning_rate": 8.594472361809046e-05, "loss": 1.3213, "step": 2897 }, { "epoch": 0.29647817079720706, "grad_norm": 0.296875, "learning_rate": 8.593969849246232e-05, "loss": 1.1436, "step": 2898 }, { "epoch": 0.2965804752039694, "grad_norm": 0.298828125, "learning_rate": 8.593467336683417e-05, "loss": 1.1766, "step": 2899 }, { "epoch": 0.29668277961073175, "grad_norm": 0.310546875, "learning_rate": 8.592964824120603e-05, "loss": 1.3176, "step": 2900 }, { "epoch": 0.29678508401749404, "grad_norm": 0.318359375, "learning_rate": 8.592462311557789e-05, "loss": 1.3699, "step": 2901 }, { "epoch": 0.2968873884242564, "grad_norm": 0.279296875, "learning_rate": 8.591959798994976e-05, "loss": 1.176, "step": 2902 }, { "epoch": 0.29698969283101867, "grad_norm": 0.287109375, "learning_rate": 8.59145728643216e-05, "loss": 1.1511, "step": 2903 }, { "epoch": 0.297091997237781, "grad_norm": 0.322265625, "learning_rate": 8.590954773869348e-05, "loss": 1.1833, "step": 2904 }, { "epoch": 0.29719430164454336, "grad_norm": 0.296875, "learning_rate": 8.590452261306533e-05, "loss": 1.1632, "step": 2905 }, { "epoch": 0.29729660605130565, "grad_norm": 0.3046875, "learning_rate": 8.589949748743719e-05, "loss": 1.334, "step": 2906 }, { "epoch": 0.297398910458068, "grad_norm": 0.326171875, "learning_rate": 8.589447236180905e-05, "loss": 1.2012, "step": 2907 }, { "epoch": 0.2975012148648303, "grad_norm": 0.294921875, "learning_rate": 8.588944723618091e-05, "loss": 1.2721, "step": 2908 }, { "epoch": 0.2976035192715926, "grad_norm": 0.3046875, "learning_rate": 8.588442211055276e-05, "loss": 1.4865, "step": 2909 }, { "epoch": 0.29770582367835496, "grad_norm": 0.291015625, "learning_rate": 8.587939698492464e-05, "loss": 1.1921, "step": 2910 }, { "epoch": 0.29780812808511725, "grad_norm": 0.265625, "learning_rate": 8.587437185929649e-05, "loss": 1.1223, "step": 2911 }, { "epoch": 0.2979104324918796, "grad_norm": 0.271484375, "learning_rate": 8.586934673366835e-05, "loss": 1.3118, "step": 2912 }, { "epoch": 0.2980127368986419, "grad_norm": 0.287109375, "learning_rate": 8.586432160804021e-05, "loss": 1.2715, "step": 2913 }, { "epoch": 0.2981150413054042, "grad_norm": 0.326171875, "learning_rate": 8.585929648241207e-05, "loss": 1.3761, "step": 2914 }, { "epoch": 0.29821734571216657, "grad_norm": 0.28515625, "learning_rate": 8.585427135678392e-05, "loss": 1.2553, "step": 2915 }, { "epoch": 0.29831965011892886, "grad_norm": 0.345703125, "learning_rate": 8.584924623115578e-05, "loss": 1.2189, "step": 2916 }, { "epoch": 0.2984219545256912, "grad_norm": 0.3125, "learning_rate": 8.584422110552765e-05, "loss": 1.3271, "step": 2917 }, { "epoch": 0.2985242589324535, "grad_norm": 0.310546875, "learning_rate": 8.583919597989951e-05, "loss": 1.4403, "step": 2918 }, { "epoch": 0.29862656333921583, "grad_norm": 0.361328125, "learning_rate": 8.583417085427136e-05, "loss": 1.1422, "step": 2919 }, { "epoch": 0.2987288677459782, "grad_norm": 0.267578125, "learning_rate": 8.582914572864322e-05, "loss": 1.0654, "step": 2920 }, { "epoch": 0.29883117215274047, "grad_norm": 0.298828125, "learning_rate": 8.582412060301508e-05, "loss": 1.2248, "step": 2921 }, { "epoch": 0.2989334765595028, "grad_norm": 0.306640625, "learning_rate": 8.581909547738693e-05, "loss": 1.3322, "step": 2922 }, { "epoch": 0.2990357809662651, "grad_norm": 0.31640625, "learning_rate": 8.58140703517588e-05, "loss": 1.3889, "step": 2923 }, { "epoch": 0.29913808537302744, "grad_norm": 0.3203125, "learning_rate": 8.580904522613065e-05, "loss": 1.2363, "step": 2924 }, { "epoch": 0.2992403897797898, "grad_norm": 0.302734375, "learning_rate": 8.580402010050251e-05, "loss": 1.2473, "step": 2925 }, { "epoch": 0.29934269418655207, "grad_norm": 0.28515625, "learning_rate": 8.579899497487438e-05, "loss": 1.2672, "step": 2926 }, { "epoch": 0.2994449985933144, "grad_norm": 0.298828125, "learning_rate": 8.579396984924624e-05, "loss": 1.1607, "step": 2927 }, { "epoch": 0.2995473030000767, "grad_norm": 0.29296875, "learning_rate": 8.578894472361809e-05, "loss": 1.2407, "step": 2928 }, { "epoch": 0.29964960740683905, "grad_norm": 0.26953125, "learning_rate": 8.578391959798996e-05, "loss": 1.0481, "step": 2929 }, { "epoch": 0.2997519118136014, "grad_norm": 0.27734375, "learning_rate": 8.577889447236181e-05, "loss": 1.252, "step": 2930 }, { "epoch": 0.2998542162203637, "grad_norm": 0.3125, "learning_rate": 8.577386934673367e-05, "loss": 1.301, "step": 2931 }, { "epoch": 0.299956520627126, "grad_norm": 0.294921875, "learning_rate": 8.576884422110552e-05, "loss": 1.3103, "step": 2932 }, { "epoch": 0.3000588250338883, "grad_norm": 0.34765625, "learning_rate": 8.57638190954774e-05, "loss": 1.4189, "step": 2933 }, { "epoch": 0.30016112944065065, "grad_norm": 0.33203125, "learning_rate": 8.575879396984925e-05, "loss": 1.3079, "step": 2934 }, { "epoch": 0.300263433847413, "grad_norm": 0.310546875, "learning_rate": 8.575376884422111e-05, "loss": 1.3566, "step": 2935 }, { "epoch": 0.3003657382541753, "grad_norm": 0.703125, "learning_rate": 8.574874371859297e-05, "loss": 1.2263, "step": 2936 }, { "epoch": 0.30046804266093763, "grad_norm": 0.306640625, "learning_rate": 8.574371859296483e-05, "loss": 1.25, "step": 2937 }, { "epoch": 0.3005703470676999, "grad_norm": 0.310546875, "learning_rate": 8.573869346733668e-05, "loss": 1.3851, "step": 2938 }, { "epoch": 0.30067265147446226, "grad_norm": 0.279296875, "learning_rate": 8.573366834170856e-05, "loss": 1.1802, "step": 2939 }, { "epoch": 0.3007749558812246, "grad_norm": 0.43359375, "learning_rate": 8.57286432160804e-05, "loss": 1.2666, "step": 2940 }, { "epoch": 0.3008772602879869, "grad_norm": 0.28125, "learning_rate": 8.572361809045227e-05, "loss": 1.3127, "step": 2941 }, { "epoch": 0.30097956469474924, "grad_norm": 0.30078125, "learning_rate": 8.571859296482413e-05, "loss": 1.3647, "step": 2942 }, { "epoch": 0.3010818691015115, "grad_norm": 0.3125, "learning_rate": 8.571356783919599e-05, "loss": 1.341, "step": 2943 }, { "epoch": 0.30118417350827387, "grad_norm": 0.32421875, "learning_rate": 8.570854271356784e-05, "loss": 1.3837, "step": 2944 }, { "epoch": 0.3012864779150362, "grad_norm": 0.279296875, "learning_rate": 8.570351758793971e-05, "loss": 1.2315, "step": 2945 }, { "epoch": 0.3013887823217985, "grad_norm": 0.318359375, "learning_rate": 8.569849246231156e-05, "loss": 1.2005, "step": 2946 }, { "epoch": 0.30149108672856084, "grad_norm": 0.31640625, "learning_rate": 8.569346733668342e-05, "loss": 1.3744, "step": 2947 }, { "epoch": 0.30159339113532313, "grad_norm": 0.31640625, "learning_rate": 8.568844221105527e-05, "loss": 1.1906, "step": 2948 }, { "epoch": 0.3016956955420855, "grad_norm": 0.26953125, "learning_rate": 8.568341708542715e-05, "loss": 1.1509, "step": 2949 }, { "epoch": 0.3017979999488478, "grad_norm": 0.373046875, "learning_rate": 8.5678391959799e-05, "loss": 1.461, "step": 2950 }, { "epoch": 0.3019003043556101, "grad_norm": 0.34375, "learning_rate": 8.567336683417086e-05, "loss": 1.3664, "step": 2951 }, { "epoch": 0.30200260876237245, "grad_norm": 0.294921875, "learning_rate": 8.566834170854272e-05, "loss": 1.2378, "step": 2952 }, { "epoch": 0.30210491316913474, "grad_norm": 0.333984375, "learning_rate": 8.566331658291458e-05, "loss": 1.4271, "step": 2953 }, { "epoch": 0.3022072175758971, "grad_norm": 0.3125, "learning_rate": 8.565829145728643e-05, "loss": 1.3261, "step": 2954 }, { "epoch": 0.3023095219826594, "grad_norm": 0.3046875, "learning_rate": 8.56532663316583e-05, "loss": 1.308, "step": 2955 }, { "epoch": 0.3024118263894217, "grad_norm": 0.27734375, "learning_rate": 8.564824120603016e-05, "loss": 1.2081, "step": 2956 }, { "epoch": 0.30251413079618406, "grad_norm": 0.298828125, "learning_rate": 8.5643216080402e-05, "loss": 1.1675, "step": 2957 }, { "epoch": 0.30261643520294634, "grad_norm": 0.376953125, "learning_rate": 8.563819095477388e-05, "loss": 1.3165, "step": 2958 }, { "epoch": 0.3027187396097087, "grad_norm": 0.3125, "learning_rate": 8.563316582914573e-05, "loss": 1.4183, "step": 2959 }, { "epoch": 0.30282104401647103, "grad_norm": 0.29296875, "learning_rate": 8.562814070351759e-05, "loss": 1.2871, "step": 2960 }, { "epoch": 0.3029233484232333, "grad_norm": 0.314453125, "learning_rate": 8.562311557788945e-05, "loss": 1.2425, "step": 2961 }, { "epoch": 0.30302565282999566, "grad_norm": 0.287109375, "learning_rate": 8.561809045226131e-05, "loss": 1.2865, "step": 2962 }, { "epoch": 0.30312795723675795, "grad_norm": 0.37890625, "learning_rate": 8.561306532663316e-05, "loss": 1.3066, "step": 2963 }, { "epoch": 0.3032302616435203, "grad_norm": 0.27734375, "learning_rate": 8.560804020100503e-05, "loss": 1.3558, "step": 2964 }, { "epoch": 0.30333256605028264, "grad_norm": 0.283203125, "learning_rate": 8.560301507537689e-05, "loss": 1.2626, "step": 2965 }, { "epoch": 0.3034348704570449, "grad_norm": 0.31640625, "learning_rate": 8.559798994974875e-05, "loss": 1.3718, "step": 2966 }, { "epoch": 0.30353717486380727, "grad_norm": 0.326171875, "learning_rate": 8.55929648241206e-05, "loss": 1.149, "step": 2967 }, { "epoch": 0.30363947927056956, "grad_norm": 0.2890625, "learning_rate": 8.558793969849247e-05, "loss": 1.3239, "step": 2968 }, { "epoch": 0.3037417836773319, "grad_norm": 0.2890625, "learning_rate": 8.558291457286432e-05, "loss": 1.3682, "step": 2969 }, { "epoch": 0.30384408808409424, "grad_norm": 0.255859375, "learning_rate": 8.557788944723618e-05, "loss": 1.1596, "step": 2970 }, { "epoch": 0.30394639249085653, "grad_norm": 0.2734375, "learning_rate": 8.557286432160805e-05, "loss": 1.0663, "step": 2971 }, { "epoch": 0.3040486968976189, "grad_norm": 0.3359375, "learning_rate": 8.556783919597991e-05, "loss": 1.34, "step": 2972 }, { "epoch": 0.30415100130438116, "grad_norm": 0.322265625, "learning_rate": 8.556281407035176e-05, "loss": 1.3952, "step": 2973 }, { "epoch": 0.3042533057111435, "grad_norm": 0.30078125, "learning_rate": 8.555778894472363e-05, "loss": 1.4323, "step": 2974 }, { "epoch": 0.30435561011790585, "grad_norm": 0.296875, "learning_rate": 8.555276381909548e-05, "loss": 1.2063, "step": 2975 }, { "epoch": 0.30445791452466814, "grad_norm": 0.3203125, "learning_rate": 8.554773869346734e-05, "loss": 1.3, "step": 2976 }, { "epoch": 0.3045602189314305, "grad_norm": 0.28515625, "learning_rate": 8.55427135678392e-05, "loss": 1.4402, "step": 2977 }, { "epoch": 0.30466252333819277, "grad_norm": 0.283203125, "learning_rate": 8.553768844221107e-05, "loss": 1.2818, "step": 2978 }, { "epoch": 0.3047648277449551, "grad_norm": 0.33203125, "learning_rate": 8.553266331658292e-05, "loss": 1.2363, "step": 2979 }, { "epoch": 0.30486713215171746, "grad_norm": 0.302734375, "learning_rate": 8.552763819095478e-05, "loss": 1.2065, "step": 2980 }, { "epoch": 0.30496943655847975, "grad_norm": 0.337890625, "learning_rate": 8.552261306532664e-05, "loss": 1.3125, "step": 2981 }, { "epoch": 0.3050717409652421, "grad_norm": 0.30078125, "learning_rate": 8.55175879396985e-05, "loss": 1.1799, "step": 2982 }, { "epoch": 0.3051740453720044, "grad_norm": 0.296875, "learning_rate": 8.551256281407035e-05, "loss": 1.0289, "step": 2983 }, { "epoch": 0.3052763497787667, "grad_norm": 0.310546875, "learning_rate": 8.550753768844223e-05, "loss": 1.3337, "step": 2984 }, { "epoch": 0.30537865418552906, "grad_norm": 0.30859375, "learning_rate": 8.550251256281407e-05, "loss": 1.4562, "step": 2985 }, { "epoch": 0.30548095859229135, "grad_norm": 0.294921875, "learning_rate": 8.549748743718594e-05, "loss": 1.2575, "step": 2986 }, { "epoch": 0.3055832629990537, "grad_norm": 0.34375, "learning_rate": 8.54924623115578e-05, "loss": 1.268, "step": 2987 }, { "epoch": 0.305685567405816, "grad_norm": 0.375, "learning_rate": 8.548743718592965e-05, "loss": 1.3316, "step": 2988 }, { "epoch": 0.3057878718125783, "grad_norm": 0.294921875, "learning_rate": 8.548241206030151e-05, "loss": 1.3248, "step": 2989 }, { "epoch": 0.30589017621934067, "grad_norm": 0.33203125, "learning_rate": 8.547738693467337e-05, "loss": 1.0458, "step": 2990 }, { "epoch": 0.30599248062610296, "grad_norm": 0.259765625, "learning_rate": 8.547236180904523e-05, "loss": 1.1364, "step": 2991 }, { "epoch": 0.3060947850328653, "grad_norm": 0.32421875, "learning_rate": 8.546733668341708e-05, "loss": 1.2603, "step": 2992 }, { "epoch": 0.3061970894396276, "grad_norm": 0.3203125, "learning_rate": 8.546231155778894e-05, "loss": 1.2379, "step": 2993 }, { "epoch": 0.30629939384638993, "grad_norm": 0.302734375, "learning_rate": 8.54572864321608e-05, "loss": 1.219, "step": 2994 }, { "epoch": 0.3064016982531523, "grad_norm": 0.3203125, "learning_rate": 8.545226130653267e-05, "loss": 1.2087, "step": 2995 }, { "epoch": 0.30650400265991457, "grad_norm": 0.29296875, "learning_rate": 8.544723618090452e-05, "loss": 1.2962, "step": 2996 }, { "epoch": 0.3066063070666769, "grad_norm": 0.337890625, "learning_rate": 8.544221105527639e-05, "loss": 1.333, "step": 2997 }, { "epoch": 0.3067086114734392, "grad_norm": 0.361328125, "learning_rate": 8.543718592964824e-05, "loss": 1.2234, "step": 2998 }, { "epoch": 0.30681091588020154, "grad_norm": 0.318359375, "learning_rate": 8.54321608040201e-05, "loss": 1.3074, "step": 2999 }, { "epoch": 0.3069132202869639, "grad_norm": 0.2890625, "learning_rate": 8.542713567839196e-05, "loss": 1.2326, "step": 3000 } ], "logging_steps": 1, "max_steps": 20000, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.74943499354112e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }