diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100755--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7042 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4342162396873643, + "eval_steps": 500, + "global_step": 1000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0004342162396873643, + "grad_norm": 0.9606487154960632, + "learning_rate": 5e-05, + "loss": 9.188, + "step": 1 + }, + { + "epoch": 0.0008684324793747286, + "grad_norm": 0.9193044304847717, + "learning_rate": 0.0001, + "loss": 9.1897, + "step": 2 + }, + { + "epoch": 0.0013026487190620929, + "grad_norm": 0.9406704306602478, + "learning_rate": 0.00015, + "loss": 9.0726, + "step": 3 + }, + { + "epoch": 0.0017368649587494573, + "grad_norm": 0.9401040077209473, + "learning_rate": 0.0002, + "loss": 8.7129, + "step": 4 + }, + { + "epoch": 0.0021710811984368217, + "grad_norm": 0.8919561505317688, + "learning_rate": 0.00025, + "loss": 8.1364, + "step": 5 + }, + { + "epoch": 0.0026052974381241857, + "grad_norm": 0.8310884237289429, + "learning_rate": 0.0003, + "loss": 7.4088, + "step": 6 + }, + { + "epoch": 0.00303951367781155, + "grad_norm": 0.748382031917572, + "learning_rate": 0.00035, + "loss": 6.6918, + "step": 7 + }, + { + "epoch": 0.0034737299174989146, + "grad_norm": 0.6353039145469666, + "learning_rate": 0.0004, + "loss": 5.8753, + "step": 8 + }, + { + "epoch": 0.0039079461571862786, + "grad_norm": 0.5517627000808716, + "learning_rate": 0.00045000000000000004, + "loss": 5.4347, + "step": 9 + }, + { + "epoch": 0.004342162396873643, + "grad_norm": 0.5389806628227234, + "learning_rate": 0.0005, + "loss": 5.0181, + "step": 10 + }, + { + "epoch": 0.004776378636561007, + "grad_norm": 0.6018754243850708, + "learning_rate": 0.00055, + "loss": 4.9487, + "step": 11 + }, + { + "epoch": 0.005210594876248371, + "grad_norm": 0.6734316349029541, + "learning_rate": 0.0006, + "loss": 4.9233, + "step": 12 + }, + { + "epoch": 0.005644811115935736, + "grad_norm": 0.6842392086982727, + "learning_rate": 0.0006500000000000001, + "loss": 4.8078, + "step": 13 + }, + { + "epoch": 0.0060790273556231, + "grad_norm": 0.6519020199775696, + "learning_rate": 0.0007, + "loss": 4.6065, + "step": 14 + }, + { + "epoch": 0.006513243595310464, + "grad_norm": 0.5788787603378296, + "learning_rate": 0.00075, + "loss": 4.3967, + "step": 15 + }, + { + "epoch": 0.006947459834997829, + "grad_norm": 0.5370550751686096, + "learning_rate": 0.0008, + "loss": 4.1971, + "step": 16 + }, + { + "epoch": 0.007381676074685193, + "grad_norm": 0.4862042963504791, + "learning_rate": 0.00085, + "loss": 3.8583, + "step": 17 + }, + { + "epoch": 0.007815892314372557, + "grad_norm": 0.4636250436306, + "learning_rate": 0.0009000000000000001, + "loss": 3.6372, + "step": 18 + }, + { + "epoch": 0.008250108554059922, + "grad_norm": 0.4613461494445801, + "learning_rate": 0.00095, + "loss": 3.5642, + "step": 19 + }, + { + "epoch": 0.008684324793747287, + "grad_norm": 0.4593052566051483, + "learning_rate": 0.001, + "loss": 3.4463, + "step": 20 + }, + { + "epoch": 0.00911854103343465, + "grad_norm": 0.4367975890636444, + "learning_rate": 0.0009999974308631955, + "loss": 3.2621, + "step": 21 + }, + { + "epoch": 0.009552757273122015, + "grad_norm": 0.41558611392974854, + "learning_rate": 0.000999989723479183, + "loss": 3.0598, + "step": 22 + }, + { + "epoch": 0.00998697351280938, + "grad_norm": 0.3967452049255371, + "learning_rate": 0.0009999768779271685, + "loss": 2.936, + "step": 23 + }, + { + "epoch": 0.010421189752496743, + "grad_norm": 0.3664601147174835, + "learning_rate": 0.0009999588943391596, + "loss": 2.6913, + "step": 24 + }, + { + "epoch": 0.010855405992184108, + "grad_norm": 0.359183669090271, + "learning_rate": 0.0009999357728999656, + "loss": 2.5639, + "step": 25 + }, + { + "epoch": 0.011289622231871473, + "grad_norm": 0.3342216908931732, + "learning_rate": 0.000999907513847195, + "loss": 2.401, + "step": 26 + }, + { + "epoch": 0.011723838471558836, + "grad_norm": 0.36653760075569153, + "learning_rate": 0.0009998741174712534, + "loss": 2.6216, + "step": 27 + }, + { + "epoch": 0.0121580547112462, + "grad_norm": 0.30254194140434265, + "learning_rate": 0.00099983558411534, + "loss": 2.2086, + "step": 28 + }, + { + "epoch": 0.012592270950933565, + "grad_norm": 0.2917495369911194, + "learning_rate": 0.0009997919141754449, + "loss": 2.1274, + "step": 29 + }, + { + "epoch": 0.013026487190620929, + "grad_norm": 0.2690434455871582, + "learning_rate": 0.000999743108100344, + "loss": 2.0581, + "step": 30 + }, + { + "epoch": 0.013460703430308293, + "grad_norm": 0.26215028762817383, + "learning_rate": 0.0009996891663915954, + "loss": 1.9764, + "step": 31 + }, + { + "epoch": 0.013894919669995658, + "grad_norm": 0.2453673779964447, + "learning_rate": 0.0009996300896035338, + "loss": 1.9219, + "step": 32 + }, + { + "epoch": 0.014329135909683021, + "grad_norm": 0.24287743866443634, + "learning_rate": 0.0009995658783432644, + "loss": 1.8433, + "step": 33 + }, + { + "epoch": 0.014763352149370386, + "grad_norm": 0.2258252203464508, + "learning_rate": 0.0009994965332706573, + "loss": 1.7392, + "step": 34 + }, + { + "epoch": 0.015197568389057751, + "grad_norm": 0.21900461614131927, + "learning_rate": 0.0009994220550983404, + "loss": 1.7185, + "step": 35 + }, + { + "epoch": 0.015631784628745114, + "grad_norm": 0.21046680212020874, + "learning_rate": 0.0009993424445916923, + "loss": 1.6541, + "step": 36 + }, + { + "epoch": 0.01606600086843248, + "grad_norm": 0.19736959040164948, + "learning_rate": 0.0009992577025688338, + "loss": 1.6321, + "step": 37 + }, + { + "epoch": 0.016500217108119844, + "grad_norm": 0.19068720936775208, + "learning_rate": 0.0009991678299006205, + "loss": 1.6236, + "step": 38 + }, + { + "epoch": 0.01693443334780721, + "grad_norm": 0.18818727135658264, + "learning_rate": 0.000999072827510633, + "loss": 1.5791, + "step": 39 + }, + { + "epoch": 0.017368649587494574, + "grad_norm": 0.17485196888446808, + "learning_rate": 0.0009989726963751681, + "loss": 1.6086, + "step": 40 + }, + { + "epoch": 0.017802865827181935, + "grad_norm": 0.1762186735868454, + "learning_rate": 0.000998867437523228, + "loss": 1.5328, + "step": 41 + }, + { + "epoch": 0.0182370820668693, + "grad_norm": 0.19658854603767395, + "learning_rate": 0.0009987570520365104, + "loss": 1.4491, + "step": 42 + }, + { + "epoch": 0.018671298306556665, + "grad_norm": 0.1639322191476822, + "learning_rate": 0.0009986415410493965, + "loss": 1.4447, + "step": 43 + }, + { + "epoch": 0.01910551454624403, + "grad_norm": 0.1552889496088028, + "learning_rate": 0.000998520905748941, + "loss": 1.4345, + "step": 44 + }, + { + "epoch": 0.019539730785931395, + "grad_norm": 0.14960208535194397, + "learning_rate": 0.0009983951473748578, + "loss": 1.364, + "step": 45 + }, + { + "epoch": 0.01997394702561876, + "grad_norm": 0.14552171528339386, + "learning_rate": 0.0009982642672195091, + "loss": 1.3422, + "step": 46 + }, + { + "epoch": 0.02040816326530612, + "grad_norm": 0.14638382196426392, + "learning_rate": 0.0009981282666278908, + "loss": 1.325, + "step": 47 + }, + { + "epoch": 0.020842379504993486, + "grad_norm": 0.1372004598379135, + "learning_rate": 0.0009979871469976197, + "loss": 1.2685, + "step": 48 + }, + { + "epoch": 0.02127659574468085, + "grad_norm": 0.12978093326091766, + "learning_rate": 0.0009978409097789176, + "loss": 1.3268, + "step": 49 + }, + { + "epoch": 0.021710811984368215, + "grad_norm": 0.13076020777225494, + "learning_rate": 0.0009976895564745992, + "loss": 1.2211, + "step": 50 + }, + { + "epoch": 0.02214502822405558, + "grad_norm": 0.1264386773109436, + "learning_rate": 0.000997533088640053, + "loss": 1.2597, + "step": 51 + }, + { + "epoch": 0.022579244463742945, + "grad_norm": 0.12510421872138977, + "learning_rate": 0.0009973715078832287, + "loss": 1.178, + "step": 52 + }, + { + "epoch": 0.02301346070343031, + "grad_norm": 0.13376790285110474, + "learning_rate": 0.0009972048158646183, + "loss": 1.1614, + "step": 53 + }, + { + "epoch": 0.02344767694311767, + "grad_norm": 0.11965133249759674, + "learning_rate": 0.00099703301429724, + "loss": 1.1549, + "step": 54 + }, + { + "epoch": 0.023881893182805036, + "grad_norm": 0.10872318595647812, + "learning_rate": 0.0009968561049466214, + "loss": 1.143, + "step": 55 + }, + { + "epoch": 0.0243161094224924, + "grad_norm": 0.12483104318380356, + "learning_rate": 0.000996674089630779, + "loss": 1.2124, + "step": 56 + }, + { + "epoch": 0.024750325662179766, + "grad_norm": 0.10931334644556046, + "learning_rate": 0.0009964869702202021, + "loss": 1.1176, + "step": 57 + }, + { + "epoch": 0.02518454190186713, + "grad_norm": 0.12575986981391907, + "learning_rate": 0.0009962947486378325, + "loss": 1.0858, + "step": 58 + }, + { + "epoch": 0.025618758141554496, + "grad_norm": 0.11589810997247696, + "learning_rate": 0.0009960974268590438, + "loss": 1.0867, + "step": 59 + }, + { + "epoch": 0.026052974381241857, + "grad_norm": 0.10477310419082642, + "learning_rate": 0.000995895006911623, + "loss": 1.1162, + "step": 60 + }, + { + "epoch": 0.026487190620929222, + "grad_norm": 0.09671472758054733, + "learning_rate": 0.0009956874908757481, + "loss": 1.0662, + "step": 61 + }, + { + "epoch": 0.026921406860616587, + "grad_norm": 0.10130201280117035, + "learning_rate": 0.0009954748808839674, + "loss": 1.0826, + "step": 62 + }, + { + "epoch": 0.02735562310030395, + "grad_norm": 0.10421687364578247, + "learning_rate": 0.0009952571791211776, + "loss": 1.0302, + "step": 63 + }, + { + "epoch": 0.027789839339991317, + "grad_norm": 0.09363257884979248, + "learning_rate": 0.0009950343878246009, + "loss": 1.0769, + "step": 64 + }, + { + "epoch": 0.02822405557967868, + "grad_norm": 0.10000266879796982, + "learning_rate": 0.000994806509283763, + "loss": 1.0048, + "step": 65 + }, + { + "epoch": 0.028658271819366043, + "grad_norm": 0.091400146484375, + "learning_rate": 0.0009945735458404682, + "loss": 0.9903, + "step": 66 + }, + { + "epoch": 0.029092488059053408, + "grad_norm": 0.09233607351779938, + "learning_rate": 0.0009943354998887763, + "loss": 1.0563, + "step": 67 + }, + { + "epoch": 0.029526704298740773, + "grad_norm": 0.0991148054599762, + "learning_rate": 0.0009940923738749779, + "loss": 1.0117, + "step": 68 + }, + { + "epoch": 0.029960920538428137, + "grad_norm": 0.08926694095134735, + "learning_rate": 0.0009938441702975688, + "loss": 0.974, + "step": 69 + }, + { + "epoch": 0.030395136778115502, + "grad_norm": 0.08934075385332108, + "learning_rate": 0.0009935908917072251, + "loss": 1.0283, + "step": 70 + }, + { + "epoch": 0.030829353017802867, + "grad_norm": 0.11343932151794434, + "learning_rate": 0.000993332540706776, + "loss": 0.998, + "step": 71 + }, + { + "epoch": 0.03126356925749023, + "grad_norm": 0.11343089491128922, + "learning_rate": 0.0009930691199511775, + "loss": 0.993, + "step": 72 + }, + { + "epoch": 0.0316977854971776, + "grad_norm": 0.09531175345182419, + "learning_rate": 0.000992800632147486, + "loss": 1.0179, + "step": 73 + }, + { + "epoch": 0.03213200173686496, + "grad_norm": 0.13536381721496582, + "learning_rate": 0.0009925270800548284, + "loss": 0.9539, + "step": 74 + }, + { + "epoch": 0.03256621797655232, + "grad_norm": 0.09377475082874298, + "learning_rate": 0.0009922484664843763, + "loss": 0.9811, + "step": 75 + }, + { + "epoch": 0.03300043421623969, + "grad_norm": 0.09711369127035141, + "learning_rate": 0.0009919647942993148, + "loss": 0.9223, + "step": 76 + }, + { + "epoch": 0.03343465045592705, + "grad_norm": 0.11353003978729248, + "learning_rate": 0.0009916760664148148, + "loss": 0.9736, + "step": 77 + }, + { + "epoch": 0.03386886669561442, + "grad_norm": 0.09144185483455658, + "learning_rate": 0.0009913822857980019, + "loss": 0.9576, + "step": 78 + }, + { + "epoch": 0.03430308293530178, + "grad_norm": 0.08733920753002167, + "learning_rate": 0.0009910834554679266, + "loss": 0.9016, + "step": 79 + }, + { + "epoch": 0.03473729917498915, + "grad_norm": 0.08889693766832352, + "learning_rate": 0.0009907795784955327, + "loss": 0.8981, + "step": 80 + }, + { + "epoch": 0.03517151541467651, + "grad_norm": 0.081461600959301, + "learning_rate": 0.0009904706580036264, + "loss": 0.9246, + "step": 81 + }, + { + "epoch": 0.03560573165436387, + "grad_norm": 0.08032266050577164, + "learning_rate": 0.0009901566971668436, + "loss": 0.901, + "step": 82 + }, + { + "epoch": 0.03603994789405124, + "grad_norm": 0.08888774365186691, + "learning_rate": 0.0009898376992116178, + "loss": 0.916, + "step": 83 + }, + { + "epoch": 0.0364741641337386, + "grad_norm": 0.08942156285047531, + "learning_rate": 0.0009895136674161465, + "loss": 0.9122, + "step": 84 + }, + { + "epoch": 0.03690838037342597, + "grad_norm": 0.08162346482276917, + "learning_rate": 0.0009891846051103576, + "loss": 0.8893, + "step": 85 + }, + { + "epoch": 0.03734259661311333, + "grad_norm": 0.08491238951683044, + "learning_rate": 0.0009888505156758758, + "loss": 0.8572, + "step": 86 + }, + { + "epoch": 0.0377768128528007, + "grad_norm": 0.07837095111608505, + "learning_rate": 0.0009885114025459864, + "loss": 0.851, + "step": 87 + }, + { + "epoch": 0.03821102909248806, + "grad_norm": 0.08016610145568848, + "learning_rate": 0.0009881672692056021, + "loss": 0.8985, + "step": 88 + }, + { + "epoch": 0.03864524533217542, + "grad_norm": 0.07894952595233917, + "learning_rate": 0.000987818119191225, + "loss": 0.8719, + "step": 89 + }, + { + "epoch": 0.03907946157186279, + "grad_norm": 0.07604355365037918, + "learning_rate": 0.0009874639560909118, + "loss": 0.883, + "step": 90 + }, + { + "epoch": 0.03951367781155015, + "grad_norm": 0.07936527580022812, + "learning_rate": 0.0009871047835442364, + "loss": 0.8385, + "step": 91 + }, + { + "epoch": 0.03994789405123752, + "grad_norm": 0.08259500563144684, + "learning_rate": 0.0009867406052422524, + "loss": 0.8862, + "step": 92 + }, + { + "epoch": 0.04038211029092488, + "grad_norm": 0.09410865604877472, + "learning_rate": 0.0009863714249274552, + "loss": 0.8654, + "step": 93 + }, + { + "epoch": 0.04081632653061224, + "grad_norm": 0.08195285499095917, + "learning_rate": 0.000985997246393744, + "loss": 0.8616, + "step": 94 + }, + { + "epoch": 0.04125054277029961, + "grad_norm": 0.07515832781791687, + "learning_rate": 0.000985618073486382, + "loss": 0.858, + "step": 95 + }, + { + "epoch": 0.04168475900998697, + "grad_norm": 0.07317376881837845, + "learning_rate": 0.0009852339101019574, + "loss": 0.8699, + "step": 96 + }, + { + "epoch": 0.04211897524967434, + "grad_norm": 0.088593028485775, + "learning_rate": 0.0009848447601883434, + "loss": 0.8326, + "step": 97 + }, + { + "epoch": 0.0425531914893617, + "grad_norm": 0.07599225640296936, + "learning_rate": 0.0009844506277446577, + "loss": 0.7984, + "step": 98 + }, + { + "epoch": 0.04298740772904907, + "grad_norm": 0.08955889940261841, + "learning_rate": 0.0009840515168212207, + "loss": 0.8489, + "step": 99 + }, + { + "epoch": 0.04342162396873643, + "grad_norm": 0.07896170765161514, + "learning_rate": 0.0009836474315195148, + "loss": 0.8133, + "step": 100 + }, + { + "epoch": 0.04385584020842379, + "grad_norm": 0.08088196814060211, + "learning_rate": 0.0009832383759921415, + "loss": 0.8356, + "step": 101 + }, + { + "epoch": 0.04429005644811116, + "grad_norm": 0.08699057996273041, + "learning_rate": 0.0009828243544427796, + "loss": 0.8226, + "step": 102 + }, + { + "epoch": 0.04472427268779852, + "grad_norm": 0.08965849131345749, + "learning_rate": 0.0009824053711261404, + "loss": 0.8137, + "step": 103 + }, + { + "epoch": 0.04515848892748589, + "grad_norm": 0.09515663236379623, + "learning_rate": 0.0009819814303479266, + "loss": 0.7937, + "step": 104 + }, + { + "epoch": 0.04559270516717325, + "grad_norm": 0.06896204501390457, + "learning_rate": 0.0009815525364647853, + "loss": 0.8056, + "step": 105 + }, + { + "epoch": 0.04602692140686062, + "grad_norm": 0.08081001043319702, + "learning_rate": 0.0009811186938842645, + "loss": 0.8465, + "step": 106 + }, + { + "epoch": 0.04646113764654798, + "grad_norm": 0.07806787639856339, + "learning_rate": 0.000980679907064768, + "loss": 0.7897, + "step": 107 + }, + { + "epoch": 0.04689535388623534, + "grad_norm": 0.07646839320659637, + "learning_rate": 0.0009802361805155097, + "loss": 0.7929, + "step": 108 + }, + { + "epoch": 0.04732957012592271, + "grad_norm": 0.08139129728078842, + "learning_rate": 0.000979787518796466, + "loss": 0.8097, + "step": 109 + }, + { + "epoch": 0.04776378636561007, + "grad_norm": 0.07595842331647873, + "learning_rate": 0.0009793339265183304, + "loss": 0.8028, + "step": 110 + }, + { + "epoch": 0.04819800260529744, + "grad_norm": 0.07882706820964813, + "learning_rate": 0.0009788754083424652, + "loss": 0.7825, + "step": 111 + }, + { + "epoch": 0.0486322188449848, + "grad_norm": 0.09278842061758041, + "learning_rate": 0.0009784119689808543, + "loss": 0.7897, + "step": 112 + }, + { + "epoch": 0.049066435084672164, + "grad_norm": 0.07860872894525528, + "learning_rate": 0.0009779436131960543, + "loss": 0.8193, + "step": 113 + }, + { + "epoch": 0.04950065132435953, + "grad_norm": 0.09767181426286697, + "learning_rate": 0.0009774703458011453, + "loss": 0.7653, + "step": 114 + }, + { + "epoch": 0.04993486756404689, + "grad_norm": 0.09374384582042694, + "learning_rate": 0.000976992171659682, + "loss": 0.7541, + "step": 115 + }, + { + "epoch": 0.05036908380373426, + "grad_norm": 0.07632128894329071, + "learning_rate": 0.0009765090956856436, + "loss": 0.7518, + "step": 116 + }, + { + "epoch": 0.05080330004342162, + "grad_norm": 0.07544126361608505, + "learning_rate": 0.0009760211228433832, + "loss": 0.8217, + "step": 117 + }, + { + "epoch": 0.05123751628310899, + "grad_norm": 0.07379148155450821, + "learning_rate": 0.0009755282581475768, + "loss": 0.7902, + "step": 118 + }, + { + "epoch": 0.05167173252279635, + "grad_norm": 0.07656335830688477, + "learning_rate": 0.0009750305066631716, + "loss": 0.7357, + "step": 119 + }, + { + "epoch": 0.052105948762483714, + "grad_norm": 0.07699162513017654, + "learning_rate": 0.0009745278735053344, + "loss": 0.7329, + "step": 120 + }, + { + "epoch": 0.05254016500217108, + "grad_norm": 0.07921908050775528, + "learning_rate": 0.0009740203638393983, + "loss": 0.7368, + "step": 121 + }, + { + "epoch": 0.052974381241858444, + "grad_norm": 0.06974276900291443, + "learning_rate": 0.0009735079828808107, + "loss": 0.7593, + "step": 122 + }, + { + "epoch": 0.05340859748154581, + "grad_norm": 0.06961283832788467, + "learning_rate": 0.0009729907358950785, + "loss": 0.7899, + "step": 123 + }, + { + "epoch": 0.053842813721233174, + "grad_norm": 0.07279238104820251, + "learning_rate": 0.0009724686281977146, + "loss": 0.7455, + "step": 124 + }, + { + "epoch": 0.054277029960920535, + "grad_norm": 0.07039442658424377, + "learning_rate": 0.0009719416651541838, + "loss": 0.7755, + "step": 125 + }, + { + "epoch": 0.0547112462006079, + "grad_norm": 0.07618307322263718, + "learning_rate": 0.0009714098521798464, + "loss": 0.7631, + "step": 126 + }, + { + "epoch": 0.055145462440295265, + "grad_norm": 0.06525314599275589, + "learning_rate": 0.0009708731947399038, + "loss": 0.7605, + "step": 127 + }, + { + "epoch": 0.05557967867998263, + "grad_norm": 0.07441145926713943, + "learning_rate": 0.0009703316983493413, + "loss": 0.7402, + "step": 128 + }, + { + "epoch": 0.056013894919669995, + "grad_norm": 0.06484754383563995, + "learning_rate": 0.0009697853685728721, + "loss": 0.7238, + "step": 129 + }, + { + "epoch": 0.05644811115935736, + "grad_norm": 0.07414862513542175, + "learning_rate": 0.0009692342110248802, + "loss": 0.7633, + "step": 130 + }, + { + "epoch": 0.056882327399044724, + "grad_norm": 0.07470294088125229, + "learning_rate": 0.0009686782313693621, + "loss": 0.7467, + "step": 131 + }, + { + "epoch": 0.057316543638732086, + "grad_norm": 0.07042668759822845, + "learning_rate": 0.0009681174353198686, + "loss": 0.7309, + "step": 132 + }, + { + "epoch": 0.057750759878419454, + "grad_norm": 0.07646633684635162, + "learning_rate": 0.0009675518286394472, + "loss": 0.722, + "step": 133 + }, + { + "epoch": 0.058184976118106815, + "grad_norm": 0.07386059314012527, + "learning_rate": 0.0009669814171405816, + "loss": 0.7028, + "step": 134 + }, + { + "epoch": 0.058619192357794184, + "grad_norm": 0.07046350091695786, + "learning_rate": 0.0009664062066851325, + "loss": 0.711, + "step": 135 + }, + { + "epoch": 0.059053408597481545, + "grad_norm": 0.07790421694517136, + "learning_rate": 0.000965826203184277, + "loss": 0.7183, + "step": 136 + }, + { + "epoch": 0.05948762483716891, + "grad_norm": 0.08006292581558228, + "learning_rate": 0.000965241412598449, + "loss": 0.7534, + "step": 137 + }, + { + "epoch": 0.059921841076856275, + "grad_norm": 0.06924112141132355, + "learning_rate": 0.0009646518409372759, + "loss": 0.7401, + "step": 138 + }, + { + "epoch": 0.060356057316543636, + "grad_norm": 0.07095640152692795, + "learning_rate": 0.0009640574942595195, + "loss": 0.7056, + "step": 139 + }, + { + "epoch": 0.060790273556231005, + "grad_norm": 0.0727936401963234, + "learning_rate": 0.0009634583786730109, + "loss": 0.6811, + "step": 140 + }, + { + "epoch": 0.061224489795918366, + "grad_norm": 0.0759989321231842, + "learning_rate": 0.0009628545003345899, + "loss": 0.719, + "step": 141 + }, + { + "epoch": 0.061658706035605734, + "grad_norm": 0.07433141767978668, + "learning_rate": 0.0009622458654500408, + "loss": 0.724, + "step": 142 + }, + { + "epoch": 0.062092922275293096, + "grad_norm": 0.06923756748437881, + "learning_rate": 0.0009616324802740286, + "loss": 0.7562, + "step": 143 + }, + { + "epoch": 0.06252713851498046, + "grad_norm": 0.10080388188362122, + "learning_rate": 0.0009610143511100353, + "loss": 0.7116, + "step": 144 + }, + { + "epoch": 0.06296135475466783, + "grad_norm": 0.07774697989225388, + "learning_rate": 0.0009603914843102941, + "loss": 0.7277, + "step": 145 + }, + { + "epoch": 0.0633955709943552, + "grad_norm": 0.07297523319721222, + "learning_rate": 0.0009597638862757254, + "loss": 0.6817, + "step": 146 + }, + { + "epoch": 0.06382978723404255, + "grad_norm": 0.07876230776309967, + "learning_rate": 0.0009591315634558697, + "loss": 0.7015, + "step": 147 + }, + { + "epoch": 0.06426400347372992, + "grad_norm": 0.07196466624736786, + "learning_rate": 0.0009584945223488226, + "loss": 0.6995, + "step": 148 + }, + { + "epoch": 0.06469821971341728, + "grad_norm": 0.07307913899421692, + "learning_rate": 0.0009578527695011669, + "loss": 0.6931, + "step": 149 + }, + { + "epoch": 0.06513243595310464, + "grad_norm": 0.0770203247666359, + "learning_rate": 0.0009572063115079062, + "loss": 0.7135, + "step": 150 + }, + { + "epoch": 0.06556665219279201, + "grad_norm": 0.06902311742305756, + "learning_rate": 0.0009565551550123967, + "loss": 0.6906, + "step": 151 + }, + { + "epoch": 0.06600086843247938, + "grad_norm": 0.10059111565351486, + "learning_rate": 0.0009558993067062785, + "loss": 0.7237, + "step": 152 + }, + { + "epoch": 0.06643508467216674, + "grad_norm": 0.07180771976709366, + "learning_rate": 0.000955238773329408, + "loss": 0.6929, + "step": 153 + }, + { + "epoch": 0.0668693009118541, + "grad_norm": 0.07852017134428024, + "learning_rate": 0.0009545735616697875, + "loss": 0.6531, + "step": 154 + }, + { + "epoch": 0.06730351715154147, + "grad_norm": 0.07011571526527405, + "learning_rate": 0.0009539036785634961, + "loss": 0.6949, + "step": 155 + }, + { + "epoch": 0.06773773339122884, + "grad_norm": 0.07384242117404938, + "learning_rate": 0.000953229130894619, + "loss": 0.688, + "step": 156 + }, + { + "epoch": 0.06817194963091619, + "grad_norm": 0.07481442391872406, + "learning_rate": 0.0009525499255951775, + "loss": 0.6969, + "step": 157 + }, + { + "epoch": 0.06860616587060356, + "grad_norm": 0.06441578269004822, + "learning_rate": 0.0009518660696450568, + "loss": 0.6478, + "step": 158 + }, + { + "epoch": 0.06904038211029093, + "grad_norm": 0.06720245629549026, + "learning_rate": 0.0009511775700719346, + "loss": 0.677, + "step": 159 + }, + { + "epoch": 0.0694745983499783, + "grad_norm": 0.08833543956279755, + "learning_rate": 0.0009504844339512095, + "loss": 0.6912, + "step": 160 + }, + { + "epoch": 0.06990881458966565, + "grad_norm": 0.08247049152851105, + "learning_rate": 0.0009497866684059277, + "loss": 0.6833, + "step": 161 + }, + { + "epoch": 0.07034303082935302, + "grad_norm": 0.06711483001708984, + "learning_rate": 0.0009490842806067095, + "loss": 0.6695, + "step": 162 + }, + { + "epoch": 0.07077724706904039, + "grad_norm": 0.06512755900621414, + "learning_rate": 0.0009483772777716766, + "loss": 0.695, + "step": 163 + }, + { + "epoch": 0.07121146330872774, + "grad_norm": 0.06651882827281952, + "learning_rate": 0.0009476656671663766, + "loss": 0.6608, + "step": 164 + }, + { + "epoch": 0.07164567954841511, + "grad_norm": 0.06416794657707214, + "learning_rate": 0.0009469494561037098, + "loss": 0.6918, + "step": 165 + }, + { + "epoch": 0.07207989578810248, + "grad_norm": 0.06808705627918243, + "learning_rate": 0.000946228651943853, + "loss": 0.6959, + "step": 166 + }, + { + "epoch": 0.07251411202778985, + "grad_norm": 0.077076256275177, + "learning_rate": 0.0009455032620941839, + "loss": 0.6737, + "step": 167 + }, + { + "epoch": 0.0729483282674772, + "grad_norm": 0.09289199858903885, + "learning_rate": 0.000944773294009206, + "loss": 0.651, + "step": 168 + }, + { + "epoch": 0.07338254450716457, + "grad_norm": 0.07365357875823975, + "learning_rate": 0.0009440387551904703, + "loss": 0.6778, + "step": 169 + }, + { + "epoch": 0.07381676074685194, + "grad_norm": 0.0635438859462738, + "learning_rate": 0.0009432996531865001, + "loss": 0.7081, + "step": 170 + }, + { + "epoch": 0.07425097698653929, + "grad_norm": 0.07002374529838562, + "learning_rate": 0.0009425559955927117, + "loss": 0.6505, + "step": 171 + }, + { + "epoch": 0.07468519322622666, + "grad_norm": 0.08253403007984161, + "learning_rate": 0.0009418077900513376, + "loss": 0.6752, + "step": 172 + }, + { + "epoch": 0.07511940946591403, + "grad_norm": 0.06753169745206833, + "learning_rate": 0.0009410550442513475, + "loss": 0.6695, + "step": 173 + }, + { + "epoch": 0.0755536257056014, + "grad_norm": 0.06907963007688522, + "learning_rate": 0.000940297765928369, + "loss": 0.6441, + "step": 174 + }, + { + "epoch": 0.07598784194528875, + "grad_norm": 0.0881882980465889, + "learning_rate": 0.0009395359628646086, + "loss": 0.6456, + "step": 175 + }, + { + "epoch": 0.07642205818497612, + "grad_norm": 0.06529023498296738, + "learning_rate": 0.0009387696428887715, + "loss": 0.6689, + "step": 176 + }, + { + "epoch": 0.07685627442466349, + "grad_norm": 0.06664346903562546, + "learning_rate": 0.0009379988138759809, + "loss": 0.6485, + "step": 177 + }, + { + "epoch": 0.07729049066435084, + "grad_norm": 0.06951016932725906, + "learning_rate": 0.0009372234837476979, + "loss": 0.68, + "step": 178 + }, + { + "epoch": 0.07772470690403821, + "grad_norm": 0.060999054461717606, + "learning_rate": 0.0009364436604716389, + "loss": 0.6706, + "step": 179 + }, + { + "epoch": 0.07815892314372558, + "grad_norm": 0.07197262346744537, + "learning_rate": 0.0009356593520616947, + "loss": 0.6277, + "step": 180 + }, + { + "epoch": 0.07859313938341293, + "grad_norm": 0.06676187366247177, + "learning_rate": 0.0009348705665778478, + "loss": 0.6492, + "step": 181 + }, + { + "epoch": 0.0790273556231003, + "grad_norm": 0.07378092408180237, + "learning_rate": 0.0009340773121260893, + "loss": 0.6375, + "step": 182 + }, + { + "epoch": 0.07946157186278767, + "grad_norm": 0.06367610394954681, + "learning_rate": 0.000933279596858336, + "loss": 0.6221, + "step": 183 + }, + { + "epoch": 0.07989578810247504, + "grad_norm": 0.0669325664639473, + "learning_rate": 0.0009324774289723468, + "loss": 0.6294, + "step": 184 + }, + { + "epoch": 0.08033000434216239, + "grad_norm": 0.06983687728643417, + "learning_rate": 0.0009316708167116377, + "loss": 0.6487, + "step": 185 + }, + { + "epoch": 0.08076422058184976, + "grad_norm": 0.06569714099168777, + "learning_rate": 0.0009308597683653976, + "loss": 0.651, + "step": 186 + }, + { + "epoch": 0.08119843682153713, + "grad_norm": 0.06825820356607437, + "learning_rate": 0.0009300442922684032, + "loss": 0.6193, + "step": 187 + }, + { + "epoch": 0.08163265306122448, + "grad_norm": 0.07318625599145889, + "learning_rate": 0.000929224396800933, + "loss": 0.7024, + "step": 188 + }, + { + "epoch": 0.08206686930091185, + "grad_norm": 0.1047971323132515, + "learning_rate": 0.0009284000903886818, + "loss": 0.6637, + "step": 189 + }, + { + "epoch": 0.08250108554059922, + "grad_norm": 0.06583527475595474, + "learning_rate": 0.0009275713815026732, + "loss": 0.6419, + "step": 190 + }, + { + "epoch": 0.08293530178028659, + "grad_norm": 0.09143144637346268, + "learning_rate": 0.000926738278659173, + "loss": 0.6295, + "step": 191 + }, + { + "epoch": 0.08336951801997394, + "grad_norm": 0.06756502389907837, + "learning_rate": 0.0009259007904196022, + "loss": 0.6718, + "step": 192 + }, + { + "epoch": 0.08380373425966131, + "grad_norm": 0.06964406371116638, + "learning_rate": 0.000925058925390448, + "loss": 0.6542, + "step": 193 + }, + { + "epoch": 0.08423795049934868, + "grad_norm": 0.07476377487182617, + "learning_rate": 0.0009242126922231762, + "loss": 0.6266, + "step": 194 + }, + { + "epoch": 0.08467216673903603, + "grad_norm": 0.07210344076156616, + "learning_rate": 0.0009233620996141421, + "loss": 0.6721, + "step": 195 + }, + { + "epoch": 0.0851063829787234, + "grad_norm": 0.07828934490680695, + "learning_rate": 0.0009225071563045006, + "loss": 0.6223, + "step": 196 + }, + { + "epoch": 0.08554059921841077, + "grad_norm": 0.07536309212446213, + "learning_rate": 0.000921647871080117, + "loss": 0.6563, + "step": 197 + }, + { + "epoch": 0.08597481545809814, + "grad_norm": 0.07678169757127762, + "learning_rate": 0.0009207842527714766, + "loss": 0.6646, + "step": 198 + }, + { + "epoch": 0.0864090316977855, + "grad_norm": 0.07808982580900192, + "learning_rate": 0.0009199163102535937, + "loss": 0.6497, + "step": 199 + }, + { + "epoch": 0.08684324793747286, + "grad_norm": 0.07775306701660156, + "learning_rate": 0.0009190440524459203, + "loss": 0.6083, + "step": 200 + }, + { + "epoch": 0.08727746417716023, + "grad_norm": 0.06569099426269531, + "learning_rate": 0.0009181674883122553, + "loss": 0.6424, + "step": 201 + }, + { + "epoch": 0.08771168041684758, + "grad_norm": 0.08122856914997101, + "learning_rate": 0.0009172866268606513, + "loss": 0.6054, + "step": 202 + }, + { + "epoch": 0.08814589665653495, + "grad_norm": 0.08045452833175659, + "learning_rate": 0.0009164014771433226, + "loss": 0.6372, + "step": 203 + }, + { + "epoch": 0.08858011289622232, + "grad_norm": 0.0744997188448906, + "learning_rate": 0.000915512048256552, + "loss": 0.6341, + "step": 204 + }, + { + "epoch": 0.08901432913590969, + "grad_norm": 0.07247951626777649, + "learning_rate": 0.0009146183493405975, + "loss": 0.5922, + "step": 205 + }, + { + "epoch": 0.08944854537559704, + "grad_norm": 0.08652166277170181, + "learning_rate": 0.0009137203895795982, + "loss": 0.6111, + "step": 206 + }, + { + "epoch": 0.08988276161528441, + "grad_norm": 0.06636743992567062, + "learning_rate": 0.0009128181782014801, + "loss": 0.6287, + "step": 207 + }, + { + "epoch": 0.09031697785497178, + "grad_norm": 0.08643815666437149, + "learning_rate": 0.0009119117244778608, + "loss": 0.6091, + "step": 208 + }, + { + "epoch": 0.09075119409465914, + "grad_norm": 0.08439099788665771, + "learning_rate": 0.0009110010377239551, + "loss": 0.6735, + "step": 209 + }, + { + "epoch": 0.0911854103343465, + "grad_norm": 0.0863455981016159, + "learning_rate": 0.0009100861272984779, + "loss": 0.6049, + "step": 210 + }, + { + "epoch": 0.09161962657403387, + "grad_norm": 0.06976283341646194, + "learning_rate": 0.0009091670026035499, + "loss": 0.5934, + "step": 211 + }, + { + "epoch": 0.09205384281372124, + "grad_norm": 0.10158892720937729, + "learning_rate": 0.0009082436730845993, + "loss": 0.6314, + "step": 212 + }, + { + "epoch": 0.0924880590534086, + "grad_norm": 0.06954232603311539, + "learning_rate": 0.0009073161482302654, + "loss": 0.6461, + "step": 213 + }, + { + "epoch": 0.09292227529309596, + "grad_norm": 0.06838048249483109, + "learning_rate": 0.0009063844375723014, + "loss": 0.5993, + "step": 214 + }, + { + "epoch": 0.09335649153278333, + "grad_norm": 0.06874913722276688, + "learning_rate": 0.0009054485506854755, + "loss": 0.6035, + "step": 215 + }, + { + "epoch": 0.09379070777247069, + "grad_norm": 0.07478207349777222, + "learning_rate": 0.0009045084971874737, + "loss": 0.6123, + "step": 216 + }, + { + "epoch": 0.09422492401215805, + "grad_norm": 0.06408224254846573, + "learning_rate": 0.0009035642867388002, + "loss": 0.6242, + "step": 217 + }, + { + "epoch": 0.09465914025184542, + "grad_norm": 0.07433444261550903, + "learning_rate": 0.0009026159290426781, + "loss": 0.6055, + "step": 218 + }, + { + "epoch": 0.09509335649153278, + "grad_norm": 0.06954223662614822, + "learning_rate": 0.0009016634338449503, + "loss": 0.6257, + "step": 219 + }, + { + "epoch": 0.09552757273122015, + "grad_norm": 0.06509328633546829, + "learning_rate": 0.0009007068109339783, + "loss": 0.5749, + "step": 220 + }, + { + "epoch": 0.09596178897090751, + "grad_norm": 0.08184398710727692, + "learning_rate": 0.000899746070140543, + "loss": 0.6166, + "step": 221 + }, + { + "epoch": 0.09639600521059488, + "grad_norm": 0.06200498715043068, + "learning_rate": 0.0008987812213377423, + "loss": 0.615, + "step": 222 + }, + { + "epoch": 0.09683022145028224, + "grad_norm": 0.06602746993303299, + "learning_rate": 0.0008978122744408905, + "loss": 0.6087, + "step": 223 + }, + { + "epoch": 0.0972644376899696, + "grad_norm": 0.0869906097650528, + "learning_rate": 0.0008968392394074163, + "loss": 0.5725, + "step": 224 + }, + { + "epoch": 0.09769865392965697, + "grad_norm": 0.064040407538414, + "learning_rate": 0.0008958621262367599, + "loss": 0.6116, + "step": 225 + }, + { + "epoch": 0.09813287016934433, + "grad_norm": 0.0657719075679779, + "learning_rate": 0.0008948809449702712, + "loss": 0.6219, + "step": 226 + }, + { + "epoch": 0.0985670864090317, + "grad_norm": 0.07608803361654282, + "learning_rate": 0.0008938957056911057, + "loss": 0.6018, + "step": 227 + }, + { + "epoch": 0.09900130264871906, + "grad_norm": 0.07249824702739716, + "learning_rate": 0.0008929064185241212, + "loss": 0.6027, + "step": 228 + }, + { + "epoch": 0.09943551888840643, + "grad_norm": 0.06748249381780624, + "learning_rate": 0.0008919130936357742, + "loss": 0.6032, + "step": 229 + }, + { + "epoch": 0.09986973512809379, + "grad_norm": 0.08020893484354019, + "learning_rate": 0.000890915741234015, + "loss": 0.607, + "step": 230 + }, + { + "epoch": 0.10030395136778116, + "grad_norm": 0.0669545903801918, + "learning_rate": 0.0008899143715681822, + "loss": 0.5974, + "step": 231 + }, + { + "epoch": 0.10073816760746852, + "grad_norm": 0.06186607852578163, + "learning_rate": 0.0008889089949288987, + "loss": 0.6186, + "step": 232 + }, + { + "epoch": 0.10117238384715588, + "grad_norm": 0.06386630982160568, + "learning_rate": 0.0008878996216479651, + "loss": 0.5995, + "step": 233 + }, + { + "epoch": 0.10160660008684325, + "grad_norm": 0.06557858735322952, + "learning_rate": 0.0008868862620982534, + "loss": 0.6066, + "step": 234 + }, + { + "epoch": 0.10204081632653061, + "grad_norm": 0.06579785794019699, + "learning_rate": 0.0008858689266936008, + "loss": 0.6144, + "step": 235 + }, + { + "epoch": 0.10247503256621798, + "grad_norm": 0.06238356605172157, + "learning_rate": 0.0008848476258887031, + "loss": 0.5899, + "step": 236 + }, + { + "epoch": 0.10290924880590534, + "grad_norm": 0.061628226190805435, + "learning_rate": 0.0008838223701790055, + "loss": 0.6092, + "step": 237 + }, + { + "epoch": 0.1033434650455927, + "grad_norm": 0.06545189768075943, + "learning_rate": 0.0008827931701005973, + "loss": 0.6014, + "step": 238 + }, + { + "epoch": 0.10377768128528007, + "grad_norm": 0.056564707309007645, + "learning_rate": 0.0008817600362301017, + "loss": 0.5959, + "step": 239 + }, + { + "epoch": 0.10421189752496743, + "grad_norm": 0.0715203583240509, + "learning_rate": 0.0008807229791845672, + "loss": 0.5942, + "step": 240 + }, + { + "epoch": 0.1046461137646548, + "grad_norm": 0.06504496932029724, + "learning_rate": 0.00087968200962136, + "loss": 0.5652, + "step": 241 + }, + { + "epoch": 0.10508033000434217, + "grad_norm": 0.06740335375070572, + "learning_rate": 0.0008786371382380527, + "loss": 0.6159, + "step": 242 + }, + { + "epoch": 0.10551454624402953, + "grad_norm": 0.06163683906197548, + "learning_rate": 0.0008775883757723155, + "loss": 0.5913, + "step": 243 + }, + { + "epoch": 0.10594876248371689, + "grad_norm": 0.07384572178125381, + "learning_rate": 0.0008765357330018055, + "loss": 0.5829, + "step": 244 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 0.06352236866950989, + "learning_rate": 0.0008754792207440556, + "loss": 0.6069, + "step": 245 + }, + { + "epoch": 0.10681719496309162, + "grad_norm": 0.0605890229344368, + "learning_rate": 0.0008744188498563641, + "loss": 0.5685, + "step": 246 + }, + { + "epoch": 0.10725141120277898, + "grad_norm": 0.06994660943746567, + "learning_rate": 0.0008733546312356824, + "loss": 0.581, + "step": 247 + }, + { + "epoch": 0.10768562744246635, + "grad_norm": 0.08128967881202698, + "learning_rate": 0.0008722865758185036, + "loss": 0.5754, + "step": 248 + }, + { + "epoch": 0.10811984368215372, + "grad_norm": 0.06497329473495483, + "learning_rate": 0.0008712146945807493, + "loss": 0.598, + "step": 249 + }, + { + "epoch": 0.10855405992184107, + "grad_norm": 0.06091492623090744, + "learning_rate": 0.0008701389985376577, + "loss": 0.5807, + "step": 250 + }, + { + "epoch": 0.10898827616152844, + "grad_norm": 0.06555697321891785, + "learning_rate": 0.0008690594987436704, + "loss": 0.5768, + "step": 251 + }, + { + "epoch": 0.1094224924012158, + "grad_norm": 0.07119280844926834, + "learning_rate": 0.0008679762062923176, + "loss": 0.6088, + "step": 252 + }, + { + "epoch": 0.10985670864090318, + "grad_norm": 0.06837710738182068, + "learning_rate": 0.0008668891323161053, + "loss": 0.5905, + "step": 253 + }, + { + "epoch": 0.11029092488059053, + "grad_norm": 0.067451111972332, + "learning_rate": 0.0008657982879864007, + "loss": 0.5974, + "step": 254 + }, + { + "epoch": 0.1107251411202779, + "grad_norm": 0.06649015098810196, + "learning_rate": 0.0008647036845133172, + "loss": 0.5684, + "step": 255 + }, + { + "epoch": 0.11115935735996527, + "grad_norm": 0.06311678886413574, + "learning_rate": 0.0008636053331455986, + "loss": 0.58, + "step": 256 + }, + { + "epoch": 0.11159357359965262, + "grad_norm": 0.06087026745080948, + "learning_rate": 0.0008625032451705052, + "loss": 0.5735, + "step": 257 + }, + { + "epoch": 0.11202778983933999, + "grad_norm": 0.061150964349508286, + "learning_rate": 0.0008613974319136957, + "loss": 0.6044, + "step": 258 + }, + { + "epoch": 0.11246200607902736, + "grad_norm": 0.06023947522044182, + "learning_rate": 0.0008602879047391126, + "loss": 0.551, + "step": 259 + }, + { + "epoch": 0.11289622231871473, + "grad_norm": 0.06919502466917038, + "learning_rate": 0.0008591746750488638, + "loss": 0.5744, + "step": 260 + }, + { + "epoch": 0.11333043855840208, + "grad_norm": 0.06585050374269485, + "learning_rate": 0.0008580577542831072, + "loss": 0.6053, + "step": 261 + }, + { + "epoch": 0.11376465479808945, + "grad_norm": 0.06967171281576157, + "learning_rate": 0.0008569371539199316, + "loss": 0.5896, + "step": 262 + }, + { + "epoch": 0.11419887103777682, + "grad_norm": 0.062438324093818665, + "learning_rate": 0.0008558128854752396, + "loss": 0.585, + "step": 263 + }, + { + "epoch": 0.11463308727746417, + "grad_norm": 0.06600722670555115, + "learning_rate": 0.0008546849605026289, + "loss": 0.5809, + "step": 264 + }, + { + "epoch": 0.11506730351715154, + "grad_norm": 0.06594278663396835, + "learning_rate": 0.0008535533905932737, + "loss": 0.5854, + "step": 265 + }, + { + "epoch": 0.11550151975683891, + "grad_norm": 0.06402723491191864, + "learning_rate": 0.0008524181873758059, + "loss": 0.5612, + "step": 266 + }, + { + "epoch": 0.11593573599652628, + "grad_norm": 0.06593722105026245, + "learning_rate": 0.0008512793625161946, + "loss": 0.5825, + "step": 267 + }, + { + "epoch": 0.11636995223621363, + "grad_norm": 0.057960718870162964, + "learning_rate": 0.0008501369277176275, + "loss": 0.5972, + "step": 268 + }, + { + "epoch": 0.116804168475901, + "grad_norm": 0.06126386299729347, + "learning_rate": 0.0008489908947203897, + "loss": 0.5675, + "step": 269 + }, + { + "epoch": 0.11723838471558837, + "grad_norm": 0.06619458645582199, + "learning_rate": 0.0008478412753017432, + "loss": 0.5724, + "step": 270 + }, + { + "epoch": 0.11767260095527572, + "grad_norm": 0.05950764939188957, + "learning_rate": 0.0008466880812758065, + "loss": 0.5732, + "step": 271 + }, + { + "epoch": 0.11810681719496309, + "grad_norm": 0.0596248060464859, + "learning_rate": 0.0008455313244934324, + "loss": 0.5844, + "step": 272 + }, + { + "epoch": 0.11854103343465046, + "grad_norm": 0.06557915359735489, + "learning_rate": 0.0008443710168420865, + "loss": 0.5706, + "step": 273 + }, + { + "epoch": 0.11897524967433783, + "grad_norm": 0.06171106547117233, + "learning_rate": 0.0008432071702457253, + "loss": 0.561, + "step": 274 + }, + { + "epoch": 0.11940946591402518, + "grad_norm": 0.06009051948785782, + "learning_rate": 0.0008420397966646731, + "loss": 0.5764, + "step": 275 + }, + { + "epoch": 0.11984368215371255, + "grad_norm": 0.06843849271535873, + "learning_rate": 0.0008408689080954998, + "loss": 0.5833, + "step": 276 + }, + { + "epoch": 0.12027789839339992, + "grad_norm": 0.06061152368783951, + "learning_rate": 0.0008396945165708972, + "loss": 0.5962, + "step": 277 + }, + { + "epoch": 0.12071211463308727, + "grad_norm": 0.05997680127620697, + "learning_rate": 0.0008385166341595548, + "loss": 0.5738, + "step": 278 + }, + { + "epoch": 0.12114633087277464, + "grad_norm": 0.06045431271195412, + "learning_rate": 0.0008373352729660373, + "loss": 0.5738, + "step": 279 + }, + { + "epoch": 0.12158054711246201, + "grad_norm": 0.06007954850792885, + "learning_rate": 0.0008361504451306584, + "loss": 0.5753, + "step": 280 + }, + { + "epoch": 0.12201476335214936, + "grad_norm": 0.05806737020611763, + "learning_rate": 0.0008349621628293577, + "loss": 0.5703, + "step": 281 + }, + { + "epoch": 0.12244897959183673, + "grad_norm": 0.06177588179707527, + "learning_rate": 0.0008337704382735741, + "loss": 0.5533, + "step": 282 + }, + { + "epoch": 0.1228831958315241, + "grad_norm": 0.06487903743982315, + "learning_rate": 0.0008325752837101213, + "loss": 0.5655, + "step": 283 + }, + { + "epoch": 0.12331741207121147, + "grad_norm": 0.06531866639852524, + "learning_rate": 0.0008313767114210615, + "loss": 0.5694, + "step": 284 + }, + { + "epoch": 0.12375162831089882, + "grad_norm": 0.08110279589891434, + "learning_rate": 0.0008301747337235797, + "loss": 0.5517, + "step": 285 + }, + { + "epoch": 0.12418584455058619, + "grad_norm": 0.06367824971675873, + "learning_rate": 0.0008289693629698564, + "loss": 0.5474, + "step": 286 + }, + { + "epoch": 0.12462006079027356, + "grad_norm": 0.07917732000350952, + "learning_rate": 0.0008277606115469409, + "loss": 0.5685, + "step": 287 + }, + { + "epoch": 0.12505427702996091, + "grad_norm": 0.06585974991321564, + "learning_rate": 0.0008265484918766243, + "loss": 0.5547, + "step": 288 + }, + { + "epoch": 0.12548849326964828, + "grad_norm": 0.0627543181180954, + "learning_rate": 0.0008253330164153117, + "loss": 0.5795, + "step": 289 + }, + { + "epoch": 0.12592270950933565, + "grad_norm": 0.06639768183231354, + "learning_rate": 0.0008241141976538943, + "loss": 0.5669, + "step": 290 + }, + { + "epoch": 0.12635692574902302, + "grad_norm": 0.06217719987034798, + "learning_rate": 0.0008228920481176202, + "loss": 0.5552, + "step": 291 + }, + { + "epoch": 0.1267911419887104, + "grad_norm": 0.06806106120347977, + "learning_rate": 0.0008216665803659671, + "loss": 0.5553, + "step": 292 + }, + { + "epoch": 0.12722535822839776, + "grad_norm": 0.058773159980773926, + "learning_rate": 0.000820437806992512, + "loss": 0.5678, + "step": 293 + }, + { + "epoch": 0.1276595744680851, + "grad_norm": 0.06487017124891281, + "learning_rate": 0.0008192057406248028, + "loss": 0.5493, + "step": 294 + }, + { + "epoch": 0.12809379070777246, + "grad_norm": 0.05514706298708916, + "learning_rate": 0.0008179703939242276, + "loss": 0.5834, + "step": 295 + }, + { + "epoch": 0.12852800694745983, + "grad_norm": 0.05893951281905174, + "learning_rate": 0.0008167317795858851, + "loss": 0.5668, + "step": 296 + }, + { + "epoch": 0.1289622231871472, + "grad_norm": 0.06080016866326332, + "learning_rate": 0.0008154899103384537, + "loss": 0.6112, + "step": 297 + }, + { + "epoch": 0.12939643942683457, + "grad_norm": 0.06676195561885834, + "learning_rate": 0.0008142447989440618, + "loss": 0.5897, + "step": 298 + }, + { + "epoch": 0.12983065566652194, + "grad_norm": 0.055440496653318405, + "learning_rate": 0.0008129964581981553, + "loss": 0.5199, + "step": 299 + }, + { + "epoch": 0.13026487190620928, + "grad_norm": 0.0642077848315239, + "learning_rate": 0.0008117449009293668, + "loss": 0.5688, + "step": 300 + }, + { + "epoch": 0.13069908814589665, + "grad_norm": 0.056085068732500076, + "learning_rate": 0.0008104901399993836, + "loss": 0.5626, + "step": 301 + }, + { + "epoch": 0.13113330438558402, + "grad_norm": 0.07240674644708633, + "learning_rate": 0.0008092321883028157, + "loss": 0.5705, + "step": 302 + }, + { + "epoch": 0.13156752062527138, + "grad_norm": 0.06402257084846497, + "learning_rate": 0.0008079710587670633, + "loss": 0.5608, + "step": 303 + }, + { + "epoch": 0.13200173686495875, + "grad_norm": 0.05889822542667389, + "learning_rate": 0.0008067067643521834, + "loss": 0.5552, + "step": 304 + }, + { + "epoch": 0.13243595310464612, + "grad_norm": 0.07787332683801651, + "learning_rate": 0.0008054393180507572, + "loss": 0.5526, + "step": 305 + }, + { + "epoch": 0.1328701693443335, + "grad_norm": 0.06866924464702606, + "learning_rate": 0.0008041687328877566, + "loss": 0.5726, + "step": 306 + }, + { + "epoch": 0.13330438558402083, + "grad_norm": 0.06494660675525665, + "learning_rate": 0.00080289502192041, + "loss": 0.5605, + "step": 307 + }, + { + "epoch": 0.1337386018237082, + "grad_norm": 0.06333691626787186, + "learning_rate": 0.0008016181982380681, + "loss": 0.5668, + "step": 308 + }, + { + "epoch": 0.13417281806339557, + "grad_norm": 0.05925397574901581, + "learning_rate": 0.0008003382749620702, + "loss": 0.539, + "step": 309 + }, + { + "epoch": 0.13460703430308293, + "grad_norm": 0.06211424991488457, + "learning_rate": 0.000799055265245608, + "loss": 0.5581, + "step": 310 + }, + { + "epoch": 0.1350412505427703, + "grad_norm": 0.06610213220119476, + "learning_rate": 0.0007977691822735914, + "loss": 0.5726, + "step": 311 + }, + { + "epoch": 0.13547546678245767, + "grad_norm": 0.06418436765670776, + "learning_rate": 0.0007964800392625129, + "loss": 0.5616, + "step": 312 + }, + { + "epoch": 0.13590968302214504, + "grad_norm": 0.05993996933102608, + "learning_rate": 0.0007951878494603115, + "loss": 0.5392, + "step": 313 + }, + { + "epoch": 0.13634389926183238, + "grad_norm": 0.05776641145348549, + "learning_rate": 0.0007938926261462366, + "loss": 0.5402, + "step": 314 + }, + { + "epoch": 0.13677811550151975, + "grad_norm": 0.06915484368801117, + "learning_rate": 0.0007925943826307118, + "loss": 0.5336, + "step": 315 + }, + { + "epoch": 0.13721233174120712, + "grad_norm": 0.06394239515066147, + "learning_rate": 0.0007912931322551981, + "loss": 0.5647, + "step": 316 + }, + { + "epoch": 0.13764654798089448, + "grad_norm": 0.05997829511761665, + "learning_rate": 0.000789988888392056, + "loss": 0.5588, + "step": 317 + }, + { + "epoch": 0.13808076422058185, + "grad_norm": 0.07337518781423569, + "learning_rate": 0.0007886816644444098, + "loss": 0.5825, + "step": 318 + }, + { + "epoch": 0.13851498046026922, + "grad_norm": 0.07288537919521332, + "learning_rate": 0.0007873714738460075, + "loss": 0.5615, + "step": 319 + }, + { + "epoch": 0.1389491966999566, + "grad_norm": 0.06601582467556, + "learning_rate": 0.0007860583300610849, + "loss": 0.5545, + "step": 320 + }, + { + "epoch": 0.13938341293964393, + "grad_norm": 0.06303299218416214, + "learning_rate": 0.000784742246584226, + "loss": 0.5579, + "step": 321 + }, + { + "epoch": 0.1398176291793313, + "grad_norm": 0.06565733999013901, + "learning_rate": 0.000783423236940225, + "loss": 0.5728, + "step": 322 + }, + { + "epoch": 0.14025184541901867, + "grad_norm": 0.06590873748064041, + "learning_rate": 0.0007821013146839467, + "loss": 0.5552, + "step": 323 + }, + { + "epoch": 0.14068606165870604, + "grad_norm": 0.06782017648220062, + "learning_rate": 0.0007807764934001874, + "loss": 0.5299, + "step": 324 + }, + { + "epoch": 0.1411202778983934, + "grad_norm": 0.06369118392467499, + "learning_rate": 0.0007794487867035358, + "loss": 0.524, + "step": 325 + }, + { + "epoch": 0.14155449413808077, + "grad_norm": 0.06337650865316391, + "learning_rate": 0.0007781182082382324, + "loss": 0.5802, + "step": 326 + }, + { + "epoch": 0.14198871037776814, + "grad_norm": 0.07011637836694717, + "learning_rate": 0.0007767847716780296, + "loss": 0.5569, + "step": 327 + }, + { + "epoch": 0.14242292661745548, + "grad_norm": 0.06176460534334183, + "learning_rate": 0.0007754484907260512, + "loss": 0.5266, + "step": 328 + }, + { + "epoch": 0.14285714285714285, + "grad_norm": 0.06650613993406296, + "learning_rate": 0.0007741093791146516, + "loss": 0.5628, + "step": 329 + }, + { + "epoch": 0.14329135909683022, + "grad_norm": 0.07864295691251755, + "learning_rate": 0.0007727674506052743, + "loss": 0.5683, + "step": 330 + }, + { + "epoch": 0.14372557533651759, + "grad_norm": 0.06307731568813324, + "learning_rate": 0.0007714227189883112, + "loss": 0.5705, + "step": 331 + }, + { + "epoch": 0.14415979157620495, + "grad_norm": 0.07688986510038376, + "learning_rate": 0.0007700751980829601, + "loss": 0.5714, + "step": 332 + }, + { + "epoch": 0.14459400781589232, + "grad_norm": 0.06042364984750748, + "learning_rate": 0.0007687249017370832, + "loss": 0.5629, + "step": 333 + }, + { + "epoch": 0.1450282240555797, + "grad_norm": 0.06250355392694473, + "learning_rate": 0.0007673718438270648, + "loss": 0.5318, + "step": 334 + }, + { + "epoch": 0.14546244029526703, + "grad_norm": 0.06661161035299301, + "learning_rate": 0.0007660160382576683, + "loss": 0.5261, + "step": 335 + }, + { + "epoch": 0.1458966565349544, + "grad_norm": 0.07277819514274597, + "learning_rate": 0.0007646574989618937, + "loss": 0.5427, + "step": 336 + }, + { + "epoch": 0.14633087277464177, + "grad_norm": 0.05938050523400307, + "learning_rate": 0.0007632962399008341, + "loss": 0.5469, + "step": 337 + }, + { + "epoch": 0.14676508901432914, + "grad_norm": 0.0626835823059082, + "learning_rate": 0.0007619322750635327, + "loss": 0.5321, + "step": 338 + }, + { + "epoch": 0.1471993052540165, + "grad_norm": 0.06853507459163666, + "learning_rate": 0.0007605656184668384, + "loss": 0.5319, + "step": 339 + }, + { + "epoch": 0.14763352149370387, + "grad_norm": 0.06294318288564682, + "learning_rate": 0.0007591962841552626, + "loss": 0.5735, + "step": 340 + }, + { + "epoch": 0.14806773773339124, + "grad_norm": 0.06769460439682007, + "learning_rate": 0.0007578242862008336, + "loss": 0.5734, + "step": 341 + }, + { + "epoch": 0.14850195397307858, + "grad_norm": 0.06945870816707611, + "learning_rate": 0.0007564496387029531, + "loss": 0.5486, + "step": 342 + }, + { + "epoch": 0.14893617021276595, + "grad_norm": 0.055839743465185165, + "learning_rate": 0.0007550723557882513, + "loss": 0.5646, + "step": 343 + }, + { + "epoch": 0.14937038645245332, + "grad_norm": 0.05766444280743599, + "learning_rate": 0.0007536924516104411, + "loss": 0.5671, + "step": 344 + }, + { + "epoch": 0.1498046026921407, + "grad_norm": 0.05658617615699768, + "learning_rate": 0.000752309940350173, + "loss": 0.5573, + "step": 345 + }, + { + "epoch": 0.15023881893182806, + "grad_norm": 0.06456328183412552, + "learning_rate": 0.0007509248362148888, + "loss": 0.5413, + "step": 346 + }, + { + "epoch": 0.15067303517151542, + "grad_norm": 0.05971672758460045, + "learning_rate": 0.000749537153438677, + "loss": 0.528, + "step": 347 + }, + { + "epoch": 0.1511072514112028, + "grad_norm": 0.059805817902088165, + "learning_rate": 0.0007481469062821251, + "loss": 0.5517, + "step": 348 + }, + { + "epoch": 0.15154146765089013, + "grad_norm": 0.05796041712164879, + "learning_rate": 0.0007467541090321735, + "loss": 0.5651, + "step": 349 + }, + { + "epoch": 0.1519756838905775, + "grad_norm": 0.0807691365480423, + "learning_rate": 0.000745358776001969, + "loss": 0.5632, + "step": 350 + }, + { + "epoch": 0.15240990013026487, + "grad_norm": 0.056193944066762924, + "learning_rate": 0.0007439609215307173, + "loss": 0.5544, + "step": 351 + }, + { + "epoch": 0.15284411636995224, + "grad_norm": 0.06092338636517525, + "learning_rate": 0.000742560559983536, + "loss": 0.5378, + "step": 352 + }, + { + "epoch": 0.1532783326096396, + "grad_norm": 0.06231565400958061, + "learning_rate": 0.0007411577057513066, + "loss": 0.5584, + "step": 353 + }, + { + "epoch": 0.15371254884932697, + "grad_norm": 0.06663519144058228, + "learning_rate": 0.000739752373250527, + "loss": 0.5246, + "step": 354 + }, + { + "epoch": 0.15414676508901434, + "grad_norm": 0.060423221439123154, + "learning_rate": 0.0007383445769231627, + "loss": 0.5203, + "step": 355 + }, + { + "epoch": 0.15458098132870168, + "grad_norm": 0.06870024651288986, + "learning_rate": 0.0007369343312364993, + "loss": 0.52, + "step": 356 + }, + { + "epoch": 0.15501519756838905, + "grad_norm": 0.06956614553928375, + "learning_rate": 0.0007355216506829932, + "loss": 0.5816, + "step": 357 + }, + { + "epoch": 0.15544941380807642, + "grad_norm": 0.060565389692783356, + "learning_rate": 0.0007341065497801229, + "loss": 0.5508, + "step": 358 + }, + { + "epoch": 0.1558836300477638, + "grad_norm": 0.06745211035013199, + "learning_rate": 0.0007326890430702396, + "loss": 0.5589, + "step": 359 + }, + { + "epoch": 0.15631784628745116, + "grad_norm": 0.06009297072887421, + "learning_rate": 0.0007312691451204177, + "loss": 0.5259, + "step": 360 + }, + { + "epoch": 0.15675206252713852, + "grad_norm": 0.05682894587516785, + "learning_rate": 0.0007298468705223059, + "loss": 0.5689, + "step": 361 + }, + { + "epoch": 0.15718627876682587, + "grad_norm": 0.06062963977456093, + "learning_rate": 0.0007284222338919758, + "loss": 0.5339, + "step": 362 + }, + { + "epoch": 0.15762049500651323, + "grad_norm": 0.058002885431051254, + "learning_rate": 0.0007269952498697733, + "loss": 0.5714, + "step": 363 + }, + { + "epoch": 0.1580547112462006, + "grad_norm": 0.056898582726716995, + "learning_rate": 0.0007255659331201672, + "loss": 0.5811, + "step": 364 + }, + { + "epoch": 0.15848892748588797, + "grad_norm": 0.06246396526694298, + "learning_rate": 0.0007241342983315984, + "loss": 0.5554, + "step": 365 + }, + { + "epoch": 0.15892314372557534, + "grad_norm": 0.05486922711133957, + "learning_rate": 0.0007227003602163296, + "loss": 0.5567, + "step": 366 + }, + { + "epoch": 0.1593573599652627, + "grad_norm": 0.057243555784225464, + "learning_rate": 0.0007212641335102932, + "loss": 0.5193, + "step": 367 + }, + { + "epoch": 0.15979157620495008, + "grad_norm": 0.05640234053134918, + "learning_rate": 0.0007198256329729411, + "loss": 0.5098, + "step": 368 + }, + { + "epoch": 0.16022579244463742, + "grad_norm": 0.05819845199584961, + "learning_rate": 0.0007183848733870917, + "loss": 0.5366, + "step": 369 + }, + { + "epoch": 0.16066000868432478, + "grad_norm": 0.05703158304095268, + "learning_rate": 0.0007169418695587791, + "loss": 0.5374, + "step": 370 + }, + { + "epoch": 0.16109422492401215, + "grad_norm": 0.05363311246037483, + "learning_rate": 0.0007154966363171003, + "loss": 0.555, + "step": 371 + }, + { + "epoch": 0.16152844116369952, + "grad_norm": 0.060146309435367584, + "learning_rate": 0.0007140491885140629, + "loss": 0.5395, + "step": 372 + }, + { + "epoch": 0.1619626574033869, + "grad_norm": 0.05488353967666626, + "learning_rate": 0.0007125995410244324, + "loss": 0.5117, + "step": 373 + }, + { + "epoch": 0.16239687364307426, + "grad_norm": 0.05473377928137779, + "learning_rate": 0.00071114770874558, + "loss": 0.5463, + "step": 374 + }, + { + "epoch": 0.16283108988276163, + "grad_norm": 0.05373527109622955, + "learning_rate": 0.0007096937065973285, + "loss": 0.522, + "step": 375 + }, + { + "epoch": 0.16326530612244897, + "grad_norm": 0.054411571472883224, + "learning_rate": 0.0007082375495217995, + "loss": 0.519, + "step": 376 + }, + { + "epoch": 0.16369952236213633, + "grad_norm": 0.059973325580358505, + "learning_rate": 0.0007067792524832604, + "loss": 0.5546, + "step": 377 + }, + { + "epoch": 0.1641337386018237, + "grad_norm": 0.07386507093906403, + "learning_rate": 0.000705318830467969, + "loss": 0.548, + "step": 378 + }, + { + "epoch": 0.16456795484151107, + "grad_norm": 0.05318816378712654, + "learning_rate": 0.0007038562984840216, + "loss": 0.5325, + "step": 379 + }, + { + "epoch": 0.16500217108119844, + "grad_norm": 0.06362178921699524, + "learning_rate": 0.0007023916715611969, + "loss": 0.5537, + "step": 380 + }, + { + "epoch": 0.1654363873208858, + "grad_norm": 0.05186508968472481, + "learning_rate": 0.0007009249647508028, + "loss": 0.5534, + "step": 381 + }, + { + "epoch": 0.16587060356057318, + "grad_norm": 0.05456273630261421, + "learning_rate": 0.0006994561931255209, + "loss": 0.533, + "step": 382 + }, + { + "epoch": 0.16630481980026052, + "grad_norm": 0.07069144397974014, + "learning_rate": 0.0006979853717792523, + "loss": 0.5591, + "step": 383 + }, + { + "epoch": 0.16673903603994789, + "grad_norm": 0.13737140595912933, + "learning_rate": 0.0006965125158269618, + "loss": 0.8425, + "step": 384 + }, + { + "epoch": 0.16717325227963525, + "grad_norm": 0.08646462857723236, + "learning_rate": 0.0006950376404045235, + "loss": 0.5356, + "step": 385 + }, + { + "epoch": 0.16760746851932262, + "grad_norm": 0.059485238045454025, + "learning_rate": 0.0006935607606685641, + "loss": 0.571, + "step": 386 + }, + { + "epoch": 0.16804168475901, + "grad_norm": 0.07368376851081848, + "learning_rate": 0.000692081891796308, + "loss": 0.5346, + "step": 387 + }, + { + "epoch": 0.16847590099869736, + "grad_norm": 0.0615500770509243, + "learning_rate": 0.0006906010489854209, + "loss": 0.5257, + "step": 388 + }, + { + "epoch": 0.16891011723838473, + "grad_norm": 0.064671091735363, + "learning_rate": 0.0006891182474538539, + "loss": 0.5492, + "step": 389 + }, + { + "epoch": 0.16934433347807207, + "grad_norm": 0.06681560724973679, + "learning_rate": 0.0006876335024396872, + "loss": 0.5595, + "step": 390 + }, + { + "epoch": 0.16977854971775944, + "grad_norm": 0.0575130432844162, + "learning_rate": 0.0006861468292009726, + "loss": 0.5895, + "step": 391 + }, + { + "epoch": 0.1702127659574468, + "grad_norm": 0.05460560694336891, + "learning_rate": 0.0006846582430155782, + "loss": 0.5374, + "step": 392 + }, + { + "epoch": 0.17064698219713417, + "grad_norm": 0.06030447408556938, + "learning_rate": 0.0006831677591810301, + "loss": 0.6073, + "step": 393 + }, + { + "epoch": 0.17108119843682154, + "grad_norm": 0.0582847036421299, + "learning_rate": 0.0006816753930143557, + "loss": 0.5347, + "step": 394 + }, + { + "epoch": 0.1715154146765089, + "grad_norm": 0.05572697892785072, + "learning_rate": 0.0006801811598519267, + "loss": 0.5042, + "step": 395 + }, + { + "epoch": 0.17194963091619628, + "grad_norm": 0.05724466219544411, + "learning_rate": 0.0006786850750493006, + "loss": 0.5613, + "step": 396 + }, + { + "epoch": 0.17238384715588362, + "grad_norm": 0.06665828824043274, + "learning_rate": 0.0006771871539810632, + "loss": 0.5841, + "step": 397 + }, + { + "epoch": 0.172818063395571, + "grad_norm": 0.05465397238731384, + "learning_rate": 0.0006756874120406714, + "loss": 0.5636, + "step": 398 + }, + { + "epoch": 0.17325227963525835, + "grad_norm": 0.07157056033611298, + "learning_rate": 0.000674185864640294, + "loss": 0.561, + "step": 399 + }, + { + "epoch": 0.17368649587494572, + "grad_norm": 0.05561946704983711, + "learning_rate": 0.0006726825272106538, + "loss": 0.5351, + "step": 400 + }, + { + "epoch": 0.1741207121146331, + "grad_norm": 0.05475137010216713, + "learning_rate": 0.0006711774152008689, + "loss": 0.5445, + "step": 401 + }, + { + "epoch": 0.17455492835432046, + "grad_norm": 0.07090818136930466, + "learning_rate": 0.0006696705440782938, + "loss": 0.5279, + "step": 402 + }, + { + "epoch": 0.17498914459400783, + "grad_norm": 0.051751721650362015, + "learning_rate": 0.0006681619293283609, + "loss": 0.4934, + "step": 403 + }, + { + "epoch": 0.17542336083369517, + "grad_norm": 0.06089828908443451, + "learning_rate": 0.0006666515864544209, + "loss": 0.5413, + "step": 404 + }, + { + "epoch": 0.17585757707338254, + "grad_norm": 0.052899301052093506, + "learning_rate": 0.0006651395309775837, + "loss": 0.5187, + "step": 405 + }, + { + "epoch": 0.1762917933130699, + "grad_norm": 0.05137433856725693, + "learning_rate": 0.0006636257784365584, + "loss": 0.5636, + "step": 406 + }, + { + "epoch": 0.17672600955275727, + "grad_norm": 0.060585830360651016, + "learning_rate": 0.0006621103443874949, + "loss": 0.5311, + "step": 407 + }, + { + "epoch": 0.17716022579244464, + "grad_norm": 0.052542008459568024, + "learning_rate": 0.0006605932444038228, + "loss": 0.5344, + "step": 408 + }, + { + "epoch": 0.177594442032132, + "grad_norm": 0.052282754331827164, + "learning_rate": 0.0006590744940760914, + "loss": 0.5616, + "step": 409 + }, + { + "epoch": 0.17802865827181938, + "grad_norm": 0.055943962186574936, + "learning_rate": 0.0006575541090118104, + "loss": 0.5657, + "step": 410 + }, + { + "epoch": 0.17846287451150672, + "grad_norm": 0.05422157794237137, + "learning_rate": 0.0006560321048352886, + "loss": 0.5482, + "step": 411 + }, + { + "epoch": 0.1788970907511941, + "grad_norm": 0.053238000720739365, + "learning_rate": 0.0006545084971874737, + "loss": 0.5477, + "step": 412 + }, + { + "epoch": 0.17933130699088146, + "grad_norm": 0.05666307359933853, + "learning_rate": 0.0006529833017257919, + "loss": 0.5593, + "step": 413 + }, + { + "epoch": 0.17976552323056882, + "grad_norm": 0.05702383071184158, + "learning_rate": 0.000651456534123986, + "loss": 0.5229, + "step": 414 + }, + { + "epoch": 0.1801997394702562, + "grad_norm": 0.05041234940290451, + "learning_rate": 0.0006499282100719558, + "loss": 0.5197, + "step": 415 + }, + { + "epoch": 0.18063395570994356, + "grad_norm": 0.05691606178879738, + "learning_rate": 0.0006483983452755952, + "loss": 0.5282, + "step": 416 + }, + { + "epoch": 0.18106817194963093, + "grad_norm": 0.05236973240971565, + "learning_rate": 0.0006468669554566324, + "loss": 0.574, + "step": 417 + }, + { + "epoch": 0.18150238818931827, + "grad_norm": 0.05194111168384552, + "learning_rate": 0.0006453340563524669, + "loss": 0.521, + "step": 418 + }, + { + "epoch": 0.18193660442900564, + "grad_norm": 0.05307863652706146, + "learning_rate": 0.0006437996637160086, + "loss": 0.5474, + "step": 419 + }, + { + "epoch": 0.182370820668693, + "grad_norm": 0.05186399817466736, + "learning_rate": 0.0006422637933155162, + "loss": 0.5584, + "step": 420 + }, + { + "epoch": 0.18280503690838038, + "grad_norm": 0.053380340337753296, + "learning_rate": 0.0006407264609344343, + "loss": 0.5309, + "step": 421 + }, + { + "epoch": 0.18323925314806774, + "grad_norm": 0.05926523730158806, + "learning_rate": 0.0006391876823712317, + "loss": 0.5362, + "step": 422 + }, + { + "epoch": 0.1836734693877551, + "grad_norm": 0.053086958825588226, + "learning_rate": 0.0006376474734392387, + "loss": 0.5345, + "step": 423 + }, + { + "epoch": 0.18410768562744248, + "grad_norm": 0.05506280064582825, + "learning_rate": 0.0006361058499664855, + "loss": 0.5301, + "step": 424 + }, + { + "epoch": 0.18454190186712982, + "grad_norm": 0.0567692294716835, + "learning_rate": 0.0006345628277955385, + "loss": 0.5336, + "step": 425 + }, + { + "epoch": 0.1849761181068172, + "grad_norm": 0.04989070072770119, + "learning_rate": 0.0006330184227833376, + "loss": 0.53, + "step": 426 + }, + { + "epoch": 0.18541033434650456, + "grad_norm": 0.05129670351743698, + "learning_rate": 0.000631472650801034, + "loss": 0.523, + "step": 427 + }, + { + "epoch": 0.18584455058619193, + "grad_norm": 0.052287109196186066, + "learning_rate": 0.0006299255277338265, + "loss": 0.5309, + "step": 428 + }, + { + "epoch": 0.1862787668258793, + "grad_norm": 0.051617227494716644, + "learning_rate": 0.0006283770694807982, + "loss": 0.5406, + "step": 429 + }, + { + "epoch": 0.18671298306556666, + "grad_norm": 0.06112902611494064, + "learning_rate": 0.0006268272919547536, + "loss": 0.5148, + "step": 430 + }, + { + "epoch": 0.187147199305254, + "grad_norm": 0.04904035106301308, + "learning_rate": 0.0006252762110820547, + "loss": 0.5277, + "step": 431 + }, + { + "epoch": 0.18758141554494137, + "grad_norm": 0.05905110388994217, + "learning_rate": 0.0006237238428024572, + "loss": 0.5148, + "step": 432 + }, + { + "epoch": 0.18801563178462874, + "grad_norm": 0.05047300457954407, + "learning_rate": 0.000622170203068947, + "loss": 0.5119, + "step": 433 + }, + { + "epoch": 0.1884498480243161, + "grad_norm": 0.0518287792801857, + "learning_rate": 0.0006206153078475762, + "loss": 0.5276, + "step": 434 + }, + { + "epoch": 0.18888406426400348, + "grad_norm": 0.04822736606001854, + "learning_rate": 0.0006190591731172991, + "loss": 0.5347, + "step": 435 + }, + { + "epoch": 0.18931828050369084, + "grad_norm": 0.052807264029979706, + "learning_rate": 0.0006175018148698076, + "loss": 0.5343, + "step": 436 + }, + { + "epoch": 0.1897524967433782, + "grad_norm": 0.054650746285915375, + "learning_rate": 0.0006159432491093672, + "loss": 0.5079, + "step": 437 + }, + { + "epoch": 0.19018671298306555, + "grad_norm": 0.047450825572013855, + "learning_rate": 0.0006143834918526527, + "loss": 0.5532, + "step": 438 + }, + { + "epoch": 0.19062092922275292, + "grad_norm": 0.047222550958395004, + "learning_rate": 0.0006128225591285831, + "loss": 0.5376, + "step": 439 + }, + { + "epoch": 0.1910551454624403, + "grad_norm": 0.05076510086655617, + "learning_rate": 0.0006112604669781572, + "loss": 0.5152, + "step": 440 + }, + { + "epoch": 0.19148936170212766, + "grad_norm": 0.06690964847803116, + "learning_rate": 0.0006096972314542889, + "loss": 0.5395, + "step": 441 + }, + { + "epoch": 0.19192357794181503, + "grad_norm": 0.049380868673324585, + "learning_rate": 0.0006081328686216418, + "loss": 0.4996, + "step": 442 + }, + { + "epoch": 0.1923577941815024, + "grad_norm": 0.06767404824495316, + "learning_rate": 0.0006065673945564642, + "loss": 0.5217, + "step": 443 + }, + { + "epoch": 0.19279201042118976, + "grad_norm": 0.054420776665210724, + "learning_rate": 0.0006050008253464246, + "loss": 0.5587, + "step": 444 + }, + { + "epoch": 0.1932262266608771, + "grad_norm": 0.0557849258184433, + "learning_rate": 0.0006034331770904454, + "loss": 0.5441, + "step": 445 + }, + { + "epoch": 0.19366044290056447, + "grad_norm": 0.06514862924814224, + "learning_rate": 0.0006018644658985379, + "loss": 0.5666, + "step": 446 + }, + { + "epoch": 0.19409465914025184, + "grad_norm": 0.061305928975343704, + "learning_rate": 0.0006002947078916364, + "loss": 0.5576, + "step": 447 + }, + { + "epoch": 0.1945288753799392, + "grad_norm": 0.05214373767375946, + "learning_rate": 0.0005987239192014335, + "loss": 0.5424, + "step": 448 + }, + { + "epoch": 0.19496309161962658, + "grad_norm": 0.05650056526064873, + "learning_rate": 0.0005971521159702136, + "loss": 0.5361, + "step": 449 + }, + { + "epoch": 0.19539730785931395, + "grad_norm": 0.05745421350002289, + "learning_rate": 0.0005955793143506863, + "loss": 0.5273, + "step": 450 + }, + { + "epoch": 0.19583152409900131, + "grad_norm": 0.05364866554737091, + "learning_rate": 0.0005940055305058219, + "loss": 0.5026, + "step": 451 + }, + { + "epoch": 0.19626574033868865, + "grad_norm": 0.047814685851335526, + "learning_rate": 0.0005924307806086844, + "loss": 0.5439, + "step": 452 + }, + { + "epoch": 0.19669995657837602, + "grad_norm": 0.051681190729141235, + "learning_rate": 0.0005908550808422655, + "loss": 0.5271, + "step": 453 + }, + { + "epoch": 0.1971341728180634, + "grad_norm": 0.05405402183532715, + "learning_rate": 0.0005892784473993184, + "loss": 0.4996, + "step": 454 + }, + { + "epoch": 0.19756838905775076, + "grad_norm": 0.04617345333099365, + "learning_rate": 0.0005877008964821908, + "loss": 0.5409, + "step": 455 + }, + { + "epoch": 0.19800260529743813, + "grad_norm": 0.05039280652999878, + "learning_rate": 0.0005861224443026595, + "loss": 0.5433, + "step": 456 + }, + { + "epoch": 0.1984368215371255, + "grad_norm": 0.05184314772486687, + "learning_rate": 0.0005845431070817626, + "loss": 0.5422, + "step": 457 + }, + { + "epoch": 0.19887103777681286, + "grad_norm": 0.056363269686698914, + "learning_rate": 0.000582962901049634, + "loss": 0.5202, + "step": 458 + }, + { + "epoch": 0.1993052540165002, + "grad_norm": 0.04747706279158592, + "learning_rate": 0.0005813818424453351, + "loss": 0.5287, + "step": 459 + }, + { + "epoch": 0.19973947025618757, + "grad_norm": 0.05112398415803909, + "learning_rate": 0.0005797999475166897, + "loss": 0.563, + "step": 460 + }, + { + "epoch": 0.20017368649587494, + "grad_norm": 0.050983451306819916, + "learning_rate": 0.0005782172325201155, + "loss": 0.5309, + "step": 461 + }, + { + "epoch": 0.2006079027355623, + "grad_norm": 0.05590641498565674, + "learning_rate": 0.000576633713720458, + "loss": 0.54, + "step": 462 + }, + { + "epoch": 0.20104211897524968, + "grad_norm": 0.04705791547894478, + "learning_rate": 0.000575049407390823, + "loss": 0.5291, + "step": 463 + }, + { + "epoch": 0.20147633521493705, + "grad_norm": 0.05234816297888756, + "learning_rate": 0.000573464329812409, + "loss": 0.5331, + "step": 464 + }, + { + "epoch": 0.20191055145462442, + "grad_norm": 0.04806197062134743, + "learning_rate": 0.0005718784972743409, + "loss": 0.5141, + "step": 465 + }, + { + "epoch": 0.20234476769431176, + "grad_norm": 0.0483107753098011, + "learning_rate": 0.0005702919260735014, + "loss": 0.5066, + "step": 466 + }, + { + "epoch": 0.20277898393399912, + "grad_norm": 0.05739190801978111, + "learning_rate": 0.0005687046325143647, + "loss": 0.5282, + "step": 467 + }, + { + "epoch": 0.2032132001736865, + "grad_norm": 0.04906076937913895, + "learning_rate": 0.0005671166329088278, + "loss": 0.5257, + "step": 468 + }, + { + "epoch": 0.20364741641337386, + "grad_norm": 0.04784254729747772, + "learning_rate": 0.0005655279435760435, + "loss": 0.5193, + "step": 469 + }, + { + "epoch": 0.20408163265306123, + "grad_norm": 0.06473550200462341, + "learning_rate": 0.000563938580842253, + "loss": 0.5257, + "step": 470 + }, + { + "epoch": 0.2045158488927486, + "grad_norm": 0.047775350511074066, + "learning_rate": 0.0005623485610406173, + "loss": 0.5008, + "step": 471 + }, + { + "epoch": 0.20495006513243597, + "grad_norm": 0.05449296161532402, + "learning_rate": 0.0005607579005110502, + "loss": 0.5316, + "step": 472 + }, + { + "epoch": 0.2053842813721233, + "grad_norm": 0.05296491086483002, + "learning_rate": 0.0005591666156000494, + "loss": 0.5493, + "step": 473 + }, + { + "epoch": 0.20581849761181067, + "grad_norm": 0.045328423380851746, + "learning_rate": 0.0005575747226605297, + "loss": 0.508, + "step": 474 + }, + { + "epoch": 0.20625271385149804, + "grad_norm": 0.05045296251773834, + "learning_rate": 0.0005559822380516539, + "loss": 0.5118, + "step": 475 + }, + { + "epoch": 0.2066869300911854, + "grad_norm": 0.04571797326207161, + "learning_rate": 0.0005543891781386656, + "loss": 0.5168, + "step": 476 + }, + { + "epoch": 0.20712114633087278, + "grad_norm": 0.05250009894371033, + "learning_rate": 0.0005527955592927197, + "loss": 0.5175, + "step": 477 + }, + { + "epoch": 0.20755536257056015, + "grad_norm": 0.048049163073301315, + "learning_rate": 0.0005512013978907157, + "loss": 0.5279, + "step": 478 + }, + { + "epoch": 0.20798957881024752, + "grad_norm": 0.04621399939060211, + "learning_rate": 0.0005496067103151288, + "loss": 0.5304, + "step": 479 + }, + { + "epoch": 0.20842379504993486, + "grad_norm": 0.04643552750349045, + "learning_rate": 0.0005480115129538409, + "loss": 0.5137, + "step": 480 + }, + { + "epoch": 0.20885801128962223, + "grad_norm": 0.05422172695398331, + "learning_rate": 0.0005464158221999731, + "loss": 0.525, + "step": 481 + }, + { + "epoch": 0.2092922275293096, + "grad_norm": 0.04593772441148758, + "learning_rate": 0.0005448196544517168, + "loss": 0.5229, + "step": 482 + }, + { + "epoch": 0.20972644376899696, + "grad_norm": 0.05278801918029785, + "learning_rate": 0.0005432230261121651, + "loss": 0.5256, + "step": 483 + }, + { + "epoch": 0.21016066000868433, + "grad_norm": 0.04598787799477577, + "learning_rate": 0.0005416259535891447, + "loss": 0.5275, + "step": 484 + }, + { + "epoch": 0.2105948762483717, + "grad_norm": 0.054117828607559204, + "learning_rate": 0.0005400284532950467, + "loss": 0.5179, + "step": 485 + }, + { + "epoch": 0.21102909248805907, + "grad_norm": 0.046420540660619736, + "learning_rate": 0.0005384305416466584, + "loss": 0.5399, + "step": 486 + }, + { + "epoch": 0.2114633087277464, + "grad_norm": 0.048214301466941833, + "learning_rate": 0.0005368322350649942, + "loss": 0.5202, + "step": 487 + }, + { + "epoch": 0.21189752496743378, + "grad_norm": 0.05419298633933067, + "learning_rate": 0.0005352335499751269, + "loss": 0.4925, + "step": 488 + }, + { + "epoch": 0.21233174120712114, + "grad_norm": 0.04759565740823746, + "learning_rate": 0.0005336345028060199, + "loss": 0.5162, + "step": 489 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 0.05174829810857773, + "learning_rate": 0.0005320351099903565, + "loss": 0.5043, + "step": 490 + }, + { + "epoch": 0.21320017368649588, + "grad_norm": 0.06595347821712494, + "learning_rate": 0.0005304353879643726, + "loss": 0.5029, + "step": 491 + }, + { + "epoch": 0.21363438992618325, + "grad_norm": 0.04908424988389015, + "learning_rate": 0.0005288353531676872, + "loss": 0.5043, + "step": 492 + }, + { + "epoch": 0.21406860616587062, + "grad_norm": 0.05055024474859238, + "learning_rate": 0.0005272350220431334, + "loss": 0.5192, + "step": 493 + }, + { + "epoch": 0.21450282240555796, + "grad_norm": 0.049286942929029465, + "learning_rate": 0.0005256344110365896, + "loss": 0.5041, + "step": 494 + }, + { + "epoch": 0.21493703864524533, + "grad_norm": 0.04700294882059097, + "learning_rate": 0.0005240335365968104, + "loss": 0.5258, + "step": 495 + }, + { + "epoch": 0.2153712548849327, + "grad_norm": 0.0560920424759388, + "learning_rate": 0.0005224324151752575, + "loss": 0.5388, + "step": 496 + }, + { + "epoch": 0.21580547112462006, + "grad_norm": 0.05053974315524101, + "learning_rate": 0.0005208310632259308, + "loss": 0.5045, + "step": 497 + }, + { + "epoch": 0.21623968736430743, + "grad_norm": 0.06237909570336342, + "learning_rate": 0.0005192294972051992, + "loss": 0.463, + "step": 498 + }, + { + "epoch": 0.2166739036039948, + "grad_norm": 0.05128193646669388, + "learning_rate": 0.0005176277335716317, + "loss": 0.5209, + "step": 499 + }, + { + "epoch": 0.21710811984368214, + "grad_norm": 0.04433543235063553, + "learning_rate": 0.0005160257887858277, + "loss": 0.5046, + "step": 500 + }, + { + "epoch": 0.2175423360833695, + "grad_norm": 0.05125072970986366, + "learning_rate": 0.0005144236793102484, + "loss": 0.4894, + "step": 501 + }, + { + "epoch": 0.21797655232305688, + "grad_norm": 0.04465992748737335, + "learning_rate": 0.0005128214216090478, + "loss": 0.4987, + "step": 502 + }, + { + "epoch": 0.21841076856274425, + "grad_norm": 0.04684034362435341, + "learning_rate": 0.0005112190321479025, + "loss": 0.5498, + "step": 503 + }, + { + "epoch": 0.2188449848024316, + "grad_norm": 0.04876565933227539, + "learning_rate": 0.0005096165273938436, + "loss": 0.5071, + "step": 504 + }, + { + "epoch": 0.21927920104211898, + "grad_norm": 0.04982059448957443, + "learning_rate": 0.0005080139238150869, + "loss": 0.5084, + "step": 505 + }, + { + "epoch": 0.21971341728180635, + "grad_norm": 0.052202560007572174, + "learning_rate": 0.0005064112378808637, + "loss": 0.5233, + "step": 506 + }, + { + "epoch": 0.2201476335214937, + "grad_norm": 0.04747169837355614, + "learning_rate": 0.0005048084860612516, + "loss": 0.5263, + "step": 507 + }, + { + "epoch": 0.22058184976118106, + "grad_norm": 0.0462409108877182, + "learning_rate": 0.0005032056848270056, + "loss": 0.5237, + "step": 508 + }, + { + "epoch": 0.22101606600086843, + "grad_norm": 0.04612501338124275, + "learning_rate": 0.000501602850649388, + "loss": 0.5507, + "step": 509 + }, + { + "epoch": 0.2214502822405558, + "grad_norm": 0.04581817239522934, + "learning_rate": 0.0005, + "loss": 0.4986, + "step": 510 + }, + { + "epoch": 0.22188449848024316, + "grad_norm": 0.0477612242102623, + "learning_rate": 0.0004983971493506121, + "loss": 0.51, + "step": 511 + }, + { + "epoch": 0.22231871471993053, + "grad_norm": 0.05176355689764023, + "learning_rate": 0.0004967943151729944, + "loss": 0.5115, + "step": 512 + }, + { + "epoch": 0.2227529309596179, + "grad_norm": 0.04731719568371773, + "learning_rate": 0.0004951915139387483, + "loss": 0.5564, + "step": 513 + }, + { + "epoch": 0.22318714719930524, + "grad_norm": 0.04438960924744606, + "learning_rate": 0.0004935887621191363, + "loss": 0.5502, + "step": 514 + }, + { + "epoch": 0.2236213634389926, + "grad_norm": 0.06116556003689766, + "learning_rate": 0.0004919860761849132, + "loss": 0.5417, + "step": 515 + }, + { + "epoch": 0.22405557967867998, + "grad_norm": 0.046353522688150406, + "learning_rate": 0.0004903834726061564, + "loss": 0.5285, + "step": 516 + }, + { + "epoch": 0.22448979591836735, + "grad_norm": 0.04643326997756958, + "learning_rate": 0.0004887809678520976, + "loss": 0.4774, + "step": 517 + }, + { + "epoch": 0.22492401215805471, + "grad_norm": 0.0530591681599617, + "learning_rate": 0.0004871785783909523, + "loss": 0.5076, + "step": 518 + }, + { + "epoch": 0.22535822839774208, + "grad_norm": 0.04789597541093826, + "learning_rate": 0.0004855763206897516, + "loss": 0.4989, + "step": 519 + }, + { + "epoch": 0.22579244463742945, + "grad_norm": 0.047394026070833206, + "learning_rate": 0.0004839742112141724, + "loss": 0.5069, + "step": 520 + }, + { + "epoch": 0.2262266608771168, + "grad_norm": 0.044425006955862045, + "learning_rate": 0.0004823722664283684, + "loss": 0.5212, + "step": 521 + }, + { + "epoch": 0.22666087711680416, + "grad_norm": 0.044534265995025635, + "learning_rate": 0.0004807705027948008, + "loss": 0.5184, + "step": 522 + }, + { + "epoch": 0.22709509335649153, + "grad_norm": 0.04293319582939148, + "learning_rate": 0.0004791689367740692, + "loss": 0.4807, + "step": 523 + }, + { + "epoch": 0.2275293095961789, + "grad_norm": 0.04485393315553665, + "learning_rate": 0.0004775675848247427, + "loss": 0.4746, + "step": 524 + }, + { + "epoch": 0.22796352583586627, + "grad_norm": 0.043829191476106644, + "learning_rate": 0.0004759664634031897, + "loss": 0.4936, + "step": 525 + }, + { + "epoch": 0.22839774207555363, + "grad_norm": 0.05110118165612221, + "learning_rate": 0.00047436558896341046, + "loss": 0.5849, + "step": 526 + }, + { + "epoch": 0.228831958315241, + "grad_norm": 0.04783101752400398, + "learning_rate": 0.0004727649779568666, + "loss": 0.5034, + "step": 527 + }, + { + "epoch": 0.22926617455492834, + "grad_norm": 0.047072965651750565, + "learning_rate": 0.00047116464683231285, + "loss": 0.5155, + "step": 528 + }, + { + "epoch": 0.2297003907946157, + "grad_norm": 0.0427449531853199, + "learning_rate": 0.0004695646120356274, + "loss": 0.5138, + "step": 529 + }, + { + "epoch": 0.23013460703430308, + "grad_norm": 0.04664051905274391, + "learning_rate": 0.0004679648900096436, + "loss": 0.4727, + "step": 530 + }, + { + "epoch": 0.23056882327399045, + "grad_norm": 0.04555191844701767, + "learning_rate": 0.0004663654971939802, + "loss": 0.5386, + "step": 531 + }, + { + "epoch": 0.23100303951367782, + "grad_norm": 0.05113929882645607, + "learning_rate": 0.00046476645002487297, + "loss": 0.5018, + "step": 532 + }, + { + "epoch": 0.23143725575336518, + "grad_norm": 0.04245986044406891, + "learning_rate": 0.00046316776493500613, + "loss": 0.4949, + "step": 533 + }, + { + "epoch": 0.23187147199305255, + "grad_norm": 0.047864172607660294, + "learning_rate": 0.0004615694583533418, + "loss": 0.5032, + "step": 534 + }, + { + "epoch": 0.2323056882327399, + "grad_norm": 0.06898235529661179, + "learning_rate": 0.0004599715467049534, + "loss": 0.4922, + "step": 535 + }, + { + "epoch": 0.23273990447242726, + "grad_norm": 0.04230332002043724, + "learning_rate": 0.0004583740464108554, + "loss": 0.5164, + "step": 536 + }, + { + "epoch": 0.23317412071211463, + "grad_norm": 0.043854959309101105, + "learning_rate": 0.00045677697388783495, + "loss": 0.4894, + "step": 537 + }, + { + "epoch": 0.233608336951802, + "grad_norm": 0.05595362186431885, + "learning_rate": 0.0004551803455482833, + "loss": 0.5107, + "step": 538 + }, + { + "epoch": 0.23404255319148937, + "grad_norm": 0.05336504802107811, + "learning_rate": 0.0004535841778000269, + "loss": 0.4845, + "step": 539 + }, + { + "epoch": 0.23447676943117673, + "grad_norm": 0.052477382123470306, + "learning_rate": 0.0004519884870461591, + "loss": 0.5215, + "step": 540 + }, + { + "epoch": 0.2349109856708641, + "grad_norm": 0.04965275526046753, + "learning_rate": 0.00045039328968487125, + "loss": 0.499, + "step": 541 + }, + { + "epoch": 0.23534520191055144, + "grad_norm": 0.07106837630271912, + "learning_rate": 0.00044879860210928434, + "loss": 0.5159, + "step": 542 + }, + { + "epoch": 0.2357794181502388, + "grad_norm": 0.06256967037916183, + "learning_rate": 0.0004472044407072805, + "loss": 0.505, + "step": 543 + }, + { + "epoch": 0.23621363438992618, + "grad_norm": 0.04907793179154396, + "learning_rate": 0.0004456108218613346, + "loss": 0.5117, + "step": 544 + }, + { + "epoch": 0.23664785062961355, + "grad_norm": 0.06353382021188736, + "learning_rate": 0.0004440177619483461, + "loss": 0.4941, + "step": 545 + }, + { + "epoch": 0.23708206686930092, + "grad_norm": 0.05265835300087929, + "learning_rate": 0.0004424252773394704, + "loss": 0.5179, + "step": 546 + }, + { + "epoch": 0.23751628310898829, + "grad_norm": 0.05525769665837288, + "learning_rate": 0.0004408333843999506, + "loss": 0.4938, + "step": 547 + }, + { + "epoch": 0.23795049934867565, + "grad_norm": 0.06166142225265503, + "learning_rate": 0.00043924209948894985, + "loss": 0.5287, + "step": 548 + }, + { + "epoch": 0.238384715588363, + "grad_norm": 0.04812607169151306, + "learning_rate": 0.0004376514389593826, + "loss": 0.5401, + "step": 549 + }, + { + "epoch": 0.23881893182805036, + "grad_norm": 0.048272911459207535, + "learning_rate": 0.00043606141915774693, + "loss": 0.5258, + "step": 550 + }, + { + "epoch": 0.23925314806773773, + "grad_norm": 0.04806293919682503, + "learning_rate": 0.00043447205642395664, + "loss": 0.5023, + "step": 551 + }, + { + "epoch": 0.2396873643074251, + "grad_norm": 0.04957396164536476, + "learning_rate": 0.0004328833670911724, + "loss": 0.5015, + "step": 552 + }, + { + "epoch": 0.24012158054711247, + "grad_norm": 0.04413224756717682, + "learning_rate": 0.00043129536748563545, + "loss": 0.5336, + "step": 553 + }, + { + "epoch": 0.24055579678679984, + "grad_norm": 0.04619096592068672, + "learning_rate": 0.0004297080739264987, + "loss": 0.4832, + "step": 554 + }, + { + "epoch": 0.2409900130264872, + "grad_norm": 0.05157000198960304, + "learning_rate": 0.0004281215027256592, + "loss": 0.529, + "step": 555 + }, + { + "epoch": 0.24142422926617454, + "grad_norm": 0.04697950556874275, + "learning_rate": 0.00042653567018759105, + "loss": 0.5034, + "step": 556 + }, + { + "epoch": 0.2418584455058619, + "grad_norm": 0.04245249181985855, + "learning_rate": 0.0004249505926091771, + "loss": 0.4839, + "step": 557 + }, + { + "epoch": 0.24229266174554928, + "grad_norm": 0.042621515691280365, + "learning_rate": 0.000423366286279542, + "loss": 0.5639, + "step": 558 + }, + { + "epoch": 0.24272687798523665, + "grad_norm": 0.054264895617961884, + "learning_rate": 0.0004217827674798845, + "loss": 0.5077, + "step": 559 + }, + { + "epoch": 0.24316109422492402, + "grad_norm": 0.04171125963330269, + "learning_rate": 0.0004202000524833105, + "loss": 0.5001, + "step": 560 + }, + { + "epoch": 0.2435953104646114, + "grad_norm": 0.047443173825740814, + "learning_rate": 0.00041861815755466506, + "loss": 0.5153, + "step": 561 + }, + { + "epoch": 0.24402952670429873, + "grad_norm": 0.04359521344304085, + "learning_rate": 0.0004170370989503662, + "loss": 0.516, + "step": 562 + }, + { + "epoch": 0.2444637429439861, + "grad_norm": 0.042067334055900574, + "learning_rate": 0.0004154568929182374, + "loss": 0.4783, + "step": 563 + }, + { + "epoch": 0.24489795918367346, + "grad_norm": 0.042515527456998825, + "learning_rate": 0.00041387755569734057, + "loss": 0.5112, + "step": 564 + }, + { + "epoch": 0.24533217542336083, + "grad_norm": 0.04880267009139061, + "learning_rate": 0.00041229910351780926, + "loss": 0.5108, + "step": 565 + }, + { + "epoch": 0.2457663916630482, + "grad_norm": 0.048264361917972565, + "learning_rate": 0.0004107215526006817, + "loss": 0.4802, + "step": 566 + }, + { + "epoch": 0.24620060790273557, + "grad_norm": 0.0446770153939724, + "learning_rate": 0.00040914491915773453, + "loss": 0.5461, + "step": 567 + }, + { + "epoch": 0.24663482414242294, + "grad_norm": 0.05128175765275955, + "learning_rate": 0.00040756921939131565, + "loss": 0.5288, + "step": 568 + }, + { + "epoch": 0.24706904038211028, + "grad_norm": 0.05804765224456787, + "learning_rate": 0.0004059944694941783, + "loss": 0.4981, + "step": 569 + }, + { + "epoch": 0.24750325662179765, + "grad_norm": 0.04847874119877815, + "learning_rate": 0.00040442068564931397, + "loss": 0.5185, + "step": 570 + }, + { + "epoch": 0.24793747286148501, + "grad_norm": 0.04028409346938133, + "learning_rate": 0.0004028478840297866, + "loss": 0.493, + "step": 571 + }, + { + "epoch": 0.24837168910117238, + "grad_norm": 0.04243363067507744, + "learning_rate": 0.00040127608079856646, + "loss": 0.5079, + "step": 572 + }, + { + "epoch": 0.24880590534085975, + "grad_norm": 0.04515422508120537, + "learning_rate": 0.00039970529210836363, + "loss": 0.5121, + "step": 573 + }, + { + "epoch": 0.24924012158054712, + "grad_norm": 0.05250799283385277, + "learning_rate": 0.00039813553410146226, + "loss": 0.4975, + "step": 574 + }, + { + "epoch": 0.2496743378202345, + "grad_norm": 0.043340008705854416, + "learning_rate": 0.00039656682290955457, + "loss": 0.5094, + "step": 575 + }, + { + "epoch": 0.25010855405992183, + "grad_norm": 0.04954323545098305, + "learning_rate": 0.00039499917465357534, + "loss": 0.5327, + "step": 576 + }, + { + "epoch": 0.2505427702996092, + "grad_norm": 0.04032951593399048, + "learning_rate": 0.00039343260544353573, + "loss": 0.5035, + "step": 577 + }, + { + "epoch": 0.25097698653929656, + "grad_norm": 0.04386662319302559, + "learning_rate": 0.0003918671313783583, + "loss": 0.5142, + "step": 578 + }, + { + "epoch": 0.25141120277898393, + "grad_norm": 0.04662942886352539, + "learning_rate": 0.00039030276854571115, + "loss": 0.4996, + "step": 579 + }, + { + "epoch": 0.2518454190186713, + "grad_norm": 0.046797946095466614, + "learning_rate": 0.00038873953302184284, + "loss": 0.4728, + "step": 580 + }, + { + "epoch": 0.25227963525835867, + "grad_norm": 0.041140951216220856, + "learning_rate": 0.000387177440871417, + "loss": 0.546, + "step": 581 + }, + { + "epoch": 0.25271385149804604, + "grad_norm": 0.04658018797636032, + "learning_rate": 0.00038561650814734736, + "loss": 0.4804, + "step": 582 + }, + { + "epoch": 0.2531480677377334, + "grad_norm": 0.04261363297700882, + "learning_rate": 0.0003840567508906328, + "loss": 0.4833, + "step": 583 + }, + { + "epoch": 0.2535822839774208, + "grad_norm": 0.046267736703157425, + "learning_rate": 0.0003824981851301924, + "loss": 0.4809, + "step": 584 + }, + { + "epoch": 0.25401650021710814, + "grad_norm": 0.04030013084411621, + "learning_rate": 0.0003809408268827009, + "loss": 0.4878, + "step": 585 + }, + { + "epoch": 0.2544507164567955, + "grad_norm": 0.041512105613946915, + "learning_rate": 0.0003793846921524237, + "loss": 0.5165, + "step": 586 + }, + { + "epoch": 0.2548849326964828, + "grad_norm": 0.04772355034947395, + "learning_rate": 0.00037782979693105293, + "loss": 0.5193, + "step": 587 + }, + { + "epoch": 0.2553191489361702, + "grad_norm": 0.040986839681863785, + "learning_rate": 0.00037627615719754295, + "loss": 0.5028, + "step": 588 + }, + { + "epoch": 0.25575336517585756, + "grad_norm": 0.04139196127653122, + "learning_rate": 0.00037472378891794533, + "loss": 0.4998, + "step": 589 + }, + { + "epoch": 0.25618758141554493, + "grad_norm": 0.04017603024840355, + "learning_rate": 0.0003731727080452464, + "loss": 0.5239, + "step": 590 + }, + { + "epoch": 0.2566217976552323, + "grad_norm": 0.04039851576089859, + "learning_rate": 0.00037162293051920184, + "loss": 0.4739, + "step": 591 + }, + { + "epoch": 0.25705601389491967, + "grad_norm": 0.04436818137764931, + "learning_rate": 0.0003700744722661736, + "loss": 0.5153, + "step": 592 + }, + { + "epoch": 0.25749023013460703, + "grad_norm": 0.03977528214454651, + "learning_rate": 0.0003685273491989661, + "loss": 0.5217, + "step": 593 + }, + { + "epoch": 0.2579244463742944, + "grad_norm": 0.038483258336782455, + "learning_rate": 0.0003669815772166625, + "loss": 0.4904, + "step": 594 + }, + { + "epoch": 0.25835866261398177, + "grad_norm": 0.04520029202103615, + "learning_rate": 0.0003654371722044616, + "loss": 0.5303, + "step": 595 + }, + { + "epoch": 0.25879287885366914, + "grad_norm": 0.04642047360539436, + "learning_rate": 0.00036389415003351444, + "loss": 0.4783, + "step": 596 + }, + { + "epoch": 0.2592270950933565, + "grad_norm": 0.037170182913541794, + "learning_rate": 0.00036235252656076133, + "loss": 0.5254, + "step": 597 + }, + { + "epoch": 0.2596613113330439, + "grad_norm": 0.037870265543460846, + "learning_rate": 0.00036081231762876846, + "loss": 0.4933, + "step": 598 + }, + { + "epoch": 0.26009552757273124, + "grad_norm": 0.03957228735089302, + "learning_rate": 0.0003592735390655658, + "loss": 0.5194, + "step": 599 + }, + { + "epoch": 0.26052974381241856, + "grad_norm": 0.04471513256430626, + "learning_rate": 0.0003577362066844838, + "loss": 0.4737, + "step": 600 + }, + { + "epoch": 0.2609639600521059, + "grad_norm": 0.0370667465031147, + "learning_rate": 0.0003562003362839914, + "loss": 0.5072, + "step": 601 + }, + { + "epoch": 0.2613981762917933, + "grad_norm": 0.04277738928794861, + "learning_rate": 0.00035466594364753326, + "loss": 0.5093, + "step": 602 + }, + { + "epoch": 0.26183239253148066, + "grad_norm": 0.04622077941894531, + "learning_rate": 0.00035313304454336763, + "loss": 0.5008, + "step": 603 + }, + { + "epoch": 0.26226660877116803, + "grad_norm": 0.0380844846367836, + "learning_rate": 0.0003516016547244047, + "loss": 0.4899, + "step": 604 + }, + { + "epoch": 0.2627008250108554, + "grad_norm": 0.03964554890990257, + "learning_rate": 0.0003500717899280442, + "loss": 0.544, + "step": 605 + }, + { + "epoch": 0.26313504125054277, + "grad_norm": 0.041507843881845474, + "learning_rate": 0.00034854346587601397, + "loss": 0.5341, + "step": 606 + }, + { + "epoch": 0.26356925749023014, + "grad_norm": 0.04041410982608795, + "learning_rate": 0.0003470166982742082, + "loss": 0.5325, + "step": 607 + }, + { + "epoch": 0.2640034737299175, + "grad_norm": 0.03915121778845787, + "learning_rate": 0.00034549150281252633, + "loss": 0.5164, + "step": 608 + }, + { + "epoch": 0.26443768996960487, + "grad_norm": 0.04141729697585106, + "learning_rate": 0.0003439678951647115, + "loss": 0.4988, + "step": 609 + }, + { + "epoch": 0.26487190620929224, + "grad_norm": 0.043251313269138336, + "learning_rate": 0.0003424458909881897, + "loss": 0.5319, + "step": 610 + }, + { + "epoch": 0.2653061224489796, + "grad_norm": 0.045615166425704956, + "learning_rate": 0.00034092550592390857, + "loss": 0.4668, + "step": 611 + }, + { + "epoch": 0.265740338688667, + "grad_norm": 0.03943486511707306, + "learning_rate": 0.00033940675559617726, + "loss": 0.4991, + "step": 612 + }, + { + "epoch": 0.26617455492835435, + "grad_norm": 0.0382043793797493, + "learning_rate": 0.000337889655612505, + "loss": 0.5117, + "step": 613 + }, + { + "epoch": 0.26660877116804166, + "grad_norm": 0.039549313485622406, + "learning_rate": 0.00033637422156344155, + "loss": 0.5029, + "step": 614 + }, + { + "epoch": 0.267042987407729, + "grad_norm": 0.03972559794783592, + "learning_rate": 0.00033486046902241664, + "loss": 0.5073, + "step": 615 + }, + { + "epoch": 0.2674772036474164, + "grad_norm": 0.03974078595638275, + "learning_rate": 0.0003333484135455792, + "loss": 0.5083, + "step": 616 + }, + { + "epoch": 0.26791141988710376, + "grad_norm": 0.039242375642061234, + "learning_rate": 0.00033183807067163916, + "loss": 0.4979, + "step": 617 + }, + { + "epoch": 0.26834563612679113, + "grad_norm": 0.045767784118652344, + "learning_rate": 0.0003303294559217063, + "loss": 0.4835, + "step": 618 + }, + { + "epoch": 0.2687798523664785, + "grad_norm": 0.03927744925022125, + "learning_rate": 0.0003288225847991312, + "loss": 0.5103, + "step": 619 + }, + { + "epoch": 0.26921406860616587, + "grad_norm": 0.03815273195505142, + "learning_rate": 0.0003273174727893463, + "loss": 0.5097, + "step": 620 + }, + { + "epoch": 0.26964828484585324, + "grad_norm": 0.0421292670071125, + "learning_rate": 0.00032581413535970593, + "loss": 0.4996, + "step": 621 + }, + { + "epoch": 0.2700825010855406, + "grad_norm": 0.0482180081307888, + "learning_rate": 0.0003243125879593286, + "loss": 0.4785, + "step": 622 + }, + { + "epoch": 0.270516717325228, + "grad_norm": 0.03935736045241356, + "learning_rate": 0.0003228128460189368, + "loss": 0.5063, + "step": 623 + }, + { + "epoch": 0.27095093356491534, + "grad_norm": 0.037722837179899216, + "learning_rate": 0.0003213149249506997, + "loss": 0.4979, + "step": 624 + }, + { + "epoch": 0.2713851498046027, + "grad_norm": 0.03764335811138153, + "learning_rate": 0.0003198188401480734, + "loss": 0.4818, + "step": 625 + }, + { + "epoch": 0.2718193660442901, + "grad_norm": 0.04136700555682182, + "learning_rate": 0.0003183246069856443, + "loss": 0.5144, + "step": 626 + }, + { + "epoch": 0.27225358228397745, + "grad_norm": 0.04152214527130127, + "learning_rate": 0.00031683224081897, + "loss": 0.5217, + "step": 627 + }, + { + "epoch": 0.27268779852366476, + "grad_norm": 0.036686960607767105, + "learning_rate": 0.00031534175698442194, + "loss": 0.4938, + "step": 628 + }, + { + "epoch": 0.27312201476335213, + "grad_norm": 0.0376245342195034, + "learning_rate": 0.00031385317079902743, + "loss": 0.5255, + "step": 629 + }, + { + "epoch": 0.2735562310030395, + "grad_norm": 0.04143936559557915, + "learning_rate": 0.000312366497560313, + "loss": 0.5005, + "step": 630 + }, + { + "epoch": 0.27399044724272686, + "grad_norm": 0.036747369915246964, + "learning_rate": 0.00031088175254614616, + "loss": 0.4621, + "step": 631 + }, + { + "epoch": 0.27442466348241423, + "grad_norm": 0.039392732083797455, + "learning_rate": 0.00030939895101457916, + "loss": 0.5333, + "step": 632 + }, + { + "epoch": 0.2748588797221016, + "grad_norm": 0.03783806785941124, + "learning_rate": 0.0003079181082036922, + "loss": 0.5075, + "step": 633 + }, + { + "epoch": 0.27529309596178897, + "grad_norm": 0.03799246624112129, + "learning_rate": 0.000306439239331436, + "loss": 0.5086, + "step": 634 + }, + { + "epoch": 0.27572731220147634, + "grad_norm": 0.040941022336483, + "learning_rate": 0.0003049623595954766, + "loss": 0.5125, + "step": 635 + }, + { + "epoch": 0.2761615284411637, + "grad_norm": 0.045769453048706055, + "learning_rate": 0.0003034874841730382, + "loss": 0.4913, + "step": 636 + }, + { + "epoch": 0.2765957446808511, + "grad_norm": 0.044418562203645706, + "learning_rate": 0.00030201462822074786, + "loss": 0.5225, + "step": 637 + }, + { + "epoch": 0.27702996092053844, + "grad_norm": 0.036289479583501816, + "learning_rate": 0.0003005438068744792, + "loss": 0.5016, + "step": 638 + }, + { + "epoch": 0.2774641771602258, + "grad_norm": 0.03927507996559143, + "learning_rate": 0.0002990750352491973, + "loss": 0.497, + "step": 639 + }, + { + "epoch": 0.2778983933999132, + "grad_norm": 0.052069611847400665, + "learning_rate": 0.0002976083284388031, + "loss": 0.5026, + "step": 640 + }, + { + "epoch": 0.27833260963960055, + "grad_norm": 0.0408078134059906, + "learning_rate": 0.00029614370151597835, + "loss": 0.5069, + "step": 641 + }, + { + "epoch": 0.27876682587928786, + "grad_norm": 0.040615539997816086, + "learning_rate": 0.0002946811695320311, + "loss": 0.4878, + "step": 642 + }, + { + "epoch": 0.27920104211897523, + "grad_norm": 0.04145561158657074, + "learning_rate": 0.00029322074751673977, + "loss": 0.4547, + "step": 643 + }, + { + "epoch": 0.2796352583586626, + "grad_norm": 0.044274091720581055, + "learning_rate": 0.00029176245047820063, + "loss": 0.5051, + "step": 644 + }, + { + "epoch": 0.28006947459834997, + "grad_norm": 0.055839963257312775, + "learning_rate": 0.0002903062934026716, + "loss": 0.495, + "step": 645 + }, + { + "epoch": 0.28050369083803733, + "grad_norm": 0.038132019340991974, + "learning_rate": 0.0002888522912544202, + "loss": 0.4746, + "step": 646 + }, + { + "epoch": 0.2809379070777247, + "grad_norm": 0.039124827831983566, + "learning_rate": 0.00028740045897556767, + "loss": 0.4978, + "step": 647 + }, + { + "epoch": 0.28137212331741207, + "grad_norm": 0.03960909694433212, + "learning_rate": 0.0002859508114859374, + "loss": 0.5043, + "step": 648 + }, + { + "epoch": 0.28180633955709944, + "grad_norm": 0.03705562651157379, + "learning_rate": 0.00028450336368289974, + "loss": 0.5041, + "step": 649 + }, + { + "epoch": 0.2822405557967868, + "grad_norm": 0.03920963406562805, + "learning_rate": 0.00028305813044122096, + "loss": 0.5055, + "step": 650 + }, + { + "epoch": 0.2826747720364742, + "grad_norm": 0.04196755215525627, + "learning_rate": 0.00028161512661290845, + "loss": 0.4962, + "step": 651 + }, + { + "epoch": 0.28310898827616154, + "grad_norm": 0.03515279293060303, + "learning_rate": 0.00028017436702705903, + "loss": 0.4708, + "step": 652 + }, + { + "epoch": 0.2835432045158489, + "grad_norm": 0.042391590774059296, + "learning_rate": 0.0002787358664897068, + "loss": 0.5268, + "step": 653 + }, + { + "epoch": 0.2839774207555363, + "grad_norm": 0.03707970678806305, + "learning_rate": 0.0002772996397836704, + "loss": 0.5252, + "step": 654 + }, + { + "epoch": 0.28441163699522365, + "grad_norm": 0.039679594337940216, + "learning_rate": 0.0002758657016684015, + "loss": 0.4954, + "step": 655 + }, + { + "epoch": 0.28484585323491096, + "grad_norm": 0.03491951525211334, + "learning_rate": 0.0002744340668798326, + "loss": 0.4859, + "step": 656 + }, + { + "epoch": 0.28528006947459833, + "grad_norm": 0.033995699137449265, + "learning_rate": 0.00027300475013022663, + "loss": 0.4917, + "step": 657 + }, + { + "epoch": 0.2857142857142857, + "grad_norm": 0.0408964566886425, + "learning_rate": 0.00027157776610802414, + "loss": 0.5176, + "step": 658 + }, + { + "epoch": 0.28614850195397307, + "grad_norm": 0.035840053111314774, + "learning_rate": 0.0002701531294776943, + "loss": 0.4757, + "step": 659 + }, + { + "epoch": 0.28658271819366044, + "grad_norm": 0.03554424270987511, + "learning_rate": 0.0002687308548795825, + "loss": 0.5043, + "step": 660 + }, + { + "epoch": 0.2870169344333478, + "grad_norm": 0.03549895063042641, + "learning_rate": 0.00026731095692976073, + "loss": 0.4924, + "step": 661 + }, + { + "epoch": 0.28745115067303517, + "grad_norm": 0.0438251756131649, + "learning_rate": 0.00026589345021987723, + "loss": 0.4986, + "step": 662 + }, + { + "epoch": 0.28788536691272254, + "grad_norm": 0.037839341908693314, + "learning_rate": 0.00026447834931700686, + "loss": 0.4716, + "step": 663 + }, + { + "epoch": 0.2883195831524099, + "grad_norm": 0.03682737424969673, + "learning_rate": 0.0002630656687635007, + "loss": 0.4991, + "step": 664 + }, + { + "epoch": 0.2887537993920973, + "grad_norm": 0.03697109594941139, + "learning_rate": 0.0002616554230768374, + "loss": 0.5104, + "step": 665 + }, + { + "epoch": 0.28918801563178465, + "grad_norm": 0.036725230515003204, + "learning_rate": 0.0002602476267494731, + "loss": 0.5163, + "step": 666 + }, + { + "epoch": 0.289622231871472, + "grad_norm": 0.04177143797278404, + "learning_rate": 0.0002588422942486932, + "loss": 0.4567, + "step": 667 + }, + { + "epoch": 0.2900564481111594, + "grad_norm": 0.04078909382224083, + "learning_rate": 0.0002574394400164639, + "loss": 0.4946, + "step": 668 + }, + { + "epoch": 0.2904906643508467, + "grad_norm": 0.04146347567439079, + "learning_rate": 0.0002560390784692828, + "loss": 0.4947, + "step": 669 + }, + { + "epoch": 0.29092488059053406, + "grad_norm": 0.03871840611100197, + "learning_rate": 0.00025464122399803123, + "loss": 0.5238, + "step": 670 + }, + { + "epoch": 0.29135909683022143, + "grad_norm": 0.04148663207888603, + "learning_rate": 0.00025324589096782657, + "loss": 0.5246, + "step": 671 + }, + { + "epoch": 0.2917933130699088, + "grad_norm": 0.03910430520772934, + "learning_rate": 0.0002518530937178751, + "loss": 0.4846, + "step": 672 + }, + { + "epoch": 0.29222752930959617, + "grad_norm": 0.036824680864810944, + "learning_rate": 0.000250462846561323, + "loss": 0.4654, + "step": 673 + }, + { + "epoch": 0.29266174554928354, + "grad_norm": 0.036250337958335876, + "learning_rate": 0.00024907516378511136, + "loss": 0.5106, + "step": 674 + }, + { + "epoch": 0.2930959617889709, + "grad_norm": 0.03372564911842346, + "learning_rate": 0.00024769005964982713, + "loss": 0.5222, + "step": 675 + }, + { + "epoch": 0.2935301780286583, + "grad_norm": 0.03601228445768356, + "learning_rate": 0.000246307548389559, + "loss": 0.5099, + "step": 676 + }, + { + "epoch": 0.29396439426834564, + "grad_norm": 0.039937473833560944, + "learning_rate": 0.0002449276442117486, + "loss": 0.5262, + "step": 677 + }, + { + "epoch": 0.294398610508033, + "grad_norm": 0.03719143569469452, + "learning_rate": 0.000243550361297047, + "loss": 0.5089, + "step": 678 + }, + { + "epoch": 0.2948328267477204, + "grad_norm": 0.034734416753053665, + "learning_rate": 0.00024217571379916668, + "loss": 0.4962, + "step": 679 + }, + { + "epoch": 0.29526704298740775, + "grad_norm": 0.03447417914867401, + "learning_rate": 0.00024080371584473748, + "loss": 0.4952, + "step": 680 + }, + { + "epoch": 0.2957012592270951, + "grad_norm": 0.03631613776087761, + "learning_rate": 0.00023943438153316155, + "loss": 0.497, + "step": 681 + }, + { + "epoch": 0.2961354754667825, + "grad_norm": 0.03350459039211273, + "learning_rate": 0.00023806772493646723, + "loss": 0.4775, + "step": 682 + }, + { + "epoch": 0.2965696917064698, + "grad_norm": 0.032677456736564636, + "learning_rate": 0.00023670376009916595, + "loss": 0.5001, + "step": 683 + }, + { + "epoch": 0.29700390794615716, + "grad_norm": 0.03544562682509422, + "learning_rate": 0.00023534250103810628, + "loss": 0.4796, + "step": 684 + }, + { + "epoch": 0.29743812418584453, + "grad_norm": 0.03991761803627014, + "learning_rate": 0.00023398396174233177, + "loss": 0.5307, + "step": 685 + }, + { + "epoch": 0.2978723404255319, + "grad_norm": 0.037654612213373184, + "learning_rate": 0.00023262815617293515, + "loss": 0.4727, + "step": 686 + }, + { + "epoch": 0.29830655666521927, + "grad_norm": 0.03447496145963669, + "learning_rate": 0.00023127509826291698, + "loss": 0.5037, + "step": 687 + }, + { + "epoch": 0.29874077290490664, + "grad_norm": 0.03626209869980812, + "learning_rate": 0.00022992480191704002, + "loss": 0.4758, + "step": 688 + }, + { + "epoch": 0.299174989144594, + "grad_norm": 0.03519487380981445, + "learning_rate": 0.000228577281011689, + "loss": 0.5046, + "step": 689 + }, + { + "epoch": 0.2996092053842814, + "grad_norm": 0.036514561623334885, + "learning_rate": 0.0002272325493947257, + "loss": 0.5014, + "step": 690 + }, + { + "epoch": 0.30004342162396874, + "grad_norm": 0.036256395280361176, + "learning_rate": 0.00022589062088534834, + "loss": 0.5311, + "step": 691 + }, + { + "epoch": 0.3004776378636561, + "grad_norm": 0.036770399659872055, + "learning_rate": 0.0002245515092739488, + "loss": 0.4922, + "step": 692 + }, + { + "epoch": 0.3009118541033435, + "grad_norm": 0.037295542657375336, + "learning_rate": 0.00022321522832197032, + "loss": 0.4817, + "step": 693 + }, + { + "epoch": 0.30134607034303085, + "grad_norm": 0.03913348540663719, + "learning_rate": 0.00022188179176176764, + "loss": 0.4565, + "step": 694 + }, + { + "epoch": 0.3017802865827182, + "grad_norm": 0.04223249852657318, + "learning_rate": 0.00022055121329646416, + "loss": 0.5297, + "step": 695 + }, + { + "epoch": 0.3022145028224056, + "grad_norm": 0.03583509102463722, + "learning_rate": 0.0002192235065998126, + "loss": 0.4808, + "step": 696 + }, + { + "epoch": 0.3026487190620929, + "grad_norm": 0.04437342286109924, + "learning_rate": 0.0002178986853160535, + "loss": 0.4794, + "step": 697 + }, + { + "epoch": 0.30308293530178027, + "grad_norm": 0.043945055454969406, + "learning_rate": 0.0002165767630597752, + "loss": 0.4959, + "step": 698 + }, + { + "epoch": 0.30351715154146763, + "grad_norm": 0.04786218702793121, + "learning_rate": 0.00021525775341577403, + "loss": 0.5047, + "step": 699 + }, + { + "epoch": 0.303951367781155, + "grad_norm": 0.041139792650938034, + "learning_rate": 0.0002139416699389153, + "loss": 0.5134, + "step": 700 + }, + { + "epoch": 0.30438558402084237, + "grad_norm": 0.034917134791612625, + "learning_rate": 0.00021262852615399258, + "loss": 0.4924, + "step": 701 + }, + { + "epoch": 0.30481980026052974, + "grad_norm": 0.03301383554935455, + "learning_rate": 0.0002113183355555904, + "loss": 0.4862, + "step": 702 + }, + { + "epoch": 0.3052540165002171, + "grad_norm": 0.0419352687895298, + "learning_rate": 0.00021001111160794383, + "loss": 0.5147, + "step": 703 + }, + { + "epoch": 0.3056882327399045, + "grad_norm": 0.03364208713173866, + "learning_rate": 0.00020870686774480197, + "loss": 0.5236, + "step": 704 + }, + { + "epoch": 0.30612244897959184, + "grad_norm": 0.03912827745079994, + "learning_rate": 0.0002074056173692881, + "loss": 0.5042, + "step": 705 + }, + { + "epoch": 0.3065566652192792, + "grad_norm": 0.03657132387161255, + "learning_rate": 0.00020610737385376348, + "loss": 0.5011, + "step": 706 + }, + { + "epoch": 0.3069908814589666, + "grad_norm": 0.03518640622496605, + "learning_rate": 0.00020481215053968872, + "loss": 0.4798, + "step": 707 + }, + { + "epoch": 0.30742509769865395, + "grad_norm": 0.03507522493600845, + "learning_rate": 0.00020351996073748714, + "loss": 0.5313, + "step": 708 + }, + { + "epoch": 0.3078593139383413, + "grad_norm": 0.03609062731266022, + "learning_rate": 0.00020223081772640866, + "loss": 0.4785, + "step": 709 + }, + { + "epoch": 0.3082935301780287, + "grad_norm": 0.03479000926017761, + "learning_rate": 0.000200944734754392, + "loss": 0.4835, + "step": 710 + }, + { + "epoch": 0.308727746417716, + "grad_norm": 0.044929634779691696, + "learning_rate": 0.00019966172503792985, + "loss": 0.5122, + "step": 711 + }, + { + "epoch": 0.30916196265740337, + "grad_norm": 0.03376319631934166, + "learning_rate": 0.00019838180176193177, + "loss": 0.4876, + "step": 712 + }, + { + "epoch": 0.30959617889709073, + "grad_norm": 0.03246736153960228, + "learning_rate": 0.0001971049780795901, + "loss": 0.4804, + "step": 713 + }, + { + "epoch": 0.3100303951367781, + "grad_norm": 0.031609781086444855, + "learning_rate": 0.00019583126711224342, + "loss": 0.545, + "step": 714 + }, + { + "epoch": 0.31046461137646547, + "grad_norm": 0.034195881336927414, + "learning_rate": 0.00019456068194924288, + "loss": 0.5096, + "step": 715 + }, + { + "epoch": 0.31089882761615284, + "grad_norm": 0.03344978764653206, + "learning_rate": 0.0001932932356478168, + "loss": 0.5079, + "step": 716 + }, + { + "epoch": 0.3113330438558402, + "grad_norm": 0.03219503536820412, + "learning_rate": 0.00019202894123293674, + "loss": 0.5339, + "step": 717 + }, + { + "epoch": 0.3117672600955276, + "grad_norm": 0.03737034648656845, + "learning_rate": 0.00019076781169718426, + "loss": 0.5131, + "step": 718 + }, + { + "epoch": 0.31220147633521494, + "grad_norm": 0.03394628316164017, + "learning_rate": 0.00018950986000061638, + "loss": 0.4969, + "step": 719 + }, + { + "epoch": 0.3126356925749023, + "grad_norm": 0.032426606863737106, + "learning_rate": 0.00018825509907063325, + "loss": 0.49, + "step": 720 + }, + { + "epoch": 0.3130699088145897, + "grad_norm": 0.03836345672607422, + "learning_rate": 0.00018700354180184464, + "loss": 0.4925, + "step": 721 + }, + { + "epoch": 0.31350412505427705, + "grad_norm": 0.035867128521203995, + "learning_rate": 0.0001857552010559382, + "loss": 0.5466, + "step": 722 + }, + { + "epoch": 0.3139383412939644, + "grad_norm": 0.03811186924576759, + "learning_rate": 0.0001845100896615462, + "loss": 0.4825, + "step": 723 + }, + { + "epoch": 0.31437255753365173, + "grad_norm": 0.032081685960292816, + "learning_rate": 0.00018326822041411523, + "loss": 0.4858, + "step": 724 + }, + { + "epoch": 0.3148067737733391, + "grad_norm": 0.03794045001268387, + "learning_rate": 0.00018202960607577247, + "loss": 0.5023, + "step": 725 + }, + { + "epoch": 0.31524099001302647, + "grad_norm": 0.033024583011865616, + "learning_rate": 0.00018079425937519728, + "loss": 0.4744, + "step": 726 + }, + { + "epoch": 0.31567520625271384, + "grad_norm": 0.03800756111741066, + "learning_rate": 0.00017956219300748795, + "loss": 0.4813, + "step": 727 + }, + { + "epoch": 0.3161094224924012, + "grad_norm": 0.03822485730051994, + "learning_rate": 0.0001783334196340331, + "loss": 0.4984, + "step": 728 + }, + { + "epoch": 0.3165436387320886, + "grad_norm": 0.03531699255108833, + "learning_rate": 0.00017710795188237988, + "loss": 0.5095, + "step": 729 + }, + { + "epoch": 0.31697785497177594, + "grad_norm": 0.03231954947113991, + "learning_rate": 0.0001758858023461059, + "loss": 0.4795, + "step": 730 + }, + { + "epoch": 0.3174120712114633, + "grad_norm": 0.04151546210050583, + "learning_rate": 0.00017466698358468825, + "loss": 0.4852, + "step": 731 + }, + { + "epoch": 0.3178462874511507, + "grad_norm": 0.03277713060379028, + "learning_rate": 0.00017345150812337563, + "loss": 0.4736, + "step": 732 + }, + { + "epoch": 0.31828050369083805, + "grad_norm": 0.03749159723520279, + "learning_rate": 0.0001722393884530593, + "loss": 0.5042, + "step": 733 + }, + { + "epoch": 0.3187147199305254, + "grad_norm": 0.029695043340325356, + "learning_rate": 0.00017103063703014372, + "loss": 0.511, + "step": 734 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.033079251646995544, + "learning_rate": 0.00016982526627642042, + "loss": 0.4871, + "step": 735 + }, + { + "epoch": 0.31958315240990015, + "grad_norm": 0.03277754411101341, + "learning_rate": 0.00016862328857893854, + "loss": 0.4805, + "step": 736 + }, + { + "epoch": 0.3200173686495875, + "grad_norm": 0.031276337802410126, + "learning_rate": 0.00016742471628987892, + "loss": 0.5195, + "step": 737 + }, + { + "epoch": 0.32045158488927483, + "grad_norm": 0.031646616756916046, + "learning_rate": 0.000166229561726426, + "loss": 0.4899, + "step": 738 + }, + { + "epoch": 0.3208858011289622, + "grad_norm": 0.03364792838692665, + "learning_rate": 0.00016503783717064247, + "loss": 0.5011, + "step": 739 + }, + { + "epoch": 0.32132001736864957, + "grad_norm": 0.032123491168022156, + "learning_rate": 0.00016384955486934156, + "loss": 0.5083, + "step": 740 + }, + { + "epoch": 0.32175423360833694, + "grad_norm": 0.03388531878590584, + "learning_rate": 0.00016266472703396284, + "loss": 0.5127, + "step": 741 + }, + { + "epoch": 0.3221884498480243, + "grad_norm": 0.03697797656059265, + "learning_rate": 0.00016148336584044537, + "loss": 0.5167, + "step": 742 + }, + { + "epoch": 0.3226226660877117, + "grad_norm": 0.03302115574479103, + "learning_rate": 0.000160305483429103, + "loss": 0.483, + "step": 743 + }, + { + "epoch": 0.32305688232739904, + "grad_norm": 0.03180578723549843, + "learning_rate": 0.0001591310919045003, + "loss": 0.5144, + "step": 744 + }, + { + "epoch": 0.3234910985670864, + "grad_norm": 0.033885449171066284, + "learning_rate": 0.00015796020333532697, + "loss": 0.484, + "step": 745 + }, + { + "epoch": 0.3239253148067738, + "grad_norm": 0.03529027849435806, + "learning_rate": 0.0001567928297542749, + "loss": 0.4967, + "step": 746 + }, + { + "epoch": 0.32435953104646115, + "grad_norm": 0.030900994315743446, + "learning_rate": 0.00015562898315791353, + "loss": 0.5177, + "step": 747 + }, + { + "epoch": 0.3247937472861485, + "grad_norm": 0.03337928652763367, + "learning_rate": 0.00015446867550656767, + "loss": 0.4902, + "step": 748 + }, + { + "epoch": 0.3252279635258359, + "grad_norm": 0.028887873515486717, + "learning_rate": 0.00015331191872419348, + "loss": 0.4841, + "step": 749 + }, + { + "epoch": 0.32566217976552325, + "grad_norm": 0.03361840918660164, + "learning_rate": 0.0001521587246982568, + "loss": 0.5171, + "step": 750 + }, + { + "epoch": 0.3260963960052106, + "grad_norm": 0.030776534229516983, + "learning_rate": 0.00015100910527961049, + "loss": 0.4592, + "step": 751 + }, + { + "epoch": 0.32653061224489793, + "grad_norm": 0.028537072241306305, + "learning_rate": 0.00014986307228237266, + "loss": 0.4894, + "step": 752 + }, + { + "epoch": 0.3269648284845853, + "grad_norm": 0.02929234690964222, + "learning_rate": 0.00014872063748380543, + "loss": 0.4832, + "step": 753 + }, + { + "epoch": 0.32739904472427267, + "grad_norm": 0.02816353552043438, + "learning_rate": 0.00014758181262419424, + "loss": 0.4788, + "step": 754 + }, + { + "epoch": 0.32783326096396004, + "grad_norm": 0.03165418654680252, + "learning_rate": 0.00014644660940672628, + "loss": 0.5128, + "step": 755 + }, + { + "epoch": 0.3282674772036474, + "grad_norm": 0.0279870443046093, + "learning_rate": 0.00014531503949737106, + "loss": 0.4966, + "step": 756 + }, + { + "epoch": 0.3287016934433348, + "grad_norm": 0.029440978541970253, + "learning_rate": 0.00014418711452476046, + "loss": 0.4785, + "step": 757 + }, + { + "epoch": 0.32913590968302214, + "grad_norm": 0.032016150653362274, + "learning_rate": 0.00014306284608006837, + "loss": 0.4813, + "step": 758 + }, + { + "epoch": 0.3295701259227095, + "grad_norm": 0.029074901714920998, + "learning_rate": 0.00014194224571689284, + "loss": 0.5165, + "step": 759 + }, + { + "epoch": 0.3300043421623969, + "grad_norm": 0.030839256942272186, + "learning_rate": 0.00014082532495113625, + "loss": 0.4615, + "step": 760 + }, + { + "epoch": 0.33043855840208425, + "grad_norm": 0.03017420321702957, + "learning_rate": 0.00013971209526088762, + "loss": 0.473, + "step": 761 + }, + { + "epoch": 0.3308727746417716, + "grad_norm": 0.03192548081278801, + "learning_rate": 0.00013860256808630427, + "loss": 0.5315, + "step": 762 + }, + { + "epoch": 0.331306990881459, + "grad_norm": 0.029673421755433083, + "learning_rate": 0.00013749675482949486, + "loss": 0.4978, + "step": 763 + }, + { + "epoch": 0.33174120712114635, + "grad_norm": 0.035233914852142334, + "learning_rate": 0.00013639466685440134, + "loss": 0.5174, + "step": 764 + }, + { + "epoch": 0.3321754233608337, + "grad_norm": 0.03196287900209427, + "learning_rate": 0.00013529631548668298, + "loss": 0.5121, + "step": 765 + }, + { + "epoch": 0.33260963960052103, + "grad_norm": 0.028510358184576035, + "learning_rate": 0.0001342017120135993, + "loss": 0.4629, + "step": 766 + }, + { + "epoch": 0.3330438558402084, + "grad_norm": 0.032123226672410965, + "learning_rate": 0.0001331108676838948, + "loss": 0.5538, + "step": 767 + }, + { + "epoch": 0.33347807207989577, + "grad_norm": 0.0301981084048748, + "learning_rate": 0.00013202379370768252, + "loss": 0.5278, + "step": 768 + }, + { + "epoch": 0.33391228831958314, + "grad_norm": 0.029721124097704887, + "learning_rate": 0.00013094050125632973, + "loss": 0.5353, + "step": 769 + }, + { + "epoch": 0.3343465045592705, + "grad_norm": 0.029476812109351158, + "learning_rate": 0.00012986100146234231, + "loss": 0.488, + "step": 770 + }, + { + "epoch": 0.3347807207989579, + "grad_norm": 0.029208241030573845, + "learning_rate": 0.00012878530541925076, + "loss": 0.5384, + "step": 771 + }, + { + "epoch": 0.33521493703864524, + "grad_norm": 0.029901737347245216, + "learning_rate": 0.00012771342418149657, + "loss": 0.4675, + "step": 772 + }, + { + "epoch": 0.3356491532783326, + "grad_norm": 0.030133001506328583, + "learning_rate": 0.00012664536876431755, + "loss": 0.5175, + "step": 773 + }, + { + "epoch": 0.33608336951802, + "grad_norm": 0.028815865516662598, + "learning_rate": 0.0001255811501436359, + "loss": 0.5012, + "step": 774 + }, + { + "epoch": 0.33651758575770735, + "grad_norm": 0.02868812158703804, + "learning_rate": 0.00012452077925594434, + "loss": 0.521, + "step": 775 + }, + { + "epoch": 0.3369518019973947, + "grad_norm": 0.035325054079294205, + "learning_rate": 0.00012346426699819457, + "loss": 0.5075, + "step": 776 + }, + { + "epoch": 0.3373860182370821, + "grad_norm": 0.03588160499930382, + "learning_rate": 0.00012241162422768444, + "loss": 0.5091, + "step": 777 + }, + { + "epoch": 0.33782023447676945, + "grad_norm": 0.03347684070467949, + "learning_rate": 0.00012136286176194744, + "loss": 0.5039, + "step": 778 + }, + { + "epoch": 0.3382544507164568, + "grad_norm": 0.03358198329806328, + "learning_rate": 0.00012031799037864011, + "loss": 0.5014, + "step": 779 + }, + { + "epoch": 0.33868866695614414, + "grad_norm": 0.02983028069138527, + "learning_rate": 0.0001192770208154328, + "loss": 0.4823, + "step": 780 + }, + { + "epoch": 0.3391228831958315, + "grad_norm": 0.028506051748991013, + "learning_rate": 0.00011823996376989849, + "loss": 0.5135, + "step": 781 + }, + { + "epoch": 0.33955709943551887, + "grad_norm": 0.031358085572719574, + "learning_rate": 0.00011720682989940262, + "loss": 0.4677, + "step": 782 + }, + { + "epoch": 0.33999131567520624, + "grad_norm": 0.031006425619125366, + "learning_rate": 0.00011617762982099444, + "loss": 0.5188, + "step": 783 + }, + { + "epoch": 0.3404255319148936, + "grad_norm": 0.0309920571744442, + "learning_rate": 0.00011515237411129698, + "loss": 0.5087, + "step": 784 + }, + { + "epoch": 0.340859748154581, + "grad_norm": 0.027928480878472328, + "learning_rate": 0.0001141310733063991, + "loss": 0.5005, + "step": 785 + }, + { + "epoch": 0.34129396439426835, + "grad_norm": 0.028440937399864197, + "learning_rate": 0.00011311373790174655, + "loss": 0.5055, + "step": 786 + }, + { + "epoch": 0.3417281806339557, + "grad_norm": 0.03135620057582855, + "learning_rate": 0.00011210037835203507, + "loss": 0.4947, + "step": 787 + }, + { + "epoch": 0.3421623968736431, + "grad_norm": 0.029479345306754112, + "learning_rate": 0.00011109100507110132, + "loss": 0.5048, + "step": 788 + }, + { + "epoch": 0.34259661311333045, + "grad_norm": 0.030894558876752853, + "learning_rate": 0.00011008562843181796, + "loss": 0.4726, + "step": 789 + }, + { + "epoch": 0.3430308293530178, + "grad_norm": 0.03545878455042839, + "learning_rate": 0.0001090842587659851, + "loss": 0.5165, + "step": 790 + }, + { + "epoch": 0.3434650455927052, + "grad_norm": 0.028414282947778702, + "learning_rate": 0.00010808690636422585, + "loss": 0.5008, + "step": 791 + }, + { + "epoch": 0.34389926183239256, + "grad_norm": 0.029046185314655304, + "learning_rate": 0.00010709358147587883, + "loss": 0.5067, + "step": 792 + }, + { + "epoch": 0.34433347807207987, + "grad_norm": 0.029067449271678925, + "learning_rate": 0.00010610429430889452, + "loss": 0.5039, + "step": 793 + }, + { + "epoch": 0.34476769431176724, + "grad_norm": 0.03735330328345299, + "learning_rate": 0.00010511905502972886, + "loss": 0.4874, + "step": 794 + }, + { + "epoch": 0.3452019105514546, + "grad_norm": 0.025629056617617607, + "learning_rate": 0.00010413787376324018, + "loss": 0.4743, + "step": 795 + }, + { + "epoch": 0.345636126791142, + "grad_norm": 0.03742838650941849, + "learning_rate": 0.00010316076059258389, + "loss": 0.4706, + "step": 796 + }, + { + "epoch": 0.34607034303082934, + "grad_norm": 0.028506414964795113, + "learning_rate": 0.00010218772555910954, + "loss": 0.4677, + "step": 797 + }, + { + "epoch": 0.3465045592705167, + "grad_norm": 0.028791414573788643, + "learning_rate": 0.00010121877866225782, + "loss": 0.4786, + "step": 798 + }, + { + "epoch": 0.3469387755102041, + "grad_norm": 0.027565713971853256, + "learning_rate": 0.00010025392985945703, + "loss": 0.4756, + "step": 799 + }, + { + "epoch": 0.34737299174989145, + "grad_norm": 0.037437453866004944, + "learning_rate": 9.929318906602175e-05, + "loss": 0.4941, + "step": 800 + }, + { + "epoch": 0.3478072079895788, + "grad_norm": 0.028768004849553108, + "learning_rate": 9.833656615504976e-05, + "loss": 0.5263, + "step": 801 + }, + { + "epoch": 0.3482414242292662, + "grad_norm": 0.030494874343276024, + "learning_rate": 9.738407095732193e-05, + "loss": 0.4933, + "step": 802 + }, + { + "epoch": 0.34867564046895355, + "grad_norm": 0.02905452810227871, + "learning_rate": 9.643571326119983e-05, + "loss": 0.4751, + "step": 803 + }, + { + "epoch": 0.3491098567086409, + "grad_norm": 0.02821405418217182, + "learning_rate": 9.549150281252633e-05, + "loss": 0.4874, + "step": 804 + }, + { + "epoch": 0.3495440729483283, + "grad_norm": 0.02889527939260006, + "learning_rate": 9.455144931452459e-05, + "loss": 0.4732, + "step": 805 + }, + { + "epoch": 0.34997828918801566, + "grad_norm": 0.02988606132566929, + "learning_rate": 9.36155624276987e-05, + "loss": 0.5044, + "step": 806 + }, + { + "epoch": 0.35041250542770297, + "grad_norm": 0.03306758031249046, + "learning_rate": 9.26838517697346e-05, + "loss": 0.492, + "step": 807 + }, + { + "epoch": 0.35084672166739034, + "grad_norm": 0.028606578707695007, + "learning_rate": 9.175632691540064e-05, + "loss": 0.4696, + "step": 808 + }, + { + "epoch": 0.3512809379070777, + "grad_norm": 0.027804942801594734, + "learning_rate": 9.083299739645007e-05, + "loss": 0.504, + "step": 809 + }, + { + "epoch": 0.3517151541467651, + "grad_norm": 0.02756122685968876, + "learning_rate": 8.991387270152201e-05, + "loss": 0.5129, + "step": 810 + }, + { + "epoch": 0.35214937038645244, + "grad_norm": 0.03051147237420082, + "learning_rate": 8.899896227604509e-05, + "loss": 0.5054, + "step": 811 + }, + { + "epoch": 0.3525835866261398, + "grad_norm": 0.032840099185705185, + "learning_rate": 8.808827552213916e-05, + "loss": 0.4926, + "step": 812 + }, + { + "epoch": 0.3530178028658272, + "grad_norm": 0.026586757972836494, + "learning_rate": 8.718182179851997e-05, + "loss": 0.4992, + "step": 813 + }, + { + "epoch": 0.35345201910551455, + "grad_norm": 0.030486930161714554, + "learning_rate": 8.627961042040184e-05, + "loss": 0.5221, + "step": 814 + }, + { + "epoch": 0.3538862353452019, + "grad_norm": 0.03037872351706028, + "learning_rate": 8.538165065940262e-05, + "loss": 0.4801, + "step": 815 + }, + { + "epoch": 0.3543204515848893, + "grad_norm": 0.027922067791223526, + "learning_rate": 8.448795174344803e-05, + "loss": 0.5011, + "step": 816 + }, + { + "epoch": 0.35475466782457665, + "grad_norm": 0.030396727845072746, + "learning_rate": 8.359852285667752e-05, + "loss": 0.4937, + "step": 817 + }, + { + "epoch": 0.355188884064264, + "grad_norm": 0.028359906747937202, + "learning_rate": 8.271337313934868e-05, + "loss": 0.4905, + "step": 818 + }, + { + "epoch": 0.3556231003039514, + "grad_norm": 0.041622817516326904, + "learning_rate": 8.183251168774475e-05, + "loss": 0.4829, + "step": 819 + }, + { + "epoch": 0.35605731654363876, + "grad_norm": 0.030735066160559654, + "learning_rate": 8.09559475540797e-05, + "loss": 0.5038, + "step": 820 + }, + { + "epoch": 0.35649153278332607, + "grad_norm": 0.030377600342035294, + "learning_rate": 8.008368974640634e-05, + "loss": 0.523, + "step": 821 + }, + { + "epoch": 0.35692574902301344, + "grad_norm": 0.027815300971269608, + "learning_rate": 7.921574722852342e-05, + "loss": 0.4758, + "step": 822 + }, + { + "epoch": 0.3573599652627008, + "grad_norm": 0.026980141177773476, + "learning_rate": 7.835212891988292e-05, + "loss": 0.5151, + "step": 823 + }, + { + "epoch": 0.3577941815023882, + "grad_norm": 0.028305258601903915, + "learning_rate": 7.749284369549953e-05, + "loss": 0.4968, + "step": 824 + }, + { + "epoch": 0.35822839774207554, + "grad_norm": 0.028862129896879196, + "learning_rate": 7.663790038585794e-05, + "loss": 0.5093, + "step": 825 + }, + { + "epoch": 0.3586626139817629, + "grad_norm": 0.031168105080723763, + "learning_rate": 7.578730777682385e-05, + "loss": 0.507, + "step": 826 + }, + { + "epoch": 0.3590968302214503, + "grad_norm": 0.02467159368097782, + "learning_rate": 7.494107460955207e-05, + "loss": 0.5031, + "step": 827 + }, + { + "epoch": 0.35953104646113765, + "grad_norm": 0.028088459745049477, + "learning_rate": 7.409920958039795e-05, + "loss": 0.5031, + "step": 828 + }, + { + "epoch": 0.359965262700825, + "grad_norm": 0.027428364381194115, + "learning_rate": 7.326172134082704e-05, + "loss": 0.5262, + "step": 829 + }, + { + "epoch": 0.3603994789405124, + "grad_norm": 0.033019062131643295, + "learning_rate": 7.242861849732695e-05, + "loss": 0.4722, + "step": 830 + }, + { + "epoch": 0.36083369518019975, + "grad_norm": 0.028655540198087692, + "learning_rate": 7.159990961131818e-05, + "loss": 0.4899, + "step": 831 + }, + { + "epoch": 0.3612679114198871, + "grad_norm": 0.02457267791032791, + "learning_rate": 7.077560319906695e-05, + "loss": 0.4989, + "step": 832 + }, + { + "epoch": 0.3617021276595745, + "grad_norm": 0.027305081486701965, + "learning_rate": 6.995570773159692e-05, + "loss": 0.4885, + "step": 833 + }, + { + "epoch": 0.36213634389926186, + "grad_norm": 0.026189139112830162, + "learning_rate": 6.914023163460248e-05, + "loss": 0.4727, + "step": 834 + }, + { + "epoch": 0.36257056013894917, + "grad_norm": 0.028178861364722252, + "learning_rate": 6.832918328836246e-05, + "loss": 0.5538, + "step": 835 + }, + { + "epoch": 0.36300477637863654, + "grad_norm": 0.02753199078142643, + "learning_rate": 6.752257102765324e-05, + "loss": 0.4991, + "step": 836 + }, + { + "epoch": 0.3634389926183239, + "grad_norm": 0.02777719311416149, + "learning_rate": 6.6720403141664e-05, + "loss": 0.4607, + "step": 837 + }, + { + "epoch": 0.3638732088580113, + "grad_norm": 0.026312079280614853, + "learning_rate": 6.592268787391076e-05, + "loss": 0.5031, + "step": 838 + }, + { + "epoch": 0.36430742509769865, + "grad_norm": 0.028296776115894318, + "learning_rate": 6.512943342215233e-05, + "loss": 0.4825, + "step": 839 + }, + { + "epoch": 0.364741641337386, + "grad_norm": 0.026355035603046417, + "learning_rate": 6.43406479383053e-05, + "loss": 0.4871, + "step": 840 + }, + { + "epoch": 0.3651758575770734, + "grad_norm": 0.025023045018315315, + "learning_rate": 6.355633952836115e-05, + "loss": 0.4797, + "step": 841 + }, + { + "epoch": 0.36561007381676075, + "grad_norm": 0.027957437559962273, + "learning_rate": 6.277651625230219e-05, + "loss": 0.5078, + "step": 842 + }, + { + "epoch": 0.3660442900564481, + "grad_norm": 0.028170911595225334, + "learning_rate": 6.200118612401917e-05, + "loss": 0.4781, + "step": 843 + }, + { + "epoch": 0.3664785062961355, + "grad_norm": 0.032493725419044495, + "learning_rate": 6.123035711122859e-05, + "loss": 0.5078, + "step": 844 + }, + { + "epoch": 0.36691272253582286, + "grad_norm": 0.029985696077346802, + "learning_rate": 6.046403713539139e-05, + "loss": 0.478, + "step": 845 + }, + { + "epoch": 0.3673469387755102, + "grad_norm": 0.03212130442261696, + "learning_rate": 5.9702234071631e-05, + "loss": 0.5146, + "step": 846 + }, + { + "epoch": 0.3677811550151976, + "grad_norm": 0.024975216016173363, + "learning_rate": 5.89449557486525e-05, + "loss": 0.4564, + "step": 847 + }, + { + "epoch": 0.36821537125488496, + "grad_norm": 0.026881275698542595, + "learning_rate": 5.8192209948662365e-05, + "loss": 0.518, + "step": 848 + }, + { + "epoch": 0.3686495874945723, + "grad_norm": 0.04243510961532593, + "learning_rate": 5.7444004407288256e-05, + "loss": 0.5238, + "step": 849 + }, + { + "epoch": 0.36908380373425964, + "grad_norm": 0.02634822018444538, + "learning_rate": 5.6700346813499945e-05, + "loss": 0.5126, + "step": 850 + }, + { + "epoch": 0.369518019973947, + "grad_norm": 0.029794232919812202, + "learning_rate": 5.5961244809529746e-05, + "loss": 0.5017, + "step": 851 + }, + { + "epoch": 0.3699522362136344, + "grad_norm": 0.028596797958016396, + "learning_rate": 5.5226705990794155e-05, + "loss": 0.508, + "step": 852 + }, + { + "epoch": 0.37038645245332175, + "grad_norm": 0.04081004112958908, + "learning_rate": 5.449673790581611e-05, + "loss": 0.4946, + "step": 853 + }, + { + "epoch": 0.3708206686930091, + "grad_norm": 0.02494877576828003, + "learning_rate": 5.377134805614714e-05, + "loss": 0.5008, + "step": 854 + }, + { + "epoch": 0.3712548849326965, + "grad_norm": 0.029261523857712746, + "learning_rate": 5.305054389629022e-05, + "loss": 0.5155, + "step": 855 + }, + { + "epoch": 0.37168910117238385, + "grad_norm": 0.026129230856895447, + "learning_rate": 5.2334332833623486e-05, + "loss": 0.4979, + "step": 856 + }, + { + "epoch": 0.3721233174120712, + "grad_norm": 0.03483575955033302, + "learning_rate": 5.162272222832348e-05, + "loss": 0.4956, + "step": 857 + }, + { + "epoch": 0.3725575336517586, + "grad_norm": 0.028554193675518036, + "learning_rate": 5.0915719393290485e-05, + "loss": 0.4816, + "step": 858 + }, + { + "epoch": 0.37299174989144596, + "grad_norm": 0.028302082791924477, + "learning_rate": 5.021333159407232e-05, + "loss": 0.4734, + "step": 859 + }, + { + "epoch": 0.3734259661311333, + "grad_norm": 0.03633783757686615, + "learning_rate": 4.9515566048790485e-05, + "loss": 0.5152, + "step": 860 + }, + { + "epoch": 0.3738601823708207, + "grad_norm": 0.025067532435059547, + "learning_rate": 4.882242992806546e-05, + "loss": 0.505, + "step": 861 + }, + { + "epoch": 0.374294398610508, + "grad_norm": 0.03426215052604675, + "learning_rate": 4.813393035494329e-05, + "loss": 0.4811, + "step": 862 + }, + { + "epoch": 0.3747286148501954, + "grad_norm": 0.024589484557509422, + "learning_rate": 4.745007440482252e-05, + "loss": 0.4881, + "step": 863 + }, + { + "epoch": 0.37516283108988274, + "grad_norm": 0.032448723912239075, + "learning_rate": 4.677086910538092e-05, + "loss": 0.5057, + "step": 864 + }, + { + "epoch": 0.3755970473295701, + "grad_norm": 0.029285568743944168, + "learning_rate": 4.6096321436503994e-05, + "loss": 0.5035, + "step": 865 + }, + { + "epoch": 0.3760312635692575, + "grad_norm": 0.02594076097011566, + "learning_rate": 4.5426438330212535e-05, + "loss": 0.4789, + "step": 866 + }, + { + "epoch": 0.37646547980894485, + "grad_norm": 0.028019646182656288, + "learning_rate": 4.476122667059207e-05, + "loss": 0.4841, + "step": 867 + }, + { + "epoch": 0.3768996960486322, + "grad_norm": 0.022888783365488052, + "learning_rate": 4.410069329372152e-05, + "loss": 0.4609, + "step": 868 + }, + { + "epoch": 0.3773339122883196, + "grad_norm": 0.025522053241729736, + "learning_rate": 4.3444844987603426e-05, + "loss": 0.5148, + "step": 869 + }, + { + "epoch": 0.37776812852800695, + "grad_norm": 0.027691571041941643, + "learning_rate": 4.279368849209381e-05, + "loss": 0.4689, + "step": 870 + }, + { + "epoch": 0.3782023447676943, + "grad_norm": 0.02531832829117775, + "learning_rate": 4.214723049883307e-05, + "loss": 0.5024, + "step": 871 + }, + { + "epoch": 0.3786365610073817, + "grad_norm": 0.025689370930194855, + "learning_rate": 4.150547765117746e-05, + "loss": 0.4839, + "step": 872 + }, + { + "epoch": 0.37907077724706906, + "grad_norm": 0.031849976629018784, + "learning_rate": 4.086843654413031e-05, + "loss": 0.4888, + "step": 873 + }, + { + "epoch": 0.3795049934867564, + "grad_norm": 0.024920670315623283, + "learning_rate": 4.0236113724274713e-05, + "loss": 0.5088, + "step": 874 + }, + { + "epoch": 0.3799392097264438, + "grad_norm": 0.03232395276427269, + "learning_rate": 3.960851568970586e-05, + "loss": 0.5039, + "step": 875 + }, + { + "epoch": 0.3803734259661311, + "grad_norm": 0.025085026398301125, + "learning_rate": 3.898564888996475e-05, + "loss": 0.4753, + "step": 876 + }, + { + "epoch": 0.3808076422058185, + "grad_norm": 0.03306965157389641, + "learning_rate": 3.83675197259713e-05, + "loss": 0.493, + "step": 877 + }, + { + "epoch": 0.38124185844550584, + "grad_norm": 0.02512296289205551, + "learning_rate": 3.7754134549959293e-05, + "loss": 0.5048, + "step": 878 + }, + { + "epoch": 0.3816760746851932, + "grad_norm": 0.02608814276754856, + "learning_rate": 3.714549966541014e-05, + "loss": 0.4893, + "step": 879 + }, + { + "epoch": 0.3821102909248806, + "grad_norm": 0.02658063732087612, + "learning_rate": 3.654162132698918e-05, + "loss": 0.5069, + "step": 880 + }, + { + "epoch": 0.38254450716456795, + "grad_norm": 0.027712546288967133, + "learning_rate": 3.594250574048058e-05, + "loss": 0.4778, + "step": 881 + }, + { + "epoch": 0.3829787234042553, + "grad_norm": 0.02457358129322529, + "learning_rate": 3.534815906272404e-05, + "loss": 0.4946, + "step": 882 + }, + { + "epoch": 0.3834129396439427, + "grad_norm": 0.032855454832315445, + "learning_rate": 3.475858740155108e-05, + "loss": 0.4796, + "step": 883 + }, + { + "epoch": 0.38384715588363005, + "grad_norm": 0.02618669532239437, + "learning_rate": 3.417379681572297e-05, + "loss": 0.4944, + "step": 884 + }, + { + "epoch": 0.3842813721233174, + "grad_norm": 0.02916988916695118, + "learning_rate": 3.3593793314867616e-05, + "loss": 0.4865, + "step": 885 + }, + { + "epoch": 0.3847155883630048, + "grad_norm": 0.02979261800646782, + "learning_rate": 3.3018582859418446e-05, + "loss": 0.5272, + "step": 886 + }, + { + "epoch": 0.38514980460269216, + "grad_norm": 0.024905268102884293, + "learning_rate": 3.2448171360552835e-05, + "loss": 0.4878, + "step": 887 + }, + { + "epoch": 0.3855840208423795, + "grad_norm": 0.025121575221419334, + "learning_rate": 3.18825646801314e-05, + "loss": 0.4887, + "step": 888 + }, + { + "epoch": 0.3860182370820669, + "grad_norm": 0.02593301795423031, + "learning_rate": 3.132176863063807e-05, + "loss": 0.5255, + "step": 889 + }, + { + "epoch": 0.3864524533217542, + "grad_norm": 0.02763954922556877, + "learning_rate": 3.076578897511978e-05, + "loss": 0.4672, + "step": 890 + }, + { + "epoch": 0.3868866695614416, + "grad_norm": 0.026684967800974846, + "learning_rate": 3.0214631427127882e-05, + "loss": 0.5277, + "step": 891 + }, + { + "epoch": 0.38732088580112894, + "grad_norm": 0.030234767124056816, + "learning_rate": 2.9668301650658758e-05, + "loss": 0.502, + "step": 892 + }, + { + "epoch": 0.3877551020408163, + "grad_norm": 0.023924171924591064, + "learning_rate": 2.9126805260096256e-05, + "loss": 0.4813, + "step": 893 + }, + { + "epoch": 0.3881893182805037, + "grad_norm": 0.03323635831475258, + "learning_rate": 2.859014782015351e-05, + "loss": 0.5422, + "step": 894 + }, + { + "epoch": 0.38862353452019105, + "grad_norm": 0.02363520860671997, + "learning_rate": 2.8058334845816213e-05, + "loss": 0.5091, + "step": 895 + }, + { + "epoch": 0.3890577507598784, + "grad_norm": 0.024630311876535416, + "learning_rate": 2.7531371802285433e-05, + "loss": 0.4642, + "step": 896 + }, + { + "epoch": 0.3894919669995658, + "grad_norm": 0.025726784020662308, + "learning_rate": 2.7009264104921606e-05, + "loss": 0.511, + "step": 897 + }, + { + "epoch": 0.38992618323925315, + "grad_norm": 0.02972118742763996, + "learning_rate": 2.6492017119189414e-05, + "loss": 0.4825, + "step": 898 + }, + { + "epoch": 0.3903603994789405, + "grad_norm": 0.024715179577469826, + "learning_rate": 2.597963616060167e-05, + "loss": 0.4879, + "step": 899 + }, + { + "epoch": 0.3907946157186279, + "grad_norm": 0.02915225736796856, + "learning_rate": 2.547212649466568e-05, + "loss": 0.4903, + "step": 900 + }, + { + "epoch": 0.39122883195831526, + "grad_norm": 0.02453678846359253, + "learning_rate": 2.4969493336828353e-05, + "loss": 0.5024, + "step": 901 + }, + { + "epoch": 0.39166304819800263, + "grad_norm": 0.025468653067946434, + "learning_rate": 2.4471741852423235e-05, + "loss": 0.4964, + "step": 902 + }, + { + "epoch": 0.39209726443769, + "grad_norm": 0.024076983332633972, + "learning_rate": 2.3978877156616784e-05, + "loss": 0.4872, + "step": 903 + }, + { + "epoch": 0.3925314806773773, + "grad_norm": 0.02444782853126526, + "learning_rate": 2.349090431435641e-05, + "loss": 0.482, + "step": 904 + }, + { + "epoch": 0.3929656969170647, + "grad_norm": 0.025424372404813766, + "learning_rate": 2.3007828340318116e-05, + "loss": 0.4907, + "step": 905 + }, + { + "epoch": 0.39339991315675205, + "grad_norm": 0.023807033896446228, + "learning_rate": 2.252965419885483e-05, + "loss": 0.5007, + "step": 906 + }, + { + "epoch": 0.3938341293964394, + "grad_norm": 0.022182505577802658, + "learning_rate": 2.2056386803945726e-05, + "loss": 0.4712, + "step": 907 + }, + { + "epoch": 0.3942683456361268, + "grad_norm": 0.02266152761876583, + "learning_rate": 2.1588031019145637e-05, + "loss": 0.479, + "step": 908 + }, + { + "epoch": 0.39470256187581415, + "grad_norm": 0.024663135409355164, + "learning_rate": 2.1124591657534777e-05, + "loss": 0.514, + "step": 909 + }, + { + "epoch": 0.3951367781155015, + "grad_norm": 0.023098768666386604, + "learning_rate": 2.0666073481669712e-05, + "loss": 0.4857, + "step": 910 + }, + { + "epoch": 0.3955709943551889, + "grad_norm": 0.025155888870358467, + "learning_rate": 2.0212481203534084e-05, + "loss": 0.5054, + "step": 911 + }, + { + "epoch": 0.39600521059487626, + "grad_norm": 0.030762221664190292, + "learning_rate": 1.9763819484490355e-05, + "loss": 0.4967, + "step": 912 + }, + { + "epoch": 0.3964394268345636, + "grad_norm": 0.029380332678556442, + "learning_rate": 1.932009293523196e-05, + "loss": 0.5412, + "step": 913 + }, + { + "epoch": 0.396873643074251, + "grad_norm": 0.023896733298897743, + "learning_rate": 1.888130611573563e-05, + "loss": 0.5002, + "step": 914 + }, + { + "epoch": 0.39730785931393836, + "grad_norm": 0.024988802149891853, + "learning_rate": 1.844746353521487e-05, + "loss": 0.4888, + "step": 915 + }, + { + "epoch": 0.39774207555362573, + "grad_norm": 0.02422316186130047, + "learning_rate": 1.801856965207338e-05, + "loss": 0.4818, + "step": 916 + }, + { + "epoch": 0.3981762917933131, + "grad_norm": 0.031409528106451035, + "learning_rate": 1.7594628873859486e-05, + "loss": 0.4923, + "step": 917 + }, + { + "epoch": 0.3986105080330004, + "grad_norm": 0.023978758603334427, + "learning_rate": 1.7175645557220565e-05, + "loss": 0.4798, + "step": 918 + }, + { + "epoch": 0.3990447242726878, + "grad_norm": 0.02715397998690605, + "learning_rate": 1.6761624007858522e-05, + "loss": 0.4474, + "step": 919 + }, + { + "epoch": 0.39947894051237515, + "grad_norm": 0.02678792178630829, + "learning_rate": 1.6352568480485276e-05, + "loss": 0.514, + "step": 920 + }, + { + "epoch": 0.3999131567520625, + "grad_norm": 0.02234644442796707, + "learning_rate": 1.594848317877934e-05, + "loss": 0.4869, + "step": 921 + }, + { + "epoch": 0.4003473729917499, + "grad_norm": 0.029807301238179207, + "learning_rate": 1.5549372255342365e-05, + "loss": 0.5242, + "step": 922 + }, + { + "epoch": 0.40078158923143725, + "grad_norm": 0.03037095069885254, + "learning_rate": 1.5155239811656562e-05, + "loss": 0.5035, + "step": 923 + }, + { + "epoch": 0.4012158054711246, + "grad_norm": 0.022767795249819756, + "learning_rate": 1.4766089898042678e-05, + "loss": 0.4903, + "step": 924 + }, + { + "epoch": 0.401650021710812, + "grad_norm": 0.026207391172647476, + "learning_rate": 1.4381926513618138e-05, + "loss": 0.4676, + "step": 925 + }, + { + "epoch": 0.40208423795049936, + "grad_norm": 0.025653038173913956, + "learning_rate": 1.400275360625608e-05, + "loss": 0.4804, + "step": 926 + }, + { + "epoch": 0.4025184541901867, + "grad_norm": 0.024481656029820442, + "learning_rate": 1.362857507254478e-05, + "loss": 0.4871, + "step": 927 + }, + { + "epoch": 0.4029526704298741, + "grad_norm": 0.02475779317319393, + "learning_rate": 1.3259394757747678e-05, + "loss": 0.5199, + "step": 928 + }, + { + "epoch": 0.40338688666956146, + "grad_norm": 0.02371501363813877, + "learning_rate": 1.289521645576358e-05, + "loss": 0.528, + "step": 929 + }, + { + "epoch": 0.40382110290924883, + "grad_norm": 0.023029111325740814, + "learning_rate": 1.2536043909088191e-05, + "loss": 0.4801, + "step": 930 + }, + { + "epoch": 0.40425531914893614, + "grad_norm": 0.025304608047008514, + "learning_rate": 1.2181880808775025e-05, + "loss": 0.4793, + "step": 931 + }, + { + "epoch": 0.4046895353886235, + "grad_norm": 0.02734360471367836, + "learning_rate": 1.183273079439795e-05, + "loss": 0.5022, + "step": 932 + }, + { + "epoch": 0.4051237516283109, + "grad_norm": 0.024530354887247086, + "learning_rate": 1.1488597454013538e-05, + "loss": 0.5032, + "step": 933 + }, + { + "epoch": 0.40555796786799825, + "grad_norm": 0.02369566634297371, + "learning_rate": 1.1149484324124327e-05, + "loss": 0.4633, + "step": 934 + }, + { + "epoch": 0.4059921841076856, + "grad_norm": 0.024855339899659157, + "learning_rate": 1.0815394889642338e-05, + "loss": 0.5175, + "step": 935 + }, + { + "epoch": 0.406426400347373, + "grad_norm": 0.023386195302009583, + "learning_rate": 1.0486332583853564e-05, + "loss": 0.5014, + "step": 936 + }, + { + "epoch": 0.40686061658706035, + "grad_norm": 0.029485292732715607, + "learning_rate": 1.0162300788382261e-05, + "loss": 0.4942, + "step": 937 + }, + { + "epoch": 0.4072948328267477, + "grad_norm": 0.023054329678416252, + "learning_rate": 9.843302833156376e-06, + "loss": 0.4683, + "step": 938 + }, + { + "epoch": 0.4077290490664351, + "grad_norm": 0.031031260266900063, + "learning_rate": 9.529341996373674e-06, + "loss": 0.5173, + "step": 939 + }, + { + "epoch": 0.40816326530612246, + "grad_norm": 0.023358464241027832, + "learning_rate": 9.22042150446728e-06, + "loss": 0.4722, + "step": 940 + }, + { + "epoch": 0.4085974815458098, + "grad_norm": 0.025074800476431847, + "learning_rate": 8.916544532073411e-06, + "loss": 0.4784, + "step": 941 + }, + { + "epoch": 0.4090316977854972, + "grad_norm": 0.02519945055246353, + "learning_rate": 8.617714201998083e-06, + "loss": 0.5022, + "step": 942 + }, + { + "epoch": 0.40946591402518456, + "grad_norm": 0.027328645810484886, + "learning_rate": 8.323933585185184e-06, + "loss": 0.4896, + "step": 943 + }, + { + "epoch": 0.40990013026487193, + "grad_norm": 0.02542622946202755, + "learning_rate": 8.035205700685167e-06, + "loss": 0.4676, + "step": 944 + }, + { + "epoch": 0.41033434650455924, + "grad_norm": 0.027195794507861137, + "learning_rate": 7.751533515623799e-06, + "loss": 0.4903, + "step": 945 + }, + { + "epoch": 0.4107685627442466, + "grad_norm": 0.02717754803597927, + "learning_rate": 7.47291994517163e-06, + "loss": 0.4812, + "step": 946 + }, + { + "epoch": 0.411202778983934, + "grad_norm": 0.02495565079152584, + "learning_rate": 7.199367852514238e-06, + "loss": 0.5014, + "step": 947 + }, + { + "epoch": 0.41163699522362135, + "grad_norm": 0.024513866752386093, + "learning_rate": 6.93088004882253e-06, + "loss": 0.5071, + "step": 948 + }, + { + "epoch": 0.4120712114633087, + "grad_norm": 0.03526155650615692, + "learning_rate": 6.667459293224154e-06, + "loss": 0.531, + "step": 949 + }, + { + "epoch": 0.4125054277029961, + "grad_norm": 0.023175369948148727, + "learning_rate": 6.4091082927749125e-06, + "loss": 0.4811, + "step": 950 + }, + { + "epoch": 0.41293964394268345, + "grad_norm": 0.026199523359537125, + "learning_rate": 6.15582970243117e-06, + "loss": 0.5087, + "step": 951 + }, + { + "epoch": 0.4133738601823708, + "grad_norm": 0.024454379454255104, + "learning_rate": 5.907626125022158e-06, + "loss": 0.479, + "step": 952 + }, + { + "epoch": 0.4138080764220582, + "grad_norm": 0.02492498978972435, + "learning_rate": 5.664500111223769e-06, + "loss": 0.4977, + "step": 953 + }, + { + "epoch": 0.41424229266174556, + "grad_norm": 0.023200612515211105, + "learning_rate": 5.426454159531913e-06, + "loss": 0.4803, + "step": 954 + }, + { + "epoch": 0.4146765089014329, + "grad_norm": 0.024389084428548813, + "learning_rate": 5.193490716237037e-06, + "loss": 0.4733, + "step": 955 + }, + { + "epoch": 0.4151107251411203, + "grad_norm": 0.027189958840608597, + "learning_rate": 4.9656121753990924e-06, + "loss": 0.4994, + "step": 956 + }, + { + "epoch": 0.41554494138080766, + "grad_norm": 0.02239627204835415, + "learning_rate": 4.7428208788224955e-06, + "loss": 0.4978, + "step": 957 + }, + { + "epoch": 0.41597915762049503, + "grad_norm": 0.021820807829499245, + "learning_rate": 4.52511911603265e-06, + "loss": 0.4871, + "step": 958 + }, + { + "epoch": 0.41641337386018235, + "grad_norm": 0.02775476686656475, + "learning_rate": 4.312509124251906e-06, + "loss": 0.4918, + "step": 959 + }, + { + "epoch": 0.4168475900998697, + "grad_norm": 0.023040220141410828, + "learning_rate": 4.104993088376974e-06, + "loss": 0.4497, + "step": 960 + }, + { + "epoch": 0.4172818063395571, + "grad_norm": 0.025357872247695923, + "learning_rate": 3.9025731409561e-06, + "loss": 0.4945, + "step": 961 + }, + { + "epoch": 0.41771602257924445, + "grad_norm": 0.02308405004441738, + "learning_rate": 3.7052513621674833e-06, + "loss": 0.4945, + "step": 962 + }, + { + "epoch": 0.4181502388189318, + "grad_norm": 0.02784230373799801, + "learning_rate": 3.513029779797783e-06, + "loss": 0.4543, + "step": 963 + }, + { + "epoch": 0.4185844550586192, + "grad_norm": 0.026198869571089745, + "learning_rate": 3.325910369220975e-06, + "loss": 0.4975, + "step": 964 + }, + { + "epoch": 0.41901867129830656, + "grad_norm": 0.023575875908136368, + "learning_rate": 3.143895053378698e-06, + "loss": 0.4987, + "step": 965 + }, + { + "epoch": 0.4194528875379939, + "grad_norm": 0.02861775830388069, + "learning_rate": 2.966985702759828e-06, + "loss": 0.4857, + "step": 966 + }, + { + "epoch": 0.4198871037776813, + "grad_norm": 0.033127009868621826, + "learning_rate": 2.7951841353817676e-06, + "loss": 0.465, + "step": 967 + }, + { + "epoch": 0.42032132001736866, + "grad_norm": 0.02451484464108944, + "learning_rate": 2.6284921167712973e-06, + "loss": 0.4863, + "step": 968 + }, + { + "epoch": 0.42075553625705603, + "grad_norm": 0.02445485256612301, + "learning_rate": 2.466911359946977e-06, + "loss": 0.4896, + "step": 969 + }, + { + "epoch": 0.4211897524967434, + "grad_norm": 0.02243492379784584, + "learning_rate": 2.3104435254008848e-06, + "loss": 0.4839, + "step": 970 + }, + { + "epoch": 0.42162396873643077, + "grad_norm": 0.026183119043707848, + "learning_rate": 2.1590902210822937e-06, + "loss": 0.4594, + "step": 971 + }, + { + "epoch": 0.42205818497611813, + "grad_norm": 0.025734947994351387, + "learning_rate": 2.012853002380466e-06, + "loss": 0.4684, + "step": 972 + }, + { + "epoch": 0.42249240121580545, + "grad_norm": 0.027873944491147995, + "learning_rate": 1.8717333721091634e-06, + "loss": 0.5056, + "step": 973 + }, + { + "epoch": 0.4229266174554928, + "grad_norm": 0.02484060265123844, + "learning_rate": 1.735732780490884e-06, + "loss": 0.5153, + "step": 974 + }, + { + "epoch": 0.4233608336951802, + "grad_norm": 0.022599438205361366, + "learning_rate": 1.6048526251421502e-06, + "loss": 0.4936, + "step": 975 + }, + { + "epoch": 0.42379504993486755, + "grad_norm": 0.020465485751628876, + "learning_rate": 1.4790942510590765e-06, + "loss": 0.4848, + "step": 976 + }, + { + "epoch": 0.4242292661745549, + "grad_norm": 0.027462849393486977, + "learning_rate": 1.3584589506034362e-06, + "loss": 0.4666, + "step": 977 + }, + { + "epoch": 0.4246634824142423, + "grad_norm": 0.02553846500813961, + "learning_rate": 1.2429479634897267e-06, + "loss": 0.4899, + "step": 978 + }, + { + "epoch": 0.42509769865392966, + "grad_norm": 0.024199191480875015, + "learning_rate": 1.132562476771959e-06, + "loss": 0.4792, + "step": 979 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 0.023328760638833046, + "learning_rate": 1.0273036248318324e-06, + "loss": 0.4973, + "step": 980 + }, + { + "epoch": 0.4259661311333044, + "grad_norm": 0.02446364425122738, + "learning_rate": 9.271724893669121e-07, + "loss": 0.4654, + "step": 981 + }, + { + "epoch": 0.42640034737299176, + "grad_norm": 0.02676616981625557, + "learning_rate": 8.321700993795811e-07, + "loss": 0.4881, + "step": 982 + }, + { + "epoch": 0.42683456361267913, + "grad_norm": 0.02478119358420372, + "learning_rate": 7.422974311662722e-07, + "loss": 0.5055, + "step": 983 + }, + { + "epoch": 0.4272687798523665, + "grad_norm": 0.03248545154929161, + "learning_rate": 6.575554083078084e-07, + "loss": 0.4949, + "step": 984 + }, + { + "epoch": 0.42770299609205387, + "grad_norm": 0.024504756554961205, + "learning_rate": 5.779449016595773e-07, + "loss": 0.4986, + "step": 985 + }, + { + "epoch": 0.42813721233174123, + "grad_norm": 0.02531745657324791, + "learning_rate": 5.034667293427053e-07, + "loss": 0.4851, + "step": 986 + }, + { + "epoch": 0.42857142857142855, + "grad_norm": 0.024369893595576286, + "learning_rate": 4.3412165673556436e-07, + "loss": 0.5044, + "step": 987 + }, + { + "epoch": 0.4290056448111159, + "grad_norm": 0.02750435471534729, + "learning_rate": 3.6991039646616656e-07, + "loss": 0.468, + "step": 988 + }, + { + "epoch": 0.4294398610508033, + "grad_norm": 0.02262257970869541, + "learning_rate": 3.1083360840455934e-07, + "loss": 0.498, + "step": 989 + }, + { + "epoch": 0.42987407729049065, + "grad_norm": 0.024247439578175545, + "learning_rate": 2.568918996560532e-07, + "loss": 0.4885, + "step": 990 + }, + { + "epoch": 0.430308293530178, + "grad_norm": 0.024190355092287064, + "learning_rate": 2.0808582455528192e-07, + "loss": 0.541, + "step": 991 + }, + { + "epoch": 0.4307425097698654, + "grad_norm": 0.023794230073690414, + "learning_rate": 1.6441588466009628e-07, + "loss": 0.4616, + "step": 992 + }, + { + "epoch": 0.43117672600955276, + "grad_norm": 0.023178860545158386, + "learning_rate": 1.2588252874673466e-07, + "loss": 0.5077, + "step": 993 + }, + { + "epoch": 0.4316109422492401, + "grad_norm": 0.0218358151614666, + "learning_rate": 9.248615280499361e-08, + "loss": 0.517, + "step": 994 + }, + { + "epoch": 0.4320451584889275, + "grad_norm": 0.021379638463258743, + "learning_rate": 6.422710003439747e-08, + "loss": 0.4589, + "step": 995 + }, + { + "epoch": 0.43247937472861486, + "grad_norm": 0.024375783279538155, + "learning_rate": 4.1105660840368156e-08, + "loss": 0.4963, + "step": 996 + }, + { + "epoch": 0.43291359096830223, + "grad_norm": 0.02347472496330738, + "learning_rate": 2.3122072831505136e-08, + "loss": 0.4855, + "step": 997 + }, + { + "epoch": 0.4333478072079896, + "grad_norm": 0.02579214982688427, + "learning_rate": 1.0276520816976388e-08, + "loss": 0.4786, + "step": 998 + }, + { + "epoch": 0.43378202344767697, + "grad_norm": 0.023950695991516113, + "learning_rate": 2.5691368046865116e-09, + "loss": 0.5068, + "step": 999 + }, + { + "epoch": 0.4342162396873643, + "grad_norm": 0.022179216146469116, + "learning_rate": 0.0, + "loss": 0.4841, + "step": 1000 + }, + { + "epoch": 0.4342162396873643, + "step": 1000, + "total_flos": 1.0100654119636173e+20, + "train_loss": 0.710578792899847, + "train_runtime": 20977.6715, + "train_samples_per_second": 0.763, + "train_steps_per_second": 0.048 + } + ], + "logging_steps": 1.0, + "max_steps": 1000, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.0100654119636173e+20, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}