diff --git "a/checkpoint-900/trainer_state.json" "b/checkpoint-900/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-900/trainer_state.json" @@ -0,0 +1,6630 @@ +{ + "best_metric": 0.0008496911614201963, + "best_model_checkpoint": "/home/paperspace/Data/models/akoul_whitehorseliquidity_25c/llm3br256/checkpoint-900", + "epoch": 2.909090909090909, + "eval_steps": 25, + "global_step": 900, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0032323232323232323, + "grad_norm": 0.03562889248132706, + "learning_rate": 1.2936610608020701e-07, + "loss": 0.0088, + "step": 1 + }, + { + "epoch": 0.006464646464646465, + "grad_norm": 0.03539334237575531, + "learning_rate": 2.5873221216041403e-07, + "loss": 0.0098, + "step": 2 + }, + { + "epoch": 0.009696969696969697, + "grad_norm": 0.03888387605547905, + "learning_rate": 3.8809831824062096e-07, + "loss": 0.0103, + "step": 3 + }, + { + "epoch": 0.01292929292929293, + "grad_norm": 0.03283314406871796, + "learning_rate": 5.174644243208281e-07, + "loss": 0.0123, + "step": 4 + }, + { + "epoch": 0.01616161616161616, + "grad_norm": 0.033392902463674545, + "learning_rate": 6.468305304010349e-07, + "loss": 0.0104, + "step": 5 + }, + { + "epoch": 0.019393939393939394, + "grad_norm": 0.041360873728990555, + "learning_rate": 7.761966364812419e-07, + "loss": 0.0135, + "step": 6 + }, + { + "epoch": 0.022626262626262626, + "grad_norm": 0.03297152742743492, + "learning_rate": 9.055627425614489e-07, + "loss": 0.01, + "step": 7 + }, + { + "epoch": 0.02585858585858586, + "grad_norm": 0.03385542333126068, + "learning_rate": 1.0349288486416561e-06, + "loss": 0.0092, + "step": 8 + }, + { + "epoch": 0.02909090909090909, + "grad_norm": 0.038460493087768555, + "learning_rate": 1.164294954721863e-06, + "loss": 0.0132, + "step": 9 + }, + { + "epoch": 0.03232323232323232, + "grad_norm": 0.030043406412005424, + "learning_rate": 1.2936610608020699e-06, + "loss": 0.0088, + "step": 10 + }, + { + "epoch": 0.035555555555555556, + "grad_norm": 0.039398446679115295, + "learning_rate": 1.423027166882277e-06, + "loss": 0.0118, + "step": 11 + }, + { + "epoch": 0.03878787878787879, + "grad_norm": 0.03809528797864914, + "learning_rate": 1.5523932729624839e-06, + "loss": 0.0146, + "step": 12 + }, + { + "epoch": 0.04202020202020202, + "grad_norm": 0.0388377383351326, + "learning_rate": 1.6817593790426907e-06, + "loss": 0.0102, + "step": 13 + }, + { + "epoch": 0.04525252525252525, + "grad_norm": 0.043977439403533936, + "learning_rate": 1.8111254851228978e-06, + "loss": 0.0099, + "step": 14 + }, + { + "epoch": 0.048484848484848485, + "grad_norm": 0.036720160394907, + "learning_rate": 1.9404915912031045e-06, + "loss": 0.0143, + "step": 15 + }, + { + "epoch": 0.05171717171717172, + "grad_norm": 0.03513036668300629, + "learning_rate": 2.0698576972833122e-06, + "loss": 0.0092, + "step": 16 + }, + { + "epoch": 0.05494949494949495, + "grad_norm": 0.03713906928896904, + "learning_rate": 2.199223803363519e-06, + "loss": 0.0144, + "step": 17 + }, + { + "epoch": 0.05818181818181818, + "grad_norm": 0.0338338240981102, + "learning_rate": 2.328589909443726e-06, + "loss": 0.0133, + "step": 18 + }, + { + "epoch": 0.061414141414141414, + "grad_norm": 0.0263836532831192, + "learning_rate": 2.457956015523933e-06, + "loss": 0.0113, + "step": 19 + }, + { + "epoch": 0.06464646464646465, + "grad_norm": 0.01817137934267521, + "learning_rate": 2.5873221216041398e-06, + "loss": 0.0074, + "step": 20 + }, + { + "epoch": 0.06787878787878789, + "grad_norm": 0.018132060766220093, + "learning_rate": 2.716688227684347e-06, + "loss": 0.008, + "step": 21 + }, + { + "epoch": 0.07111111111111111, + "grad_norm": 0.02117246761918068, + "learning_rate": 2.846054333764554e-06, + "loss": 0.0091, + "step": 22 + }, + { + "epoch": 0.07434343434343435, + "grad_norm": 0.01761500909924507, + "learning_rate": 2.975420439844761e-06, + "loss": 0.008, + "step": 23 + }, + { + "epoch": 0.07757575757575758, + "grad_norm": 0.01908932812511921, + "learning_rate": 3.1047865459249677e-06, + "loss": 0.008, + "step": 24 + }, + { + "epoch": 0.08080808080808081, + "grad_norm": 0.01619753986597061, + "learning_rate": 3.234152652005175e-06, + "loss": 0.0078, + "step": 25 + }, + { + "epoch": 0.08080808080808081, + "eval_loss": 0.007939654402434826, + "eval_runtime": 20.513, + "eval_samples_per_second": 4.875, + "eval_steps_per_second": 1.219, + "step": 25 + }, + { + "epoch": 0.08404040404040404, + "grad_norm": 0.01418386958539486, + "learning_rate": 3.3635187580853815e-06, + "loss": 0.0067, + "step": 26 + }, + { + "epoch": 0.08727272727272728, + "grad_norm": 0.018947051838040352, + "learning_rate": 3.492884864165589e-06, + "loss": 0.0086, + "step": 27 + }, + { + "epoch": 0.0905050505050505, + "grad_norm": 0.016921700909733772, + "learning_rate": 3.6222509702457957e-06, + "loss": 0.0085, + "step": 28 + }, + { + "epoch": 0.09373737373737374, + "grad_norm": 0.01551737543195486, + "learning_rate": 3.751617076326003e-06, + "loss": 0.0084, + "step": 29 + }, + { + "epoch": 0.09696969696969697, + "grad_norm": 0.014143792912364006, + "learning_rate": 3.880983182406209e-06, + "loss": 0.0078, + "step": 30 + }, + { + "epoch": 0.10020202020202021, + "grad_norm": 0.011397017166018486, + "learning_rate": 4.010349288486417e-06, + "loss": 0.0053, + "step": 31 + }, + { + "epoch": 0.10343434343434343, + "grad_norm": 0.015161341056227684, + "learning_rate": 4.1397153945666245e-06, + "loss": 0.0073, + "step": 32 + }, + { + "epoch": 0.10666666666666667, + "grad_norm": 0.012180107645690441, + "learning_rate": 4.2690815006468305e-06, + "loss": 0.0063, + "step": 33 + }, + { + "epoch": 0.1098989898989899, + "grad_norm": 0.012770959176123142, + "learning_rate": 4.398447606727038e-06, + "loss": 0.0079, + "step": 34 + }, + { + "epoch": 0.11313131313131314, + "grad_norm": 0.014727453701198101, + "learning_rate": 4.527813712807244e-06, + "loss": 0.0122, + "step": 35 + }, + { + "epoch": 0.11636363636363636, + "grad_norm": 0.012778300791978836, + "learning_rate": 4.657179818887452e-06, + "loss": 0.0071, + "step": 36 + }, + { + "epoch": 0.1195959595959596, + "grad_norm": 0.012685113586485386, + "learning_rate": 4.786545924967659e-06, + "loss": 0.0065, + "step": 37 + }, + { + "epoch": 0.12282828282828283, + "grad_norm": 0.014156874269247055, + "learning_rate": 4.915912031047866e-06, + "loss": 0.0082, + "step": 38 + }, + { + "epoch": 0.12606060606060607, + "grad_norm": 0.01115910243242979, + "learning_rate": 5.045278137128073e-06, + "loss": 0.007, + "step": 39 + }, + { + "epoch": 0.1292929292929293, + "grad_norm": 0.014089164324104786, + "learning_rate": 5.1746442432082795e-06, + "loss": 0.0073, + "step": 40 + }, + { + "epoch": 0.13252525252525252, + "grad_norm": 0.013440214097499847, + "learning_rate": 5.304010349288486e-06, + "loss": 0.0082, + "step": 41 + }, + { + "epoch": 0.13575757575757577, + "grad_norm": 0.01083499938249588, + "learning_rate": 5.433376455368694e-06, + "loss": 0.006, + "step": 42 + }, + { + "epoch": 0.138989898989899, + "grad_norm": 0.011082631535828114, + "learning_rate": 5.5627425614489e-06, + "loss": 0.007, + "step": 43 + }, + { + "epoch": 0.14222222222222222, + "grad_norm": 0.012490961700677872, + "learning_rate": 5.692108667529108e-06, + "loss": 0.0062, + "step": 44 + }, + { + "epoch": 0.14545454545454545, + "grad_norm": 0.012171825394034386, + "learning_rate": 5.821474773609315e-06, + "loss": 0.0066, + "step": 45 + }, + { + "epoch": 0.1486868686868687, + "grad_norm": 0.010148804634809494, + "learning_rate": 5.950840879689522e-06, + "loss": 0.0042, + "step": 46 + }, + { + "epoch": 0.15191919191919193, + "grad_norm": 0.013537143357098103, + "learning_rate": 6.0802069857697286e-06, + "loss": 0.0073, + "step": 47 + }, + { + "epoch": 0.15515151515151515, + "grad_norm": 0.011110929772257805, + "learning_rate": 6.2095730918499354e-06, + "loss": 0.0072, + "step": 48 + }, + { + "epoch": 0.15838383838383838, + "grad_norm": 0.00937278475612402, + "learning_rate": 6.338939197930142e-06, + "loss": 0.0054, + "step": 49 + }, + { + "epoch": 0.16161616161616163, + "grad_norm": 0.011824984103441238, + "learning_rate": 6.46830530401035e-06, + "loss": 0.0119, + "step": 50 + }, + { + "epoch": 0.16161616161616163, + "eval_loss": 0.005149205215275288, + "eval_runtime": 18.6062, + "eval_samples_per_second": 5.375, + "eval_steps_per_second": 1.344, + "step": 50 + }, + { + "epoch": 0.16484848484848486, + "grad_norm": 0.009701536037027836, + "learning_rate": 6.597671410090557e-06, + "loss": 0.0054, + "step": 51 + }, + { + "epoch": 0.16808080808080808, + "grad_norm": 0.010364706628024578, + "learning_rate": 6.727037516170763e-06, + "loss": 0.0049, + "step": 52 + }, + { + "epoch": 0.1713131313131313, + "grad_norm": 0.009346776641905308, + "learning_rate": 6.856403622250971e-06, + "loss": 0.0054, + "step": 53 + }, + { + "epoch": 0.17454545454545456, + "grad_norm": 0.008815059438347816, + "learning_rate": 6.985769728331178e-06, + "loss": 0.0069, + "step": 54 + }, + { + "epoch": 0.17777777777777778, + "grad_norm": 0.009431697428226471, + "learning_rate": 7.115135834411385e-06, + "loss": 0.0083, + "step": 55 + }, + { + "epoch": 0.181010101010101, + "grad_norm": 0.012181616388261318, + "learning_rate": 7.244501940491591e-06, + "loss": 0.0065, + "step": 56 + }, + { + "epoch": 0.18424242424242424, + "grad_norm": 0.009227719157934189, + "learning_rate": 7.373868046571798e-06, + "loss": 0.0052, + "step": 57 + }, + { + "epoch": 0.1874747474747475, + "grad_norm": 0.009855546988546848, + "learning_rate": 7.503234152652006e-06, + "loss": 0.0082, + "step": 58 + }, + { + "epoch": 0.1907070707070707, + "grad_norm": 0.008454745635390282, + "learning_rate": 7.632600258732213e-06, + "loss": 0.0044, + "step": 59 + }, + { + "epoch": 0.19393939393939394, + "grad_norm": 0.010128876194357872, + "learning_rate": 7.761966364812418e-06, + "loss": 0.0092, + "step": 60 + }, + { + "epoch": 0.19717171717171716, + "grad_norm": 0.009267722256481647, + "learning_rate": 7.891332470892627e-06, + "loss": 0.0049, + "step": 61 + }, + { + "epoch": 0.20040404040404042, + "grad_norm": 0.009394655004143715, + "learning_rate": 8.020698576972833e-06, + "loss": 0.0059, + "step": 62 + }, + { + "epoch": 0.20363636363636364, + "grad_norm": 0.008983040228486061, + "learning_rate": 8.15006468305304e-06, + "loss": 0.0041, + "step": 63 + }, + { + "epoch": 0.20686868686868687, + "grad_norm": 0.009920783340930939, + "learning_rate": 8.279430789133249e-06, + "loss": 0.0058, + "step": 64 + }, + { + "epoch": 0.2101010101010101, + "grad_norm": 0.007694128435105085, + "learning_rate": 8.408796895213454e-06, + "loss": 0.0033, + "step": 65 + }, + { + "epoch": 0.21333333333333335, + "grad_norm": 0.01226822566241026, + "learning_rate": 8.538163001293661e-06, + "loss": 0.0099, + "step": 66 + }, + { + "epoch": 0.21656565656565657, + "grad_norm": 0.010597337037324905, + "learning_rate": 8.66752910737387e-06, + "loss": 0.0081, + "step": 67 + }, + { + "epoch": 0.2197979797979798, + "grad_norm": 0.009066218510270119, + "learning_rate": 8.796895213454076e-06, + "loss": 0.0044, + "step": 68 + }, + { + "epoch": 0.22303030303030302, + "grad_norm": 0.01163114607334137, + "learning_rate": 8.926261319534282e-06, + "loss": 0.0129, + "step": 69 + }, + { + "epoch": 0.22626262626262628, + "grad_norm": 0.012832598760724068, + "learning_rate": 9.055627425614489e-06, + "loss": 0.0071, + "step": 70 + }, + { + "epoch": 0.2294949494949495, + "grad_norm": 0.010620299726724625, + "learning_rate": 9.184993531694697e-06, + "loss": 0.004, + "step": 71 + }, + { + "epoch": 0.23272727272727273, + "grad_norm": 0.00913357175886631, + "learning_rate": 9.314359637774904e-06, + "loss": 0.0058, + "step": 72 + }, + { + "epoch": 0.23595959595959595, + "grad_norm": 0.011772734113037586, + "learning_rate": 9.44372574385511e-06, + "loss": 0.0036, + "step": 73 + }, + { + "epoch": 0.2391919191919192, + "grad_norm": 0.011326112784445286, + "learning_rate": 9.573091849935318e-06, + "loss": 0.0071, + "step": 74 + }, + { + "epoch": 0.24242424242424243, + "grad_norm": 0.010667567141354084, + "learning_rate": 9.702457956015525e-06, + "loss": 0.0036, + "step": 75 + }, + { + "epoch": 0.24242424242424243, + "eval_loss": 0.0031650287564843893, + "eval_runtime": 18.6391, + "eval_samples_per_second": 5.365, + "eval_steps_per_second": 1.341, + "step": 75 + }, + { + "epoch": 0.24565656565656566, + "grad_norm": 0.007108451332896948, + "learning_rate": 9.831824062095732e-06, + "loss": 0.0023, + "step": 76 + }, + { + "epoch": 0.24888888888888888, + "grad_norm": 0.008278781548142433, + "learning_rate": 9.961190168175938e-06, + "loss": 0.0026, + "step": 77 + }, + { + "epoch": 0.25212121212121213, + "grad_norm": 0.010979422368109226, + "learning_rate": 1.0090556274256145e-05, + "loss": 0.0036, + "step": 78 + }, + { + "epoch": 0.25535353535353533, + "grad_norm": 0.007666402496397495, + "learning_rate": 1.0219922380336352e-05, + "loss": 0.0015, + "step": 79 + }, + { + "epoch": 0.2585858585858586, + "grad_norm": 0.009133332408964634, + "learning_rate": 1.0349288486416559e-05, + "loss": 0.0021, + "step": 80 + }, + { + "epoch": 0.26181818181818184, + "grad_norm": 0.009566927328705788, + "learning_rate": 1.0478654592496766e-05, + "loss": 0.0038, + "step": 81 + }, + { + "epoch": 0.26505050505050504, + "grad_norm": 0.011765814386308193, + "learning_rate": 1.0608020698576973e-05, + "loss": 0.0061, + "step": 82 + }, + { + "epoch": 0.2682828282828283, + "grad_norm": 0.009488740935921669, + "learning_rate": 1.073738680465718e-05, + "loss": 0.0023, + "step": 83 + }, + { + "epoch": 0.27151515151515154, + "grad_norm": 0.007137450389564037, + "learning_rate": 1.0866752910737388e-05, + "loss": 0.0021, + "step": 84 + }, + { + "epoch": 0.27474747474747474, + "grad_norm": 0.008984182961285114, + "learning_rate": 1.0996119016817593e-05, + "loss": 0.0028, + "step": 85 + }, + { + "epoch": 0.277979797979798, + "grad_norm": 0.010403821244835854, + "learning_rate": 1.11254851228978e-05, + "loss": 0.0057, + "step": 86 + }, + { + "epoch": 0.2812121212121212, + "grad_norm": 0.0071556540206074715, + "learning_rate": 1.1254851228978009e-05, + "loss": 0.0017, + "step": 87 + }, + { + "epoch": 0.28444444444444444, + "grad_norm": 0.012338937260210514, + "learning_rate": 1.1384217335058216e-05, + "loss": 0.0029, + "step": 88 + }, + { + "epoch": 0.2876767676767677, + "grad_norm": 0.010343515314161777, + "learning_rate": 1.1513583441138421e-05, + "loss": 0.0032, + "step": 89 + }, + { + "epoch": 0.2909090909090909, + "grad_norm": 0.009991390630602837, + "learning_rate": 1.164294954721863e-05, + "loss": 0.0035, + "step": 90 + }, + { + "epoch": 0.29414141414141415, + "grad_norm": 0.009020160883665085, + "learning_rate": 1.1772315653298836e-05, + "loss": 0.0033, + "step": 91 + }, + { + "epoch": 0.2973737373737374, + "grad_norm": 0.0067694829776883125, + "learning_rate": 1.1901681759379043e-05, + "loss": 0.0019, + "step": 92 + }, + { + "epoch": 0.3006060606060606, + "grad_norm": 0.013036763295531273, + "learning_rate": 1.203104786545925e-05, + "loss": 0.0041, + "step": 93 + }, + { + "epoch": 0.30383838383838385, + "grad_norm": 0.009171461686491966, + "learning_rate": 1.2160413971539457e-05, + "loss": 0.006, + "step": 94 + }, + { + "epoch": 0.30707070707070705, + "grad_norm": 0.007229423616081476, + "learning_rate": 1.2289780077619664e-05, + "loss": 0.002, + "step": 95 + }, + { + "epoch": 0.3103030303030303, + "grad_norm": 0.009065698832273483, + "learning_rate": 1.2419146183699871e-05, + "loss": 0.0028, + "step": 96 + }, + { + "epoch": 0.31353535353535356, + "grad_norm": 0.0068460931070148945, + "learning_rate": 1.254851228978008e-05, + "loss": 0.0024, + "step": 97 + }, + { + "epoch": 0.31676767676767675, + "grad_norm": 0.12772586941719055, + "learning_rate": 1.2677878395860285e-05, + "loss": 0.0046, + "step": 98 + }, + { + "epoch": 0.32, + "grad_norm": 0.008569791913032532, + "learning_rate": 1.2807244501940493e-05, + "loss": 0.0024, + "step": 99 + }, + { + "epoch": 0.32323232323232326, + "grad_norm": 0.010226168669760227, + "learning_rate": 1.29366106080207e-05, + "loss": 0.004, + "step": 100 + }, + { + "epoch": 0.32323232323232326, + "eval_loss": 0.002474932000041008, + "eval_runtime": 18.5931, + "eval_samples_per_second": 5.378, + "eval_steps_per_second": 1.345, + "step": 100 + }, + { + "epoch": 0.32646464646464646, + "grad_norm": 0.010962710715830326, + "learning_rate": 1.3065976714100905e-05, + "loss": 0.0018, + "step": 101 + }, + { + "epoch": 0.3296969696969697, + "grad_norm": 0.01257998775690794, + "learning_rate": 1.3195342820181114e-05, + "loss": 0.0033, + "step": 102 + }, + { + "epoch": 0.3329292929292929, + "grad_norm": 0.013269903138279915, + "learning_rate": 1.332470892626132e-05, + "loss": 0.0036, + "step": 103 + }, + { + "epoch": 0.33616161616161616, + "grad_norm": 0.009509469382464886, + "learning_rate": 1.3454075032341526e-05, + "loss": 0.0034, + "step": 104 + }, + { + "epoch": 0.3393939393939394, + "grad_norm": 0.00902635883539915, + "learning_rate": 1.3583441138421735e-05, + "loss": 0.0028, + "step": 105 + }, + { + "epoch": 0.3426262626262626, + "grad_norm": 0.007760872133076191, + "learning_rate": 1.3712807244501941e-05, + "loss": 0.0024, + "step": 106 + }, + { + "epoch": 0.34585858585858587, + "grad_norm": 0.009263481013476849, + "learning_rate": 1.384217335058215e-05, + "loss": 0.0031, + "step": 107 + }, + { + "epoch": 0.3490909090909091, + "grad_norm": 0.00697364890947938, + "learning_rate": 1.3971539456662355e-05, + "loss": 0.0021, + "step": 108 + }, + { + "epoch": 0.3523232323232323, + "grad_norm": 0.007600918412208557, + "learning_rate": 1.4100905562742562e-05, + "loss": 0.0018, + "step": 109 + }, + { + "epoch": 0.35555555555555557, + "grad_norm": 0.00856352224946022, + "learning_rate": 1.423027166882277e-05, + "loss": 0.0031, + "step": 110 + }, + { + "epoch": 0.35878787878787877, + "grad_norm": 0.007947358302772045, + "learning_rate": 1.4359637774902976e-05, + "loss": 0.0025, + "step": 111 + }, + { + "epoch": 0.362020202020202, + "grad_norm": 0.007706194184720516, + "learning_rate": 1.4489003880983183e-05, + "loss": 0.0023, + "step": 112 + }, + { + "epoch": 0.3652525252525253, + "grad_norm": 0.008598407730460167, + "learning_rate": 1.4618369987063391e-05, + "loss": 0.0028, + "step": 113 + }, + { + "epoch": 0.36848484848484847, + "grad_norm": 0.00845644623041153, + "learning_rate": 1.4747736093143596e-05, + "loss": 0.0029, + "step": 114 + }, + { + "epoch": 0.3717171717171717, + "grad_norm": 0.008533057756721973, + "learning_rate": 1.4877102199223805e-05, + "loss": 0.0039, + "step": 115 + }, + { + "epoch": 0.374949494949495, + "grad_norm": 0.007429471705108881, + "learning_rate": 1.5006468305304012e-05, + "loss": 0.0024, + "step": 116 + }, + { + "epoch": 0.3781818181818182, + "grad_norm": 0.0069627161137759686, + "learning_rate": 1.5135834411384217e-05, + "loss": 0.0018, + "step": 117 + }, + { + "epoch": 0.3814141414141414, + "grad_norm": 0.007471222430467606, + "learning_rate": 1.5265200517464426e-05, + "loss": 0.0024, + "step": 118 + }, + { + "epoch": 0.3846464646464646, + "grad_norm": 0.006873416714370251, + "learning_rate": 1.5394566623544633e-05, + "loss": 0.0019, + "step": 119 + }, + { + "epoch": 0.3878787878787879, + "grad_norm": 0.009251467883586884, + "learning_rate": 1.5523932729624836e-05, + "loss": 0.0027, + "step": 120 + }, + { + "epoch": 0.39111111111111113, + "grad_norm": 0.007302634883671999, + "learning_rate": 1.5653298835705046e-05, + "loss": 0.0017, + "step": 121 + }, + { + "epoch": 0.39434343434343433, + "grad_norm": 0.006746005266904831, + "learning_rate": 1.5782664941785253e-05, + "loss": 0.0019, + "step": 122 + }, + { + "epoch": 0.3975757575757576, + "grad_norm": 0.008011849597096443, + "learning_rate": 1.591203104786546e-05, + "loss": 0.0023, + "step": 123 + }, + { + "epoch": 0.40080808080808084, + "grad_norm": 0.008799983188509941, + "learning_rate": 1.6041397153945667e-05, + "loss": 0.0023, + "step": 124 + }, + { + "epoch": 0.40404040404040403, + "grad_norm": 0.010617449879646301, + "learning_rate": 1.6170763260025874e-05, + "loss": 0.0019, + "step": 125 + }, + { + "epoch": 0.40404040404040403, + "eval_loss": 0.002037045545876026, + "eval_runtime": 18.5899, + "eval_samples_per_second": 5.379, + "eval_steps_per_second": 1.345, + "step": 125 + }, + { + "epoch": 0.4072727272727273, + "grad_norm": 0.006759752053767443, + "learning_rate": 1.630012936610608e-05, + "loss": 0.0018, + "step": 126 + }, + { + "epoch": 0.4105050505050505, + "grad_norm": 0.008995631709694862, + "learning_rate": 1.6429495472186288e-05, + "loss": 0.0022, + "step": 127 + }, + { + "epoch": 0.41373737373737374, + "grad_norm": 0.00835257675498724, + "learning_rate": 1.6558861578266498e-05, + "loss": 0.0026, + "step": 128 + }, + { + "epoch": 0.416969696969697, + "grad_norm": 0.009772442281246185, + "learning_rate": 1.66882276843467e-05, + "loss": 0.0028, + "step": 129 + }, + { + "epoch": 0.4202020202020202, + "grad_norm": 0.006931178271770477, + "learning_rate": 1.6817593790426908e-05, + "loss": 0.002, + "step": 130 + }, + { + "epoch": 0.42343434343434344, + "grad_norm": 0.012796151451766491, + "learning_rate": 1.694695989650712e-05, + "loss": 0.0021, + "step": 131 + }, + { + "epoch": 0.4266666666666667, + "grad_norm": 0.007226724177598953, + "learning_rate": 1.7076326002587322e-05, + "loss": 0.0033, + "step": 132 + }, + { + "epoch": 0.4298989898989899, + "grad_norm": 0.0066243987530469894, + "learning_rate": 1.720569210866753e-05, + "loss": 0.002, + "step": 133 + }, + { + "epoch": 0.43313131313131314, + "grad_norm": 0.007128287572413683, + "learning_rate": 1.733505821474774e-05, + "loss": 0.0021, + "step": 134 + }, + { + "epoch": 0.43636363636363634, + "grad_norm": 0.007028148043900728, + "learning_rate": 1.7464424320827943e-05, + "loss": 0.0028, + "step": 135 + }, + { + "epoch": 0.4395959595959596, + "grad_norm": 0.00866060983389616, + "learning_rate": 1.7593790426908153e-05, + "loss": 0.0053, + "step": 136 + }, + { + "epoch": 0.44282828282828285, + "grad_norm": 0.005735491868108511, + "learning_rate": 1.7723156532988356e-05, + "loss": 0.0023, + "step": 137 + }, + { + "epoch": 0.44606060606060605, + "grad_norm": 0.005858046934008598, + "learning_rate": 1.7852522639068563e-05, + "loss": 0.0019, + "step": 138 + }, + { + "epoch": 0.4492929292929293, + "grad_norm": 0.007398343179374933, + "learning_rate": 1.7981888745148774e-05, + "loss": 0.0024, + "step": 139 + }, + { + "epoch": 0.45252525252525255, + "grad_norm": 0.009320907294750214, + "learning_rate": 1.8111254851228977e-05, + "loss": 0.0063, + "step": 140 + }, + { + "epoch": 0.45575757575757575, + "grad_norm": 0.008654161356389523, + "learning_rate": 1.8240620957309184e-05, + "loss": 0.0025, + "step": 141 + }, + { + "epoch": 0.458989898989899, + "grad_norm": 0.004506159573793411, + "learning_rate": 1.8369987063389394e-05, + "loss": 0.0013, + "step": 142 + }, + { + "epoch": 0.4622222222222222, + "grad_norm": 0.009703525342047215, + "learning_rate": 1.8499353169469598e-05, + "loss": 0.0034, + "step": 143 + }, + { + "epoch": 0.46545454545454545, + "grad_norm": 0.007734385784715414, + "learning_rate": 1.8628719275549808e-05, + "loss": 0.0029, + "step": 144 + }, + { + "epoch": 0.4686868686868687, + "grad_norm": 0.010392666794359684, + "learning_rate": 1.8758085381630015e-05, + "loss": 0.0028, + "step": 145 + }, + { + "epoch": 0.4719191919191919, + "grad_norm": 0.01011224091053009, + "learning_rate": 1.888745148771022e-05, + "loss": 0.0036, + "step": 146 + }, + { + "epoch": 0.47515151515151516, + "grad_norm": 0.005182855296880007, + "learning_rate": 1.901681759379043e-05, + "loss": 0.0014, + "step": 147 + }, + { + "epoch": 0.4783838383838384, + "grad_norm": 0.009993299841880798, + "learning_rate": 1.9146183699870636e-05, + "loss": 0.003, + "step": 148 + }, + { + "epoch": 0.4816161616161616, + "grad_norm": 0.008879208005964756, + "learning_rate": 1.927554980595084e-05, + "loss": 0.002, + "step": 149 + }, + { + "epoch": 0.48484848484848486, + "grad_norm": 0.006573604419827461, + "learning_rate": 1.940491591203105e-05, + "loss": 0.0021, + "step": 150 + }, + { + "epoch": 0.48484848484848486, + "eval_loss": 0.0017716821748763323, + "eval_runtime": 18.6403, + "eval_samples_per_second": 5.365, + "eval_steps_per_second": 1.341, + "step": 150 + }, + { + "epoch": 0.48808080808080806, + "grad_norm": 0.005542725790292025, + "learning_rate": 1.9534282018111256e-05, + "loss": 0.0017, + "step": 151 + }, + { + "epoch": 0.4913131313131313, + "grad_norm": 0.008134805597364902, + "learning_rate": 1.9663648124191463e-05, + "loss": 0.002, + "step": 152 + }, + { + "epoch": 0.49454545454545457, + "grad_norm": 0.009408293291926384, + "learning_rate": 1.979301423027167e-05, + "loss": 0.0031, + "step": 153 + }, + { + "epoch": 0.49777777777777776, + "grad_norm": 0.007265687920153141, + "learning_rate": 1.9922380336351877e-05, + "loss": 0.0021, + "step": 154 + }, + { + "epoch": 0.501010101010101, + "grad_norm": 0.009999910369515419, + "learning_rate": 2.0051746442432084e-05, + "loss": 0.0014, + "step": 155 + }, + { + "epoch": 0.5042424242424243, + "grad_norm": 0.007147953379899263, + "learning_rate": 2.018111254851229e-05, + "loss": 0.0022, + "step": 156 + }, + { + "epoch": 0.5074747474747475, + "grad_norm": 0.0072359428741037846, + "learning_rate": 2.0310478654592497e-05, + "loss": 0.0034, + "step": 157 + }, + { + "epoch": 0.5107070707070707, + "grad_norm": 0.005821447819471359, + "learning_rate": 2.0439844760672704e-05, + "loss": 0.0018, + "step": 158 + }, + { + "epoch": 0.5139393939393939, + "grad_norm": 0.007256666664034128, + "learning_rate": 2.056921086675291e-05, + "loss": 0.0027, + "step": 159 + }, + { + "epoch": 0.5171717171717172, + "grad_norm": 0.00663521233946085, + "learning_rate": 2.0698576972833118e-05, + "loss": 0.0023, + "step": 160 + }, + { + "epoch": 0.5204040404040404, + "grad_norm": 0.007504627574235201, + "learning_rate": 2.0827943078913325e-05, + "loss": 0.004, + "step": 161 + }, + { + "epoch": 0.5236363636363637, + "grad_norm": 0.005167617462575436, + "learning_rate": 2.0957309184993532e-05, + "loss": 0.0015, + "step": 162 + }, + { + "epoch": 0.5268686868686868, + "grad_norm": 0.00587807409465313, + "learning_rate": 2.108667529107374e-05, + "loss": 0.0016, + "step": 163 + }, + { + "epoch": 0.5301010101010101, + "grad_norm": 0.005344127304852009, + "learning_rate": 2.1216041397153946e-05, + "loss": 0.0011, + "step": 164 + }, + { + "epoch": 0.5333333333333333, + "grad_norm": 0.005351161118596792, + "learning_rate": 2.1345407503234156e-05, + "loss": 0.0016, + "step": 165 + }, + { + "epoch": 0.5365656565656566, + "grad_norm": 0.006603873800486326, + "learning_rate": 2.147477360931436e-05, + "loss": 0.0017, + "step": 166 + }, + { + "epoch": 0.5397979797979798, + "grad_norm": 0.006440309341996908, + "learning_rate": 2.1604139715394566e-05, + "loss": 0.0016, + "step": 167 + }, + { + "epoch": 0.5430303030303031, + "grad_norm": 0.007731405086815357, + "learning_rate": 2.1733505821474777e-05, + "loss": 0.0024, + "step": 168 + }, + { + "epoch": 0.5462626262626262, + "grad_norm": 0.0070298160426318645, + "learning_rate": 2.186287192755498e-05, + "loss": 0.0019, + "step": 169 + }, + { + "epoch": 0.5494949494949495, + "grad_norm": 0.008382032625377178, + "learning_rate": 2.1992238033635187e-05, + "loss": 0.0017, + "step": 170 + }, + { + "epoch": 0.5527272727272727, + "grad_norm": 0.007320054341107607, + "learning_rate": 2.2121604139715397e-05, + "loss": 0.0017, + "step": 171 + }, + { + "epoch": 0.555959595959596, + "grad_norm": 0.008367806673049927, + "learning_rate": 2.22509702457956e-05, + "loss": 0.0036, + "step": 172 + }, + { + "epoch": 0.5591919191919192, + "grad_norm": 0.008995486423373222, + "learning_rate": 2.238033635187581e-05, + "loss": 0.0019, + "step": 173 + }, + { + "epoch": 0.5624242424242424, + "grad_norm": 0.007300530560314655, + "learning_rate": 2.2509702457956018e-05, + "loss": 0.0036, + "step": 174 + }, + { + "epoch": 0.5656565656565656, + "grad_norm": 0.005388753954321146, + "learning_rate": 2.263906856403622e-05, + "loss": 0.0016, + "step": 175 + }, + { + "epoch": 0.5656565656565656, + "eval_loss": 0.0016459682956337929, + "eval_runtime": 18.6118, + "eval_samples_per_second": 5.373, + "eval_steps_per_second": 1.343, + "step": 175 + }, + { + "epoch": 0.5688888888888889, + "grad_norm": 0.005545719526708126, + "learning_rate": 2.276843467011643e-05, + "loss": 0.0016, + "step": 176 + }, + { + "epoch": 0.5721212121212121, + "grad_norm": 0.007394594140350819, + "learning_rate": 2.289780077619664e-05, + "loss": 0.0021, + "step": 177 + }, + { + "epoch": 0.5753535353535354, + "grad_norm": 0.005089296959340572, + "learning_rate": 2.3027166882276842e-05, + "loss": 0.0013, + "step": 178 + }, + { + "epoch": 0.5785858585858585, + "grad_norm": 0.00883107353001833, + "learning_rate": 2.3156532988357052e-05, + "loss": 0.0022, + "step": 179 + }, + { + "epoch": 0.5818181818181818, + "grad_norm": 0.012213426642119884, + "learning_rate": 2.328589909443726e-05, + "loss": 0.005, + "step": 180 + }, + { + "epoch": 0.585050505050505, + "grad_norm": 0.007255645003169775, + "learning_rate": 2.3415265200517466e-05, + "loss": 0.001, + "step": 181 + }, + { + "epoch": 0.5882828282828283, + "grad_norm": 0.006818380672484636, + "learning_rate": 2.3544631306597673e-05, + "loss": 0.0015, + "step": 182 + }, + { + "epoch": 0.5915151515151515, + "grad_norm": 0.006983195431530476, + "learning_rate": 2.367399741267788e-05, + "loss": 0.0027, + "step": 183 + }, + { + "epoch": 0.5947474747474748, + "grad_norm": 0.012661872431635857, + "learning_rate": 2.3803363518758087e-05, + "loss": 0.0053, + "step": 184 + }, + { + "epoch": 0.597979797979798, + "grad_norm": 0.005046526901423931, + "learning_rate": 2.3932729624838294e-05, + "loss": 0.0019, + "step": 185 + }, + { + "epoch": 0.6012121212121212, + "grad_norm": 0.004457033704966307, + "learning_rate": 2.40620957309185e-05, + "loss": 0.0012, + "step": 186 + }, + { + "epoch": 0.6044444444444445, + "grad_norm": 0.00732004176825285, + "learning_rate": 2.4191461836998707e-05, + "loss": 0.0017, + "step": 187 + }, + { + "epoch": 0.6076767676767677, + "grad_norm": 0.006394708063453436, + "learning_rate": 2.4320827943078914e-05, + "loss": 0.0019, + "step": 188 + }, + { + "epoch": 0.610909090909091, + "grad_norm": 0.007961345836520195, + "learning_rate": 2.445019404915912e-05, + "loss": 0.0034, + "step": 189 + }, + { + "epoch": 0.6141414141414141, + "grad_norm": 0.0068709347397089005, + "learning_rate": 2.4579560155239328e-05, + "loss": 0.0013, + "step": 190 + }, + { + "epoch": 0.6173737373737374, + "grad_norm": 0.009581703692674637, + "learning_rate": 2.4708926261319535e-05, + "loss": 0.0026, + "step": 191 + }, + { + "epoch": 0.6206060606060606, + "grad_norm": 0.007400548551231623, + "learning_rate": 2.4838292367399742e-05, + "loss": 0.0024, + "step": 192 + }, + { + "epoch": 0.6238383838383839, + "grad_norm": 0.00970857497304678, + "learning_rate": 2.496765847347995e-05, + "loss": 0.0013, + "step": 193 + }, + { + "epoch": 0.6270707070707071, + "grad_norm": 0.007368095684796572, + "learning_rate": 2.509702457956016e-05, + "loss": 0.0016, + "step": 194 + }, + { + "epoch": 0.6303030303030303, + "grad_norm": 0.007306639105081558, + "learning_rate": 2.5226390685640362e-05, + "loss": 0.0015, + "step": 195 + }, + { + "epoch": 0.6335353535353535, + "grad_norm": 0.008257709443569183, + "learning_rate": 2.535575679172057e-05, + "loss": 0.0017, + "step": 196 + }, + { + "epoch": 0.6367676767676768, + "grad_norm": 0.0050062090158462524, + "learning_rate": 2.548512289780078e-05, + "loss": 0.0012, + "step": 197 + }, + { + "epoch": 0.64, + "grad_norm": 0.007328143808990717, + "learning_rate": 2.5614489003880986e-05, + "loss": 0.0025, + "step": 198 + }, + { + "epoch": 0.6432323232323233, + "grad_norm": 0.007282217964529991, + "learning_rate": 2.574385510996119e-05, + "loss": 0.0018, + "step": 199 + }, + { + "epoch": 0.6464646464646465, + "grad_norm": 0.007711055688560009, + "learning_rate": 2.58732212160414e-05, + "loss": 0.0013, + "step": 200 + }, + { + "epoch": 0.6464646464646465, + "eval_loss": 0.0015150802209973335, + "eval_runtime": 18.6339, + "eval_samples_per_second": 5.367, + "eval_steps_per_second": 1.342, + "step": 200 + }, + { + "epoch": 0.6496969696969697, + "grad_norm": 0.006153900176286697, + "learning_rate": 2.6002587322121607e-05, + "loss": 0.0016, + "step": 201 + }, + { + "epoch": 0.6529292929292929, + "grad_norm": 0.0070861754938960075, + "learning_rate": 2.613195342820181e-05, + "loss": 0.002, + "step": 202 + }, + { + "epoch": 0.6561616161616162, + "grad_norm": 0.0048012156039476395, + "learning_rate": 2.626131953428202e-05, + "loss": 0.0012, + "step": 203 + }, + { + "epoch": 0.6593939393939394, + "grad_norm": 0.006167956627905369, + "learning_rate": 2.6390685640362228e-05, + "loss": 0.0018, + "step": 204 + }, + { + "epoch": 0.6626262626262627, + "grad_norm": 0.0051431735046207905, + "learning_rate": 2.652005174644243e-05, + "loss": 0.0013, + "step": 205 + }, + { + "epoch": 0.6658585858585858, + "grad_norm": 0.009899413213133812, + "learning_rate": 2.664941785252264e-05, + "loss": 0.0031, + "step": 206 + }, + { + "epoch": 0.6690909090909091, + "grad_norm": 0.007686274591833353, + "learning_rate": 2.677878395860285e-05, + "loss": 0.0022, + "step": 207 + }, + { + "epoch": 0.6723232323232323, + "grad_norm": 0.004816455766558647, + "learning_rate": 2.6908150064683052e-05, + "loss": 0.0011, + "step": 208 + }, + { + "epoch": 0.6755555555555556, + "grad_norm": 0.004873145837336779, + "learning_rate": 2.7037516170763262e-05, + "loss": 0.0011, + "step": 209 + }, + { + "epoch": 0.6787878787878788, + "grad_norm": 0.006136384792625904, + "learning_rate": 2.716688227684347e-05, + "loss": 0.0016, + "step": 210 + }, + { + "epoch": 0.682020202020202, + "grad_norm": 0.004508585669100285, + "learning_rate": 2.7296248382923673e-05, + "loss": 0.0012, + "step": 211 + }, + { + "epoch": 0.6852525252525252, + "grad_norm": 0.005723617039620876, + "learning_rate": 2.7425614489003883e-05, + "loss": 0.0016, + "step": 212 + }, + { + "epoch": 0.6884848484848485, + "grad_norm": 0.0072562843561172485, + "learning_rate": 2.755498059508409e-05, + "loss": 0.0015, + "step": 213 + }, + { + "epoch": 0.6917171717171717, + "grad_norm": 0.006079181097447872, + "learning_rate": 2.76843467011643e-05, + "loss": 0.0017, + "step": 214 + }, + { + "epoch": 0.694949494949495, + "grad_norm": 0.011977693997323513, + "learning_rate": 2.7813712807244503e-05, + "loss": 0.0048, + "step": 215 + }, + { + "epoch": 0.6981818181818182, + "grad_norm": 0.007406435441225767, + "learning_rate": 2.794307891332471e-05, + "loss": 0.0022, + "step": 216 + }, + { + "epoch": 0.7014141414141414, + "grad_norm": 0.006909268908202648, + "learning_rate": 2.807244501940492e-05, + "loss": 0.0022, + "step": 217 + }, + { + "epoch": 0.7046464646464646, + "grad_norm": 0.0051758866757154465, + "learning_rate": 2.8201811125485124e-05, + "loss": 0.0012, + "step": 218 + }, + { + "epoch": 0.7078787878787879, + "grad_norm": 0.00657995231449604, + "learning_rate": 2.833117723156533e-05, + "loss": 0.0014, + "step": 219 + }, + { + "epoch": 0.7111111111111111, + "grad_norm": 0.006561241112649441, + "learning_rate": 2.846054333764554e-05, + "loss": 0.0018, + "step": 220 + }, + { + "epoch": 0.7143434343434344, + "grad_norm": 0.005906842183321714, + "learning_rate": 2.8589909443725745e-05, + "loss": 0.0013, + "step": 221 + }, + { + "epoch": 0.7175757575757575, + "grad_norm": 0.004000976216048002, + "learning_rate": 2.871927554980595e-05, + "loss": 0.0009, + "step": 222 + }, + { + "epoch": 0.7208080808080808, + "grad_norm": 0.006313610821962357, + "learning_rate": 2.8848641655886162e-05, + "loss": 0.0015, + "step": 223 + }, + { + "epoch": 0.724040404040404, + "grad_norm": 0.005339731462299824, + "learning_rate": 2.8978007761966365e-05, + "loss": 0.0014, + "step": 224 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 0.00704893097281456, + "learning_rate": 2.9107373868046572e-05, + "loss": 0.0017, + "step": 225 + }, + { + "epoch": 0.7272727272727273, + "eval_loss": 0.0014526441227644682, + "eval_runtime": 18.5916, + "eval_samples_per_second": 5.379, + "eval_steps_per_second": 1.345, + "step": 225 + }, + { + "epoch": 0.7305050505050505, + "grad_norm": 0.007863117381930351, + "learning_rate": 2.9236739974126783e-05, + "loss": 0.002, + "step": 226 + }, + { + "epoch": 0.7337373737373737, + "grad_norm": 0.007929647341370583, + "learning_rate": 2.936610608020699e-05, + "loss": 0.0014, + "step": 227 + }, + { + "epoch": 0.7369696969696969, + "grad_norm": 0.008834928274154663, + "learning_rate": 2.9495472186287193e-05, + "loss": 0.0033, + "step": 228 + }, + { + "epoch": 0.7402020202020202, + "grad_norm": 0.005479221232235432, + "learning_rate": 2.9624838292367403e-05, + "loss": 0.0013, + "step": 229 + }, + { + "epoch": 0.7434343434343434, + "grad_norm": 0.008690214715898037, + "learning_rate": 2.975420439844761e-05, + "loss": 0.0031, + "step": 230 + }, + { + "epoch": 0.7466666666666667, + "grad_norm": 0.006373909767717123, + "learning_rate": 2.9883570504527814e-05, + "loss": 0.0013, + "step": 231 + }, + { + "epoch": 0.74989898989899, + "grad_norm": 0.005924658849835396, + "learning_rate": 3.0012936610608024e-05, + "loss": 0.0014, + "step": 232 + }, + { + "epoch": 0.7531313131313131, + "grad_norm": 0.007938825525343418, + "learning_rate": 3.014230271668823e-05, + "loss": 0.0027, + "step": 233 + }, + { + "epoch": 0.7563636363636363, + "grad_norm": 0.007831827737390995, + "learning_rate": 3.0271668822768434e-05, + "loss": 0.0025, + "step": 234 + }, + { + "epoch": 0.7595959595959596, + "grad_norm": 0.008570835925638676, + "learning_rate": 3.0401034928848644e-05, + "loss": 0.003, + "step": 235 + }, + { + "epoch": 0.7628282828282829, + "grad_norm": 0.005871222820132971, + "learning_rate": 3.053040103492885e-05, + "loss": 0.0013, + "step": 236 + }, + { + "epoch": 0.7660606060606061, + "grad_norm": 0.007041990291327238, + "learning_rate": 3.0659767141009055e-05, + "loss": 0.0028, + "step": 237 + }, + { + "epoch": 0.7692929292929293, + "grad_norm": 0.0073891859501600266, + "learning_rate": 3.0789133247089265e-05, + "loss": 0.0023, + "step": 238 + }, + { + "epoch": 0.7725252525252525, + "grad_norm": 0.005527487024664879, + "learning_rate": 3.0918499353169475e-05, + "loss": 0.0014, + "step": 239 + }, + { + "epoch": 0.7757575757575758, + "grad_norm": 0.007551091257482767, + "learning_rate": 3.104786545924967e-05, + "loss": 0.002, + "step": 240 + }, + { + "epoch": 0.778989898989899, + "grad_norm": 0.005599314346909523, + "learning_rate": 3.117723156532988e-05, + "loss": 0.002, + "step": 241 + }, + { + "epoch": 0.7822222222222223, + "grad_norm": 0.007462210953235626, + "learning_rate": 3.130659767141009e-05, + "loss": 0.0019, + "step": 242 + }, + { + "epoch": 0.7854545454545454, + "grad_norm": 0.004866201896220446, + "learning_rate": 3.14359637774903e-05, + "loss": 0.0012, + "step": 243 + }, + { + "epoch": 0.7886868686868687, + "grad_norm": 0.005268635228276253, + "learning_rate": 3.1565329883570506e-05, + "loss": 0.0013, + "step": 244 + }, + { + "epoch": 0.7919191919191919, + "grad_norm": 0.007261104416102171, + "learning_rate": 3.169469598965072e-05, + "loss": 0.0031, + "step": 245 + }, + { + "epoch": 0.7951515151515152, + "grad_norm": 0.006308354903012514, + "learning_rate": 3.182406209573092e-05, + "loss": 0.0014, + "step": 246 + }, + { + "epoch": 0.7983838383838384, + "grad_norm": 0.0074616689234972, + "learning_rate": 3.1953428201811124e-05, + "loss": 0.0015, + "step": 247 + }, + { + "epoch": 0.8016161616161617, + "grad_norm": 0.007333788555115461, + "learning_rate": 3.2082794307891334e-05, + "loss": 0.0014, + "step": 248 + }, + { + "epoch": 0.8048484848484848, + "grad_norm": 0.006965926848351955, + "learning_rate": 3.2212160413971544e-05, + "loss": 0.0018, + "step": 249 + }, + { + "epoch": 0.8080808080808081, + "grad_norm": 0.00535194855183363, + "learning_rate": 3.234152652005175e-05, + "loss": 0.0015, + "step": 250 + }, + { + "epoch": 0.8080808080808081, + "eval_loss": 0.0013973835157230496, + "eval_runtime": 18.6113, + "eval_samples_per_second": 5.373, + "eval_steps_per_second": 1.343, + "step": 250 + }, + { + "epoch": 0.8113131313131313, + "grad_norm": 0.006415734998881817, + "learning_rate": 3.247089262613196e-05, + "loss": 0.0014, + "step": 251 + }, + { + "epoch": 0.8145454545454546, + "grad_norm": 0.00617300346493721, + "learning_rate": 3.260025873221216e-05, + "loss": 0.0017, + "step": 252 + }, + { + "epoch": 0.8177777777777778, + "grad_norm": 0.007527848239988089, + "learning_rate": 3.2729624838292365e-05, + "loss": 0.0017, + "step": 253 + }, + { + "epoch": 0.821010101010101, + "grad_norm": 0.00747877499088645, + "learning_rate": 3.2858990944372575e-05, + "loss": 0.0014, + "step": 254 + }, + { + "epoch": 0.8242424242424242, + "grad_norm": 0.005451089236885309, + "learning_rate": 3.2988357050452786e-05, + "loss": 0.0012, + "step": 255 + }, + { + "epoch": 0.8274747474747475, + "grad_norm": 0.006235205102711916, + "learning_rate": 3.3117723156532996e-05, + "loss": 0.0014, + "step": 256 + }, + { + "epoch": 0.8307070707070707, + "grad_norm": 0.005764327012002468, + "learning_rate": 3.324708926261319e-05, + "loss": 0.0013, + "step": 257 + }, + { + "epoch": 0.833939393939394, + "grad_norm": 0.007408145349472761, + "learning_rate": 3.33764553686934e-05, + "loss": 0.0013, + "step": 258 + }, + { + "epoch": 0.8371717171717171, + "grad_norm": 0.004855205304920673, + "learning_rate": 3.350582147477361e-05, + "loss": 0.0009, + "step": 259 + }, + { + "epoch": 0.8404040404040404, + "grad_norm": 0.010664681904017925, + "learning_rate": 3.3635187580853817e-05, + "loss": 0.0016, + "step": 260 + }, + { + "epoch": 0.8436363636363636, + "grad_norm": 0.006055675912648439, + "learning_rate": 3.376455368693403e-05, + "loss": 0.0016, + "step": 261 + }, + { + "epoch": 0.8468686868686869, + "grad_norm": 0.01056759525090456, + "learning_rate": 3.389391979301424e-05, + "loss": 0.0016, + "step": 262 + }, + { + "epoch": 0.8501010101010101, + "grad_norm": 0.007450331002473831, + "learning_rate": 3.4023285899094434e-05, + "loss": 0.0016, + "step": 263 + }, + { + "epoch": 0.8533333333333334, + "grad_norm": 0.0053154826164245605, + "learning_rate": 3.4152652005174644e-05, + "loss": 0.0011, + "step": 264 + }, + { + "epoch": 0.8565656565656565, + "grad_norm": 0.0060371761210262775, + "learning_rate": 3.4282018111254854e-05, + "loss": 0.0013, + "step": 265 + }, + { + "epoch": 0.8597979797979798, + "grad_norm": 0.005466979928314686, + "learning_rate": 3.441138421733506e-05, + "loss": 0.0012, + "step": 266 + }, + { + "epoch": 0.863030303030303, + "grad_norm": 0.00831522885710001, + "learning_rate": 3.454075032341527e-05, + "loss": 0.0014, + "step": 267 + }, + { + "epoch": 0.8662626262626263, + "grad_norm": 0.007963420823216438, + "learning_rate": 3.467011642949548e-05, + "loss": 0.0025, + "step": 268 + }, + { + "epoch": 0.8694949494949495, + "grad_norm": 0.004937573801726103, + "learning_rate": 3.4799482535575675e-05, + "loss": 0.0011, + "step": 269 + }, + { + "epoch": 0.8727272727272727, + "grad_norm": 0.0058564417995512486, + "learning_rate": 3.4928848641655885e-05, + "loss": 0.0016, + "step": 270 + }, + { + "epoch": 0.8759595959595959, + "grad_norm": 0.007912451401352882, + "learning_rate": 3.5058214747736096e-05, + "loss": 0.0026, + "step": 271 + }, + { + "epoch": 0.8791919191919192, + "grad_norm": 0.003929544240236282, + "learning_rate": 3.5187580853816306e-05, + "loss": 0.0011, + "step": 272 + }, + { + "epoch": 0.8824242424242424, + "grad_norm": 0.008150831796228886, + "learning_rate": 3.531694695989651e-05, + "loss": 0.0018, + "step": 273 + }, + { + "epoch": 0.8856565656565657, + "grad_norm": 0.009232031181454659, + "learning_rate": 3.544631306597671e-05, + "loss": 0.0042, + "step": 274 + }, + { + "epoch": 0.8888888888888888, + "grad_norm": 0.007469587959349155, + "learning_rate": 3.557567917205692e-05, + "loss": 0.0023, + "step": 275 + }, + { + "epoch": 0.8888888888888888, + "eval_loss": 0.0013014678843319416, + "eval_runtime": 18.7103, + "eval_samples_per_second": 5.345, + "eval_steps_per_second": 1.336, + "step": 275 + }, + { + "epoch": 0.8921212121212121, + "grad_norm": 0.006450500804930925, + "learning_rate": 3.570504527813713e-05, + "loss": 0.0014, + "step": 276 + }, + { + "epoch": 0.8953535353535353, + "grad_norm": 0.004984740167856216, + "learning_rate": 3.583441138421734e-05, + "loss": 0.0014, + "step": 277 + }, + { + "epoch": 0.8985858585858586, + "grad_norm": 0.006635539699345827, + "learning_rate": 3.596377749029755e-05, + "loss": 0.0016, + "step": 278 + }, + { + "epoch": 0.9018181818181819, + "grad_norm": 0.010307137854397297, + "learning_rate": 3.609314359637775e-05, + "loss": 0.0025, + "step": 279 + }, + { + "epoch": 0.9050505050505051, + "grad_norm": 0.008318118751049042, + "learning_rate": 3.6222509702457954e-05, + "loss": 0.0028, + "step": 280 + }, + { + "epoch": 0.9082828282828282, + "grad_norm": 0.007523260544985533, + "learning_rate": 3.6351875808538164e-05, + "loss": 0.0022, + "step": 281 + }, + { + "epoch": 0.9115151515151515, + "grad_norm": 0.00529656745493412, + "learning_rate": 3.648124191461837e-05, + "loss": 0.001, + "step": 282 + }, + { + "epoch": 0.9147474747474748, + "grad_norm": 0.0050112344324588776, + "learning_rate": 3.661060802069858e-05, + "loss": 0.0012, + "step": 283 + }, + { + "epoch": 0.917979797979798, + "grad_norm": 0.006318412255495787, + "learning_rate": 3.673997412677879e-05, + "loss": 0.0017, + "step": 284 + }, + { + "epoch": 0.9212121212121213, + "grad_norm": 0.006407878361642361, + "learning_rate": 3.6869340232859e-05, + "loss": 0.0022, + "step": 285 + }, + { + "epoch": 0.9244444444444444, + "grad_norm": 0.004356731195002794, + "learning_rate": 3.6998706338939195e-05, + "loss": 0.0007, + "step": 286 + }, + { + "epoch": 0.9276767676767677, + "grad_norm": 0.004966625012457371, + "learning_rate": 3.7128072445019406e-05, + "loss": 0.0012, + "step": 287 + }, + { + "epoch": 0.9309090909090909, + "grad_norm": 0.0055012330412864685, + "learning_rate": 3.7257438551099616e-05, + "loss": 0.0011, + "step": 288 + }, + { + "epoch": 0.9341414141414142, + "grad_norm": 0.006766812410205603, + "learning_rate": 3.738680465717982e-05, + "loss": 0.0017, + "step": 289 + }, + { + "epoch": 0.9373737373737374, + "grad_norm": 0.006011553108692169, + "learning_rate": 3.751617076326003e-05, + "loss": 0.0022, + "step": 290 + }, + { + "epoch": 0.9406060606060606, + "grad_norm": 0.0056586177088320255, + "learning_rate": 3.764553686934023e-05, + "loss": 0.0014, + "step": 291 + }, + { + "epoch": 0.9438383838383838, + "grad_norm": 0.0064941453747451305, + "learning_rate": 3.777490297542044e-05, + "loss": 0.0011, + "step": 292 + }, + { + "epoch": 0.9470707070707071, + "grad_norm": 0.006994554307311773, + "learning_rate": 3.790426908150065e-05, + "loss": 0.0021, + "step": 293 + }, + { + "epoch": 0.9503030303030303, + "grad_norm": 0.006325817201286554, + "learning_rate": 3.803363518758086e-05, + "loss": 0.0014, + "step": 294 + }, + { + "epoch": 0.9535353535353536, + "grad_norm": 0.006501290015876293, + "learning_rate": 3.816300129366106e-05, + "loss": 0.0017, + "step": 295 + }, + { + "epoch": 0.9567676767676768, + "grad_norm": 0.006014223676174879, + "learning_rate": 3.829236739974127e-05, + "loss": 0.0013, + "step": 296 + }, + { + "epoch": 0.96, + "grad_norm": 0.005024549085646868, + "learning_rate": 3.8421733505821475e-05, + "loss": 0.0013, + "step": 297 + }, + { + "epoch": 0.9632323232323232, + "grad_norm": 0.00840625911951065, + "learning_rate": 3.855109961190168e-05, + "loss": 0.0017, + "step": 298 + }, + { + "epoch": 0.9664646464646465, + "grad_norm": 0.005791939329355955, + "learning_rate": 3.868046571798189e-05, + "loss": 0.0014, + "step": 299 + }, + { + "epoch": 0.9696969696969697, + "grad_norm": 0.0047224657610058784, + "learning_rate": 3.88098318240621e-05, + "loss": 0.0012, + "step": 300 + }, + { + "epoch": 0.9696969696969697, + "eval_loss": 0.0012624365044757724, + "eval_runtime": 18.7089, + "eval_samples_per_second": 5.345, + "eval_steps_per_second": 1.336, + "step": 300 + }, + { + "epoch": 0.972929292929293, + "grad_norm": 0.005727376788854599, + "learning_rate": 3.893919793014231e-05, + "loss": 0.0013, + "step": 301 + }, + { + "epoch": 0.9761616161616161, + "grad_norm": 0.007024977821856737, + "learning_rate": 3.906856403622251e-05, + "loss": 0.0018, + "step": 302 + }, + { + "epoch": 0.9793939393939394, + "grad_norm": 0.004745373502373695, + "learning_rate": 3.9197930142302716e-05, + "loss": 0.001, + "step": 303 + }, + { + "epoch": 0.9826262626262626, + "grad_norm": 0.0050310962833464146, + "learning_rate": 3.9327296248382926e-05, + "loss": 0.0013, + "step": 304 + }, + { + "epoch": 0.9858585858585859, + "grad_norm": 0.008561142720282078, + "learning_rate": 3.945666235446313e-05, + "loss": 0.0018, + "step": 305 + }, + { + "epoch": 0.9890909090909091, + "grad_norm": 0.0055971029214560986, + "learning_rate": 3.958602846054334e-05, + "loss": 0.0013, + "step": 306 + }, + { + "epoch": 0.9923232323232323, + "grad_norm": 0.007602418772876263, + "learning_rate": 3.971539456662355e-05, + "loss": 0.0014, + "step": 307 + }, + { + "epoch": 0.9955555555555555, + "grad_norm": 0.0054479725658893585, + "learning_rate": 3.9844760672703754e-05, + "loss": 0.0014, + "step": 308 + }, + { + "epoch": 0.9987878787878788, + "grad_norm": 0.006376450881361961, + "learning_rate": 3.997412677878396e-05, + "loss": 0.0013, + "step": 309 + }, + { + "epoch": 1.002020202020202, + "grad_norm": 0.00963524729013443, + "learning_rate": 4.010349288486417e-05, + "loss": 0.0025, + "step": 310 + }, + { + "epoch": 1.0052525252525253, + "grad_norm": 0.004950121510773897, + "learning_rate": 4.023285899094437e-05, + "loss": 0.0009, + "step": 311 + }, + { + "epoch": 1.0084848484848485, + "grad_norm": 0.0047010756097733974, + "learning_rate": 4.036222509702458e-05, + "loss": 0.0012, + "step": 312 + }, + { + "epoch": 1.0117171717171718, + "grad_norm": 0.004061982501298189, + "learning_rate": 4.049159120310479e-05, + "loss": 0.0008, + "step": 313 + }, + { + "epoch": 1.014949494949495, + "grad_norm": 0.006061443593353033, + "learning_rate": 4.0620957309184995e-05, + "loss": 0.0022, + "step": 314 + }, + { + "epoch": 1.018181818181818, + "grad_norm": 0.005154592916369438, + "learning_rate": 4.07503234152652e-05, + "loss": 0.0023, + "step": 315 + }, + { + "epoch": 1.0214141414141413, + "grad_norm": 0.004801798611879349, + "learning_rate": 4.087968952134541e-05, + "loss": 0.0012, + "step": 316 + }, + { + "epoch": 1.0246464646464646, + "grad_norm": 0.004275497514754534, + "learning_rate": 4.100905562742562e-05, + "loss": 0.0012, + "step": 317 + }, + { + "epoch": 1.0278787878787878, + "grad_norm": 0.006128865294158459, + "learning_rate": 4.113842173350582e-05, + "loss": 0.0022, + "step": 318 + }, + { + "epoch": 1.031111111111111, + "grad_norm": 0.004465613514184952, + "learning_rate": 4.126778783958603e-05, + "loss": 0.0019, + "step": 319 + }, + { + "epoch": 1.0343434343434343, + "grad_norm": 0.004612160846590996, + "learning_rate": 4.1397153945666236e-05, + "loss": 0.0013, + "step": 320 + }, + { + "epoch": 1.0375757575757576, + "grad_norm": 0.00512789748609066, + "learning_rate": 4.152652005174644e-05, + "loss": 0.0012, + "step": 321 + }, + { + "epoch": 1.0408080808080808, + "grad_norm": 0.006197880953550339, + "learning_rate": 4.165588615782665e-05, + "loss": 0.0015, + "step": 322 + }, + { + "epoch": 1.044040404040404, + "grad_norm": 0.005857696291059256, + "learning_rate": 4.178525226390686e-05, + "loss": 0.0014, + "step": 323 + }, + { + "epoch": 1.0472727272727274, + "grad_norm": 0.010157118551433086, + "learning_rate": 4.1914618369987064e-05, + "loss": 0.0015, + "step": 324 + }, + { + "epoch": 1.0505050505050506, + "grad_norm": 0.006577750667929649, + "learning_rate": 4.2043984476067274e-05, + "loss": 0.0011, + "step": 325 + }, + { + "epoch": 1.0505050505050506, + "eval_loss": 0.0012751913163810968, + "eval_runtime": 18.6453, + "eval_samples_per_second": 5.363, + "eval_steps_per_second": 1.341, + "step": 325 + }, + { + "epoch": 1.0537373737373736, + "grad_norm": 0.004863920155912638, + "learning_rate": 4.217335058214748e-05, + "loss": 0.0009, + "step": 326 + }, + { + "epoch": 1.056969696969697, + "grad_norm": 0.005144843365997076, + "learning_rate": 4.230271668822768e-05, + "loss": 0.0013, + "step": 327 + }, + { + "epoch": 1.0602020202020201, + "grad_norm": 0.005289267282932997, + "learning_rate": 4.243208279430789e-05, + "loss": 0.0012, + "step": 328 + }, + { + "epoch": 1.0634343434343434, + "grad_norm": 0.005019436590373516, + "learning_rate": 4.25614489003881e-05, + "loss": 0.0011, + "step": 329 + }, + { + "epoch": 1.0666666666666667, + "grad_norm": 0.0038786993827670813, + "learning_rate": 4.269081500646831e-05, + "loss": 0.0008, + "step": 330 + }, + { + "epoch": 1.06989898989899, + "grad_norm": 0.005253227427601814, + "learning_rate": 4.2820181112548515e-05, + "loss": 0.0012, + "step": 331 + }, + { + "epoch": 1.0731313131313132, + "grad_norm": 0.005880107171833515, + "learning_rate": 4.294954721862872e-05, + "loss": 0.0016, + "step": 332 + }, + { + "epoch": 1.0763636363636364, + "grad_norm": 0.006426130421459675, + "learning_rate": 4.307891332470893e-05, + "loss": 0.0009, + "step": 333 + }, + { + "epoch": 1.0795959595959597, + "grad_norm": 0.0038312741089612246, + "learning_rate": 4.320827943078913e-05, + "loss": 0.0009, + "step": 334 + }, + { + "epoch": 1.082828282828283, + "grad_norm": 0.005139067303389311, + "learning_rate": 4.333764553686934e-05, + "loss": 0.0011, + "step": 335 + }, + { + "epoch": 1.086060606060606, + "grad_norm": 0.006610156502574682, + "learning_rate": 4.346701164294955e-05, + "loss": 0.0019, + "step": 336 + }, + { + "epoch": 1.0892929292929292, + "grad_norm": 0.003907322883605957, + "learning_rate": 4.359637774902976e-05, + "loss": 0.001, + "step": 337 + }, + { + "epoch": 1.0925252525252525, + "grad_norm": 0.006776052061468363, + "learning_rate": 4.372574385510996e-05, + "loss": 0.0011, + "step": 338 + }, + { + "epoch": 1.0957575757575757, + "grad_norm": 0.007625575177371502, + "learning_rate": 4.385510996119017e-05, + "loss": 0.0016, + "step": 339 + }, + { + "epoch": 1.098989898989899, + "grad_norm": 0.012086856178939342, + "learning_rate": 4.3984476067270374e-05, + "loss": 0.0035, + "step": 340 + }, + { + "epoch": 1.1022222222222222, + "grad_norm": 0.007610958535224199, + "learning_rate": 4.4113842173350584e-05, + "loss": 0.0014, + "step": 341 + }, + { + "epoch": 1.1054545454545455, + "grad_norm": 0.006693006958812475, + "learning_rate": 4.4243208279430794e-05, + "loss": 0.0013, + "step": 342 + }, + { + "epoch": 1.1086868686868687, + "grad_norm": 0.0045067970640957355, + "learning_rate": 4.4372574385511e-05, + "loss": 0.0011, + "step": 343 + }, + { + "epoch": 1.111919191919192, + "grad_norm": 0.006877266336232424, + "learning_rate": 4.45019404915912e-05, + "loss": 0.001, + "step": 344 + }, + { + "epoch": 1.1151515151515152, + "grad_norm": 0.0043833632953464985, + "learning_rate": 4.463130659767141e-05, + "loss": 0.0011, + "step": 345 + }, + { + "epoch": 1.1183838383838385, + "grad_norm": 0.004103951156139374, + "learning_rate": 4.476067270375162e-05, + "loss": 0.0009, + "step": 346 + }, + { + "epoch": 1.1216161616161617, + "grad_norm": 0.006397966295480728, + "learning_rate": 4.4890038809831825e-05, + "loss": 0.0008, + "step": 347 + }, + { + "epoch": 1.1248484848484848, + "grad_norm": 0.007373814936727285, + "learning_rate": 4.5019404915912036e-05, + "loss": 0.0016, + "step": 348 + }, + { + "epoch": 1.128080808080808, + "grad_norm": 0.005406413692981005, + "learning_rate": 4.514877102199224e-05, + "loss": 0.0012, + "step": 349 + }, + { + "epoch": 1.1313131313131313, + "grad_norm": 0.0059000300243496895, + "learning_rate": 4.527813712807244e-05, + "loss": 0.0011, + "step": 350 + }, + { + "epoch": 1.1313131313131313, + "eval_loss": 0.0012826790334656835, + "eval_runtime": 18.7277, + "eval_samples_per_second": 5.34, + "eval_steps_per_second": 1.335, + "step": 350 + }, + { + "epoch": 1.1345454545454545, + "grad_norm": 0.007405865006148815, + "learning_rate": 4.540750323415265e-05, + "loss": 0.0026, + "step": 351 + }, + { + "epoch": 1.1377777777777778, + "grad_norm": 0.005577345844358206, + "learning_rate": 4.553686934023286e-05, + "loss": 0.0018, + "step": 352 + }, + { + "epoch": 1.141010101010101, + "grad_norm": 0.006122430320829153, + "learning_rate": 4.566623544631307e-05, + "loss": 0.0015, + "step": 353 + }, + { + "epoch": 1.1442424242424243, + "grad_norm": 0.007260697893798351, + "learning_rate": 4.579560155239328e-05, + "loss": 0.0012, + "step": 354 + }, + { + "epoch": 1.1474747474747475, + "grad_norm": 0.007470736745744944, + "learning_rate": 4.592496765847348e-05, + "loss": 0.0032, + "step": 355 + }, + { + "epoch": 1.1507070707070708, + "grad_norm": 0.0061160847544670105, + "learning_rate": 4.6054333764553684e-05, + "loss": 0.0013, + "step": 356 + }, + { + "epoch": 1.1539393939393938, + "grad_norm": 0.006069364957511425, + "learning_rate": 4.6183699870633894e-05, + "loss": 0.0011, + "step": 357 + }, + { + "epoch": 1.157171717171717, + "grad_norm": 0.004093985538929701, + "learning_rate": 4.6313065976714105e-05, + "loss": 0.0005, + "step": 358 + }, + { + "epoch": 1.1604040404040403, + "grad_norm": 0.0061876364052295685, + "learning_rate": 4.6442432082794315e-05, + "loss": 0.0017, + "step": 359 + }, + { + "epoch": 1.1636363636363636, + "grad_norm": 0.0048703462816774845, + "learning_rate": 4.657179818887452e-05, + "loss": 0.0009, + "step": 360 + }, + { + "epoch": 1.1668686868686868, + "grad_norm": 0.007618950214236975, + "learning_rate": 4.670116429495472e-05, + "loss": 0.0019, + "step": 361 + }, + { + "epoch": 1.17010101010101, + "grad_norm": 0.004160857293754816, + "learning_rate": 4.683053040103493e-05, + "loss": 0.001, + "step": 362 + }, + { + "epoch": 1.1733333333333333, + "grad_norm": 0.007051107473671436, + "learning_rate": 4.6959896507115136e-05, + "loss": 0.0019, + "step": 363 + }, + { + "epoch": 1.1765656565656566, + "grad_norm": 0.004829261917620897, + "learning_rate": 4.7089262613195346e-05, + "loss": 0.001, + "step": 364 + }, + { + "epoch": 1.1797979797979798, + "grad_norm": 0.005092508625239134, + "learning_rate": 4.7218628719275556e-05, + "loss": 0.0023, + "step": 365 + }, + { + "epoch": 1.183030303030303, + "grad_norm": 0.004417267628014088, + "learning_rate": 4.734799482535576e-05, + "loss": 0.001, + "step": 366 + }, + { + "epoch": 1.1862626262626264, + "grad_norm": 0.005750073119997978, + "learning_rate": 4.747736093143596e-05, + "loss": 0.0013, + "step": 367 + }, + { + "epoch": 1.1894949494949496, + "grad_norm": 0.0033691064454615116, + "learning_rate": 4.760672703751617e-05, + "loss": 0.0008, + "step": 368 + }, + { + "epoch": 1.1927272727272726, + "grad_norm": 0.0044763232581317425, + "learning_rate": 4.773609314359638e-05, + "loss": 0.0009, + "step": 369 + }, + { + "epoch": 1.195959595959596, + "grad_norm": 0.007102371193468571, + "learning_rate": 4.786545924967659e-05, + "loss": 0.0012, + "step": 370 + }, + { + "epoch": 1.1991919191919191, + "grad_norm": 0.0041382270865142345, + "learning_rate": 4.79948253557568e-05, + "loss": 0.001, + "step": 371 + }, + { + "epoch": 1.2024242424242424, + "grad_norm": 0.006266591139137745, + "learning_rate": 4.8124191461837e-05, + "loss": 0.0015, + "step": 372 + }, + { + "epoch": 1.2056565656565656, + "grad_norm": 0.00487930653616786, + "learning_rate": 4.8253557567917204e-05, + "loss": 0.0012, + "step": 373 + }, + { + "epoch": 1.208888888888889, + "grad_norm": 0.005311232525855303, + "learning_rate": 4.8382923673997415e-05, + "loss": 0.0014, + "step": 374 + }, + { + "epoch": 1.2121212121212122, + "grad_norm": 0.004289830103516579, + "learning_rate": 4.8512289780077625e-05, + "loss": 0.0009, + "step": 375 + }, + { + "epoch": 1.2121212121212122, + "eval_loss": 0.0011982680298388004, + "eval_runtime": 18.7329, + "eval_samples_per_second": 5.338, + "eval_steps_per_second": 1.335, + "step": 375 + }, + { + "epoch": 1.2153535353535354, + "grad_norm": 0.0062402524054050446, + "learning_rate": 4.864165588615783e-05, + "loss": 0.0011, + "step": 376 + }, + { + "epoch": 1.2185858585858587, + "grad_norm": 0.0034229829907417297, + "learning_rate": 4.877102199223804e-05, + "loss": 0.0008, + "step": 377 + }, + { + "epoch": 1.221818181818182, + "grad_norm": 0.0063551426865160465, + "learning_rate": 4.890038809831824e-05, + "loss": 0.0027, + "step": 378 + }, + { + "epoch": 1.225050505050505, + "grad_norm": 0.004902714863419533, + "learning_rate": 4.9029754204398446e-05, + "loss": 0.0015, + "step": 379 + }, + { + "epoch": 1.2282828282828282, + "grad_norm": 0.0044431439600884914, + "learning_rate": 4.9159120310478656e-05, + "loss": 0.0005, + "step": 380 + }, + { + "epoch": 1.2315151515151515, + "grad_norm": 0.007968113757669926, + "learning_rate": 4.9288486416558866e-05, + "loss": 0.0013, + "step": 381 + }, + { + "epoch": 1.2347474747474747, + "grad_norm": 0.007964645512402058, + "learning_rate": 4.941785252263907e-05, + "loss": 0.0033, + "step": 382 + }, + { + "epoch": 1.237979797979798, + "grad_norm": 0.004984153900295496, + "learning_rate": 4.954721862871928e-05, + "loss": 0.0008, + "step": 383 + }, + { + "epoch": 1.2412121212121212, + "grad_norm": 0.005040554329752922, + "learning_rate": 4.9676584734799483e-05, + "loss": 0.0008, + "step": 384 + }, + { + "epoch": 1.2444444444444445, + "grad_norm": 0.005269059911370277, + "learning_rate": 4.980595084087969e-05, + "loss": 0.0016, + "step": 385 + }, + { + "epoch": 1.2476767676767677, + "grad_norm": 0.004820747766643763, + "learning_rate": 4.99353169469599e-05, + "loss": 0.0013, + "step": 386 + }, + { + "epoch": 1.250909090909091, + "grad_norm": 0.003650473430752754, + "learning_rate": 5.006468305304011e-05, + "loss": 0.001, + "step": 387 + }, + { + "epoch": 1.2541414141414142, + "grad_norm": 0.003410038072615862, + "learning_rate": 5.019404915912032e-05, + "loss": 0.0007, + "step": 388 + }, + { + "epoch": 1.2573737373737375, + "grad_norm": 0.00503958947956562, + "learning_rate": 5.032341526520052e-05, + "loss": 0.0013, + "step": 389 + }, + { + "epoch": 1.2606060606060607, + "grad_norm": 0.006551372352987528, + "learning_rate": 5.0452781371280725e-05, + "loss": 0.0017, + "step": 390 + }, + { + "epoch": 1.2638383838383838, + "grad_norm": 0.004784473218023777, + "learning_rate": 5.058214747736093e-05, + "loss": 0.0011, + "step": 391 + }, + { + "epoch": 1.267070707070707, + "grad_norm": 0.009283591993153095, + "learning_rate": 5.071151358344114e-05, + "loss": 0.0017, + "step": 392 + }, + { + "epoch": 1.2703030303030303, + "grad_norm": 0.005054951179772615, + "learning_rate": 5.084087968952135e-05, + "loss": 0.0014, + "step": 393 + }, + { + "epoch": 1.2735353535353535, + "grad_norm": 0.007423613220453262, + "learning_rate": 5.097024579560156e-05, + "loss": 0.0012, + "step": 394 + }, + { + "epoch": 1.2767676767676768, + "grad_norm": 0.004986994434148073, + "learning_rate": 5.109961190168176e-05, + "loss": 0.0017, + "step": 395 + }, + { + "epoch": 1.28, + "grad_norm": 0.004347431473433971, + "learning_rate": 5.122897800776197e-05, + "loss": 0.0012, + "step": 396 + }, + { + "epoch": 1.2832323232323233, + "grad_norm": 0.006622066255658865, + "learning_rate": 5.135834411384217e-05, + "loss": 0.0015, + "step": 397 + }, + { + "epoch": 1.2864646464646465, + "grad_norm": 0.006094200070947409, + "learning_rate": 5.148771021992238e-05, + "loss": 0.0014, + "step": 398 + }, + { + "epoch": 1.2896969696969696, + "grad_norm": 0.004648863337934017, + "learning_rate": 5.161707632600259e-05, + "loss": 0.0012, + "step": 399 + }, + { + "epoch": 1.2929292929292928, + "grad_norm": 0.005023527424782515, + "learning_rate": 5.17464424320828e-05, + "loss": 0.0015, + "step": 400 + }, + { + "epoch": 1.2929292929292928, + "eval_loss": 0.001097235712222755, + "eval_runtime": 18.6509, + "eval_samples_per_second": 5.362, + "eval_steps_per_second": 1.34, + "step": 400 + }, + { + "epoch": 1.296161616161616, + "grad_norm": 0.004932883661240339, + "learning_rate": 5.1875808538163004e-05, + "loss": 0.0012, + "step": 401 + }, + { + "epoch": 1.2993939393939393, + "grad_norm": 0.004028505180031061, + "learning_rate": 5.2005174644243214e-05, + "loss": 0.0013, + "step": 402 + }, + { + "epoch": 1.3026262626262626, + "grad_norm": 0.004800683818757534, + "learning_rate": 5.213454075032341e-05, + "loss": 0.0014, + "step": 403 + }, + { + "epoch": 1.3058585858585858, + "grad_norm": 0.0042937519028782845, + "learning_rate": 5.226390685640362e-05, + "loss": 0.001, + "step": 404 + }, + { + "epoch": 1.309090909090909, + "grad_norm": 0.004741652403026819, + "learning_rate": 5.239327296248383e-05, + "loss": 0.0011, + "step": 405 + }, + { + "epoch": 1.3123232323232323, + "grad_norm": 0.004214128013700247, + "learning_rate": 5.252263906856404e-05, + "loss": 0.0009, + "step": 406 + }, + { + "epoch": 1.3155555555555556, + "grad_norm": 0.004764182958751917, + "learning_rate": 5.2652005174644245e-05, + "loss": 0.0011, + "step": 407 + }, + { + "epoch": 1.3187878787878788, + "grad_norm": 0.005405279342085123, + "learning_rate": 5.2781371280724455e-05, + "loss": 0.0011, + "step": 408 + }, + { + "epoch": 1.322020202020202, + "grad_norm": 0.005355632398277521, + "learning_rate": 5.2910737386804666e-05, + "loss": 0.0012, + "step": 409 + }, + { + "epoch": 1.3252525252525253, + "grad_norm": 0.004391840659081936, + "learning_rate": 5.304010349288486e-05, + "loss": 0.0011, + "step": 410 + }, + { + "epoch": 1.3284848484848486, + "grad_norm": 0.002719137817621231, + "learning_rate": 5.316946959896507e-05, + "loss": 0.0005, + "step": 411 + }, + { + "epoch": 1.3317171717171719, + "grad_norm": 0.005910648964345455, + "learning_rate": 5.329883570504528e-05, + "loss": 0.0014, + "step": 412 + }, + { + "epoch": 1.3349494949494949, + "grad_norm": 0.004575495608150959, + "learning_rate": 5.3428201811125486e-05, + "loss": 0.0013, + "step": 413 + }, + { + "epoch": 1.3381818181818181, + "grad_norm": 0.004968920256942511, + "learning_rate": 5.35575679172057e-05, + "loss": 0.0012, + "step": 414 + }, + { + "epoch": 1.3414141414141414, + "grad_norm": 0.005680674687027931, + "learning_rate": 5.368693402328591e-05, + "loss": 0.0015, + "step": 415 + }, + { + "epoch": 1.3446464646464646, + "grad_norm": 0.005805244669318199, + "learning_rate": 5.3816300129366104e-05, + "loss": 0.0017, + "step": 416 + }, + { + "epoch": 1.347878787878788, + "grad_norm": 0.004365504253655672, + "learning_rate": 5.3945666235446314e-05, + "loss": 0.001, + "step": 417 + }, + { + "epoch": 1.3511111111111112, + "grad_norm": 0.004444981925189495, + "learning_rate": 5.4075032341526524e-05, + "loss": 0.0011, + "step": 418 + }, + { + "epoch": 1.3543434343434344, + "grad_norm": 0.0048050908371806145, + "learning_rate": 5.420439844760673e-05, + "loss": 0.0011, + "step": 419 + }, + { + "epoch": 1.3575757575757577, + "grad_norm": 0.0112074613571167, + "learning_rate": 5.433376455368694e-05, + "loss": 0.0009, + "step": 420 + }, + { + "epoch": 1.3608080808080807, + "grad_norm": 0.005805397406220436, + "learning_rate": 5.446313065976715e-05, + "loss": 0.001, + "step": 421 + }, + { + "epoch": 1.364040404040404, + "grad_norm": 0.006793915294110775, + "learning_rate": 5.4592496765847345e-05, + "loss": 0.0018, + "step": 422 + }, + { + "epoch": 1.3672727272727272, + "grad_norm": 0.005133500322699547, + "learning_rate": 5.4721862871927555e-05, + "loss": 0.0015, + "step": 423 + }, + { + "epoch": 1.3705050505050504, + "grad_norm": 0.0034578884951770306, + "learning_rate": 5.4851228978007766e-05, + "loss": 0.0008, + "step": 424 + }, + { + "epoch": 1.3737373737373737, + "grad_norm": 0.005579716991633177, + "learning_rate": 5.498059508408797e-05, + "loss": 0.0025, + "step": 425 + }, + { + "epoch": 1.3737373737373737, + "eval_loss": 0.0011115286033600569, + "eval_runtime": 18.7263, + "eval_samples_per_second": 5.34, + "eval_steps_per_second": 1.335, + "step": 425 + }, + { + "epoch": 1.376969696969697, + "grad_norm": 0.005262942984700203, + "learning_rate": 5.510996119016818e-05, + "loss": 0.0011, + "step": 426 + }, + { + "epoch": 1.3802020202020202, + "grad_norm": 0.004813515581190586, + "learning_rate": 5.523932729624839e-05, + "loss": 0.0011, + "step": 427 + }, + { + "epoch": 1.3834343434343435, + "grad_norm": 0.005556623917073011, + "learning_rate": 5.53686934023286e-05, + "loss": 0.0014, + "step": 428 + }, + { + "epoch": 1.3866666666666667, + "grad_norm": 0.0028185530100017786, + "learning_rate": 5.5498059508408797e-05, + "loss": 0.0007, + "step": 429 + }, + { + "epoch": 1.38989898989899, + "grad_norm": 0.004296639934182167, + "learning_rate": 5.562742561448901e-05, + "loss": 0.001, + "step": 430 + }, + { + "epoch": 1.3931313131313132, + "grad_norm": 0.004092990420758724, + "learning_rate": 5.575679172056921e-05, + "loss": 0.001, + "step": 431 + }, + { + "epoch": 1.3963636363636365, + "grad_norm": 0.005713030230253935, + "learning_rate": 5.588615782664942e-05, + "loss": 0.0015, + "step": 432 + }, + { + "epoch": 1.3995959595959597, + "grad_norm": 0.005011020693928003, + "learning_rate": 5.601552393272963e-05, + "loss": 0.0012, + "step": 433 + }, + { + "epoch": 1.4028282828282828, + "grad_norm": 0.004393043462187052, + "learning_rate": 5.614489003880984e-05, + "loss": 0.0009, + "step": 434 + }, + { + "epoch": 1.406060606060606, + "grad_norm": 0.0047727166675031185, + "learning_rate": 5.627425614489004e-05, + "loss": 0.0011, + "step": 435 + }, + { + "epoch": 1.4092929292929293, + "grad_norm": 0.0034502504859119654, + "learning_rate": 5.640362225097025e-05, + "loss": 0.001, + "step": 436 + }, + { + "epoch": 1.4125252525252525, + "grad_norm": 0.004155455157160759, + "learning_rate": 5.653298835705045e-05, + "loss": 0.0008, + "step": 437 + }, + { + "epoch": 1.4157575757575758, + "grad_norm": 0.005455946549773216, + "learning_rate": 5.666235446313066e-05, + "loss": 0.0012, + "step": 438 + }, + { + "epoch": 1.418989898989899, + "grad_norm": 0.005093955434858799, + "learning_rate": 5.679172056921087e-05, + "loss": 0.0015, + "step": 439 + }, + { + "epoch": 1.4222222222222223, + "grad_norm": 0.004076706245541573, + "learning_rate": 5.692108667529108e-05, + "loss": 0.0012, + "step": 440 + }, + { + "epoch": 1.4254545454545455, + "grad_norm": 0.004360303282737732, + "learning_rate": 5.7050452781371286e-05, + "loss": 0.001, + "step": 441 + }, + { + "epoch": 1.4286868686868686, + "grad_norm": 0.005434586200863123, + "learning_rate": 5.717981888745149e-05, + "loss": 0.002, + "step": 442 + }, + { + "epoch": 1.4319191919191918, + "grad_norm": 0.006644332781434059, + "learning_rate": 5.730918499353169e-05, + "loss": 0.0033, + "step": 443 + }, + { + "epoch": 1.435151515151515, + "grad_norm": 0.006485591642558575, + "learning_rate": 5.74385510996119e-05, + "loss": 0.002, + "step": 444 + }, + { + "epoch": 1.4383838383838383, + "grad_norm": 0.006898663938045502, + "learning_rate": 5.7567917205692113e-05, + "loss": 0.001, + "step": 445 + }, + { + "epoch": 1.4416161616161616, + "grad_norm": 0.006443925201892853, + "learning_rate": 5.7697283311772324e-05, + "loss": 0.0015, + "step": 446 + }, + { + "epoch": 1.4448484848484848, + "grad_norm": 0.003106453223153949, + "learning_rate": 5.782664941785253e-05, + "loss": 0.0009, + "step": 447 + }, + { + "epoch": 1.448080808080808, + "grad_norm": 0.0036977077834308147, + "learning_rate": 5.795601552393273e-05, + "loss": 0.0009, + "step": 448 + }, + { + "epoch": 1.4513131313131313, + "grad_norm": 0.0050973836332559586, + "learning_rate": 5.8085381630012934e-05, + "loss": 0.0014, + "step": 449 + }, + { + "epoch": 1.4545454545454546, + "grad_norm": 0.0057764300145208836, + "learning_rate": 5.8214747736093145e-05, + "loss": 0.0016, + "step": 450 + }, + { + "epoch": 1.4545454545454546, + "eval_loss": 0.0010863245697692037, + "eval_runtime": 18.7321, + "eval_samples_per_second": 5.338, + "eval_steps_per_second": 1.335, + "step": 450 + }, + { + "epoch": 1.4577777777777778, + "grad_norm": 0.004166171886026859, + "learning_rate": 5.8344113842173355e-05, + "loss": 0.0009, + "step": 451 + }, + { + "epoch": 1.461010101010101, + "grad_norm": 0.004900315310806036, + "learning_rate": 5.8473479948253565e-05, + "loss": 0.0013, + "step": 452 + }, + { + "epoch": 1.4642424242424243, + "grad_norm": 0.003387290518730879, + "learning_rate": 5.860284605433377e-05, + "loss": 0.0007, + "step": 453 + }, + { + "epoch": 1.4674747474747476, + "grad_norm": 0.004271840676665306, + "learning_rate": 5.873221216041398e-05, + "loss": 0.0011, + "step": 454 + }, + { + "epoch": 1.4707070707070706, + "grad_norm": 0.0032306541688740253, + "learning_rate": 5.8861578266494176e-05, + "loss": 0.0007, + "step": 455 + }, + { + "epoch": 1.4739393939393939, + "grad_norm": 0.0022131800651550293, + "learning_rate": 5.8990944372574386e-05, + "loss": 0.0006, + "step": 456 + }, + { + "epoch": 1.4771717171717171, + "grad_norm": 0.005533823277801275, + "learning_rate": 5.9120310478654596e-05, + "loss": 0.0017, + "step": 457 + }, + { + "epoch": 1.4804040404040404, + "grad_norm": 0.004672409035265446, + "learning_rate": 5.9249676584734806e-05, + "loss": 0.0014, + "step": 458 + }, + { + "epoch": 1.4836363636363636, + "grad_norm": 0.006192138884216547, + "learning_rate": 5.937904269081501e-05, + "loss": 0.0015, + "step": 459 + }, + { + "epoch": 1.486868686868687, + "grad_norm": 0.011036567389965057, + "learning_rate": 5.950840879689522e-05, + "loss": 0.0025, + "step": 460 + }, + { + "epoch": 1.4901010101010101, + "grad_norm": 0.0055129267275333405, + "learning_rate": 5.963777490297542e-05, + "loss": 0.0015, + "step": 461 + }, + { + "epoch": 1.4933333333333334, + "grad_norm": 0.005369866266846657, + "learning_rate": 5.976714100905563e-05, + "loss": 0.0027, + "step": 462 + }, + { + "epoch": 1.4965656565656564, + "grad_norm": 0.005017601884901524, + "learning_rate": 5.989650711513584e-05, + "loss": 0.0011, + "step": 463 + }, + { + "epoch": 1.4997979797979797, + "grad_norm": 0.00623415969312191, + "learning_rate": 6.002587322121605e-05, + "loss": 0.0028, + "step": 464 + }, + { + "epoch": 1.503030303030303, + "grad_norm": 0.005604151636362076, + "learning_rate": 6.015523932729625e-05, + "loss": 0.0023, + "step": 465 + }, + { + "epoch": 1.5062626262626262, + "grad_norm": 0.004366429056972265, + "learning_rate": 6.028460543337646e-05, + "loss": 0.001, + "step": 466 + }, + { + "epoch": 1.5094949494949494, + "grad_norm": 0.004982110112905502, + "learning_rate": 6.041397153945667e-05, + "loss": 0.0014, + "step": 467 + }, + { + "epoch": 1.5127272727272727, + "grad_norm": 0.005147982854396105, + "learning_rate": 6.054333764553687e-05, + "loss": 0.0013, + "step": 468 + }, + { + "epoch": 1.515959595959596, + "grad_norm": 0.004790551029145718, + "learning_rate": 6.067270375161708e-05, + "loss": 0.0011, + "step": 469 + }, + { + "epoch": 1.5191919191919192, + "grad_norm": 0.0038922505918890238, + "learning_rate": 6.080206985769729e-05, + "loss": 0.0011, + "step": 470 + }, + { + "epoch": 1.5224242424242425, + "grad_norm": 0.006303661502897739, + "learning_rate": 6.093143596377749e-05, + "loss": 0.0023, + "step": 471 + }, + { + "epoch": 1.5256565656565657, + "grad_norm": 0.003573813708499074, + "learning_rate": 6.10608020698577e-05, + "loss": 0.0011, + "step": 472 + }, + { + "epoch": 1.528888888888889, + "grad_norm": 0.005556274671107531, + "learning_rate": 6.119016817593791e-05, + "loss": 0.0033, + "step": 473 + }, + { + "epoch": 1.5321212121212122, + "grad_norm": 0.004455295857042074, + "learning_rate": 6.131953428201811e-05, + "loss": 0.0014, + "step": 474 + }, + { + "epoch": 1.5353535353535355, + "grad_norm": 0.003466435242444277, + "learning_rate": 6.144890038809832e-05, + "loss": 0.001, + "step": 475 + }, + { + "epoch": 1.5353535353535355, + "eval_loss": 0.0010681893909350038, + "eval_runtime": 18.6082, + "eval_samples_per_second": 5.374, + "eval_steps_per_second": 1.343, + "step": 475 + }, + { + "epoch": 1.5385858585858587, + "grad_norm": 0.004641688894480467, + "learning_rate": 6.157826649417853e-05, + "loss": 0.0029, + "step": 476 + }, + { + "epoch": 1.541818181818182, + "grad_norm": 0.004081512801349163, + "learning_rate": 6.170763260025874e-05, + "loss": 0.0011, + "step": 477 + }, + { + "epoch": 1.545050505050505, + "grad_norm": 0.005640064366161823, + "learning_rate": 6.183699870633895e-05, + "loss": 0.0024, + "step": 478 + }, + { + "epoch": 1.5482828282828283, + "grad_norm": 0.0031765501480549574, + "learning_rate": 6.196636481241915e-05, + "loss": 0.0009, + "step": 479 + }, + { + "epoch": 1.5515151515151515, + "grad_norm": 0.0049113016575574875, + "learning_rate": 6.209573091849934e-05, + "loss": 0.0019, + "step": 480 + }, + { + "epoch": 1.5547474747474748, + "grad_norm": 0.003700861008837819, + "learning_rate": 6.222509702457955e-05, + "loss": 0.0012, + "step": 481 + }, + { + "epoch": 1.557979797979798, + "grad_norm": 0.004294991493225098, + "learning_rate": 6.235446313065976e-05, + "loss": 0.0014, + "step": 482 + }, + { + "epoch": 1.561212121212121, + "grad_norm": 0.003475453006103635, + "learning_rate": 6.248382923673998e-05, + "loss": 0.0007, + "step": 483 + }, + { + "epoch": 1.5644444444444443, + "grad_norm": 0.007765649352222681, + "learning_rate": 6.261319534282019e-05, + "loss": 0.001, + "step": 484 + }, + { + "epoch": 1.5676767676767676, + "grad_norm": 0.0032908658031374216, + "learning_rate": 6.27425614489004e-05, + "loss": 0.0008, + "step": 485 + }, + { + "epoch": 1.5709090909090908, + "grad_norm": 0.00453177560120821, + "learning_rate": 6.28719275549806e-05, + "loss": 0.0013, + "step": 486 + }, + { + "epoch": 1.574141414141414, + "grad_norm": 0.0038091707974672318, + "learning_rate": 6.30012936610608e-05, + "loss": 0.0008, + "step": 487 + }, + { + "epoch": 1.5773737373737373, + "grad_norm": 0.004123839549720287, + "learning_rate": 6.313065976714101e-05, + "loss": 0.0011, + "step": 488 + }, + { + "epoch": 1.5806060606060606, + "grad_norm": 0.003308449639007449, + "learning_rate": 6.326002587322122e-05, + "loss": 0.0012, + "step": 489 + }, + { + "epoch": 1.5838383838383838, + "grad_norm": 0.005438206251710653, + "learning_rate": 6.338939197930143e-05, + "loss": 0.0023, + "step": 490 + }, + { + "epoch": 1.587070707070707, + "grad_norm": 0.005823109764605761, + "learning_rate": 6.351875808538163e-05, + "loss": 0.0014, + "step": 491 + }, + { + "epoch": 1.5903030303030303, + "grad_norm": 0.0052726129069924355, + "learning_rate": 6.364812419146184e-05, + "loss": 0.0012, + "step": 492 + }, + { + "epoch": 1.5935353535353536, + "grad_norm": 0.003648497397080064, + "learning_rate": 6.377749029754204e-05, + "loss": 0.0013, + "step": 493 + }, + { + "epoch": 1.5967676767676768, + "grad_norm": 0.003007176099345088, + "learning_rate": 6.390685640362225e-05, + "loss": 0.0007, + "step": 494 + }, + { + "epoch": 1.6, + "grad_norm": 0.004037541802972555, + "learning_rate": 6.403622250970246e-05, + "loss": 0.0011, + "step": 495 + }, + { + "epoch": 1.6032323232323233, + "grad_norm": 0.0036643114872276783, + "learning_rate": 6.416558861578267e-05, + "loss": 0.0009, + "step": 496 + }, + { + "epoch": 1.6064646464646466, + "grad_norm": 0.0038799517787992954, + "learning_rate": 6.429495472186288e-05, + "loss": 0.001, + "step": 497 + }, + { + "epoch": 1.6096969696969698, + "grad_norm": 0.002598103601485491, + "learning_rate": 6.442432082794309e-05, + "loss": 0.0005, + "step": 498 + }, + { + "epoch": 1.6129292929292929, + "grad_norm": 0.006756095215678215, + "learning_rate": 6.45536869340233e-05, + "loss": 0.0014, + "step": 499 + }, + { + "epoch": 1.6161616161616161, + "grad_norm": 0.0038995088543742895, + "learning_rate": 6.46830530401035e-05, + "loss": 0.0007, + "step": 500 + }, + { + "epoch": 1.6161616161616161, + "eval_loss": 0.0010785168269649148, + "eval_runtime": 18.596, + "eval_samples_per_second": 5.378, + "eval_steps_per_second": 1.344, + "step": 500 + }, + { + "epoch": 1.6193939393939394, + "grad_norm": 0.004936838988214731, + "learning_rate": 6.48124191461837e-05, + "loss": 0.0012, + "step": 501 + }, + { + "epoch": 1.6226262626262626, + "grad_norm": 0.006074307020753622, + "learning_rate": 6.494178525226392e-05, + "loss": 0.0022, + "step": 502 + }, + { + "epoch": 1.625858585858586, + "grad_norm": 0.0044588991440832615, + "learning_rate": 6.507115135834411e-05, + "loss": 0.0011, + "step": 503 + }, + { + "epoch": 1.6290909090909091, + "grad_norm": 0.004507533740252256, + "learning_rate": 6.520051746442432e-05, + "loss": 0.0011, + "step": 504 + }, + { + "epoch": 1.6323232323232322, + "grad_norm": 0.0048494781367480755, + "learning_rate": 6.532988357050453e-05, + "loss": 0.0008, + "step": 505 + }, + { + "epoch": 1.6355555555555554, + "grad_norm": 0.0033034745138138533, + "learning_rate": 6.545924967658473e-05, + "loss": 0.0007, + "step": 506 + }, + { + "epoch": 1.6387878787878787, + "grad_norm": 0.0053002117201685905, + "learning_rate": 6.558861578266494e-05, + "loss": 0.0009, + "step": 507 + }, + { + "epoch": 1.642020202020202, + "grad_norm": 0.00404641218483448, + "learning_rate": 6.571798188874515e-05, + "loss": 0.0008, + "step": 508 + }, + { + "epoch": 1.6452525252525252, + "grad_norm": 0.0036740771029144526, + "learning_rate": 6.584734799482536e-05, + "loss": 0.0009, + "step": 509 + }, + { + "epoch": 1.6484848484848484, + "grad_norm": 0.005331697873771191, + "learning_rate": 6.597671410090557e-05, + "loss": 0.0014, + "step": 510 + }, + { + "epoch": 1.6517171717171717, + "grad_norm": 0.004965492524206638, + "learning_rate": 6.610608020698578e-05, + "loss": 0.0019, + "step": 511 + }, + { + "epoch": 1.654949494949495, + "grad_norm": 0.0037727411836385727, + "learning_rate": 6.623544631306599e-05, + "loss": 0.001, + "step": 512 + }, + { + "epoch": 1.6581818181818182, + "grad_norm": 0.0048839072696864605, + "learning_rate": 6.636481241914619e-05, + "loss": 0.0013, + "step": 513 + }, + { + "epoch": 1.6614141414141415, + "grad_norm": 0.004607974551618099, + "learning_rate": 6.649417852522638e-05, + "loss": 0.0017, + "step": 514 + }, + { + "epoch": 1.6646464646464647, + "grad_norm": 0.004647100809961557, + "learning_rate": 6.66235446313066e-05, + "loss": 0.0024, + "step": 515 + }, + { + "epoch": 1.667878787878788, + "grad_norm": 0.004482895601540804, + "learning_rate": 6.67529107373868e-05, + "loss": 0.0023, + "step": 516 + }, + { + "epoch": 1.6711111111111112, + "grad_norm": 0.004681951366364956, + "learning_rate": 6.688227684346702e-05, + "loss": 0.0031, + "step": 517 + }, + { + "epoch": 1.6743434343434345, + "grad_norm": 0.004061279818415642, + "learning_rate": 6.701164294954723e-05, + "loss": 0.0012, + "step": 518 + }, + { + "epoch": 1.6775757575757577, + "grad_norm": 0.005936305969953537, + "learning_rate": 6.714100905562742e-05, + "loss": 0.0018, + "step": 519 + }, + { + "epoch": 1.680808080808081, + "grad_norm": 0.003031841479241848, + "learning_rate": 6.727037516170763e-05, + "loss": 0.0009, + "step": 520 + }, + { + "epoch": 1.684040404040404, + "grad_norm": 0.0044912113808095455, + "learning_rate": 6.739974126778784e-05, + "loss": 0.0013, + "step": 521 + }, + { + "epoch": 1.6872727272727273, + "grad_norm": 0.003882101271301508, + "learning_rate": 6.752910737386805e-05, + "loss": 0.0014, + "step": 522 + }, + { + "epoch": 1.6905050505050505, + "grad_norm": 0.0031570433638989925, + "learning_rate": 6.765847347994826e-05, + "loss": 0.0011, + "step": 523 + }, + { + "epoch": 1.6937373737373738, + "grad_norm": 0.004184515681117773, + "learning_rate": 6.778783958602847e-05, + "loss": 0.0015, + "step": 524 + }, + { + "epoch": 1.696969696969697, + "grad_norm": 0.002800683258101344, + "learning_rate": 6.791720569210867e-05, + "loss": 0.0008, + "step": 525 + }, + { + "epoch": 1.696969696969697, + "eval_loss": 0.0010542384115979075, + "eval_runtime": 18.5989, + "eval_samples_per_second": 5.377, + "eval_steps_per_second": 1.344, + "step": 525 + }, + { + "epoch": 1.70020202020202, + "grad_norm": 0.003550174878910184, + "learning_rate": 6.804657179818887e-05, + "loss": 0.0014, + "step": 526 + }, + { + "epoch": 1.7034343434343433, + "grad_norm": 0.0030985972844064236, + "learning_rate": 6.817593790426908e-05, + "loss": 0.0008, + "step": 527 + }, + { + "epoch": 1.7066666666666666, + "grad_norm": 0.0048317620530724525, + "learning_rate": 6.830530401034929e-05, + "loss": 0.0018, + "step": 528 + }, + { + "epoch": 1.7098989898989898, + "grad_norm": 0.005726094823330641, + "learning_rate": 6.84346701164295e-05, + "loss": 0.0025, + "step": 529 + }, + { + "epoch": 1.713131313131313, + "grad_norm": 0.0024808107409626245, + "learning_rate": 6.856403622250971e-05, + "loss": 0.0007, + "step": 530 + }, + { + "epoch": 1.7163636363636363, + "grad_norm": 0.003422652604058385, + "learning_rate": 6.869340232858992e-05, + "loss": 0.001, + "step": 531 + }, + { + "epoch": 1.7195959595959596, + "grad_norm": 0.0037957008462399244, + "learning_rate": 6.882276843467012e-05, + "loss": 0.0009, + "step": 532 + }, + { + "epoch": 1.7228282828282828, + "grad_norm": 0.0028711955528706312, + "learning_rate": 6.895213454075033e-05, + "loss": 0.0008, + "step": 533 + }, + { + "epoch": 1.726060606060606, + "grad_norm": 0.002812835620716214, + "learning_rate": 6.908150064683054e-05, + "loss": 0.0009, + "step": 534 + }, + { + "epoch": 1.7292929292929293, + "grad_norm": 0.003739473642781377, + "learning_rate": 6.921086675291075e-05, + "loss": 0.0016, + "step": 535 + }, + { + "epoch": 1.7325252525252526, + "grad_norm": 0.004807054530829191, + "learning_rate": 6.934023285899096e-05, + "loss": 0.0023, + "step": 536 + }, + { + "epoch": 1.7357575757575758, + "grad_norm": 0.004576352424919605, + "learning_rate": 6.946959896507115e-05, + "loss": 0.0024, + "step": 537 + }, + { + "epoch": 1.738989898989899, + "grad_norm": 0.0030553669203072786, + "learning_rate": 6.959896507115135e-05, + "loss": 0.0009, + "step": 538 + }, + { + "epoch": 1.7422222222222223, + "grad_norm": 0.003879109164699912, + "learning_rate": 6.972833117723156e-05, + "loss": 0.0012, + "step": 539 + }, + { + "epoch": 1.7454545454545456, + "grad_norm": 0.0031888741068542004, + "learning_rate": 6.985769728331177e-05, + "loss": 0.0013, + "step": 540 + }, + { + "epoch": 1.7486868686868688, + "grad_norm": 0.0037957336753606796, + "learning_rate": 6.998706338939198e-05, + "loss": 0.0008, + "step": 541 + }, + { + "epoch": 1.7519191919191919, + "grad_norm": 0.002901956904679537, + "learning_rate": 7.011642949547219e-05, + "loss": 0.0007, + "step": 542 + }, + { + "epoch": 1.7551515151515151, + "grad_norm": 0.0029531833715736866, + "learning_rate": 7.02457956015524e-05, + "loss": 0.0008, + "step": 543 + }, + { + "epoch": 1.7583838383838384, + "grad_norm": 0.004063300788402557, + "learning_rate": 7.037516170763261e-05, + "loss": 0.0012, + "step": 544 + }, + { + "epoch": 1.7616161616161616, + "grad_norm": 0.0041914028115570545, + "learning_rate": 7.050452781371281e-05, + "loss": 0.001, + "step": 545 + }, + { + "epoch": 1.7648484848484849, + "grad_norm": 0.004884886089712381, + "learning_rate": 7.063389391979302e-05, + "loss": 0.0021, + "step": 546 + }, + { + "epoch": 1.768080808080808, + "grad_norm": 0.005683641415089369, + "learning_rate": 7.076326002587323e-05, + "loss": 0.0024, + "step": 547 + }, + { + "epoch": 1.7713131313131312, + "grad_norm": 0.0042356885969638824, + "learning_rate": 7.089262613195343e-05, + "loss": 0.0015, + "step": 548 + }, + { + "epoch": 1.7745454545454544, + "grad_norm": 0.005650636274367571, + "learning_rate": 7.102199223803364e-05, + "loss": 0.0013, + "step": 549 + }, + { + "epoch": 1.7777777777777777, + "grad_norm": 0.00460450816899538, + "learning_rate": 7.115135834411385e-05, + "loss": 0.001, + "step": 550 + }, + { + "epoch": 1.7777777777777777, + "eval_loss": 0.0010012522106990218, + "eval_runtime": 18.6807, + "eval_samples_per_second": 5.353, + "eval_steps_per_second": 1.338, + "step": 550 + }, + { + "epoch": 1.781010101010101, + "grad_norm": 0.004403329454362392, + "learning_rate": 7.128072445019404e-05, + "loss": 0.0012, + "step": 551 + }, + { + "epoch": 1.7842424242424242, + "grad_norm": 0.0032920974772423506, + "learning_rate": 7.141009055627425e-05, + "loss": 0.0008, + "step": 552 + }, + { + "epoch": 1.7874747474747474, + "grad_norm": 0.0038138548843562603, + "learning_rate": 7.153945666235446e-05, + "loss": 0.0011, + "step": 553 + }, + { + "epoch": 1.7907070707070707, + "grad_norm": 0.00296188285574317, + "learning_rate": 7.166882276843467e-05, + "loss": 0.0008, + "step": 554 + }, + { + "epoch": 1.793939393939394, + "grad_norm": 0.0031317227985709906, + "learning_rate": 7.179818887451488e-05, + "loss": 0.001, + "step": 555 + }, + { + "epoch": 1.7971717171717172, + "grad_norm": 0.003491588868200779, + "learning_rate": 7.19275549805951e-05, + "loss": 0.0007, + "step": 556 + }, + { + "epoch": 1.8004040404040405, + "grad_norm": 0.0031343346927314997, + "learning_rate": 7.20569210866753e-05, + "loss": 0.0005, + "step": 557 + }, + { + "epoch": 1.8036363636363637, + "grad_norm": 0.004438635893166065, + "learning_rate": 7.21862871927555e-05, + "loss": 0.0011, + "step": 558 + }, + { + "epoch": 1.806868686868687, + "grad_norm": 0.00411741528660059, + "learning_rate": 7.231565329883571e-05, + "loss": 0.0008, + "step": 559 + }, + { + "epoch": 1.8101010101010102, + "grad_norm": 0.004590868018567562, + "learning_rate": 7.244501940491591e-05, + "loss": 0.0009, + "step": 560 + }, + { + "epoch": 1.8133333333333335, + "grad_norm": 0.00454489141702652, + "learning_rate": 7.257438551099612e-05, + "loss": 0.001, + "step": 561 + }, + { + "epoch": 1.8165656565656567, + "grad_norm": 0.004015767015516758, + "learning_rate": 7.270375161707633e-05, + "loss": 0.0007, + "step": 562 + }, + { + "epoch": 1.8197979797979797, + "grad_norm": 0.003563391976058483, + "learning_rate": 7.283311772315654e-05, + "loss": 0.0008, + "step": 563 + }, + { + "epoch": 1.823030303030303, + "grad_norm": 0.0037166401743888855, + "learning_rate": 7.296248382923674e-05, + "loss": 0.0009, + "step": 564 + }, + { + "epoch": 1.8262626262626263, + "grad_norm": 0.004116456024348736, + "learning_rate": 7.309184993531695e-05, + "loss": 0.0009, + "step": 565 + }, + { + "epoch": 1.8294949494949495, + "grad_norm": 0.004104538355022669, + "learning_rate": 7.322121604139716e-05, + "loss": 0.0013, + "step": 566 + }, + { + "epoch": 1.8327272727272728, + "grad_norm": 0.005509045906364918, + "learning_rate": 7.335058214747737e-05, + "loss": 0.0023, + "step": 567 + }, + { + "epoch": 1.835959595959596, + "grad_norm": 0.0035526566207408905, + "learning_rate": 7.347994825355758e-05, + "loss": 0.0008, + "step": 568 + }, + { + "epoch": 1.839191919191919, + "grad_norm": 0.003440872300416231, + "learning_rate": 7.360931435963779e-05, + "loss": 0.0007, + "step": 569 + }, + { + "epoch": 1.8424242424242423, + "grad_norm": 0.0029369727708399296, + "learning_rate": 7.3738680465718e-05, + "loss": 0.0008, + "step": 570 + }, + { + "epoch": 1.8456565656565656, + "grad_norm": 0.004200720228254795, + "learning_rate": 7.38680465717982e-05, + "loss": 0.0011, + "step": 571 + }, + { + "epoch": 1.8488888888888888, + "grad_norm": 0.005561929661780596, + "learning_rate": 7.399741267787839e-05, + "loss": 0.0018, + "step": 572 + }, + { + "epoch": 1.852121212121212, + "grad_norm": 0.0033285904210060835, + "learning_rate": 7.41267787839586e-05, + "loss": 0.0009, + "step": 573 + }, + { + "epoch": 1.8553535353535353, + "grad_norm": 0.004428067244589329, + "learning_rate": 7.425614489003881e-05, + "loss": 0.0015, + "step": 574 + }, + { + "epoch": 1.8585858585858586, + "grad_norm": 0.003229588968679309, + "learning_rate": 7.438551099611902e-05, + "loss": 0.0007, + "step": 575 + }, + { + "epoch": 1.8585858585858586, + "eval_loss": 0.0009781663538888097, + "eval_runtime": 18.7511, + "eval_samples_per_second": 5.333, + "eval_steps_per_second": 1.333, + "step": 575 + }, + { + "epoch": 1.8618181818181818, + "grad_norm": 0.003516717813909054, + "learning_rate": 7.451487710219923e-05, + "loss": 0.0011, + "step": 576 + }, + { + "epoch": 1.865050505050505, + "grad_norm": 0.002692221663892269, + "learning_rate": 7.464424320827943e-05, + "loss": 0.0008, + "step": 577 + }, + { + "epoch": 1.8682828282828283, + "grad_norm": 0.003864881582558155, + "learning_rate": 7.477360931435964e-05, + "loss": 0.0013, + "step": 578 + }, + { + "epoch": 1.8715151515151516, + "grad_norm": 0.003865283913910389, + "learning_rate": 7.490297542043985e-05, + "loss": 0.0015, + "step": 579 + }, + { + "epoch": 1.8747474747474748, + "grad_norm": 0.0035206254106014967, + "learning_rate": 7.503234152652006e-05, + "loss": 0.0012, + "step": 580 + }, + { + "epoch": 1.877979797979798, + "grad_norm": 0.004602121654897928, + "learning_rate": 7.516170763260027e-05, + "loss": 0.0018, + "step": 581 + }, + { + "epoch": 1.8812121212121213, + "grad_norm": 0.0029543524142354727, + "learning_rate": 7.529107373868047e-05, + "loss": 0.0009, + "step": 582 + }, + { + "epoch": 1.8844444444444446, + "grad_norm": 0.0023640701547265053, + "learning_rate": 7.542043984476068e-05, + "loss": 0.0007, + "step": 583 + }, + { + "epoch": 1.8876767676767678, + "grad_norm": 0.0040098559111356735, + "learning_rate": 7.554980595084087e-05, + "loss": 0.0012, + "step": 584 + }, + { + "epoch": 1.8909090909090909, + "grad_norm": 0.004274715203791857, + "learning_rate": 7.567917205692108e-05, + "loss": 0.0015, + "step": 585 + }, + { + "epoch": 1.8941414141414141, + "grad_norm": 0.0031054418068379164, + "learning_rate": 7.58085381630013e-05, + "loss": 0.0008, + "step": 586 + }, + { + "epoch": 1.8973737373737374, + "grad_norm": 0.003954428713768721, + "learning_rate": 7.59379042690815e-05, + "loss": 0.0009, + "step": 587 + }, + { + "epoch": 1.9006060606060606, + "grad_norm": 0.003786911489441991, + "learning_rate": 7.606727037516171e-05, + "loss": 0.0006, + "step": 588 + }, + { + "epoch": 1.9038383838383839, + "grad_norm": 0.0041464087553322315, + "learning_rate": 7.619663648124192e-05, + "loss": 0.0011, + "step": 589 + }, + { + "epoch": 1.907070707070707, + "grad_norm": 0.004568018950521946, + "learning_rate": 7.632600258732212e-05, + "loss": 0.0019, + "step": 590 + }, + { + "epoch": 1.9103030303030302, + "grad_norm": 0.003288006642833352, + "learning_rate": 7.645536869340233e-05, + "loss": 0.0009, + "step": 591 + }, + { + "epoch": 1.9135353535353534, + "grad_norm": 0.004277890548110008, + "learning_rate": 7.658473479948254e-05, + "loss": 0.0008, + "step": 592 + }, + { + "epoch": 1.9167676767676767, + "grad_norm": 0.004379922058433294, + "learning_rate": 7.671410090556275e-05, + "loss": 0.001, + "step": 593 + }, + { + "epoch": 1.92, + "grad_norm": 0.003021504729986191, + "learning_rate": 7.684346701164295e-05, + "loss": 0.0006, + "step": 594 + }, + { + "epoch": 1.9232323232323232, + "grad_norm": 0.0032772128470242023, + "learning_rate": 7.697283311772316e-05, + "loss": 0.0008, + "step": 595 + }, + { + "epoch": 1.9264646464646464, + "grad_norm": 0.004533900413662195, + "learning_rate": 7.710219922380336e-05, + "loss": 0.0015, + "step": 596 + }, + { + "epoch": 1.9296969696969697, + "grad_norm": 0.005284740123897791, + "learning_rate": 7.723156532988357e-05, + "loss": 0.0011, + "step": 597 + }, + { + "epoch": 1.932929292929293, + "grad_norm": 0.005875818431377411, + "learning_rate": 7.736093143596378e-05, + "loss": 0.0013, + "step": 598 + }, + { + "epoch": 1.9361616161616162, + "grad_norm": 0.004483737051486969, + "learning_rate": 7.749029754204399e-05, + "loss": 0.0009, + "step": 599 + }, + { + "epoch": 1.9393939393939394, + "grad_norm": 0.004435641225427389, + "learning_rate": 7.76196636481242e-05, + "loss": 0.0013, + "step": 600 + }, + { + "epoch": 1.9393939393939394, + "eval_loss": 0.0009266917477361858, + "eval_runtime": 18.7646, + "eval_samples_per_second": 5.329, + "eval_steps_per_second": 1.332, + "step": 600 + }, + { + "epoch": 1.9426262626262627, + "grad_norm": 0.003876154311001301, + "learning_rate": 7.774902975420441e-05, + "loss": 0.001, + "step": 601 + }, + { + "epoch": 1.945858585858586, + "grad_norm": 0.004851337987929583, + "learning_rate": 7.787839586028462e-05, + "loss": 0.0014, + "step": 602 + }, + { + "epoch": 1.9490909090909092, + "grad_norm": 0.003333737375214696, + "learning_rate": 7.800776196636481e-05, + "loss": 0.001, + "step": 603 + }, + { + "epoch": 1.9523232323232325, + "grad_norm": 0.0032944250851869583, + "learning_rate": 7.813712807244502e-05, + "loss": 0.0007, + "step": 604 + }, + { + "epoch": 1.9555555555555557, + "grad_norm": 0.004762569442391396, + "learning_rate": 7.826649417852523e-05, + "loss": 0.0027, + "step": 605 + }, + { + "epoch": 1.9587878787878787, + "grad_norm": 0.0037507396191358566, + "learning_rate": 7.839586028460543e-05, + "loss": 0.0014, + "step": 606 + }, + { + "epoch": 1.962020202020202, + "grad_norm": 0.005207899492233992, + "learning_rate": 7.852522639068564e-05, + "loss": 0.0031, + "step": 607 + }, + { + "epoch": 1.9652525252525253, + "grad_norm": 0.0035236128605902195, + "learning_rate": 7.865459249676585e-05, + "loss": 0.001, + "step": 608 + }, + { + "epoch": 1.9684848484848485, + "grad_norm": 0.004186858423054218, + "learning_rate": 7.878395860284605e-05, + "loss": 0.0012, + "step": 609 + }, + { + "epoch": 1.9717171717171718, + "grad_norm": 0.0038515774067491293, + "learning_rate": 7.891332470892626e-05, + "loss": 0.0022, + "step": 610 + }, + { + "epoch": 1.9749494949494948, + "grad_norm": 0.0030782243702560663, + "learning_rate": 7.904269081500647e-05, + "loss": 0.0006, + "step": 611 + }, + { + "epoch": 1.978181818181818, + "grad_norm": 0.005162171553820372, + "learning_rate": 7.917205692108668e-05, + "loss": 0.0013, + "step": 612 + }, + { + "epoch": 1.9814141414141413, + "grad_norm": 0.004226456396281719, + "learning_rate": 7.930142302716689e-05, + "loss": 0.0009, + "step": 613 + }, + { + "epoch": 1.9846464646464645, + "grad_norm": 0.003615841967985034, + "learning_rate": 7.94307891332471e-05, + "loss": 0.0022, + "step": 614 + }, + { + "epoch": 1.9878787878787878, + "grad_norm": 0.00408519571647048, + "learning_rate": 7.956015523932731e-05, + "loss": 0.0015, + "step": 615 + }, + { + "epoch": 1.991111111111111, + "grad_norm": 0.0042512728832662106, + "learning_rate": 7.968952134540751e-05, + "loss": 0.0014, + "step": 616 + }, + { + "epoch": 1.9943434343434343, + "grad_norm": 0.003178700339049101, + "learning_rate": 7.981888745148772e-05, + "loss": 0.0008, + "step": 617 + }, + { + "epoch": 1.9975757575757576, + "grad_norm": 0.003822662867605686, + "learning_rate": 7.994825355756791e-05, + "loss": 0.001, + "step": 618 + }, + { + "epoch": 2.000808080808081, + "grad_norm": 0.0046024019829928875, + "learning_rate": 8.007761966364812e-05, + "loss": 0.0011, + "step": 619 + }, + { + "epoch": 2.004040404040404, + "grad_norm": 0.0038011916913092136, + "learning_rate": 8.020698576972833e-05, + "loss": 0.0008, + "step": 620 + }, + { + "epoch": 2.0072727272727273, + "grad_norm": 0.006909184157848358, + "learning_rate": 8.033635187580855e-05, + "loss": 0.0026, + "step": 621 + }, + { + "epoch": 2.0105050505050506, + "grad_norm": 0.005042599979788065, + "learning_rate": 8.046571798188874e-05, + "loss": 0.0011, + "step": 622 + }, + { + "epoch": 2.013737373737374, + "grad_norm": 0.004002322442829609, + "learning_rate": 8.059508408796895e-05, + "loss": 0.0011, + "step": 623 + }, + { + "epoch": 2.016969696969697, + "grad_norm": 0.0031209783628582954, + "learning_rate": 8.072445019404916e-05, + "loss": 0.0009, + "step": 624 + }, + { + "epoch": 2.0202020202020203, + "grad_norm": 0.0026158462278544903, + "learning_rate": 8.085381630012937e-05, + "loss": 0.0007, + "step": 625 + }, + { + "epoch": 2.0202020202020203, + "eval_loss": 0.0009721739334054291, + "eval_runtime": 18.7204, + "eval_samples_per_second": 5.342, + "eval_steps_per_second": 1.335, + "step": 625 + }, + { + "epoch": 2.0234343434343436, + "grad_norm": 0.003943873103708029, + "learning_rate": 8.098318240620958e-05, + "loss": 0.001, + "step": 626 + }, + { + "epoch": 2.026666666666667, + "grad_norm": 0.0038578941021114588, + "learning_rate": 8.111254851228979e-05, + "loss": 0.0017, + "step": 627 + }, + { + "epoch": 2.02989898989899, + "grad_norm": 0.002991423010826111, + "learning_rate": 8.124191461836999e-05, + "loss": 0.0007, + "step": 628 + }, + { + "epoch": 2.0331313131313133, + "grad_norm": 0.002815463813021779, + "learning_rate": 8.13712807244502e-05, + "loss": 0.0009, + "step": 629 + }, + { + "epoch": 2.036363636363636, + "grad_norm": 0.0025943731889128685, + "learning_rate": 8.15006468305304e-05, + "loss": 0.0006, + "step": 630 + }, + { + "epoch": 2.0395959595959594, + "grad_norm": 0.0029918155632913113, + "learning_rate": 8.163001293661061e-05, + "loss": 0.0008, + "step": 631 + }, + { + "epoch": 2.0428282828282827, + "grad_norm": 0.00398919777944684, + "learning_rate": 8.175937904269082e-05, + "loss": 0.0011, + "step": 632 + }, + { + "epoch": 2.046060606060606, + "grad_norm": 0.003045483957976103, + "learning_rate": 8.188874514877103e-05, + "loss": 0.0007, + "step": 633 + }, + { + "epoch": 2.049292929292929, + "grad_norm": 0.005803203675895929, + "learning_rate": 8.201811125485124e-05, + "loss": 0.0008, + "step": 634 + }, + { + "epoch": 2.0525252525252524, + "grad_norm": 0.003953828942030668, + "learning_rate": 8.214747736093143e-05, + "loss": 0.001, + "step": 635 + }, + { + "epoch": 2.0557575757575757, + "grad_norm": 0.003973971586674452, + "learning_rate": 8.227684346701164e-05, + "loss": 0.0006, + "step": 636 + }, + { + "epoch": 2.058989898989899, + "grad_norm": 0.0025888276286423206, + "learning_rate": 8.240620957309186e-05, + "loss": 0.0005, + "step": 637 + }, + { + "epoch": 2.062222222222222, + "grad_norm": 0.003989651799201965, + "learning_rate": 8.253557567917207e-05, + "loss": 0.0008, + "step": 638 + }, + { + "epoch": 2.0654545454545454, + "grad_norm": 0.0030330433510243893, + "learning_rate": 8.266494178525228e-05, + "loss": 0.0009, + "step": 639 + }, + { + "epoch": 2.0686868686868687, + "grad_norm": 0.0044325897470116615, + "learning_rate": 8.279430789133247e-05, + "loss": 0.0016, + "step": 640 + }, + { + "epoch": 2.071919191919192, + "grad_norm": 0.002264161128550768, + "learning_rate": 8.292367399741268e-05, + "loss": 0.0006, + "step": 641 + }, + { + "epoch": 2.075151515151515, + "grad_norm": 0.0033110452350229025, + "learning_rate": 8.305304010349288e-05, + "loss": 0.001, + "step": 642 + }, + { + "epoch": 2.0783838383838384, + "grad_norm": 0.0029199772980064154, + "learning_rate": 8.318240620957309e-05, + "loss": 0.0006, + "step": 643 + }, + { + "epoch": 2.0816161616161617, + "grad_norm": 0.0031028217636048794, + "learning_rate": 8.33117723156533e-05, + "loss": 0.0007, + "step": 644 + }, + { + "epoch": 2.084848484848485, + "grad_norm": 0.004011626821011305, + "learning_rate": 8.344113842173351e-05, + "loss": 0.0009, + "step": 645 + }, + { + "epoch": 2.088080808080808, + "grad_norm": 0.003449072130024433, + "learning_rate": 8.357050452781372e-05, + "loss": 0.0009, + "step": 646 + }, + { + "epoch": 2.0913131313131315, + "grad_norm": 0.003367091529071331, + "learning_rate": 8.369987063389393e-05, + "loss": 0.001, + "step": 647 + }, + { + "epoch": 2.0945454545454547, + "grad_norm": 0.003321894910186529, + "learning_rate": 8.382923673997413e-05, + "loss": 0.0016, + "step": 648 + }, + { + "epoch": 2.097777777777778, + "grad_norm": 0.003046546597033739, + "learning_rate": 8.395860284605434e-05, + "loss": 0.0008, + "step": 649 + }, + { + "epoch": 2.101010101010101, + "grad_norm": 0.0023006205447018147, + "learning_rate": 8.408796895213455e-05, + "loss": 0.0006, + "step": 650 + }, + { + "epoch": 2.101010101010101, + "eval_loss": 0.0009433354716747999, + "eval_runtime": 18.7374, + "eval_samples_per_second": 5.337, + "eval_steps_per_second": 1.334, + "step": 650 + }, + { + "epoch": 2.1042424242424245, + "grad_norm": 0.004013998433947563, + "learning_rate": 8.421733505821476e-05, + "loss": 0.001, + "step": 651 + }, + { + "epoch": 2.1074747474747473, + "grad_norm": 0.0035437876358628273, + "learning_rate": 8.434670116429496e-05, + "loss": 0.0007, + "step": 652 + }, + { + "epoch": 2.1107070707070705, + "grad_norm": 0.004152386449277401, + "learning_rate": 8.447606727037517e-05, + "loss": 0.001, + "step": 653 + }, + { + "epoch": 2.113939393939394, + "grad_norm": 0.0031009165104478598, + "learning_rate": 8.460543337645536e-05, + "loss": 0.0007, + "step": 654 + }, + { + "epoch": 2.117171717171717, + "grad_norm": 0.002030389616265893, + "learning_rate": 8.473479948253557e-05, + "loss": 0.0005, + "step": 655 + }, + { + "epoch": 2.1204040404040403, + "grad_norm": 0.003986351191997528, + "learning_rate": 8.486416558861578e-05, + "loss": 0.0011, + "step": 656 + }, + { + "epoch": 2.1236363636363635, + "grad_norm": 0.0037633986212313175, + "learning_rate": 8.499353169469599e-05, + "loss": 0.0017, + "step": 657 + }, + { + "epoch": 2.126868686868687, + "grad_norm": 0.003191509749740362, + "learning_rate": 8.51228978007762e-05, + "loss": 0.0009, + "step": 658 + }, + { + "epoch": 2.13010101010101, + "grad_norm": 0.003234416712075472, + "learning_rate": 8.525226390685641e-05, + "loss": 0.0015, + "step": 659 + }, + { + "epoch": 2.1333333333333333, + "grad_norm": 0.00314782140776515, + "learning_rate": 8.538163001293662e-05, + "loss": 0.0008, + "step": 660 + }, + { + "epoch": 2.1365656565656566, + "grad_norm": 0.004626644309610128, + "learning_rate": 8.551099611901682e-05, + "loss": 0.0017, + "step": 661 + }, + { + "epoch": 2.13979797979798, + "grad_norm": 0.0037515638396143913, + "learning_rate": 8.564036222509703e-05, + "loss": 0.001, + "step": 662 + }, + { + "epoch": 2.143030303030303, + "grad_norm": 0.0028777304105460644, + "learning_rate": 8.576972833117724e-05, + "loss": 0.0009, + "step": 663 + }, + { + "epoch": 2.1462626262626263, + "grad_norm": 0.0035881204530596733, + "learning_rate": 8.589909443725744e-05, + "loss": 0.0008, + "step": 664 + }, + { + "epoch": 2.1494949494949496, + "grad_norm": 0.002268304582685232, + "learning_rate": 8.602846054333765e-05, + "loss": 0.0005, + "step": 665 + }, + { + "epoch": 2.152727272727273, + "grad_norm": 0.003100323723629117, + "learning_rate": 8.615782664941786e-05, + "loss": 0.0008, + "step": 666 + }, + { + "epoch": 2.155959595959596, + "grad_norm": 0.003520137397572398, + "learning_rate": 8.628719275549805e-05, + "loss": 0.0009, + "step": 667 + }, + { + "epoch": 2.1591919191919193, + "grad_norm": 0.0042273253202438354, + "learning_rate": 8.641655886157827e-05, + "loss": 0.0015, + "step": 668 + }, + { + "epoch": 2.1624242424242426, + "grad_norm": 0.0030146201606839895, + "learning_rate": 8.654592496765848e-05, + "loss": 0.0009, + "step": 669 + }, + { + "epoch": 2.165656565656566, + "grad_norm": 0.0055503263138234615, + "learning_rate": 8.667529107373869e-05, + "loss": 0.0025, + "step": 670 + }, + { + "epoch": 2.168888888888889, + "grad_norm": 0.0036939766723662615, + "learning_rate": 8.68046571798189e-05, + "loss": 0.002, + "step": 671 + }, + { + "epoch": 2.172121212121212, + "grad_norm": 0.0032185425516217947, + "learning_rate": 8.69340232858991e-05, + "loss": 0.0009, + "step": 672 + }, + { + "epoch": 2.175353535353535, + "grad_norm": 0.0030356363859027624, + "learning_rate": 8.706338939197932e-05, + "loss": 0.0007, + "step": 673 + }, + { + "epoch": 2.1785858585858584, + "grad_norm": 0.006692877039313316, + "learning_rate": 8.719275549805951e-05, + "loss": 0.0008, + "step": 674 + }, + { + "epoch": 2.1818181818181817, + "grad_norm": 0.0026474855840206146, + "learning_rate": 8.732212160413972e-05, + "loss": 0.0007, + "step": 675 + }, + { + "epoch": 2.1818181818181817, + "eval_loss": 0.0009410877246409655, + "eval_runtime": 18.7748, + "eval_samples_per_second": 5.326, + "eval_steps_per_second": 1.332, + "step": 675 + }, + { + "epoch": 2.185050505050505, + "grad_norm": 0.003672214224934578, + "learning_rate": 8.745148771021992e-05, + "loss": 0.0018, + "step": 676 + }, + { + "epoch": 2.188282828282828, + "grad_norm": 0.0031880387105047703, + "learning_rate": 8.758085381630013e-05, + "loss": 0.0012, + "step": 677 + }, + { + "epoch": 2.1915151515151514, + "grad_norm": 0.0033309224527329206, + "learning_rate": 8.771021992238034e-05, + "loss": 0.001, + "step": 678 + }, + { + "epoch": 2.1947474747474747, + "grad_norm": 0.003027291502803564, + "learning_rate": 8.783958602846055e-05, + "loss": 0.0008, + "step": 679 + }, + { + "epoch": 2.197979797979798, + "grad_norm": 0.0036887172609567642, + "learning_rate": 8.796895213454075e-05, + "loss": 0.001, + "step": 680 + }, + { + "epoch": 2.201212121212121, + "grad_norm": 0.003997828811407089, + "learning_rate": 8.809831824062096e-05, + "loss": 0.0011, + "step": 681 + }, + { + "epoch": 2.2044444444444444, + "grad_norm": 0.003314135130494833, + "learning_rate": 8.822768434670117e-05, + "loss": 0.0017, + "step": 682 + }, + { + "epoch": 2.2076767676767677, + "grad_norm": 0.0036913359072059393, + "learning_rate": 8.835705045278138e-05, + "loss": 0.0009, + "step": 683 + }, + { + "epoch": 2.210909090909091, + "grad_norm": 0.0030470637138932943, + "learning_rate": 8.848641655886159e-05, + "loss": 0.0008, + "step": 684 + }, + { + "epoch": 2.214141414141414, + "grad_norm": 0.004008392803370953, + "learning_rate": 8.86157826649418e-05, + "loss": 0.0012, + "step": 685 + }, + { + "epoch": 2.2173737373737374, + "grad_norm": 0.0055717285722494125, + "learning_rate": 8.8745148771022e-05, + "loss": 0.0019, + "step": 686 + }, + { + "epoch": 2.2206060606060607, + "grad_norm": 0.002990501932799816, + "learning_rate": 8.88745148771022e-05, + "loss": 0.0009, + "step": 687 + }, + { + "epoch": 2.223838383838384, + "grad_norm": 0.002291604643687606, + "learning_rate": 8.90038809831824e-05, + "loss": 0.0006, + "step": 688 + }, + { + "epoch": 2.227070707070707, + "grad_norm": 0.002231605350971222, + "learning_rate": 8.913324708926261e-05, + "loss": 0.0006, + "step": 689 + }, + { + "epoch": 2.2303030303030305, + "grad_norm": 0.002650222275406122, + "learning_rate": 8.926261319534282e-05, + "loss": 0.0007, + "step": 690 + }, + { + "epoch": 2.2335353535353537, + "grad_norm": 0.0019864251371473074, + "learning_rate": 8.939197930142303e-05, + "loss": 0.0006, + "step": 691 + }, + { + "epoch": 2.236767676767677, + "grad_norm": 0.0028250280302017927, + "learning_rate": 8.952134540750324e-05, + "loss": 0.0006, + "step": 692 + }, + { + "epoch": 2.24, + "grad_norm": 0.0035562533885240555, + "learning_rate": 8.965071151358344e-05, + "loss": 0.0018, + "step": 693 + }, + { + "epoch": 2.2432323232323235, + "grad_norm": 0.0030060771387070417, + "learning_rate": 8.978007761966365e-05, + "loss": 0.0007, + "step": 694 + }, + { + "epoch": 2.2464646464646463, + "grad_norm": 0.002113576978445053, + "learning_rate": 8.990944372574386e-05, + "loss": 0.0005, + "step": 695 + }, + { + "epoch": 2.2496969696969695, + "grad_norm": 0.002114512724801898, + "learning_rate": 9.003880983182407e-05, + "loss": 0.0006, + "step": 696 + }, + { + "epoch": 2.252929292929293, + "grad_norm": 0.0029473064932972193, + "learning_rate": 9.016817593790428e-05, + "loss": 0.0006, + "step": 697 + }, + { + "epoch": 2.256161616161616, + "grad_norm": 0.004192824941128492, + "learning_rate": 9.029754204398448e-05, + "loss": 0.001, + "step": 698 + }, + { + "epoch": 2.2593939393939393, + "grad_norm": 0.003509392263367772, + "learning_rate": 9.042690815006469e-05, + "loss": 0.0009, + "step": 699 + }, + { + "epoch": 2.2626262626262625, + "grad_norm": 0.004787352867424488, + "learning_rate": 9.055627425614489e-05, + "loss": 0.001, + "step": 700 + }, + { + "epoch": 2.2626262626262625, + "eval_loss": 0.000903558568097651, + "eval_runtime": 18.694, + "eval_samples_per_second": 5.349, + "eval_steps_per_second": 1.337, + "step": 700 + }, + { + "epoch": 2.265858585858586, + "grad_norm": 0.0027677167672663927, + "learning_rate": 9.06856403622251e-05, + "loss": 0.0009, + "step": 701 + }, + { + "epoch": 2.269090909090909, + "grad_norm": 0.0026491908356547356, + "learning_rate": 9.08150064683053e-05, + "loss": 0.0007, + "step": 702 + }, + { + "epoch": 2.2723232323232323, + "grad_norm": 0.0025004090275615454, + "learning_rate": 9.094437257438552e-05, + "loss": 0.0007, + "step": 703 + }, + { + "epoch": 2.2755555555555556, + "grad_norm": 0.0028262247797101736, + "learning_rate": 9.107373868046573e-05, + "loss": 0.0008, + "step": 704 + }, + { + "epoch": 2.278787878787879, + "grad_norm": 0.003089368110522628, + "learning_rate": 9.120310478654594e-05, + "loss": 0.0009, + "step": 705 + }, + { + "epoch": 2.282020202020202, + "grad_norm": 0.002824244322255254, + "learning_rate": 9.133247089262613e-05, + "loss": 0.0008, + "step": 706 + }, + { + "epoch": 2.2852525252525253, + "grad_norm": 0.0023714362177997828, + "learning_rate": 9.146183699870634e-05, + "loss": 0.0007, + "step": 707 + }, + { + "epoch": 2.2884848484848486, + "grad_norm": 0.0033874395303428173, + "learning_rate": 9.159120310478655e-05, + "loss": 0.0017, + "step": 708 + }, + { + "epoch": 2.291717171717172, + "grad_norm": 0.0033185749780386686, + "learning_rate": 9.172056921086676e-05, + "loss": 0.0017, + "step": 709 + }, + { + "epoch": 2.294949494949495, + "grad_norm": 0.00360478856600821, + "learning_rate": 9.184993531694696e-05, + "loss": 0.0009, + "step": 710 + }, + { + "epoch": 2.2981818181818183, + "grad_norm": 0.0032836326863616705, + "learning_rate": 9.197930142302717e-05, + "loss": 0.0009, + "step": 711 + }, + { + "epoch": 2.3014141414141416, + "grad_norm": 0.0029353760182857513, + "learning_rate": 9.210866752910737e-05, + "loss": 0.0008, + "step": 712 + }, + { + "epoch": 2.304646464646465, + "grad_norm": 0.004789955448359251, + "learning_rate": 9.223803363518758e-05, + "loss": 0.0015, + "step": 713 + }, + { + "epoch": 2.3078787878787876, + "grad_norm": 0.00258410326205194, + "learning_rate": 9.236739974126779e-05, + "loss": 0.0007, + "step": 714 + }, + { + "epoch": 2.311111111111111, + "grad_norm": 0.003211255418136716, + "learning_rate": 9.2496765847348e-05, + "loss": 0.0009, + "step": 715 + }, + { + "epoch": 2.314343434343434, + "grad_norm": 0.003458111546933651, + "learning_rate": 9.262613195342821e-05, + "loss": 0.0008, + "step": 716 + }, + { + "epoch": 2.3175757575757574, + "grad_norm": 0.0020549860782921314, + "learning_rate": 9.275549805950842e-05, + "loss": 0.0005, + "step": 717 + }, + { + "epoch": 2.3208080808080807, + "grad_norm": 0.0034006584901362658, + "learning_rate": 9.288486416558863e-05, + "loss": 0.0011, + "step": 718 + }, + { + "epoch": 2.324040404040404, + "grad_norm": 0.003893098793923855, + "learning_rate": 9.301423027166883e-05, + "loss": 0.0008, + "step": 719 + }, + { + "epoch": 2.327272727272727, + "grad_norm": 0.004374553449451923, + "learning_rate": 9.314359637774904e-05, + "loss": 0.0014, + "step": 720 + }, + { + "epoch": 2.3305050505050504, + "grad_norm": 0.004447360523045063, + "learning_rate": 9.327296248382925e-05, + "loss": 0.0016, + "step": 721 + }, + { + "epoch": 2.3337373737373737, + "grad_norm": 0.003019913798198104, + "learning_rate": 9.340232858990944e-05, + "loss": 0.0007, + "step": 722 + }, + { + "epoch": 2.336969696969697, + "grad_norm": 0.00266702170483768, + "learning_rate": 9.353169469598965e-05, + "loss": 0.0006, + "step": 723 + }, + { + "epoch": 2.34020202020202, + "grad_norm": 0.004037540405988693, + "learning_rate": 9.366106080206986e-05, + "loss": 0.001, + "step": 724 + }, + { + "epoch": 2.3434343434343434, + "grad_norm": 0.003830693429335952, + "learning_rate": 9.379042690815006e-05, + "loss": 0.0015, + "step": 725 + }, + { + "epoch": 2.3434343434343434, + "eval_loss": 0.0009221473592333496, + "eval_runtime": 18.682, + "eval_samples_per_second": 5.353, + "eval_steps_per_second": 1.338, + "step": 725 + }, + { + "epoch": 2.3466666666666667, + "grad_norm": 0.002714785747230053, + "learning_rate": 9.391979301423027e-05, + "loss": 0.0007, + "step": 726 + }, + { + "epoch": 2.34989898989899, + "grad_norm": 0.0034484020434319973, + "learning_rate": 9.404915912031048e-05, + "loss": 0.0021, + "step": 727 + }, + { + "epoch": 2.353131313131313, + "grad_norm": 0.0027550137601792812, + "learning_rate": 9.417852522639069e-05, + "loss": 0.0014, + "step": 728 + }, + { + "epoch": 2.3563636363636364, + "grad_norm": 0.004323527216911316, + "learning_rate": 9.43078913324709e-05, + "loss": 0.0021, + "step": 729 + }, + { + "epoch": 2.3595959595959597, + "grad_norm": 0.0036909414920955896, + "learning_rate": 9.443725743855111e-05, + "loss": 0.0009, + "step": 730 + }, + { + "epoch": 2.362828282828283, + "grad_norm": 0.0026773421559482813, + "learning_rate": 9.456662354463132e-05, + "loss": 0.0007, + "step": 731 + }, + { + "epoch": 2.366060606060606, + "grad_norm": 0.0027335931081324816, + "learning_rate": 9.469598965071152e-05, + "loss": 0.0011, + "step": 732 + }, + { + "epoch": 2.3692929292929295, + "grad_norm": 0.0037821868900209665, + "learning_rate": 9.482535575679173e-05, + "loss": 0.0009, + "step": 733 + }, + { + "epoch": 2.3725252525252527, + "grad_norm": 0.0027637695893645287, + "learning_rate": 9.495472186287193e-05, + "loss": 0.0008, + "step": 734 + }, + { + "epoch": 2.375757575757576, + "grad_norm": 0.0028088942635804415, + "learning_rate": 9.508408796895214e-05, + "loss": 0.0008, + "step": 735 + }, + { + "epoch": 2.378989898989899, + "grad_norm": 0.004683246370404959, + "learning_rate": 9.521345407503235e-05, + "loss": 0.0011, + "step": 736 + }, + { + "epoch": 2.3822222222222225, + "grad_norm": 0.0035344541538506746, + "learning_rate": 9.534282018111256e-05, + "loss": 0.001, + "step": 737 + }, + { + "epoch": 2.3854545454545453, + "grad_norm": 0.003326730104163289, + "learning_rate": 9.547218628719275e-05, + "loss": 0.001, + "step": 738 + }, + { + "epoch": 2.3886868686868685, + "grad_norm": 0.0027331975288689137, + "learning_rate": 9.560155239327296e-05, + "loss": 0.001, + "step": 739 + }, + { + "epoch": 2.391919191919192, + "grad_norm": 0.0023180183488875628, + "learning_rate": 9.573091849935317e-05, + "loss": 0.0005, + "step": 740 + }, + { + "epoch": 2.395151515151515, + "grad_norm": 0.00545561034232378, + "learning_rate": 9.586028460543338e-05, + "loss": 0.0019, + "step": 741 + }, + { + "epoch": 2.3983838383838383, + "grad_norm": 0.002278296509757638, + "learning_rate": 9.59896507115136e-05, + "loss": 0.0008, + "step": 742 + }, + { + "epoch": 2.4016161616161615, + "grad_norm": 0.0036029706243425608, + "learning_rate": 9.61190168175938e-05, + "loss": 0.0011, + "step": 743 + }, + { + "epoch": 2.404848484848485, + "grad_norm": 0.0028302932623773813, + "learning_rate": 9.6248382923674e-05, + "loss": 0.001, + "step": 744 + }, + { + "epoch": 2.408080808080808, + "grad_norm": 0.005623773206025362, + "learning_rate": 9.63777490297542e-05, + "loss": 0.0025, + "step": 745 + }, + { + "epoch": 2.4113131313131313, + "grad_norm": 0.003791616763919592, + "learning_rate": 9.650711513583441e-05, + "loss": 0.001, + "step": 746 + }, + { + "epoch": 2.4145454545454546, + "grad_norm": 0.003847538959234953, + "learning_rate": 9.663648124191462e-05, + "loss": 0.002, + "step": 747 + }, + { + "epoch": 2.417777777777778, + "grad_norm": 0.003144781803712249, + "learning_rate": 9.676584734799483e-05, + "loss": 0.0008, + "step": 748 + }, + { + "epoch": 2.421010101010101, + "grad_norm": 0.0030476360116153955, + "learning_rate": 9.689521345407504e-05, + "loss": 0.002, + "step": 749 + }, + { + "epoch": 2.4242424242424243, + "grad_norm": 0.023542512208223343, + "learning_rate": 9.702457956015525e-05, + "loss": 0.0012, + "step": 750 + }, + { + "epoch": 2.4242424242424243, + "eval_loss": 0.0009755383944138885, + "eval_runtime": 18.7174, + "eval_samples_per_second": 5.343, + "eval_steps_per_second": 1.336, + "step": 750 + }, + { + "epoch": 2.4274747474747476, + "grad_norm": 0.005616334266960621, + "learning_rate": 9.715394566623545e-05, + "loss": 0.0009, + "step": 751 + }, + { + "epoch": 2.430707070707071, + "grad_norm": 0.007717492058873177, + "learning_rate": 9.728331177231566e-05, + "loss": 0.0008, + "step": 752 + }, + { + "epoch": 2.433939393939394, + "grad_norm": 0.008256220258772373, + "learning_rate": 9.741267787839587e-05, + "loss": 0.0011, + "step": 753 + }, + { + "epoch": 2.4371717171717173, + "grad_norm": 0.005972269922494888, + "learning_rate": 9.754204398447608e-05, + "loss": 0.0004, + "step": 754 + }, + { + "epoch": 2.4404040404040406, + "grad_norm": 0.006538939196616411, + "learning_rate": 9.767141009055629e-05, + "loss": 0.0009, + "step": 755 + }, + { + "epoch": 2.443636363636364, + "grad_norm": 0.003234976204112172, + "learning_rate": 9.780077619663648e-05, + "loss": 0.0011, + "step": 756 + }, + { + "epoch": 2.4468686868686866, + "grad_norm": 0.0033180294558405876, + "learning_rate": 9.793014230271668e-05, + "loss": 0.0009, + "step": 757 + }, + { + "epoch": 2.45010101010101, + "grad_norm": 0.0037402757443487644, + "learning_rate": 9.805950840879689e-05, + "loss": 0.0007, + "step": 758 + }, + { + "epoch": 2.453333333333333, + "grad_norm": 0.0028039240278303623, + "learning_rate": 9.81888745148771e-05, + "loss": 0.0008, + "step": 759 + }, + { + "epoch": 2.4565656565656564, + "grad_norm": 0.0029848841950297356, + "learning_rate": 9.831824062095731e-05, + "loss": 0.0007, + "step": 760 + }, + { + "epoch": 2.4597979797979797, + "grad_norm": 0.002809871220961213, + "learning_rate": 9.844760672703752e-05, + "loss": 0.0007, + "step": 761 + }, + { + "epoch": 2.463030303030303, + "grad_norm": 0.004293316043913364, + "learning_rate": 9.857697283311773e-05, + "loss": 0.0012, + "step": 762 + }, + { + "epoch": 2.466262626262626, + "grad_norm": 0.003171822987496853, + "learning_rate": 9.870633893919794e-05, + "loss": 0.0008, + "step": 763 + }, + { + "epoch": 2.4694949494949494, + "grad_norm": 0.0035177527461200953, + "learning_rate": 9.883570504527814e-05, + "loss": 0.0008, + "step": 764 + }, + { + "epoch": 2.4727272727272727, + "grad_norm": 0.0037470462266355753, + "learning_rate": 9.896507115135835e-05, + "loss": 0.0011, + "step": 765 + }, + { + "epoch": 2.475959595959596, + "grad_norm": 0.004131955560296774, + "learning_rate": 9.909443725743856e-05, + "loss": 0.0015, + "step": 766 + }, + { + "epoch": 2.479191919191919, + "grad_norm": 0.002823730930685997, + "learning_rate": 9.922380336351877e-05, + "loss": 0.0007, + "step": 767 + }, + { + "epoch": 2.4824242424242424, + "grad_norm": 0.0037240665405988693, + "learning_rate": 9.935316946959897e-05, + "loss": 0.001, + "step": 768 + }, + { + "epoch": 2.4856565656565657, + "grad_norm": 0.0038744148332625628, + "learning_rate": 9.948253557567918e-05, + "loss": 0.0019, + "step": 769 + }, + { + "epoch": 2.488888888888889, + "grad_norm": 0.003629435319453478, + "learning_rate": 9.961190168175937e-05, + "loss": 0.0008, + "step": 770 + }, + { + "epoch": 2.492121212121212, + "grad_norm": 0.004831044003367424, + "learning_rate": 9.974126778783958e-05, + "loss": 0.0011, + "step": 771 + }, + { + "epoch": 2.4953535353535354, + "grad_norm": 0.003547330852597952, + "learning_rate": 9.98706338939198e-05, + "loss": 0.0009, + "step": 772 + }, + { + "epoch": 2.4985858585858587, + "grad_norm": 0.003263116115704179, + "learning_rate": 0.0001, + "loss": 0.0008, + "step": 773 + }, + { + "epoch": 2.501818181818182, + "grad_norm": 0.0030386094003915787, + "learning_rate": 9.999999489471233e-05, + "loss": 0.0007, + "step": 774 + }, + { + "epoch": 2.505050505050505, + "grad_norm": 0.0036390256136655807, + "learning_rate": 9.99999795788503e-05, + "loss": 0.0012, + "step": 775 + }, + { + "epoch": 2.505050505050505, + "eval_loss": 0.000891694042365998, + "eval_runtime": 18.7365, + "eval_samples_per_second": 5.337, + "eval_steps_per_second": 1.334, + "step": 775 + }, + { + "epoch": 2.5082828282828284, + "grad_norm": 0.0027272645384073257, + "learning_rate": 9.99999540524171e-05, + "loss": 0.0006, + "step": 776 + }, + { + "epoch": 2.5115151515151517, + "grad_norm": 0.004212843254208565, + "learning_rate": 9.999991831541789e-05, + "loss": 0.0023, + "step": 777 + }, + { + "epoch": 2.514747474747475, + "grad_norm": 0.0034173503518104553, + "learning_rate": 9.999987236786e-05, + "loss": 0.0008, + "step": 778 + }, + { + "epoch": 2.517979797979798, + "grad_norm": 0.0028145266696810722, + "learning_rate": 9.999981620975281e-05, + "loss": 0.0007, + "step": 779 + }, + { + "epoch": 2.5212121212121215, + "grad_norm": 0.0030887024477124214, + "learning_rate": 9.999974984110779e-05, + "loss": 0.0009, + "step": 780 + }, + { + "epoch": 2.5244444444444447, + "grad_norm": 0.003968552686274052, + "learning_rate": 9.999967326193847e-05, + "loss": 0.0009, + "step": 781 + }, + { + "epoch": 2.5276767676767675, + "grad_norm": 0.00264795683324337, + "learning_rate": 9.999958647226049e-05, + "loss": 0.0006, + "step": 782 + }, + { + "epoch": 2.5309090909090908, + "grad_norm": 0.004550054203718901, + "learning_rate": 9.999948947209162e-05, + "loss": 0.002, + "step": 783 + }, + { + "epoch": 2.534141414141414, + "grad_norm": 0.0030430385377258062, + "learning_rate": 9.999938226145161e-05, + "loss": 0.0007, + "step": 784 + }, + { + "epoch": 2.5373737373737373, + "grad_norm": 0.003766052657738328, + "learning_rate": 9.999926484036237e-05, + "loss": 0.001, + "step": 785 + }, + { + "epoch": 2.5406060606060605, + "grad_norm": 0.003455114783719182, + "learning_rate": 9.999913720884791e-05, + "loss": 0.002, + "step": 786 + }, + { + "epoch": 2.543838383838384, + "grad_norm": 0.002563537796959281, + "learning_rate": 9.999899936693426e-05, + "loss": 0.0008, + "step": 787 + }, + { + "epoch": 2.547070707070707, + "grad_norm": 0.0025122894439846277, + "learning_rate": 9.99988513146496e-05, + "loss": 0.0006, + "step": 788 + }, + { + "epoch": 2.5503030303030303, + "grad_norm": 0.004589984659105539, + "learning_rate": 9.999869305202412e-05, + "loss": 0.0015, + "step": 789 + }, + { + "epoch": 2.5535353535353535, + "grad_norm": 0.0041996450163424015, + "learning_rate": 9.999852457909018e-05, + "loss": 0.0019, + "step": 790 + }, + { + "epoch": 2.556767676767677, + "grad_norm": 0.0033383311238139868, + "learning_rate": 9.999834589588217e-05, + "loss": 0.0013, + "step": 791 + }, + { + "epoch": 2.56, + "grad_norm": 0.0018662047805264592, + "learning_rate": 9.999815700243656e-05, + "loss": 0.0006, + "step": 792 + }, + { + "epoch": 2.5632323232323233, + "grad_norm": 0.0032810529228299856, + "learning_rate": 9.999795789879196e-05, + "loss": 0.0009, + "step": 793 + }, + { + "epoch": 2.5664646464646466, + "grad_norm": 0.003441553795710206, + "learning_rate": 9.9997748584989e-05, + "loss": 0.0013, + "step": 794 + }, + { + "epoch": 2.56969696969697, + "grad_norm": 0.0025200124364346266, + "learning_rate": 9.999752906107042e-05, + "loss": 0.0008, + "step": 795 + }, + { + "epoch": 2.572929292929293, + "grad_norm": 0.0030850358307361603, + "learning_rate": 9.999729932708109e-05, + "loss": 0.0009, + "step": 796 + }, + { + "epoch": 2.5761616161616163, + "grad_norm": 0.0029367755632847548, + "learning_rate": 9.999705938306789e-05, + "loss": 0.0007, + "step": 797 + }, + { + "epoch": 2.579393939393939, + "grad_norm": 0.003588201245293021, + "learning_rate": 9.999680922907982e-05, + "loss": 0.0026, + "step": 798 + }, + { + "epoch": 2.5826262626262624, + "grad_norm": 0.0031253646593540907, + "learning_rate": 9.999654886516798e-05, + "loss": 0.0009, + "step": 799 + }, + { + "epoch": 2.5858585858585856, + "grad_norm": 0.0036176196299493313, + "learning_rate": 9.999627829138554e-05, + "loss": 0.0015, + "step": 800 + }, + { + "epoch": 2.5858585858585856, + "eval_loss": 0.0009534513228572905, + "eval_runtime": 18.6223, + "eval_samples_per_second": 5.37, + "eval_steps_per_second": 1.342, + "step": 800 + }, + { + "epoch": 2.589090909090909, + "grad_norm": 0.00325636169873178, + "learning_rate": 9.999599750778772e-05, + "loss": 0.0007, + "step": 801 + }, + { + "epoch": 2.592323232323232, + "grad_norm": 0.003968099132180214, + "learning_rate": 9.999570651443191e-05, + "loss": 0.0011, + "step": 802 + }, + { + "epoch": 2.5955555555555554, + "grad_norm": 0.00244711060076952, + "learning_rate": 9.99954053113775e-05, + "loss": 0.0006, + "step": 803 + }, + { + "epoch": 2.5987878787878786, + "grad_norm": 0.003342804964631796, + "learning_rate": 9.9995093898686e-05, + "loss": 0.0014, + "step": 804 + }, + { + "epoch": 2.602020202020202, + "grad_norm": 0.006182527635246515, + "learning_rate": 9.999477227642103e-05, + "loss": 0.0025, + "step": 805 + }, + { + "epoch": 2.605252525252525, + "grad_norm": 0.0032223479356616735, + "learning_rate": 9.999444044464823e-05, + "loss": 0.0007, + "step": 806 + }, + { + "epoch": 2.6084848484848484, + "grad_norm": 0.002208409830927849, + "learning_rate": 9.999409840343539e-05, + "loss": 0.0007, + "step": 807 + }, + { + "epoch": 2.6117171717171717, + "grad_norm": 0.002247196389362216, + "learning_rate": 9.999374615285236e-05, + "loss": 0.0007, + "step": 808 + }, + { + "epoch": 2.614949494949495, + "grad_norm": 0.0024588643573224545, + "learning_rate": 9.999338369297106e-05, + "loss": 0.0005, + "step": 809 + }, + { + "epoch": 2.618181818181818, + "grad_norm": 0.003213444259017706, + "learning_rate": 9.999301102386553e-05, + "loss": 0.0008, + "step": 810 + }, + { + "epoch": 2.6214141414141414, + "grad_norm": 0.0028623330872505903, + "learning_rate": 9.999262814561185e-05, + "loss": 0.0009, + "step": 811 + }, + { + "epoch": 2.6246464646464647, + "grad_norm": 0.0031820102594792843, + "learning_rate": 9.999223505828821e-05, + "loss": 0.001, + "step": 812 + }, + { + "epoch": 2.627878787878788, + "grad_norm": 0.0017483988776803017, + "learning_rate": 9.999183176197491e-05, + "loss": 0.0006, + "step": 813 + }, + { + "epoch": 2.631111111111111, + "grad_norm": 0.0020935633219778538, + "learning_rate": 9.999141825675426e-05, + "loss": 0.0006, + "step": 814 + }, + { + "epoch": 2.6343434343434344, + "grad_norm": 0.003117109416052699, + "learning_rate": 9.999099454271074e-05, + "loss": 0.0008, + "step": 815 + }, + { + "epoch": 2.6375757575757577, + "grad_norm": 0.0037443467881530523, + "learning_rate": 9.999056061993089e-05, + "loss": 0.001, + "step": 816 + }, + { + "epoch": 2.640808080808081, + "grad_norm": 0.0032844438683241606, + "learning_rate": 9.999011648850329e-05, + "loss": 0.0009, + "step": 817 + }, + { + "epoch": 2.644040404040404, + "grad_norm": 0.0027299323119223118, + "learning_rate": 9.998966214851864e-05, + "loss": 0.001, + "step": 818 + }, + { + "epoch": 2.6472727272727274, + "grad_norm": 0.003157002152875066, + "learning_rate": 9.998919760006972e-05, + "loss": 0.001, + "step": 819 + }, + { + "epoch": 2.6505050505050507, + "grad_norm": 0.003021983429789543, + "learning_rate": 9.998872284325142e-05, + "loss": 0.0011, + "step": 820 + }, + { + "epoch": 2.653737373737374, + "grad_norm": 0.0036499700509011745, + "learning_rate": 9.998823787816066e-05, + "loss": 0.001, + "step": 821 + }, + { + "epoch": 2.656969696969697, + "grad_norm": 0.0031615635380148888, + "learning_rate": 9.99877427048965e-05, + "loss": 0.0009, + "step": 822 + }, + { + "epoch": 2.6602020202020205, + "grad_norm": 0.0032300378661602736, + "learning_rate": 9.998723732356006e-05, + "loss": 0.0009, + "step": 823 + }, + { + "epoch": 2.6634343434343437, + "grad_norm": 0.00247188750654459, + "learning_rate": 9.998672173425452e-05, + "loss": 0.0008, + "step": 824 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 0.004531141836196184, + "learning_rate": 9.998619593708518e-05, + "loss": 0.0011, + "step": 825 + }, + { + "epoch": 2.6666666666666665, + "eval_loss": 0.0008804297540336847, + "eval_runtime": 18.6211, + "eval_samples_per_second": 5.37, + "eval_steps_per_second": 1.343, + "step": 825 + }, + { + "epoch": 2.6698989898989898, + "grad_norm": 0.0034359728451818228, + "learning_rate": 9.998565993215943e-05, + "loss": 0.0009, + "step": 826 + }, + { + "epoch": 2.673131313131313, + "grad_norm": 0.002785489894449711, + "learning_rate": 9.998511371958672e-05, + "loss": 0.0007, + "step": 827 + }, + { + "epoch": 2.6763636363636363, + "grad_norm": 0.0025934847071766853, + "learning_rate": 9.998455729947858e-05, + "loss": 0.0008, + "step": 828 + }, + { + "epoch": 2.6795959595959595, + "grad_norm": 0.0026367492973804474, + "learning_rate": 9.998399067194864e-05, + "loss": 0.0008, + "step": 829 + }, + { + "epoch": 2.682828282828283, + "grad_norm": 0.003330858191475272, + "learning_rate": 9.998341383711263e-05, + "loss": 0.0007, + "step": 830 + }, + { + "epoch": 2.686060606060606, + "grad_norm": 0.003155591432005167, + "learning_rate": 9.998282679508835e-05, + "loss": 0.0017, + "step": 831 + }, + { + "epoch": 2.6892929292929293, + "grad_norm": 0.0030347639694809914, + "learning_rate": 9.998222954599563e-05, + "loss": 0.0007, + "step": 832 + }, + { + "epoch": 2.6925252525252525, + "grad_norm": 0.0032036558259278536, + "learning_rate": 9.99816220899565e-05, + "loss": 0.0009, + "step": 833 + }, + { + "epoch": 2.695757575757576, + "grad_norm": 0.003590774955227971, + "learning_rate": 9.998100442709497e-05, + "loss": 0.0009, + "step": 834 + }, + { + "epoch": 2.698989898989899, + "grad_norm": 0.002222040668129921, + "learning_rate": 9.998037655753717e-05, + "loss": 0.0009, + "step": 835 + }, + { + "epoch": 2.7022222222222223, + "grad_norm": 0.0027599751483649015, + "learning_rate": 9.997973848141137e-05, + "loss": 0.0009, + "step": 836 + }, + { + "epoch": 2.7054545454545456, + "grad_norm": 0.00284950970672071, + "learning_rate": 9.997909019884781e-05, + "loss": 0.0008, + "step": 837 + }, + { + "epoch": 2.708686868686869, + "grad_norm": 0.0024073810782283545, + "learning_rate": 9.99784317099789e-05, + "loss": 0.001, + "step": 838 + }, + { + "epoch": 2.711919191919192, + "grad_norm": 0.0033184473868459463, + "learning_rate": 9.997776301493914e-05, + "loss": 0.0016, + "step": 839 + }, + { + "epoch": 2.7151515151515153, + "grad_norm": 0.0023880742955952883, + "learning_rate": 9.997708411386501e-05, + "loss": 0.0015, + "step": 840 + }, + { + "epoch": 2.718383838383838, + "grad_norm": 0.0031732122879475355, + "learning_rate": 9.997639500689523e-05, + "loss": 0.0007, + "step": 841 + }, + { + "epoch": 2.7216161616161614, + "grad_norm": 0.0036892895586788654, + "learning_rate": 9.997569569417049e-05, + "loss": 0.002, + "step": 842 + }, + { + "epoch": 2.7248484848484846, + "grad_norm": 0.002280700486153364, + "learning_rate": 9.997498617583358e-05, + "loss": 0.0018, + "step": 843 + }, + { + "epoch": 2.728080808080808, + "grad_norm": 0.003923532087355852, + "learning_rate": 9.997426645202943e-05, + "loss": 0.0006, + "step": 844 + }, + { + "epoch": 2.731313131313131, + "grad_norm": 0.0025437732692807913, + "learning_rate": 9.9973536522905e-05, + "loss": 0.0009, + "step": 845 + }, + { + "epoch": 2.7345454545454544, + "grad_norm": 0.002573117846623063, + "learning_rate": 9.997279638860933e-05, + "loss": 0.0005, + "step": 846 + }, + { + "epoch": 2.7377777777777776, + "grad_norm": 0.0021492561791092157, + "learning_rate": 9.99720460492936e-05, + "loss": 0.0008, + "step": 847 + }, + { + "epoch": 2.741010101010101, + "grad_norm": 0.0021558962762355804, + "learning_rate": 9.997128550511099e-05, + "loss": 0.0006, + "step": 848 + }, + { + "epoch": 2.744242424242424, + "grad_norm": 0.0031675281934440136, + "learning_rate": 9.997051475621687e-05, + "loss": 0.001, + "step": 849 + }, + { + "epoch": 2.7474747474747474, + "grad_norm": 0.003528911853209138, + "learning_rate": 9.996973380276857e-05, + "loss": 0.0007, + "step": 850 + }, + { + "epoch": 2.7474747474747474, + "eval_loss": 0.000889240182004869, + "eval_runtime": 18.6302, + "eval_samples_per_second": 5.368, + "eval_steps_per_second": 1.342, + "step": 850 + }, + { + "epoch": 2.7507070707070707, + "grad_norm": 0.003921836614608765, + "learning_rate": 9.996894264492563e-05, + "loss": 0.001, + "step": 851 + }, + { + "epoch": 2.753939393939394, + "grad_norm": 0.002436129143461585, + "learning_rate": 9.99681412828496e-05, + "loss": 0.0007, + "step": 852 + }, + { + "epoch": 2.757171717171717, + "grad_norm": 0.002964465180411935, + "learning_rate": 9.996732971670408e-05, + "loss": 0.0008, + "step": 853 + }, + { + "epoch": 2.7604040404040404, + "grad_norm": 0.0034572421573102474, + "learning_rate": 9.996650794665487e-05, + "loss": 0.0007, + "step": 854 + }, + { + "epoch": 2.7636363636363637, + "grad_norm": 0.0021866310853511095, + "learning_rate": 9.996567597286974e-05, + "loss": 0.0005, + "step": 855 + }, + { + "epoch": 2.766868686868687, + "grad_norm": 0.0026556740049272776, + "learning_rate": 9.996483379551861e-05, + "loss": 0.0014, + "step": 856 + }, + { + "epoch": 2.77010101010101, + "grad_norm": 0.003023721743375063, + "learning_rate": 9.996398141477344e-05, + "loss": 0.0007, + "step": 857 + }, + { + "epoch": 2.7733333333333334, + "grad_norm": 0.0022941052448004484, + "learning_rate": 9.996311883080832e-05, + "loss": 0.0006, + "step": 858 + }, + { + "epoch": 2.7765656565656567, + "grad_norm": 0.003141100751236081, + "learning_rate": 9.996224604379938e-05, + "loss": 0.0007, + "step": 859 + }, + { + "epoch": 2.77979797979798, + "grad_norm": 0.002024593763053417, + "learning_rate": 9.996136305392487e-05, + "loss": 0.0006, + "step": 860 + }, + { + "epoch": 2.783030303030303, + "grad_norm": 0.004872876685112715, + "learning_rate": 9.996046986136509e-05, + "loss": 0.0026, + "step": 861 + }, + { + "epoch": 2.7862626262626264, + "grad_norm": 0.002601329004392028, + "learning_rate": 9.995956646630246e-05, + "loss": 0.0009, + "step": 862 + }, + { + "epoch": 2.7894949494949497, + "grad_norm": 0.0028132593724876642, + "learning_rate": 9.995865286892145e-05, + "loss": 0.0009, + "step": 863 + }, + { + "epoch": 2.792727272727273, + "grad_norm": 0.002662686165422201, + "learning_rate": 9.995772906940864e-05, + "loss": 0.0007, + "step": 864 + }, + { + "epoch": 2.795959595959596, + "grad_norm": 0.0027195930015295744, + "learning_rate": 9.995679506795264e-05, + "loss": 0.0007, + "step": 865 + }, + { + "epoch": 2.7991919191919195, + "grad_norm": 0.003179864026606083, + "learning_rate": 9.995585086474424e-05, + "loss": 0.0014, + "step": 866 + }, + { + "epoch": 2.8024242424242423, + "grad_norm": 0.0034151843283325434, + "learning_rate": 9.995489645997622e-05, + "loss": 0.0009, + "step": 867 + }, + { + "epoch": 2.8056565656565655, + "grad_norm": 0.0026823675725609064, + "learning_rate": 9.99539318538435e-05, + "loss": 0.0007, + "step": 868 + }, + { + "epoch": 2.8088888888888888, + "grad_norm": 0.0021664374507963657, + "learning_rate": 9.995295704654304e-05, + "loss": 0.0012, + "step": 869 + }, + { + "epoch": 2.812121212121212, + "grad_norm": 0.003440326079726219, + "learning_rate": 9.995197203827393e-05, + "loss": 0.0009, + "step": 870 + }, + { + "epoch": 2.8153535353535353, + "grad_norm": 0.003980646841228008, + "learning_rate": 9.995097682923733e-05, + "loss": 0.0018, + "step": 871 + }, + { + "epoch": 2.8185858585858585, + "grad_norm": 0.003389423480257392, + "learning_rate": 9.994997141963644e-05, + "loss": 0.001, + "step": 872 + }, + { + "epoch": 2.821818181818182, + "grad_norm": 0.0027842579875141382, + "learning_rate": 9.994895580967658e-05, + "loss": 0.0007, + "step": 873 + }, + { + "epoch": 2.825050505050505, + "grad_norm": 0.0033308009151369333, + "learning_rate": 9.994792999956518e-05, + "loss": 0.0011, + "step": 874 + }, + { + "epoch": 2.8282828282828283, + "grad_norm": 0.002514815656468272, + "learning_rate": 9.994689398951169e-05, + "loss": 0.0009, + "step": 875 + }, + { + "epoch": 2.8282828282828283, + "eval_loss": 0.0008539481204934418, + "eval_runtime": 18.6059, + "eval_samples_per_second": 5.375, + "eval_steps_per_second": 1.344, + "step": 875 + }, + { + "epoch": 2.8315151515151515, + "grad_norm": 0.0024293591268360615, + "learning_rate": 9.994584777972769e-05, + "loss": 0.0008, + "step": 876 + }, + { + "epoch": 2.834747474747475, + "grad_norm": 0.0035905223339796066, + "learning_rate": 9.994479137042683e-05, + "loss": 0.002, + "step": 877 + }, + { + "epoch": 2.837979797979798, + "grad_norm": 0.002918825950473547, + "learning_rate": 9.994372476182484e-05, + "loss": 0.0009, + "step": 878 + }, + { + "epoch": 2.8412121212121213, + "grad_norm": 0.0031527606770396233, + "learning_rate": 9.994264795413953e-05, + "loss": 0.0008, + "step": 879 + }, + { + "epoch": 2.8444444444444446, + "grad_norm": 0.0035410267300903797, + "learning_rate": 9.99415609475908e-05, + "loss": 0.0021, + "step": 880 + }, + { + "epoch": 2.847676767676768, + "grad_norm": 0.0020397889893501997, + "learning_rate": 9.994046374240062e-05, + "loss": 0.0007, + "step": 881 + }, + { + "epoch": 2.850909090909091, + "grad_norm": 0.002342833438888192, + "learning_rate": 9.993935633879306e-05, + "loss": 0.0007, + "step": 882 + }, + { + "epoch": 2.854141414141414, + "grad_norm": 0.003542246064171195, + "learning_rate": 9.993823873699426e-05, + "loss": 0.0008, + "step": 883 + }, + { + "epoch": 2.857373737373737, + "grad_norm": 0.0024878752883523703, + "learning_rate": 9.993711093723245e-05, + "loss": 0.0008, + "step": 884 + }, + { + "epoch": 2.8606060606060604, + "grad_norm": 0.0030815114732831717, + "learning_rate": 9.993597293973796e-05, + "loss": 0.0018, + "step": 885 + }, + { + "epoch": 2.8638383838383836, + "grad_norm": 0.004479612223803997, + "learning_rate": 9.993482474474314e-05, + "loss": 0.0009, + "step": 886 + }, + { + "epoch": 2.867070707070707, + "grad_norm": 0.0033372335601598024, + "learning_rate": 9.99336663524825e-05, + "loss": 0.0009, + "step": 887 + }, + { + "epoch": 2.87030303030303, + "grad_norm": 0.0021939794532954693, + "learning_rate": 9.993249776319258e-05, + "loss": 0.0006, + "step": 888 + }, + { + "epoch": 2.8735353535353534, + "grad_norm": 0.0028632464818656445, + "learning_rate": 9.993131897711202e-05, + "loss": 0.0009, + "step": 889 + }, + { + "epoch": 2.8767676767676766, + "grad_norm": 0.002615751465782523, + "learning_rate": 9.993012999448154e-05, + "loss": 0.0009, + "step": 890 + }, + { + "epoch": 2.88, + "grad_norm": 0.0017363366205245256, + "learning_rate": 9.992893081554397e-05, + "loss": 0.0005, + "step": 891 + }, + { + "epoch": 2.883232323232323, + "grad_norm": 0.0032067778520286083, + "learning_rate": 9.992772144054415e-05, + "loss": 0.0009, + "step": 892 + }, + { + "epoch": 2.8864646464646464, + "grad_norm": 0.0025459511671215296, + "learning_rate": 9.992650186972909e-05, + "loss": 0.0008, + "step": 893 + }, + { + "epoch": 2.8896969696969697, + "grad_norm": 0.001871303771622479, + "learning_rate": 9.99252721033478e-05, + "loss": 0.0008, + "step": 894 + }, + { + "epoch": 2.892929292929293, + "grad_norm": 0.001998218474909663, + "learning_rate": 9.992403214165147e-05, + "loss": 0.0005, + "step": 895 + }, + { + "epoch": 2.896161616161616, + "grad_norm": 0.0023115125950425863, + "learning_rate": 9.992278198489327e-05, + "loss": 0.0007, + "step": 896 + }, + { + "epoch": 2.8993939393939394, + "grad_norm": 0.0019588919822126627, + "learning_rate": 9.99215216333285e-05, + "loss": 0.0007, + "step": 897 + }, + { + "epoch": 2.9026262626262627, + "grad_norm": 0.001903701457194984, + "learning_rate": 9.992025108721454e-05, + "loss": 0.0004, + "step": 898 + }, + { + "epoch": 2.905858585858586, + "grad_norm": 0.0023024296388030052, + "learning_rate": 9.991897034681087e-05, + "loss": 0.0006, + "step": 899 + }, + { + "epoch": 2.909090909090909, + "grad_norm": 0.002626319881528616, + "learning_rate": 9.9917679412379e-05, + "loss": 0.0009, + "step": 900 + }, + { + "epoch": 2.909090909090909, + "eval_loss": 0.0008496911614201963, + "eval_runtime": 18.6252, + "eval_samples_per_second": 5.369, + "eval_steps_per_second": 1.342, + "step": 900 + } + ], + "logging_steps": 1, + "max_steps": 7725, + "num_input_tokens_seen": 0, + "num_train_epochs": 25, + "save_steps": 100, + "stateful_callbacks": { + "EarlyStoppingCallback": { + "args": { + "early_stopping_patience": 5, + "early_stopping_threshold": 0.0 + }, + "attributes": { + "early_stopping_patience_counter": 0 + } + }, + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.268707775185879e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}