{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.18851918182675087, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00018851918182675087, "grad_norm": 1.6484375, "learning_rate": 5e-06, "loss": 2.7659, "step": 1 }, { "epoch": 0.00037703836365350174, "grad_norm": 1.6640625, "learning_rate": 1e-05, "loss": 2.5842, "step": 2 }, { "epoch": 0.0005655575454802526, "grad_norm": 1.609375, "learning_rate": 1.5e-05, "loss": 2.8169, "step": 3 }, { "epoch": 0.0007540767273070035, "grad_norm": 1.640625, "learning_rate": 2e-05, "loss": 2.6938, "step": 4 }, { "epoch": 0.0009425959091337543, "grad_norm": 1.546875, "learning_rate": 2.5e-05, "loss": 2.7862, "step": 5 }, { "epoch": 0.0011311150909605052, "grad_norm": 1.671875, "learning_rate": 3e-05, "loss": 2.8844, "step": 6 }, { "epoch": 0.0013196342727872562, "grad_norm": 1.5703125, "learning_rate": 3.5000000000000004e-05, "loss": 2.8254, "step": 7 }, { "epoch": 0.001508153454614007, "grad_norm": 1.5625, "learning_rate": 4e-05, "loss": 2.7735, "step": 8 }, { "epoch": 0.001696672636440758, "grad_norm": 1.609375, "learning_rate": 4.4999999999999996e-05, "loss": 2.8222, "step": 9 }, { "epoch": 0.0018851918182675087, "grad_norm": 1.6328125, "learning_rate": 5e-05, "loss": 2.6943, "step": 10 }, { "epoch": 0.0020737110000942595, "grad_norm": 1.5546875, "learning_rate": 5.5e-05, "loss": 2.6735, "step": 11 }, { "epoch": 0.0022622301819210104, "grad_norm": 1.6640625, "learning_rate": 6e-05, "loss": 2.6482, "step": 12 }, { "epoch": 0.0024507493637477614, "grad_norm": 1.4375, "learning_rate": 6.500000000000001e-05, "loss": 2.8788, "step": 13 }, { "epoch": 0.0026392685455745124, "grad_norm": 1.6328125, "learning_rate": 7.000000000000001e-05, "loss": 2.7531, "step": 14 }, { "epoch": 0.002827787727401263, "grad_norm": 1.53125, "learning_rate": 7.5e-05, "loss": 2.7911, "step": 15 }, { "epoch": 0.003016306909228014, "grad_norm": 1.5703125, "learning_rate": 8e-05, "loss": 2.7358, "step": 16 }, { "epoch": 0.003204826091054765, "grad_norm": 1.5859375, "learning_rate": 8.5e-05, "loss": 2.7272, "step": 17 }, { "epoch": 0.003393345272881516, "grad_norm": 1.515625, "learning_rate": 8.999999999999999e-05, "loss": 2.7176, "step": 18 }, { "epoch": 0.0035818644547082664, "grad_norm": 1.5, "learning_rate": 9.5e-05, "loss": 2.8573, "step": 19 }, { "epoch": 0.0037703836365350174, "grad_norm": 1.5234375, "learning_rate": 0.0001, "loss": 2.7512, "step": 20 }, { "epoch": 0.003958902818361768, "grad_norm": 1.296875, "learning_rate": 0.000105, "loss": 2.7962, "step": 21 }, { "epoch": 0.004147422000188519, "grad_norm": 1.3671875, "learning_rate": 0.00011, "loss": 2.7, "step": 22 }, { "epoch": 0.00433594118201527, "grad_norm": 1.296875, "learning_rate": 0.000115, "loss": 2.7128, "step": 23 }, { "epoch": 0.004524460363842021, "grad_norm": 1.2890625, "learning_rate": 0.00012, "loss": 2.729, "step": 24 }, { "epoch": 0.004712979545668771, "grad_norm": 1.25, "learning_rate": 0.000125, "loss": 2.698, "step": 25 }, { "epoch": 0.004901498727495523, "grad_norm": 1.3046875, "learning_rate": 0.00013000000000000002, "loss": 2.7461, "step": 26 }, { "epoch": 0.005090017909322273, "grad_norm": 1.1796875, "learning_rate": 0.000135, "loss": 2.7315, "step": 27 }, { "epoch": 0.005278537091149025, "grad_norm": 1.15625, "learning_rate": 0.00014000000000000001, "loss": 2.7089, "step": 28 }, { "epoch": 0.005467056272975775, "grad_norm": 1.1875, "learning_rate": 0.000145, "loss": 2.6724, "step": 29 }, { "epoch": 0.005655575454802526, "grad_norm": 1.0859375, "learning_rate": 0.00015, "loss": 2.799, "step": 30 }, { "epoch": 0.005844094636629277, "grad_norm": 1.0546875, "learning_rate": 0.000155, "loss": 2.7939, "step": 31 }, { "epoch": 0.006032613818456028, "grad_norm": 1.078125, "learning_rate": 0.00016, "loss": 2.8004, "step": 32 }, { "epoch": 0.006221133000282778, "grad_norm": 1.0390625, "learning_rate": 0.000165, "loss": 2.6322, "step": 33 }, { "epoch": 0.00640965218210953, "grad_norm": 1.0, "learning_rate": 0.00017, "loss": 2.7095, "step": 34 }, { "epoch": 0.00659817136393628, "grad_norm": 1.0234375, "learning_rate": 0.000175, "loss": 2.7111, "step": 35 }, { "epoch": 0.006786690545763032, "grad_norm": 0.91796875, "learning_rate": 0.00017999999999999998, "loss": 2.7666, "step": 36 }, { "epoch": 0.006975209727589782, "grad_norm": 0.921875, "learning_rate": 0.000185, "loss": 2.7779, "step": 37 }, { "epoch": 0.007163728909416533, "grad_norm": 0.91015625, "learning_rate": 0.00019, "loss": 2.7684, "step": 38 }, { "epoch": 0.007352248091243284, "grad_norm": 0.875, "learning_rate": 0.00019500000000000002, "loss": 2.8674, "step": 39 }, { "epoch": 0.007540767273070035, "grad_norm": 0.8515625, "learning_rate": 0.0002, "loss": 2.7694, "step": 40 }, { "epoch": 0.007729286454896786, "grad_norm": 0.85546875, "learning_rate": 0.000205, "loss": 2.6799, "step": 41 }, { "epoch": 0.007917805636723537, "grad_norm": 0.8046875, "learning_rate": 0.00021, "loss": 2.6289, "step": 42 }, { "epoch": 0.008106324818550288, "grad_norm": 0.859375, "learning_rate": 0.000215, "loss": 2.7937, "step": 43 }, { "epoch": 0.008294844000377038, "grad_norm": 0.80859375, "learning_rate": 0.00022, "loss": 2.78, "step": 44 }, { "epoch": 0.00848336318220379, "grad_norm": 0.765625, "learning_rate": 0.00022500000000000002, "loss": 2.6351, "step": 45 }, { "epoch": 0.00867188236403054, "grad_norm": 0.7890625, "learning_rate": 0.00023, "loss": 2.8156, "step": 46 }, { "epoch": 0.00886040154585729, "grad_norm": 0.7890625, "learning_rate": 0.000235, "loss": 2.8304, "step": 47 }, { "epoch": 0.009048920727684042, "grad_norm": 0.76171875, "learning_rate": 0.00024, "loss": 2.7148, "step": 48 }, { "epoch": 0.009237439909510793, "grad_norm": 0.75390625, "learning_rate": 0.000245, "loss": 2.7169, "step": 49 }, { "epoch": 0.009425959091337543, "grad_norm": 0.76953125, "learning_rate": 0.00025, "loss": 2.8345, "step": 50 }, { "epoch": 0.009614478273164294, "grad_norm": 0.73046875, "learning_rate": 0.000255, "loss": 2.8149, "step": 51 }, { "epoch": 0.009802997454991046, "grad_norm": 0.7421875, "learning_rate": 0.00026000000000000003, "loss": 2.8182, "step": 52 }, { "epoch": 0.009991516636817797, "grad_norm": 0.7578125, "learning_rate": 0.00026500000000000004, "loss": 2.8114, "step": 53 }, { "epoch": 0.010180035818644547, "grad_norm": 0.71875, "learning_rate": 0.00027, "loss": 2.803, "step": 54 }, { "epoch": 0.010368555000471298, "grad_norm": 0.73046875, "learning_rate": 0.000275, "loss": 2.7979, "step": 55 }, { "epoch": 0.01055707418229805, "grad_norm": 0.73828125, "learning_rate": 0.00028000000000000003, "loss": 2.8062, "step": 56 }, { "epoch": 0.0107455933641248, "grad_norm": 0.76171875, "learning_rate": 0.000285, "loss": 2.6728, "step": 57 }, { "epoch": 0.01093411254595155, "grad_norm": 0.73046875, "learning_rate": 0.00029, "loss": 2.7547, "step": 58 }, { "epoch": 0.011122631727778302, "grad_norm": 0.71875, "learning_rate": 0.000295, "loss": 2.6773, "step": 59 }, { "epoch": 0.011311150909605052, "grad_norm": 0.7109375, "learning_rate": 0.0003, "loss": 2.7238, "step": 60 }, { "epoch": 0.011499670091431803, "grad_norm": 0.77734375, "learning_rate": 0.000305, "loss": 2.6842, "step": 61 }, { "epoch": 0.011688189273258555, "grad_norm": 0.703125, "learning_rate": 0.00031, "loss": 2.8449, "step": 62 }, { "epoch": 0.011876708455085304, "grad_norm": 0.703125, "learning_rate": 0.000315, "loss": 2.6828, "step": 63 }, { "epoch": 0.012065227636912056, "grad_norm": 0.6796875, "learning_rate": 0.00032, "loss": 2.7663, "step": 64 }, { "epoch": 0.012253746818738807, "grad_norm": 0.69140625, "learning_rate": 0.00032500000000000004, "loss": 2.6127, "step": 65 }, { "epoch": 0.012442266000565557, "grad_norm": 0.70703125, "learning_rate": 0.00033, "loss": 2.6333, "step": 66 }, { "epoch": 0.012630785182392308, "grad_norm": 0.6796875, "learning_rate": 0.000335, "loss": 2.7669, "step": 67 }, { "epoch": 0.01281930436421906, "grad_norm": 0.7265625, "learning_rate": 0.00034, "loss": 2.7363, "step": 68 }, { "epoch": 0.013007823546045811, "grad_norm": 0.7109375, "learning_rate": 0.000345, "loss": 2.6626, "step": 69 }, { "epoch": 0.01319634272787256, "grad_norm": 0.71875, "learning_rate": 0.00035, "loss": 2.7896, "step": 70 }, { "epoch": 0.013384861909699312, "grad_norm": 0.69140625, "learning_rate": 0.000355, "loss": 2.7407, "step": 71 }, { "epoch": 0.013573381091526063, "grad_norm": 0.6953125, "learning_rate": 0.00035999999999999997, "loss": 2.804, "step": 72 }, { "epoch": 0.013761900273352813, "grad_norm": 0.69140625, "learning_rate": 0.000365, "loss": 2.781, "step": 73 }, { "epoch": 0.013950419455179565, "grad_norm": 0.6875, "learning_rate": 0.00037, "loss": 2.5436, "step": 74 }, { "epoch": 0.014138938637006316, "grad_norm": 0.66796875, "learning_rate": 0.000375, "loss": 2.7272, "step": 75 }, { "epoch": 0.014327457818833066, "grad_norm": 0.68359375, "learning_rate": 0.00038, "loss": 2.6777, "step": 76 }, { "epoch": 0.014515977000659817, "grad_norm": 0.7734375, "learning_rate": 0.00038500000000000003, "loss": 2.8211, "step": 77 }, { "epoch": 0.014704496182486568, "grad_norm": 0.6953125, "learning_rate": 0.00039000000000000005, "loss": 2.7639, "step": 78 }, { "epoch": 0.014893015364313318, "grad_norm": 0.6953125, "learning_rate": 0.000395, "loss": 2.6884, "step": 79 }, { "epoch": 0.01508153454614007, "grad_norm": 0.76171875, "learning_rate": 0.0004, "loss": 2.6492, "step": 80 }, { "epoch": 0.015270053727966821, "grad_norm": 0.6796875, "learning_rate": 0.00040500000000000003, "loss": 2.8072, "step": 81 }, { "epoch": 0.015458572909793572, "grad_norm": 0.7109375, "learning_rate": 0.00041, "loss": 2.7446, "step": 82 }, { "epoch": 0.015647092091620324, "grad_norm": 0.69140625, "learning_rate": 0.000415, "loss": 2.7554, "step": 83 }, { "epoch": 0.015835611273447073, "grad_norm": 0.70703125, "learning_rate": 0.00042, "loss": 2.7212, "step": 84 }, { "epoch": 0.016024130455273823, "grad_norm": 0.72265625, "learning_rate": 0.000425, "loss": 2.6933, "step": 85 }, { "epoch": 0.016212649637100576, "grad_norm": 0.69140625, "learning_rate": 0.00043, "loss": 2.7461, "step": 86 }, { "epoch": 0.016401168818927326, "grad_norm": 0.671875, "learning_rate": 0.000435, "loss": 2.7079, "step": 87 }, { "epoch": 0.016589688000754076, "grad_norm": 0.66796875, "learning_rate": 0.00044, "loss": 2.8562, "step": 88 }, { "epoch": 0.01677820718258083, "grad_norm": 0.671875, "learning_rate": 0.00044500000000000003, "loss": 2.6606, "step": 89 }, { "epoch": 0.01696672636440758, "grad_norm": 0.69140625, "learning_rate": 0.00045000000000000004, "loss": 2.7817, "step": 90 }, { "epoch": 0.017155245546234328, "grad_norm": 0.74609375, "learning_rate": 0.000455, "loss": 2.7714, "step": 91 }, { "epoch": 0.01734376472806108, "grad_norm": 0.68359375, "learning_rate": 0.00046, "loss": 2.7217, "step": 92 }, { "epoch": 0.01753228390988783, "grad_norm": 0.671875, "learning_rate": 0.000465, "loss": 2.6855, "step": 93 }, { "epoch": 0.01772080309171458, "grad_norm": 0.734375, "learning_rate": 0.00047, "loss": 2.7111, "step": 94 }, { "epoch": 0.017909322273541334, "grad_norm": 0.71484375, "learning_rate": 0.000475, "loss": 2.6868, "step": 95 }, { "epoch": 0.018097841455368083, "grad_norm": 0.71484375, "learning_rate": 0.00048, "loss": 2.7355, "step": 96 }, { "epoch": 0.018286360637194833, "grad_norm": 0.69921875, "learning_rate": 0.00048499999999999997, "loss": 2.7172, "step": 97 }, { "epoch": 0.018474879819021586, "grad_norm": 0.6796875, "learning_rate": 0.00049, "loss": 2.8204, "step": 98 }, { "epoch": 0.018663399000848336, "grad_norm": 0.6875, "learning_rate": 0.000495, "loss": 2.6965, "step": 99 }, { "epoch": 0.018851918182675086, "grad_norm": 0.6875, "learning_rate": 0.0005, "loss": 2.7988, "step": 100 }, { "epoch": 0.01904043736450184, "grad_norm": 0.72265625, "learning_rate": 0.000505, "loss": 2.7069, "step": 101 }, { "epoch": 0.01922895654632859, "grad_norm": 0.6796875, "learning_rate": 0.00051, "loss": 2.6942, "step": 102 }, { "epoch": 0.019417475728155338, "grad_norm": 0.66015625, "learning_rate": 0.000515, "loss": 2.7497, "step": 103 }, { "epoch": 0.01960599490998209, "grad_norm": 0.6875, "learning_rate": 0.0005200000000000001, "loss": 2.6381, "step": 104 }, { "epoch": 0.01979451409180884, "grad_norm": 0.6796875, "learning_rate": 0.0005250000000000001, "loss": 2.6969, "step": 105 }, { "epoch": 0.019983033273635594, "grad_norm": 0.66015625, "learning_rate": 0.0005300000000000001, "loss": 2.7247, "step": 106 }, { "epoch": 0.020171552455462344, "grad_norm": 0.72265625, "learning_rate": 0.000535, "loss": 2.828, "step": 107 }, { "epoch": 0.020360071637289093, "grad_norm": 0.7265625, "learning_rate": 0.00054, "loss": 2.7309, "step": 108 }, { "epoch": 0.020548590819115847, "grad_norm": 0.71484375, "learning_rate": 0.000545, "loss": 2.8354, "step": 109 }, { "epoch": 0.020737110000942596, "grad_norm": 0.69140625, "learning_rate": 0.00055, "loss": 2.8101, "step": 110 }, { "epoch": 0.020925629182769346, "grad_norm": 0.69921875, "learning_rate": 0.000555, "loss": 2.7837, "step": 111 }, { "epoch": 0.0211141483645961, "grad_norm": 0.69140625, "learning_rate": 0.0005600000000000001, "loss": 2.6813, "step": 112 }, { "epoch": 0.02130266754642285, "grad_norm": 0.65234375, "learning_rate": 0.000565, "loss": 2.7035, "step": 113 }, { "epoch": 0.0214911867282496, "grad_norm": 0.75390625, "learning_rate": 0.00057, "loss": 2.6901, "step": 114 }, { "epoch": 0.02167970591007635, "grad_norm": 0.75, "learning_rate": 0.000575, "loss": 2.7001, "step": 115 }, { "epoch": 0.0218682250919031, "grad_norm": 0.70703125, "learning_rate": 0.00058, "loss": 2.7508, "step": 116 }, { "epoch": 0.02205674427372985, "grad_norm": 0.65625, "learning_rate": 0.000585, "loss": 2.7348, "step": 117 }, { "epoch": 0.022245263455556604, "grad_norm": 0.6796875, "learning_rate": 0.00059, "loss": 2.7434, "step": 118 }, { "epoch": 0.022433782637383354, "grad_norm": 0.66015625, "learning_rate": 0.0005949999999999999, "loss": 2.6735, "step": 119 }, { "epoch": 0.022622301819210103, "grad_norm": 0.703125, "learning_rate": 0.0006, "loss": 2.6258, "step": 120 }, { "epoch": 0.022810821001036857, "grad_norm": 0.75, "learning_rate": 0.000605, "loss": 2.7676, "step": 121 }, { "epoch": 0.022999340182863606, "grad_norm": 0.6875, "learning_rate": 0.00061, "loss": 2.7045, "step": 122 }, { "epoch": 0.023187859364690356, "grad_norm": 0.66015625, "learning_rate": 0.000615, "loss": 2.6322, "step": 123 }, { "epoch": 0.02337637854651711, "grad_norm": 0.7109375, "learning_rate": 0.00062, "loss": 2.6953, "step": 124 }, { "epoch": 0.02356489772834386, "grad_norm": 0.71875, "learning_rate": 0.000625, "loss": 2.6045, "step": 125 }, { "epoch": 0.02375341691017061, "grad_norm": 0.71484375, "learning_rate": 0.00063, "loss": 2.6551, "step": 126 }, { "epoch": 0.02394193609199736, "grad_norm": 0.69921875, "learning_rate": 0.000635, "loss": 2.656, "step": 127 }, { "epoch": 0.02413045527382411, "grad_norm": 0.81640625, "learning_rate": 0.00064, "loss": 2.791, "step": 128 }, { "epoch": 0.02431897445565086, "grad_norm": 0.69921875, "learning_rate": 0.0006450000000000001, "loss": 2.6599, "step": 129 }, { "epoch": 0.024507493637477614, "grad_norm": 0.7421875, "learning_rate": 0.0006500000000000001, "loss": 2.633, "step": 130 }, { "epoch": 0.024696012819304364, "grad_norm": 0.74609375, "learning_rate": 0.0006550000000000001, "loss": 2.6002, "step": 131 }, { "epoch": 0.024884532001131113, "grad_norm": 0.68359375, "learning_rate": 0.00066, "loss": 2.7593, "step": 132 }, { "epoch": 0.025073051182957867, "grad_norm": 0.69921875, "learning_rate": 0.000665, "loss": 2.706, "step": 133 }, { "epoch": 0.025261570364784616, "grad_norm": 0.7109375, "learning_rate": 0.00067, "loss": 2.7094, "step": 134 }, { "epoch": 0.02545008954661137, "grad_norm": 0.796875, "learning_rate": 0.000675, "loss": 2.6961, "step": 135 }, { "epoch": 0.02563860872843812, "grad_norm": 0.74609375, "learning_rate": 0.00068, "loss": 2.7805, "step": 136 }, { "epoch": 0.02582712791026487, "grad_norm": 0.71875, "learning_rate": 0.0006850000000000001, "loss": 2.6559, "step": 137 }, { "epoch": 0.026015647092091622, "grad_norm": 0.6796875, "learning_rate": 0.00069, "loss": 2.7455, "step": 138 }, { "epoch": 0.02620416627391837, "grad_norm": 0.734375, "learning_rate": 0.000695, "loss": 2.7533, "step": 139 }, { "epoch": 0.02639268545574512, "grad_norm": 0.75, "learning_rate": 0.0007, "loss": 2.7434, "step": 140 }, { "epoch": 0.026581204637571874, "grad_norm": 0.6953125, "learning_rate": 0.000705, "loss": 2.7018, "step": 141 }, { "epoch": 0.026769723819398624, "grad_norm": 0.70703125, "learning_rate": 0.00071, "loss": 2.6182, "step": 142 }, { "epoch": 0.026958243001225374, "grad_norm": 0.71875, "learning_rate": 0.000715, "loss": 2.5742, "step": 143 }, { "epoch": 0.027146762183052127, "grad_norm": 0.68359375, "learning_rate": 0.0007199999999999999, "loss": 2.6547, "step": 144 }, { "epoch": 0.027335281364878877, "grad_norm": 0.74609375, "learning_rate": 0.000725, "loss": 2.7054, "step": 145 }, { "epoch": 0.027523800546705626, "grad_norm": 0.75390625, "learning_rate": 0.00073, "loss": 2.5809, "step": 146 }, { "epoch": 0.02771231972853238, "grad_norm": 0.71484375, "learning_rate": 0.000735, "loss": 2.6474, "step": 147 }, { "epoch": 0.02790083891035913, "grad_norm": 0.71484375, "learning_rate": 0.00074, "loss": 2.7606, "step": 148 }, { "epoch": 0.02808935809218588, "grad_norm": 0.7734375, "learning_rate": 0.000745, "loss": 2.6923, "step": 149 }, { "epoch": 0.028277877274012632, "grad_norm": 0.7421875, "learning_rate": 0.00075, "loss": 2.782, "step": 150 }, { "epoch": 0.02846639645583938, "grad_norm": 0.69140625, "learning_rate": 0.000755, "loss": 2.7369, "step": 151 }, { "epoch": 0.02865491563766613, "grad_norm": 0.74609375, "learning_rate": 0.00076, "loss": 2.6287, "step": 152 }, { "epoch": 0.028843434819492884, "grad_norm": 0.765625, "learning_rate": 0.0007650000000000001, "loss": 2.6649, "step": 153 }, { "epoch": 0.029031954001319634, "grad_norm": 0.875, "learning_rate": 0.0007700000000000001, "loss": 2.7421, "step": 154 }, { "epoch": 0.029220473183146384, "grad_norm": 0.734375, "learning_rate": 0.0007750000000000001, "loss": 2.5988, "step": 155 }, { "epoch": 0.029408992364973137, "grad_norm": 0.734375, "learning_rate": 0.0007800000000000001, "loss": 2.6876, "step": 156 }, { "epoch": 0.029597511546799887, "grad_norm": 0.796875, "learning_rate": 0.000785, "loss": 2.6846, "step": 157 }, { "epoch": 0.029786030728626636, "grad_norm": 0.7734375, "learning_rate": 0.00079, "loss": 2.7869, "step": 158 }, { "epoch": 0.02997454991045339, "grad_norm": 0.68359375, "learning_rate": 0.000795, "loss": 2.6972, "step": 159 }, { "epoch": 0.03016306909228014, "grad_norm": 0.73828125, "learning_rate": 0.0008, "loss": 2.7664, "step": 160 }, { "epoch": 0.03035158827410689, "grad_norm": 0.69140625, "learning_rate": 0.000805, "loss": 2.6554, "step": 161 }, { "epoch": 0.030540107455933642, "grad_norm": 0.69140625, "learning_rate": 0.0008100000000000001, "loss": 2.662, "step": 162 }, { "epoch": 0.03072862663776039, "grad_norm": 0.734375, "learning_rate": 0.000815, "loss": 2.622, "step": 163 }, { "epoch": 0.030917145819587145, "grad_norm": 0.73828125, "learning_rate": 0.00082, "loss": 2.6071, "step": 164 }, { "epoch": 0.031105665001413894, "grad_norm": 0.7421875, "learning_rate": 0.000825, "loss": 2.6724, "step": 165 }, { "epoch": 0.03129418418324065, "grad_norm": 0.71875, "learning_rate": 0.00083, "loss": 2.5888, "step": 166 }, { "epoch": 0.031482703365067394, "grad_norm": 0.75390625, "learning_rate": 0.000835, "loss": 2.7932, "step": 167 }, { "epoch": 0.03167122254689415, "grad_norm": 0.734375, "learning_rate": 0.00084, "loss": 2.6234, "step": 168 }, { "epoch": 0.0318597417287209, "grad_norm": 0.73828125, "learning_rate": 0.0008449999999999999, "loss": 2.6725, "step": 169 }, { "epoch": 0.032048260910547646, "grad_norm": 0.8515625, "learning_rate": 0.00085, "loss": 2.6502, "step": 170 }, { "epoch": 0.0322367800923744, "grad_norm": 0.7109375, "learning_rate": 0.000855, "loss": 2.7151, "step": 171 }, { "epoch": 0.03242529927420115, "grad_norm": 0.84375, "learning_rate": 0.00086, "loss": 2.8332, "step": 172 }, { "epoch": 0.0326138184560279, "grad_norm": 0.72265625, "learning_rate": 0.000865, "loss": 2.8183, "step": 173 }, { "epoch": 0.03280233763785465, "grad_norm": 0.73046875, "learning_rate": 0.00087, "loss": 2.6777, "step": 174 }, { "epoch": 0.032990856819681405, "grad_norm": 0.7421875, "learning_rate": 0.000875, "loss": 2.6281, "step": 175 }, { "epoch": 0.03317937600150815, "grad_norm": 0.734375, "learning_rate": 0.00088, "loss": 2.7047, "step": 176 }, { "epoch": 0.033367895183334904, "grad_norm": 0.73046875, "learning_rate": 0.000885, "loss": 2.6637, "step": 177 }, { "epoch": 0.03355641436516166, "grad_norm": 0.7734375, "learning_rate": 0.0008900000000000001, "loss": 2.7817, "step": 178 }, { "epoch": 0.033744933546988404, "grad_norm": 0.7265625, "learning_rate": 0.0008950000000000001, "loss": 2.6216, "step": 179 }, { "epoch": 0.03393345272881516, "grad_norm": 0.734375, "learning_rate": 0.0009000000000000001, "loss": 2.6608, "step": 180 }, { "epoch": 0.03412197191064191, "grad_norm": 0.6953125, "learning_rate": 0.0009050000000000001, "loss": 2.712, "step": 181 }, { "epoch": 0.034310491092468656, "grad_norm": 0.71484375, "learning_rate": 0.00091, "loss": 2.6812, "step": 182 }, { "epoch": 0.03449901027429541, "grad_norm": 0.73828125, "learning_rate": 0.000915, "loss": 2.6181, "step": 183 }, { "epoch": 0.03468752945612216, "grad_norm": 0.77734375, "learning_rate": 0.00092, "loss": 2.5939, "step": 184 }, { "epoch": 0.03487604863794891, "grad_norm": 0.75, "learning_rate": 0.000925, "loss": 2.6378, "step": 185 }, { "epoch": 0.03506456781977566, "grad_norm": 0.78515625, "learning_rate": 0.00093, "loss": 2.658, "step": 186 }, { "epoch": 0.035253087001602415, "grad_norm": 0.7578125, "learning_rate": 0.0009350000000000001, "loss": 2.6324, "step": 187 }, { "epoch": 0.03544160618342916, "grad_norm": 0.73046875, "learning_rate": 0.00094, "loss": 2.7615, "step": 188 }, { "epoch": 0.035630125365255914, "grad_norm": 0.7421875, "learning_rate": 0.000945, "loss": 2.8334, "step": 189 }, { "epoch": 0.03581864454708267, "grad_norm": 0.7421875, "learning_rate": 0.00095, "loss": 2.8026, "step": 190 }, { "epoch": 0.036007163728909414, "grad_norm": 0.71484375, "learning_rate": 0.000955, "loss": 2.6532, "step": 191 }, { "epoch": 0.03619568291073617, "grad_norm": 0.73046875, "learning_rate": 0.00096, "loss": 2.5541, "step": 192 }, { "epoch": 0.03638420209256292, "grad_norm": 0.703125, "learning_rate": 0.000965, "loss": 2.6375, "step": 193 }, { "epoch": 0.036572721274389666, "grad_norm": 0.703125, "learning_rate": 0.0009699999999999999, "loss": 2.5705, "step": 194 }, { "epoch": 0.03676124045621642, "grad_norm": 0.73828125, "learning_rate": 0.000975, "loss": 2.6405, "step": 195 }, { "epoch": 0.03694975963804317, "grad_norm": 0.72265625, "learning_rate": 0.00098, "loss": 2.7821, "step": 196 }, { "epoch": 0.03713827881986992, "grad_norm": 0.69140625, "learning_rate": 0.000985, "loss": 2.6889, "step": 197 }, { "epoch": 0.03732679800169667, "grad_norm": 0.7578125, "learning_rate": 0.00099, "loss": 2.6658, "step": 198 }, { "epoch": 0.037515317183523425, "grad_norm": 0.72265625, "learning_rate": 0.000995, "loss": 2.6969, "step": 199 }, { "epoch": 0.03770383636535017, "grad_norm": 0.71484375, "learning_rate": 0.001, "loss": 2.5479, "step": 200 }, { "epoch": 0.037892355547176924, "grad_norm": 0.78125, "learning_rate": 0.0009998040752351098, "loss": 2.7177, "step": 201 }, { "epoch": 0.03808087472900368, "grad_norm": 0.76171875, "learning_rate": 0.0009996081504702195, "loss": 2.7224, "step": 202 }, { "epoch": 0.038269393910830424, "grad_norm": 0.6796875, "learning_rate": 0.0009994122257053293, "loss": 2.6316, "step": 203 }, { "epoch": 0.03845791309265718, "grad_norm": 0.76171875, "learning_rate": 0.0009992163009404388, "loss": 2.8178, "step": 204 }, { "epoch": 0.03864643227448393, "grad_norm": 0.73046875, "learning_rate": 0.0009990203761755486, "loss": 2.7619, "step": 205 }, { "epoch": 0.038834951456310676, "grad_norm": 0.671875, "learning_rate": 0.0009988244514106584, "loss": 2.5739, "step": 206 }, { "epoch": 0.03902347063813743, "grad_norm": 0.76171875, "learning_rate": 0.0009986285266457681, "loss": 2.7797, "step": 207 }, { "epoch": 0.03921198981996418, "grad_norm": 0.7734375, "learning_rate": 0.0009984326018808779, "loss": 2.695, "step": 208 }, { "epoch": 0.039400509001790936, "grad_norm": 0.69921875, "learning_rate": 0.0009982366771159876, "loss": 2.7551, "step": 209 }, { "epoch": 0.03958902818361768, "grad_norm": 0.73828125, "learning_rate": 0.0009980407523510972, "loss": 2.7898, "step": 210 }, { "epoch": 0.039777547365444435, "grad_norm": 0.765625, "learning_rate": 0.000997844827586207, "loss": 2.6824, "step": 211 }, { "epoch": 0.03996606654727119, "grad_norm": 0.703125, "learning_rate": 0.0009976489028213167, "loss": 2.8341, "step": 212 }, { "epoch": 0.040154585729097934, "grad_norm": 0.671875, "learning_rate": 0.0009974529780564262, "loss": 2.6885, "step": 213 }, { "epoch": 0.04034310491092469, "grad_norm": 0.70703125, "learning_rate": 0.000997257053291536, "loss": 2.5722, "step": 214 }, { "epoch": 0.04053162409275144, "grad_norm": 0.6875, "learning_rate": 0.0009970611285266457, "loss": 2.7023, "step": 215 }, { "epoch": 0.04072014327457819, "grad_norm": 0.734375, "learning_rate": 0.0009968652037617555, "loss": 2.6429, "step": 216 }, { "epoch": 0.04090866245640494, "grad_norm": 0.76171875, "learning_rate": 0.0009966692789968653, "loss": 2.7053, "step": 217 }, { "epoch": 0.04109718163823169, "grad_norm": 0.73046875, "learning_rate": 0.000996473354231975, "loss": 2.7841, "step": 218 }, { "epoch": 0.04128570082005844, "grad_norm": 0.6875, "learning_rate": 0.0009962774294670846, "loss": 2.6687, "step": 219 }, { "epoch": 0.04147422000188519, "grad_norm": 0.71875, "learning_rate": 0.0009960815047021943, "loss": 2.7893, "step": 220 }, { "epoch": 0.041662739183711946, "grad_norm": 0.703125, "learning_rate": 0.000995885579937304, "loss": 2.5992, "step": 221 }, { "epoch": 0.04185125836553869, "grad_norm": 0.71484375, "learning_rate": 0.0009956896551724138, "loss": 2.7238, "step": 222 }, { "epoch": 0.042039777547365445, "grad_norm": 0.71484375, "learning_rate": 0.0009954937304075236, "loss": 2.7477, "step": 223 }, { "epoch": 0.0422282967291922, "grad_norm": 0.73828125, "learning_rate": 0.0009952978056426334, "loss": 2.6079, "step": 224 }, { "epoch": 0.042416815911018944, "grad_norm": 0.71875, "learning_rate": 0.000995101880877743, "loss": 2.6389, "step": 225 }, { "epoch": 0.0426053350928457, "grad_norm": 0.703125, "learning_rate": 0.0009949059561128527, "loss": 2.6014, "step": 226 }, { "epoch": 0.04279385427467245, "grad_norm": 0.69921875, "learning_rate": 0.0009947100313479624, "loss": 2.6708, "step": 227 }, { "epoch": 0.0429823734564992, "grad_norm": 0.703125, "learning_rate": 0.0009945141065830722, "loss": 2.7032, "step": 228 }, { "epoch": 0.04317089263832595, "grad_norm": 0.71484375, "learning_rate": 0.0009943181818181817, "loss": 2.7911, "step": 229 }, { "epoch": 0.0433594118201527, "grad_norm": 0.734375, "learning_rate": 0.0009941222570532915, "loss": 2.5071, "step": 230 }, { "epoch": 0.04354793100197945, "grad_norm": 0.734375, "learning_rate": 0.0009939263322884012, "loss": 2.695, "step": 231 }, { "epoch": 0.0437364501838062, "grad_norm": 0.72265625, "learning_rate": 0.000993730407523511, "loss": 2.5969, "step": 232 }, { "epoch": 0.043924969365632956, "grad_norm": 0.6875, "learning_rate": 0.0009935344827586207, "loss": 2.6602, "step": 233 }, { "epoch": 0.0441134885474597, "grad_norm": 0.69140625, "learning_rate": 0.0009933385579937305, "loss": 2.6561, "step": 234 }, { "epoch": 0.044302007729286455, "grad_norm": 0.734375, "learning_rate": 0.00099314263322884, "loss": 2.6442, "step": 235 }, { "epoch": 0.04449052691111321, "grad_norm": 0.859375, "learning_rate": 0.0009929467084639498, "loss": 2.7465, "step": 236 }, { "epoch": 0.044679046092939954, "grad_norm": 0.76171875, "learning_rate": 0.0009927507836990596, "loss": 2.7102, "step": 237 }, { "epoch": 0.04486756527476671, "grad_norm": 0.69140625, "learning_rate": 0.0009925548589341693, "loss": 2.7074, "step": 238 }, { "epoch": 0.04505608445659346, "grad_norm": 0.8125, "learning_rate": 0.000992358934169279, "loss": 2.6626, "step": 239 }, { "epoch": 0.04524460363842021, "grad_norm": 0.671875, "learning_rate": 0.0009921630094043888, "loss": 2.5579, "step": 240 }, { "epoch": 0.04543312282024696, "grad_norm": 0.73828125, "learning_rate": 0.0009919670846394984, "loss": 2.7225, "step": 241 }, { "epoch": 0.04562164200207371, "grad_norm": 0.703125, "learning_rate": 0.0009917711598746081, "loss": 2.6952, "step": 242 }, { "epoch": 0.04581016118390046, "grad_norm": 0.76171875, "learning_rate": 0.000991575235109718, "loss": 2.6886, "step": 243 }, { "epoch": 0.04599868036572721, "grad_norm": 0.68359375, "learning_rate": 0.0009913793103448277, "loss": 2.6096, "step": 244 }, { "epoch": 0.046187199547553966, "grad_norm": 0.75390625, "learning_rate": 0.0009911833855799374, "loss": 2.7612, "step": 245 }, { "epoch": 0.04637571872938071, "grad_norm": 0.71484375, "learning_rate": 0.0009909874608150472, "loss": 2.6082, "step": 246 }, { "epoch": 0.046564237911207465, "grad_norm": 0.7734375, "learning_rate": 0.0009907915360501567, "loss": 2.7621, "step": 247 }, { "epoch": 0.04675275709303422, "grad_norm": 0.6953125, "learning_rate": 0.0009905956112852665, "loss": 2.6764, "step": 248 }, { "epoch": 0.046941276274860964, "grad_norm": 0.71484375, "learning_rate": 0.0009903996865203762, "loss": 2.6527, "step": 249 }, { "epoch": 0.04712979545668772, "grad_norm": 0.67578125, "learning_rate": 0.0009902037617554858, "loss": 2.5762, "step": 250 }, { "epoch": 0.04731831463851447, "grad_norm": 0.7421875, "learning_rate": 0.0009900078369905955, "loss": 2.7241, "step": 251 }, { "epoch": 0.04750683382034122, "grad_norm": 0.69921875, "learning_rate": 0.0009898119122257053, "loss": 2.6935, "step": 252 }, { "epoch": 0.04769535300216797, "grad_norm": 0.75390625, "learning_rate": 0.000989615987460815, "loss": 2.776, "step": 253 }, { "epoch": 0.04788387218399472, "grad_norm": 0.7109375, "learning_rate": 0.0009894200626959248, "loss": 2.7799, "step": 254 }, { "epoch": 0.04807239136582147, "grad_norm": 0.69921875, "learning_rate": 0.0009892241379310346, "loss": 2.7589, "step": 255 }, { "epoch": 0.04826091054764822, "grad_norm": 0.69921875, "learning_rate": 0.0009890282131661443, "loss": 2.646, "step": 256 }, { "epoch": 0.048449429729474976, "grad_norm": 0.6953125, "learning_rate": 0.0009888322884012539, "loss": 2.7226, "step": 257 }, { "epoch": 0.04863794891130172, "grad_norm": 0.66796875, "learning_rate": 0.0009886363636363636, "loss": 2.6825, "step": 258 }, { "epoch": 0.048826468093128475, "grad_norm": 0.671875, "learning_rate": 0.0009884404388714734, "loss": 2.6494, "step": 259 }, { "epoch": 0.04901498727495523, "grad_norm": 0.703125, "learning_rate": 0.0009882445141065831, "loss": 2.7586, "step": 260 }, { "epoch": 0.049203506456781974, "grad_norm": 0.73046875, "learning_rate": 0.000988048589341693, "loss": 2.7986, "step": 261 }, { "epoch": 0.04939202563860873, "grad_norm": 0.68359375, "learning_rate": 0.0009878526645768027, "loss": 2.624, "step": 262 }, { "epoch": 0.04958054482043548, "grad_norm": 0.66015625, "learning_rate": 0.0009876567398119122, "loss": 2.4967, "step": 263 }, { "epoch": 0.04976906400226223, "grad_norm": 0.69140625, "learning_rate": 0.000987460815047022, "loss": 2.5694, "step": 264 }, { "epoch": 0.04995758318408898, "grad_norm": 0.71484375, "learning_rate": 0.0009872648902821317, "loss": 2.7369, "step": 265 }, { "epoch": 0.05014610236591573, "grad_norm": 0.73828125, "learning_rate": 0.0009870689655172413, "loss": 2.641, "step": 266 }, { "epoch": 0.05033462154774248, "grad_norm": 0.6953125, "learning_rate": 0.000986873040752351, "loss": 2.5988, "step": 267 }, { "epoch": 0.05052314072956923, "grad_norm": 0.7734375, "learning_rate": 0.0009866771159874608, "loss": 2.6935, "step": 268 }, { "epoch": 0.050711659911395986, "grad_norm": 0.71875, "learning_rate": 0.0009864811912225705, "loss": 2.6573, "step": 269 }, { "epoch": 0.05090017909322274, "grad_norm": 0.71875, "learning_rate": 0.0009862852664576803, "loss": 2.5501, "step": 270 }, { "epoch": 0.051088698275049485, "grad_norm": 0.8515625, "learning_rate": 0.00098608934169279, "loss": 2.7173, "step": 271 }, { "epoch": 0.05127721745687624, "grad_norm": 0.765625, "learning_rate": 0.0009858934169278996, "loss": 2.7147, "step": 272 }, { "epoch": 0.05146573663870299, "grad_norm": 0.69921875, "learning_rate": 0.0009856974921630094, "loss": 2.6823, "step": 273 }, { "epoch": 0.05165425582052974, "grad_norm": 0.71875, "learning_rate": 0.0009855015673981191, "loss": 2.7399, "step": 274 }, { "epoch": 0.05184277500235649, "grad_norm": 0.7734375, "learning_rate": 0.0009853056426332289, "loss": 2.8052, "step": 275 }, { "epoch": 0.052031294184183244, "grad_norm": 0.78125, "learning_rate": 0.0009851097178683386, "loss": 2.6471, "step": 276 }, { "epoch": 0.05221981336600999, "grad_norm": 0.703125, "learning_rate": 0.0009849137931034484, "loss": 2.5997, "step": 277 }, { "epoch": 0.05240833254783674, "grad_norm": 0.671875, "learning_rate": 0.000984717868338558, "loss": 2.6933, "step": 278 }, { "epoch": 0.052596851729663496, "grad_norm": 0.71484375, "learning_rate": 0.0009845219435736677, "loss": 2.7849, "step": 279 }, { "epoch": 0.05278537091149024, "grad_norm": 0.76953125, "learning_rate": 0.0009843260188087774, "loss": 2.7277, "step": 280 }, { "epoch": 0.052973890093316996, "grad_norm": 0.671875, "learning_rate": 0.0009841300940438872, "loss": 2.7328, "step": 281 }, { "epoch": 0.05316240927514375, "grad_norm": 0.69921875, "learning_rate": 0.000983934169278997, "loss": 2.8041, "step": 282 }, { "epoch": 0.053350928456970495, "grad_norm": 0.6796875, "learning_rate": 0.0009837382445141067, "loss": 2.6497, "step": 283 }, { "epoch": 0.05353944763879725, "grad_norm": 0.7265625, "learning_rate": 0.0009835423197492165, "loss": 2.6852, "step": 284 }, { "epoch": 0.053727966820624, "grad_norm": 0.71484375, "learning_rate": 0.000983346394984326, "loss": 2.6116, "step": 285 }, { "epoch": 0.05391648600245075, "grad_norm": 0.72265625, "learning_rate": 0.0009831504702194358, "loss": 2.5864, "step": 286 }, { "epoch": 0.0541050051842775, "grad_norm": 0.68359375, "learning_rate": 0.0009829545454545455, "loss": 2.6291, "step": 287 }, { "epoch": 0.054293524366104254, "grad_norm": 0.69921875, "learning_rate": 0.000982758620689655, "loss": 2.672, "step": 288 }, { "epoch": 0.054482043547931, "grad_norm": 0.75390625, "learning_rate": 0.0009825626959247648, "loss": 2.6036, "step": 289 }, { "epoch": 0.05467056272975775, "grad_norm": 0.6484375, "learning_rate": 0.0009823667711598746, "loss": 2.4802, "step": 290 }, { "epoch": 0.054859081911584506, "grad_norm": 0.7578125, "learning_rate": 0.0009821708463949844, "loss": 2.721, "step": 291 }, { "epoch": 0.05504760109341125, "grad_norm": 0.703125, "learning_rate": 0.0009819749216300941, "loss": 2.6039, "step": 292 }, { "epoch": 0.055236120275238006, "grad_norm": 0.69140625, "learning_rate": 0.0009817789968652039, "loss": 2.7125, "step": 293 }, { "epoch": 0.05542463945706476, "grad_norm": 0.71875, "learning_rate": 0.0009815830721003134, "loss": 2.7176, "step": 294 }, { "epoch": 0.055613158638891505, "grad_norm": 0.73828125, "learning_rate": 0.0009813871473354232, "loss": 2.7061, "step": 295 }, { "epoch": 0.05580167782071826, "grad_norm": 0.69140625, "learning_rate": 0.000981191222570533, "loss": 2.7324, "step": 296 }, { "epoch": 0.05599019700254501, "grad_norm": 0.69140625, "learning_rate": 0.0009809952978056427, "loss": 2.6318, "step": 297 }, { "epoch": 0.05617871618437176, "grad_norm": 0.6796875, "learning_rate": 0.0009807993730407524, "loss": 2.637, "step": 298 }, { "epoch": 0.05636723536619851, "grad_norm": 0.6875, "learning_rate": 0.0009806034482758622, "loss": 2.5645, "step": 299 }, { "epoch": 0.056555754548025264, "grad_norm": 0.6953125, "learning_rate": 0.0009804075235109717, "loss": 2.7011, "step": 300 }, { "epoch": 0.05674427372985201, "grad_norm": 0.6796875, "learning_rate": 0.0009802115987460815, "loss": 2.7293, "step": 301 }, { "epoch": 0.05693279291167876, "grad_norm": 0.68359375, "learning_rate": 0.0009800156739811913, "loss": 2.5779, "step": 302 }, { "epoch": 0.057121312093505516, "grad_norm": 0.73046875, "learning_rate": 0.000979819749216301, "loss": 2.7574, "step": 303 }, { "epoch": 0.05730983127533226, "grad_norm": 0.69140625, "learning_rate": 0.0009796238244514106, "loss": 2.7168, "step": 304 }, { "epoch": 0.057498350457159016, "grad_norm": 0.6953125, "learning_rate": 0.0009794278996865203, "loss": 2.6531, "step": 305 }, { "epoch": 0.05768686963898577, "grad_norm": 0.67578125, "learning_rate": 0.00097923197492163, "loss": 2.6852, "step": 306 }, { "epoch": 0.057875388820812515, "grad_norm": 0.69921875, "learning_rate": 0.0009790360501567398, "loss": 2.8098, "step": 307 }, { "epoch": 0.05806390800263927, "grad_norm": 0.68359375, "learning_rate": 0.0009788401253918496, "loss": 2.5938, "step": 308 }, { "epoch": 0.05825242718446602, "grad_norm": 0.69921875, "learning_rate": 0.0009786442006269591, "loss": 2.6858, "step": 309 }, { "epoch": 0.05844094636629277, "grad_norm": 0.7265625, "learning_rate": 0.000978448275862069, "loss": 2.6455, "step": 310 }, { "epoch": 0.05862946554811952, "grad_norm": 0.72265625, "learning_rate": 0.0009782523510971787, "loss": 2.7194, "step": 311 }, { "epoch": 0.058817984729946274, "grad_norm": 0.71484375, "learning_rate": 0.0009780564263322884, "loss": 2.5933, "step": 312 }, { "epoch": 0.05900650391177302, "grad_norm": 0.68359375, "learning_rate": 0.0009778605015673982, "loss": 2.7103, "step": 313 }, { "epoch": 0.05919502309359977, "grad_norm": 0.73046875, "learning_rate": 0.000977664576802508, "loss": 2.7317, "step": 314 }, { "epoch": 0.059383542275426526, "grad_norm": 0.796875, "learning_rate": 0.0009774686520376177, "loss": 2.6629, "step": 315 }, { "epoch": 0.05957206145725327, "grad_norm": 0.71484375, "learning_rate": 0.0009772727272727272, "loss": 2.811, "step": 316 }, { "epoch": 0.059760580639080026, "grad_norm": 0.71875, "learning_rate": 0.000977076802507837, "loss": 2.679, "step": 317 }, { "epoch": 0.05994909982090678, "grad_norm": 0.70703125, "learning_rate": 0.0009768808777429468, "loss": 2.7421, "step": 318 }, { "epoch": 0.060137619002733525, "grad_norm": 0.77734375, "learning_rate": 0.0009766849529780565, "loss": 2.7717, "step": 319 }, { "epoch": 0.06032613818456028, "grad_norm": 0.74609375, "learning_rate": 0.0009764890282131662, "loss": 2.7456, "step": 320 }, { "epoch": 0.06051465736638703, "grad_norm": 0.73046875, "learning_rate": 0.0009762931034482759, "loss": 2.6342, "step": 321 }, { "epoch": 0.06070317654821378, "grad_norm": 0.80078125, "learning_rate": 0.0009760971786833856, "loss": 2.7088, "step": 322 }, { "epoch": 0.06089169573004053, "grad_norm": 0.72265625, "learning_rate": 0.0009759012539184952, "loss": 2.651, "step": 323 }, { "epoch": 0.061080214911867284, "grad_norm": 0.890625, "learning_rate": 0.000975705329153605, "loss": 2.7472, "step": 324 }, { "epoch": 0.06126873409369403, "grad_norm": 0.7421875, "learning_rate": 0.0009755094043887147, "loss": 2.8417, "step": 325 }, { "epoch": 0.06145725327552078, "grad_norm": 0.671875, "learning_rate": 0.0009753134796238245, "loss": 2.7298, "step": 326 }, { "epoch": 0.061645772457347536, "grad_norm": 0.7109375, "learning_rate": 0.0009751175548589341, "loss": 2.4934, "step": 327 }, { "epoch": 0.06183429163917429, "grad_norm": 0.76953125, "learning_rate": 0.0009749216300940439, "loss": 2.6646, "step": 328 }, { "epoch": 0.062022810821001036, "grad_norm": 0.75390625, "learning_rate": 0.0009747257053291537, "loss": 2.6718, "step": 329 }, { "epoch": 0.06221133000282779, "grad_norm": 0.66796875, "learning_rate": 0.0009745297805642633, "loss": 2.5689, "step": 330 }, { "epoch": 0.06239984918465454, "grad_norm": 0.66796875, "learning_rate": 0.0009743338557993731, "loss": 2.5188, "step": 331 }, { "epoch": 0.0625883683664813, "grad_norm": 0.7734375, "learning_rate": 0.0009741379310344828, "loss": 2.7064, "step": 332 }, { "epoch": 0.06277688754830804, "grad_norm": 0.73046875, "learning_rate": 0.0009739420062695925, "loss": 2.726, "step": 333 }, { "epoch": 0.06296540673013479, "grad_norm": 0.69140625, "learning_rate": 0.0009737460815047022, "loss": 2.7389, "step": 334 }, { "epoch": 0.06315392591196155, "grad_norm": 0.75390625, "learning_rate": 0.000973550156739812, "loss": 2.8134, "step": 335 }, { "epoch": 0.0633424450937883, "grad_norm": 0.73046875, "learning_rate": 0.0009733542319749216, "loss": 2.7394, "step": 336 }, { "epoch": 0.06353096427561504, "grad_norm": 0.66015625, "learning_rate": 0.0009731583072100314, "loss": 2.6256, "step": 337 }, { "epoch": 0.0637194834574418, "grad_norm": 0.76171875, "learning_rate": 0.0009729623824451412, "loss": 2.7413, "step": 338 }, { "epoch": 0.06390800263926855, "grad_norm": 0.7109375, "learning_rate": 0.0009727664576802508, "loss": 2.7725, "step": 339 }, { "epoch": 0.06409652182109529, "grad_norm": 0.69140625, "learning_rate": 0.0009725705329153606, "loss": 2.8092, "step": 340 }, { "epoch": 0.06428504100292205, "grad_norm": 0.703125, "learning_rate": 0.0009723746081504702, "loss": 2.7276, "step": 341 }, { "epoch": 0.0644735601847488, "grad_norm": 0.66015625, "learning_rate": 0.0009721786833855799, "loss": 2.5861, "step": 342 }, { "epoch": 0.06466207936657555, "grad_norm": 0.75390625, "learning_rate": 0.0009719827586206896, "loss": 2.6467, "step": 343 }, { "epoch": 0.0648505985484023, "grad_norm": 0.70703125, "learning_rate": 0.0009717868338557994, "loss": 2.7404, "step": 344 }, { "epoch": 0.06503911773022905, "grad_norm": 0.6875, "learning_rate": 0.000971590909090909, "loss": 2.6333, "step": 345 }, { "epoch": 0.0652276369120558, "grad_norm": 0.7265625, "learning_rate": 0.0009713949843260188, "loss": 2.6079, "step": 346 }, { "epoch": 0.06541615609388256, "grad_norm": 0.6796875, "learning_rate": 0.0009711990595611286, "loss": 2.5708, "step": 347 }, { "epoch": 0.0656046752757093, "grad_norm": 0.71484375, "learning_rate": 0.0009710031347962382, "loss": 2.6675, "step": 348 }, { "epoch": 0.06579319445753605, "grad_norm": 0.734375, "learning_rate": 0.000970807210031348, "loss": 2.7782, "step": 349 }, { "epoch": 0.06598171363936281, "grad_norm": 0.66015625, "learning_rate": 0.0009706112852664577, "loss": 2.6853, "step": 350 }, { "epoch": 0.06617023282118956, "grad_norm": 0.70703125, "learning_rate": 0.0009704153605015674, "loss": 2.7684, "step": 351 }, { "epoch": 0.0663587520030163, "grad_norm": 0.68359375, "learning_rate": 0.0009702194357366771, "loss": 2.5759, "step": 352 }, { "epoch": 0.06654727118484306, "grad_norm": 0.73828125, "learning_rate": 0.0009700235109717869, "loss": 2.7151, "step": 353 }, { "epoch": 0.06673579036666981, "grad_norm": 0.7265625, "learning_rate": 0.0009698275862068966, "loss": 2.6346, "step": 354 }, { "epoch": 0.06692430954849656, "grad_norm": 0.70703125, "learning_rate": 0.0009696316614420063, "loss": 2.5878, "step": 355 }, { "epoch": 0.06711282873032332, "grad_norm": 0.71875, "learning_rate": 0.0009694357366771161, "loss": 2.6841, "step": 356 }, { "epoch": 0.06730134791215006, "grad_norm": 0.72265625, "learning_rate": 0.0009692398119122258, "loss": 2.5688, "step": 357 }, { "epoch": 0.06748986709397681, "grad_norm": 0.6875, "learning_rate": 0.0009690438871473355, "loss": 2.5057, "step": 358 }, { "epoch": 0.06767838627580357, "grad_norm": 0.74609375, "learning_rate": 0.0009688479623824452, "loss": 2.6444, "step": 359 }, { "epoch": 0.06786690545763031, "grad_norm": 0.73046875, "learning_rate": 0.0009686520376175549, "loss": 2.6894, "step": 360 }, { "epoch": 0.06805542463945706, "grad_norm": 0.65234375, "learning_rate": 0.0009684561128526645, "loss": 2.5921, "step": 361 }, { "epoch": 0.06824394382128382, "grad_norm": 0.76171875, "learning_rate": 0.0009682601880877743, "loss": 2.7547, "step": 362 }, { "epoch": 0.06843246300311057, "grad_norm": 0.8125, "learning_rate": 0.000968064263322884, "loss": 2.7235, "step": 363 }, { "epoch": 0.06862098218493731, "grad_norm": 0.69921875, "learning_rate": 0.0009678683385579937, "loss": 2.6726, "step": 364 }, { "epoch": 0.06880950136676407, "grad_norm": 0.77734375, "learning_rate": 0.0009676724137931034, "loss": 2.7688, "step": 365 }, { "epoch": 0.06899802054859082, "grad_norm": 0.6953125, "learning_rate": 0.0009674764890282132, "loss": 2.6567, "step": 366 }, { "epoch": 0.06918653973041756, "grad_norm": 0.69921875, "learning_rate": 0.0009672805642633229, "loss": 2.7241, "step": 367 }, { "epoch": 0.06937505891224433, "grad_norm": 0.703125, "learning_rate": 0.0009670846394984326, "loss": 2.603, "step": 368 }, { "epoch": 0.06956357809407107, "grad_norm": 0.6796875, "learning_rate": 0.0009668887147335424, "loss": 2.6863, "step": 369 }, { "epoch": 0.06975209727589782, "grad_norm": 0.6796875, "learning_rate": 0.000966692789968652, "loss": 2.6655, "step": 370 }, { "epoch": 0.06994061645772458, "grad_norm": 0.66015625, "learning_rate": 0.0009664968652037618, "loss": 2.5301, "step": 371 }, { "epoch": 0.07012913563955132, "grad_norm": 0.71484375, "learning_rate": 0.0009663009404388715, "loss": 2.7405, "step": 372 }, { "epoch": 0.07031765482137807, "grad_norm": 0.671875, "learning_rate": 0.0009661050156739812, "loss": 2.7326, "step": 373 }, { "epoch": 0.07050617400320483, "grad_norm": 0.7265625, "learning_rate": 0.000965909090909091, "loss": 2.6685, "step": 374 }, { "epoch": 0.07069469318503158, "grad_norm": 0.71484375, "learning_rate": 0.0009657131661442007, "loss": 2.6664, "step": 375 }, { "epoch": 0.07088321236685832, "grad_norm": 0.7109375, "learning_rate": 0.0009655172413793104, "loss": 2.5892, "step": 376 }, { "epoch": 0.07107173154868508, "grad_norm": 0.7109375, "learning_rate": 0.0009653213166144201, "loss": 2.6351, "step": 377 }, { "epoch": 0.07126025073051183, "grad_norm": 0.75390625, "learning_rate": 0.0009651253918495299, "loss": 2.6587, "step": 378 }, { "epoch": 0.07144876991233857, "grad_norm": 0.69140625, "learning_rate": 0.0009649294670846394, "loss": 2.7744, "step": 379 }, { "epoch": 0.07163728909416534, "grad_norm": 0.69921875, "learning_rate": 0.0009647335423197492, "loss": 2.7516, "step": 380 }, { "epoch": 0.07182580827599208, "grad_norm": 0.7734375, "learning_rate": 0.0009645376175548589, "loss": 2.6607, "step": 381 }, { "epoch": 0.07201432745781883, "grad_norm": 0.73046875, "learning_rate": 0.0009643416927899687, "loss": 2.7513, "step": 382 }, { "epoch": 0.07220284663964559, "grad_norm": 0.65625, "learning_rate": 0.0009641457680250783, "loss": 2.607, "step": 383 }, { "epoch": 0.07239136582147233, "grad_norm": 0.6875, "learning_rate": 0.0009639498432601881, "loss": 2.5463, "step": 384 }, { "epoch": 0.07257988500329908, "grad_norm": 0.7421875, "learning_rate": 0.0009637539184952979, "loss": 2.6368, "step": 385 }, { "epoch": 0.07276840418512584, "grad_norm": 0.734375, "learning_rate": 0.0009635579937304075, "loss": 2.5846, "step": 386 }, { "epoch": 0.07295692336695259, "grad_norm": 0.73046875, "learning_rate": 0.0009633620689655173, "loss": 2.7072, "step": 387 }, { "epoch": 0.07314544254877933, "grad_norm": 0.6953125, "learning_rate": 0.000963166144200627, "loss": 2.6918, "step": 388 }, { "epoch": 0.07333396173060609, "grad_norm": 0.69140625, "learning_rate": 0.0009629702194357367, "loss": 2.6682, "step": 389 }, { "epoch": 0.07352248091243284, "grad_norm": 0.6953125, "learning_rate": 0.0009627742946708464, "loss": 2.6512, "step": 390 }, { "epoch": 0.07371100009425958, "grad_norm": 0.796875, "learning_rate": 0.0009625783699059562, "loss": 2.718, "step": 391 }, { "epoch": 0.07389951927608635, "grad_norm": 0.79296875, "learning_rate": 0.0009623824451410658, "loss": 2.7208, "step": 392 }, { "epoch": 0.07408803845791309, "grad_norm": 0.734375, "learning_rate": 0.0009621865203761756, "loss": 2.7411, "step": 393 }, { "epoch": 0.07427655763973984, "grad_norm": 0.69921875, "learning_rate": 0.0009619905956112854, "loss": 2.6763, "step": 394 }, { "epoch": 0.0744650768215666, "grad_norm": 0.703125, "learning_rate": 0.000961794670846395, "loss": 2.6919, "step": 395 }, { "epoch": 0.07465359600339334, "grad_norm": 0.7421875, "learning_rate": 0.0009615987460815048, "loss": 2.767, "step": 396 }, { "epoch": 0.07484211518522009, "grad_norm": 0.66796875, "learning_rate": 0.0009614028213166145, "loss": 2.6868, "step": 397 }, { "epoch": 0.07503063436704685, "grad_norm": 0.67578125, "learning_rate": 0.0009612068965517241, "loss": 2.6393, "step": 398 }, { "epoch": 0.0752191535488736, "grad_norm": 0.7265625, "learning_rate": 0.0009610109717868338, "loss": 2.5917, "step": 399 }, { "epoch": 0.07540767273070034, "grad_norm": 0.70703125, "learning_rate": 0.0009608150470219436, "loss": 2.709, "step": 400 }, { "epoch": 0.0755961919125271, "grad_norm": 0.70703125, "learning_rate": 0.0009606191222570532, "loss": 2.6591, "step": 401 }, { "epoch": 0.07578471109435385, "grad_norm": 0.70703125, "learning_rate": 0.000960423197492163, "loss": 2.7638, "step": 402 }, { "epoch": 0.0759732302761806, "grad_norm": 0.67578125, "learning_rate": 0.0009602272727272728, "loss": 2.58, "step": 403 }, { "epoch": 0.07616174945800736, "grad_norm": 0.7109375, "learning_rate": 0.0009600313479623824, "loss": 2.5257, "step": 404 }, { "epoch": 0.0763502686398341, "grad_norm": 0.75390625, "learning_rate": 0.0009598354231974922, "loss": 2.6512, "step": 405 }, { "epoch": 0.07653878782166085, "grad_norm": 0.67578125, "learning_rate": 0.0009596394984326019, "loss": 2.6432, "step": 406 }, { "epoch": 0.07672730700348761, "grad_norm": 0.65625, "learning_rate": 0.0009594435736677116, "loss": 2.6028, "step": 407 }, { "epoch": 0.07691582618531435, "grad_norm": 0.87890625, "learning_rate": 0.0009592476489028213, "loss": 2.707, "step": 408 }, { "epoch": 0.0771043453671411, "grad_norm": 0.75390625, "learning_rate": 0.0009590517241379311, "loss": 2.5831, "step": 409 }, { "epoch": 0.07729286454896786, "grad_norm": 0.72265625, "learning_rate": 0.0009588557993730408, "loss": 2.7447, "step": 410 }, { "epoch": 0.0774813837307946, "grad_norm": 0.65234375, "learning_rate": 0.0009586598746081505, "loss": 2.6452, "step": 411 }, { "epoch": 0.07766990291262135, "grad_norm": 0.75, "learning_rate": 0.0009584639498432603, "loss": 2.7138, "step": 412 }, { "epoch": 0.07785842209444811, "grad_norm": 0.76953125, "learning_rate": 0.00095826802507837, "loss": 2.5726, "step": 413 }, { "epoch": 0.07804694127627486, "grad_norm": 0.71484375, "learning_rate": 0.0009580721003134797, "loss": 2.7128, "step": 414 }, { "epoch": 0.0782354604581016, "grad_norm": 0.6640625, "learning_rate": 0.0009578761755485894, "loss": 2.5482, "step": 415 }, { "epoch": 0.07842397963992837, "grad_norm": 0.7421875, "learning_rate": 0.0009576802507836991, "loss": 2.768, "step": 416 }, { "epoch": 0.07861249882175511, "grad_norm": 0.7265625, "learning_rate": 0.0009574843260188087, "loss": 2.774, "step": 417 }, { "epoch": 0.07880101800358187, "grad_norm": 0.67578125, "learning_rate": 0.0009572884012539185, "loss": 2.7388, "step": 418 }, { "epoch": 0.07898953718540862, "grad_norm": 0.63671875, "learning_rate": 0.0009570924764890282, "loss": 2.5905, "step": 419 }, { "epoch": 0.07917805636723536, "grad_norm": 0.73046875, "learning_rate": 0.0009568965517241379, "loss": 2.7478, "step": 420 }, { "epoch": 0.07936657554906212, "grad_norm": 0.6953125, "learning_rate": 0.0009567006269592476, "loss": 2.699, "step": 421 }, { "epoch": 0.07955509473088887, "grad_norm": 0.66796875, "learning_rate": 0.0009565047021943574, "loss": 2.5614, "step": 422 }, { "epoch": 0.07974361391271562, "grad_norm": 0.69921875, "learning_rate": 0.0009563087774294671, "loss": 2.6036, "step": 423 }, { "epoch": 0.07993213309454238, "grad_norm": 0.69921875, "learning_rate": 0.0009561128526645768, "loss": 2.6515, "step": 424 }, { "epoch": 0.08012065227636912, "grad_norm": 0.69921875, "learning_rate": 0.0009559169278996866, "loss": 2.7026, "step": 425 }, { "epoch": 0.08030917145819587, "grad_norm": 0.68359375, "learning_rate": 0.0009557210031347962, "loss": 2.7497, "step": 426 }, { "epoch": 0.08049769064002263, "grad_norm": 0.7109375, "learning_rate": 0.000955525078369906, "loss": 2.6129, "step": 427 }, { "epoch": 0.08068620982184938, "grad_norm": 0.70703125, "learning_rate": 0.0009553291536050157, "loss": 2.6113, "step": 428 }, { "epoch": 0.08087472900367612, "grad_norm": 0.734375, "learning_rate": 0.0009551332288401254, "loss": 2.4547, "step": 429 }, { "epoch": 0.08106324818550288, "grad_norm": 0.671875, "learning_rate": 0.0009549373040752351, "loss": 2.6197, "step": 430 }, { "epoch": 0.08125176736732963, "grad_norm": 0.66796875, "learning_rate": 0.0009547413793103449, "loss": 2.684, "step": 431 }, { "epoch": 0.08144028654915637, "grad_norm": 0.71875, "learning_rate": 0.0009545454545454546, "loss": 2.6874, "step": 432 }, { "epoch": 0.08162880573098313, "grad_norm": 0.734375, "learning_rate": 0.0009543495297805643, "loss": 2.6019, "step": 433 }, { "epoch": 0.08181732491280988, "grad_norm": 0.6796875, "learning_rate": 0.0009541536050156741, "loss": 2.6309, "step": 434 }, { "epoch": 0.08200584409463663, "grad_norm": 0.7734375, "learning_rate": 0.0009539576802507836, "loss": 2.6848, "step": 435 }, { "epoch": 0.08219436327646339, "grad_norm": 0.72265625, "learning_rate": 0.0009537617554858934, "loss": 2.7124, "step": 436 }, { "epoch": 0.08238288245829013, "grad_norm": 0.66015625, "learning_rate": 0.0009535658307210031, "loss": 2.5744, "step": 437 }, { "epoch": 0.08257140164011688, "grad_norm": 0.75, "learning_rate": 0.0009533699059561129, "loss": 2.7689, "step": 438 }, { "epoch": 0.08275992082194364, "grad_norm": 0.7265625, "learning_rate": 0.0009531739811912225, "loss": 2.7709, "step": 439 }, { "epoch": 0.08294844000377039, "grad_norm": 0.71484375, "learning_rate": 0.0009529780564263323, "loss": 2.5495, "step": 440 }, { "epoch": 0.08313695918559713, "grad_norm": 0.66796875, "learning_rate": 0.0009527821316614421, "loss": 2.696, "step": 441 }, { "epoch": 0.08332547836742389, "grad_norm": 0.68359375, "learning_rate": 0.0009525862068965517, "loss": 2.6657, "step": 442 }, { "epoch": 0.08351399754925064, "grad_norm": 0.67578125, "learning_rate": 0.0009523902821316615, "loss": 2.6998, "step": 443 }, { "epoch": 0.08370251673107738, "grad_norm": 0.6875, "learning_rate": 0.0009521943573667712, "loss": 2.7154, "step": 444 }, { "epoch": 0.08389103591290414, "grad_norm": 0.68359375, "learning_rate": 0.0009519984326018809, "loss": 2.6478, "step": 445 }, { "epoch": 0.08407955509473089, "grad_norm": 0.671875, "learning_rate": 0.0009518025078369906, "loss": 2.6899, "step": 446 }, { "epoch": 0.08426807427655764, "grad_norm": 0.6640625, "learning_rate": 0.0009516065830721004, "loss": 2.7137, "step": 447 }, { "epoch": 0.0844565934583844, "grad_norm": 0.6875, "learning_rate": 0.00095141065830721, "loss": 2.6207, "step": 448 }, { "epoch": 0.08464511264021114, "grad_norm": 0.73046875, "learning_rate": 0.0009512147335423198, "loss": 2.7149, "step": 449 }, { "epoch": 0.08483363182203789, "grad_norm": 0.6796875, "learning_rate": 0.0009510188087774296, "loss": 2.7011, "step": 450 }, { "epoch": 0.08502215100386465, "grad_norm": 0.6484375, "learning_rate": 0.0009508228840125392, "loss": 2.6496, "step": 451 }, { "epoch": 0.0852106701856914, "grad_norm": 0.70703125, "learning_rate": 0.000950626959247649, "loss": 2.6714, "step": 452 }, { "epoch": 0.08539918936751814, "grad_norm": 0.6484375, "learning_rate": 0.0009504310344827587, "loss": 2.6271, "step": 453 }, { "epoch": 0.0855877085493449, "grad_norm": 0.7109375, "learning_rate": 0.0009502351097178683, "loss": 2.6513, "step": 454 }, { "epoch": 0.08577622773117165, "grad_norm": 0.671875, "learning_rate": 0.000950039184952978, "loss": 2.6638, "step": 455 }, { "epoch": 0.0859647469129984, "grad_norm": 0.6953125, "learning_rate": 0.0009498432601880878, "loss": 2.7398, "step": 456 }, { "epoch": 0.08615326609482515, "grad_norm": 0.69921875, "learning_rate": 0.0009496473354231974, "loss": 2.7013, "step": 457 }, { "epoch": 0.0863417852766519, "grad_norm": 0.69140625, "learning_rate": 0.0009494514106583072, "loss": 2.6336, "step": 458 }, { "epoch": 0.08653030445847865, "grad_norm": 0.65625, "learning_rate": 0.000949255485893417, "loss": 2.5915, "step": 459 }, { "epoch": 0.0867188236403054, "grad_norm": 0.6796875, "learning_rate": 0.0009490595611285266, "loss": 2.6545, "step": 460 }, { "epoch": 0.08690734282213215, "grad_norm": 0.73046875, "learning_rate": 0.0009488636363636364, "loss": 2.6792, "step": 461 }, { "epoch": 0.0870958620039589, "grad_norm": 0.66796875, "learning_rate": 0.0009486677115987461, "loss": 2.6238, "step": 462 }, { "epoch": 0.08728438118578566, "grad_norm": 0.70703125, "learning_rate": 0.0009484717868338558, "loss": 2.6929, "step": 463 }, { "epoch": 0.0874729003676124, "grad_norm": 0.66796875, "learning_rate": 0.0009482758620689655, "loss": 2.7269, "step": 464 }, { "epoch": 0.08766141954943915, "grad_norm": 0.6875, "learning_rate": 0.0009480799373040753, "loss": 2.6728, "step": 465 }, { "epoch": 0.08784993873126591, "grad_norm": 0.87890625, "learning_rate": 0.000947884012539185, "loss": 2.667, "step": 466 }, { "epoch": 0.08803845791309266, "grad_norm": 0.75390625, "learning_rate": 0.0009476880877742947, "loss": 2.7706, "step": 467 }, { "epoch": 0.0882269770949194, "grad_norm": 0.65234375, "learning_rate": 0.0009474921630094045, "loss": 2.7464, "step": 468 }, { "epoch": 0.08841549627674616, "grad_norm": 0.76953125, "learning_rate": 0.0009472962382445142, "loss": 2.6004, "step": 469 }, { "epoch": 0.08860401545857291, "grad_norm": 0.671875, "learning_rate": 0.0009471003134796239, "loss": 2.6237, "step": 470 }, { "epoch": 0.08879253464039966, "grad_norm": 0.69140625, "learning_rate": 0.0009469043887147336, "loss": 2.6628, "step": 471 }, { "epoch": 0.08898105382222642, "grad_norm": 0.734375, "learning_rate": 0.0009467084639498434, "loss": 2.7066, "step": 472 }, { "epoch": 0.08916957300405316, "grad_norm": 0.73046875, "learning_rate": 0.0009465125391849529, "loss": 2.6655, "step": 473 }, { "epoch": 0.08935809218587991, "grad_norm": 0.7265625, "learning_rate": 0.0009463166144200627, "loss": 2.6333, "step": 474 }, { "epoch": 0.08954661136770667, "grad_norm": 0.72265625, "learning_rate": 0.0009461206896551724, "loss": 2.5766, "step": 475 }, { "epoch": 0.08973513054953342, "grad_norm": 0.7734375, "learning_rate": 0.0009459247648902821, "loss": 2.7387, "step": 476 }, { "epoch": 0.08992364973136016, "grad_norm": 0.7890625, "learning_rate": 0.0009457288401253918, "loss": 2.7342, "step": 477 }, { "epoch": 0.09011216891318692, "grad_norm": 0.75, "learning_rate": 0.0009455329153605016, "loss": 2.6416, "step": 478 }, { "epoch": 0.09030068809501367, "grad_norm": 0.70703125, "learning_rate": 0.0009453369905956113, "loss": 2.6143, "step": 479 }, { "epoch": 0.09048920727684041, "grad_norm": 0.73046875, "learning_rate": 0.000945141065830721, "loss": 2.7185, "step": 480 }, { "epoch": 0.09067772645866717, "grad_norm": 0.8359375, "learning_rate": 0.0009449451410658308, "loss": 2.6152, "step": 481 }, { "epoch": 0.09086624564049392, "grad_norm": 0.6875, "learning_rate": 0.0009447492163009404, "loss": 2.6592, "step": 482 }, { "epoch": 0.09105476482232067, "grad_norm": 0.6875, "learning_rate": 0.0009445532915360502, "loss": 2.5181, "step": 483 }, { "epoch": 0.09124328400414743, "grad_norm": 0.6953125, "learning_rate": 0.0009443573667711599, "loss": 2.6332, "step": 484 }, { "epoch": 0.09143180318597417, "grad_norm": 0.73828125, "learning_rate": 0.0009441614420062696, "loss": 2.521, "step": 485 }, { "epoch": 0.09162032236780092, "grad_norm": 0.71484375, "learning_rate": 0.0009439655172413793, "loss": 2.6339, "step": 486 }, { "epoch": 0.09180884154962768, "grad_norm": 0.70703125, "learning_rate": 0.0009437695924764891, "loss": 2.6627, "step": 487 }, { "epoch": 0.09199736073145443, "grad_norm": 0.703125, "learning_rate": 0.0009435736677115988, "loss": 2.6227, "step": 488 }, { "epoch": 0.09218587991328117, "grad_norm": 0.734375, "learning_rate": 0.0009433777429467085, "loss": 2.784, "step": 489 }, { "epoch": 0.09237439909510793, "grad_norm": 2.09375, "learning_rate": 0.0009431818181818183, "loss": 2.5622, "step": 490 }, { "epoch": 0.09256291827693468, "grad_norm": 0.7265625, "learning_rate": 0.0009429858934169278, "loss": 2.6712, "step": 491 }, { "epoch": 0.09275143745876142, "grad_norm": 0.71875, "learning_rate": 0.0009427899686520376, "loss": 2.5781, "step": 492 }, { "epoch": 0.09293995664058818, "grad_norm": 0.7734375, "learning_rate": 0.0009425940438871473, "loss": 2.6193, "step": 493 }, { "epoch": 0.09312847582241493, "grad_norm": 0.703125, "learning_rate": 0.0009423981191222571, "loss": 2.716, "step": 494 }, { "epoch": 0.09331699500424168, "grad_norm": 0.73828125, "learning_rate": 0.0009422021943573667, "loss": 2.745, "step": 495 }, { "epoch": 0.09350551418606844, "grad_norm": 0.67578125, "learning_rate": 0.0009420062695924765, "loss": 2.5251, "step": 496 }, { "epoch": 0.09369403336789518, "grad_norm": 0.69921875, "learning_rate": 0.0009418103448275863, "loss": 2.7023, "step": 497 }, { "epoch": 0.09388255254972193, "grad_norm": 0.6953125, "learning_rate": 0.0009416144200626959, "loss": 2.7697, "step": 498 }, { "epoch": 0.09407107173154869, "grad_norm": 0.72265625, "learning_rate": 0.0009414184952978057, "loss": 2.6253, "step": 499 }, { "epoch": 0.09425959091337544, "grad_norm": 0.640625, "learning_rate": 0.0009412225705329154, "loss": 2.6668, "step": 500 }, { "epoch": 0.09425959091337544, "eval_runtime": 58.5785, "eval_samples_per_second": 17.481, "eval_steps_per_second": 0.546, "step": 500 }, { "epoch": 0.09425959091337544, "eval/hellaswag_acc": 0.37572196773551086, "eval/hellaswag_acc_norm": 0.4714200358494324, "eval_hellaswag_elapsed_time": 195.95180106163025, "step": 500 }, { "epoch": 0.09444811009520218, "grad_norm": 0.67578125, "learning_rate": 0.0009410266457680251, "loss": 2.6645, "step": 501 }, { "epoch": 0.09463662927702894, "grad_norm": 0.77734375, "learning_rate": 0.0009408307210031348, "loss": 2.7233, "step": 502 }, { "epoch": 0.09482514845885569, "grad_norm": 0.68359375, "learning_rate": 0.0009406347962382446, "loss": 2.6959, "step": 503 }, { "epoch": 0.09501366764068243, "grad_norm": 0.6796875, "learning_rate": 0.0009404388714733542, "loss": 2.747, "step": 504 }, { "epoch": 0.0952021868225092, "grad_norm": 0.63671875, "learning_rate": 0.000940242946708464, "loss": 2.521, "step": 505 }, { "epoch": 0.09539070600433594, "grad_norm": 0.69140625, "learning_rate": 0.0009400470219435738, "loss": 2.7368, "step": 506 }, { "epoch": 0.09557922518616269, "grad_norm": 0.7109375, "learning_rate": 0.0009398510971786834, "loss": 2.6509, "step": 507 }, { "epoch": 0.09576774436798945, "grad_norm": 0.73046875, "learning_rate": 0.0009396551724137932, "loss": 2.785, "step": 508 }, { "epoch": 0.09595626354981619, "grad_norm": 0.66796875, "learning_rate": 0.0009394592476489029, "loss": 2.5647, "step": 509 }, { "epoch": 0.09614478273164294, "grad_norm": 0.70703125, "learning_rate": 0.0009392633228840125, "loss": 2.6087, "step": 510 }, { "epoch": 0.0963333019134697, "grad_norm": 0.67578125, "learning_rate": 0.0009390673981191222, "loss": 2.6032, "step": 511 }, { "epoch": 0.09652182109529645, "grad_norm": 0.703125, "learning_rate": 0.000938871473354232, "loss": 2.6934, "step": 512 }, { "epoch": 0.09671034027712319, "grad_norm": 0.6953125, "learning_rate": 0.0009386755485893416, "loss": 2.7077, "step": 513 }, { "epoch": 0.09689885945894995, "grad_norm": 0.71875, "learning_rate": 0.0009384796238244514, "loss": 2.7372, "step": 514 }, { "epoch": 0.0970873786407767, "grad_norm": 0.69921875, "learning_rate": 0.0009382836990595611, "loss": 2.5907, "step": 515 }, { "epoch": 0.09727589782260344, "grad_norm": 0.6484375, "learning_rate": 0.0009380877742946708, "loss": 2.5623, "step": 516 }, { "epoch": 0.0974644170044302, "grad_norm": 0.703125, "learning_rate": 0.0009378918495297806, "loss": 2.6949, "step": 517 }, { "epoch": 0.09765293618625695, "grad_norm": 0.78515625, "learning_rate": 0.0009376959247648903, "loss": 2.6505, "step": 518 }, { "epoch": 0.0978414553680837, "grad_norm": 0.72265625, "learning_rate": 0.0009375, "loss": 2.6902, "step": 519 }, { "epoch": 0.09802997454991046, "grad_norm": 0.71484375, "learning_rate": 0.0009373040752351097, "loss": 2.6529, "step": 520 }, { "epoch": 0.0982184937317372, "grad_norm": 0.69921875, "learning_rate": 0.0009371081504702195, "loss": 2.5571, "step": 521 }, { "epoch": 0.09840701291356395, "grad_norm": 0.71875, "learning_rate": 0.0009369122257053292, "loss": 2.7718, "step": 522 }, { "epoch": 0.09859553209539071, "grad_norm": 0.734375, "learning_rate": 0.0009367163009404389, "loss": 2.7112, "step": 523 }, { "epoch": 0.09878405127721746, "grad_norm": 0.6875, "learning_rate": 0.0009365203761755486, "loss": 2.5318, "step": 524 }, { "epoch": 0.0989725704590442, "grad_norm": 0.66796875, "learning_rate": 0.0009363244514106584, "loss": 2.6242, "step": 525 }, { "epoch": 0.09916108964087096, "grad_norm": 0.6875, "learning_rate": 0.0009361285266457681, "loss": 2.6603, "step": 526 }, { "epoch": 0.09934960882269771, "grad_norm": 0.7109375, "learning_rate": 0.0009359326018808778, "loss": 2.7204, "step": 527 }, { "epoch": 0.09953812800452445, "grad_norm": 0.68359375, "learning_rate": 0.0009357366771159876, "loss": 2.6355, "step": 528 }, { "epoch": 0.09972664718635121, "grad_norm": 0.69140625, "learning_rate": 0.0009355407523510971, "loss": 2.6126, "step": 529 }, { "epoch": 0.09991516636817796, "grad_norm": 0.6484375, "learning_rate": 0.0009353448275862069, "loss": 2.5042, "step": 530 }, { "epoch": 0.1001036855500047, "grad_norm": 0.6953125, "learning_rate": 0.0009351489028213166, "loss": 2.639, "step": 531 }, { "epoch": 0.10029220473183147, "grad_norm": 0.69140625, "learning_rate": 0.0009349529780564263, "loss": 2.6981, "step": 532 }, { "epoch": 0.10048072391365821, "grad_norm": 0.66796875, "learning_rate": 0.000934757053291536, "loss": 2.6578, "step": 533 }, { "epoch": 0.10066924309548496, "grad_norm": 0.734375, "learning_rate": 0.0009345611285266458, "loss": 2.7651, "step": 534 }, { "epoch": 0.10085776227731172, "grad_norm": 0.69140625, "learning_rate": 0.0009343652037617555, "loss": 2.6639, "step": 535 }, { "epoch": 0.10104628145913847, "grad_norm": 0.70703125, "learning_rate": 0.0009341692789968652, "loss": 2.6911, "step": 536 }, { "epoch": 0.10123480064096523, "grad_norm": 0.69921875, "learning_rate": 0.000933973354231975, "loss": 2.6213, "step": 537 }, { "epoch": 0.10142331982279197, "grad_norm": 0.7109375, "learning_rate": 0.0009337774294670846, "loss": 2.6084, "step": 538 }, { "epoch": 0.10161183900461872, "grad_norm": 0.72265625, "learning_rate": 0.0009335815047021944, "loss": 2.6893, "step": 539 }, { "epoch": 0.10180035818644548, "grad_norm": 0.65625, "learning_rate": 0.0009333855799373041, "loss": 2.547, "step": 540 }, { "epoch": 0.10198887736827222, "grad_norm": 0.72265625, "learning_rate": 0.0009331896551724138, "loss": 2.7084, "step": 541 }, { "epoch": 0.10217739655009897, "grad_norm": 0.65625, "learning_rate": 0.0009329937304075235, "loss": 2.6611, "step": 542 }, { "epoch": 0.10236591573192573, "grad_norm": 0.703125, "learning_rate": 0.0009327978056426333, "loss": 2.658, "step": 543 }, { "epoch": 0.10255443491375248, "grad_norm": 0.70703125, "learning_rate": 0.000932601880877743, "loss": 2.7423, "step": 544 }, { "epoch": 0.10274295409557922, "grad_norm": 0.63671875, "learning_rate": 0.0009324059561128527, "loss": 2.6237, "step": 545 }, { "epoch": 0.10293147327740598, "grad_norm": 0.671875, "learning_rate": 0.0009322100313479625, "loss": 2.6846, "step": 546 }, { "epoch": 0.10311999245923273, "grad_norm": 0.66015625, "learning_rate": 0.0009320141065830722, "loss": 2.5963, "step": 547 }, { "epoch": 0.10330851164105948, "grad_norm": 0.7265625, "learning_rate": 0.0009318181818181818, "loss": 2.6334, "step": 548 }, { "epoch": 0.10349703082288624, "grad_norm": 0.69921875, "learning_rate": 0.0009316222570532915, "loss": 2.6657, "step": 549 }, { "epoch": 0.10368555000471298, "grad_norm": 0.671875, "learning_rate": 0.0009314263322884013, "loss": 2.7307, "step": 550 }, { "epoch": 0.10387406918653973, "grad_norm": 0.67578125, "learning_rate": 0.0009312304075235109, "loss": 2.7773, "step": 551 }, { "epoch": 0.10406258836836649, "grad_norm": 0.7109375, "learning_rate": 0.0009310344827586207, "loss": 2.7054, "step": 552 }, { "epoch": 0.10425110755019323, "grad_norm": 0.68359375, "learning_rate": 0.0009308385579937305, "loss": 2.6325, "step": 553 }, { "epoch": 0.10443962673201998, "grad_norm": 0.65625, "learning_rate": 0.0009306426332288401, "loss": 2.5011, "step": 554 }, { "epoch": 0.10462814591384674, "grad_norm": 0.67578125, "learning_rate": 0.0009304467084639499, "loss": 2.6252, "step": 555 }, { "epoch": 0.10481666509567349, "grad_norm": 0.7109375, "learning_rate": 0.0009302507836990596, "loss": 2.6176, "step": 556 }, { "epoch": 0.10500518427750023, "grad_norm": 0.69921875, "learning_rate": 0.0009300548589341693, "loss": 2.699, "step": 557 }, { "epoch": 0.10519370345932699, "grad_norm": 0.71875, "learning_rate": 0.000929858934169279, "loss": 2.5058, "step": 558 }, { "epoch": 0.10538222264115374, "grad_norm": 0.72265625, "learning_rate": 0.0009296630094043888, "loss": 2.6034, "step": 559 }, { "epoch": 0.10557074182298049, "grad_norm": 0.640625, "learning_rate": 0.0009294670846394984, "loss": 2.5785, "step": 560 }, { "epoch": 0.10575926100480725, "grad_norm": 0.7578125, "learning_rate": 0.0009292711598746082, "loss": 2.7846, "step": 561 }, { "epoch": 0.10594778018663399, "grad_norm": 0.86328125, "learning_rate": 0.000929075235109718, "loss": 2.7049, "step": 562 }, { "epoch": 0.10613629936846074, "grad_norm": 0.703125, "learning_rate": 0.0009288793103448276, "loss": 2.72, "step": 563 }, { "epoch": 0.1063248185502875, "grad_norm": 0.72265625, "learning_rate": 0.0009286833855799374, "loss": 2.7468, "step": 564 }, { "epoch": 0.10651333773211424, "grad_norm": 0.703125, "learning_rate": 0.0009284874608150471, "loss": 2.706, "step": 565 }, { "epoch": 0.10670185691394099, "grad_norm": 0.71484375, "learning_rate": 0.0009282915360501567, "loss": 2.6664, "step": 566 }, { "epoch": 0.10689037609576775, "grad_norm": 0.734375, "learning_rate": 0.0009280956112852664, "loss": 2.6685, "step": 567 }, { "epoch": 0.1070788952775945, "grad_norm": 0.76953125, "learning_rate": 0.0009278996865203762, "loss": 2.6295, "step": 568 }, { "epoch": 0.10726741445942124, "grad_norm": 0.75, "learning_rate": 0.0009277037617554858, "loss": 2.7556, "step": 569 }, { "epoch": 0.107455933641248, "grad_norm": 0.6875, "learning_rate": 0.0009275078369905956, "loss": 2.6027, "step": 570 }, { "epoch": 0.10764445282307475, "grad_norm": 0.73828125, "learning_rate": 0.0009273119122257053, "loss": 2.622, "step": 571 }, { "epoch": 0.1078329720049015, "grad_norm": 0.7421875, "learning_rate": 0.000927115987460815, "loss": 2.6136, "step": 572 }, { "epoch": 0.10802149118672826, "grad_norm": 0.7421875, "learning_rate": 0.0009269200626959248, "loss": 2.6196, "step": 573 }, { "epoch": 0.108210010368555, "grad_norm": 0.65234375, "learning_rate": 0.0009267241379310345, "loss": 2.6569, "step": 574 }, { "epoch": 0.10839852955038175, "grad_norm": 0.69921875, "learning_rate": 0.0009265282131661443, "loss": 2.7018, "step": 575 }, { "epoch": 0.10858704873220851, "grad_norm": 0.70703125, "learning_rate": 0.0009263322884012539, "loss": 2.5521, "step": 576 }, { "epoch": 0.10877556791403525, "grad_norm": 0.67578125, "learning_rate": 0.0009261363636363637, "loss": 2.6091, "step": 577 }, { "epoch": 0.108964087095862, "grad_norm": 0.6796875, "learning_rate": 0.0009259404388714734, "loss": 2.5969, "step": 578 }, { "epoch": 0.10915260627768876, "grad_norm": 0.70703125, "learning_rate": 0.0009257445141065831, "loss": 2.6032, "step": 579 }, { "epoch": 0.1093411254595155, "grad_norm": 0.671875, "learning_rate": 0.0009255485893416928, "loss": 2.6755, "step": 580 }, { "epoch": 0.10952964464134225, "grad_norm": 0.67578125, "learning_rate": 0.0009253526645768026, "loss": 2.6504, "step": 581 }, { "epoch": 0.10971816382316901, "grad_norm": 0.7109375, "learning_rate": 0.0009251567398119123, "loss": 2.7218, "step": 582 }, { "epoch": 0.10990668300499576, "grad_norm": 0.66796875, "learning_rate": 0.000924960815047022, "loss": 2.6733, "step": 583 }, { "epoch": 0.1100952021868225, "grad_norm": 0.703125, "learning_rate": 0.0009247648902821318, "loss": 2.7395, "step": 584 }, { "epoch": 0.11028372136864927, "grad_norm": 0.6953125, "learning_rate": 0.0009245689655172413, "loss": 2.7391, "step": 585 }, { "epoch": 0.11047224055047601, "grad_norm": 0.66015625, "learning_rate": 0.0009243730407523511, "loss": 2.5436, "step": 586 }, { "epoch": 0.11066075973230276, "grad_norm": 0.6796875, "learning_rate": 0.0009241771159874608, "loss": 2.6671, "step": 587 }, { "epoch": 0.11084927891412952, "grad_norm": 0.71484375, "learning_rate": 0.0009239811912225705, "loss": 2.7918, "step": 588 }, { "epoch": 0.11103779809595626, "grad_norm": 0.65625, "learning_rate": 0.0009237852664576802, "loss": 2.6591, "step": 589 }, { "epoch": 0.11122631727778301, "grad_norm": 0.6640625, "learning_rate": 0.00092358934169279, "loss": 2.5917, "step": 590 }, { "epoch": 0.11141483645960977, "grad_norm": 0.69921875, "learning_rate": 0.0009233934169278996, "loss": 2.5555, "step": 591 }, { "epoch": 0.11160335564143652, "grad_norm": 0.7421875, "learning_rate": 0.0009231974921630094, "loss": 2.7138, "step": 592 }, { "epoch": 0.11179187482326326, "grad_norm": 0.7109375, "learning_rate": 0.0009230015673981192, "loss": 2.7638, "step": 593 }, { "epoch": 0.11198039400509002, "grad_norm": 0.65234375, "learning_rate": 0.0009228056426332288, "loss": 2.6662, "step": 594 }, { "epoch": 0.11216891318691677, "grad_norm": 0.65625, "learning_rate": 0.0009226097178683386, "loss": 2.7772, "step": 595 }, { "epoch": 0.11235743236874352, "grad_norm": 0.78125, "learning_rate": 0.0009224137931034483, "loss": 2.7929, "step": 596 }, { "epoch": 0.11254595155057028, "grad_norm": 0.75, "learning_rate": 0.000922217868338558, "loss": 2.7971, "step": 597 }, { "epoch": 0.11273447073239702, "grad_norm": 0.6953125, "learning_rate": 0.0009220219435736677, "loss": 2.7352, "step": 598 }, { "epoch": 0.11292298991422377, "grad_norm": 0.71875, "learning_rate": 0.0009218260188087775, "loss": 2.6516, "step": 599 }, { "epoch": 0.11311150909605053, "grad_norm": 0.65625, "learning_rate": 0.0009216300940438871, "loss": 2.7391, "step": 600 }, { "epoch": 0.11330002827787727, "grad_norm": 0.65234375, "learning_rate": 0.0009214341692789969, "loss": 2.6186, "step": 601 }, { "epoch": 0.11348854745970402, "grad_norm": 0.66796875, "learning_rate": 0.0009212382445141067, "loss": 2.7238, "step": 602 }, { "epoch": 0.11367706664153078, "grad_norm": 0.64453125, "learning_rate": 0.0009210423197492164, "loss": 2.5381, "step": 603 }, { "epoch": 0.11386558582335753, "grad_norm": 0.83203125, "learning_rate": 0.000920846394984326, "loss": 2.6816, "step": 604 }, { "epoch": 0.11405410500518427, "grad_norm": 0.6640625, "learning_rate": 0.0009206504702194357, "loss": 2.5738, "step": 605 }, { "epoch": 0.11424262418701103, "grad_norm": 0.671875, "learning_rate": 0.0009204545454545455, "loss": 2.6701, "step": 606 }, { "epoch": 0.11443114336883778, "grad_norm": 0.65234375, "learning_rate": 0.0009202586206896551, "loss": 2.668, "step": 607 }, { "epoch": 0.11461966255066453, "grad_norm": 0.7734375, "learning_rate": 0.0009200626959247649, "loss": 2.6853, "step": 608 }, { "epoch": 0.11480818173249129, "grad_norm": 0.71875, "learning_rate": 0.0009198667711598747, "loss": 2.6926, "step": 609 }, { "epoch": 0.11499670091431803, "grad_norm": 0.6796875, "learning_rate": 0.0009196708463949843, "loss": 2.7387, "step": 610 }, { "epoch": 0.11518522009614478, "grad_norm": 0.64453125, "learning_rate": 0.0009194749216300941, "loss": 2.613, "step": 611 }, { "epoch": 0.11537373927797154, "grad_norm": 0.69921875, "learning_rate": 0.0009192789968652038, "loss": 2.5144, "step": 612 }, { "epoch": 0.11556225845979828, "grad_norm": 0.765625, "learning_rate": 0.0009190830721003135, "loss": 2.7777, "step": 613 }, { "epoch": 0.11575077764162503, "grad_norm": 0.7265625, "learning_rate": 0.0009188871473354232, "loss": 2.6103, "step": 614 }, { "epoch": 0.11593929682345179, "grad_norm": 0.65625, "learning_rate": 0.000918691222570533, "loss": 2.6806, "step": 615 }, { "epoch": 0.11612781600527854, "grad_norm": 0.73046875, "learning_rate": 0.0009184952978056426, "loss": 2.6122, "step": 616 }, { "epoch": 0.11631633518710528, "grad_norm": 0.671875, "learning_rate": 0.0009182993730407524, "loss": 2.6198, "step": 617 }, { "epoch": 0.11650485436893204, "grad_norm": 0.703125, "learning_rate": 0.0009181034482758622, "loss": 2.5324, "step": 618 }, { "epoch": 0.11669337355075879, "grad_norm": 0.6328125, "learning_rate": 0.0009179075235109718, "loss": 2.6424, "step": 619 }, { "epoch": 0.11688189273258554, "grad_norm": 0.67578125, "learning_rate": 0.0009177115987460816, "loss": 2.6544, "step": 620 }, { "epoch": 0.1170704119144123, "grad_norm": 0.67578125, "learning_rate": 0.0009175156739811913, "loss": 2.6725, "step": 621 }, { "epoch": 0.11725893109623904, "grad_norm": 0.6640625, "learning_rate": 0.000917319749216301, "loss": 2.6113, "step": 622 }, { "epoch": 0.11744745027806579, "grad_norm": 0.7109375, "learning_rate": 0.0009171238244514106, "loss": 2.6232, "step": 623 }, { "epoch": 0.11763596945989255, "grad_norm": 0.68359375, "learning_rate": 0.0009169278996865204, "loss": 2.5457, "step": 624 }, { "epoch": 0.1178244886417193, "grad_norm": 0.65234375, "learning_rate": 0.00091673197492163, "loss": 2.5777, "step": 625 }, { "epoch": 0.11801300782354604, "grad_norm": 0.6640625, "learning_rate": 0.0009165360501567398, "loss": 2.6668, "step": 626 }, { "epoch": 0.1182015270053728, "grad_norm": 0.65234375, "learning_rate": 0.0009163401253918495, "loss": 2.7571, "step": 627 }, { "epoch": 0.11839004618719955, "grad_norm": 0.671875, "learning_rate": 0.0009161442006269592, "loss": 2.6457, "step": 628 }, { "epoch": 0.11857856536902629, "grad_norm": 0.66796875, "learning_rate": 0.000915948275862069, "loss": 2.5717, "step": 629 }, { "epoch": 0.11876708455085305, "grad_norm": 0.6640625, "learning_rate": 0.0009157523510971787, "loss": 2.6518, "step": 630 }, { "epoch": 0.1189556037326798, "grad_norm": 0.69921875, "learning_rate": 0.0009155564263322885, "loss": 2.6215, "step": 631 }, { "epoch": 0.11914412291450655, "grad_norm": 0.66015625, "learning_rate": 0.0009153605015673981, "loss": 2.6829, "step": 632 }, { "epoch": 0.1193326420963333, "grad_norm": 0.75, "learning_rate": 0.0009151645768025079, "loss": 2.6905, "step": 633 }, { "epoch": 0.11952116127816005, "grad_norm": 0.71875, "learning_rate": 0.0009149686520376176, "loss": 2.5993, "step": 634 }, { "epoch": 0.1197096804599868, "grad_norm": 0.6796875, "learning_rate": 0.0009147727272727273, "loss": 2.6115, "step": 635 }, { "epoch": 0.11989819964181356, "grad_norm": 0.66796875, "learning_rate": 0.000914576802507837, "loss": 2.7553, "step": 636 }, { "epoch": 0.1200867188236403, "grad_norm": 0.6640625, "learning_rate": 0.0009143808777429468, "loss": 2.685, "step": 637 }, { "epoch": 0.12027523800546705, "grad_norm": 0.66796875, "learning_rate": 0.0009141849529780565, "loss": 2.6026, "step": 638 }, { "epoch": 0.12046375718729381, "grad_norm": 0.73046875, "learning_rate": 0.0009139890282131662, "loss": 2.7128, "step": 639 }, { "epoch": 0.12065227636912056, "grad_norm": 0.6796875, "learning_rate": 0.000913793103448276, "loss": 2.5892, "step": 640 }, { "epoch": 0.1208407955509473, "grad_norm": 0.6953125, "learning_rate": 0.0009135971786833855, "loss": 2.6472, "step": 641 }, { "epoch": 0.12102931473277406, "grad_norm": 0.6875, "learning_rate": 0.0009134012539184953, "loss": 2.6738, "step": 642 }, { "epoch": 0.12121783391460081, "grad_norm": 0.65234375, "learning_rate": 0.000913205329153605, "loss": 2.6913, "step": 643 }, { "epoch": 0.12140635309642756, "grad_norm": 0.67578125, "learning_rate": 0.0009130094043887147, "loss": 2.6355, "step": 644 }, { "epoch": 0.12159487227825432, "grad_norm": 0.65234375, "learning_rate": 0.0009128134796238244, "loss": 2.5979, "step": 645 }, { "epoch": 0.12178339146008106, "grad_norm": 0.6875, "learning_rate": 0.0009126175548589342, "loss": 2.7308, "step": 646 }, { "epoch": 0.12197191064190781, "grad_norm": 0.69140625, "learning_rate": 0.0009124216300940438, "loss": 2.6809, "step": 647 }, { "epoch": 0.12216042982373457, "grad_norm": 0.67578125, "learning_rate": 0.0009122257053291536, "loss": 2.7713, "step": 648 }, { "epoch": 0.12234894900556131, "grad_norm": 0.63671875, "learning_rate": 0.0009120297805642634, "loss": 2.4518, "step": 649 }, { "epoch": 0.12253746818738806, "grad_norm": 0.66015625, "learning_rate": 0.000911833855799373, "loss": 2.5884, "step": 650 }, { "epoch": 0.12272598736921482, "grad_norm": 0.6875, "learning_rate": 0.0009116379310344828, "loss": 2.6641, "step": 651 }, { "epoch": 0.12291450655104157, "grad_norm": 0.69140625, "learning_rate": 0.0009114420062695925, "loss": 2.5444, "step": 652 }, { "epoch": 0.12310302573286831, "grad_norm": 0.69140625, "learning_rate": 0.0009112460815047022, "loss": 2.6773, "step": 653 }, { "epoch": 0.12329154491469507, "grad_norm": 0.6796875, "learning_rate": 0.0009110501567398119, "loss": 2.6689, "step": 654 }, { "epoch": 0.12348006409652182, "grad_norm": 0.6953125, "learning_rate": 0.0009108542319749217, "loss": 2.8371, "step": 655 }, { "epoch": 0.12366858327834858, "grad_norm": 0.671875, "learning_rate": 0.0009106583072100313, "loss": 2.6491, "step": 656 }, { "epoch": 0.12385710246017533, "grad_norm": 0.70703125, "learning_rate": 0.0009104623824451411, "loss": 2.7192, "step": 657 }, { "epoch": 0.12404562164200207, "grad_norm": 0.69140625, "learning_rate": 0.0009102664576802509, "loss": 2.7254, "step": 658 }, { "epoch": 0.12423414082382883, "grad_norm": 0.65625, "learning_rate": 0.0009100705329153606, "loss": 2.6403, "step": 659 }, { "epoch": 0.12442266000565558, "grad_norm": 0.67578125, "learning_rate": 0.0009098746081504702, "loss": 2.6217, "step": 660 }, { "epoch": 0.12461117918748232, "grad_norm": 0.7265625, "learning_rate": 0.0009096786833855799, "loss": 2.7402, "step": 661 }, { "epoch": 0.12479969836930908, "grad_norm": 0.671875, "learning_rate": 0.0009094827586206897, "loss": 2.7237, "step": 662 }, { "epoch": 0.12498821755113583, "grad_norm": 0.7109375, "learning_rate": 0.0009092868338557993, "loss": 2.5321, "step": 663 }, { "epoch": 0.1251767367329626, "grad_norm": 0.69140625, "learning_rate": 0.0009090909090909091, "loss": 2.4766, "step": 664 }, { "epoch": 0.12536525591478934, "grad_norm": 0.625, "learning_rate": 0.0009088949843260188, "loss": 2.5655, "step": 665 }, { "epoch": 0.12555377509661608, "grad_norm": 0.6796875, "learning_rate": 0.0009086990595611285, "loss": 2.4495, "step": 666 }, { "epoch": 0.12574229427844283, "grad_norm": 0.67578125, "learning_rate": 0.0009085031347962383, "loss": 2.7035, "step": 667 }, { "epoch": 0.12593081346026958, "grad_norm": 0.671875, "learning_rate": 0.000908307210031348, "loss": 2.5528, "step": 668 }, { "epoch": 0.12611933264209632, "grad_norm": 0.66015625, "learning_rate": 0.0009081112852664577, "loss": 2.5787, "step": 669 }, { "epoch": 0.1263078518239231, "grad_norm": 0.6640625, "learning_rate": 0.0009079153605015674, "loss": 2.6167, "step": 670 }, { "epoch": 0.12649637100574984, "grad_norm": 0.66015625, "learning_rate": 0.0009077194357366772, "loss": 2.7147, "step": 671 }, { "epoch": 0.1266848901875766, "grad_norm": 0.7109375, "learning_rate": 0.0009075235109717868, "loss": 2.7819, "step": 672 }, { "epoch": 0.12687340936940333, "grad_norm": 0.69921875, "learning_rate": 0.0009073275862068966, "loss": 2.5718, "step": 673 }, { "epoch": 0.12706192855123008, "grad_norm": 0.70703125, "learning_rate": 0.0009071316614420063, "loss": 2.6887, "step": 674 }, { "epoch": 0.12725044773305683, "grad_norm": 0.69921875, "learning_rate": 0.000906935736677116, "loss": 2.6037, "step": 675 }, { "epoch": 0.1274389669148836, "grad_norm": 0.65234375, "learning_rate": 0.0009067398119122258, "loss": 2.6148, "step": 676 }, { "epoch": 0.12762748609671035, "grad_norm": 0.796875, "learning_rate": 0.0009065438871473355, "loss": 2.6737, "step": 677 }, { "epoch": 0.1278160052785371, "grad_norm": 0.75390625, "learning_rate": 0.0009063479623824452, "loss": 2.7679, "step": 678 }, { "epoch": 0.12800452446036384, "grad_norm": 0.703125, "learning_rate": 0.0009061520376175548, "loss": 2.6948, "step": 679 }, { "epoch": 0.12819304364219059, "grad_norm": 0.671875, "learning_rate": 0.0009059561128526646, "loss": 2.6185, "step": 680 }, { "epoch": 0.12838156282401733, "grad_norm": 0.7578125, "learning_rate": 0.0009057601880877742, "loss": 2.7351, "step": 681 }, { "epoch": 0.1285700820058441, "grad_norm": 0.6796875, "learning_rate": 0.000905564263322884, "loss": 2.6394, "step": 682 }, { "epoch": 0.12875860118767085, "grad_norm": 0.6875, "learning_rate": 0.0009053683385579937, "loss": 2.7473, "step": 683 }, { "epoch": 0.1289471203694976, "grad_norm": 0.68359375, "learning_rate": 0.0009051724137931034, "loss": 2.5965, "step": 684 }, { "epoch": 0.12913563955132434, "grad_norm": 0.671875, "learning_rate": 0.0009049764890282132, "loss": 2.8092, "step": 685 }, { "epoch": 0.1293241587331511, "grad_norm": 0.72265625, "learning_rate": 0.0009047805642633229, "loss": 2.58, "step": 686 }, { "epoch": 0.12951267791497784, "grad_norm": 0.69921875, "learning_rate": 0.0009045846394984327, "loss": 2.6549, "step": 687 }, { "epoch": 0.1297011970968046, "grad_norm": 0.66796875, "learning_rate": 0.0009043887147335423, "loss": 2.6374, "step": 688 }, { "epoch": 0.12988971627863136, "grad_norm": 0.6875, "learning_rate": 0.0009041927899686521, "loss": 2.5968, "step": 689 }, { "epoch": 0.1300782354604581, "grad_norm": 0.66015625, "learning_rate": 0.0009039968652037618, "loss": 2.6792, "step": 690 }, { "epoch": 0.13026675464228485, "grad_norm": 0.72265625, "learning_rate": 0.0009038009404388715, "loss": 2.6631, "step": 691 }, { "epoch": 0.1304552738241116, "grad_norm": 0.7578125, "learning_rate": 0.0009036050156739812, "loss": 2.7008, "step": 692 }, { "epoch": 0.13064379300593834, "grad_norm": 0.6640625, "learning_rate": 0.000903409090909091, "loss": 2.5822, "step": 693 }, { "epoch": 0.13083231218776512, "grad_norm": 0.65625, "learning_rate": 0.0009032131661442007, "loss": 2.5259, "step": 694 }, { "epoch": 0.13102083136959186, "grad_norm": 0.6796875, "learning_rate": 0.0009030172413793104, "loss": 2.5717, "step": 695 }, { "epoch": 0.1312093505514186, "grad_norm": 0.71875, "learning_rate": 0.0009028213166144202, "loss": 2.7658, "step": 696 }, { "epoch": 0.13139786973324535, "grad_norm": 0.76953125, "learning_rate": 0.0009026253918495298, "loss": 2.6304, "step": 697 }, { "epoch": 0.1315863889150721, "grad_norm": 0.6875, "learning_rate": 0.0009024294670846395, "loss": 2.6773, "step": 698 }, { "epoch": 0.13177490809689885, "grad_norm": 0.75390625, "learning_rate": 0.0009022335423197492, "loss": 2.6761, "step": 699 }, { "epoch": 0.13196342727872562, "grad_norm": 0.8984375, "learning_rate": 0.0009020376175548589, "loss": 2.6741, "step": 700 }, { "epoch": 0.13215194646055237, "grad_norm": 0.67578125, "learning_rate": 0.0009018416927899686, "loss": 2.5493, "step": 701 }, { "epoch": 0.1323404656423791, "grad_norm": 0.66015625, "learning_rate": 0.0009016457680250784, "loss": 2.6143, "step": 702 }, { "epoch": 0.13252898482420586, "grad_norm": 0.76953125, "learning_rate": 0.000901449843260188, "loss": 2.6671, "step": 703 }, { "epoch": 0.1327175040060326, "grad_norm": 0.69921875, "learning_rate": 0.0009012539184952978, "loss": 2.6108, "step": 704 }, { "epoch": 0.13290602318785935, "grad_norm": 0.6953125, "learning_rate": 0.0009010579937304076, "loss": 2.6649, "step": 705 }, { "epoch": 0.13309454236968613, "grad_norm": 0.66796875, "learning_rate": 0.0009008620689655172, "loss": 2.6925, "step": 706 }, { "epoch": 0.13328306155151287, "grad_norm": 0.703125, "learning_rate": 0.000900666144200627, "loss": 2.6506, "step": 707 }, { "epoch": 0.13347158073333962, "grad_norm": 0.8046875, "learning_rate": 0.0009004702194357367, "loss": 2.7378, "step": 708 }, { "epoch": 0.13366009991516636, "grad_norm": 0.6953125, "learning_rate": 0.0009002742946708464, "loss": 2.6705, "step": 709 }, { "epoch": 0.1338486190969931, "grad_norm": 0.71875, "learning_rate": 0.0009000783699059561, "loss": 2.7019, "step": 710 }, { "epoch": 0.13403713827881986, "grad_norm": 0.65625, "learning_rate": 0.0008998824451410659, "loss": 2.6219, "step": 711 }, { "epoch": 0.13422565746064663, "grad_norm": 0.7734375, "learning_rate": 0.0008996865203761755, "loss": 2.5641, "step": 712 }, { "epoch": 0.13441417664247338, "grad_norm": 0.76171875, "learning_rate": 0.0008994905956112853, "loss": 2.6977, "step": 713 }, { "epoch": 0.13460269582430012, "grad_norm": 0.68359375, "learning_rate": 0.0008992946708463951, "loss": 2.7332, "step": 714 }, { "epoch": 0.13479121500612687, "grad_norm": 0.67578125, "learning_rate": 0.0008990987460815048, "loss": 2.5868, "step": 715 }, { "epoch": 0.13497973418795361, "grad_norm": 0.67578125, "learning_rate": 0.0008989028213166145, "loss": 2.6311, "step": 716 }, { "epoch": 0.1351682533697804, "grad_norm": 0.6640625, "learning_rate": 0.0008987068965517241, "loss": 2.6369, "step": 717 }, { "epoch": 0.13535677255160714, "grad_norm": 0.69921875, "learning_rate": 0.0008985109717868339, "loss": 2.778, "step": 718 }, { "epoch": 0.13554529173343388, "grad_norm": 0.7421875, "learning_rate": 0.0008983150470219435, "loss": 2.6355, "step": 719 }, { "epoch": 0.13573381091526063, "grad_norm": 0.67578125, "learning_rate": 0.0008981191222570533, "loss": 2.6524, "step": 720 }, { "epoch": 0.13592233009708737, "grad_norm": 0.6875, "learning_rate": 0.000897923197492163, "loss": 2.7543, "step": 721 }, { "epoch": 0.13611084927891412, "grad_norm": 0.71875, "learning_rate": 0.0008977272727272727, "loss": 2.7062, "step": 722 }, { "epoch": 0.1362993684607409, "grad_norm": 0.75390625, "learning_rate": 0.0008975313479623825, "loss": 2.7506, "step": 723 }, { "epoch": 0.13648788764256764, "grad_norm": 0.6484375, "learning_rate": 0.0008973354231974922, "loss": 2.6015, "step": 724 }, { "epoch": 0.1366764068243944, "grad_norm": 0.640625, "learning_rate": 0.0008971394984326019, "loss": 2.6391, "step": 725 }, { "epoch": 0.13686492600622113, "grad_norm": 0.7421875, "learning_rate": 0.0008969435736677116, "loss": 2.7587, "step": 726 }, { "epoch": 0.13705344518804788, "grad_norm": 0.6640625, "learning_rate": 0.0008967476489028214, "loss": 2.7175, "step": 727 }, { "epoch": 0.13724196436987462, "grad_norm": 0.68359375, "learning_rate": 0.000896551724137931, "loss": 2.6185, "step": 728 }, { "epoch": 0.1374304835517014, "grad_norm": 0.6796875, "learning_rate": 0.0008963557993730408, "loss": 2.5831, "step": 729 }, { "epoch": 0.13761900273352815, "grad_norm": 0.64453125, "learning_rate": 0.0008961598746081505, "loss": 2.6364, "step": 730 }, { "epoch": 0.1378075219153549, "grad_norm": 0.65625, "learning_rate": 0.0008959639498432602, "loss": 2.5928, "step": 731 }, { "epoch": 0.13799604109718164, "grad_norm": 0.66796875, "learning_rate": 0.00089576802507837, "loss": 2.6144, "step": 732 }, { "epoch": 0.13818456027900838, "grad_norm": 0.671875, "learning_rate": 0.0008955721003134797, "loss": 2.5887, "step": 733 }, { "epoch": 0.13837307946083513, "grad_norm": 0.62109375, "learning_rate": 0.0008953761755485894, "loss": 2.5888, "step": 734 }, { "epoch": 0.1385615986426619, "grad_norm": 0.63671875, "learning_rate": 0.000895180250783699, "loss": 2.5938, "step": 735 }, { "epoch": 0.13875011782448865, "grad_norm": 0.640625, "learning_rate": 0.0008949843260188088, "loss": 2.5876, "step": 736 }, { "epoch": 0.1389386370063154, "grad_norm": 0.62890625, "learning_rate": 0.0008947884012539184, "loss": 2.6677, "step": 737 }, { "epoch": 0.13912715618814214, "grad_norm": 0.65234375, "learning_rate": 0.0008945924764890282, "loss": 2.5932, "step": 738 }, { "epoch": 0.1393156753699689, "grad_norm": 0.68359375, "learning_rate": 0.0008943965517241379, "loss": 2.6358, "step": 739 }, { "epoch": 0.13950419455179563, "grad_norm": 0.671875, "learning_rate": 0.0008942006269592476, "loss": 2.6529, "step": 740 }, { "epoch": 0.1396927137336224, "grad_norm": 0.6796875, "learning_rate": 0.0008940047021943573, "loss": 2.6557, "step": 741 }, { "epoch": 0.13988123291544916, "grad_norm": 0.6796875, "learning_rate": 0.0008938087774294671, "loss": 2.6333, "step": 742 }, { "epoch": 0.1400697520972759, "grad_norm": 0.734375, "learning_rate": 0.0008936128526645769, "loss": 2.5974, "step": 743 }, { "epoch": 0.14025827127910265, "grad_norm": 0.71875, "learning_rate": 0.0008934169278996865, "loss": 2.6484, "step": 744 }, { "epoch": 0.1404467904609294, "grad_norm": 0.97265625, "learning_rate": 0.0008932210031347963, "loss": 2.617, "step": 745 }, { "epoch": 0.14063530964275614, "grad_norm": 0.765625, "learning_rate": 0.000893025078369906, "loss": 2.6803, "step": 746 }, { "epoch": 0.14082382882458291, "grad_norm": 0.6875, "learning_rate": 0.0008928291536050157, "loss": 2.6882, "step": 747 }, { "epoch": 0.14101234800640966, "grad_norm": 0.79296875, "learning_rate": 0.0008926332288401254, "loss": 2.6814, "step": 748 }, { "epoch": 0.1412008671882364, "grad_norm": 0.75, "learning_rate": 0.0008924373040752352, "loss": 2.7618, "step": 749 }, { "epoch": 0.14138938637006315, "grad_norm": 0.69140625, "learning_rate": 0.0008922413793103448, "loss": 2.7598, "step": 750 }, { "epoch": 0.1415779055518899, "grad_norm": 0.91015625, "learning_rate": 0.0008920454545454546, "loss": 2.6827, "step": 751 }, { "epoch": 0.14176642473371664, "grad_norm": 0.7265625, "learning_rate": 0.0008918495297805644, "loss": 2.7531, "step": 752 }, { "epoch": 0.14195494391554342, "grad_norm": 0.6484375, "learning_rate": 0.000891653605015674, "loss": 2.5641, "step": 753 }, { "epoch": 0.14214346309737017, "grad_norm": 0.65625, "learning_rate": 0.0008914576802507837, "loss": 2.6333, "step": 754 }, { "epoch": 0.1423319822791969, "grad_norm": 0.69140625, "learning_rate": 0.0008912617554858934, "loss": 2.7232, "step": 755 }, { "epoch": 0.14252050146102366, "grad_norm": 0.65625, "learning_rate": 0.0008910658307210031, "loss": 2.6999, "step": 756 }, { "epoch": 0.1427090206428504, "grad_norm": 0.6953125, "learning_rate": 0.0008908699059561128, "loss": 2.6267, "step": 757 }, { "epoch": 0.14289753982467715, "grad_norm": 0.765625, "learning_rate": 0.0008906739811912226, "loss": 2.6947, "step": 758 }, { "epoch": 0.14308605900650392, "grad_norm": 0.76953125, "learning_rate": 0.0008904780564263322, "loss": 2.6448, "step": 759 }, { "epoch": 0.14327457818833067, "grad_norm": 0.703125, "learning_rate": 0.000890282131661442, "loss": 2.7433, "step": 760 }, { "epoch": 0.14346309737015742, "grad_norm": 0.71875, "learning_rate": 0.0008900862068965518, "loss": 2.6372, "step": 761 }, { "epoch": 0.14365161655198416, "grad_norm": 0.68359375, "learning_rate": 0.0008898902821316614, "loss": 2.7054, "step": 762 }, { "epoch": 0.1438401357338109, "grad_norm": 0.65625, "learning_rate": 0.0008896943573667712, "loss": 2.6433, "step": 763 }, { "epoch": 0.14402865491563765, "grad_norm": 0.671875, "learning_rate": 0.0008894984326018809, "loss": 2.6648, "step": 764 }, { "epoch": 0.14421717409746443, "grad_norm": 0.7109375, "learning_rate": 0.0008893025078369906, "loss": 2.6961, "step": 765 }, { "epoch": 0.14440569327929118, "grad_norm": 0.69921875, "learning_rate": 0.0008891065830721003, "loss": 2.5835, "step": 766 }, { "epoch": 0.14459421246111792, "grad_norm": 0.67578125, "learning_rate": 0.0008889106583072101, "loss": 2.6922, "step": 767 }, { "epoch": 0.14478273164294467, "grad_norm": 0.65625, "learning_rate": 0.0008887147335423197, "loss": 2.5899, "step": 768 }, { "epoch": 0.1449712508247714, "grad_norm": 0.70703125, "learning_rate": 0.0008885188087774295, "loss": 2.6164, "step": 769 }, { "epoch": 0.14515977000659816, "grad_norm": 0.6953125, "learning_rate": 0.0008883228840125393, "loss": 2.708, "step": 770 }, { "epoch": 0.14534828918842493, "grad_norm": 0.69140625, "learning_rate": 0.000888126959247649, "loss": 2.59, "step": 771 }, { "epoch": 0.14553680837025168, "grad_norm": 0.69140625, "learning_rate": 0.0008879310344827587, "loss": 2.5905, "step": 772 }, { "epoch": 0.14572532755207843, "grad_norm": 0.66015625, "learning_rate": 0.0008877351097178683, "loss": 2.6069, "step": 773 }, { "epoch": 0.14591384673390517, "grad_norm": 0.72265625, "learning_rate": 0.0008875391849529781, "loss": 2.695, "step": 774 }, { "epoch": 0.14610236591573192, "grad_norm": 0.68359375, "learning_rate": 0.0008873432601880877, "loss": 2.7418, "step": 775 }, { "epoch": 0.14629088509755866, "grad_norm": 0.66796875, "learning_rate": 0.0008871473354231975, "loss": 2.6499, "step": 776 }, { "epoch": 0.14647940427938544, "grad_norm": 0.66015625, "learning_rate": 0.0008869514106583072, "loss": 2.6959, "step": 777 }, { "epoch": 0.14666792346121219, "grad_norm": 0.6875, "learning_rate": 0.0008867554858934169, "loss": 2.6255, "step": 778 }, { "epoch": 0.14685644264303893, "grad_norm": 0.6875, "learning_rate": 0.0008865595611285267, "loss": 2.6885, "step": 779 }, { "epoch": 0.14704496182486568, "grad_norm": 0.671875, "learning_rate": 0.0008863636363636364, "loss": 2.6154, "step": 780 }, { "epoch": 0.14723348100669242, "grad_norm": 0.63671875, "learning_rate": 0.0008861677115987461, "loss": 2.6715, "step": 781 }, { "epoch": 0.14742200018851917, "grad_norm": 0.69140625, "learning_rate": 0.0008859717868338558, "loss": 2.6356, "step": 782 }, { "epoch": 0.14761051937034594, "grad_norm": 0.6953125, "learning_rate": 0.0008857758620689656, "loss": 2.6181, "step": 783 }, { "epoch": 0.1477990385521727, "grad_norm": 0.70703125, "learning_rate": 0.0008855799373040752, "loss": 2.6921, "step": 784 }, { "epoch": 0.14798755773399944, "grad_norm": 0.62109375, "learning_rate": 0.000885384012539185, "loss": 2.5067, "step": 785 }, { "epoch": 0.14817607691582618, "grad_norm": 0.7109375, "learning_rate": 0.0008851880877742947, "loss": 2.729, "step": 786 }, { "epoch": 0.14836459609765293, "grad_norm": 0.72265625, "learning_rate": 0.0008849921630094044, "loss": 2.6272, "step": 787 }, { "epoch": 0.14855311527947967, "grad_norm": 0.63671875, "learning_rate": 0.0008847962382445142, "loss": 2.5603, "step": 788 }, { "epoch": 0.14874163446130645, "grad_norm": 0.671875, "learning_rate": 0.0008846003134796239, "loss": 2.7072, "step": 789 }, { "epoch": 0.1489301536431332, "grad_norm": 0.6953125, "learning_rate": 0.0008844043887147336, "loss": 2.6431, "step": 790 }, { "epoch": 0.14911867282495994, "grad_norm": 0.703125, "learning_rate": 0.0008842084639498433, "loss": 2.6058, "step": 791 }, { "epoch": 0.1493071920067867, "grad_norm": 0.6484375, "learning_rate": 0.000884012539184953, "loss": 2.6369, "step": 792 }, { "epoch": 0.14949571118861343, "grad_norm": 0.63671875, "learning_rate": 0.0008838166144200626, "loss": 2.5803, "step": 793 }, { "epoch": 0.14968423037044018, "grad_norm": 0.65625, "learning_rate": 0.0008836206896551724, "loss": 2.6228, "step": 794 }, { "epoch": 0.14987274955226695, "grad_norm": 0.65234375, "learning_rate": 0.0008834247648902821, "loss": 2.5875, "step": 795 }, { "epoch": 0.1500612687340937, "grad_norm": 0.63671875, "learning_rate": 0.0008832288401253918, "loss": 2.6705, "step": 796 }, { "epoch": 0.15024978791592045, "grad_norm": 0.63671875, "learning_rate": 0.0008830329153605015, "loss": 2.6356, "step": 797 }, { "epoch": 0.1504383070977472, "grad_norm": 0.65625, "learning_rate": 0.0008828369905956113, "loss": 2.6354, "step": 798 }, { "epoch": 0.15062682627957394, "grad_norm": 0.66796875, "learning_rate": 0.0008826410658307211, "loss": 2.7054, "step": 799 }, { "epoch": 0.15081534546140068, "grad_norm": 0.671875, "learning_rate": 0.0008824451410658307, "loss": 2.5472, "step": 800 }, { "epoch": 0.15100386464322746, "grad_norm": 0.64453125, "learning_rate": 0.0008822492163009405, "loss": 2.5917, "step": 801 }, { "epoch": 0.1511923838250542, "grad_norm": 0.671875, "learning_rate": 0.0008820532915360502, "loss": 2.6797, "step": 802 }, { "epoch": 0.15138090300688095, "grad_norm": 0.68359375, "learning_rate": 0.0008818573667711599, "loss": 2.6894, "step": 803 }, { "epoch": 0.1515694221887077, "grad_norm": 0.75390625, "learning_rate": 0.0008816614420062696, "loss": 2.6542, "step": 804 }, { "epoch": 0.15175794137053444, "grad_norm": 0.6640625, "learning_rate": 0.0008814655172413794, "loss": 2.7165, "step": 805 }, { "epoch": 0.1519464605523612, "grad_norm": 0.6875, "learning_rate": 0.000881269592476489, "loss": 2.6396, "step": 806 }, { "epoch": 0.15213497973418796, "grad_norm": 0.66015625, "learning_rate": 0.0008810736677115988, "loss": 2.5562, "step": 807 }, { "epoch": 0.1523234989160147, "grad_norm": 0.6953125, "learning_rate": 0.0008808777429467086, "loss": 2.5097, "step": 808 }, { "epoch": 0.15251201809784146, "grad_norm": 0.7109375, "learning_rate": 0.0008806818181818182, "loss": 2.6022, "step": 809 }, { "epoch": 0.1527005372796682, "grad_norm": 0.67578125, "learning_rate": 0.0008804858934169279, "loss": 2.6144, "step": 810 }, { "epoch": 0.15288905646149495, "grad_norm": 0.68359375, "learning_rate": 0.0008802899686520376, "loss": 2.7956, "step": 811 }, { "epoch": 0.1530775756433217, "grad_norm": 0.66015625, "learning_rate": 0.0008800940438871473, "loss": 2.6763, "step": 812 }, { "epoch": 0.15326609482514847, "grad_norm": 0.68359375, "learning_rate": 0.000879898119122257, "loss": 2.7261, "step": 813 }, { "epoch": 0.15345461400697522, "grad_norm": 0.6484375, "learning_rate": 0.0008797021943573668, "loss": 2.5969, "step": 814 }, { "epoch": 0.15364313318880196, "grad_norm": 0.68359375, "learning_rate": 0.0008795062695924764, "loss": 2.5595, "step": 815 }, { "epoch": 0.1538316523706287, "grad_norm": 0.66796875, "learning_rate": 0.0008793103448275862, "loss": 2.4956, "step": 816 }, { "epoch": 0.15402017155245545, "grad_norm": 0.69140625, "learning_rate": 0.000879114420062696, "loss": 2.6691, "step": 817 }, { "epoch": 0.1542086907342822, "grad_norm": 0.65234375, "learning_rate": 0.0008789184952978056, "loss": 2.5205, "step": 818 }, { "epoch": 0.15439720991610897, "grad_norm": 0.6875, "learning_rate": 0.0008787225705329154, "loss": 2.5277, "step": 819 }, { "epoch": 0.15458572909793572, "grad_norm": 0.62890625, "learning_rate": 0.0008785266457680251, "loss": 2.6255, "step": 820 }, { "epoch": 0.15477424827976247, "grad_norm": 0.68359375, "learning_rate": 0.0008783307210031348, "loss": 2.6624, "step": 821 }, { "epoch": 0.1549627674615892, "grad_norm": 0.6171875, "learning_rate": 0.0008781347962382445, "loss": 2.6192, "step": 822 }, { "epoch": 0.15515128664341596, "grad_norm": 0.671875, "learning_rate": 0.0008779388714733543, "loss": 2.6637, "step": 823 }, { "epoch": 0.1553398058252427, "grad_norm": 0.703125, "learning_rate": 0.000877742946708464, "loss": 2.6116, "step": 824 }, { "epoch": 0.15552832500706948, "grad_norm": 0.66015625, "learning_rate": 0.0008775470219435737, "loss": 2.7295, "step": 825 }, { "epoch": 0.15571684418889623, "grad_norm": 0.69140625, "learning_rate": 0.0008773510971786835, "loss": 2.6604, "step": 826 }, { "epoch": 0.15590536337072297, "grad_norm": 0.703125, "learning_rate": 0.0008771551724137932, "loss": 2.661, "step": 827 }, { "epoch": 0.15609388255254972, "grad_norm": 0.6875, "learning_rate": 0.0008769592476489029, "loss": 2.6784, "step": 828 }, { "epoch": 0.15628240173437646, "grad_norm": 0.69140625, "learning_rate": 0.0008767633228840125, "loss": 2.6322, "step": 829 }, { "epoch": 0.1564709209162032, "grad_norm": 0.67578125, "learning_rate": 0.0008765673981191223, "loss": 2.6958, "step": 830 }, { "epoch": 0.15665944009802998, "grad_norm": 0.640625, "learning_rate": 0.0008763714733542319, "loss": 2.614, "step": 831 }, { "epoch": 0.15684795927985673, "grad_norm": 0.67578125, "learning_rate": 0.0008761755485893417, "loss": 2.61, "step": 832 }, { "epoch": 0.15703647846168348, "grad_norm": 0.703125, "learning_rate": 0.0008759796238244514, "loss": 2.6638, "step": 833 }, { "epoch": 0.15722499764351022, "grad_norm": 0.69921875, "learning_rate": 0.0008757836990595611, "loss": 2.5964, "step": 834 }, { "epoch": 0.15741351682533697, "grad_norm": 0.6953125, "learning_rate": 0.0008755877742946709, "loss": 2.6957, "step": 835 }, { "epoch": 0.15760203600716374, "grad_norm": 0.640625, "learning_rate": 0.0008753918495297806, "loss": 2.5888, "step": 836 }, { "epoch": 0.1577905551889905, "grad_norm": 0.6640625, "learning_rate": 0.0008751959247648903, "loss": 2.644, "step": 837 }, { "epoch": 0.15797907437081724, "grad_norm": 0.69921875, "learning_rate": 0.000875, "loss": 2.772, "step": 838 }, { "epoch": 0.15816759355264398, "grad_norm": 0.70703125, "learning_rate": 0.0008748040752351098, "loss": 2.6483, "step": 839 }, { "epoch": 0.15835611273447073, "grad_norm": 0.69140625, "learning_rate": 0.0008746081504702194, "loss": 2.5755, "step": 840 }, { "epoch": 0.15854463191629747, "grad_norm": 0.71875, "learning_rate": 0.0008744122257053292, "loss": 2.6584, "step": 841 }, { "epoch": 0.15873315109812425, "grad_norm": 0.6796875, "learning_rate": 0.0008742163009404389, "loss": 2.6448, "step": 842 }, { "epoch": 0.158921670279951, "grad_norm": 0.6640625, "learning_rate": 0.0008740203761755486, "loss": 2.6187, "step": 843 }, { "epoch": 0.15911018946177774, "grad_norm": 0.6953125, "learning_rate": 0.0008738244514106584, "loss": 2.7043, "step": 844 }, { "epoch": 0.1592987086436045, "grad_norm": 0.6875, "learning_rate": 0.0008736285266457681, "loss": 2.5718, "step": 845 }, { "epoch": 0.15948722782543123, "grad_norm": 0.66015625, "learning_rate": 0.0008734326018808778, "loss": 2.6218, "step": 846 }, { "epoch": 0.15967574700725798, "grad_norm": 0.640625, "learning_rate": 0.0008732366771159875, "loss": 2.7162, "step": 847 }, { "epoch": 0.15986426618908475, "grad_norm": 0.69140625, "learning_rate": 0.0008730407523510972, "loss": 2.7002, "step": 848 }, { "epoch": 0.1600527853709115, "grad_norm": 0.984375, "learning_rate": 0.0008728448275862068, "loss": 2.5591, "step": 849 }, { "epoch": 0.16024130455273825, "grad_norm": 0.70703125, "learning_rate": 0.0008726489028213166, "loss": 2.68, "step": 850 }, { "epoch": 0.160429823734565, "grad_norm": 0.703125, "learning_rate": 0.0008724529780564263, "loss": 2.6753, "step": 851 }, { "epoch": 0.16061834291639174, "grad_norm": 0.6953125, "learning_rate": 0.000872257053291536, "loss": 2.6676, "step": 852 }, { "epoch": 0.16080686209821848, "grad_norm": 0.75390625, "learning_rate": 0.0008720611285266457, "loss": 2.6622, "step": 853 }, { "epoch": 0.16099538128004526, "grad_norm": 0.76953125, "learning_rate": 0.0008718652037617555, "loss": 2.8021, "step": 854 }, { "epoch": 0.161183900461872, "grad_norm": 0.66796875, "learning_rate": 0.0008716692789968653, "loss": 2.6823, "step": 855 }, { "epoch": 0.16137241964369875, "grad_norm": 0.73828125, "learning_rate": 0.0008714733542319749, "loss": 2.6915, "step": 856 }, { "epoch": 0.1615609388255255, "grad_norm": 0.7109375, "learning_rate": 0.0008712774294670847, "loss": 2.6373, "step": 857 }, { "epoch": 0.16174945800735224, "grad_norm": 0.67578125, "learning_rate": 0.0008710815047021944, "loss": 2.6126, "step": 858 }, { "epoch": 0.161937977189179, "grad_norm": 0.6953125, "learning_rate": 0.0008708855799373041, "loss": 2.544, "step": 859 }, { "epoch": 0.16212649637100576, "grad_norm": 0.6796875, "learning_rate": 0.0008706896551724138, "loss": 2.5462, "step": 860 }, { "epoch": 0.1623150155528325, "grad_norm": 0.68359375, "learning_rate": 0.0008704937304075236, "loss": 2.6703, "step": 861 }, { "epoch": 0.16250353473465926, "grad_norm": 0.74609375, "learning_rate": 0.0008702978056426332, "loss": 2.606, "step": 862 }, { "epoch": 0.162692053916486, "grad_norm": 0.71484375, "learning_rate": 0.000870101880877743, "loss": 2.5978, "step": 863 }, { "epoch": 0.16288057309831275, "grad_norm": 0.734375, "learning_rate": 0.0008699059561128528, "loss": 2.6027, "step": 864 }, { "epoch": 0.1630690922801395, "grad_norm": 0.65234375, "learning_rate": 0.0008697100313479624, "loss": 2.654, "step": 865 }, { "epoch": 0.16325761146196627, "grad_norm": 0.640625, "learning_rate": 0.0008695141065830722, "loss": 2.4922, "step": 866 }, { "epoch": 0.16344613064379301, "grad_norm": 0.6875, "learning_rate": 0.0008693181818181818, "loss": 2.54, "step": 867 }, { "epoch": 0.16363464982561976, "grad_norm": 0.6640625, "learning_rate": 0.0008691222570532915, "loss": 2.6802, "step": 868 }, { "epoch": 0.1638231690074465, "grad_norm": 0.76953125, "learning_rate": 0.0008689263322884012, "loss": 2.6451, "step": 869 }, { "epoch": 0.16401168818927325, "grad_norm": 0.6484375, "learning_rate": 0.000868730407523511, "loss": 2.6301, "step": 870 }, { "epoch": 0.1642002073711, "grad_norm": 0.65625, "learning_rate": 0.0008685344827586206, "loss": 2.6155, "step": 871 }, { "epoch": 0.16438872655292677, "grad_norm": 0.65234375, "learning_rate": 0.0008683385579937304, "loss": 2.5815, "step": 872 }, { "epoch": 0.16457724573475352, "grad_norm": 0.6640625, "learning_rate": 0.0008681426332288402, "loss": 2.6225, "step": 873 }, { "epoch": 0.16476576491658027, "grad_norm": 0.671875, "learning_rate": 0.0008679467084639498, "loss": 2.6562, "step": 874 }, { "epoch": 0.164954284098407, "grad_norm": 0.6796875, "learning_rate": 0.0008677507836990596, "loss": 2.6155, "step": 875 }, { "epoch": 0.16514280328023376, "grad_norm": 0.69140625, "learning_rate": 0.0008675548589341693, "loss": 2.6055, "step": 876 }, { "epoch": 0.1653313224620605, "grad_norm": 0.64453125, "learning_rate": 0.000867358934169279, "loss": 2.637, "step": 877 }, { "epoch": 0.16551984164388728, "grad_norm": 0.69140625, "learning_rate": 0.0008671630094043887, "loss": 2.6913, "step": 878 }, { "epoch": 0.16570836082571402, "grad_norm": 0.6484375, "learning_rate": 0.0008669670846394985, "loss": 2.5859, "step": 879 }, { "epoch": 0.16589688000754077, "grad_norm": 0.68359375, "learning_rate": 0.0008667711598746082, "loss": 2.7039, "step": 880 }, { "epoch": 0.16608539918936752, "grad_norm": 0.67578125, "learning_rate": 0.0008665752351097179, "loss": 2.662, "step": 881 }, { "epoch": 0.16627391837119426, "grad_norm": 0.65625, "learning_rate": 0.0008663793103448277, "loss": 2.6888, "step": 882 }, { "epoch": 0.166462437553021, "grad_norm": 0.625, "learning_rate": 0.0008661833855799374, "loss": 2.6346, "step": 883 }, { "epoch": 0.16665095673484778, "grad_norm": 0.65625, "learning_rate": 0.0008659874608150471, "loss": 2.6536, "step": 884 }, { "epoch": 0.16683947591667453, "grad_norm": 0.65625, "learning_rate": 0.0008657915360501567, "loss": 2.5995, "step": 885 }, { "epoch": 0.16702799509850128, "grad_norm": 0.6484375, "learning_rate": 0.0008655956112852665, "loss": 2.7111, "step": 886 }, { "epoch": 0.16721651428032802, "grad_norm": 0.71484375, "learning_rate": 0.0008653996865203761, "loss": 2.6629, "step": 887 }, { "epoch": 0.16740503346215477, "grad_norm": 0.65625, "learning_rate": 0.0008652037617554859, "loss": 2.5553, "step": 888 }, { "epoch": 0.1675935526439815, "grad_norm": 0.6796875, "learning_rate": 0.0008650078369905956, "loss": 2.7594, "step": 889 }, { "epoch": 0.1677820718258083, "grad_norm": 0.69140625, "learning_rate": 0.0008648119122257053, "loss": 2.6729, "step": 890 }, { "epoch": 0.16797059100763503, "grad_norm": 0.68359375, "learning_rate": 0.000864615987460815, "loss": 2.7915, "step": 891 }, { "epoch": 0.16815911018946178, "grad_norm": 0.6640625, "learning_rate": 0.0008644200626959248, "loss": 2.5758, "step": 892 }, { "epoch": 0.16834762937128853, "grad_norm": 0.6484375, "learning_rate": 0.0008642241379310345, "loss": 2.5708, "step": 893 }, { "epoch": 0.16853614855311527, "grad_norm": 0.69140625, "learning_rate": 0.0008640282131661442, "loss": 2.6254, "step": 894 }, { "epoch": 0.16872466773494202, "grad_norm": 0.6796875, "learning_rate": 0.000863832288401254, "loss": 2.5748, "step": 895 }, { "epoch": 0.1689131869167688, "grad_norm": 0.6875, "learning_rate": 0.0008636363636363636, "loss": 2.7726, "step": 896 }, { "epoch": 0.16910170609859554, "grad_norm": 0.68359375, "learning_rate": 0.0008634404388714734, "loss": 2.6658, "step": 897 }, { "epoch": 0.16929022528042229, "grad_norm": 0.65234375, "learning_rate": 0.0008632445141065831, "loss": 2.6508, "step": 898 }, { "epoch": 0.16947874446224903, "grad_norm": 0.7421875, "learning_rate": 0.0008630485893416928, "loss": 2.6266, "step": 899 }, { "epoch": 0.16966726364407578, "grad_norm": 0.71875, "learning_rate": 0.0008628526645768026, "loss": 2.7459, "step": 900 }, { "epoch": 0.16985578282590252, "grad_norm": 0.68359375, "learning_rate": 0.0008626567398119123, "loss": 2.587, "step": 901 }, { "epoch": 0.1700443020077293, "grad_norm": 0.6484375, "learning_rate": 0.000862460815047022, "loss": 2.6079, "step": 902 }, { "epoch": 0.17023282118955604, "grad_norm": 0.7109375, "learning_rate": 0.0008622648902821317, "loss": 2.7354, "step": 903 }, { "epoch": 0.1704213403713828, "grad_norm": 0.76953125, "learning_rate": 0.0008620689655172414, "loss": 2.5455, "step": 904 }, { "epoch": 0.17060985955320954, "grad_norm": 0.66015625, "learning_rate": 0.000861873040752351, "loss": 2.6302, "step": 905 }, { "epoch": 0.17079837873503628, "grad_norm": 0.63671875, "learning_rate": 0.0008616771159874608, "loss": 2.4757, "step": 906 }, { "epoch": 0.17098689791686303, "grad_norm": 0.75, "learning_rate": 0.0008614811912225705, "loss": 2.6884, "step": 907 }, { "epoch": 0.1711754170986898, "grad_norm": 0.68359375, "learning_rate": 0.0008612852664576803, "loss": 2.5877, "step": 908 }, { "epoch": 0.17136393628051655, "grad_norm": 0.70703125, "learning_rate": 0.0008610893416927899, "loss": 2.6323, "step": 909 }, { "epoch": 0.1715524554623433, "grad_norm": 0.66796875, "learning_rate": 0.0008608934169278997, "loss": 2.657, "step": 910 }, { "epoch": 0.17174097464417004, "grad_norm": 0.671875, "learning_rate": 0.0008606974921630095, "loss": 2.6591, "step": 911 }, { "epoch": 0.1719294938259968, "grad_norm": 0.68359375, "learning_rate": 0.0008605015673981191, "loss": 2.6385, "step": 912 }, { "epoch": 0.17211801300782353, "grad_norm": 1.859375, "learning_rate": 0.0008603056426332289, "loss": 2.7853, "step": 913 }, { "epoch": 0.1723065321896503, "grad_norm": 0.70703125, "learning_rate": 0.0008601097178683386, "loss": 2.7299, "step": 914 }, { "epoch": 0.17249505137147705, "grad_norm": 0.68359375, "learning_rate": 0.0008599137931034483, "loss": 2.7032, "step": 915 }, { "epoch": 0.1726835705533038, "grad_norm": 0.68359375, "learning_rate": 0.000859717868338558, "loss": 2.6568, "step": 916 }, { "epoch": 0.17287208973513055, "grad_norm": 0.6640625, "learning_rate": 0.0008595219435736678, "loss": 2.8574, "step": 917 }, { "epoch": 0.1730606089169573, "grad_norm": 0.7109375, "learning_rate": 0.0008593260188087774, "loss": 2.7283, "step": 918 }, { "epoch": 0.17324912809878404, "grad_norm": 0.72265625, "learning_rate": 0.0008591300940438872, "loss": 2.712, "step": 919 }, { "epoch": 0.1734376472806108, "grad_norm": 0.67578125, "learning_rate": 0.000858934169278997, "loss": 2.5807, "step": 920 }, { "epoch": 0.17362616646243756, "grad_norm": 0.67578125, "learning_rate": 0.0008587382445141066, "loss": 2.6998, "step": 921 }, { "epoch": 0.1738146856442643, "grad_norm": 0.66796875, "learning_rate": 0.0008585423197492164, "loss": 2.5734, "step": 922 }, { "epoch": 0.17400320482609105, "grad_norm": 0.6953125, "learning_rate": 0.000858346394984326, "loss": 2.6374, "step": 923 }, { "epoch": 0.1741917240079178, "grad_norm": 0.640625, "learning_rate": 0.0008581504702194357, "loss": 2.5408, "step": 924 }, { "epoch": 0.17438024318974454, "grad_norm": 0.71484375, "learning_rate": 0.0008579545454545454, "loss": 2.661, "step": 925 }, { "epoch": 0.17456876237157132, "grad_norm": 0.68359375, "learning_rate": 0.0008577586206896552, "loss": 2.701, "step": 926 }, { "epoch": 0.17475728155339806, "grad_norm": 0.69140625, "learning_rate": 0.0008575626959247648, "loss": 2.71, "step": 927 }, { "epoch": 0.1749458007352248, "grad_norm": 0.671875, "learning_rate": 0.0008573667711598746, "loss": 2.6468, "step": 928 }, { "epoch": 0.17513431991705156, "grad_norm": 0.64453125, "learning_rate": 0.0008571708463949844, "loss": 2.5039, "step": 929 }, { "epoch": 0.1753228390988783, "grad_norm": 0.6875, "learning_rate": 0.000856974921630094, "loss": 2.6299, "step": 930 }, { "epoch": 0.17551135828070505, "grad_norm": 0.7109375, "learning_rate": 0.0008567789968652038, "loss": 2.7097, "step": 931 }, { "epoch": 0.17569987746253182, "grad_norm": 0.71875, "learning_rate": 0.0008565830721003135, "loss": 2.662, "step": 932 }, { "epoch": 0.17588839664435857, "grad_norm": 0.6875, "learning_rate": 0.0008563871473354232, "loss": 2.6554, "step": 933 }, { "epoch": 0.17607691582618532, "grad_norm": 0.69140625, "learning_rate": 0.0008561912225705329, "loss": 2.5157, "step": 934 }, { "epoch": 0.17626543500801206, "grad_norm": 0.703125, "learning_rate": 0.0008559952978056427, "loss": 2.658, "step": 935 }, { "epoch": 0.1764539541898388, "grad_norm": 0.6796875, "learning_rate": 0.0008557993730407524, "loss": 2.5697, "step": 936 }, { "epoch": 0.17664247337166555, "grad_norm": 0.6953125, "learning_rate": 0.0008556034482758621, "loss": 2.6467, "step": 937 }, { "epoch": 0.17683099255349233, "grad_norm": 0.671875, "learning_rate": 0.0008554075235109719, "loss": 2.7308, "step": 938 }, { "epoch": 0.17701951173531907, "grad_norm": 0.703125, "learning_rate": 0.0008552115987460816, "loss": 2.5552, "step": 939 }, { "epoch": 0.17720803091714582, "grad_norm": 0.70703125, "learning_rate": 0.0008550156739811913, "loss": 2.7391, "step": 940 }, { "epoch": 0.17739655009897257, "grad_norm": 0.6328125, "learning_rate": 0.000854819749216301, "loss": 2.5914, "step": 941 }, { "epoch": 0.1775850692807993, "grad_norm": 0.66796875, "learning_rate": 0.0008546238244514107, "loss": 2.6652, "step": 942 }, { "epoch": 0.17777358846262606, "grad_norm": 0.67578125, "learning_rate": 0.0008544278996865203, "loss": 2.7119, "step": 943 }, { "epoch": 0.17796210764445283, "grad_norm": 0.67578125, "learning_rate": 0.0008542319749216301, "loss": 2.7099, "step": 944 }, { "epoch": 0.17815062682627958, "grad_norm": 0.69921875, "learning_rate": 0.0008540360501567398, "loss": 2.6452, "step": 945 }, { "epoch": 0.17833914600810633, "grad_norm": 0.66796875, "learning_rate": 0.0008538401253918495, "loss": 2.5148, "step": 946 }, { "epoch": 0.17852766518993307, "grad_norm": 0.69140625, "learning_rate": 0.0008536442006269592, "loss": 2.6379, "step": 947 }, { "epoch": 0.17871618437175982, "grad_norm": 0.66796875, "learning_rate": 0.000853448275862069, "loss": 2.6155, "step": 948 }, { "epoch": 0.17890470355358656, "grad_norm": 0.68359375, "learning_rate": 0.0008532523510971787, "loss": 2.5122, "step": 949 }, { "epoch": 0.17909322273541334, "grad_norm": 0.66015625, "learning_rate": 0.0008530564263322884, "loss": 2.7232, "step": 950 }, { "epoch": 0.17928174191724008, "grad_norm": 0.703125, "learning_rate": 0.0008528605015673982, "loss": 2.7196, "step": 951 }, { "epoch": 0.17947026109906683, "grad_norm": 0.6484375, "learning_rate": 0.0008526645768025078, "loss": 2.5332, "step": 952 }, { "epoch": 0.17965878028089358, "grad_norm": 0.6484375, "learning_rate": 0.0008524686520376176, "loss": 2.7149, "step": 953 }, { "epoch": 0.17984729946272032, "grad_norm": 0.6640625, "learning_rate": 0.0008522727272727273, "loss": 2.7601, "step": 954 }, { "epoch": 0.1800358186445471, "grad_norm": 0.671875, "learning_rate": 0.000852076802507837, "loss": 2.6729, "step": 955 }, { "epoch": 0.18022433782637384, "grad_norm": 0.69140625, "learning_rate": 0.0008518808777429467, "loss": 2.5701, "step": 956 }, { "epoch": 0.1804128570082006, "grad_norm": 0.7109375, "learning_rate": 0.0008516849529780565, "loss": 2.6001, "step": 957 }, { "epoch": 0.18060137619002734, "grad_norm": 0.6796875, "learning_rate": 0.0008514890282131662, "loss": 2.7079, "step": 958 }, { "epoch": 0.18078989537185408, "grad_norm": 0.7109375, "learning_rate": 0.0008512931034482759, "loss": 2.6712, "step": 959 }, { "epoch": 0.18097841455368083, "grad_norm": 0.7265625, "learning_rate": 0.0008510971786833856, "loss": 2.722, "step": 960 }, { "epoch": 0.1811669337355076, "grad_norm": 0.65625, "learning_rate": 0.0008509012539184952, "loss": 2.5951, "step": 961 }, { "epoch": 0.18135545291733435, "grad_norm": 0.6484375, "learning_rate": 0.000850705329153605, "loss": 2.6927, "step": 962 }, { "epoch": 0.1815439720991611, "grad_norm": 0.68359375, "learning_rate": 0.0008505094043887147, "loss": 2.6872, "step": 963 }, { "epoch": 0.18173249128098784, "grad_norm": 0.671875, "learning_rate": 0.0008503134796238245, "loss": 2.498, "step": 964 }, { "epoch": 0.1819210104628146, "grad_norm": 0.69140625, "learning_rate": 0.0008501175548589341, "loss": 2.7254, "step": 965 }, { "epoch": 0.18210952964464133, "grad_norm": 0.67578125, "learning_rate": 0.0008499216300940439, "loss": 2.6044, "step": 966 }, { "epoch": 0.1822980488264681, "grad_norm": 0.63671875, "learning_rate": 0.0008497257053291537, "loss": 2.6202, "step": 967 }, { "epoch": 0.18248656800829485, "grad_norm": 0.7109375, "learning_rate": 0.0008495297805642633, "loss": 2.6671, "step": 968 }, { "epoch": 0.1826750871901216, "grad_norm": 0.6875, "learning_rate": 0.0008493338557993731, "loss": 2.7925, "step": 969 }, { "epoch": 0.18286360637194835, "grad_norm": 0.71875, "learning_rate": 0.0008491379310344828, "loss": 2.5893, "step": 970 }, { "epoch": 0.1830521255537751, "grad_norm": 0.6328125, "learning_rate": 0.0008489420062695925, "loss": 2.6704, "step": 971 }, { "epoch": 0.18324064473560184, "grad_norm": 0.68359375, "learning_rate": 0.0008487460815047022, "loss": 2.6642, "step": 972 }, { "epoch": 0.1834291639174286, "grad_norm": 0.72265625, "learning_rate": 0.000848550156739812, "loss": 2.6579, "step": 973 }, { "epoch": 0.18361768309925536, "grad_norm": 0.62890625, "learning_rate": 0.0008483542319749216, "loss": 2.47, "step": 974 }, { "epoch": 0.1838062022810821, "grad_norm": 0.66796875, "learning_rate": 0.0008481583072100314, "loss": 2.5976, "step": 975 }, { "epoch": 0.18399472146290885, "grad_norm": 0.67578125, "learning_rate": 0.0008479623824451412, "loss": 2.7708, "step": 976 }, { "epoch": 0.1841832406447356, "grad_norm": 0.68359375, "learning_rate": 0.0008477664576802508, "loss": 2.6823, "step": 977 }, { "epoch": 0.18437175982656234, "grad_norm": 0.640625, "learning_rate": 0.0008475705329153606, "loss": 2.6927, "step": 978 }, { "epoch": 0.18456027900838912, "grad_norm": 0.69921875, "learning_rate": 0.0008473746081504702, "loss": 2.682, "step": 979 }, { "epoch": 0.18474879819021586, "grad_norm": 0.6953125, "learning_rate": 0.0008471786833855799, "loss": 2.6509, "step": 980 }, { "epoch": 0.1849373173720426, "grad_norm": 0.66015625, "learning_rate": 0.0008469827586206896, "loss": 2.7625, "step": 981 }, { "epoch": 0.18512583655386936, "grad_norm": 0.6953125, "learning_rate": 0.0008467868338557994, "loss": 2.7144, "step": 982 }, { "epoch": 0.1853143557356961, "grad_norm": 0.69140625, "learning_rate": 0.000846590909090909, "loss": 2.7065, "step": 983 }, { "epoch": 0.18550287491752285, "grad_norm": 0.625, "learning_rate": 0.0008463949843260188, "loss": 2.5608, "step": 984 }, { "epoch": 0.18569139409934962, "grad_norm": 0.63671875, "learning_rate": 0.0008461990595611286, "loss": 2.587, "step": 985 }, { "epoch": 0.18587991328117637, "grad_norm": 0.69140625, "learning_rate": 0.0008460031347962382, "loss": 2.5719, "step": 986 }, { "epoch": 0.18606843246300311, "grad_norm": 0.68359375, "learning_rate": 0.000845807210031348, "loss": 2.6672, "step": 987 }, { "epoch": 0.18625695164482986, "grad_norm": 0.6640625, "learning_rate": 0.0008456112852664577, "loss": 2.6857, "step": 988 }, { "epoch": 0.1864454708266566, "grad_norm": 0.6796875, "learning_rate": 0.0008454153605015674, "loss": 2.6708, "step": 989 }, { "epoch": 0.18663399000848335, "grad_norm": 0.69921875, "learning_rate": 0.0008452194357366771, "loss": 2.6183, "step": 990 }, { "epoch": 0.18682250919031013, "grad_norm": 0.66015625, "learning_rate": 0.0008450235109717869, "loss": 2.6452, "step": 991 }, { "epoch": 0.18701102837213687, "grad_norm": 0.7421875, "learning_rate": 0.0008448275862068966, "loss": 2.704, "step": 992 }, { "epoch": 0.18719954755396362, "grad_norm": 0.68359375, "learning_rate": 0.0008446316614420063, "loss": 2.674, "step": 993 }, { "epoch": 0.18738806673579037, "grad_norm": 0.6953125, "learning_rate": 0.000844435736677116, "loss": 2.7137, "step": 994 }, { "epoch": 0.1875765859176171, "grad_norm": 0.66015625, "learning_rate": 0.0008442398119122258, "loss": 2.7582, "step": 995 }, { "epoch": 0.18776510509944386, "grad_norm": 0.69140625, "learning_rate": 0.0008440438871473355, "loss": 2.6009, "step": 996 }, { "epoch": 0.18795362428127063, "grad_norm": 0.71875, "learning_rate": 0.0008438479623824452, "loss": 2.6595, "step": 997 }, { "epoch": 0.18814214346309738, "grad_norm": 0.66015625, "learning_rate": 0.0008436520376175549, "loss": 2.552, "step": 998 }, { "epoch": 0.18833066264492412, "grad_norm": 0.68359375, "learning_rate": 0.0008434561128526645, "loss": 2.5806, "step": 999 }, { "epoch": 0.18851918182675087, "grad_norm": 0.66796875, "learning_rate": 0.0008432601880877743, "loss": 2.6469, "step": 1000 }, { "epoch": 0.18851918182675087, "eval_runtime": 16.219, "eval_samples_per_second": 63.136, "eval_steps_per_second": 1.973, "step": 1000 }, { "epoch": 0.18851918182675087, "eval/hellaswag_acc": 0.3743278231428002, "eval/hellaswag_acc_norm": 0.4706233817964549, "eval_hellaswag_elapsed_time": 116.27660393714905, "step": 1000 } ], "logging_steps": 1, "max_steps": 5304, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.325965577388032e+18, "train_batch_size": 12, "trial_name": null, "trial_params": null }