{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.17578125, "eval_steps": 500, "global_step": 1800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.765625e-05, "grad_norm": 13.125093460083008, "learning_rate": 1.6666666666666669e-06, "loss": 6.0473, "step": 1 }, { "epoch": 0.0001953125, "grad_norm": 13.109691619873047, "learning_rate": 3.3333333333333337e-06, "loss": 6.0151, "step": 2 }, { "epoch": 0.00029296875, "grad_norm": 12.975897789001465, "learning_rate": 5e-06, "loss": 6.0374, "step": 3 }, { "epoch": 0.000390625, "grad_norm": 12.861470222473145, "learning_rate": 6.6666666666666675e-06, "loss": 5.9534, "step": 4 }, { "epoch": 0.00048828125, "grad_norm": 11.551366806030273, "learning_rate": 8.333333333333334e-06, "loss": 5.7805, "step": 5 }, { "epoch": 0.0005859375, "grad_norm": 7.968364238739014, "learning_rate": 1e-05, "loss": 5.5664, "step": 6 }, { "epoch": 0.00068359375, "grad_norm": 6.854646682739258, "learning_rate": 1.1666666666666668e-05, "loss": 5.4982, "step": 7 }, { "epoch": 0.00078125, "grad_norm": 3.759798049926758, "learning_rate": 1.3333333333333335e-05, "loss": 5.1824, "step": 8 }, { "epoch": 0.00087890625, "grad_norm": 3.48822283744812, "learning_rate": 1.5e-05, "loss": 5.0667, "step": 9 }, { "epoch": 0.0009765625, "grad_norm": 3.1011886596679688, "learning_rate": 1.6666666666666667e-05, "loss": 5.0028, "step": 10 }, { "epoch": 0.00107421875, "grad_norm": 2.5868849754333496, "learning_rate": 1.8333333333333333e-05, "loss": 4.7787, "step": 11 }, { "epoch": 0.001171875, "grad_norm": 2.378497838973999, "learning_rate": 2e-05, "loss": 4.6855, "step": 12 }, { "epoch": 0.00126953125, "grad_norm": 2.2101738452911377, "learning_rate": 2.1666666666666667e-05, "loss": 4.6042, "step": 13 }, { "epoch": 0.0013671875, "grad_norm": 1.9938961267471313, "learning_rate": 2.3333333333333336e-05, "loss": 4.3843, "step": 14 }, { "epoch": 0.00146484375, "grad_norm": 1.9345463514328003, "learning_rate": 2.5e-05, "loss": 4.2548, "step": 15 }, { "epoch": 0.0015625, "grad_norm": 1.7730776071548462, "learning_rate": 2.666666666666667e-05, "loss": 4.1363, "step": 16 }, { "epoch": 0.00166015625, "grad_norm": 1.58334481716156, "learning_rate": 2.8333333333333332e-05, "loss": 3.9704, "step": 17 }, { "epoch": 0.0017578125, "grad_norm": 1.4827001094818115, "learning_rate": 3e-05, "loss": 3.81, "step": 18 }, { "epoch": 0.00185546875, "grad_norm": 1.4140430688858032, "learning_rate": 3.166666666666667e-05, "loss": 3.7305, "step": 19 }, { "epoch": 0.001953125, "grad_norm": 1.3934496641159058, "learning_rate": 3.3333333333333335e-05, "loss": 3.5375, "step": 20 }, { "epoch": 0.00205078125, "grad_norm": 1.3384771347045898, "learning_rate": 3.5000000000000004e-05, "loss": 3.5217, "step": 21 }, { "epoch": 0.0021484375, "grad_norm": 1.2228978872299194, "learning_rate": 3.6666666666666666e-05, "loss": 3.3742, "step": 22 }, { "epoch": 0.00224609375, "grad_norm": 1.1775028705596924, "learning_rate": 3.8333333333333334e-05, "loss": 3.2626, "step": 23 }, { "epoch": 0.00234375, "grad_norm": 1.0886842012405396, "learning_rate": 4e-05, "loss": 3.1864, "step": 24 }, { "epoch": 0.00244140625, "grad_norm": 1.0981535911560059, "learning_rate": 4.1666666666666665e-05, "loss": 3.0962, "step": 25 }, { "epoch": 0.0025390625, "grad_norm": 1.009294867515564, "learning_rate": 4.3333333333333334e-05, "loss": 3.0507, "step": 26 }, { "epoch": 0.00263671875, "grad_norm": 0.9390103816986084, "learning_rate": 4.4999999999999996e-05, "loss": 2.9579, "step": 27 }, { "epoch": 0.002734375, "grad_norm": 0.8647847175598145, "learning_rate": 4.666666666666667e-05, "loss": 2.8166, "step": 28 }, { "epoch": 0.00283203125, "grad_norm": 0.8606237769126892, "learning_rate": 4.8333333333333334e-05, "loss": 2.7919, "step": 29 }, { "epoch": 0.0029296875, "grad_norm": 0.8069576025009155, "learning_rate": 5e-05, "loss": 2.7426, "step": 30 }, { "epoch": 0.00302734375, "grad_norm": 0.7007808685302734, "learning_rate": 5.1666666666666664e-05, "loss": 2.6926, "step": 31 }, { "epoch": 0.003125, "grad_norm": 0.6825646162033081, "learning_rate": 5.333333333333334e-05, "loss": 2.6783, "step": 32 }, { "epoch": 0.00322265625, "grad_norm": 0.7054751515388489, "learning_rate": 5.5e-05, "loss": 2.6239, "step": 33 }, { "epoch": 0.0033203125, "grad_norm": 0.7101700305938721, "learning_rate": 5.6666666666666664e-05, "loss": 2.5993, "step": 34 }, { "epoch": 0.00341796875, "grad_norm": 0.56828773021698, "learning_rate": 5.833333333333333e-05, "loss": 2.5258, "step": 35 }, { "epoch": 0.003515625, "grad_norm": 0.5513983368873596, "learning_rate": 6e-05, "loss": 2.5018, "step": 36 }, { "epoch": 0.00361328125, "grad_norm": 0.573391318321228, "learning_rate": 6.166666666666667e-05, "loss": 2.4881, "step": 37 }, { "epoch": 0.0037109375, "grad_norm": 0.4520920515060425, "learning_rate": 6.333333333333335e-05, "loss": 2.4489, "step": 38 }, { "epoch": 0.00380859375, "grad_norm": 0.43038809299468994, "learning_rate": 6.500000000000001e-05, "loss": 2.4387, "step": 39 }, { "epoch": 0.00390625, "grad_norm": 0.4071808457374573, "learning_rate": 6.666666666666667e-05, "loss": 2.4261, "step": 40 }, { "epoch": 0.00400390625, "grad_norm": 0.3631410300731659, "learning_rate": 6.833333333333333e-05, "loss": 2.3656, "step": 41 }, { "epoch": 0.0041015625, "grad_norm": 0.33457818627357483, "learning_rate": 7.000000000000001e-05, "loss": 2.4011, "step": 42 }, { "epoch": 0.00419921875, "grad_norm": 0.3582305610179901, "learning_rate": 7.166666666666667e-05, "loss": 2.3662, "step": 43 }, { "epoch": 0.004296875, "grad_norm": 0.4090467691421509, "learning_rate": 7.333333333333333e-05, "loss": 2.3282, "step": 44 }, { "epoch": 0.00439453125, "grad_norm": 0.34438556432724, "learning_rate": 7.5e-05, "loss": 2.2991, "step": 45 }, { "epoch": 0.0044921875, "grad_norm": 0.29381102323532104, "learning_rate": 7.666666666666667e-05, "loss": 2.2506, "step": 46 }, { "epoch": 0.00458984375, "grad_norm": 0.4019562304019928, "learning_rate": 7.833333333333334e-05, "loss": 2.2494, "step": 47 }, { "epoch": 0.0046875, "grad_norm": 0.3183911144733429, "learning_rate": 8e-05, "loss": 2.2853, "step": 48 }, { "epoch": 0.00478515625, "grad_norm": 0.33041393756866455, "learning_rate": 8.166666666666667e-05, "loss": 2.2251, "step": 49 }, { "epoch": 0.0048828125, "grad_norm": 0.28936123847961426, "learning_rate": 8.333333333333333e-05, "loss": 2.2308, "step": 50 }, { "epoch": 0.00498046875, "grad_norm": 0.36125779151916504, "learning_rate": 8.5e-05, "loss": 2.2473, "step": 51 }, { "epoch": 0.005078125, "grad_norm": 0.2832348942756653, "learning_rate": 8.666666666666667e-05, "loss": 2.201, "step": 52 }, { "epoch": 0.00517578125, "grad_norm": 0.3250133693218231, "learning_rate": 8.833333333333333e-05, "loss": 2.1596, "step": 53 }, { "epoch": 0.0052734375, "grad_norm": 0.31502828001976013, "learning_rate": 8.999999999999999e-05, "loss": 2.2183, "step": 54 }, { "epoch": 0.00537109375, "grad_norm": 0.33068645000457764, "learning_rate": 9.166666666666667e-05, "loss": 2.2254, "step": 55 }, { "epoch": 0.00546875, "grad_norm": 0.30684661865234375, "learning_rate": 9.333333333333334e-05, "loss": 2.2572, "step": 56 }, { "epoch": 0.00556640625, "grad_norm": 0.31540846824645996, "learning_rate": 9.5e-05, "loss": 2.1594, "step": 57 }, { "epoch": 0.0056640625, "grad_norm": 0.23259752988815308, "learning_rate": 9.666666666666667e-05, "loss": 2.172, "step": 58 }, { "epoch": 0.00576171875, "grad_norm": 0.26666638255119324, "learning_rate": 9.833333333333333e-05, "loss": 2.1961, "step": 59 }, { "epoch": 0.005859375, "grad_norm": 0.26295527815818787, "learning_rate": 0.0001, "loss": 2.1811, "step": 60 }, { "epoch": 0.00595703125, "grad_norm": 0.2302207201719284, "learning_rate": 0.00010166666666666667, "loss": 2.1431, "step": 61 }, { "epoch": 0.0060546875, "grad_norm": 0.468537837266922, "learning_rate": 0.00010333333333333333, "loss": 2.133, "step": 62 }, { "epoch": 0.00615234375, "grad_norm": 0.27946797013282776, "learning_rate": 0.000105, "loss": 2.1917, "step": 63 }, { "epoch": 0.00625, "grad_norm": 0.26789286732673645, "learning_rate": 0.00010666666666666668, "loss": 2.1417, "step": 64 }, { "epoch": 0.00634765625, "grad_norm": 0.22977805137634277, "learning_rate": 0.00010833333333333334, "loss": 2.1371, "step": 65 }, { "epoch": 0.0064453125, "grad_norm": 0.2590682804584503, "learning_rate": 0.00011, "loss": 2.113, "step": 66 }, { "epoch": 0.00654296875, "grad_norm": 0.2660805583000183, "learning_rate": 0.00011166666666666667, "loss": 2.1346, "step": 67 }, { "epoch": 0.006640625, "grad_norm": 0.24538874626159668, "learning_rate": 0.00011333333333333333, "loss": 2.1078, "step": 68 }, { "epoch": 0.00673828125, "grad_norm": 0.2986001968383789, "learning_rate": 0.000115, "loss": 2.1048, "step": 69 }, { "epoch": 0.0068359375, "grad_norm": 0.30193910002708435, "learning_rate": 0.00011666666666666667, "loss": 2.135, "step": 70 }, { "epoch": 0.00693359375, "grad_norm": 0.25153177976608276, "learning_rate": 0.00011833333333333334, "loss": 2.1101, "step": 71 }, { "epoch": 0.00703125, "grad_norm": 0.7685809135437012, "learning_rate": 0.00012, "loss": 2.0833, "step": 72 }, { "epoch": 0.00712890625, "grad_norm": 0.24468126893043518, "learning_rate": 0.00012166666666666668, "loss": 2.1409, "step": 73 }, { "epoch": 0.0072265625, "grad_norm": 0.3070752024650574, "learning_rate": 0.00012333333333333334, "loss": 2.1402, "step": 74 }, { "epoch": 0.00732421875, "grad_norm": 0.352273553609848, "learning_rate": 0.000125, "loss": 2.171, "step": 75 }, { "epoch": 0.007421875, "grad_norm": 0.34806111454963684, "learning_rate": 0.0001266666666666667, "loss": 2.1155, "step": 76 }, { "epoch": 0.00751953125, "grad_norm": 0.317414253950119, "learning_rate": 0.00012833333333333333, "loss": 2.0812, "step": 77 }, { "epoch": 0.0076171875, "grad_norm": 0.3320539891719818, "learning_rate": 0.00013000000000000002, "loss": 2.0987, "step": 78 }, { "epoch": 0.00771484375, "grad_norm": 0.3315901756286621, "learning_rate": 0.00013166666666666665, "loss": 2.1025, "step": 79 }, { "epoch": 0.0078125, "grad_norm": 0.29021286964416504, "learning_rate": 0.00013333333333333334, "loss": 2.1474, "step": 80 }, { "epoch": 0.00791015625, "grad_norm": 0.2593044936656952, "learning_rate": 0.000135, "loss": 2.0667, "step": 81 }, { "epoch": 0.0080078125, "grad_norm": 0.3192054033279419, "learning_rate": 0.00013666666666666666, "loss": 2.0764, "step": 82 }, { "epoch": 0.00810546875, "grad_norm": 0.26053664088249207, "learning_rate": 0.00013833333333333333, "loss": 2.0698, "step": 83 }, { "epoch": 0.008203125, "grad_norm": 0.24790963530540466, "learning_rate": 0.00014000000000000001, "loss": 2.124, "step": 84 }, { "epoch": 0.00830078125, "grad_norm": 0.2546316087245941, "learning_rate": 0.00014166666666666668, "loss": 2.1025, "step": 85 }, { "epoch": 0.0083984375, "grad_norm": 0.22275258600711823, "learning_rate": 0.00014333333333333334, "loss": 2.0778, "step": 86 }, { "epoch": 0.00849609375, "grad_norm": 0.22855599224567413, "learning_rate": 0.000145, "loss": 2.1113, "step": 87 }, { "epoch": 0.00859375, "grad_norm": 0.2456902116537094, "learning_rate": 0.00014666666666666666, "loss": 2.1399, "step": 88 }, { "epoch": 0.00869140625, "grad_norm": 0.22144420444965363, "learning_rate": 0.00014833333333333335, "loss": 2.0211, "step": 89 }, { "epoch": 0.0087890625, "grad_norm": 0.254894495010376, "learning_rate": 0.00015, "loss": 2.1382, "step": 90 }, { "epoch": 0.00888671875, "grad_norm": 0.2729082703590393, "learning_rate": 0.00015166666666666668, "loss": 2.1271, "step": 91 }, { "epoch": 0.008984375, "grad_norm": 0.2564642131328583, "learning_rate": 0.00015333333333333334, "loss": 2.0423, "step": 92 }, { "epoch": 0.00908203125, "grad_norm": 0.24927419424057007, "learning_rate": 0.000155, "loss": 2.0603, "step": 93 }, { "epoch": 0.0091796875, "grad_norm": 0.2591923475265503, "learning_rate": 0.0001566666666666667, "loss": 2.0522, "step": 94 }, { "epoch": 0.00927734375, "grad_norm": 0.3350547254085541, "learning_rate": 0.00015833333333333332, "loss": 2.1095, "step": 95 }, { "epoch": 0.009375, "grad_norm": 0.30227839946746826, "learning_rate": 0.00016, "loss": 2.121, "step": 96 }, { "epoch": 0.00947265625, "grad_norm": 0.3027198314666748, "learning_rate": 0.00016166666666666665, "loss": 2.0288, "step": 97 }, { "epoch": 0.0095703125, "grad_norm": 0.29791951179504395, "learning_rate": 0.00016333333333333334, "loss": 2.0471, "step": 98 }, { "epoch": 0.00966796875, "grad_norm": 0.30918803811073303, "learning_rate": 0.000165, "loss": 2.1036, "step": 99 }, { "epoch": 0.009765625, "grad_norm": 0.2510216534137726, "learning_rate": 0.00016666666666666666, "loss": 2.0217, "step": 100 }, { "epoch": 0.00986328125, "grad_norm": 0.288231760263443, "learning_rate": 0.00016833333333333335, "loss": 2.0178, "step": 101 }, { "epoch": 0.0099609375, "grad_norm": 0.3326691687107086, "learning_rate": 0.00017, "loss": 2.0947, "step": 102 }, { "epoch": 0.01005859375, "grad_norm": 0.3552946150302887, "learning_rate": 0.00017166666666666667, "loss": 1.9953, "step": 103 }, { "epoch": 0.01015625, "grad_norm": 0.34882935881614685, "learning_rate": 0.00017333333333333334, "loss": 2.1223, "step": 104 }, { "epoch": 0.01025390625, "grad_norm": 0.35487979650497437, "learning_rate": 0.000175, "loss": 2.0599, "step": 105 }, { "epoch": 0.0103515625, "grad_norm": 0.33561578392982483, "learning_rate": 0.00017666666666666666, "loss": 2.0398, "step": 106 }, { "epoch": 0.01044921875, "grad_norm": 0.27415409684181213, "learning_rate": 0.00017833333333333335, "loss": 2.0982, "step": 107 }, { "epoch": 0.010546875, "grad_norm": 0.28592920303344727, "learning_rate": 0.00017999999999999998, "loss": 2.0571, "step": 108 }, { "epoch": 0.01064453125, "grad_norm": 0.3228552043437958, "learning_rate": 0.00018166666666666667, "loss": 1.9923, "step": 109 }, { "epoch": 0.0107421875, "grad_norm": 0.29005661606788635, "learning_rate": 0.00018333333333333334, "loss": 2.0394, "step": 110 }, { "epoch": 0.01083984375, "grad_norm": 0.36677825450897217, "learning_rate": 0.000185, "loss": 2.0153, "step": 111 }, { "epoch": 0.0109375, "grad_norm": 0.2562806308269501, "learning_rate": 0.0001866666666666667, "loss": 2.05, "step": 112 }, { "epoch": 0.01103515625, "grad_norm": 0.2748093008995056, "learning_rate": 0.00018833333333333332, "loss": 2.0449, "step": 113 }, { "epoch": 0.0111328125, "grad_norm": 0.3924459218978882, "learning_rate": 0.00019, "loss": 2.0409, "step": 114 }, { "epoch": 0.01123046875, "grad_norm": 0.4368191063404083, "learning_rate": 0.00019166666666666667, "loss": 2.0178, "step": 115 }, { "epoch": 0.011328125, "grad_norm": 0.4236035943031311, "learning_rate": 0.00019333333333333333, "loss": 2.0518, "step": 116 }, { "epoch": 0.01142578125, "grad_norm": 0.28912851214408875, "learning_rate": 0.00019500000000000002, "loss": 2.0058, "step": 117 }, { "epoch": 0.0115234375, "grad_norm": 0.38847798109054565, "learning_rate": 0.00019666666666666666, "loss": 2.0818, "step": 118 }, { "epoch": 0.01162109375, "grad_norm": 0.29429811239242554, "learning_rate": 0.00019833333333333335, "loss": 2.004, "step": 119 }, { "epoch": 0.01171875, "grad_norm": 0.2969271242618561, "learning_rate": 0.0002, "loss": 2.046, "step": 120 }, { "epoch": 0.01181640625, "grad_norm": 0.297894150018692, "learning_rate": 0.00020166666666666667, "loss": 2.0204, "step": 121 }, { "epoch": 0.0119140625, "grad_norm": 0.28794237971305847, "learning_rate": 0.00020333333333333333, "loss": 2.0663, "step": 122 }, { "epoch": 0.01201171875, "grad_norm": 0.2766033411026001, "learning_rate": 0.000205, "loss": 2.0399, "step": 123 }, { "epoch": 0.012109375, "grad_norm": 0.3516612648963928, "learning_rate": 0.00020666666666666666, "loss": 2.0389, "step": 124 }, { "epoch": 0.01220703125, "grad_norm": 0.28204381465911865, "learning_rate": 0.00020833333333333335, "loss": 2.0084, "step": 125 }, { "epoch": 0.0123046875, "grad_norm": 0.3306240737438202, "learning_rate": 0.00021, "loss": 2.013, "step": 126 }, { "epoch": 0.01240234375, "grad_norm": 0.3221111595630646, "learning_rate": 0.00021166666666666667, "loss": 2.0312, "step": 127 }, { "epoch": 0.0125, "grad_norm": 0.2742249071598053, "learning_rate": 0.00021333333333333336, "loss": 2.0375, "step": 128 }, { "epoch": 0.01259765625, "grad_norm": 0.2785228490829468, "learning_rate": 0.000215, "loss": 2.0449, "step": 129 }, { "epoch": 0.0126953125, "grad_norm": 0.2666397988796234, "learning_rate": 0.00021666666666666668, "loss": 2.0355, "step": 130 }, { "epoch": 0.01279296875, "grad_norm": 0.22522135078907013, "learning_rate": 0.00021833333333333332, "loss": 1.9693, "step": 131 }, { "epoch": 0.012890625, "grad_norm": 0.2724483013153076, "learning_rate": 0.00022, "loss": 2.0671, "step": 132 }, { "epoch": 0.01298828125, "grad_norm": 0.2655040919780731, "learning_rate": 0.00022166666666666667, "loss": 1.9805, "step": 133 }, { "epoch": 0.0130859375, "grad_norm": 0.3194504678249359, "learning_rate": 0.00022333333333333333, "loss": 2.067, "step": 134 }, { "epoch": 0.01318359375, "grad_norm": 0.25511813163757324, "learning_rate": 0.00022500000000000002, "loss": 2.0141, "step": 135 }, { "epoch": 0.01328125, "grad_norm": 0.3995087444782257, "learning_rate": 0.00022666666666666666, "loss": 2.0278, "step": 136 }, { "epoch": 0.01337890625, "grad_norm": 0.2623380422592163, "learning_rate": 0.00022833333333333334, "loss": 1.9751, "step": 137 }, { "epoch": 0.0134765625, "grad_norm": 0.23814889788627625, "learning_rate": 0.00023, "loss": 1.969, "step": 138 }, { "epoch": 0.01357421875, "grad_norm": 0.23566491901874542, "learning_rate": 0.00023166666666666667, "loss": 2.0135, "step": 139 }, { "epoch": 0.013671875, "grad_norm": 0.2437373697757721, "learning_rate": 0.00023333333333333333, "loss": 2.0044, "step": 140 }, { "epoch": 0.01376953125, "grad_norm": 0.2861543595790863, "learning_rate": 0.000235, "loss": 2.0378, "step": 141 }, { "epoch": 0.0138671875, "grad_norm": 0.318050354719162, "learning_rate": 0.00023666666666666668, "loss": 2.0857, "step": 142 }, { "epoch": 0.01396484375, "grad_norm": 0.39669227600097656, "learning_rate": 0.00023833333333333334, "loss": 2.0535, "step": 143 }, { "epoch": 0.0140625, "grad_norm": 0.4359401762485504, "learning_rate": 0.00024, "loss": 2.0432, "step": 144 }, { "epoch": 0.01416015625, "grad_norm": 0.4532039165496826, "learning_rate": 0.00024166666666666667, "loss": 1.9909, "step": 145 }, { "epoch": 0.0142578125, "grad_norm": 0.4570695161819458, "learning_rate": 0.00024333333333333336, "loss": 2.0123, "step": 146 }, { "epoch": 0.01435546875, "grad_norm": 0.36623403429985046, "learning_rate": 0.000245, "loss": 2.0179, "step": 147 }, { "epoch": 0.014453125, "grad_norm": 0.3069714307785034, "learning_rate": 0.0002466666666666667, "loss": 2.0014, "step": 148 }, { "epoch": 0.01455078125, "grad_norm": 0.3980304002761841, "learning_rate": 0.0002483333333333333, "loss": 2.0489, "step": 149 }, { "epoch": 0.0146484375, "grad_norm": 0.31907564401626587, "learning_rate": 0.00025, "loss": 2.013, "step": 150 }, { "epoch": 0.01474609375, "grad_norm": 0.2952549159526825, "learning_rate": 0.00025166666666666664, "loss": 2.0709, "step": 151 }, { "epoch": 0.01484375, "grad_norm": 0.29451197385787964, "learning_rate": 0.0002533333333333334, "loss": 1.9613, "step": 152 }, { "epoch": 0.01494140625, "grad_norm": 0.2893507778644562, "learning_rate": 0.000255, "loss": 2.0048, "step": 153 }, { "epoch": 0.0150390625, "grad_norm": 0.24850639700889587, "learning_rate": 0.00025666666666666665, "loss": 2.0198, "step": 154 }, { "epoch": 0.01513671875, "grad_norm": 0.24297639727592468, "learning_rate": 0.00025833333333333334, "loss": 2.0561, "step": 155 }, { "epoch": 0.015234375, "grad_norm": 0.2777438461780548, "learning_rate": 0.00026000000000000003, "loss": 2.0206, "step": 156 }, { "epoch": 0.01533203125, "grad_norm": 0.28714093565940857, "learning_rate": 0.00026166666666666667, "loss": 1.9891, "step": 157 }, { "epoch": 0.0154296875, "grad_norm": 0.25796255469322205, "learning_rate": 0.0002633333333333333, "loss": 2.0369, "step": 158 }, { "epoch": 0.01552734375, "grad_norm": 0.2391008883714676, "learning_rate": 0.00026500000000000004, "loss": 2.0015, "step": 159 }, { "epoch": 0.015625, "grad_norm": 0.3203892409801483, "learning_rate": 0.0002666666666666667, "loss": 2.0213, "step": 160 }, { "epoch": 0.01572265625, "grad_norm": 0.3396870791912079, "learning_rate": 0.0002683333333333333, "loss": 2.0333, "step": 161 }, { "epoch": 0.0158203125, "grad_norm": 0.315060555934906, "learning_rate": 0.00027, "loss": 2.0195, "step": 162 }, { "epoch": 0.01591796875, "grad_norm": 0.2672436535358429, "learning_rate": 0.0002716666666666667, "loss": 1.9946, "step": 163 }, { "epoch": 0.016015625, "grad_norm": 0.2996402382850647, "learning_rate": 0.00027333333333333333, "loss": 2.0112, "step": 164 }, { "epoch": 0.01611328125, "grad_norm": 0.2894189953804016, "learning_rate": 0.000275, "loss": 2.0157, "step": 165 }, { "epoch": 0.0162109375, "grad_norm": 0.26241254806518555, "learning_rate": 0.00027666666666666665, "loss": 2.0177, "step": 166 }, { "epoch": 0.01630859375, "grad_norm": 0.22900305688381195, "learning_rate": 0.00027833333333333334, "loss": 1.9834, "step": 167 }, { "epoch": 0.01640625, "grad_norm": 0.2373427301645279, "learning_rate": 0.00028000000000000003, "loss": 1.9792, "step": 168 }, { "epoch": 0.01650390625, "grad_norm": 0.2663004696369171, "learning_rate": 0.00028166666666666666, "loss": 2.0203, "step": 169 }, { "epoch": 0.0166015625, "grad_norm": 0.31653544306755066, "learning_rate": 0.00028333333333333335, "loss": 2.0216, "step": 170 }, { "epoch": 0.01669921875, "grad_norm": 0.3077234923839569, "learning_rate": 0.000285, "loss": 2.0361, "step": 171 }, { "epoch": 0.016796875, "grad_norm": 0.25555703043937683, "learning_rate": 0.0002866666666666667, "loss": 2.0102, "step": 172 }, { "epoch": 0.01689453125, "grad_norm": 0.29817435145378113, "learning_rate": 0.0002883333333333333, "loss": 1.972, "step": 173 }, { "epoch": 0.0169921875, "grad_norm": 0.3075692355632782, "learning_rate": 0.00029, "loss": 2.0195, "step": 174 }, { "epoch": 0.01708984375, "grad_norm": 0.29917964339256287, "learning_rate": 0.0002916666666666667, "loss": 1.9972, "step": 175 }, { "epoch": 0.0171875, "grad_norm": 0.32018229365348816, "learning_rate": 0.0002933333333333333, "loss": 1.9895, "step": 176 }, { "epoch": 0.01728515625, "grad_norm": 0.2907097339630127, "learning_rate": 0.000295, "loss": 1.9777, "step": 177 }, { "epoch": 0.0173828125, "grad_norm": 0.22390642762184143, "learning_rate": 0.0002966666666666667, "loss": 1.9961, "step": 178 }, { "epoch": 0.01748046875, "grad_norm": 0.25350186228752136, "learning_rate": 0.00029833333333333334, "loss": 1.9875, "step": 179 }, { "epoch": 0.017578125, "grad_norm": 0.22856706380844116, "learning_rate": 0.0003, "loss": 2.0169, "step": 180 }, { "epoch": 0.01767578125, "grad_norm": 0.2288493812084198, "learning_rate": 0.0003016666666666667, "loss": 2.0238, "step": 181 }, { "epoch": 0.0177734375, "grad_norm": 0.27326855063438416, "learning_rate": 0.00030333333333333335, "loss": 2.0134, "step": 182 }, { "epoch": 0.01787109375, "grad_norm": 0.2447524517774582, "learning_rate": 0.000305, "loss": 1.9852, "step": 183 }, { "epoch": 0.01796875, "grad_norm": 0.4363366663455963, "learning_rate": 0.0003066666666666667, "loss": 1.9921, "step": 184 }, { "epoch": 0.01806640625, "grad_norm": 0.4569666385650635, "learning_rate": 0.00030833333333333337, "loss": 1.9997, "step": 185 }, { "epoch": 0.0181640625, "grad_norm": 0.43348655104637146, "learning_rate": 0.00031, "loss": 2.0584, "step": 186 }, { "epoch": 0.01826171875, "grad_norm": 0.3844921588897705, "learning_rate": 0.00031166666666666663, "loss": 2.0035, "step": 187 }, { "epoch": 0.018359375, "grad_norm": 0.3427641987800598, "learning_rate": 0.0003133333333333334, "loss": 1.9536, "step": 188 }, { "epoch": 0.01845703125, "grad_norm": 0.33557865023612976, "learning_rate": 0.000315, "loss": 1.972, "step": 189 }, { "epoch": 0.0185546875, "grad_norm": 0.4006612300872803, "learning_rate": 0.00031666666666666665, "loss": 2.0652, "step": 190 }, { "epoch": 0.01865234375, "grad_norm": 0.3158099055290222, "learning_rate": 0.00031833333333333334, "loss": 2.0516, "step": 191 }, { "epoch": 0.01875, "grad_norm": 0.3799190819263458, "learning_rate": 0.00032, "loss": 2.011, "step": 192 }, { "epoch": 0.01884765625, "grad_norm": 0.2948876619338989, "learning_rate": 0.00032166666666666666, "loss": 2.0109, "step": 193 }, { "epoch": 0.0189453125, "grad_norm": 0.24561335146427155, "learning_rate": 0.0003233333333333333, "loss": 2.0264, "step": 194 }, { "epoch": 0.01904296875, "grad_norm": 0.24896866083145142, "learning_rate": 0.00032500000000000004, "loss": 1.9726, "step": 195 }, { "epoch": 0.019140625, "grad_norm": 0.26887547969818115, "learning_rate": 0.0003266666666666667, "loss": 2.0036, "step": 196 }, { "epoch": 0.01923828125, "grad_norm": 0.3186735212802887, "learning_rate": 0.0003283333333333333, "loss": 2.0174, "step": 197 }, { "epoch": 0.0193359375, "grad_norm": 0.3317165672779083, "learning_rate": 0.00033, "loss": 2.0047, "step": 198 }, { "epoch": 0.01943359375, "grad_norm": 0.3068574070930481, "learning_rate": 0.0003316666666666667, "loss": 2.0032, "step": 199 }, { "epoch": 0.01953125, "grad_norm": 0.29292526841163635, "learning_rate": 0.0003333333333333333, "loss": 2.0533, "step": 200 }, { "epoch": 0.01962890625, "grad_norm": 0.2519834041595459, "learning_rate": 0.000335, "loss": 2.0113, "step": 201 }, { "epoch": 0.0197265625, "grad_norm": 0.25766584277153015, "learning_rate": 0.0003366666666666667, "loss": 2.0278, "step": 202 }, { "epoch": 0.01982421875, "grad_norm": 0.2704983055591583, "learning_rate": 0.00033833333333333334, "loss": 1.9725, "step": 203 }, { "epoch": 0.019921875, "grad_norm": 0.2882053256034851, "learning_rate": 0.00034, "loss": 2.0706, "step": 204 }, { "epoch": 0.02001953125, "grad_norm": 0.34524375200271606, "learning_rate": 0.00034166666666666666, "loss": 1.9868, "step": 205 }, { "epoch": 0.0201171875, "grad_norm": 0.3718552887439728, "learning_rate": 0.00034333333333333335, "loss": 1.9596, "step": 206 }, { "epoch": 0.02021484375, "grad_norm": 0.2747247815132141, "learning_rate": 0.000345, "loss": 2.0125, "step": 207 }, { "epoch": 0.0203125, "grad_norm": 0.3062858283519745, "learning_rate": 0.00034666666666666667, "loss": 2.0224, "step": 208 }, { "epoch": 0.02041015625, "grad_norm": 0.32505863904953003, "learning_rate": 0.00034833333333333336, "loss": 2.0376, "step": 209 }, { "epoch": 0.0205078125, "grad_norm": 0.35048386454582214, "learning_rate": 0.00035, "loss": 2.0268, "step": 210 }, { "epoch": 0.02060546875, "grad_norm": 0.31204426288604736, "learning_rate": 0.0003516666666666667, "loss": 2.0198, "step": 211 }, { "epoch": 0.020703125, "grad_norm": 0.24253524839878082, "learning_rate": 0.0003533333333333333, "loss": 2.028, "step": 212 }, { "epoch": 0.02080078125, "grad_norm": 0.286915123462677, "learning_rate": 0.000355, "loss": 1.9358, "step": 213 }, { "epoch": 0.0208984375, "grad_norm": 0.2800680994987488, "learning_rate": 0.0003566666666666667, "loss": 1.9905, "step": 214 }, { "epoch": 0.02099609375, "grad_norm": 0.2718358635902405, "learning_rate": 0.00035833333333333333, "loss": 1.9902, "step": 215 }, { "epoch": 0.02109375, "grad_norm": 0.28583604097366333, "learning_rate": 0.00035999999999999997, "loss": 1.991, "step": 216 }, { "epoch": 0.02119140625, "grad_norm": 0.2911478281021118, "learning_rate": 0.0003616666666666667, "loss": 1.973, "step": 217 }, { "epoch": 0.0212890625, "grad_norm": 0.3601188361644745, "learning_rate": 0.00036333333333333335, "loss": 1.9727, "step": 218 }, { "epoch": 0.02138671875, "grad_norm": 0.2888337969779968, "learning_rate": 0.000365, "loss": 1.988, "step": 219 }, { "epoch": 0.021484375, "grad_norm": 0.25628700852394104, "learning_rate": 0.00036666666666666667, "loss": 1.9984, "step": 220 }, { "epoch": 0.02158203125, "grad_norm": 0.2637641429901123, "learning_rate": 0.00036833333333333336, "loss": 2.0229, "step": 221 }, { "epoch": 0.0216796875, "grad_norm": 0.23845899105072021, "learning_rate": 0.00037, "loss": 1.9985, "step": 222 }, { "epoch": 0.02177734375, "grad_norm": 0.28519535064697266, "learning_rate": 0.00037166666666666663, "loss": 2.0061, "step": 223 }, { "epoch": 0.021875, "grad_norm": 0.31845173239707947, "learning_rate": 0.0003733333333333334, "loss": 2.0081, "step": 224 }, { "epoch": 0.02197265625, "grad_norm": 0.3725838363170624, "learning_rate": 0.000375, "loss": 2.0032, "step": 225 }, { "epoch": 0.0220703125, "grad_norm": 0.49783870577812195, "learning_rate": 0.00037666666666666664, "loss": 2.0404, "step": 226 }, { "epoch": 0.02216796875, "grad_norm": 0.5059479475021362, "learning_rate": 0.0003783333333333334, "loss": 2.0498, "step": 227 }, { "epoch": 0.022265625, "grad_norm": 0.461291640996933, "learning_rate": 0.00038, "loss": 2.0078, "step": 228 }, { "epoch": 0.02236328125, "grad_norm": 0.3970203399658203, "learning_rate": 0.00038166666666666666, "loss": 1.9966, "step": 229 }, { "epoch": 0.0224609375, "grad_norm": 0.3155679404735565, "learning_rate": 0.00038333333333333334, "loss": 2.0263, "step": 230 }, { "epoch": 0.02255859375, "grad_norm": 0.32979920506477356, "learning_rate": 0.00038500000000000003, "loss": 1.9967, "step": 231 }, { "epoch": 0.02265625, "grad_norm": 0.27470117807388306, "learning_rate": 0.00038666666666666667, "loss": 2.0461, "step": 232 }, { "epoch": 0.02275390625, "grad_norm": 0.2981088161468506, "learning_rate": 0.0003883333333333333, "loss": 1.9944, "step": 233 }, { "epoch": 0.0228515625, "grad_norm": 0.3496599495410919, "learning_rate": 0.00039000000000000005, "loss": 1.9907, "step": 234 }, { "epoch": 0.02294921875, "grad_norm": 0.3318106532096863, "learning_rate": 0.0003916666666666667, "loss": 2.0576, "step": 235 }, { "epoch": 0.023046875, "grad_norm": 0.29498377442359924, "learning_rate": 0.0003933333333333333, "loss": 2.0242, "step": 236 }, { "epoch": 0.02314453125, "grad_norm": 0.2970214784145355, "learning_rate": 0.000395, "loss": 2.0087, "step": 237 }, { "epoch": 0.0232421875, "grad_norm": 0.37431418895721436, "learning_rate": 0.0003966666666666667, "loss": 2.0657, "step": 238 }, { "epoch": 0.02333984375, "grad_norm": 0.30095174908638, "learning_rate": 0.00039833333333333333, "loss": 2.0217, "step": 239 }, { "epoch": 0.0234375, "grad_norm": 0.24695053696632385, "learning_rate": 0.0004, "loss": 1.9833, "step": 240 }, { "epoch": 0.02353515625, "grad_norm": 0.2923540771007538, "learning_rate": 0.00040166666666666665, "loss": 2.0272, "step": 241 }, { "epoch": 0.0236328125, "grad_norm": 0.2788209915161133, "learning_rate": 0.00040333333333333334, "loss": 2.0104, "step": 242 }, { "epoch": 0.02373046875, "grad_norm": 0.2529614567756653, "learning_rate": 0.00040500000000000003, "loss": 2.003, "step": 243 }, { "epoch": 0.023828125, "grad_norm": 0.2551966905593872, "learning_rate": 0.00040666666666666667, "loss": 2.001, "step": 244 }, { "epoch": 0.02392578125, "grad_norm": 0.2613292634487152, "learning_rate": 0.00040833333333333336, "loss": 1.9822, "step": 245 }, { "epoch": 0.0240234375, "grad_norm": 0.3060430884361267, "learning_rate": 0.00041, "loss": 2.0024, "step": 246 }, { "epoch": 0.02412109375, "grad_norm": 0.33755916357040405, "learning_rate": 0.0004116666666666667, "loss": 2.0023, "step": 247 }, { "epoch": 0.02421875, "grad_norm": 0.33021774888038635, "learning_rate": 0.0004133333333333333, "loss": 1.9086, "step": 248 }, { "epoch": 0.02431640625, "grad_norm": 0.26662060618400574, "learning_rate": 0.000415, "loss": 2.009, "step": 249 }, { "epoch": 0.0244140625, "grad_norm": 0.27698251605033875, "learning_rate": 0.0004166666666666667, "loss": 2.0183, "step": 250 }, { "epoch": 0.02451171875, "grad_norm": 0.2582184970378876, "learning_rate": 0.00041833333333333333, "loss": 1.9932, "step": 251 }, { "epoch": 0.024609375, "grad_norm": 0.28684699535369873, "learning_rate": 0.00042, "loss": 2.0021, "step": 252 }, { "epoch": 0.02470703125, "grad_norm": 0.33535540103912354, "learning_rate": 0.0004216666666666667, "loss": 2.0387, "step": 253 }, { "epoch": 0.0248046875, "grad_norm": 0.3330588638782501, "learning_rate": 0.00042333333333333334, "loss": 1.9776, "step": 254 }, { "epoch": 0.02490234375, "grad_norm": 0.27919256687164307, "learning_rate": 0.000425, "loss": 2.0111, "step": 255 }, { "epoch": 0.025, "grad_norm": 0.25296416878700256, "learning_rate": 0.0004266666666666667, "loss": 1.9755, "step": 256 }, { "epoch": 0.02509765625, "grad_norm": 0.31288138031959534, "learning_rate": 0.00042833333333333335, "loss": 2.032, "step": 257 }, { "epoch": 0.0251953125, "grad_norm": 0.3360923230648041, "learning_rate": 0.00043, "loss": 2.0837, "step": 258 }, { "epoch": 0.02529296875, "grad_norm": 0.36317816376686096, "learning_rate": 0.0004316666666666667, "loss": 1.9696, "step": 259 }, { "epoch": 0.025390625, "grad_norm": 0.366953581571579, "learning_rate": 0.00043333333333333337, "loss": 2.031, "step": 260 }, { "epoch": 0.02548828125, "grad_norm": 0.34289368987083435, "learning_rate": 0.000435, "loss": 1.9968, "step": 261 }, { "epoch": 0.0255859375, "grad_norm": 0.35170793533325195, "learning_rate": 0.00043666666666666664, "loss": 1.9963, "step": 262 }, { "epoch": 0.02568359375, "grad_norm": 0.28625521063804626, "learning_rate": 0.0004383333333333334, "loss": 1.9932, "step": 263 }, { "epoch": 0.02578125, "grad_norm": 0.2861610949039459, "learning_rate": 0.00044, "loss": 2.0297, "step": 264 }, { "epoch": 0.02587890625, "grad_norm": 0.30467647314071655, "learning_rate": 0.00044166666666666665, "loss": 1.992, "step": 265 }, { "epoch": 0.0259765625, "grad_norm": 0.31711357831954956, "learning_rate": 0.00044333333333333334, "loss": 2.0135, "step": 266 }, { "epoch": 0.02607421875, "grad_norm": 0.341530978679657, "learning_rate": 0.00044500000000000003, "loss": 1.983, "step": 267 }, { "epoch": 0.026171875, "grad_norm": 0.373901903629303, "learning_rate": 0.00044666666666666666, "loss": 2.0048, "step": 268 }, { "epoch": 0.02626953125, "grad_norm": 0.3105134665966034, "learning_rate": 0.0004483333333333333, "loss": 2.0368, "step": 269 }, { "epoch": 0.0263671875, "grad_norm": 0.29363134503364563, "learning_rate": 0.00045000000000000004, "loss": 1.978, "step": 270 }, { "epoch": 0.02646484375, "grad_norm": 0.3060167133808136, "learning_rate": 0.0004516666666666667, "loss": 1.9479, "step": 271 }, { "epoch": 0.0265625, "grad_norm": 0.30803290009498596, "learning_rate": 0.0004533333333333333, "loss": 1.9662, "step": 272 }, { "epoch": 0.02666015625, "grad_norm": 0.3324045240879059, "learning_rate": 0.000455, "loss": 2.0299, "step": 273 }, { "epoch": 0.0267578125, "grad_norm": 0.39051148295402527, "learning_rate": 0.0004566666666666667, "loss": 1.9856, "step": 274 }, { "epoch": 0.02685546875, "grad_norm": 0.4288715124130249, "learning_rate": 0.0004583333333333333, "loss": 2.0264, "step": 275 }, { "epoch": 0.026953125, "grad_norm": 0.34478962421417236, "learning_rate": 0.00046, "loss": 1.9824, "step": 276 }, { "epoch": 0.02705078125, "grad_norm": 0.2766290009021759, "learning_rate": 0.0004616666666666667, "loss": 2.0066, "step": 277 }, { "epoch": 0.0271484375, "grad_norm": 0.2508682608604431, "learning_rate": 0.00046333333333333334, "loss": 1.9663, "step": 278 }, { "epoch": 0.02724609375, "grad_norm": 0.26924827694892883, "learning_rate": 0.000465, "loss": 1.9903, "step": 279 }, { "epoch": 0.02734375, "grad_norm": 0.27668496966362, "learning_rate": 0.00046666666666666666, "loss": 2.0097, "step": 280 }, { "epoch": 0.02744140625, "grad_norm": 0.25026220083236694, "learning_rate": 0.00046833333333333335, "loss": 2.0583, "step": 281 }, { "epoch": 0.0275390625, "grad_norm": 0.2158055454492569, "learning_rate": 0.00047, "loss": 2.0137, "step": 282 }, { "epoch": 0.02763671875, "grad_norm": 0.22540244460105896, "learning_rate": 0.0004716666666666667, "loss": 1.994, "step": 283 }, { "epoch": 0.027734375, "grad_norm": 0.26405519247055054, "learning_rate": 0.00047333333333333336, "loss": 2.0221, "step": 284 }, { "epoch": 0.02783203125, "grad_norm": 0.2979099452495575, "learning_rate": 0.000475, "loss": 2.0047, "step": 285 }, { "epoch": 0.0279296875, "grad_norm": 0.34131935238838196, "learning_rate": 0.0004766666666666667, "loss": 1.9907, "step": 286 }, { "epoch": 0.02802734375, "grad_norm": 0.37178686261177063, "learning_rate": 0.0004783333333333333, "loss": 1.9806, "step": 287 }, { "epoch": 0.028125, "grad_norm": 0.36835598945617676, "learning_rate": 0.00048, "loss": 2.0134, "step": 288 }, { "epoch": 0.02822265625, "grad_norm": 0.29690125584602356, "learning_rate": 0.0004816666666666667, "loss": 2.0261, "step": 289 }, { "epoch": 0.0283203125, "grad_norm": 0.2690771818161011, "learning_rate": 0.00048333333333333334, "loss": 1.9718, "step": 290 }, { "epoch": 0.02841796875, "grad_norm": 0.3377201557159424, "learning_rate": 0.00048499999999999997, "loss": 1.99, "step": 291 }, { "epoch": 0.028515625, "grad_norm": 0.34973010420799255, "learning_rate": 0.0004866666666666667, "loss": 1.9721, "step": 292 }, { "epoch": 0.02861328125, "grad_norm": 0.3172457218170166, "learning_rate": 0.0004883333333333333, "loss": 1.9928, "step": 293 }, { "epoch": 0.0287109375, "grad_norm": 0.34357598423957825, "learning_rate": 0.00049, "loss": 1.9995, "step": 294 }, { "epoch": 0.02880859375, "grad_norm": 0.3824540376663208, "learning_rate": 0.0004916666666666666, "loss": 1.9772, "step": 295 }, { "epoch": 0.02890625, "grad_norm": 0.3704535663127899, "learning_rate": 0.0004933333333333334, "loss": 1.9829, "step": 296 }, { "epoch": 0.02900390625, "grad_norm": 0.2571757733821869, "learning_rate": 0.000495, "loss": 1.9966, "step": 297 }, { "epoch": 0.0291015625, "grad_norm": 0.3970927894115448, "learning_rate": 0.0004966666666666666, "loss": 1.9841, "step": 298 }, { "epoch": 0.02919921875, "grad_norm": 0.3420144319534302, "learning_rate": 0.0004983333333333334, "loss": 1.9747, "step": 299 }, { "epoch": 0.029296875, "grad_norm": 0.30147823691368103, "learning_rate": 0.0005, "loss": 1.9999, "step": 300 }, { "epoch": 0.02939453125, "grad_norm": 0.33727970719337463, "learning_rate": 0.0004999999887622467, "loss": 2.0084, "step": 301 }, { "epoch": 0.0294921875, "grad_norm": 0.29407384991645813, "learning_rate": 0.0004999999550489878, "loss": 2.0144, "step": 302 }, { "epoch": 0.02958984375, "grad_norm": 0.3489755094051361, "learning_rate": 0.0004999998988602267, "loss": 2.0058, "step": 303 }, { "epoch": 0.0296875, "grad_norm": 0.3327770233154297, "learning_rate": 0.0004999998201959691, "loss": 2.0166, "step": 304 }, { "epoch": 0.02978515625, "grad_norm": 0.2923370599746704, "learning_rate": 0.0004999997190562227, "loss": 2.0148, "step": 305 }, { "epoch": 0.0298828125, "grad_norm": 0.31616437435150146, "learning_rate": 0.0004999995954409976, "loss": 1.9772, "step": 306 }, { "epoch": 0.02998046875, "grad_norm": 0.22982288897037506, "learning_rate": 0.0004999994493503064, "loss": 2.0584, "step": 307 }, { "epoch": 0.030078125, "grad_norm": 0.2886744737625122, "learning_rate": 0.0004999992807841634, "loss": 2.0114, "step": 308 }, { "epoch": 0.03017578125, "grad_norm": 0.3027271032333374, "learning_rate": 0.0004999990897425856, "loss": 2.011, "step": 309 }, { "epoch": 0.0302734375, "grad_norm": 0.3191162645816803, "learning_rate": 0.0004999988762255922, "loss": 1.9962, "step": 310 }, { "epoch": 0.03037109375, "grad_norm": 0.34986981749534607, "learning_rate": 0.0004999986402332042, "loss": 1.9612, "step": 311 }, { "epoch": 0.03046875, "grad_norm": 0.36431390047073364, "learning_rate": 0.0004999983817654454, "loss": 2.045, "step": 312 }, { "epoch": 0.03056640625, "grad_norm": 0.4198042154312134, "learning_rate": 0.0004999981008223416, "loss": 2.0132, "step": 313 }, { "epoch": 0.0306640625, "grad_norm": 0.43374890089035034, "learning_rate": 0.0004999977974039207, "loss": 2.0578, "step": 314 }, { "epoch": 0.03076171875, "grad_norm": 0.3654812276363373, "learning_rate": 0.0004999974715102132, "loss": 1.9721, "step": 315 }, { "epoch": 0.030859375, "grad_norm": 0.29420921206474304, "learning_rate": 0.0004999971231412517, "loss": 2.0296, "step": 316 }, { "epoch": 0.03095703125, "grad_norm": 0.31979072093963623, "learning_rate": 0.0004999967522970708, "loss": 1.9623, "step": 317 }, { "epoch": 0.0310546875, "grad_norm": 0.3570129871368408, "learning_rate": 0.0004999963589777076, "loss": 1.9966, "step": 318 }, { "epoch": 0.03115234375, "grad_norm": 0.29743143916130066, "learning_rate": 0.0004999959431832016, "loss": 1.9535, "step": 319 }, { "epoch": 0.03125, "grad_norm": 0.23641493916511536, "learning_rate": 0.000499995504913594, "loss": 2.0304, "step": 320 }, { "epoch": 0.03134765625, "grad_norm": 0.3556622564792633, "learning_rate": 0.0004999950441689288, "loss": 1.971, "step": 321 }, { "epoch": 0.0314453125, "grad_norm": 0.323939710855484, "learning_rate": 0.0004999945609492519, "loss": 2.0153, "step": 322 }, { "epoch": 0.03154296875, "grad_norm": 0.22797244787216187, "learning_rate": 0.0004999940552546118, "loss": 1.9807, "step": 323 }, { "epoch": 0.031640625, "grad_norm": 0.2641647458076477, "learning_rate": 0.0004999935270850587, "loss": 1.9988, "step": 324 }, { "epoch": 0.03173828125, "grad_norm": 0.25289344787597656, "learning_rate": 0.0004999929764406455, "loss": 1.9748, "step": 325 }, { "epoch": 0.0318359375, "grad_norm": 0.2377796769142151, "learning_rate": 0.0004999924033214274, "loss": 1.9983, "step": 326 }, { "epoch": 0.03193359375, "grad_norm": 0.2711915671825409, "learning_rate": 0.0004999918077274612, "loss": 1.9643, "step": 327 }, { "epoch": 0.03203125, "grad_norm": 0.2866462767124176, "learning_rate": 0.0004999911896588068, "loss": 2.0196, "step": 328 }, { "epoch": 0.03212890625, "grad_norm": 0.25075578689575195, "learning_rate": 0.0004999905491155257, "loss": 2.0426, "step": 329 }, { "epoch": 0.0322265625, "grad_norm": 0.266648530960083, "learning_rate": 0.000499989886097682, "loss": 2.0318, "step": 330 }, { "epoch": 0.03232421875, "grad_norm": 0.22494247555732727, "learning_rate": 0.0004999892006053421, "loss": 1.9839, "step": 331 }, { "epoch": 0.032421875, "grad_norm": 0.3179854452610016, "learning_rate": 0.0004999884926385741, "loss": 1.9981, "step": 332 }, { "epoch": 0.03251953125, "grad_norm": 0.2754990756511688, "learning_rate": 0.000499987762197449, "loss": 1.9879, "step": 333 }, { "epoch": 0.0326171875, "grad_norm": 0.2669137418270111, "learning_rate": 0.0004999870092820395, "loss": 2.0083, "step": 334 }, { "epoch": 0.03271484375, "grad_norm": 0.31376007199287415, "learning_rate": 0.0004999862338924212, "loss": 2.0166, "step": 335 }, { "epoch": 0.0328125, "grad_norm": 0.32645899057388306, "learning_rate": 0.0004999854360286712, "loss": 2.0019, "step": 336 }, { "epoch": 0.03291015625, "grad_norm": 0.3550071120262146, "learning_rate": 0.0004999846156908692, "loss": 1.9744, "step": 337 }, { "epoch": 0.0330078125, "grad_norm": 0.30171769857406616, "learning_rate": 0.0004999837728790975, "loss": 2.0231, "step": 338 }, { "epoch": 0.03310546875, "grad_norm": 0.23128142952919006, "learning_rate": 0.00049998290759344, "loss": 2.001, "step": 339 }, { "epoch": 0.033203125, "grad_norm": 0.2433364987373352, "learning_rate": 0.0004999820198339832, "loss": 1.9945, "step": 340 }, { "epoch": 0.03330078125, "grad_norm": 0.30562201142311096, "learning_rate": 0.0004999811096008159, "loss": 2.0228, "step": 341 }, { "epoch": 0.0333984375, "grad_norm": 0.33020487427711487, "learning_rate": 0.0004999801768940287, "loss": 1.9691, "step": 342 }, { "epoch": 0.03349609375, "grad_norm": 0.3125375807285309, "learning_rate": 0.0004999792217137151, "loss": 1.9728, "step": 343 }, { "epoch": 0.03359375, "grad_norm": 0.26769348978996277, "learning_rate": 0.0004999782440599702, "loss": 2.0066, "step": 344 }, { "epoch": 0.03369140625, "grad_norm": 0.3596431314945221, "learning_rate": 0.0004999772439328921, "loss": 2.0426, "step": 345 }, { "epoch": 0.0337890625, "grad_norm": 0.34117281436920166, "learning_rate": 0.0004999762213325803, "loss": 1.9985, "step": 346 }, { "epoch": 0.03388671875, "grad_norm": 0.2922564148902893, "learning_rate": 0.0004999751762591371, "loss": 2.0346, "step": 347 }, { "epoch": 0.033984375, "grad_norm": 0.2813419997692108, "learning_rate": 0.0004999741087126669, "loss": 2.0573, "step": 348 }, { "epoch": 0.03408203125, "grad_norm": 0.27346089482307434, "learning_rate": 0.0004999730186932764, "loss": 1.9655, "step": 349 }, { "epoch": 0.0341796875, "grad_norm": 0.24704065918922424, "learning_rate": 0.0004999719062010745, "loss": 1.9542, "step": 350 }, { "epoch": 0.03427734375, "grad_norm": 0.24761976301670074, "learning_rate": 0.000499970771236172, "loss": 1.9526, "step": 351 }, { "epoch": 0.034375, "grad_norm": 0.2508860230445862, "learning_rate": 0.0004999696137986826, "loss": 2.0119, "step": 352 }, { "epoch": 0.03447265625, "grad_norm": 0.24160990118980408, "learning_rate": 0.0004999684338887219, "loss": 2.0148, "step": 353 }, { "epoch": 0.0345703125, "grad_norm": 0.2879098355770111, "learning_rate": 0.0004999672315064076, "loss": 2.0206, "step": 354 }, { "epoch": 0.03466796875, "grad_norm": 0.28366580605506897, "learning_rate": 0.0004999660066518601, "loss": 2.0737, "step": 355 }, { "epoch": 0.034765625, "grad_norm": 0.3401612937450409, "learning_rate": 0.0004999647593252013, "loss": 1.9868, "step": 356 }, { "epoch": 0.03486328125, "grad_norm": 0.4696269631385803, "learning_rate": 0.0004999634895265562, "loss": 2.0079, "step": 357 }, { "epoch": 0.0349609375, "grad_norm": 0.596168041229248, "learning_rate": 0.0004999621972560515, "loss": 2.0084, "step": 358 }, { "epoch": 0.03505859375, "grad_norm": 0.5319205522537231, "learning_rate": 0.0004999608825138162, "loss": 2.0084, "step": 359 }, { "epoch": 0.03515625, "grad_norm": 0.24298223853111267, "learning_rate": 0.0004999595452999818, "loss": 1.9183, "step": 360 }, { "epoch": 0.03525390625, "grad_norm": 0.39650923013687134, "learning_rate": 0.0004999581856146817, "loss": 1.9886, "step": 361 }, { "epoch": 0.0353515625, "grad_norm": 0.3819667100906372, "learning_rate": 0.0004999568034580518, "loss": 2.0047, "step": 362 }, { "epoch": 0.03544921875, "grad_norm": 0.3298279345035553, "learning_rate": 0.0004999553988302303, "loss": 1.9466, "step": 363 }, { "epoch": 0.035546875, "grad_norm": 0.2343115359544754, "learning_rate": 0.0004999539717313573, "loss": 1.9828, "step": 364 }, { "epoch": 0.03564453125, "grad_norm": 0.31238943338394165, "learning_rate": 0.0004999525221615755, "loss": 2.0721, "step": 365 }, { "epoch": 0.0357421875, "grad_norm": 0.27481910586357117, "learning_rate": 0.0004999510501210295, "loss": 2.0106, "step": 366 }, { "epoch": 0.03583984375, "grad_norm": 0.24870915710926056, "learning_rate": 0.0004999495556098666, "loss": 1.9846, "step": 367 }, { "epoch": 0.0359375, "grad_norm": 0.2666539251804352, "learning_rate": 0.0004999480386282359, "loss": 1.9988, "step": 368 }, { "epoch": 0.03603515625, "grad_norm": 0.23231451213359833, "learning_rate": 0.000499946499176289, "loss": 2.0235, "step": 369 }, { "epoch": 0.0361328125, "grad_norm": 0.19276577234268188, "learning_rate": 0.0004999449372541798, "loss": 1.9832, "step": 370 }, { "epoch": 0.03623046875, "grad_norm": 0.23545822501182556, "learning_rate": 0.000499943352862064, "loss": 1.9879, "step": 371 }, { "epoch": 0.036328125, "grad_norm": 0.2608807384967804, "learning_rate": 0.0004999417460001002, "loss": 2.047, "step": 372 }, { "epoch": 0.03642578125, "grad_norm": 0.27395492792129517, "learning_rate": 0.0004999401166684487, "loss": 1.9496, "step": 373 }, { "epoch": 0.0365234375, "grad_norm": 0.24661333858966827, "learning_rate": 0.0004999384648672724, "loss": 1.9924, "step": 374 }, { "epoch": 0.03662109375, "grad_norm": 0.23559828102588654, "learning_rate": 0.0004999367905967362, "loss": 1.979, "step": 375 }, { "epoch": 0.03671875, "grad_norm": 0.2556357681751251, "learning_rate": 0.0004999350938570074, "loss": 1.9793, "step": 376 }, { "epoch": 0.03681640625, "grad_norm": 0.29081088304519653, "learning_rate": 0.0004999333746482555, "loss": 1.962, "step": 377 }, { "epoch": 0.0369140625, "grad_norm": 0.2583800256252289, "learning_rate": 0.0004999316329706521, "loss": 1.9765, "step": 378 }, { "epoch": 0.03701171875, "grad_norm": 0.2506580054759979, "learning_rate": 0.0004999298688243714, "loss": 2.0302, "step": 379 }, { "epoch": 0.037109375, "grad_norm": 0.289530485868454, "learning_rate": 0.0004999280822095895, "loss": 1.9514, "step": 380 }, { "epoch": 0.03720703125, "grad_norm": 0.24479152262210846, "learning_rate": 0.0004999262731264848, "loss": 1.9742, "step": 381 }, { "epoch": 0.0373046875, "grad_norm": 0.23676550388336182, "learning_rate": 0.0004999244415752381, "loss": 2.0161, "step": 382 }, { "epoch": 0.03740234375, "grad_norm": 0.2447502315044403, "learning_rate": 0.0004999225875560323, "loss": 2.0114, "step": 383 }, { "epoch": 0.0375, "grad_norm": 0.20993874967098236, "learning_rate": 0.0004999207110690528, "loss": 1.9786, "step": 384 }, { "epoch": 0.03759765625, "grad_norm": 0.23405557870864868, "learning_rate": 0.0004999188121144867, "loss": 2.0004, "step": 385 }, { "epoch": 0.0376953125, "grad_norm": 0.2310025990009308, "learning_rate": 0.0004999168906925238, "loss": 1.924, "step": 386 }, { "epoch": 0.03779296875, "grad_norm": 0.2507460117340088, "learning_rate": 0.0004999149468033564, "loss": 2.0301, "step": 387 }, { "epoch": 0.037890625, "grad_norm": 0.30916762351989746, "learning_rate": 0.0004999129804471782, "loss": 1.9948, "step": 388 }, { "epoch": 0.03798828125, "grad_norm": 0.42094507813453674, "learning_rate": 0.0004999109916241858, "loss": 2.0128, "step": 389 }, { "epoch": 0.0380859375, "grad_norm": 0.5398088693618774, "learning_rate": 0.0004999089803345779, "loss": 1.9866, "step": 390 }, { "epoch": 0.03818359375, "grad_norm": 0.4601620137691498, "learning_rate": 0.0004999069465785554, "loss": 1.9807, "step": 391 }, { "epoch": 0.03828125, "grad_norm": 0.24873095750808716, "learning_rate": 0.0004999048903563213, "loss": 1.9225, "step": 392 }, { "epoch": 0.03837890625, "grad_norm": 0.37363573908805847, "learning_rate": 0.0004999028116680814, "loss": 2.0212, "step": 393 }, { "epoch": 0.0384765625, "grad_norm": 0.27399107813835144, "learning_rate": 0.0004999007105140428, "loss": 1.9797, "step": 394 }, { "epoch": 0.03857421875, "grad_norm": 0.2717953622341156, "learning_rate": 0.0004998985868944158, "loss": 2.0119, "step": 395 }, { "epoch": 0.038671875, "grad_norm": 0.2766648530960083, "learning_rate": 0.0004998964408094124, "loss": 1.9936, "step": 396 }, { "epoch": 0.03876953125, "grad_norm": 0.2989843487739563, "learning_rate": 0.0004998942722592469, "loss": 2.0261, "step": 397 }, { "epoch": 0.0388671875, "grad_norm": 0.31954097747802734, "learning_rate": 0.000499892081244136, "loss": 1.9713, "step": 398 }, { "epoch": 0.03896484375, "grad_norm": 0.31199777126312256, "learning_rate": 0.0004998898677642987, "loss": 1.9548, "step": 399 }, { "epoch": 0.0390625, "grad_norm": 0.2475418746471405, "learning_rate": 0.0004998876318199557, "loss": 1.9878, "step": 400 }, { "epoch": 0.03916015625, "grad_norm": 0.25001809000968933, "learning_rate": 0.0004998853734113308, "loss": 1.991, "step": 401 }, { "epoch": 0.0392578125, "grad_norm": 0.2468329817056656, "learning_rate": 0.0004998830925386492, "loss": 1.9988, "step": 402 }, { "epoch": 0.03935546875, "grad_norm": 0.23273305594921112, "learning_rate": 0.000499880789202139, "loss": 2.0138, "step": 403 }, { "epoch": 0.039453125, "grad_norm": 0.21512626111507416, "learning_rate": 0.0004998784634020303, "loss": 1.9326, "step": 404 }, { "epoch": 0.03955078125, "grad_norm": 0.2123369723558426, "learning_rate": 0.0004998761151385554, "loss": 1.9954, "step": 405 }, { "epoch": 0.0396484375, "grad_norm": 0.20985403656959534, "learning_rate": 0.0004998737444119488, "loss": 1.996, "step": 406 }, { "epoch": 0.03974609375, "grad_norm": 0.19709332287311554, "learning_rate": 0.0004998713512224473, "loss": 1.9642, "step": 407 }, { "epoch": 0.03984375, "grad_norm": 0.21997478604316711, "learning_rate": 0.00049986893557029, "loss": 1.9925, "step": 408 }, { "epoch": 0.03994140625, "grad_norm": 0.2419801503419876, "learning_rate": 0.0004998664974557182, "loss": 2.0262, "step": 409 }, { "epoch": 0.0400390625, "grad_norm": 0.2527197003364563, "learning_rate": 0.0004998640368789754, "loss": 2.0011, "step": 410 }, { "epoch": 0.04013671875, "grad_norm": 0.2539777159690857, "learning_rate": 0.0004998615538403074, "loss": 1.9799, "step": 411 }, { "epoch": 0.040234375, "grad_norm": 0.2419251948595047, "learning_rate": 0.0004998590483399623, "loss": 2.0035, "step": 412 }, { "epoch": 0.04033203125, "grad_norm": 0.26659199595451355, "learning_rate": 0.0004998565203781904, "loss": 1.9743, "step": 413 }, { "epoch": 0.0404296875, "grad_norm": 0.3216726779937744, "learning_rate": 0.0004998539699552441, "loss": 1.979, "step": 414 }, { "epoch": 0.04052734375, "grad_norm": 0.2780512571334839, "learning_rate": 0.0004998513970713783, "loss": 1.9892, "step": 415 }, { "epoch": 0.040625, "grad_norm": 0.2947809100151062, "learning_rate": 0.00049984880172685, "loss": 1.9836, "step": 416 }, { "epoch": 0.04072265625, "grad_norm": 0.33937209844589233, "learning_rate": 0.0004998461839219182, "loss": 2.0007, "step": 417 }, { "epoch": 0.0408203125, "grad_norm": 0.3530365228652954, "learning_rate": 0.0004998435436568446, "loss": 1.9462, "step": 418 }, { "epoch": 0.04091796875, "grad_norm": 0.34433260560035706, "learning_rate": 0.000499840880931893, "loss": 1.9733, "step": 419 }, { "epoch": 0.041015625, "grad_norm": 0.323081910610199, "learning_rate": 0.0004998381957473293, "loss": 1.9522, "step": 420 }, { "epoch": 0.04111328125, "grad_norm": 0.22895868122577667, "learning_rate": 0.0004998354881034217, "loss": 1.9672, "step": 421 }, { "epoch": 0.0412109375, "grad_norm": 0.2543105185031891, "learning_rate": 0.0004998327580004408, "loss": 1.9429, "step": 422 }, { "epoch": 0.04130859375, "grad_norm": 0.2795581519603729, "learning_rate": 0.0004998300054386591, "loss": 1.9902, "step": 423 }, { "epoch": 0.04140625, "grad_norm": 0.2201048880815506, "learning_rate": 0.0004998272304183517, "loss": 1.9365, "step": 424 }, { "epoch": 0.04150390625, "grad_norm": 0.26653197407722473, "learning_rate": 0.0004998244329397958, "loss": 2.0178, "step": 425 }, { "epoch": 0.0416015625, "grad_norm": 0.26959821581840515, "learning_rate": 0.0004998216130032708, "loss": 1.9625, "step": 426 }, { "epoch": 0.04169921875, "grad_norm": 0.2673629820346832, "learning_rate": 0.0004998187706090584, "loss": 1.995, "step": 427 }, { "epoch": 0.041796875, "grad_norm": 0.3190925121307373, "learning_rate": 0.0004998159057574426, "loss": 2.0094, "step": 428 }, { "epoch": 0.04189453125, "grad_norm": 0.3376927673816681, "learning_rate": 0.0004998130184487094, "loss": 2.0535, "step": 429 }, { "epoch": 0.0419921875, "grad_norm": 0.35333067178726196, "learning_rate": 0.0004998101086831474, "loss": 2.0256, "step": 430 }, { "epoch": 0.04208984375, "grad_norm": 0.3051585853099823, "learning_rate": 0.0004998071764610471, "loss": 1.9798, "step": 431 }, { "epoch": 0.0421875, "grad_norm": 0.21881523728370667, "learning_rate": 0.0004998042217827015, "loss": 1.9726, "step": 432 }, { "epoch": 0.04228515625, "grad_norm": 0.2296576052904129, "learning_rate": 0.0004998012446484057, "loss": 1.9675, "step": 433 }, { "epoch": 0.0423828125, "grad_norm": 0.24055050313472748, "learning_rate": 0.0004997982450584572, "loss": 1.9488, "step": 434 }, { "epoch": 0.04248046875, "grad_norm": 0.22267191112041473, "learning_rate": 0.0004997952230131555, "loss": 1.9617, "step": 435 }, { "epoch": 0.042578125, "grad_norm": 0.2590855658054352, "learning_rate": 0.0004997921785128026, "loss": 1.9745, "step": 436 }, { "epoch": 0.04267578125, "grad_norm": 0.2995806336402893, "learning_rate": 0.0004997891115577025, "loss": 1.9667, "step": 437 }, { "epoch": 0.0427734375, "grad_norm": 0.3083277642726898, "learning_rate": 0.0004997860221481616, "loss": 1.9661, "step": 438 }, { "epoch": 0.04287109375, "grad_norm": 0.31743094325065613, "learning_rate": 0.0004997829102844885, "loss": 2.0172, "step": 439 }, { "epoch": 0.04296875, "grad_norm": 0.37507855892181396, "learning_rate": 0.0004997797759669941, "loss": 2.0211, "step": 440 }, { "epoch": 0.04306640625, "grad_norm": 0.35785287618637085, "learning_rate": 0.0004997766191959914, "loss": 2.038, "step": 441 }, { "epoch": 0.0431640625, "grad_norm": 0.3099903166294098, "learning_rate": 0.0004997734399717958, "loss": 1.9879, "step": 442 }, { "epoch": 0.04326171875, "grad_norm": 0.32211577892303467, "learning_rate": 0.0004997702382947248, "loss": 1.987, "step": 443 }, { "epoch": 0.043359375, "grad_norm": 0.24999700486660004, "learning_rate": 0.0004997670141650984, "loss": 1.9765, "step": 444 }, { "epoch": 0.04345703125, "grad_norm": 0.23301391303539276, "learning_rate": 0.0004997637675832386, "loss": 1.9592, "step": 445 }, { "epoch": 0.0435546875, "grad_norm": 0.22780932486057281, "learning_rate": 0.0004997604985494694, "loss": 1.9762, "step": 446 }, { "epoch": 0.04365234375, "grad_norm": 0.2347191423177719, "learning_rate": 0.0004997572070641178, "loss": 1.9791, "step": 447 }, { "epoch": 0.04375, "grad_norm": 0.3288608193397522, "learning_rate": 0.0004997538931275123, "loss": 1.9664, "step": 448 }, { "epoch": 0.04384765625, "grad_norm": 0.29897212982177734, "learning_rate": 0.000499750556739984, "loss": 1.9876, "step": 449 }, { "epoch": 0.0439453125, "grad_norm": 0.24389693140983582, "learning_rate": 0.0004997471979018663, "loss": 1.9742, "step": 450 }, { "epoch": 0.04404296875, "grad_norm": 0.3172384798526764, "learning_rate": 0.0004997438166134945, "loss": 1.969, "step": 451 }, { "epoch": 0.044140625, "grad_norm": 0.22951611876487732, "learning_rate": 0.0004997404128752065, "loss": 1.9832, "step": 452 }, { "epoch": 0.04423828125, "grad_norm": 0.2356499582529068, "learning_rate": 0.0004997369866873423, "loss": 2.0236, "step": 453 }, { "epoch": 0.0443359375, "grad_norm": 0.2687929570674896, "learning_rate": 0.000499733538050244, "loss": 1.9521, "step": 454 }, { "epoch": 0.04443359375, "grad_norm": 0.2512679398059845, "learning_rate": 0.0004997300669642564, "loss": 1.9754, "step": 455 }, { "epoch": 0.04453125, "grad_norm": 0.22110989689826965, "learning_rate": 0.0004997265734297259, "loss": 1.9931, "step": 456 }, { "epoch": 0.04462890625, "grad_norm": 0.2548464834690094, "learning_rate": 0.0004997230574470017, "loss": 2.001, "step": 457 }, { "epoch": 0.0447265625, "grad_norm": 0.3107970356941223, "learning_rate": 0.0004997195190164349, "loss": 1.9855, "step": 458 }, { "epoch": 0.04482421875, "grad_norm": 0.2989678680896759, "learning_rate": 0.0004997159581383789, "loss": 1.9646, "step": 459 }, { "epoch": 0.044921875, "grad_norm": 0.3651764690876007, "learning_rate": 0.0004997123748131896, "loss": 1.9865, "step": 460 }, { "epoch": 0.04501953125, "grad_norm": 0.41693389415740967, "learning_rate": 0.0004997087690412248, "loss": 1.9756, "step": 461 }, { "epoch": 0.0451171875, "grad_norm": 0.425959974527359, "learning_rate": 0.0004997051408228447, "loss": 1.9887, "step": 462 }, { "epoch": 0.04521484375, "grad_norm": 0.38966843485832214, "learning_rate": 0.0004997014901584118, "loss": 1.9532, "step": 463 }, { "epoch": 0.0453125, "grad_norm": 0.29931432008743286, "learning_rate": 0.0004996978170482906, "loss": 1.9806, "step": 464 }, { "epoch": 0.04541015625, "grad_norm": 0.23077397048473358, "learning_rate": 0.0004996941214928482, "loss": 2.0258, "step": 465 }, { "epoch": 0.0455078125, "grad_norm": 0.32101985812187195, "learning_rate": 0.0004996904034924536, "loss": 2.0023, "step": 466 }, { "epoch": 0.04560546875, "grad_norm": 0.3225870430469513, "learning_rate": 0.0004996866630474783, "loss": 1.9385, "step": 467 }, { "epoch": 0.045703125, "grad_norm": 0.2592551112174988, "learning_rate": 0.0004996829001582959, "loss": 2.0176, "step": 468 }, { "epoch": 0.04580078125, "grad_norm": 0.24367393553256989, "learning_rate": 0.0004996791148252824, "loss": 1.9844, "step": 469 }, { "epoch": 0.0458984375, "grad_norm": 0.26423218846321106, "learning_rate": 0.0004996753070488156, "loss": 1.9675, "step": 470 }, { "epoch": 0.04599609375, "grad_norm": 0.27801692485809326, "learning_rate": 0.0004996714768292762, "loss": 1.9265, "step": 471 }, { "epoch": 0.04609375, "grad_norm": 0.3210003077983856, "learning_rate": 0.0004996676241670467, "loss": 1.9937, "step": 472 }, { "epoch": 0.04619140625, "grad_norm": 0.3082759380340576, "learning_rate": 0.0004996637490625118, "loss": 1.9997, "step": 473 }, { "epoch": 0.0462890625, "grad_norm": 0.2486785650253296, "learning_rate": 0.0004996598515160589, "loss": 2.0687, "step": 474 }, { "epoch": 0.04638671875, "grad_norm": 0.31927499175071716, "learning_rate": 0.000499655931528077, "loss": 1.9853, "step": 475 }, { "epoch": 0.046484375, "grad_norm": 0.2736678123474121, "learning_rate": 0.0004996519890989578, "loss": 1.9916, "step": 476 }, { "epoch": 0.04658203125, "grad_norm": 0.29522353410720825, "learning_rate": 0.0004996480242290952, "loss": 1.9554, "step": 477 }, { "epoch": 0.0466796875, "grad_norm": 0.26841938495635986, "learning_rate": 0.0004996440369188851, "loss": 1.9275, "step": 478 }, { "epoch": 0.04677734375, "grad_norm": 0.2695488929748535, "learning_rate": 0.0004996400271687259, "loss": 1.9898, "step": 479 }, { "epoch": 0.046875, "grad_norm": 0.36634454131126404, "learning_rate": 0.0004996359949790181, "loss": 1.9799, "step": 480 }, { "epoch": 0.04697265625, "grad_norm": 0.37612974643707275, "learning_rate": 0.0004996319403501647, "loss": 1.9645, "step": 481 }, { "epoch": 0.0470703125, "grad_norm": 0.29007261991500854, "learning_rate": 0.0004996278632825703, "loss": 1.9486, "step": 482 }, { "epoch": 0.04716796875, "grad_norm": 0.30373314023017883, "learning_rate": 0.0004996237637766424, "loss": 1.9913, "step": 483 }, { "epoch": 0.047265625, "grad_norm": 0.2949075400829315, "learning_rate": 0.0004996196418327906, "loss": 1.9894, "step": 484 }, { "epoch": 0.04736328125, "grad_norm": 0.30601680278778076, "learning_rate": 0.0004996154974514264, "loss": 1.9835, "step": 485 }, { "epoch": 0.0474609375, "grad_norm": 0.2896345853805542, "learning_rate": 0.000499611330632964, "loss": 2.0043, "step": 486 }, { "epoch": 0.04755859375, "grad_norm": 0.3836851418018341, "learning_rate": 0.0004996071413778195, "loss": 1.9782, "step": 487 }, { "epoch": 0.04765625, "grad_norm": 0.4167245626449585, "learning_rate": 0.0004996029296864114, "loss": 1.9792, "step": 488 }, { "epoch": 0.04775390625, "grad_norm": 0.3247157335281372, "learning_rate": 0.0004995986955591606, "loss": 1.9548, "step": 489 }, { "epoch": 0.0478515625, "grad_norm": 0.36726462841033936, "learning_rate": 0.0004995944389964897, "loss": 1.9844, "step": 490 }, { "epoch": 0.04794921875, "grad_norm": 0.27440258860588074, "learning_rate": 0.0004995901599988241, "loss": 1.9608, "step": 491 }, { "epoch": 0.048046875, "grad_norm": 0.271771103143692, "learning_rate": 0.0004995858585665912, "loss": 2.022, "step": 492 }, { "epoch": 0.04814453125, "grad_norm": 0.3079121708869934, "learning_rate": 0.0004995815347002208, "loss": 1.9804, "step": 493 }, { "epoch": 0.0482421875, "grad_norm": 0.2976233661174774, "learning_rate": 0.0004995771884001445, "loss": 1.9679, "step": 494 }, { "epoch": 0.04833984375, "grad_norm": 0.31318438053131104, "learning_rate": 0.0004995728196667969, "loss": 2.0342, "step": 495 }, { "epoch": 0.0484375, "grad_norm": 0.2397848516702652, "learning_rate": 0.0004995684285006139, "loss": 1.9801, "step": 496 }, { "epoch": 0.04853515625, "grad_norm": 0.2205589860677719, "learning_rate": 0.0004995640149020346, "loss": 2.0088, "step": 497 }, { "epoch": 0.0486328125, "grad_norm": 0.23892101645469666, "learning_rate": 0.0004995595788714995, "loss": 1.9888, "step": 498 }, { "epoch": 0.04873046875, "grad_norm": 0.2999497354030609, "learning_rate": 0.000499555120409452, "loss": 1.9896, "step": 499 }, { "epoch": 0.048828125, "grad_norm": 0.3309103548526764, "learning_rate": 0.0004995506395163372, "loss": 1.9394, "step": 500 }, { "epoch": 0.04892578125, "grad_norm": 0.36243122816085815, "learning_rate": 0.000499546136192603, "loss": 1.9604, "step": 501 }, { "epoch": 0.0490234375, "grad_norm": 0.3052852153778076, "learning_rate": 0.0004995416104386991, "loss": 1.9887, "step": 502 }, { "epoch": 0.04912109375, "grad_norm": 0.2957259714603424, "learning_rate": 0.0004995370622550775, "loss": 1.9876, "step": 503 }, { "epoch": 0.04921875, "grad_norm": 0.3991003632545471, "learning_rate": 0.0004995324916421926, "loss": 1.9932, "step": 504 }, { "epoch": 0.04931640625, "grad_norm": 0.27617356181144714, "learning_rate": 0.000499527898600501, "loss": 1.9387, "step": 505 }, { "epoch": 0.0494140625, "grad_norm": 0.2820669412612915, "learning_rate": 0.0004995232831304614, "loss": 2.012, "step": 506 }, { "epoch": 0.04951171875, "grad_norm": 0.3026665449142456, "learning_rate": 0.0004995186452325351, "loss": 1.9709, "step": 507 }, { "epoch": 0.049609375, "grad_norm": 0.2331731766462326, "learning_rate": 0.000499513984907185, "loss": 1.9617, "step": 508 }, { "epoch": 0.04970703125, "grad_norm": 0.2791639268398285, "learning_rate": 0.0004995093021548768, "loss": 1.9635, "step": 509 }, { "epoch": 0.0498046875, "grad_norm": 0.257519006729126, "learning_rate": 0.0004995045969760785, "loss": 1.8908, "step": 510 }, { "epoch": 0.04990234375, "grad_norm": 0.26278433203697205, "learning_rate": 0.0004994998693712598, "loss": 1.9418, "step": 511 }, { "epoch": 0.05, "grad_norm": 0.2534024715423584, "learning_rate": 0.0004994951193408929, "loss": 1.9821, "step": 512 }, { "epoch": 0.05009765625, "grad_norm": 0.272348552942276, "learning_rate": 0.0004994903468854527, "loss": 1.982, "step": 513 }, { "epoch": 0.0501953125, "grad_norm": 0.23883765935897827, "learning_rate": 0.0004994855520054154, "loss": 2.006, "step": 514 }, { "epoch": 0.05029296875, "grad_norm": 0.25627830624580383, "learning_rate": 0.0004994807347012603, "loss": 2.0008, "step": 515 }, { "epoch": 0.050390625, "grad_norm": 0.33997049927711487, "learning_rate": 0.0004994758949734686, "loss": 2.0339, "step": 516 }, { "epoch": 0.05048828125, "grad_norm": 0.4115971028804779, "learning_rate": 0.0004994710328225236, "loss": 1.9982, "step": 517 }, { "epoch": 0.0505859375, "grad_norm": 0.4417625069618225, "learning_rate": 0.000499466148248911, "loss": 1.9742, "step": 518 }, { "epoch": 0.05068359375, "grad_norm": 0.4123833477497101, "learning_rate": 0.0004994612412531189, "loss": 2.0148, "step": 519 }, { "epoch": 0.05078125, "grad_norm": 0.2234133630990982, "learning_rate": 0.0004994563118356373, "loss": 1.9679, "step": 520 }, { "epoch": 0.05087890625, "grad_norm": 0.35439473390579224, "learning_rate": 0.0004994513599969586, "loss": 1.9602, "step": 521 }, { "epoch": 0.0509765625, "grad_norm": 0.32346996665000916, "learning_rate": 0.0004994463857375776, "loss": 2.0106, "step": 522 }, { "epoch": 0.05107421875, "grad_norm": 0.24506594240665436, "learning_rate": 0.000499441389057991, "loss": 1.993, "step": 523 }, { "epoch": 0.051171875, "grad_norm": 0.26808494329452515, "learning_rate": 0.000499436369958698, "loss": 1.9764, "step": 524 }, { "epoch": 0.05126953125, "grad_norm": 0.21936193108558655, "learning_rate": 0.0004994313284401999, "loss": 1.9619, "step": 525 }, { "epoch": 0.0513671875, "grad_norm": 0.2545083165168762, "learning_rate": 0.0004994262645030005, "loss": 1.9825, "step": 526 }, { "epoch": 0.05146484375, "grad_norm": 0.26330727338790894, "learning_rate": 0.0004994211781476055, "loss": 1.9512, "step": 527 }, { "epoch": 0.0515625, "grad_norm": 0.2708059549331665, "learning_rate": 0.0004994160693745229, "loss": 1.9761, "step": 528 }, { "epoch": 0.05166015625, "grad_norm": 0.2882954478263855, "learning_rate": 0.0004994109381842632, "loss": 2.0223, "step": 529 }, { "epoch": 0.0517578125, "grad_norm": 0.25243133306503296, "learning_rate": 0.0004994057845773389, "loss": 1.976, "step": 530 }, { "epoch": 0.05185546875, "grad_norm": 0.23666390776634216, "learning_rate": 0.0004994006085542648, "loss": 1.9729, "step": 531 }, { "epoch": 0.051953125, "grad_norm": 0.24242763221263885, "learning_rate": 0.0004993954101155578, "loss": 1.9452, "step": 532 }, { "epoch": 0.05205078125, "grad_norm": 0.259750634431839, "learning_rate": 0.0004993901892617373, "loss": 1.9582, "step": 533 }, { "epoch": 0.0521484375, "grad_norm": 0.28516751527786255, "learning_rate": 0.0004993849459933249, "loss": 1.9367, "step": 534 }, { "epoch": 0.05224609375, "grad_norm": 0.26534441113471985, "learning_rate": 0.0004993796803108442, "loss": 2.0137, "step": 535 }, { "epoch": 0.05234375, "grad_norm": 0.28771716356277466, "learning_rate": 0.0004993743922148213, "loss": 1.927, "step": 536 }, { "epoch": 0.05244140625, "grad_norm": 0.2747785449028015, "learning_rate": 0.0004993690817057844, "loss": 1.9729, "step": 537 }, { "epoch": 0.0525390625, "grad_norm": 0.29758942127227783, "learning_rate": 0.0004993637487842639, "loss": 1.9713, "step": 538 }, { "epoch": 0.05263671875, "grad_norm": 0.3059535026550293, "learning_rate": 0.0004993583934507927, "loss": 2.0313, "step": 539 }, { "epoch": 0.052734375, "grad_norm": 0.36820656061172485, "learning_rate": 0.0004993530157059056, "loss": 1.9476, "step": 540 }, { "epoch": 0.05283203125, "grad_norm": 0.39971432089805603, "learning_rate": 0.0004993476155501396, "loss": 2.0238, "step": 541 }, { "epoch": 0.0529296875, "grad_norm": 0.3369678854942322, "learning_rate": 0.0004993421929840346, "loss": 1.9502, "step": 542 }, { "epoch": 0.05302734375, "grad_norm": 0.3733840882778168, "learning_rate": 0.000499336748008132, "loss": 2.0114, "step": 543 }, { "epoch": 0.053125, "grad_norm": 0.3251579999923706, "learning_rate": 0.0004993312806229757, "loss": 1.964, "step": 544 }, { "epoch": 0.05322265625, "grad_norm": 0.4093327522277832, "learning_rate": 0.0004993257908291117, "loss": 1.9853, "step": 545 }, { "epoch": 0.0533203125, "grad_norm": 0.26201343536376953, "learning_rate": 0.0004993202786270888, "loss": 2.0297, "step": 546 }, { "epoch": 0.05341796875, "grad_norm": 0.27842357754707336, "learning_rate": 0.0004993147440174572, "loss": 2.003, "step": 547 }, { "epoch": 0.053515625, "grad_norm": 0.2930687367916107, "learning_rate": 0.00049930918700077, "loss": 1.9634, "step": 548 }, { "epoch": 0.05361328125, "grad_norm": 0.2799973785877228, "learning_rate": 0.0004993036075775821, "loss": 2.0023, "step": 549 }, { "epoch": 0.0537109375, "grad_norm": 0.35084354877471924, "learning_rate": 0.000499298005748451, "loss": 2.0508, "step": 550 }, { "epoch": 0.05380859375, "grad_norm": 0.2895161211490631, "learning_rate": 0.0004992923815139362, "loss": 2.0074, "step": 551 }, { "epoch": 0.05390625, "grad_norm": 0.2637045979499817, "learning_rate": 0.0004992867348745997, "loss": 1.9559, "step": 552 }, { "epoch": 0.05400390625, "grad_norm": 0.2837629020214081, "learning_rate": 0.0004992810658310052, "loss": 1.9603, "step": 553 }, { "epoch": 0.0541015625, "grad_norm": 0.2644243836402893, "learning_rate": 0.0004992753743837193, "loss": 1.9767, "step": 554 }, { "epoch": 0.05419921875, "grad_norm": 0.21815134584903717, "learning_rate": 0.0004992696605333103, "loss": 1.9684, "step": 555 }, { "epoch": 0.054296875, "grad_norm": 0.28167852759361267, "learning_rate": 0.0004992639242803492, "loss": 1.97, "step": 556 }, { "epoch": 0.05439453125, "grad_norm": 0.26942357420921326, "learning_rate": 0.0004992581656254087, "loss": 1.9825, "step": 557 }, { "epoch": 0.0544921875, "grad_norm": 0.25316697359085083, "learning_rate": 0.0004992523845690644, "loss": 1.9698, "step": 558 }, { "epoch": 0.05458984375, "grad_norm": 0.29587239027023315, "learning_rate": 0.0004992465811118934, "loss": 2.0063, "step": 559 }, { "epoch": 0.0546875, "grad_norm": 0.3209396004676819, "learning_rate": 0.0004992407552544757, "loss": 2.0002, "step": 560 }, { "epoch": 0.05478515625, "grad_norm": 0.2541654109954834, "learning_rate": 0.0004992349069973931, "loss": 2.0039, "step": 561 }, { "epoch": 0.0548828125, "grad_norm": 0.34102505445480347, "learning_rate": 0.0004992290363412298, "loss": 1.9181, "step": 562 }, { "epoch": 0.05498046875, "grad_norm": 0.36123788356781006, "learning_rate": 0.0004992231432865723, "loss": 2.042, "step": 563 }, { "epoch": 0.055078125, "grad_norm": 0.27344051003456116, "learning_rate": 0.0004992172278340093, "loss": 2.0236, "step": 564 }, { "epoch": 0.05517578125, "grad_norm": 0.34781453013420105, "learning_rate": 0.0004992112899841315, "loss": 2.0017, "step": 565 }, { "epoch": 0.0552734375, "grad_norm": 0.2822319567203522, "learning_rate": 0.0004992053297375322, "loss": 1.9607, "step": 566 }, { "epoch": 0.05537109375, "grad_norm": 0.27945882081985474, "learning_rate": 0.0004991993470948066, "loss": 1.9707, "step": 567 }, { "epoch": 0.05546875, "grad_norm": 0.31535595655441284, "learning_rate": 0.0004991933420565527, "loss": 1.9708, "step": 568 }, { "epoch": 0.05556640625, "grad_norm": 0.2658466398715973, "learning_rate": 0.00049918731462337, "loss": 1.9763, "step": 569 }, { "epoch": 0.0556640625, "grad_norm": 0.3618657886981964, "learning_rate": 0.0004991812647958607, "loss": 2.0154, "step": 570 }, { "epoch": 0.05576171875, "grad_norm": 0.34552058577537537, "learning_rate": 0.000499175192574629, "loss": 1.978, "step": 571 }, { "epoch": 0.055859375, "grad_norm": 0.32903313636779785, "learning_rate": 0.0004991690979602817, "loss": 1.9762, "step": 572 }, { "epoch": 0.05595703125, "grad_norm": 0.30814382433891296, "learning_rate": 0.0004991629809534275, "loss": 1.9944, "step": 573 }, { "epoch": 0.0560546875, "grad_norm": 0.21156492829322815, "learning_rate": 0.0004991568415546775, "loss": 1.9919, "step": 574 }, { "epoch": 0.05615234375, "grad_norm": 0.32283446192741394, "learning_rate": 0.0004991506797646446, "loss": 1.9697, "step": 575 }, { "epoch": 0.05625, "grad_norm": 0.2915102541446686, "learning_rate": 0.0004991444955839447, "loss": 1.9736, "step": 576 }, { "epoch": 0.05634765625, "grad_norm": 0.28996148705482483, "learning_rate": 0.0004991382890131955, "loss": 2.002, "step": 577 }, { "epoch": 0.0564453125, "grad_norm": 0.3859401345252991, "learning_rate": 0.0004991320600530168, "loss": 1.9753, "step": 578 }, { "epoch": 0.05654296875, "grad_norm": 0.28169432282447815, "learning_rate": 0.0004991258087040312, "loss": 1.9798, "step": 579 }, { "epoch": 0.056640625, "grad_norm": 0.30669984221458435, "learning_rate": 0.0004991195349668626, "loss": 2.0007, "step": 580 }, { "epoch": 0.05673828125, "grad_norm": 0.26832297444343567, "learning_rate": 0.000499113238842138, "loss": 2.0146, "step": 581 }, { "epoch": 0.0568359375, "grad_norm": 0.25695785880088806, "learning_rate": 0.0004991069203304865, "loss": 1.9814, "step": 582 }, { "epoch": 0.05693359375, "grad_norm": 0.25700533390045166, "learning_rate": 0.0004991005794325389, "loss": 1.9623, "step": 583 }, { "epoch": 0.05703125, "grad_norm": 0.20786869525909424, "learning_rate": 0.0004990942161489288, "loss": 1.9728, "step": 584 }, { "epoch": 0.05712890625, "grad_norm": 0.23988646268844604, "learning_rate": 0.0004990878304802918, "loss": 1.9749, "step": 585 }, { "epoch": 0.0572265625, "grad_norm": 0.2418268769979477, "learning_rate": 0.0004990814224272658, "loss": 1.928, "step": 586 }, { "epoch": 0.05732421875, "grad_norm": 0.27739882469177246, "learning_rate": 0.0004990749919904909, "loss": 1.964, "step": 587 }, { "epoch": 0.057421875, "grad_norm": 0.2773842513561249, "learning_rate": 0.0004990685391706094, "loss": 1.9589, "step": 588 }, { "epoch": 0.05751953125, "grad_norm": 0.24483682215213776, "learning_rate": 0.0004990620639682659, "loss": 1.9965, "step": 589 }, { "epoch": 0.0576171875, "grad_norm": 0.27345889806747437, "learning_rate": 0.0004990555663841071, "loss": 1.9444, "step": 590 }, { "epoch": 0.05771484375, "grad_norm": 0.26006460189819336, "learning_rate": 0.0004990490464187824, "loss": 1.9391, "step": 591 }, { "epoch": 0.0578125, "grad_norm": 0.1935306191444397, "learning_rate": 0.0004990425040729427, "loss": 1.9679, "step": 592 }, { "epoch": 0.05791015625, "grad_norm": 0.26823100447654724, "learning_rate": 0.0004990359393472418, "loss": 2.0055, "step": 593 }, { "epoch": 0.0580078125, "grad_norm": 0.2706959545612335, "learning_rate": 0.0004990293522423352, "loss": 1.9853, "step": 594 }, { "epoch": 0.05810546875, "grad_norm": 0.2099694162607193, "learning_rate": 0.0004990227427588811, "loss": 1.959, "step": 595 }, { "epoch": 0.058203125, "grad_norm": 0.22345881164073944, "learning_rate": 0.0004990161108975398, "loss": 1.9867, "step": 596 }, { "epoch": 0.05830078125, "grad_norm": 0.34050315618515015, "learning_rate": 0.0004990094566589734, "loss": 1.9555, "step": 597 }, { "epoch": 0.0583984375, "grad_norm": 0.5298358798027039, "learning_rate": 0.0004990027800438468, "loss": 1.9824, "step": 598 }, { "epoch": 0.05849609375, "grad_norm": 0.6408056616783142, "learning_rate": 0.0004989960810528271, "loss": 1.9647, "step": 599 }, { "epoch": 0.05859375, "grad_norm": 0.4058885872364044, "learning_rate": 0.0004989893596865833, "loss": 1.9564, "step": 600 }, { "epoch": 0.05869140625, "grad_norm": 0.2911423146724701, "learning_rate": 0.0004989826159457869, "loss": 1.9485, "step": 601 }, { "epoch": 0.0587890625, "grad_norm": 0.40600210428237915, "learning_rate": 0.0004989758498311114, "loss": 1.993, "step": 602 }, { "epoch": 0.05888671875, "grad_norm": 0.3920167088508606, "learning_rate": 0.0004989690613432327, "loss": 2.0188, "step": 603 }, { "epoch": 0.058984375, "grad_norm": 0.2924116551876068, "learning_rate": 0.0004989622504828291, "loss": 1.9749, "step": 604 }, { "epoch": 0.05908203125, "grad_norm": 0.33497628569602966, "learning_rate": 0.0004989554172505807, "loss": 1.9358, "step": 605 }, { "epoch": 0.0591796875, "grad_norm": 0.31751033663749695, "learning_rate": 0.0004989485616471702, "loss": 1.9886, "step": 606 }, { "epoch": 0.05927734375, "grad_norm": 0.2204602211713791, "learning_rate": 0.0004989416836732825, "loss": 1.9715, "step": 607 }, { "epoch": 0.059375, "grad_norm": 0.2980511784553528, "learning_rate": 0.0004989347833296044, "loss": 1.9479, "step": 608 }, { "epoch": 0.05947265625, "grad_norm": 0.2874164879322052, "learning_rate": 0.0004989278606168253, "loss": 1.9717, "step": 609 }, { "epoch": 0.0595703125, "grad_norm": 0.2526591718196869, "learning_rate": 0.0004989209155356368, "loss": 2.0009, "step": 610 }, { "epoch": 0.05966796875, "grad_norm": 0.2890442907810211, "learning_rate": 0.0004989139480867327, "loss": 1.9862, "step": 611 }, { "epoch": 0.059765625, "grad_norm": 0.26463401317596436, "learning_rate": 0.0004989069582708088, "loss": 2.0143, "step": 612 }, { "epoch": 0.05986328125, "grad_norm": 0.19906644523143768, "learning_rate": 0.0004988999460885634, "loss": 1.9845, "step": 613 }, { "epoch": 0.0599609375, "grad_norm": 0.28752684593200684, "learning_rate": 0.000498892911540697, "loss": 1.9421, "step": 614 }, { "epoch": 0.06005859375, "grad_norm": 0.2739415466785431, "learning_rate": 0.0004988858546279123, "loss": 1.967, "step": 615 }, { "epoch": 0.06015625, "grad_norm": 0.27040478587150574, "learning_rate": 0.000498878775350914, "loss": 2.0001, "step": 616 }, { "epoch": 0.06025390625, "grad_norm": 0.28345200419425964, "learning_rate": 0.0004988716737104096, "loss": 1.9637, "step": 617 }, { "epoch": 0.0603515625, "grad_norm": 0.2515316903591156, "learning_rate": 0.0004988645497071082, "loss": 1.9511, "step": 618 }, { "epoch": 0.06044921875, "grad_norm": 0.22784312069416046, "learning_rate": 0.0004988574033417216, "loss": 1.9752, "step": 619 }, { "epoch": 0.060546875, "grad_norm": 0.23128612339496613, "learning_rate": 0.0004988502346149636, "loss": 1.9803, "step": 620 }, { "epoch": 0.06064453125, "grad_norm": 0.22022667527198792, "learning_rate": 0.0004988430435275503, "loss": 1.9592, "step": 621 }, { "epoch": 0.0607421875, "grad_norm": 0.2470719963312149, "learning_rate": 0.0004988358300802, "loss": 2.0063, "step": 622 }, { "epoch": 0.06083984375, "grad_norm": 0.22507692873477936, "learning_rate": 0.0004988285942736332, "loss": 1.9873, "step": 623 }, { "epoch": 0.0609375, "grad_norm": 0.2285120040178299, "learning_rate": 0.0004988213361085729, "loss": 1.973, "step": 624 }, { "epoch": 0.06103515625, "grad_norm": 0.2436297982931137, "learning_rate": 0.000498814055585744, "loss": 1.9786, "step": 625 }, { "epoch": 0.0611328125, "grad_norm": 0.2736089825630188, "learning_rate": 0.0004988067527058737, "loss": 1.9552, "step": 626 }, { "epoch": 0.06123046875, "grad_norm": 0.2813490033149719, "learning_rate": 0.0004987994274696917, "loss": 2.0173, "step": 627 }, { "epoch": 0.061328125, "grad_norm": 0.3024250864982605, "learning_rate": 0.0004987920798779294, "loss": 1.9232, "step": 628 }, { "epoch": 0.06142578125, "grad_norm": 0.22795617580413818, "learning_rate": 0.000498784709931321, "loss": 1.9981, "step": 629 }, { "epoch": 0.0615234375, "grad_norm": 0.2585217356681824, "learning_rate": 0.0004987773176306026, "loss": 2.0197, "step": 630 }, { "epoch": 0.06162109375, "grad_norm": 0.27120688557624817, "learning_rate": 0.0004987699029765127, "loss": 1.961, "step": 631 }, { "epoch": 0.06171875, "grad_norm": 0.23527094721794128, "learning_rate": 0.000498762465969792, "loss": 1.9996, "step": 632 }, { "epoch": 0.06181640625, "grad_norm": 0.2519530653953552, "learning_rate": 0.0004987550066111832, "loss": 1.9926, "step": 633 }, { "epoch": 0.0619140625, "grad_norm": 0.3199564218521118, "learning_rate": 0.0004987475249014315, "loss": 2.0066, "step": 634 }, { "epoch": 0.06201171875, "grad_norm": 0.3830641806125641, "learning_rate": 0.0004987400208412843, "loss": 1.9368, "step": 635 }, { "epoch": 0.062109375, "grad_norm": 0.3777483403682709, "learning_rate": 0.0004987324944314913, "loss": 1.9792, "step": 636 }, { "epoch": 0.06220703125, "grad_norm": 0.420107901096344, "learning_rate": 0.0004987249456728041, "loss": 1.9765, "step": 637 }, { "epoch": 0.0623046875, "grad_norm": 0.3748738467693329, "learning_rate": 0.0004987173745659768, "loss": 1.9826, "step": 638 }, { "epoch": 0.06240234375, "grad_norm": 0.3142974078655243, "learning_rate": 0.0004987097811117658, "loss": 1.9393, "step": 639 }, { "epoch": 0.0625, "grad_norm": 0.24331647157669067, "learning_rate": 0.0004987021653109296, "loss": 1.9305, "step": 640 }, { "epoch": 0.06259765625, "grad_norm": 0.2792535424232483, "learning_rate": 0.0004986945271642289, "loss": 1.9437, "step": 641 }, { "epoch": 0.0626953125, "grad_norm": 0.24984711408615112, "learning_rate": 0.0004986868666724267, "loss": 1.9838, "step": 642 }, { "epoch": 0.06279296875, "grad_norm": 0.2477482557296753, "learning_rate": 0.0004986791838362881, "loss": 1.9981, "step": 643 }, { "epoch": 0.062890625, "grad_norm": 0.3012976348400116, "learning_rate": 0.0004986714786565807, "loss": 1.9963, "step": 644 }, { "epoch": 0.06298828125, "grad_norm": 0.2480735331773758, "learning_rate": 0.0004986637511340742, "loss": 1.9567, "step": 645 }, { "epoch": 0.0630859375, "grad_norm": 0.20119979977607727, "learning_rate": 0.0004986560012695403, "loss": 2.0012, "step": 646 }, { "epoch": 0.06318359375, "grad_norm": 0.27725833654403687, "learning_rate": 0.0004986482290637535, "loss": 1.9751, "step": 647 }, { "epoch": 0.06328125, "grad_norm": 0.27348190546035767, "learning_rate": 0.0004986404345174898, "loss": 1.976, "step": 648 }, { "epoch": 0.06337890625, "grad_norm": 0.29639676213264465, "learning_rate": 0.000498632617631528, "loss": 1.9835, "step": 649 }, { "epoch": 0.0634765625, "grad_norm": 0.2911996841430664, "learning_rate": 0.000498624778406649, "loss": 2.0107, "step": 650 }, { "epoch": 0.06357421875, "grad_norm": 0.30200451612472534, "learning_rate": 0.0004986169168436355, "loss": 1.9433, "step": 651 }, { "epoch": 0.063671875, "grad_norm": 0.288396954536438, "learning_rate": 0.0004986090329432734, "loss": 1.9759, "step": 652 }, { "epoch": 0.06376953125, "grad_norm": 0.26006969809532166, "learning_rate": 0.0004986011267063497, "loss": 1.977, "step": 653 }, { "epoch": 0.0638671875, "grad_norm": 0.2467355877161026, "learning_rate": 0.0004985931981336544, "loss": 1.9892, "step": 654 }, { "epoch": 0.06396484375, "grad_norm": 0.2180376946926117, "learning_rate": 0.0004985852472259797, "loss": 2.0012, "step": 655 }, { "epoch": 0.0640625, "grad_norm": 0.23811106383800507, "learning_rate": 0.0004985772739841193, "loss": 1.969, "step": 656 }, { "epoch": 0.06416015625, "grad_norm": 0.2517051696777344, "learning_rate": 0.00049856927840887, "loss": 1.9607, "step": 657 }, { "epoch": 0.0642578125, "grad_norm": 0.2900746464729309, "learning_rate": 0.0004985612605010305, "loss": 1.9865, "step": 658 }, { "epoch": 0.06435546875, "grad_norm": 0.2779560983181, "learning_rate": 0.0004985532202614017, "loss": 1.9787, "step": 659 }, { "epoch": 0.064453125, "grad_norm": 0.2686167061328888, "learning_rate": 0.0004985451576907865, "loss": 1.9279, "step": 660 }, { "epoch": 0.06455078125, "grad_norm": 0.25136813521385193, "learning_rate": 0.0004985370727899907, "loss": 1.9502, "step": 661 }, { "epoch": 0.0646484375, "grad_norm": 0.26508840918540955, "learning_rate": 0.0004985289655598216, "loss": 1.9588, "step": 662 }, { "epoch": 0.06474609375, "grad_norm": 0.2178860753774643, "learning_rate": 0.000498520836001089, "loss": 1.9445, "step": 663 }, { "epoch": 0.06484375, "grad_norm": 0.21833141148090363, "learning_rate": 0.0004985126841146052, "loss": 1.9963, "step": 664 }, { "epoch": 0.06494140625, "grad_norm": 0.21572785079479218, "learning_rate": 0.0004985045099011844, "loss": 1.9789, "step": 665 }, { "epoch": 0.0650390625, "grad_norm": 0.23386414349079132, "learning_rate": 0.0004984963133616432, "loss": 1.9829, "step": 666 }, { "epoch": 0.06513671875, "grad_norm": 0.25145047903060913, "learning_rate": 0.0004984880944968003, "loss": 2.0106, "step": 667 }, { "epoch": 0.065234375, "grad_norm": 0.23240795731544495, "learning_rate": 0.0004984798533074767, "loss": 1.999, "step": 668 }, { "epoch": 0.06533203125, "grad_norm": 0.22223520278930664, "learning_rate": 0.0004984715897944954, "loss": 1.9872, "step": 669 }, { "epoch": 0.0654296875, "grad_norm": 0.2768714725971222, "learning_rate": 0.0004984633039586823, "loss": 1.9656, "step": 670 }, { "epoch": 0.06552734375, "grad_norm": 0.35761523246765137, "learning_rate": 0.0004984549958008646, "loss": 1.9984, "step": 671 }, { "epoch": 0.065625, "grad_norm": 0.4054514169692993, "learning_rate": 0.0004984466653218726, "loss": 1.9406, "step": 672 }, { "epoch": 0.06572265625, "grad_norm": 0.40501922369003296, "learning_rate": 0.0004984383125225383, "loss": 1.9682, "step": 673 }, { "epoch": 0.0658203125, "grad_norm": 0.31225842237472534, "learning_rate": 0.0004984299374036961, "loss": 1.9833, "step": 674 }, { "epoch": 0.06591796875, "grad_norm": 0.24159114062786102, "learning_rate": 0.0004984215399661825, "loss": 1.9634, "step": 675 }, { "epoch": 0.066015625, "grad_norm": 0.2537291944026947, "learning_rate": 0.0004984131202108364, "loss": 1.9563, "step": 676 }, { "epoch": 0.06611328125, "grad_norm": 0.29919305443763733, "learning_rate": 0.0004984046781384988, "loss": 1.9594, "step": 677 }, { "epoch": 0.0662109375, "grad_norm": 0.28458645939826965, "learning_rate": 0.0004983962137500133, "loss": 1.9313, "step": 678 }, { "epoch": 0.06630859375, "grad_norm": 0.2523960471153259, "learning_rate": 0.0004983877270462249, "loss": 1.9616, "step": 679 }, { "epoch": 0.06640625, "grad_norm": 0.2605716288089752, "learning_rate": 0.0004983792180279816, "loss": 1.9631, "step": 680 }, { "epoch": 0.06650390625, "grad_norm": 0.280000239610672, "learning_rate": 0.0004983706866961335, "loss": 1.9945, "step": 681 }, { "epoch": 0.0666015625, "grad_norm": 0.2881770730018616, "learning_rate": 0.0004983621330515326, "loss": 1.9639, "step": 682 }, { "epoch": 0.06669921875, "grad_norm": 0.27240094542503357, "learning_rate": 0.0004983535570950334, "loss": 1.9226, "step": 683 }, { "epoch": 0.066796875, "grad_norm": 0.2580428123474121, "learning_rate": 0.0004983449588274926, "loss": 1.9625, "step": 684 }, { "epoch": 0.06689453125, "grad_norm": 0.2719865143299103, "learning_rate": 0.0004983363382497692, "loss": 1.9941, "step": 685 }, { "epoch": 0.0669921875, "grad_norm": 0.2593652904033661, "learning_rate": 0.000498327695362724, "loss": 1.9558, "step": 686 }, { "epoch": 0.06708984375, "grad_norm": 0.2575826346874237, "learning_rate": 0.0004983190301672207, "loss": 2.0188, "step": 687 }, { "epoch": 0.0671875, "grad_norm": 0.23726949095726013, "learning_rate": 0.0004983103426641246, "loss": 1.9568, "step": 688 }, { "epoch": 0.06728515625, "grad_norm": 0.3027707040309906, "learning_rate": 0.0004983016328543037, "loss": 1.9708, "step": 689 }, { "epoch": 0.0673828125, "grad_norm": 0.29431256651878357, "learning_rate": 0.0004982929007386279, "loss": 1.9668, "step": 690 }, { "epoch": 0.06748046875, "grad_norm": 0.24073566496372223, "learning_rate": 0.0004982841463179694, "loss": 1.9746, "step": 691 }, { "epoch": 0.067578125, "grad_norm": 0.27540677785873413, "learning_rate": 0.000498275369593203, "loss": 1.868, "step": 692 }, { "epoch": 0.06767578125, "grad_norm": 0.2665644884109497, "learning_rate": 0.0004982665705652052, "loss": 1.9528, "step": 693 }, { "epoch": 0.0677734375, "grad_norm": 0.2117265909910202, "learning_rate": 0.0004982577492348549, "loss": 1.9977, "step": 694 }, { "epoch": 0.06787109375, "grad_norm": 0.30092543363571167, "learning_rate": 0.0004982489056030334, "loss": 1.9636, "step": 695 }, { "epoch": 0.06796875, "grad_norm": 0.2956363558769226, "learning_rate": 0.0004982400396706239, "loss": 1.9519, "step": 696 }, { "epoch": 0.06806640625, "grad_norm": 0.3209246098995209, "learning_rate": 0.0004982311514385124, "loss": 1.9761, "step": 697 }, { "epoch": 0.0681640625, "grad_norm": 0.4527253806591034, "learning_rate": 0.0004982222409075863, "loss": 1.9405, "step": 698 }, { "epoch": 0.06826171875, "grad_norm": 0.4838014245033264, "learning_rate": 0.000498213308078736, "loss": 1.9434, "step": 699 }, { "epoch": 0.068359375, "grad_norm": 0.4865509569644928, "learning_rate": 0.0004982043529528536, "loss": 1.9706, "step": 700 }, { "epoch": 0.06845703125, "grad_norm": 0.38092029094696045, "learning_rate": 0.0004981953755308339, "loss": 1.9168, "step": 701 }, { "epoch": 0.0685546875, "grad_norm": 0.24092066287994385, "learning_rate": 0.0004981863758135735, "loss": 1.9693, "step": 702 }, { "epoch": 0.06865234375, "grad_norm": 0.3893338143825531, "learning_rate": 0.0004981773538019714, "loss": 1.9812, "step": 703 }, { "epoch": 0.06875, "grad_norm": 0.3019930422306061, "learning_rate": 0.0004981683094969287, "loss": 1.9207, "step": 704 }, { "epoch": 0.06884765625, "grad_norm": 0.2746503949165344, "learning_rate": 0.0004981592428993492, "loss": 1.9416, "step": 705 }, { "epoch": 0.0689453125, "grad_norm": 0.2998049557209015, "learning_rate": 0.0004981501540101382, "loss": 1.9685, "step": 706 }, { "epoch": 0.06904296875, "grad_norm": 0.2986246645450592, "learning_rate": 0.0004981410428302037, "loss": 1.9827, "step": 707 }, { "epoch": 0.069140625, "grad_norm": 0.28714433312416077, "learning_rate": 0.000498131909360456, "loss": 1.904, "step": 708 }, { "epoch": 0.06923828125, "grad_norm": 0.21930216252803802, "learning_rate": 0.0004981227536018073, "loss": 1.966, "step": 709 }, { "epoch": 0.0693359375, "grad_norm": 0.2309812754392624, "learning_rate": 0.0004981135755551722, "loss": 1.9983, "step": 710 }, { "epoch": 0.06943359375, "grad_norm": 0.2016872614622116, "learning_rate": 0.0004981043752214674, "loss": 1.9382, "step": 711 }, { "epoch": 0.06953125, "grad_norm": 0.235287606716156, "learning_rate": 0.0004980951526016122, "loss": 1.9914, "step": 712 }, { "epoch": 0.06962890625, "grad_norm": 0.23780131340026855, "learning_rate": 0.0004980859076965276, "loss": 1.9404, "step": 713 }, { "epoch": 0.0697265625, "grad_norm": 0.2195412665605545, "learning_rate": 0.0004980766405071372, "loss": 1.9526, "step": 714 }, { "epoch": 0.06982421875, "grad_norm": 0.2391882836818695, "learning_rate": 0.0004980673510343668, "loss": 1.9179, "step": 715 }, { "epoch": 0.069921875, "grad_norm": 0.27623385190963745, "learning_rate": 0.000498058039279144, "loss": 1.9519, "step": 716 }, { "epoch": 0.07001953125, "grad_norm": 0.27679604291915894, "learning_rate": 0.0004980487052423994, "loss": 2.0292, "step": 717 }, { "epoch": 0.0701171875, "grad_norm": 0.2968669533729553, "learning_rate": 0.0004980393489250651, "loss": 1.9162, "step": 718 }, { "epoch": 0.07021484375, "grad_norm": 0.2479199320077896, "learning_rate": 0.0004980299703280758, "loss": 1.9544, "step": 719 }, { "epoch": 0.0703125, "grad_norm": 0.2597293555736542, "learning_rate": 0.0004980205694523682, "loss": 1.9273, "step": 720 }, { "epoch": 0.07041015625, "grad_norm": 0.23868051171302795, "learning_rate": 0.0004980111462988817, "loss": 1.9527, "step": 721 }, { "epoch": 0.0705078125, "grad_norm": 0.28150075674057007, "learning_rate": 0.0004980017008685573, "loss": 1.9523, "step": 722 }, { "epoch": 0.07060546875, "grad_norm": 0.3689873516559601, "learning_rate": 0.0004979922331623387, "loss": 1.9167, "step": 723 }, { "epoch": 0.070703125, "grad_norm": 0.380206823348999, "learning_rate": 0.0004979827431811713, "loss": 1.9025, "step": 724 }, { "epoch": 0.07080078125, "grad_norm": 0.2919938266277313, "learning_rate": 0.0004979732309260036, "loss": 1.9996, "step": 725 }, { "epoch": 0.0708984375, "grad_norm": 0.302680104970932, "learning_rate": 0.0004979636963977853, "loss": 1.95, "step": 726 }, { "epoch": 0.07099609375, "grad_norm": 0.25307831168174744, "learning_rate": 0.0004979541395974691, "loss": 1.9723, "step": 727 }, { "epoch": 0.07109375, "grad_norm": 0.304147869348526, "learning_rate": 0.0004979445605260096, "loss": 2.0359, "step": 728 }, { "epoch": 0.07119140625, "grad_norm": 0.35432276129722595, "learning_rate": 0.0004979349591843635, "loss": 1.9384, "step": 729 }, { "epoch": 0.0712890625, "grad_norm": 0.23866906762123108, "learning_rate": 0.00049792533557349, "loss": 1.9341, "step": 730 }, { "epoch": 0.07138671875, "grad_norm": 0.30183711647987366, "learning_rate": 0.0004979156896943504, "loss": 1.9603, "step": 731 }, { "epoch": 0.071484375, "grad_norm": 0.30881768465042114, "learning_rate": 0.0004979060215479083, "loss": 1.982, "step": 732 }, { "epoch": 0.07158203125, "grad_norm": 0.2996484041213989, "learning_rate": 0.0004978963311351295, "loss": 1.9536, "step": 733 }, { "epoch": 0.0716796875, "grad_norm": 0.33594587445259094, "learning_rate": 0.0004978866184569818, "loss": 1.9813, "step": 734 }, { "epoch": 0.07177734375, "grad_norm": 0.30377864837646484, "learning_rate": 0.0004978768835144355, "loss": 1.9417, "step": 735 }, { "epoch": 0.071875, "grad_norm": 0.34858009219169617, "learning_rate": 0.0004978671263084631, "loss": 1.9368, "step": 736 }, { "epoch": 0.07197265625, "grad_norm": 0.28119683265686035, "learning_rate": 0.0004978573468400392, "loss": 1.9223, "step": 737 }, { "epoch": 0.0720703125, "grad_norm": 0.2478085160255432, "learning_rate": 0.0004978475451101408, "loss": 1.9744, "step": 738 }, { "epoch": 0.07216796875, "grad_norm": 0.2461906224489212, "learning_rate": 0.0004978377211197466, "loss": 1.961, "step": 739 }, { "epoch": 0.072265625, "grad_norm": 0.2495601922273636, "learning_rate": 0.0004978278748698385, "loss": 1.9602, "step": 740 }, { "epoch": 0.07236328125, "grad_norm": 0.3028644323348999, "learning_rate": 0.0004978180063613997, "loss": 1.95, "step": 741 }, { "epoch": 0.0724609375, "grad_norm": 0.3690338730812073, "learning_rate": 0.000497808115595416, "loss": 1.9625, "step": 742 }, { "epoch": 0.07255859375, "grad_norm": 0.3791324198246002, "learning_rate": 0.0004977982025728755, "loss": 1.9747, "step": 743 }, { "epoch": 0.07265625, "grad_norm": 0.29515188932418823, "learning_rate": 0.0004977882672947683, "loss": 1.9774, "step": 744 }, { "epoch": 0.07275390625, "grad_norm": 0.29207196831703186, "learning_rate": 0.000497778309762087, "loss": 1.9704, "step": 745 }, { "epoch": 0.0728515625, "grad_norm": 0.23698067665100098, "learning_rate": 0.0004977683299758262, "loss": 1.978, "step": 746 }, { "epoch": 0.07294921875, "grad_norm": 0.2712165415287018, "learning_rate": 0.0004977583279369827, "loss": 1.9524, "step": 747 }, { "epoch": 0.073046875, "grad_norm": 0.32161638140678406, "learning_rate": 0.0004977483036465558, "loss": 1.9532, "step": 748 }, { "epoch": 0.07314453125, "grad_norm": 0.25917574763298035, "learning_rate": 0.0004977382571055467, "loss": 1.9307, "step": 749 }, { "epoch": 0.0732421875, "grad_norm": 0.27562615275382996, "learning_rate": 0.000497728188314959, "loss": 1.9842, "step": 750 }, { "epoch": 0.07333984375, "grad_norm": 0.23242716491222382, "learning_rate": 0.0004977180972757986, "loss": 1.9097, "step": 751 }, { "epoch": 0.0734375, "grad_norm": 0.3111313283443451, "learning_rate": 0.0004977079839890732, "loss": 1.9812, "step": 752 }, { "epoch": 0.07353515625, "grad_norm": 0.3295767605304718, "learning_rate": 0.0004976978484557933, "loss": 1.9206, "step": 753 }, { "epoch": 0.0736328125, "grad_norm": 0.3592395782470703, "learning_rate": 0.0004976876906769713, "loss": 1.9817, "step": 754 }, { "epoch": 0.07373046875, "grad_norm": 0.3530953824520111, "learning_rate": 0.0004976775106536218, "loss": 1.9831, "step": 755 }, { "epoch": 0.073828125, "grad_norm": 0.3064524829387665, "learning_rate": 0.0004976673083867617, "loss": 1.9207, "step": 756 }, { "epoch": 0.07392578125, "grad_norm": 0.281597763299942, "learning_rate": 0.0004976570838774102, "loss": 1.9781, "step": 757 }, { "epoch": 0.0740234375, "grad_norm": 0.296041876077652, "learning_rate": 0.0004976468371265886, "loss": 1.9708, "step": 758 }, { "epoch": 0.07412109375, "grad_norm": 0.3518696427345276, "learning_rate": 0.0004976365681353203, "loss": 1.9333, "step": 759 }, { "epoch": 0.07421875, "grad_norm": 0.21943148970603943, "learning_rate": 0.0004976262769046313, "loss": 1.9416, "step": 760 }, { "epoch": 0.07431640625, "grad_norm": 0.2540123760700226, "learning_rate": 0.0004976159634355495, "loss": 1.989, "step": 761 }, { "epoch": 0.0744140625, "grad_norm": 0.3120715320110321, "learning_rate": 0.0004976056277291053, "loss": 1.95, "step": 762 }, { "epoch": 0.07451171875, "grad_norm": 0.23797550797462463, "learning_rate": 0.0004975952697863308, "loss": 1.9534, "step": 763 }, { "epoch": 0.074609375, "grad_norm": 0.30815449357032776, "learning_rate": 0.000497584889608261, "loss": 2.0283, "step": 764 }, { "epoch": 0.07470703125, "grad_norm": 0.37625932693481445, "learning_rate": 0.0004975744871959327, "loss": 1.9658, "step": 765 }, { "epoch": 0.0748046875, "grad_norm": 0.27407678961753845, "learning_rate": 0.0004975640625503848, "loss": 1.9458, "step": 766 }, { "epoch": 0.07490234375, "grad_norm": 0.26577049493789673, "learning_rate": 0.000497553615672659, "loss": 1.9667, "step": 767 }, { "epoch": 0.075, "grad_norm": 0.29821062088012695, "learning_rate": 0.0004975431465637983, "loss": 1.9805, "step": 768 }, { "epoch": 0.07509765625, "grad_norm": 0.2582527995109558, "learning_rate": 0.0004975326552248491, "loss": 1.9904, "step": 769 }, { "epoch": 0.0751953125, "grad_norm": 0.2533649802207947, "learning_rate": 0.0004975221416568591, "loss": 1.9383, "step": 770 }, { "epoch": 0.07529296875, "grad_norm": 0.253592312335968, "learning_rate": 0.0004975116058608784, "loss": 1.9855, "step": 771 }, { "epoch": 0.075390625, "grad_norm": 0.2658921182155609, "learning_rate": 0.0004975010478379596, "loss": 1.9845, "step": 772 }, { "epoch": 0.07548828125, "grad_norm": 0.25861868262290955, "learning_rate": 0.0004974904675891573, "loss": 1.9776, "step": 773 }, { "epoch": 0.0755859375, "grad_norm": 0.272682249546051, "learning_rate": 0.0004974798651155284, "loss": 1.9357, "step": 774 }, { "epoch": 0.07568359375, "grad_norm": 0.2574002742767334, "learning_rate": 0.0004974692404181319, "loss": 1.9677, "step": 775 }, { "epoch": 0.07578125, "grad_norm": 0.3008319139480591, "learning_rate": 0.0004974585934980293, "loss": 1.983, "step": 776 }, { "epoch": 0.07587890625, "grad_norm": 0.3242790699005127, "learning_rate": 0.0004974479243562839, "loss": 1.9846, "step": 777 }, { "epoch": 0.0759765625, "grad_norm": 0.29044070839881897, "learning_rate": 0.0004974372329939617, "loss": 2.0012, "step": 778 }, { "epoch": 0.07607421875, "grad_norm": 0.32275035977363586, "learning_rate": 0.0004974265194121303, "loss": 1.9842, "step": 779 }, { "epoch": 0.076171875, "grad_norm": 0.31752893328666687, "learning_rate": 0.0004974157836118604, "loss": 1.9368, "step": 780 }, { "epoch": 0.07626953125, "grad_norm": 0.27947962284088135, "learning_rate": 0.0004974050255942239, "loss": 1.9794, "step": 781 }, { "epoch": 0.0763671875, "grad_norm": 0.29937905073165894, "learning_rate": 0.000497394245360296, "loss": 1.9597, "step": 782 }, { "epoch": 0.07646484375, "grad_norm": 0.2969702482223511, "learning_rate": 0.0004973834429111529, "loss": 1.945, "step": 783 }, { "epoch": 0.0765625, "grad_norm": 0.3729354441165924, "learning_rate": 0.0004973726182478741, "loss": 1.9032, "step": 784 }, { "epoch": 0.07666015625, "grad_norm": 0.32856106758117676, "learning_rate": 0.0004973617713715408, "loss": 1.9551, "step": 785 }, { "epoch": 0.0767578125, "grad_norm": 0.23239392042160034, "learning_rate": 0.0004973509022832364, "loss": 1.9632, "step": 786 }, { "epoch": 0.07685546875, "grad_norm": 0.23353087902069092, "learning_rate": 0.000497340010984047, "loss": 1.9514, "step": 787 }, { "epoch": 0.076953125, "grad_norm": 0.23280592262744904, "learning_rate": 0.00049732909747506, "loss": 1.9696, "step": 788 }, { "epoch": 0.07705078125, "grad_norm": 0.2193138748407364, "learning_rate": 0.0004973181617573658, "loss": 1.9734, "step": 789 }, { "epoch": 0.0771484375, "grad_norm": 0.3234734833240509, "learning_rate": 0.0004973072038320569, "loss": 1.9874, "step": 790 }, { "epoch": 0.07724609375, "grad_norm": 0.35144004225730896, "learning_rate": 0.0004972962237002278, "loss": 1.9331, "step": 791 }, { "epoch": 0.07734375, "grad_norm": 0.37716230750083923, "learning_rate": 0.0004972852213629753, "loss": 1.9731, "step": 792 }, { "epoch": 0.07744140625, "grad_norm": 0.3221200108528137, "learning_rate": 0.0004972741968213985, "loss": 2.0043, "step": 793 }, { "epoch": 0.0775390625, "grad_norm": 0.23167745769023895, "learning_rate": 0.0004972631500765985, "loss": 1.9634, "step": 794 }, { "epoch": 0.07763671875, "grad_norm": 0.4177049398422241, "learning_rate": 0.0004972520811296789, "loss": 1.9533, "step": 795 }, { "epoch": 0.077734375, "grad_norm": 0.39038336277008057, "learning_rate": 0.0004972409899817455, "loss": 1.9508, "step": 796 }, { "epoch": 0.07783203125, "grad_norm": 0.4033074676990509, "learning_rate": 0.0004972298766339061, "loss": 1.965, "step": 797 }, { "epoch": 0.0779296875, "grad_norm": 0.3271097242832184, "learning_rate": 0.0004972187410872707, "loss": 1.9893, "step": 798 }, { "epoch": 0.07802734375, "grad_norm": 0.30553460121154785, "learning_rate": 0.0004972075833429516, "loss": 1.9735, "step": 799 }, { "epoch": 0.078125, "grad_norm": 0.28286099433898926, "learning_rate": 0.0004971964034020637, "loss": 1.9704, "step": 800 }, { "epoch": 0.07822265625, "grad_norm": 0.28064650297164917, "learning_rate": 0.0004971852012657235, "loss": 1.9452, "step": 801 }, { "epoch": 0.0783203125, "grad_norm": 0.2370154857635498, "learning_rate": 0.0004971739769350502, "loss": 2.0068, "step": 802 }, { "epoch": 0.07841796875, "grad_norm": 0.26157453656196594, "learning_rate": 0.0004971627304111648, "loss": 1.9209, "step": 803 }, { "epoch": 0.078515625, "grad_norm": 0.2244575023651123, "learning_rate": 0.0004971514616951908, "loss": 1.9952, "step": 804 }, { "epoch": 0.07861328125, "grad_norm": 0.241064190864563, "learning_rate": 0.0004971401707882538, "loss": 1.9616, "step": 805 }, { "epoch": 0.0787109375, "grad_norm": 0.23037207126617432, "learning_rate": 0.0004971288576914819, "loss": 1.9792, "step": 806 }, { "epoch": 0.07880859375, "grad_norm": 0.2162441909313202, "learning_rate": 0.0004971175224060049, "loss": 1.9556, "step": 807 }, { "epoch": 0.07890625, "grad_norm": 0.2614160180091858, "learning_rate": 0.0004971061649329551, "loss": 1.9247, "step": 808 }, { "epoch": 0.07900390625, "grad_norm": 0.22268271446228027, "learning_rate": 0.0004970947852734672, "loss": 1.9596, "step": 809 }, { "epoch": 0.0791015625, "grad_norm": 0.26265689730644226, "learning_rate": 0.0004970833834286779, "loss": 1.9722, "step": 810 }, { "epoch": 0.07919921875, "grad_norm": 0.29413190484046936, "learning_rate": 0.000497071959399726, "loss": 1.9808, "step": 811 }, { "epoch": 0.079296875, "grad_norm": 0.3389902114868164, "learning_rate": 0.0004970605131877527, "loss": 1.9986, "step": 812 }, { "epoch": 0.07939453125, "grad_norm": 0.3456060290336609, "learning_rate": 0.0004970490447939016, "loss": 1.9556, "step": 813 }, { "epoch": 0.0794921875, "grad_norm": 0.3039393126964569, "learning_rate": 0.0004970375542193179, "loss": 1.9359, "step": 814 }, { "epoch": 0.07958984375, "grad_norm": 0.46128201484680176, "learning_rate": 0.0004970260414651497, "loss": 1.935, "step": 815 }, { "epoch": 0.0796875, "grad_norm": 0.5321182608604431, "learning_rate": 0.000497014506532547, "loss": 1.9767, "step": 816 }, { "epoch": 0.07978515625, "grad_norm": 0.39562714099884033, "learning_rate": 0.0004970029494226618, "loss": 1.9803, "step": 817 }, { "epoch": 0.0798828125, "grad_norm": 0.2717466950416565, "learning_rate": 0.0004969913701366488, "loss": 1.9675, "step": 818 }, { "epoch": 0.07998046875, "grad_norm": 0.32809481024742126, "learning_rate": 0.0004969797686756646, "loss": 1.9869, "step": 819 }, { "epoch": 0.080078125, "grad_norm": 0.3525324761867523, "learning_rate": 0.0004969681450408681, "loss": 1.9858, "step": 820 }, { "epoch": 0.08017578125, "grad_norm": 0.30448198318481445, "learning_rate": 0.0004969564992334202, "loss": 2.0067, "step": 821 }, { "epoch": 0.0802734375, "grad_norm": 0.27088144421577454, "learning_rate": 0.0004969448312544845, "loss": 2.0068, "step": 822 }, { "epoch": 0.08037109375, "grad_norm": 0.26881691813468933, "learning_rate": 0.0004969331411052265, "loss": 1.9418, "step": 823 }, { "epoch": 0.08046875, "grad_norm": 0.31323063373565674, "learning_rate": 0.0004969214287868138, "loss": 1.9662, "step": 824 }, { "epoch": 0.08056640625, "grad_norm": 0.25669077038764954, "learning_rate": 0.0004969096943004163, "loss": 1.9733, "step": 825 }, { "epoch": 0.0806640625, "grad_norm": 0.27999046444892883, "learning_rate": 0.0004968979376472065, "loss": 1.9611, "step": 826 }, { "epoch": 0.08076171875, "grad_norm": 0.35361799597740173, "learning_rate": 0.0004968861588283584, "loss": 1.9035, "step": 827 }, { "epoch": 0.080859375, "grad_norm": 0.290341317653656, "learning_rate": 0.0004968743578450489, "loss": 2.0007, "step": 828 }, { "epoch": 0.08095703125, "grad_norm": 0.29945844411849976, "learning_rate": 0.0004968625346984565, "loss": 1.9724, "step": 829 }, { "epoch": 0.0810546875, "grad_norm": 0.2630455493927002, "learning_rate": 0.0004968506893897626, "loss": 1.9765, "step": 830 }, { "epoch": 0.08115234375, "grad_norm": 0.2183736264705658, "learning_rate": 0.0004968388219201501, "loss": 1.9414, "step": 831 }, { "epoch": 0.08125, "grad_norm": 0.24703919887542725, "learning_rate": 0.0004968269322908046, "loss": 1.9725, "step": 832 }, { "epoch": 0.08134765625, "grad_norm": 0.22808632254600525, "learning_rate": 0.0004968150205029139, "loss": 1.9121, "step": 833 }, { "epoch": 0.0814453125, "grad_norm": 0.18958304822444916, "learning_rate": 0.0004968030865576677, "loss": 1.971, "step": 834 }, { "epoch": 0.08154296875, "grad_norm": 0.22013933956623077, "learning_rate": 0.0004967911304562581, "loss": 1.9703, "step": 835 }, { "epoch": 0.081640625, "grad_norm": 0.21636781096458435, "learning_rate": 0.0004967791521998795, "loss": 1.992, "step": 836 }, { "epoch": 0.08173828125, "grad_norm": 0.20808343589305878, "learning_rate": 0.0004967671517897284, "loss": 1.9881, "step": 837 }, { "epoch": 0.0818359375, "grad_norm": 0.23455387353897095, "learning_rate": 0.0004967551292270034, "loss": 1.9635, "step": 838 }, { "epoch": 0.08193359375, "grad_norm": 0.2605065703392029, "learning_rate": 0.0004967430845129057, "loss": 1.8877, "step": 839 }, { "epoch": 0.08203125, "grad_norm": 0.30832260847091675, "learning_rate": 0.0004967310176486382, "loss": 1.9557, "step": 840 }, { "epoch": 0.08212890625, "grad_norm": 0.32088643312454224, "learning_rate": 0.0004967189286354066, "loss": 1.9352, "step": 841 }, { "epoch": 0.0822265625, "grad_norm": 0.3073631823062897, "learning_rate": 0.0004967068174744181, "loss": 1.9487, "step": 842 }, { "epoch": 0.08232421875, "grad_norm": 0.23467223346233368, "learning_rate": 0.0004966946841668828, "loss": 1.9514, "step": 843 }, { "epoch": 0.082421875, "grad_norm": 0.2447347342967987, "learning_rate": 0.0004966825287140125, "loss": 1.9435, "step": 844 }, { "epoch": 0.08251953125, "grad_norm": 0.2715449631214142, "learning_rate": 0.0004966703511170215, "loss": 1.9591, "step": 845 }, { "epoch": 0.0826171875, "grad_norm": 0.28944098949432373, "learning_rate": 0.0004966581513771263, "loss": 1.952, "step": 846 }, { "epoch": 0.08271484375, "grad_norm": 0.25582170486450195, "learning_rate": 0.0004966459294955455, "loss": 1.9486, "step": 847 }, { "epoch": 0.0828125, "grad_norm": 0.305513858795166, "learning_rate": 0.0004966336854734999, "loss": 1.9416, "step": 848 }, { "epoch": 0.08291015625, "grad_norm": 0.28799137473106384, "learning_rate": 0.0004966214193122127, "loss": 1.9417, "step": 849 }, { "epoch": 0.0830078125, "grad_norm": 0.22209681570529938, "learning_rate": 0.000496609131012909, "loss": 1.9641, "step": 850 }, { "epoch": 0.08310546875, "grad_norm": 0.35411134362220764, "learning_rate": 0.0004965968205768165, "loss": 1.9497, "step": 851 }, { "epoch": 0.083203125, "grad_norm": 0.2821413576602936, "learning_rate": 0.0004965844880051647, "loss": 1.9398, "step": 852 }, { "epoch": 0.08330078125, "grad_norm": 0.31592175364494324, "learning_rate": 0.0004965721332991856, "loss": 1.9407, "step": 853 }, { "epoch": 0.0833984375, "grad_norm": 0.3579123318195343, "learning_rate": 0.0004965597564601134, "loss": 1.9767, "step": 854 }, { "epoch": 0.08349609375, "grad_norm": 0.3371833860874176, "learning_rate": 0.0004965473574891845, "loss": 2.0025, "step": 855 }, { "epoch": 0.08359375, "grad_norm": 0.39282771944999695, "learning_rate": 0.0004965349363876371, "loss": 1.9303, "step": 856 }, { "epoch": 0.08369140625, "grad_norm": 0.34869179129600525, "learning_rate": 0.0004965224931567122, "loss": 1.9613, "step": 857 }, { "epoch": 0.0837890625, "grad_norm": 0.2918208837509155, "learning_rate": 0.0004965100277976528, "loss": 1.9242, "step": 858 }, { "epoch": 0.08388671875, "grad_norm": 0.25440847873687744, "learning_rate": 0.0004964975403117041, "loss": 1.9223, "step": 859 }, { "epoch": 0.083984375, "grad_norm": 0.27069318294525146, "learning_rate": 0.0004964850307001132, "loss": 1.9711, "step": 860 }, { "epoch": 0.08408203125, "grad_norm": 0.27405187487602234, "learning_rate": 0.0004964724989641301, "loss": 1.9646, "step": 861 }, { "epoch": 0.0841796875, "grad_norm": 0.3116970956325531, "learning_rate": 0.0004964599451050064, "loss": 1.9231, "step": 862 }, { "epoch": 0.08427734375, "grad_norm": 0.3815697729587555, "learning_rate": 0.000496447369123996, "loss": 1.9773, "step": 863 }, { "epoch": 0.084375, "grad_norm": 0.32103797793388367, "learning_rate": 0.0004964347710223555, "loss": 1.936, "step": 864 }, { "epoch": 0.08447265625, "grad_norm": 0.42826011776924133, "learning_rate": 0.0004964221508013429, "loss": 1.9315, "step": 865 }, { "epoch": 0.0845703125, "grad_norm": 0.3091580271720886, "learning_rate": 0.0004964095084622192, "loss": 1.9362, "step": 866 }, { "epoch": 0.08466796875, "grad_norm": 0.2816343605518341, "learning_rate": 0.0004963968440062471, "loss": 2.0076, "step": 867 }, { "epoch": 0.084765625, "grad_norm": 0.35303476452827454, "learning_rate": 0.0004963841574346917, "loss": 1.9399, "step": 868 }, { "epoch": 0.08486328125, "grad_norm": 0.27021846175193787, "learning_rate": 0.0004963714487488202, "loss": 1.9556, "step": 869 }, { "epoch": 0.0849609375, "grad_norm": 0.28174805641174316, "learning_rate": 0.0004963587179499021, "loss": 1.9749, "step": 870 }, { "epoch": 0.08505859375, "grad_norm": 0.2774275243282318, "learning_rate": 0.0004963459650392093, "loss": 1.9457, "step": 871 }, { "epoch": 0.08515625, "grad_norm": 0.2178190052509308, "learning_rate": 0.0004963331900180154, "loss": 2.0192, "step": 872 }, { "epoch": 0.08525390625, "grad_norm": 0.2983119487762451, "learning_rate": 0.0004963203928875967, "loss": 1.9659, "step": 873 }, { "epoch": 0.0853515625, "grad_norm": 0.2878938913345337, "learning_rate": 0.0004963075736492315, "loss": 1.9581, "step": 874 }, { "epoch": 0.08544921875, "grad_norm": 0.19974899291992188, "learning_rate": 0.0004962947323042002, "loss": 1.9898, "step": 875 }, { "epoch": 0.085546875, "grad_norm": 0.31860581040382385, "learning_rate": 0.0004962818688537857, "loss": 1.9737, "step": 876 }, { "epoch": 0.08564453125, "grad_norm": 0.2651256322860718, "learning_rate": 0.0004962689832992729, "loss": 1.9977, "step": 877 }, { "epoch": 0.0857421875, "grad_norm": 0.286821573972702, "learning_rate": 0.0004962560756419489, "loss": 1.9317, "step": 878 }, { "epoch": 0.08583984375, "grad_norm": 0.3837340772151947, "learning_rate": 0.0004962431458831032, "loss": 1.9419, "step": 879 }, { "epoch": 0.0859375, "grad_norm": 0.32222527265548706, "learning_rate": 0.0004962301940240272, "loss": 1.9603, "step": 880 }, { "epoch": 0.08603515625, "grad_norm": 0.33181512355804443, "learning_rate": 0.0004962172200660146, "loss": 1.9646, "step": 881 }, { "epoch": 0.0861328125, "grad_norm": 0.2813611626625061, "learning_rate": 0.0004962042240103617, "loss": 1.9706, "step": 882 }, { "epoch": 0.08623046875, "grad_norm": 0.2933270335197449, "learning_rate": 0.0004961912058583664, "loss": 1.9529, "step": 883 }, { "epoch": 0.086328125, "grad_norm": 0.31641921401023865, "learning_rate": 0.0004961781656113292, "loss": 1.9437, "step": 884 }, { "epoch": 0.08642578125, "grad_norm": 0.39271247386932373, "learning_rate": 0.0004961651032705528, "loss": 1.957, "step": 885 }, { "epoch": 0.0865234375, "grad_norm": 0.3720405697822571, "learning_rate": 0.000496152018837342, "loss": 1.9393, "step": 886 }, { "epoch": 0.08662109375, "grad_norm": 0.27104172110557556, "learning_rate": 0.0004961389123130037, "loss": 1.9547, "step": 887 }, { "epoch": 0.08671875, "grad_norm": 0.24182896316051483, "learning_rate": 0.0004961257836988472, "loss": 1.9456, "step": 888 }, { "epoch": 0.08681640625, "grad_norm": 0.22501450777053833, "learning_rate": 0.0004961126329961838, "loss": 1.9336, "step": 889 }, { "epoch": 0.0869140625, "grad_norm": 0.2969515323638916, "learning_rate": 0.0004960994602063273, "loss": 1.9415, "step": 890 }, { "epoch": 0.08701171875, "grad_norm": 0.3117569386959076, "learning_rate": 0.0004960862653305934, "loss": 1.8993, "step": 891 }, { "epoch": 0.087109375, "grad_norm": 0.32299551367759705, "learning_rate": 0.0004960730483703004, "loss": 1.9879, "step": 892 }, { "epoch": 0.08720703125, "grad_norm": 0.335592657327652, "learning_rate": 0.0004960598093267682, "loss": 1.9665, "step": 893 }, { "epoch": 0.0873046875, "grad_norm": 0.21160855889320374, "learning_rate": 0.0004960465482013197, "loss": 1.8927, "step": 894 }, { "epoch": 0.08740234375, "grad_norm": 0.30621418356895447, "learning_rate": 0.0004960332649952792, "loss": 1.9828, "step": 895 }, { "epoch": 0.0875, "grad_norm": 0.3541114330291748, "learning_rate": 0.0004960199597099737, "loss": 1.9478, "step": 896 }, { "epoch": 0.08759765625, "grad_norm": 0.31596067547798157, "learning_rate": 0.0004960066323467323, "loss": 1.9355, "step": 897 }, { "epoch": 0.0876953125, "grad_norm": 0.3170725107192993, "learning_rate": 0.0004959932829068863, "loss": 1.9501, "step": 898 }, { "epoch": 0.08779296875, "grad_norm": 0.26953962445259094, "learning_rate": 0.0004959799113917691, "loss": 1.9488, "step": 899 }, { "epoch": 0.087890625, "grad_norm": 0.23792020976543427, "learning_rate": 0.0004959665178027165, "loss": 1.941, "step": 900 }, { "epoch": 0.08798828125, "grad_norm": 0.2375611513853073, "learning_rate": 0.0004959531021410664, "loss": 1.9133, "step": 901 }, { "epoch": 0.0880859375, "grad_norm": 0.20919479429721832, "learning_rate": 0.0004959396644081588, "loss": 1.9624, "step": 902 }, { "epoch": 0.08818359375, "grad_norm": 0.22488316893577576, "learning_rate": 0.0004959262046053361, "loss": 1.9408, "step": 903 }, { "epoch": 0.08828125, "grad_norm": 0.2249651849269867, "learning_rate": 0.0004959127227339427, "loss": 2.0022, "step": 904 }, { "epoch": 0.08837890625, "grad_norm": 0.19234181940555573, "learning_rate": 0.0004958992187953256, "loss": 1.9221, "step": 905 }, { "epoch": 0.0884765625, "grad_norm": 0.21266022324562073, "learning_rate": 0.0004958856927908334, "loss": 1.9796, "step": 906 }, { "epoch": 0.08857421875, "grad_norm": 0.22636987268924713, "learning_rate": 0.0004958721447218174, "loss": 1.9799, "step": 907 }, { "epoch": 0.088671875, "grad_norm": 0.21743637323379517, "learning_rate": 0.0004958585745896308, "loss": 1.9156, "step": 908 }, { "epoch": 0.08876953125, "grad_norm": 0.316840797662735, "learning_rate": 0.0004958449823956294, "loss": 1.9856, "step": 909 }, { "epoch": 0.0888671875, "grad_norm": 0.2583603262901306, "learning_rate": 0.0004958313681411706, "loss": 1.9601, "step": 910 }, { "epoch": 0.08896484375, "grad_norm": 0.2872125208377838, "learning_rate": 0.0004958177318276146, "loss": 1.9526, "step": 911 }, { "epoch": 0.0890625, "grad_norm": 0.37896475195884705, "learning_rate": 0.0004958040734563235, "loss": 1.9465, "step": 912 }, { "epoch": 0.08916015625, "grad_norm": 0.3056401014328003, "learning_rate": 0.0004957903930286616, "loss": 1.9344, "step": 913 }, { "epoch": 0.0892578125, "grad_norm": 0.35092082619667053, "learning_rate": 0.0004957766905459953, "loss": 1.9294, "step": 914 }, { "epoch": 0.08935546875, "grad_norm": 0.2681847810745239, "learning_rate": 0.0004957629660096937, "loss": 1.9436, "step": 915 }, { "epoch": 0.089453125, "grad_norm": 0.25762611627578735, "learning_rate": 0.0004957492194211275, "loss": 1.931, "step": 916 }, { "epoch": 0.08955078125, "grad_norm": 0.24634388089179993, "learning_rate": 0.00049573545078167, "loss": 1.9373, "step": 917 }, { "epoch": 0.0896484375, "grad_norm": 0.27374333143234253, "learning_rate": 0.0004957216600926965, "loss": 2.046, "step": 918 }, { "epoch": 0.08974609375, "grad_norm": 0.338375061750412, "learning_rate": 0.0004957078473555845, "loss": 2.0259, "step": 919 }, { "epoch": 0.08984375, "grad_norm": 0.27678367495536804, "learning_rate": 0.0004956940125717139, "loss": 1.9947, "step": 920 }, { "epoch": 0.08994140625, "grad_norm": 0.266079306602478, "learning_rate": 0.0004956801557424666, "loss": 2.0097, "step": 921 }, { "epoch": 0.0900390625, "grad_norm": 0.28650638461112976, "learning_rate": 0.0004956662768692268, "loss": 1.9893, "step": 922 }, { "epoch": 0.09013671875, "grad_norm": 0.2717742323875427, "learning_rate": 0.0004956523759533809, "loss": 1.976, "step": 923 }, { "epoch": 0.090234375, "grad_norm": 0.3267741799354553, "learning_rate": 0.0004956384529963174, "loss": 1.9803, "step": 924 }, { "epoch": 0.09033203125, "grad_norm": 0.32973724603652954, "learning_rate": 0.0004956245079994272, "loss": 1.942, "step": 925 }, { "epoch": 0.0904296875, "grad_norm": 0.3257177770137787, "learning_rate": 0.0004956105409641031, "loss": 1.9584, "step": 926 }, { "epoch": 0.09052734375, "grad_norm": 0.31881803274154663, "learning_rate": 0.0004955965518917405, "loss": 1.9386, "step": 927 }, { "epoch": 0.090625, "grad_norm": 0.2814694046974182, "learning_rate": 0.0004955825407837365, "loss": 1.9114, "step": 928 }, { "epoch": 0.09072265625, "grad_norm": 0.24232172966003418, "learning_rate": 0.000495568507641491, "loss": 1.954, "step": 929 }, { "epoch": 0.0908203125, "grad_norm": 0.2062203288078308, "learning_rate": 0.0004955544524664056, "loss": 1.9682, "step": 930 }, { "epoch": 0.09091796875, "grad_norm": 0.2777712047100067, "learning_rate": 0.0004955403752598844, "loss": 1.983, "step": 931 }, { "epoch": 0.091015625, "grad_norm": 0.31253954768180847, "learning_rate": 0.0004955262760233335, "loss": 1.9211, "step": 932 }, { "epoch": 0.09111328125, "grad_norm": 0.292059987783432, "learning_rate": 0.0004955121547581613, "loss": 1.9745, "step": 933 }, { "epoch": 0.0912109375, "grad_norm": 0.309123694896698, "learning_rate": 0.0004954980114657783, "loss": 1.966, "step": 934 }, { "epoch": 0.09130859375, "grad_norm": 0.2765352129936218, "learning_rate": 0.0004954838461475975, "loss": 1.9535, "step": 935 }, { "epoch": 0.09140625, "grad_norm": 0.2576250731945038, "learning_rate": 0.0004954696588050338, "loss": 1.9002, "step": 936 }, { "epoch": 0.09150390625, "grad_norm": 0.3167097866535187, "learning_rate": 0.0004954554494395044, "loss": 1.9249, "step": 937 }, { "epoch": 0.0916015625, "grad_norm": 0.32505497336387634, "learning_rate": 0.0004954412180524285, "loss": 1.9669, "step": 938 }, { "epoch": 0.09169921875, "grad_norm": 0.3114219307899475, "learning_rate": 0.000495426964645228, "loss": 1.9276, "step": 939 }, { "epoch": 0.091796875, "grad_norm": 0.2786170542240143, "learning_rate": 0.0004954126892193264, "loss": 1.965, "step": 940 }, { "epoch": 0.09189453125, "grad_norm": 0.2753596901893616, "learning_rate": 0.00049539839177615, "loss": 1.9925, "step": 941 }, { "epoch": 0.0919921875, "grad_norm": 0.29401320219039917, "learning_rate": 0.0004953840723171268, "loss": 1.9772, "step": 942 }, { "epoch": 0.09208984375, "grad_norm": 0.3157658278942108, "learning_rate": 0.0004953697308436871, "loss": 2.0019, "step": 943 }, { "epoch": 0.0921875, "grad_norm": 0.29067566990852356, "learning_rate": 0.0004953553673572636, "loss": 1.9753, "step": 944 }, { "epoch": 0.09228515625, "grad_norm": 0.28693830966949463, "learning_rate": 0.0004953409818592911, "loss": 1.9436, "step": 945 }, { "epoch": 0.0923828125, "grad_norm": 0.2536534070968628, "learning_rate": 0.0004953265743512065, "loss": 1.9801, "step": 946 }, { "epoch": 0.09248046875, "grad_norm": 0.3041675090789795, "learning_rate": 0.000495312144834449, "loss": 1.9378, "step": 947 }, { "epoch": 0.092578125, "grad_norm": 0.3237321376800537, "learning_rate": 0.0004952976933104602, "loss": 1.9781, "step": 948 }, { "epoch": 0.09267578125, "grad_norm": 0.25767290592193604, "learning_rate": 0.0004952832197806835, "loss": 1.9031, "step": 949 }, { "epoch": 0.0927734375, "grad_norm": 0.25587791204452515, "learning_rate": 0.0004952687242465645, "loss": 1.9405, "step": 950 }, { "epoch": 0.09287109375, "grad_norm": 0.2901047468185425, "learning_rate": 0.0004952542067095515, "loss": 1.9567, "step": 951 }, { "epoch": 0.09296875, "grad_norm": 0.31517183780670166, "learning_rate": 0.0004952396671710945, "loss": 1.9464, "step": 952 }, { "epoch": 0.09306640625, "grad_norm": 0.30896711349487305, "learning_rate": 0.0004952251056326458, "loss": 1.9096, "step": 953 }, { "epoch": 0.0931640625, "grad_norm": 0.26281988620758057, "learning_rate": 0.0004952105220956602, "loss": 1.9616, "step": 954 }, { "epoch": 0.09326171875, "grad_norm": 0.23971204459667206, "learning_rate": 0.0004951959165615942, "loss": 1.9928, "step": 955 }, { "epoch": 0.093359375, "grad_norm": 0.22277623414993286, "learning_rate": 0.000495181289031907, "loss": 1.9652, "step": 956 }, { "epoch": 0.09345703125, "grad_norm": 0.26211845874786377, "learning_rate": 0.0004951666395080596, "loss": 1.9617, "step": 957 }, { "epoch": 0.0935546875, "grad_norm": 0.24874535202980042, "learning_rate": 0.0004951519679915155, "loss": 1.9509, "step": 958 }, { "epoch": 0.09365234375, "grad_norm": 0.21750757098197937, "learning_rate": 0.0004951372744837401, "loss": 1.966, "step": 959 }, { "epoch": 0.09375, "grad_norm": 0.2798830270767212, "learning_rate": 0.0004951225589862012, "loss": 1.9574, "step": 960 }, { "epoch": 0.09384765625, "grad_norm": 0.33166754245758057, "learning_rate": 0.0004951078215003688, "loss": 1.9672, "step": 961 }, { "epoch": 0.0939453125, "grad_norm": 0.3122573792934418, "learning_rate": 0.0004950930620277149, "loss": 1.9205, "step": 962 }, { "epoch": 0.09404296875, "grad_norm": 0.233395054936409, "learning_rate": 0.0004950782805697141, "loss": 1.9389, "step": 963 }, { "epoch": 0.094140625, "grad_norm": 0.2523380219936371, "learning_rate": 0.0004950634771278427, "loss": 1.9643, "step": 964 }, { "epoch": 0.09423828125, "grad_norm": 0.2774633765220642, "learning_rate": 0.0004950486517035796, "loss": 1.9467, "step": 965 }, { "epoch": 0.0943359375, "grad_norm": 0.2782750129699707, "learning_rate": 0.0004950338042984056, "loss": 1.9296, "step": 966 }, { "epoch": 0.09443359375, "grad_norm": 0.3408234417438507, "learning_rate": 0.0004950189349138039, "loss": 1.9705, "step": 967 }, { "epoch": 0.09453125, "grad_norm": 0.27266553044319153, "learning_rate": 0.0004950040435512598, "loss": 1.9292, "step": 968 }, { "epoch": 0.09462890625, "grad_norm": 0.2276991307735443, "learning_rate": 0.0004949891302122607, "loss": 1.9793, "step": 969 }, { "epoch": 0.0947265625, "grad_norm": 0.23914872109889984, "learning_rate": 0.0004949741948982965, "loss": 1.9632, "step": 970 }, { "epoch": 0.09482421875, "grad_norm": 0.24836085736751556, "learning_rate": 0.000494959237610859, "loss": 1.9304, "step": 971 }, { "epoch": 0.094921875, "grad_norm": 0.1932704746723175, "learning_rate": 0.0004949442583514425, "loss": 1.9483, "step": 972 }, { "epoch": 0.09501953125, "grad_norm": 0.27378395199775696, "learning_rate": 0.0004949292571215429, "loss": 1.9687, "step": 973 }, { "epoch": 0.0951171875, "grad_norm": 0.3175489008426666, "learning_rate": 0.0004949142339226591, "loss": 2.0036, "step": 974 }, { "epoch": 0.09521484375, "grad_norm": 0.34563618898391724, "learning_rate": 0.0004948991887562914, "loss": 1.9429, "step": 975 }, { "epoch": 0.0953125, "grad_norm": 0.2931719720363617, "learning_rate": 0.000494884121623943, "loss": 1.9424, "step": 976 }, { "epoch": 0.09541015625, "grad_norm": 0.29774370789527893, "learning_rate": 0.0004948690325271188, "loss": 1.9297, "step": 977 }, { "epoch": 0.0955078125, "grad_norm": 0.31513288617134094, "learning_rate": 0.0004948539214673261, "loss": 1.9917, "step": 978 }, { "epoch": 0.09560546875, "grad_norm": 0.2150924950838089, "learning_rate": 0.0004948387884460744, "loss": 1.9529, "step": 979 }, { "epoch": 0.095703125, "grad_norm": 0.27547362446784973, "learning_rate": 0.0004948236334648753, "loss": 1.9885, "step": 980 }, { "epoch": 0.09580078125, "grad_norm": 0.3271898925304413, "learning_rate": 0.0004948084565252427, "loss": 1.9232, "step": 981 }, { "epoch": 0.0958984375, "grad_norm": 0.2923772931098938, "learning_rate": 0.0004947932576286928, "loss": 1.9429, "step": 982 }, { "epoch": 0.09599609375, "grad_norm": 0.2936458885669708, "learning_rate": 0.0004947780367767434, "loss": 1.9773, "step": 983 }, { "epoch": 0.09609375, "grad_norm": 0.29706791043281555, "learning_rate": 0.0004947627939709153, "loss": 1.9284, "step": 984 }, { "epoch": 0.09619140625, "grad_norm": 0.22721995413303375, "learning_rate": 0.000494747529212731, "loss": 1.9281, "step": 985 }, { "epoch": 0.0962890625, "grad_norm": 0.27041006088256836, "learning_rate": 0.0004947322425037152, "loss": 1.9735, "step": 986 }, { "epoch": 0.09638671875, "grad_norm": 0.26037418842315674, "learning_rate": 0.0004947169338453952, "loss": 1.9426, "step": 987 }, { "epoch": 0.096484375, "grad_norm": 0.23383264243602753, "learning_rate": 0.0004947016032392999, "loss": 1.923, "step": 988 }, { "epoch": 0.09658203125, "grad_norm": 0.29238423705101013, "learning_rate": 0.0004946862506869609, "loss": 1.9803, "step": 989 }, { "epoch": 0.0966796875, "grad_norm": 0.223761647939682, "learning_rate": 0.0004946708761899116, "loss": 1.9377, "step": 990 }, { "epoch": 0.09677734375, "grad_norm": 0.2669995427131653, "learning_rate": 0.0004946554797496881, "loss": 1.9694, "step": 991 }, { "epoch": 0.096875, "grad_norm": 0.2847992181777954, "learning_rate": 0.0004946400613678279, "loss": 1.9501, "step": 992 }, { "epoch": 0.09697265625, "grad_norm": 0.25099360942840576, "learning_rate": 0.0004946246210458715, "loss": 1.9707, "step": 993 }, { "epoch": 0.0970703125, "grad_norm": 0.23306165635585785, "learning_rate": 0.0004946091587853611, "loss": 1.9174, "step": 994 }, { "epoch": 0.09716796875, "grad_norm": 0.2907635569572449, "learning_rate": 0.0004945936745878414, "loss": 1.9521, "step": 995 }, { "epoch": 0.097265625, "grad_norm": 0.3234463334083557, "learning_rate": 0.0004945781684548591, "loss": 1.9667, "step": 996 }, { "epoch": 0.09736328125, "grad_norm": 0.36099013686180115, "learning_rate": 0.0004945626403879629, "loss": 1.9257, "step": 997 }, { "epoch": 0.0974609375, "grad_norm": 0.3735397160053253, "learning_rate": 0.0004945470903887041, "loss": 1.9512, "step": 998 }, { "epoch": 0.09755859375, "grad_norm": 0.3653644025325775, "learning_rate": 0.0004945315184586359, "loss": 1.9664, "step": 999 }, { "epoch": 0.09765625, "grad_norm": 0.2931279242038727, "learning_rate": 0.000494515924599314, "loss": 1.9209, "step": 1000 }, { "epoch": 0.09775390625, "grad_norm": 0.31677475571632385, "learning_rate": 0.000494500308812296, "loss": 1.9989, "step": 1001 }, { "epoch": 0.0978515625, "grad_norm": 0.2640211582183838, "learning_rate": 0.0004944846710991416, "loss": 1.8792, "step": 1002 }, { "epoch": 0.09794921875, "grad_norm": 0.3207566738128662, "learning_rate": 0.0004944690114614132, "loss": 1.955, "step": 1003 }, { "epoch": 0.098046875, "grad_norm": 0.29306915402412415, "learning_rate": 0.0004944533299006749, "loss": 1.9264, "step": 1004 }, { "epoch": 0.09814453125, "grad_norm": 0.2783614993095398, "learning_rate": 0.000494437626418493, "loss": 1.9341, "step": 1005 }, { "epoch": 0.0982421875, "grad_norm": 0.346975713968277, "learning_rate": 0.0004944219010164361, "loss": 1.9599, "step": 1006 }, { "epoch": 0.09833984375, "grad_norm": 0.2799683213233948, "learning_rate": 0.0004944061536960755, "loss": 1.9201, "step": 1007 }, { "epoch": 0.0984375, "grad_norm": 0.27122223377227783, "learning_rate": 0.0004943903844589837, "loss": 1.9685, "step": 1008 }, { "epoch": 0.09853515625, "grad_norm": 0.34742453694343567, "learning_rate": 0.0004943745933067362, "loss": 1.9533, "step": 1009 }, { "epoch": 0.0986328125, "grad_norm": 0.2686299681663513, "learning_rate": 0.0004943587802409103, "loss": 1.9149, "step": 1010 }, { "epoch": 0.09873046875, "grad_norm": 0.3223266303539276, "learning_rate": 0.0004943429452630856, "loss": 1.8887, "step": 1011 }, { "epoch": 0.098828125, "grad_norm": 0.2767395079135895, "learning_rate": 0.0004943270883748439, "loss": 1.9756, "step": 1012 }, { "epoch": 0.09892578125, "grad_norm": 0.2761443257331848, "learning_rate": 0.0004943112095777691, "loss": 1.981, "step": 1013 }, { "epoch": 0.0990234375, "grad_norm": 0.22587864100933075, "learning_rate": 0.0004942953088734474, "loss": 1.939, "step": 1014 }, { "epoch": 0.09912109375, "grad_norm": 0.2596238851547241, "learning_rate": 0.0004942793862634671, "loss": 1.9546, "step": 1015 }, { "epoch": 0.09921875, "grad_norm": 0.3035464584827423, "learning_rate": 0.0004942634417494188, "loss": 2.0001, "step": 1016 }, { "epoch": 0.09931640625, "grad_norm": 0.32868078351020813, "learning_rate": 0.0004942474753328952, "loss": 1.9248, "step": 1017 }, { "epoch": 0.0994140625, "grad_norm": 0.34833693504333496, "learning_rate": 0.000494231487015491, "loss": 1.9148, "step": 1018 }, { "epoch": 0.09951171875, "grad_norm": 0.2557753920555115, "learning_rate": 0.0004942154767988037, "loss": 1.9585, "step": 1019 }, { "epoch": 0.099609375, "grad_norm": 0.25750526785850525, "learning_rate": 0.0004941994446844321, "loss": 2.0257, "step": 1020 }, { "epoch": 0.09970703125, "grad_norm": 0.22088824212551117, "learning_rate": 0.000494183390673978, "loss": 1.9373, "step": 1021 }, { "epoch": 0.0998046875, "grad_norm": 0.23292425274848938, "learning_rate": 0.000494167314769045, "loss": 1.9462, "step": 1022 }, { "epoch": 0.09990234375, "grad_norm": 0.24202801287174225, "learning_rate": 0.0004941512169712389, "loss": 2.0237, "step": 1023 }, { "epoch": 0.1, "grad_norm": 0.28507405519485474, "learning_rate": 0.0004941350972821678, "loss": 1.9373, "step": 1024 }, { "epoch": 0.10009765625, "grad_norm": 0.22370004653930664, "learning_rate": 0.0004941189557034417, "loss": 1.9332, "step": 1025 }, { "epoch": 0.1001953125, "grad_norm": 0.2707929313182831, "learning_rate": 0.0004941027922366732, "loss": 1.9908, "step": 1026 }, { "epoch": 0.10029296875, "grad_norm": 0.29618075489997864, "learning_rate": 0.0004940866068834769, "loss": 1.9317, "step": 1027 }, { "epoch": 0.100390625, "grad_norm": 0.27780136466026306, "learning_rate": 0.0004940703996454695, "loss": 1.9225, "step": 1028 }, { "epoch": 0.10048828125, "grad_norm": 0.33955708146095276, "learning_rate": 0.0004940541705242698, "loss": 1.9232, "step": 1029 }, { "epoch": 0.1005859375, "grad_norm": 0.3132137954235077, "learning_rate": 0.0004940379195214992, "loss": 1.976, "step": 1030 }, { "epoch": 0.10068359375, "grad_norm": 0.2742595076560974, "learning_rate": 0.000494021646638781, "loss": 1.9138, "step": 1031 }, { "epoch": 0.10078125, "grad_norm": 0.2921305298805237, "learning_rate": 0.0004940053518777406, "loss": 1.9178, "step": 1032 }, { "epoch": 0.10087890625, "grad_norm": 0.2586468458175659, "learning_rate": 0.0004939890352400058, "loss": 1.9385, "step": 1033 }, { "epoch": 0.1009765625, "grad_norm": 0.30906325578689575, "learning_rate": 0.0004939726967272064, "loss": 1.9368, "step": 1034 }, { "epoch": 0.10107421875, "grad_norm": 0.3483060300350189, "learning_rate": 0.0004939563363409745, "loss": 1.9922, "step": 1035 }, { "epoch": 0.101171875, "grad_norm": 0.3559792637825012, "learning_rate": 0.0004939399540829445, "loss": 1.9564, "step": 1036 }, { "epoch": 0.10126953125, "grad_norm": 0.33506283164024353, "learning_rate": 0.0004939235499547526, "loss": 1.9784, "step": 1037 }, { "epoch": 0.1013671875, "grad_norm": 0.23956939578056335, "learning_rate": 0.0004939071239580377, "loss": 1.9755, "step": 1038 }, { "epoch": 0.10146484375, "grad_norm": 0.33348920941352844, "learning_rate": 0.0004938906760944403, "loss": 1.9128, "step": 1039 }, { "epoch": 0.1015625, "grad_norm": 0.2843862771987915, "learning_rate": 0.0004938742063656036, "loss": 1.9297, "step": 1040 }, { "epoch": 0.10166015625, "grad_norm": 0.2671546936035156, "learning_rate": 0.0004938577147731727, "loss": 1.9391, "step": 1041 }, { "epoch": 0.1017578125, "grad_norm": 0.2899993360042572, "learning_rate": 0.000493841201318795, "loss": 1.9198, "step": 1042 }, { "epoch": 0.10185546875, "grad_norm": 0.20722518861293793, "learning_rate": 0.0004938246660041201, "loss": 1.9295, "step": 1043 }, { "epoch": 0.101953125, "grad_norm": 0.22667360305786133, "learning_rate": 0.0004938081088307997, "loss": 1.9371, "step": 1044 }, { "epoch": 0.10205078125, "grad_norm": 0.275579035282135, "learning_rate": 0.0004937915298004877, "loss": 1.9383, "step": 1045 }, { "epoch": 0.1021484375, "grad_norm": 0.2900732159614563, "learning_rate": 0.0004937749289148401, "loss": 1.9805, "step": 1046 }, { "epoch": 0.10224609375, "grad_norm": 0.26479572057724, "learning_rate": 0.0004937583061755153, "loss": 1.911, "step": 1047 }, { "epoch": 0.10234375, "grad_norm": 0.2967575490474701, "learning_rate": 0.0004937416615841739, "loss": 1.9654, "step": 1048 }, { "epoch": 0.10244140625, "grad_norm": 0.2840502858161926, "learning_rate": 0.0004937249951424782, "loss": 1.9332, "step": 1049 }, { "epoch": 0.1025390625, "grad_norm": 0.2664666175842285, "learning_rate": 0.0004937083068520931, "loss": 1.973, "step": 1050 }, { "epoch": 0.10263671875, "grad_norm": 0.2535867989063263, "learning_rate": 0.000493691596714686, "loss": 1.9421, "step": 1051 }, { "epoch": 0.102734375, "grad_norm": 0.3005189895629883, "learning_rate": 0.0004936748647319256, "loss": 1.9434, "step": 1052 }, { "epoch": 0.10283203125, "grad_norm": 0.2663765549659729, "learning_rate": 0.0004936581109054837, "loss": 1.9947, "step": 1053 }, { "epoch": 0.1029296875, "grad_norm": 0.29048940539360046, "learning_rate": 0.0004936413352370335, "loss": 1.9737, "step": 1054 }, { "epoch": 0.10302734375, "grad_norm": 0.2594558596611023, "learning_rate": 0.000493624537728251, "loss": 1.9601, "step": 1055 }, { "epoch": 0.103125, "grad_norm": 0.27011555433273315, "learning_rate": 0.000493607718380814, "loss": 1.9459, "step": 1056 }, { "epoch": 0.10322265625, "grad_norm": 0.3043757677078247, "learning_rate": 0.0004935908771964026, "loss": 1.9494, "step": 1057 }, { "epoch": 0.1033203125, "grad_norm": 0.26345208287239075, "learning_rate": 0.0004935740141766991, "loss": 1.896, "step": 1058 }, { "epoch": 0.10341796875, "grad_norm": 0.20400577783584595, "learning_rate": 0.000493557129323388, "loss": 1.9276, "step": 1059 }, { "epoch": 0.103515625, "grad_norm": 0.2836177349090576, "learning_rate": 0.0004935402226381558, "loss": 1.921, "step": 1060 }, { "epoch": 0.10361328125, "grad_norm": 0.2904714047908783, "learning_rate": 0.0004935232941226916, "loss": 1.9631, "step": 1061 }, { "epoch": 0.1037109375, "grad_norm": 0.25980454683303833, "learning_rate": 0.0004935063437786863, "loss": 1.9757, "step": 1062 }, { "epoch": 0.10380859375, "grad_norm": 0.27394455671310425, "learning_rate": 0.0004934893716078329, "loss": 1.9647, "step": 1063 }, { "epoch": 0.10390625, "grad_norm": 0.23419839143753052, "learning_rate": 0.000493472377611827, "loss": 1.962, "step": 1064 }, { "epoch": 0.10400390625, "grad_norm": 0.2323913723230362, "learning_rate": 0.0004934553617923661, "loss": 1.9759, "step": 1065 }, { "epoch": 0.1041015625, "grad_norm": 0.2700064778327942, "learning_rate": 0.0004934383241511498, "loss": 1.9206, "step": 1066 }, { "epoch": 0.10419921875, "grad_norm": 0.3530958294868469, "learning_rate": 0.0004934212646898802, "loss": 1.979, "step": 1067 }, { "epoch": 0.104296875, "grad_norm": 0.30405712127685547, "learning_rate": 0.0004934041834102612, "loss": 1.951, "step": 1068 }, { "epoch": 0.10439453125, "grad_norm": 0.2923583984375, "learning_rate": 0.0004933870803139992, "loss": 2.0158, "step": 1069 }, { "epoch": 0.1044921875, "grad_norm": 0.23782993853092194, "learning_rate": 0.0004933699554028027, "loss": 1.896, "step": 1070 }, { "epoch": 0.10458984375, "grad_norm": 0.3045034110546112, "learning_rate": 0.0004933528086783821, "loss": 1.9555, "step": 1071 }, { "epoch": 0.1046875, "grad_norm": 0.40851131081581116, "learning_rate": 0.0004933356401424505, "loss": 1.9108, "step": 1072 }, { "epoch": 0.10478515625, "grad_norm": 0.39787429571151733, "learning_rate": 0.0004933184497967226, "loss": 1.9472, "step": 1073 }, { "epoch": 0.1048828125, "grad_norm": 0.24665354192256927, "learning_rate": 0.0004933012376429158, "loss": 1.9617, "step": 1074 }, { "epoch": 0.10498046875, "grad_norm": 0.28025394678115845, "learning_rate": 0.0004932840036827492, "loss": 1.898, "step": 1075 }, { "epoch": 0.105078125, "grad_norm": 0.3374732434749603, "learning_rate": 0.0004932667479179445, "loss": 1.9323, "step": 1076 }, { "epoch": 0.10517578125, "grad_norm": 0.2941461503505707, "learning_rate": 0.0004932494703502254, "loss": 1.964, "step": 1077 }, { "epoch": 0.1052734375, "grad_norm": 0.2827446758747101, "learning_rate": 0.0004932321709813177, "loss": 1.9554, "step": 1078 }, { "epoch": 0.10537109375, "grad_norm": 0.32132452726364136, "learning_rate": 0.0004932148498129495, "loss": 1.9333, "step": 1079 }, { "epoch": 0.10546875, "grad_norm": 0.2732277810573578, "learning_rate": 0.000493197506846851, "loss": 1.932, "step": 1080 }, { "epoch": 0.10556640625, "grad_norm": 0.2594139575958252, "learning_rate": 0.0004931801420847545, "loss": 1.9447, "step": 1081 }, { "epoch": 0.1056640625, "grad_norm": 0.24399201571941376, "learning_rate": 0.0004931627555283948, "loss": 1.9163, "step": 1082 }, { "epoch": 0.10576171875, "grad_norm": 0.22696708142757416, "learning_rate": 0.0004931453471795087, "loss": 1.9734, "step": 1083 }, { "epoch": 0.105859375, "grad_norm": 0.25717827677726746, "learning_rate": 0.0004931279170398349, "loss": 1.9767, "step": 1084 }, { "epoch": 0.10595703125, "grad_norm": 0.23899690806865692, "learning_rate": 0.0004931104651111146, "loss": 1.9519, "step": 1085 }, { "epoch": 0.1060546875, "grad_norm": 0.27689501643180847, "learning_rate": 0.0004930929913950911, "loss": 1.9582, "step": 1086 }, { "epoch": 0.10615234375, "grad_norm": 0.23959669470787048, "learning_rate": 0.00049307549589351, "loss": 1.9369, "step": 1087 }, { "epoch": 0.10625, "grad_norm": 0.25673651695251465, "learning_rate": 0.0004930579786081188, "loss": 1.9711, "step": 1088 }, { "epoch": 0.10634765625, "grad_norm": 0.30868396162986755, "learning_rate": 0.0004930404395406674, "loss": 1.9539, "step": 1089 }, { "epoch": 0.1064453125, "grad_norm": 0.2462042272090912, "learning_rate": 0.0004930228786929076, "loss": 1.9061, "step": 1090 }, { "epoch": 0.10654296875, "grad_norm": 0.24275164306163788, "learning_rate": 0.0004930052960665938, "loss": 1.8992, "step": 1091 }, { "epoch": 0.106640625, "grad_norm": 0.27242550253868103, "learning_rate": 0.0004929876916634823, "loss": 1.8997, "step": 1092 }, { "epoch": 0.10673828125, "grad_norm": 0.2655372619628906, "learning_rate": 0.0004929700654853316, "loss": 1.8905, "step": 1093 }, { "epoch": 0.1068359375, "grad_norm": 0.27779421210289, "learning_rate": 0.0004929524175339024, "loss": 1.9485, "step": 1094 }, { "epoch": 0.10693359375, "grad_norm": 0.3308311402797699, "learning_rate": 0.0004929347478109576, "loss": 1.896, "step": 1095 }, { "epoch": 0.10703125, "grad_norm": 0.2988120913505554, "learning_rate": 0.0004929170563182621, "loss": 1.9508, "step": 1096 }, { "epoch": 0.10712890625, "grad_norm": 0.24604608118534088, "learning_rate": 0.0004928993430575834, "loss": 1.9912, "step": 1097 }, { "epoch": 0.1072265625, "grad_norm": 0.31434887647628784, "learning_rate": 0.0004928816080306906, "loss": 1.9592, "step": 1098 }, { "epoch": 0.10732421875, "grad_norm": 0.21227046847343445, "learning_rate": 0.0004928638512393555, "loss": 1.9523, "step": 1099 }, { "epoch": 0.107421875, "grad_norm": 0.26643091440200806, "learning_rate": 0.0004928460726853518, "loss": 1.9506, "step": 1100 }, { "epoch": 0.10751953125, "grad_norm": 0.2661186754703522, "learning_rate": 0.0004928282723704554, "loss": 1.9425, "step": 1101 }, { "epoch": 0.1076171875, "grad_norm": 0.25391289591789246, "learning_rate": 0.0004928104502964442, "loss": 1.9551, "step": 1102 }, { "epoch": 0.10771484375, "grad_norm": 0.276909202337265, "learning_rate": 0.0004927926064650988, "loss": 1.9416, "step": 1103 }, { "epoch": 0.1078125, "grad_norm": 0.25135815143585205, "learning_rate": 0.0004927747408782014, "loss": 1.9197, "step": 1104 }, { "epoch": 0.10791015625, "grad_norm": 0.2529551684856415, "learning_rate": 0.0004927568535375368, "loss": 1.9561, "step": 1105 }, { "epoch": 0.1080078125, "grad_norm": 0.2697699964046478, "learning_rate": 0.0004927389444448915, "loss": 1.9369, "step": 1106 }, { "epoch": 0.10810546875, "grad_norm": 0.2805427014827728, "learning_rate": 0.0004927210136020548, "loss": 1.9535, "step": 1107 }, { "epoch": 0.108203125, "grad_norm": 0.33747565746307373, "learning_rate": 0.0004927030610108176, "loss": 1.9582, "step": 1108 }, { "epoch": 0.10830078125, "grad_norm": 0.30215132236480713, "learning_rate": 0.0004926850866729734, "loss": 1.9245, "step": 1109 }, { "epoch": 0.1083984375, "grad_norm": 0.3483729064464569, "learning_rate": 0.0004926670905903174, "loss": 1.9959, "step": 1110 }, { "epoch": 0.10849609375, "grad_norm": 0.23263247311115265, "learning_rate": 0.0004926490727646476, "loss": 1.9224, "step": 1111 }, { "epoch": 0.10859375, "grad_norm": 0.3023853898048401, "learning_rate": 0.0004926310331977633, "loss": 1.9493, "step": 1112 }, { "epoch": 0.10869140625, "grad_norm": 0.3592170774936676, "learning_rate": 0.0004926129718914672, "loss": 1.9082, "step": 1113 }, { "epoch": 0.1087890625, "grad_norm": 0.29466864466667175, "learning_rate": 0.0004925948888475628, "loss": 2.016, "step": 1114 }, { "epoch": 0.10888671875, "grad_norm": 0.3136419355869293, "learning_rate": 0.0004925767840678569, "loss": 1.9592, "step": 1115 }, { "epoch": 0.108984375, "grad_norm": 0.21686695516109467, "learning_rate": 0.0004925586575541578, "loss": 1.9081, "step": 1116 }, { "epoch": 0.10908203125, "grad_norm": 0.2520851790904999, "learning_rate": 0.0004925405093082761, "loss": 1.9418, "step": 1117 }, { "epoch": 0.1091796875, "grad_norm": 0.28278669714927673, "learning_rate": 0.0004925223393320247, "loss": 1.9815, "step": 1118 }, { "epoch": 0.10927734375, "grad_norm": 0.2773968577384949, "learning_rate": 0.0004925041476272189, "loss": 1.8888, "step": 1119 }, { "epoch": 0.109375, "grad_norm": 0.24399970471858978, "learning_rate": 0.0004924859341956755, "loss": 1.9548, "step": 1120 }, { "epoch": 0.10947265625, "grad_norm": 0.26073598861694336, "learning_rate": 0.0004924676990392141, "loss": 1.9112, "step": 1121 }, { "epoch": 0.1095703125, "grad_norm": 0.24541547894477844, "learning_rate": 0.0004924494421596562, "loss": 1.914, "step": 1122 }, { "epoch": 0.10966796875, "grad_norm": 0.21719145774841309, "learning_rate": 0.0004924311635588254, "loss": 1.9204, "step": 1123 }, { "epoch": 0.109765625, "grad_norm": 0.295460045337677, "learning_rate": 0.0004924128632385476, "loss": 1.9545, "step": 1124 }, { "epoch": 0.10986328125, "grad_norm": 0.3058072626590729, "learning_rate": 0.0004923945412006509, "loss": 1.9048, "step": 1125 }, { "epoch": 0.1099609375, "grad_norm": 0.19512003660202026, "learning_rate": 0.0004923761974469655, "loss": 1.9581, "step": 1126 }, { "epoch": 0.11005859375, "grad_norm": 0.21684713661670685, "learning_rate": 0.0004923578319793237, "loss": 1.8765, "step": 1127 }, { "epoch": 0.11015625, "grad_norm": 0.2229132503271103, "learning_rate": 0.0004923394447995601, "loss": 1.8599, "step": 1128 }, { "epoch": 0.11025390625, "grad_norm": 0.20479273796081543, "learning_rate": 0.0004923210359095115, "loss": 1.938, "step": 1129 }, { "epoch": 0.1103515625, "grad_norm": 0.2477833479642868, "learning_rate": 0.0004923026053110167, "loss": 1.9655, "step": 1130 }, { "epoch": 0.11044921875, "grad_norm": 0.23927846550941467, "learning_rate": 0.0004922841530059167, "loss": 1.9461, "step": 1131 }, { "epoch": 0.110546875, "grad_norm": 0.24189342558383942, "learning_rate": 0.0004922656789960547, "loss": 1.9338, "step": 1132 }, { "epoch": 0.11064453125, "grad_norm": 0.20208942890167236, "learning_rate": 0.0004922471832832763, "loss": 1.973, "step": 1133 }, { "epoch": 0.1107421875, "grad_norm": 0.20967181026935577, "learning_rate": 0.0004922286658694289, "loss": 1.9391, "step": 1134 }, { "epoch": 0.11083984375, "grad_norm": 0.2538667917251587, "learning_rate": 0.0004922101267563622, "loss": 1.9159, "step": 1135 }, { "epoch": 0.1109375, "grad_norm": 0.2945930063724518, "learning_rate": 0.0004921915659459281, "loss": 1.915, "step": 1136 }, { "epoch": 0.11103515625, "grad_norm": 0.3631344437599182, "learning_rate": 0.0004921729834399808, "loss": 1.9926, "step": 1137 }, { "epoch": 0.1111328125, "grad_norm": 0.32187792658805847, "learning_rate": 0.0004921543792403765, "loss": 1.9137, "step": 1138 }, { "epoch": 0.11123046875, "grad_norm": 0.276929646730423, "learning_rate": 0.0004921357533489734, "loss": 1.9427, "step": 1139 }, { "epoch": 0.111328125, "grad_norm": 0.20455695688724518, "learning_rate": 0.0004921171057676323, "loss": 1.9118, "step": 1140 }, { "epoch": 0.11142578125, "grad_norm": 0.2824450731277466, "learning_rate": 0.0004920984364982157, "loss": 1.9471, "step": 1141 }, { "epoch": 0.1115234375, "grad_norm": 0.2931687831878662, "learning_rate": 0.0004920797455425887, "loss": 1.9546, "step": 1142 }, { "epoch": 0.11162109375, "grad_norm": 0.31234028935432434, "learning_rate": 0.0004920610329026181, "loss": 1.9466, "step": 1143 }, { "epoch": 0.11171875, "grad_norm": 0.39190223813056946, "learning_rate": 0.0004920422985801735, "loss": 1.9394, "step": 1144 }, { "epoch": 0.11181640625, "grad_norm": 0.26808232069015503, "learning_rate": 0.0004920235425771261, "loss": 1.9278, "step": 1145 }, { "epoch": 0.1119140625, "grad_norm": 0.22590310871601105, "learning_rate": 0.0004920047648953493, "loss": 1.9467, "step": 1146 }, { "epoch": 0.11201171875, "grad_norm": 0.2531168758869171, "learning_rate": 0.0004919859655367192, "loss": 1.958, "step": 1147 }, { "epoch": 0.112109375, "grad_norm": 0.207502081990242, "learning_rate": 0.0004919671445031132, "loss": 2.0148, "step": 1148 }, { "epoch": 0.11220703125, "grad_norm": 0.23596623539924622, "learning_rate": 0.0004919483017964118, "loss": 1.9617, "step": 1149 }, { "epoch": 0.1123046875, "grad_norm": 0.25810694694519043, "learning_rate": 0.000491929437418497, "loss": 2.0262, "step": 1150 }, { "epoch": 0.11240234375, "grad_norm": 0.27257680892944336, "learning_rate": 0.0004919105513712533, "loss": 1.9285, "step": 1151 }, { "epoch": 0.1125, "grad_norm": 0.2861190736293793, "learning_rate": 0.000491891643656567, "loss": 1.965, "step": 1152 }, { "epoch": 0.11259765625, "grad_norm": 0.2790696620941162, "learning_rate": 0.0004918727142763271, "loss": 1.9572, "step": 1153 }, { "epoch": 0.1126953125, "grad_norm": 0.35226336121559143, "learning_rate": 0.0004918537632324243, "loss": 1.9338, "step": 1154 }, { "epoch": 0.11279296875, "grad_norm": 0.2735171616077423, "learning_rate": 0.0004918347905267517, "loss": 1.9304, "step": 1155 }, { "epoch": 0.112890625, "grad_norm": 0.2629276514053345, "learning_rate": 0.0004918157961612046, "loss": 1.9383, "step": 1156 }, { "epoch": 0.11298828125, "grad_norm": 0.3583827614784241, "learning_rate": 0.0004917967801376802, "loss": 1.9448, "step": 1157 }, { "epoch": 0.1130859375, "grad_norm": 0.2546387016773224, "learning_rate": 0.0004917777424580781, "loss": 1.9506, "step": 1158 }, { "epoch": 0.11318359375, "grad_norm": 0.35346466302871704, "learning_rate": 0.0004917586831243001, "loss": 1.9207, "step": 1159 }, { "epoch": 0.11328125, "grad_norm": 0.2943922281265259, "learning_rate": 0.0004917396021382499, "loss": 1.961, "step": 1160 }, { "epoch": 0.11337890625, "grad_norm": 0.24972833693027496, "learning_rate": 0.0004917204995018336, "loss": 1.9271, "step": 1161 }, { "epoch": 0.1134765625, "grad_norm": 0.31200850009918213, "learning_rate": 0.0004917013752169594, "loss": 1.9444, "step": 1162 }, { "epoch": 0.11357421875, "grad_norm": 0.3580279052257538, "learning_rate": 0.0004916822292855375, "loss": 1.9132, "step": 1163 }, { "epoch": 0.113671875, "grad_norm": 0.31269049644470215, "learning_rate": 0.0004916630617094807, "loss": 1.9283, "step": 1164 }, { "epoch": 0.11376953125, "grad_norm": 0.25159773230552673, "learning_rate": 0.0004916438724907033, "loss": 1.9018, "step": 1165 }, { "epoch": 0.1138671875, "grad_norm": 0.22316916286945343, "learning_rate": 0.0004916246616311225, "loss": 1.8821, "step": 1166 }, { "epoch": 0.11396484375, "grad_norm": 0.31150946021080017, "learning_rate": 0.0004916054291326571, "loss": 1.9565, "step": 1167 }, { "epoch": 0.1140625, "grad_norm": 0.29112792015075684, "learning_rate": 0.0004915861749972281, "loss": 1.9582, "step": 1168 }, { "epoch": 0.11416015625, "grad_norm": 0.26860764622688293, "learning_rate": 0.0004915668992267593, "loss": 1.9299, "step": 1169 }, { "epoch": 0.1142578125, "grad_norm": 0.24470628798007965, "learning_rate": 0.0004915476018231757, "loss": 1.9131, "step": 1170 }, { "epoch": 0.11435546875, "grad_norm": 0.26279789209365845, "learning_rate": 0.0004915282827884051, "loss": 1.9055, "step": 1171 }, { "epoch": 0.114453125, "grad_norm": 0.21272587776184082, "learning_rate": 0.0004915089421243773, "loss": 1.9694, "step": 1172 }, { "epoch": 0.11455078125, "grad_norm": 0.21711157262325287, "learning_rate": 0.0004914895798330244, "loss": 1.9644, "step": 1173 }, { "epoch": 0.1146484375, "grad_norm": 0.21599583327770233, "learning_rate": 0.0004914701959162803, "loss": 1.9528, "step": 1174 }, { "epoch": 0.11474609375, "grad_norm": 0.22919407486915588, "learning_rate": 0.0004914507903760814, "loss": 1.9353, "step": 1175 }, { "epoch": 0.11484375, "grad_norm": 0.22974969446659088, "learning_rate": 0.0004914313632143661, "loss": 1.9148, "step": 1176 }, { "epoch": 0.11494140625, "grad_norm": 0.2370641529560089, "learning_rate": 0.000491411914433075, "loss": 1.9056, "step": 1177 }, { "epoch": 0.1150390625, "grad_norm": 0.27402329444885254, "learning_rate": 0.000491392444034151, "loss": 1.9483, "step": 1178 }, { "epoch": 0.11513671875, "grad_norm": 0.2953847050666809, "learning_rate": 0.0004913729520195388, "loss": 1.9553, "step": 1179 }, { "epoch": 0.115234375, "grad_norm": 0.26085543632507324, "learning_rate": 0.0004913534383911857, "loss": 1.9307, "step": 1180 }, { "epoch": 0.11533203125, "grad_norm": 0.2782340943813324, "learning_rate": 0.0004913339031510408, "loss": 1.9733, "step": 1181 }, { "epoch": 0.1154296875, "grad_norm": 0.2568228244781494, "learning_rate": 0.0004913143463010554, "loss": 1.9402, "step": 1182 }, { "epoch": 0.11552734375, "grad_norm": 0.2584892809391022, "learning_rate": 0.0004912947678431833, "loss": 1.9254, "step": 1183 }, { "epoch": 0.115625, "grad_norm": 0.2621140778064728, "learning_rate": 0.00049127516777938, "loss": 1.9323, "step": 1184 }, { "epoch": 0.11572265625, "grad_norm": 0.28239843249320984, "learning_rate": 0.0004912555461116035, "loss": 1.9447, "step": 1185 }, { "epoch": 0.1158203125, "grad_norm": 0.3147210478782654, "learning_rate": 0.0004912359028418138, "loss": 1.8867, "step": 1186 }, { "epoch": 0.11591796875, "grad_norm": 0.25541990995407104, "learning_rate": 0.0004912162379719731, "loss": 1.871, "step": 1187 }, { "epoch": 0.116015625, "grad_norm": 0.24328139424324036, "learning_rate": 0.0004911965515040457, "loss": 1.9316, "step": 1188 }, { "epoch": 0.11611328125, "grad_norm": 0.20230042934417725, "learning_rate": 0.0004911768434399981, "loss": 1.9298, "step": 1189 }, { "epoch": 0.1162109375, "grad_norm": 0.32389941811561584, "learning_rate": 0.0004911571137817991, "loss": 1.8941, "step": 1190 }, { "epoch": 0.11630859375, "grad_norm": 0.3173327147960663, "learning_rate": 0.0004911373625314194, "loss": 1.9222, "step": 1191 }, { "epoch": 0.11640625, "grad_norm": 0.32637453079223633, "learning_rate": 0.0004911175896908319, "loss": 1.9273, "step": 1192 }, { "epoch": 0.11650390625, "grad_norm": 0.307746022939682, "learning_rate": 0.0004910977952620119, "loss": 1.9352, "step": 1193 }, { "epoch": 0.1166015625, "grad_norm": 0.21165402233600616, "learning_rate": 0.0004910779792469365, "loss": 1.8867, "step": 1194 }, { "epoch": 0.11669921875, "grad_norm": 0.3059009909629822, "learning_rate": 0.0004910581416475854, "loss": 1.9187, "step": 1195 }, { "epoch": 0.116796875, "grad_norm": 0.2690452039241791, "learning_rate": 0.00049103828246594, "loss": 1.9723, "step": 1196 }, { "epoch": 0.11689453125, "grad_norm": 0.25487321615219116, "learning_rate": 0.0004910184017039841, "loss": 1.9525, "step": 1197 }, { "epoch": 0.1169921875, "grad_norm": 0.33227404952049255, "learning_rate": 0.0004909984993637036, "loss": 1.9062, "step": 1198 }, { "epoch": 0.11708984375, "grad_norm": 0.32993388175964355, "learning_rate": 0.0004909785754470866, "loss": 1.9327, "step": 1199 }, { "epoch": 0.1171875, "grad_norm": 0.28574901819229126, "learning_rate": 0.0004909586299561233, "loss": 1.9481, "step": 1200 }, { "epoch": 0.11728515625, "grad_norm": 0.27981141209602356, "learning_rate": 0.0004909386628928061, "loss": 1.9284, "step": 1201 }, { "epoch": 0.1173828125, "grad_norm": 0.26131895184516907, "learning_rate": 0.0004909186742591296, "loss": 1.9276, "step": 1202 }, { "epoch": 0.11748046875, "grad_norm": 0.3690519630908966, "learning_rate": 0.0004908986640570903, "loss": 1.8733, "step": 1203 }, { "epoch": 0.117578125, "grad_norm": 0.38982152938842773, "learning_rate": 0.000490878632288687, "loss": 1.9204, "step": 1204 }, { "epoch": 0.11767578125, "grad_norm": 0.2588847875595093, "learning_rate": 0.000490858578955921, "loss": 1.8941, "step": 1205 }, { "epoch": 0.1177734375, "grad_norm": 0.3161073625087738, "learning_rate": 0.0004908385040607954, "loss": 1.9077, "step": 1206 }, { "epoch": 0.11787109375, "grad_norm": 0.26444011926651, "learning_rate": 0.0004908184076053153, "loss": 1.9471, "step": 1207 }, { "epoch": 0.11796875, "grad_norm": 0.28895479440689087, "learning_rate": 0.0004907982895914883, "loss": 1.9159, "step": 1208 }, { "epoch": 0.11806640625, "grad_norm": 0.264992356300354, "learning_rate": 0.000490778150021324, "loss": 1.9383, "step": 1209 }, { "epoch": 0.1181640625, "grad_norm": 0.25046393275260925, "learning_rate": 0.0004907579888968341, "loss": 1.9756, "step": 1210 }, { "epoch": 0.11826171875, "grad_norm": 0.2875134348869324, "learning_rate": 0.0004907378062200325, "loss": 1.9368, "step": 1211 }, { "epoch": 0.118359375, "grad_norm": 0.24127423763275146, "learning_rate": 0.0004907176019929353, "loss": 1.9399, "step": 1212 }, { "epoch": 0.11845703125, "grad_norm": 0.25520575046539307, "learning_rate": 0.0004906973762175608, "loss": 1.9662, "step": 1213 }, { "epoch": 0.1185546875, "grad_norm": 0.23595774173736572, "learning_rate": 0.0004906771288959293, "loss": 1.9302, "step": 1214 }, { "epoch": 0.11865234375, "grad_norm": 0.2919490337371826, "learning_rate": 0.0004906568600300635, "loss": 1.9421, "step": 1215 }, { "epoch": 0.11875, "grad_norm": 0.23202334344387054, "learning_rate": 0.0004906365696219877, "loss": 1.8934, "step": 1216 }, { "epoch": 0.11884765625, "grad_norm": 0.2639678716659546, "learning_rate": 0.0004906162576737292, "loss": 1.9272, "step": 1217 }, { "epoch": 0.1189453125, "grad_norm": 0.24962055683135986, "learning_rate": 0.0004905959241873165, "loss": 1.9605, "step": 1218 }, { "epoch": 0.11904296875, "grad_norm": 0.2519899308681488, "learning_rate": 0.0004905755691647812, "loss": 1.9757, "step": 1219 }, { "epoch": 0.119140625, "grad_norm": 0.2044687420129776, "learning_rate": 0.0004905551926081563, "loss": 1.9817, "step": 1220 }, { "epoch": 0.11923828125, "grad_norm": 0.21942391991615295, "learning_rate": 0.0004905347945194772, "loss": 1.9705, "step": 1221 }, { "epoch": 0.1193359375, "grad_norm": 0.26614969968795776, "learning_rate": 0.0004905143749007817, "loss": 1.9375, "step": 1222 }, { "epoch": 0.11943359375, "grad_norm": 0.23212748765945435, "learning_rate": 0.0004904939337541093, "loss": 1.9728, "step": 1223 }, { "epoch": 0.11953125, "grad_norm": 0.28004127740859985, "learning_rate": 0.0004904734710815022, "loss": 1.9082, "step": 1224 }, { "epoch": 0.11962890625, "grad_norm": 0.3012614846229553, "learning_rate": 0.0004904529868850042, "loss": 1.9347, "step": 1225 }, { "epoch": 0.1197265625, "grad_norm": 0.2914253771305084, "learning_rate": 0.0004904324811666615, "loss": 1.9435, "step": 1226 }, { "epoch": 0.11982421875, "grad_norm": 0.2315436154603958, "learning_rate": 0.0004904119539285225, "loss": 1.9503, "step": 1227 }, { "epoch": 0.119921875, "grad_norm": 0.2043173611164093, "learning_rate": 0.0004903914051726377, "loss": 1.937, "step": 1228 }, { "epoch": 0.12001953125, "grad_norm": 0.26184800267219543, "learning_rate": 0.0004903708349010598, "loss": 1.9459, "step": 1229 }, { "epoch": 0.1201171875, "grad_norm": 0.24898020923137665, "learning_rate": 0.0004903502431158434, "loss": 1.935, "step": 1230 }, { "epoch": 0.12021484375, "grad_norm": 0.29706010222435, "learning_rate": 0.0004903296298190456, "loss": 1.9741, "step": 1231 }, { "epoch": 0.1203125, "grad_norm": 0.3235405385494232, "learning_rate": 0.0004903089950127254, "loss": 1.9319, "step": 1232 }, { "epoch": 0.12041015625, "grad_norm": 0.25629666447639465, "learning_rate": 0.0004902883386989441, "loss": 1.9533, "step": 1233 }, { "epoch": 0.1205078125, "grad_norm": 0.23315580189228058, "learning_rate": 0.0004902676608797649, "loss": 1.9033, "step": 1234 }, { "epoch": 0.12060546875, "grad_norm": 0.2336481362581253, "learning_rate": 0.0004902469615572537, "loss": 1.9123, "step": 1235 }, { "epoch": 0.120703125, "grad_norm": 0.27162161469459534, "learning_rate": 0.0004902262407334779, "loss": 1.9253, "step": 1236 }, { "epoch": 0.12080078125, "grad_norm": 0.2913827896118164, "learning_rate": 0.0004902054984105073, "loss": 1.9526, "step": 1237 }, { "epoch": 0.1208984375, "grad_norm": 0.30284538865089417, "learning_rate": 0.0004901847345904141, "loss": 1.9588, "step": 1238 }, { "epoch": 0.12099609375, "grad_norm": 0.2579716742038727, "learning_rate": 0.000490163949275272, "loss": 1.892, "step": 1239 }, { "epoch": 0.12109375, "grad_norm": 0.21740810573101044, "learning_rate": 0.0004901431424671577, "loss": 1.9584, "step": 1240 }, { "epoch": 0.12119140625, "grad_norm": 0.3371858298778534, "learning_rate": 0.0004901223141681496, "loss": 1.9415, "step": 1241 }, { "epoch": 0.1212890625, "grad_norm": 0.34596753120422363, "learning_rate": 0.000490101464380328, "loss": 1.9384, "step": 1242 }, { "epoch": 0.12138671875, "grad_norm": 0.2591091990470886, "learning_rate": 0.0004900805931057757, "loss": 1.9433, "step": 1243 }, { "epoch": 0.121484375, "grad_norm": 0.27759140729904175, "learning_rate": 0.0004900597003465777, "loss": 1.9124, "step": 1244 }, { "epoch": 0.12158203125, "grad_norm": 0.2503783106803894, "learning_rate": 0.0004900387861048208, "loss": 1.9426, "step": 1245 }, { "epoch": 0.1216796875, "grad_norm": 0.36085277795791626, "learning_rate": 0.0004900178503825942, "loss": 1.9141, "step": 1246 }, { "epoch": 0.12177734375, "grad_norm": 0.2831665873527527, "learning_rate": 0.0004899968931819893, "loss": 1.9515, "step": 1247 }, { "epoch": 0.121875, "grad_norm": 0.26685988903045654, "learning_rate": 0.0004899759145050996, "loss": 1.9752, "step": 1248 }, { "epoch": 0.12197265625, "grad_norm": 0.3139730393886566, "learning_rate": 0.0004899549143540205, "loss": 1.9298, "step": 1249 }, { "epoch": 0.1220703125, "grad_norm": 0.20647141337394714, "learning_rate": 0.0004899338927308497, "loss": 1.9233, "step": 1250 }, { "epoch": 0.12216796875, "grad_norm": 0.293544203042984, "learning_rate": 0.0004899128496376871, "loss": 1.9534, "step": 1251 }, { "epoch": 0.122265625, "grad_norm": 0.28824836015701294, "learning_rate": 0.0004898917850766349, "loss": 1.96, "step": 1252 }, { "epoch": 0.12236328125, "grad_norm": 0.23198992013931274, "learning_rate": 0.0004898706990497971, "loss": 1.9275, "step": 1253 }, { "epoch": 0.1224609375, "grad_norm": 0.27970659732818604, "learning_rate": 0.00048984959155928, "loss": 1.8998, "step": 1254 }, { "epoch": 0.12255859375, "grad_norm": 0.2355661541223526, "learning_rate": 0.0004898284626071921, "loss": 1.9313, "step": 1255 }, { "epoch": 0.12265625, "grad_norm": 0.3456537127494812, "learning_rate": 0.0004898073121956441, "loss": 1.9164, "step": 1256 }, { "epoch": 0.12275390625, "grad_norm": 0.32490289211273193, "learning_rate": 0.0004897861403267486, "loss": 1.9568, "step": 1257 }, { "epoch": 0.1228515625, "grad_norm": 0.37740960717201233, "learning_rate": 0.0004897649470026205, "loss": 1.9618, "step": 1258 }, { "epoch": 0.12294921875, "grad_norm": 0.34424281120300293, "learning_rate": 0.0004897437322253767, "loss": 1.9192, "step": 1259 }, { "epoch": 0.123046875, "grad_norm": 0.1956419050693512, "learning_rate": 0.0004897224959971368, "loss": 1.9128, "step": 1260 }, { "epoch": 0.12314453125, "grad_norm": 0.30533185601234436, "learning_rate": 0.0004897012383200216, "loss": 1.9378, "step": 1261 }, { "epoch": 0.1232421875, "grad_norm": 0.28745076060295105, "learning_rate": 0.0004896799591961548, "loss": 1.9734, "step": 1262 }, { "epoch": 0.12333984375, "grad_norm": 0.3063562512397766, "learning_rate": 0.000489658658627662, "loss": 1.9713, "step": 1263 }, { "epoch": 0.1234375, "grad_norm": 0.26775872707366943, "learning_rate": 0.000489637336616671, "loss": 1.9305, "step": 1264 }, { "epoch": 0.12353515625, "grad_norm": 0.2045804262161255, "learning_rate": 0.0004896159931653114, "loss": 1.9018, "step": 1265 }, { "epoch": 0.1236328125, "grad_norm": 0.26752471923828125, "learning_rate": 0.0004895946282757155, "loss": 1.9334, "step": 1266 }, { "epoch": 0.12373046875, "grad_norm": 0.3504626452922821, "learning_rate": 0.0004895732419500174, "loss": 1.9678, "step": 1267 }, { "epoch": 0.123828125, "grad_norm": 0.2588043808937073, "learning_rate": 0.0004895518341903534, "loss": 1.9336, "step": 1268 }, { "epoch": 0.12392578125, "grad_norm": 0.23393654823303223, "learning_rate": 0.0004895304049988619, "loss": 1.9386, "step": 1269 }, { "epoch": 0.1240234375, "grad_norm": 0.3162434697151184, "learning_rate": 0.0004895089543776834, "loss": 1.9752, "step": 1270 }, { "epoch": 0.12412109375, "grad_norm": 0.28987085819244385, "learning_rate": 0.0004894874823289609, "loss": 1.9346, "step": 1271 }, { "epoch": 0.12421875, "grad_norm": 0.2837126553058624, "learning_rate": 0.000489465988854839, "loss": 1.9375, "step": 1272 }, { "epoch": 0.12431640625, "grad_norm": 0.26402387022972107, "learning_rate": 0.0004894444739574649, "loss": 1.9235, "step": 1273 }, { "epoch": 0.1244140625, "grad_norm": 0.21077674627304077, "learning_rate": 0.0004894229376389875, "loss": 1.9524, "step": 1274 }, { "epoch": 0.12451171875, "grad_norm": 0.18328504264354706, "learning_rate": 0.0004894013799015583, "loss": 1.9496, "step": 1275 }, { "epoch": 0.124609375, "grad_norm": 0.22258712351322174, "learning_rate": 0.0004893798007473307, "loss": 1.9661, "step": 1276 }, { "epoch": 0.12470703125, "grad_norm": 0.20786426961421967, "learning_rate": 0.0004893582001784604, "loss": 1.9263, "step": 1277 }, { "epoch": 0.1248046875, "grad_norm": 0.1959957331418991, "learning_rate": 0.0004893365781971047, "loss": 1.9525, "step": 1278 }, { "epoch": 0.12490234375, "grad_norm": 0.2622356712818146, "learning_rate": 0.0004893149348054238, "loss": 1.9433, "step": 1279 }, { "epoch": 0.125, "grad_norm": 0.2329346388578415, "learning_rate": 0.0004892932700055797, "loss": 1.9293, "step": 1280 }, { "epoch": 0.12509765625, "grad_norm": 0.26422178745269775, "learning_rate": 0.0004892715837997361, "loss": 1.9252, "step": 1281 }, { "epoch": 0.1251953125, "grad_norm": 0.30737435817718506, "learning_rate": 0.0004892498761900597, "loss": 1.9171, "step": 1282 }, { "epoch": 0.12529296875, "grad_norm": 0.30651819705963135, "learning_rate": 0.0004892281471787187, "loss": 1.937, "step": 1283 }, { "epoch": 0.125390625, "grad_norm": 0.23027795553207397, "learning_rate": 0.0004892063967678838, "loss": 1.9491, "step": 1284 }, { "epoch": 0.12548828125, "grad_norm": 0.30681830644607544, "learning_rate": 0.0004891846249597276, "loss": 1.905, "step": 1285 }, { "epoch": 0.1255859375, "grad_norm": 0.34682074189186096, "learning_rate": 0.0004891628317564247, "loss": 1.9157, "step": 1286 }, { "epoch": 0.12568359375, "grad_norm": 0.26123619079589844, "learning_rate": 0.0004891410171601523, "loss": 1.9382, "step": 1287 }, { "epoch": 0.12578125, "grad_norm": 0.32031431794166565, "learning_rate": 0.0004891191811730894, "loss": 1.9234, "step": 1288 }, { "epoch": 0.12587890625, "grad_norm": 0.24917498230934143, "learning_rate": 0.0004890973237974172, "loss": 1.919, "step": 1289 }, { "epoch": 0.1259765625, "grad_norm": 0.29161348938941956, "learning_rate": 0.0004890754450353192, "loss": 1.9405, "step": 1290 }, { "epoch": 0.12607421875, "grad_norm": 0.31046807765960693, "learning_rate": 0.0004890535448889807, "loss": 1.9298, "step": 1291 }, { "epoch": 0.126171875, "grad_norm": 0.30187463760375977, "learning_rate": 0.0004890316233605893, "loss": 1.9329, "step": 1292 }, { "epoch": 0.12626953125, "grad_norm": 0.3013293445110321, "learning_rate": 0.000489009680452335, "loss": 1.9228, "step": 1293 }, { "epoch": 0.1263671875, "grad_norm": 0.2492775321006775, "learning_rate": 0.0004889877161664096, "loss": 1.9565, "step": 1294 }, { "epoch": 0.12646484375, "grad_norm": 0.29035332798957825, "learning_rate": 0.000488965730505007, "loss": 1.9184, "step": 1295 }, { "epoch": 0.1265625, "grad_norm": 0.1983688622713089, "learning_rate": 0.0004889437234703236, "loss": 1.97, "step": 1296 }, { "epoch": 0.12666015625, "grad_norm": 0.24074336886405945, "learning_rate": 0.0004889216950645576, "loss": 1.9289, "step": 1297 }, { "epoch": 0.1267578125, "grad_norm": 0.2601425051689148, "learning_rate": 0.0004888996452899093, "loss": 1.9374, "step": 1298 }, { "epoch": 0.12685546875, "grad_norm": 0.21484848856925964, "learning_rate": 0.0004888775741485816, "loss": 1.9075, "step": 1299 }, { "epoch": 0.126953125, "grad_norm": 0.2370765060186386, "learning_rate": 0.0004888554816427789, "loss": 1.9131, "step": 1300 }, { "epoch": 0.12705078125, "grad_norm": 0.22826816141605377, "learning_rate": 0.0004888333677747082, "loss": 1.9514, "step": 1301 }, { "epoch": 0.1271484375, "grad_norm": 0.2141917198896408, "learning_rate": 0.0004888112325465785, "loss": 1.8406, "step": 1302 }, { "epoch": 0.12724609375, "grad_norm": 0.22987738251686096, "learning_rate": 0.0004887890759606008, "loss": 1.9498, "step": 1303 }, { "epoch": 0.12734375, "grad_norm": 0.2592550814151764, "learning_rate": 0.0004887668980189885, "loss": 1.9427, "step": 1304 }, { "epoch": 0.12744140625, "grad_norm": 0.24788032472133636, "learning_rate": 0.0004887446987239567, "loss": 1.9718, "step": 1305 }, { "epoch": 0.1275390625, "grad_norm": 0.3112640380859375, "learning_rate": 0.0004887224780777233, "loss": 1.972, "step": 1306 }, { "epoch": 0.12763671875, "grad_norm": 0.2711445987224579, "learning_rate": 0.0004887002360825077, "loss": 1.9636, "step": 1307 }, { "epoch": 0.127734375, "grad_norm": 0.19959181547164917, "learning_rate": 0.0004886779727405318, "loss": 1.9142, "step": 1308 }, { "epoch": 0.12783203125, "grad_norm": 0.24054476618766785, "learning_rate": 0.0004886556880540192, "loss": 1.8724, "step": 1309 }, { "epoch": 0.1279296875, "grad_norm": 0.21965253353118896, "learning_rate": 0.0004886333820251963, "loss": 1.9295, "step": 1310 }, { "epoch": 0.12802734375, "grad_norm": 0.2383204847574234, "learning_rate": 0.0004886110546562912, "loss": 1.8945, "step": 1311 }, { "epoch": 0.128125, "grad_norm": 0.20343053340911865, "learning_rate": 0.0004885887059495342, "loss": 1.9364, "step": 1312 }, { "epoch": 0.12822265625, "grad_norm": 0.24133317172527313, "learning_rate": 0.0004885663359071576, "loss": 1.9023, "step": 1313 }, { "epoch": 0.1283203125, "grad_norm": 0.27090492844581604, "learning_rate": 0.0004885439445313962, "loss": 1.9007, "step": 1314 }, { "epoch": 0.12841796875, "grad_norm": 0.2153930515050888, "learning_rate": 0.0004885215318244865, "loss": 1.9475, "step": 1315 }, { "epoch": 0.128515625, "grad_norm": 0.3661288022994995, "learning_rate": 0.0004884990977886673, "loss": 1.9858, "step": 1316 }, { "epoch": 0.12861328125, "grad_norm": 0.4454726576805115, "learning_rate": 0.0004884766424261798, "loss": 1.9207, "step": 1317 }, { "epoch": 0.1287109375, "grad_norm": 0.4249671995639801, "learning_rate": 0.000488454165739267, "loss": 1.9889, "step": 1318 }, { "epoch": 0.12880859375, "grad_norm": 0.3520338237285614, "learning_rate": 0.0004884316677301739, "loss": 1.9564, "step": 1319 }, { "epoch": 0.12890625, "grad_norm": 0.28052136301994324, "learning_rate": 0.0004884091484011482, "loss": 1.9677, "step": 1320 }, { "epoch": 0.12900390625, "grad_norm": 0.31618887186050415, "learning_rate": 0.0004883866077544391, "loss": 1.9306, "step": 1321 }, { "epoch": 0.1291015625, "grad_norm": 0.29120492935180664, "learning_rate": 0.0004883640457922984, "loss": 1.9205, "step": 1322 }, { "epoch": 0.12919921875, "grad_norm": 0.23774267733097076, "learning_rate": 0.0004883414625169797, "loss": 1.9078, "step": 1323 }, { "epoch": 0.129296875, "grad_norm": 0.3183814287185669, "learning_rate": 0.0004883188579307389, "loss": 1.7927, "step": 1324 }, { "epoch": 0.12939453125, "grad_norm": 0.2845805585384369, "learning_rate": 0.0004882962320358341, "loss": 1.888, "step": 1325 }, { "epoch": 0.1294921875, "grad_norm": 0.2877943515777588, "learning_rate": 0.00048827358483452527, "loss": 1.9307, "step": 1326 }, { "epoch": 0.12958984375, "grad_norm": 0.2753991186618805, "learning_rate": 0.0004882509163290748, "loss": 1.9306, "step": 1327 }, { "epoch": 0.1296875, "grad_norm": 0.2753439247608185, "learning_rate": 0.00048822822652174705, "loss": 1.9573, "step": 1328 }, { "epoch": 0.12978515625, "grad_norm": 0.2690608501434326, "learning_rate": 0.00048820551541480845, "loss": 1.9229, "step": 1329 }, { "epoch": 0.1298828125, "grad_norm": 0.2276460975408554, "learning_rate": 0.00048818278301052776, "loss": 1.9871, "step": 1330 }, { "epoch": 0.12998046875, "grad_norm": 0.23562243580818176, "learning_rate": 0.00048816002931117557, "loss": 1.9489, "step": 1331 }, { "epoch": 0.130078125, "grad_norm": 0.27421703934669495, "learning_rate": 0.00048813725431902496, "loss": 1.922, "step": 1332 }, { "epoch": 0.13017578125, "grad_norm": 0.1969674974679947, "learning_rate": 0.00048811445803635087, "loss": 1.929, "step": 1333 }, { "epoch": 0.1302734375, "grad_norm": 0.22562848031520844, "learning_rate": 0.00048809164046543043, "loss": 1.9183, "step": 1334 }, { "epoch": 0.13037109375, "grad_norm": 0.2235979586839676, "learning_rate": 0.0004880688016085429, "loss": 1.9155, "step": 1335 }, { "epoch": 0.13046875, "grad_norm": 0.23072493076324463, "learning_rate": 0.00048804594146796977, "loss": 1.9647, "step": 1336 }, { "epoch": 0.13056640625, "grad_norm": 0.2528524398803711, "learning_rate": 0.00048802306004599446, "loss": 1.979, "step": 1337 }, { "epoch": 0.1306640625, "grad_norm": 0.2586860954761505, "learning_rate": 0.00048800015734490267, "loss": 1.9371, "step": 1338 }, { "epoch": 0.13076171875, "grad_norm": 0.21609428524971008, "learning_rate": 0.0004879772333669822, "loss": 1.9319, "step": 1339 }, { "epoch": 0.130859375, "grad_norm": 0.28810006380081177, "learning_rate": 0.0004879542881145229, "loss": 1.9218, "step": 1340 }, { "epoch": 0.13095703125, "grad_norm": 0.288849413394928, "learning_rate": 0.0004879313215898168, "loss": 1.929, "step": 1341 }, { "epoch": 0.1310546875, "grad_norm": 0.3168569803237915, "learning_rate": 0.00048790833379515813, "loss": 1.9276, "step": 1342 }, { "epoch": 0.13115234375, "grad_norm": 0.3000844120979309, "learning_rate": 0.00048788532473284307, "loss": 1.9337, "step": 1343 }, { "epoch": 0.13125, "grad_norm": 0.2524905204772949, "learning_rate": 0.00048786229440517004, "loss": 1.9105, "step": 1344 }, { "epoch": 0.13134765625, "grad_norm": 0.25065815448760986, "learning_rate": 0.0004878392428144396, "loss": 1.9161, "step": 1345 }, { "epoch": 0.1314453125, "grad_norm": 0.3235273063182831, "learning_rate": 0.00048781616996295443, "loss": 1.9159, "step": 1346 }, { "epoch": 0.13154296875, "grad_norm": 0.23726877570152283, "learning_rate": 0.0004877930758530192, "loss": 1.9063, "step": 1347 }, { "epoch": 0.131640625, "grad_norm": 0.2991824150085449, "learning_rate": 0.000487769960486941, "loss": 1.9885, "step": 1348 }, { "epoch": 0.13173828125, "grad_norm": 0.24819223582744598, "learning_rate": 0.0004877468238670287, "loss": 1.9311, "step": 1349 }, { "epoch": 0.1318359375, "grad_norm": 0.23198595643043518, "learning_rate": 0.00048772366599559333, "loss": 1.9278, "step": 1350 }, { "epoch": 0.13193359375, "grad_norm": 0.3012297451496124, "learning_rate": 0.0004877004868749483, "loss": 1.9258, "step": 1351 }, { "epoch": 0.13203125, "grad_norm": 0.21619653701782227, "learning_rate": 0.0004876772865074091, "loss": 1.9121, "step": 1352 }, { "epoch": 0.13212890625, "grad_norm": 0.24853789806365967, "learning_rate": 0.00048765406489529315, "loss": 1.9533, "step": 1353 }, { "epoch": 0.1322265625, "grad_norm": 0.33258283138275146, "learning_rate": 0.00048763082204091995, "loss": 1.9424, "step": 1354 }, { "epoch": 0.13232421875, "grad_norm": 0.3056190609931946, "learning_rate": 0.00048760755794661137, "loss": 1.9605, "step": 1355 }, { "epoch": 0.132421875, "grad_norm": 0.20714829862117767, "learning_rate": 0.00048758427261469135, "loss": 1.9416, "step": 1356 }, { "epoch": 0.13251953125, "grad_norm": 0.2573312520980835, "learning_rate": 0.0004875609660474857, "loss": 1.9198, "step": 1357 }, { "epoch": 0.1326171875, "grad_norm": 0.24284306168556213, "learning_rate": 0.0004875376382473228, "loss": 1.9308, "step": 1358 }, { "epoch": 0.13271484375, "grad_norm": 0.20341083407402039, "learning_rate": 0.0004875142892165327, "loss": 1.8933, "step": 1359 }, { "epoch": 0.1328125, "grad_norm": 0.3155312240123749, "learning_rate": 0.0004874909189574478, "loss": 1.8825, "step": 1360 }, { "epoch": 0.13291015625, "grad_norm": 0.2829397916793823, "learning_rate": 0.0004874675274724026, "loss": 1.9229, "step": 1361 }, { "epoch": 0.1330078125, "grad_norm": 0.34132909774780273, "learning_rate": 0.00048744411476373374, "loss": 1.8935, "step": 1362 }, { "epoch": 0.13310546875, "grad_norm": 0.41804420948028564, "learning_rate": 0.0004874206808337799, "loss": 1.9241, "step": 1363 }, { "epoch": 0.133203125, "grad_norm": 0.3885767161846161, "learning_rate": 0.0004873972256848819, "loss": 1.9391, "step": 1364 }, { "epoch": 0.13330078125, "grad_norm": 0.3422619104385376, "learning_rate": 0.00048737374931938273, "loss": 1.9113, "step": 1365 }, { "epoch": 0.1333984375, "grad_norm": 0.33838555216789246, "learning_rate": 0.00048735025173962746, "loss": 1.8967, "step": 1366 }, { "epoch": 0.13349609375, "grad_norm": 0.3656925857067108, "learning_rate": 0.0004873267329479633, "loss": 1.9405, "step": 1367 }, { "epoch": 0.13359375, "grad_norm": 0.3086741268634796, "learning_rate": 0.00048730319294673955, "loss": 1.9518, "step": 1368 }, { "epoch": 0.13369140625, "grad_norm": 0.24022363126277924, "learning_rate": 0.0004872796317383077, "loss": 1.9368, "step": 1369 }, { "epoch": 0.1337890625, "grad_norm": 0.2861846089363098, "learning_rate": 0.00048725604932502116, "loss": 1.9186, "step": 1370 }, { "epoch": 0.13388671875, "grad_norm": 0.29110145568847656, "learning_rate": 0.0004872324457092358, "loss": 1.8963, "step": 1371 }, { "epoch": 0.133984375, "grad_norm": 0.2232580929994583, "learning_rate": 0.0004872088208933093, "loss": 1.929, "step": 1372 }, { "epoch": 0.13408203125, "grad_norm": 0.3331020772457123, "learning_rate": 0.0004871851748796016, "loss": 1.936, "step": 1373 }, { "epoch": 0.1341796875, "grad_norm": 0.2353833019733429, "learning_rate": 0.00048716150767047473, "loss": 1.9047, "step": 1374 }, { "epoch": 0.13427734375, "grad_norm": 0.25953710079193115, "learning_rate": 0.0004871378192682928, "loss": 1.9812, "step": 1375 }, { "epoch": 0.134375, "grad_norm": 0.24382802844047546, "learning_rate": 0.0004871141096754222, "loss": 1.9599, "step": 1376 }, { "epoch": 0.13447265625, "grad_norm": 0.2496114820241928, "learning_rate": 0.00048709037889423107, "loss": 1.9094, "step": 1377 }, { "epoch": 0.1345703125, "grad_norm": 0.26899224519729614, "learning_rate": 0.0004870666269270901, "loss": 1.9467, "step": 1378 }, { "epoch": 0.13466796875, "grad_norm": 0.23490363359451294, "learning_rate": 0.00048704285377637184, "loss": 1.9682, "step": 1379 }, { "epoch": 0.134765625, "grad_norm": 0.23190240561962128, "learning_rate": 0.000487019059444451, "loss": 1.9324, "step": 1380 }, { "epoch": 0.13486328125, "grad_norm": 0.2239055186510086, "learning_rate": 0.0004869952439337045, "loss": 1.995, "step": 1381 }, { "epoch": 0.1349609375, "grad_norm": 0.23679253458976746, "learning_rate": 0.00048697140724651114, "loss": 1.8685, "step": 1382 }, { "epoch": 0.13505859375, "grad_norm": 0.2584695816040039, "learning_rate": 0.0004869475493852521, "loss": 1.9421, "step": 1383 }, { "epoch": 0.13515625, "grad_norm": 0.219711571931839, "learning_rate": 0.0004869236703523106, "loss": 1.9326, "step": 1384 }, { "epoch": 0.13525390625, "grad_norm": 0.2176481932401657, "learning_rate": 0.00048689977015007197, "loss": 1.8959, "step": 1385 }, { "epoch": 0.1353515625, "grad_norm": 0.22419661283493042, "learning_rate": 0.0004868758487809235, "loss": 1.9095, "step": 1386 }, { "epoch": 0.13544921875, "grad_norm": 0.255984365940094, "learning_rate": 0.0004868519062472548, "loss": 1.9756, "step": 1387 }, { "epoch": 0.135546875, "grad_norm": 0.217428520321846, "learning_rate": 0.0004868279425514575, "loss": 1.9481, "step": 1388 }, { "epoch": 0.13564453125, "grad_norm": 0.19990497827529907, "learning_rate": 0.00048680395769592534, "loss": 1.9279, "step": 1389 }, { "epoch": 0.1357421875, "grad_norm": 0.2209380716085434, "learning_rate": 0.00048677995168305426, "loss": 1.9148, "step": 1390 }, { "epoch": 0.13583984375, "grad_norm": 0.25395241379737854, "learning_rate": 0.00048675592451524226, "loss": 1.9688, "step": 1391 }, { "epoch": 0.1359375, "grad_norm": 0.1944066882133484, "learning_rate": 0.0004867318761948892, "loss": 1.9005, "step": 1392 }, { "epoch": 0.13603515625, "grad_norm": 0.22825607657432556, "learning_rate": 0.0004867078067243977, "loss": 1.9538, "step": 1393 }, { "epoch": 0.1361328125, "grad_norm": 0.23758959770202637, "learning_rate": 0.0004866837161061717, "loss": 1.9887, "step": 1394 }, { "epoch": 0.13623046875, "grad_norm": 0.26341712474823, "learning_rate": 0.00048665960434261796, "loss": 1.9217, "step": 1395 }, { "epoch": 0.136328125, "grad_norm": 0.283316045999527, "learning_rate": 0.0004866354714361448, "loss": 1.9576, "step": 1396 }, { "epoch": 0.13642578125, "grad_norm": 0.32178181409835815, "learning_rate": 0.000486611317389163, "loss": 1.9651, "step": 1397 }, { "epoch": 0.1365234375, "grad_norm": 0.4065782129764557, "learning_rate": 0.0004865871422040853, "loss": 1.897, "step": 1398 }, { "epoch": 0.13662109375, "grad_norm": 0.3661766052246094, "learning_rate": 0.00048656294588332653, "loss": 1.9472, "step": 1399 }, { "epoch": 0.13671875, "grad_norm": 0.24434155225753784, "learning_rate": 0.0004865387284293038, "loss": 1.9179, "step": 1400 }, { "epoch": 0.13681640625, "grad_norm": 0.4203675091266632, "learning_rate": 0.0004865144898444361, "loss": 1.9583, "step": 1401 }, { "epoch": 0.1369140625, "grad_norm": 0.4013305902481079, "learning_rate": 0.0004864902301311448, "loss": 1.9322, "step": 1402 }, { "epoch": 0.13701171875, "grad_norm": 0.30299681425094604, "learning_rate": 0.00048646594929185313, "loss": 1.9152, "step": 1403 }, { "epoch": 0.137109375, "grad_norm": 0.30768755078315735, "learning_rate": 0.0004864416473289865, "loss": 1.966, "step": 1404 }, { "epoch": 0.13720703125, "grad_norm": 0.2759581506252289, "learning_rate": 0.00048641732424497256, "loss": 1.9211, "step": 1405 }, { "epoch": 0.1373046875, "grad_norm": 0.2976212799549103, "learning_rate": 0.00048639298004224087, "loss": 1.9376, "step": 1406 }, { "epoch": 0.13740234375, "grad_norm": 0.21789424121379852, "learning_rate": 0.0004863686147232232, "loss": 1.994, "step": 1407 }, { "epoch": 0.1375, "grad_norm": 0.2543160617351532, "learning_rate": 0.0004863442282903535, "loss": 1.9153, "step": 1408 }, { "epoch": 0.13759765625, "grad_norm": 0.2789456248283386, "learning_rate": 0.0004863198207460678, "loss": 1.9238, "step": 1409 }, { "epoch": 0.1376953125, "grad_norm": 0.21548853814601898, "learning_rate": 0.0004862953920928041, "loss": 1.9003, "step": 1410 }, { "epoch": 0.13779296875, "grad_norm": 0.2607603967189789, "learning_rate": 0.00048627094233300266, "loss": 1.951, "step": 1411 }, { "epoch": 0.137890625, "grad_norm": 0.21658223867416382, "learning_rate": 0.00048624647146910575, "loss": 1.9818, "step": 1412 }, { "epoch": 0.13798828125, "grad_norm": 0.24333196878433228, "learning_rate": 0.00048622197950355786, "loss": 1.8933, "step": 1413 }, { "epoch": 0.1380859375, "grad_norm": 0.22348099946975708, "learning_rate": 0.0004861974664388054, "loss": 1.9302, "step": 1414 }, { "epoch": 0.13818359375, "grad_norm": 0.1976805180311203, "learning_rate": 0.0004861729322772971, "loss": 1.9044, "step": 1415 }, { "epoch": 0.13828125, "grad_norm": 0.22617708146572113, "learning_rate": 0.00048614837702148374, "loss": 1.9445, "step": 1416 }, { "epoch": 0.13837890625, "grad_norm": 0.2243851125240326, "learning_rate": 0.0004861238006738181, "loss": 1.9215, "step": 1417 }, { "epoch": 0.1384765625, "grad_norm": 0.21161694824695587, "learning_rate": 0.0004860992032367551, "loss": 1.9443, "step": 1418 }, { "epoch": 0.13857421875, "grad_norm": 0.22872799634933472, "learning_rate": 0.0004860745847127519, "loss": 1.8174, "step": 1419 }, { "epoch": 0.138671875, "grad_norm": 0.27401795983314514, "learning_rate": 0.0004860499451042677, "loss": 1.919, "step": 1420 }, { "epoch": 0.13876953125, "grad_norm": 0.2309410721063614, "learning_rate": 0.00048602528441376365, "loss": 1.9025, "step": 1421 }, { "epoch": 0.1388671875, "grad_norm": 0.21750116348266602, "learning_rate": 0.0004860006026437033, "loss": 1.9242, "step": 1422 }, { "epoch": 0.13896484375, "grad_norm": 0.24858328700065613, "learning_rate": 0.0004859758997965519, "loss": 1.9562, "step": 1423 }, { "epoch": 0.1390625, "grad_norm": 0.23306035995483398, "learning_rate": 0.00048595117587477724, "loss": 1.9504, "step": 1424 }, { "epoch": 0.13916015625, "grad_norm": 0.24605897068977356, "learning_rate": 0.00048592643088084904, "loss": 1.9502, "step": 1425 }, { "epoch": 0.1392578125, "grad_norm": 0.23687416315078735, "learning_rate": 0.0004859016648172389, "loss": 1.9035, "step": 1426 }, { "epoch": 0.13935546875, "grad_norm": 0.33239808678627014, "learning_rate": 0.00048587687768642104, "loss": 1.9469, "step": 1427 }, { "epoch": 0.139453125, "grad_norm": 0.3443644344806671, "learning_rate": 0.00048585206949087117, "loss": 1.9126, "step": 1428 }, { "epoch": 0.13955078125, "grad_norm": 0.4044901430606842, "learning_rate": 0.0004858272402330676, "loss": 1.928, "step": 1429 }, { "epoch": 0.1396484375, "grad_norm": 0.3553960919380188, "learning_rate": 0.0004858023899154905, "loss": 1.9782, "step": 1430 }, { "epoch": 0.13974609375, "grad_norm": 0.24459035694599152, "learning_rate": 0.00048577751854062207, "loss": 1.9228, "step": 1431 }, { "epoch": 0.13984375, "grad_norm": 0.26586851477622986, "learning_rate": 0.000485752626110947, "loss": 1.9607, "step": 1432 }, { "epoch": 0.13994140625, "grad_norm": 0.2949119210243225, "learning_rate": 0.0004857277126289516, "loss": 1.9418, "step": 1433 }, { "epoch": 0.1400390625, "grad_norm": 0.23309734463691711, "learning_rate": 0.0004857027780971246, "loss": 1.9036, "step": 1434 }, { "epoch": 0.14013671875, "grad_norm": 0.27574148774147034, "learning_rate": 0.00048567782251795674, "loss": 1.883, "step": 1435 }, { "epoch": 0.140234375, "grad_norm": 0.2791823446750641, "learning_rate": 0.0004856528458939409, "loss": 1.9021, "step": 1436 }, { "epoch": 0.14033203125, "grad_norm": 0.3588283061981201, "learning_rate": 0.00048562784822757187, "loss": 1.9417, "step": 1437 }, { "epoch": 0.1404296875, "grad_norm": 0.3069411814212799, "learning_rate": 0.00048560282952134687, "loss": 1.9055, "step": 1438 }, { "epoch": 0.14052734375, "grad_norm": 0.3290352523326874, "learning_rate": 0.00048557778977776495, "loss": 1.9674, "step": 1439 }, { "epoch": 0.140625, "grad_norm": 0.21290095150470734, "learning_rate": 0.0004855527289993274, "loss": 1.9099, "step": 1440 }, { "epoch": 0.14072265625, "grad_norm": 0.28244757652282715, "learning_rate": 0.00048552764718853755, "loss": 1.9821, "step": 1441 }, { "epoch": 0.1408203125, "grad_norm": 0.2707923948764801, "learning_rate": 0.0004855025443479008, "loss": 1.92, "step": 1442 }, { "epoch": 0.14091796875, "grad_norm": 0.2617279589176178, "learning_rate": 0.0004854774204799248, "loss": 1.9425, "step": 1443 }, { "epoch": 0.141015625, "grad_norm": 0.2797471880912781, "learning_rate": 0.00048545227558711914, "loss": 1.9239, "step": 1444 }, { "epoch": 0.14111328125, "grad_norm": 0.21120323240756989, "learning_rate": 0.00048542710967199563, "loss": 1.9323, "step": 1445 }, { "epoch": 0.1412109375, "grad_norm": 0.23532937467098236, "learning_rate": 0.00048540192273706807, "loss": 1.9033, "step": 1446 }, { "epoch": 0.14130859375, "grad_norm": 0.26372474431991577, "learning_rate": 0.00048537671478485233, "loss": 1.9109, "step": 1447 }, { "epoch": 0.14140625, "grad_norm": 0.2063819169998169, "learning_rate": 0.0004853514858178667, "loss": 1.8868, "step": 1448 }, { "epoch": 0.14150390625, "grad_norm": 0.2495461255311966, "learning_rate": 0.0004853262358386311, "loss": 1.9261, "step": 1449 }, { "epoch": 0.1416015625, "grad_norm": 0.24477644264698029, "learning_rate": 0.00048530096484966786, "loss": 1.8921, "step": 1450 }, { "epoch": 0.14169921875, "grad_norm": 0.20758818089962006, "learning_rate": 0.0004852756728535013, "loss": 1.9274, "step": 1451 }, { "epoch": 0.141796875, "grad_norm": 0.24920731782913208, "learning_rate": 0.000485250359852658, "loss": 1.9203, "step": 1452 }, { "epoch": 0.14189453125, "grad_norm": 0.27062875032424927, "learning_rate": 0.00048522502584966626, "loss": 1.9225, "step": 1453 }, { "epoch": 0.1419921875, "grad_norm": 0.26489493250846863, "learning_rate": 0.00048519967084705696, "loss": 1.9797, "step": 1454 }, { "epoch": 0.14208984375, "grad_norm": 0.2739894688129425, "learning_rate": 0.0004851742948473627, "loss": 1.8824, "step": 1455 }, { "epoch": 0.1421875, "grad_norm": 0.24199414253234863, "learning_rate": 0.0004851488978531184, "loss": 1.9439, "step": 1456 }, { "epoch": 0.14228515625, "grad_norm": 0.21669942140579224, "learning_rate": 0.00048512347986686094, "loss": 1.8901, "step": 1457 }, { "epoch": 0.1423828125, "grad_norm": 0.27475011348724365, "learning_rate": 0.0004850980408911294, "loss": 1.877, "step": 1458 }, { "epoch": 0.14248046875, "grad_norm": 0.24100442230701447, "learning_rate": 0.0004850725809284648, "loss": 1.9286, "step": 1459 }, { "epoch": 0.142578125, "grad_norm": 0.24343852698802948, "learning_rate": 0.0004850470999814105, "loss": 1.9503, "step": 1460 }, { "epoch": 0.14267578125, "grad_norm": 0.22666993737220764, "learning_rate": 0.0004850215980525118, "loss": 1.8722, "step": 1461 }, { "epoch": 0.1427734375, "grad_norm": 0.22877848148345947, "learning_rate": 0.00048499607514431597, "loss": 1.8726, "step": 1462 }, { "epoch": 0.14287109375, "grad_norm": 0.21896860003471375, "learning_rate": 0.0004849705312593727, "loss": 1.9109, "step": 1463 }, { "epoch": 0.14296875, "grad_norm": 0.2896305024623871, "learning_rate": 0.00048494496640023356, "loss": 1.9293, "step": 1464 }, { "epoch": 0.14306640625, "grad_norm": 0.2690875232219696, "learning_rate": 0.00048491938056945214, "loss": 1.9191, "step": 1465 }, { "epoch": 0.1431640625, "grad_norm": 0.27440106868743896, "learning_rate": 0.0004848937737695844, "loss": 1.9729, "step": 1466 }, { "epoch": 0.14326171875, "grad_norm": 0.26410380005836487, "learning_rate": 0.000484868146003188, "loss": 1.8782, "step": 1467 }, { "epoch": 0.143359375, "grad_norm": 0.2948254644870758, "learning_rate": 0.00048484249727282326, "loss": 1.9726, "step": 1468 }, { "epoch": 0.14345703125, "grad_norm": 0.34213653206825256, "learning_rate": 0.000484816827581052, "loss": 1.8893, "step": 1469 }, { "epoch": 0.1435546875, "grad_norm": 0.2864801287651062, "learning_rate": 0.0004847911369304385, "loss": 1.9717, "step": 1470 }, { "epoch": 0.14365234375, "grad_norm": 0.23252104222774506, "learning_rate": 0.00048476542532354896, "loss": 1.9256, "step": 1471 }, { "epoch": 0.14375, "grad_norm": 0.24983392655849457, "learning_rate": 0.00048473969276295175, "loss": 1.9275, "step": 1472 }, { "epoch": 0.14384765625, "grad_norm": 0.24153786897659302, "learning_rate": 0.0004847139392512175, "loss": 1.8821, "step": 1473 }, { "epoch": 0.1439453125, "grad_norm": 0.2148951143026352, "learning_rate": 0.00048468816479091853, "loss": 1.9125, "step": 1474 }, { "epoch": 0.14404296875, "grad_norm": 0.20793981850147247, "learning_rate": 0.00048466236938462956, "loss": 1.9336, "step": 1475 }, { "epoch": 0.144140625, "grad_norm": 0.22007182240486145, "learning_rate": 0.0004846365530349273, "loss": 1.9383, "step": 1476 }, { "epoch": 0.14423828125, "grad_norm": 0.21164564788341522, "learning_rate": 0.00048461071574439063, "loss": 1.9188, "step": 1477 }, { "epoch": 0.1443359375, "grad_norm": 0.26241186261177063, "learning_rate": 0.00048458485751560045, "loss": 1.9348, "step": 1478 }, { "epoch": 0.14443359375, "grad_norm": 0.19936083257198334, "learning_rate": 0.00048455897835113974, "loss": 1.8605, "step": 1479 }, { "epoch": 0.14453125, "grad_norm": 0.2293226420879364, "learning_rate": 0.0004845330782535937, "loss": 1.8666, "step": 1480 }, { "epoch": 0.14462890625, "grad_norm": 0.27344003319740295, "learning_rate": 0.0004845071572255494, "loss": 1.9281, "step": 1481 }, { "epoch": 0.1447265625, "grad_norm": 0.2886092960834503, "learning_rate": 0.0004844812152695961, "loss": 1.9106, "step": 1482 }, { "epoch": 0.14482421875, "grad_norm": 0.345197856426239, "learning_rate": 0.00048445525238832526, "loss": 1.8825, "step": 1483 }, { "epoch": 0.144921875, "grad_norm": 0.32034623622894287, "learning_rate": 0.0004844292685843303, "loss": 1.9103, "step": 1484 }, { "epoch": 0.14501953125, "grad_norm": 0.30518102645874023, "learning_rate": 0.00048440326386020685, "loss": 1.9614, "step": 1485 }, { "epoch": 0.1451171875, "grad_norm": 0.28948330879211426, "learning_rate": 0.00048437723821855244, "loss": 1.8966, "step": 1486 }, { "epoch": 0.14521484375, "grad_norm": 0.28401121497154236, "learning_rate": 0.00048435119166196684, "loss": 1.9455, "step": 1487 }, { "epoch": 0.1453125, "grad_norm": 0.2352132648229599, "learning_rate": 0.0004843251241930519, "loss": 1.9285, "step": 1488 }, { "epoch": 0.14541015625, "grad_norm": 0.231564462184906, "learning_rate": 0.0004842990358144115, "loss": 1.9204, "step": 1489 }, { "epoch": 0.1455078125, "grad_norm": 0.2393004596233368, "learning_rate": 0.00048427292652865165, "loss": 1.9449, "step": 1490 }, { "epoch": 0.14560546875, "grad_norm": 0.2513711154460907, "learning_rate": 0.00048424679633838046, "loss": 1.9092, "step": 1491 }, { "epoch": 0.145703125, "grad_norm": 0.22026249766349792, "learning_rate": 0.000484220645246208, "loss": 1.9439, "step": 1492 }, { "epoch": 0.14580078125, "grad_norm": 0.29993200302124023, "learning_rate": 0.0004841944732547467, "loss": 1.8934, "step": 1493 }, { "epoch": 0.1458984375, "grad_norm": 0.27871406078338623, "learning_rate": 0.0004841682803666108, "loss": 1.8947, "step": 1494 }, { "epoch": 0.14599609375, "grad_norm": 0.21927416324615479, "learning_rate": 0.0004841420665844166, "loss": 1.9472, "step": 1495 }, { "epoch": 0.14609375, "grad_norm": 0.24884743988513947, "learning_rate": 0.0004841158319107829, "loss": 1.9099, "step": 1496 }, { "epoch": 0.14619140625, "grad_norm": 0.256428986787796, "learning_rate": 0.0004840895763483302, "loss": 1.9266, "step": 1497 }, { "epoch": 0.1462890625, "grad_norm": 0.2777554392814636, "learning_rate": 0.0004840632998996812, "loss": 1.9764, "step": 1498 }, { "epoch": 0.14638671875, "grad_norm": 0.3199704587459564, "learning_rate": 0.00048403700256746066, "loss": 1.9391, "step": 1499 }, { "epoch": 0.146484375, "grad_norm": 0.33837810158729553, "learning_rate": 0.00048401068435429545, "loss": 1.9131, "step": 1500 }, { "epoch": 0.14658203125, "grad_norm": 0.25376418232917786, "learning_rate": 0.00048398434526281456, "loss": 1.9391, "step": 1501 }, { "epoch": 0.1466796875, "grad_norm": 0.2878507673740387, "learning_rate": 0.00048395798529564897, "loss": 1.9226, "step": 1502 }, { "epoch": 0.14677734375, "grad_norm": 0.3166121244430542, "learning_rate": 0.0004839316044554319, "loss": 1.9024, "step": 1503 }, { "epoch": 0.146875, "grad_norm": 0.30799803137779236, "learning_rate": 0.00048390520274479857, "loss": 1.9076, "step": 1504 }, { "epoch": 0.14697265625, "grad_norm": 0.3012866675853729, "learning_rate": 0.0004838787801663861, "loss": 1.9282, "step": 1505 }, { "epoch": 0.1470703125, "grad_norm": 0.29956120252609253, "learning_rate": 0.000483852336722834, "loss": 1.8934, "step": 1506 }, { "epoch": 0.14716796875, "grad_norm": 0.2174024134874344, "learning_rate": 0.0004838258724167838, "loss": 1.9434, "step": 1507 }, { "epoch": 0.147265625, "grad_norm": 0.31407755613327026, "learning_rate": 0.0004837993872508789, "loss": 1.8654, "step": 1508 }, { "epoch": 0.14736328125, "grad_norm": 0.22638960182666779, "learning_rate": 0.00048377288122776506, "loss": 1.937, "step": 1509 }, { "epoch": 0.1474609375, "grad_norm": 0.26045146584510803, "learning_rate": 0.0004837463543500899, "loss": 1.8986, "step": 1510 }, { "epoch": 0.14755859375, "grad_norm": 0.2935525178909302, "learning_rate": 0.0004837198066205034, "loss": 1.9554, "step": 1511 }, { "epoch": 0.14765625, "grad_norm": 0.20349502563476562, "learning_rate": 0.0004836932380416571, "loss": 1.893, "step": 1512 }, { "epoch": 0.14775390625, "grad_norm": 0.28056201338768005, "learning_rate": 0.0004836666486162053, "loss": 1.983, "step": 1513 }, { "epoch": 0.1478515625, "grad_norm": 0.27571672201156616, "learning_rate": 0.00048364003834680395, "loss": 1.9399, "step": 1514 }, { "epoch": 0.14794921875, "grad_norm": 0.28008097410202026, "learning_rate": 0.0004836134072361111, "loss": 1.9549, "step": 1515 }, { "epoch": 0.148046875, "grad_norm": 0.2211456298828125, "learning_rate": 0.00048358675528678704, "loss": 1.9245, "step": 1516 }, { "epoch": 0.14814453125, "grad_norm": 0.259030818939209, "learning_rate": 0.00048356008250149404, "loss": 1.8919, "step": 1517 }, { "epoch": 0.1482421875, "grad_norm": 0.2772839665412903, "learning_rate": 0.00048353338888289645, "loss": 1.9386, "step": 1518 }, { "epoch": 0.14833984375, "grad_norm": 0.3579663336277008, "learning_rate": 0.00048350667443366075, "loss": 1.9375, "step": 1519 }, { "epoch": 0.1484375, "grad_norm": 0.37625831365585327, "learning_rate": 0.0004834799391564555, "loss": 1.9508, "step": 1520 }, { "epoch": 0.14853515625, "grad_norm": 0.23583106696605682, "learning_rate": 0.0004834531830539512, "loss": 1.8934, "step": 1521 }, { "epoch": 0.1486328125, "grad_norm": 0.2464006394147873, "learning_rate": 0.00048342640612882075, "loss": 1.8932, "step": 1522 }, { "epoch": 0.14873046875, "grad_norm": 0.30233830213546753, "learning_rate": 0.0004833996083837388, "loss": 1.9117, "step": 1523 }, { "epoch": 0.148828125, "grad_norm": 0.23098327219486237, "learning_rate": 0.0004833727898213822, "loss": 1.9229, "step": 1524 }, { "epoch": 0.14892578125, "grad_norm": 0.2418309450149536, "learning_rate": 0.00048334595044442996, "loss": 1.9369, "step": 1525 }, { "epoch": 0.1490234375, "grad_norm": 0.2360529899597168, "learning_rate": 0.000483319090255563, "loss": 1.8828, "step": 1526 }, { "epoch": 0.14912109375, "grad_norm": 0.24173954129219055, "learning_rate": 0.0004832922092574645, "loss": 1.9349, "step": 1527 }, { "epoch": 0.14921875, "grad_norm": 0.21543891727924347, "learning_rate": 0.0004832653074528196, "loss": 1.9317, "step": 1528 }, { "epoch": 0.14931640625, "grad_norm": 0.21203261613845825, "learning_rate": 0.0004832383848443155, "loss": 1.9261, "step": 1529 }, { "epoch": 0.1494140625, "grad_norm": 0.19579768180847168, "learning_rate": 0.0004832114414346417, "loss": 1.878, "step": 1530 }, { "epoch": 0.14951171875, "grad_norm": 0.23105336725711823, "learning_rate": 0.0004831844772264894, "loss": 1.8957, "step": 1531 }, { "epoch": 0.149609375, "grad_norm": 0.1797371655702591, "learning_rate": 0.0004831574922225521, "loss": 1.9036, "step": 1532 }, { "epoch": 0.14970703125, "grad_norm": 0.22427886724472046, "learning_rate": 0.00048313048642552553, "loss": 1.9143, "step": 1533 }, { "epoch": 0.1498046875, "grad_norm": 0.2837050259113312, "learning_rate": 0.00048310345983810723, "loss": 1.9411, "step": 1534 }, { "epoch": 0.14990234375, "grad_norm": 0.25001662969589233, "learning_rate": 0.0004830764124629969, "loss": 1.9157, "step": 1535 }, { "epoch": 0.15, "grad_norm": 0.21143823862075806, "learning_rate": 0.00048304934430289635, "loss": 1.9654, "step": 1536 }, { "epoch": 0.15009765625, "grad_norm": 0.27027738094329834, "learning_rate": 0.00048302225536050954, "loss": 1.8961, "step": 1537 }, { "epoch": 0.1501953125, "grad_norm": 0.2757401168346405, "learning_rate": 0.00048299514563854225, "loss": 1.9537, "step": 1538 }, { "epoch": 0.15029296875, "grad_norm": 0.30425819754600525, "learning_rate": 0.0004829680151397025, "loss": 1.8713, "step": 1539 }, { "epoch": 0.150390625, "grad_norm": 0.3185950815677643, "learning_rate": 0.0004829408638667006, "loss": 1.987, "step": 1540 }, { "epoch": 0.15048828125, "grad_norm": 0.29616236686706543, "learning_rate": 0.0004829136918222486, "loss": 1.9329, "step": 1541 }, { "epoch": 0.1505859375, "grad_norm": 0.2886502146720886, "learning_rate": 0.00048288649900906066, "loss": 1.9178, "step": 1542 }, { "epoch": 0.15068359375, "grad_norm": 0.27268344163894653, "learning_rate": 0.0004828592854298532, "loss": 1.941, "step": 1543 }, { "epoch": 0.15078125, "grad_norm": 0.27281153202056885, "learning_rate": 0.00048283205108734463, "loss": 1.9668, "step": 1544 }, { "epoch": 0.15087890625, "grad_norm": 0.34445685148239136, "learning_rate": 0.0004828047959842554, "loss": 1.9153, "step": 1545 }, { "epoch": 0.1509765625, "grad_norm": 0.3492426872253418, "learning_rate": 0.0004827775201233079, "loss": 1.9123, "step": 1546 }, { "epoch": 0.15107421875, "grad_norm": 0.23972700536251068, "learning_rate": 0.000482750223507227, "loss": 1.9189, "step": 1547 }, { "epoch": 0.151171875, "grad_norm": 0.28324854373931885, "learning_rate": 0.0004827229061387393, "loss": 1.9113, "step": 1548 }, { "epoch": 0.15126953125, "grad_norm": 0.26321306824684143, "learning_rate": 0.0004826955680205734, "loss": 1.9568, "step": 1549 }, { "epoch": 0.1513671875, "grad_norm": 0.26853737235069275, "learning_rate": 0.0004826682091554604, "loss": 1.8798, "step": 1550 }, { "epoch": 0.15146484375, "grad_norm": 0.32173535227775574, "learning_rate": 0.00048264082954613296, "loss": 1.9564, "step": 1551 }, { "epoch": 0.1515625, "grad_norm": 0.2380555272102356, "learning_rate": 0.0004826134291953263, "loss": 1.883, "step": 1552 }, { "epoch": 0.15166015625, "grad_norm": 0.251919150352478, "learning_rate": 0.0004825860081057773, "loss": 1.8876, "step": 1553 }, { "epoch": 0.1517578125, "grad_norm": 0.26018527150154114, "learning_rate": 0.00048255856628022514, "loss": 1.9115, "step": 1554 }, { "epoch": 0.15185546875, "grad_norm": 0.21712715923786163, "learning_rate": 0.00048253110372141104, "loss": 1.9097, "step": 1555 }, { "epoch": 0.151953125, "grad_norm": 0.2382095605134964, "learning_rate": 0.0004825036204320782, "loss": 1.937, "step": 1556 }, { "epoch": 0.15205078125, "grad_norm": 0.231217160820961, "learning_rate": 0.000482476116414972, "loss": 1.9344, "step": 1557 }, { "epoch": 0.1521484375, "grad_norm": 0.20458881556987762, "learning_rate": 0.0004824485916728398, "loss": 1.9036, "step": 1558 }, { "epoch": 0.15224609375, "grad_norm": 0.205721914768219, "learning_rate": 0.00048242104620843125, "loss": 1.8998, "step": 1559 }, { "epoch": 0.15234375, "grad_norm": 0.23993369936943054, "learning_rate": 0.0004823934800244976, "loss": 1.8706, "step": 1560 }, { "epoch": 0.15244140625, "grad_norm": 0.23396044969558716, "learning_rate": 0.00048236589312379276, "loss": 1.8834, "step": 1561 }, { "epoch": 0.1525390625, "grad_norm": 0.24282585084438324, "learning_rate": 0.0004823382855090723, "loss": 1.8873, "step": 1562 }, { "epoch": 0.15263671875, "grad_norm": 0.25903937220573425, "learning_rate": 0.00048231065718309387, "loss": 1.9357, "step": 1563 }, { "epoch": 0.152734375, "grad_norm": 0.24471600353717804, "learning_rate": 0.0004822830081486175, "loss": 1.8975, "step": 1564 }, { "epoch": 0.15283203125, "grad_norm": 0.21125410497188568, "learning_rate": 0.00048225533840840493, "loss": 1.9513, "step": 1565 }, { "epoch": 0.1529296875, "grad_norm": 0.22737543284893036, "learning_rate": 0.00048222764796522017, "loss": 1.8962, "step": 1566 }, { "epoch": 0.15302734375, "grad_norm": 0.1857408881187439, "learning_rate": 0.0004821999368218292, "loss": 1.9054, "step": 1567 }, { "epoch": 0.153125, "grad_norm": 0.23223114013671875, "learning_rate": 0.0004821722049810002, "loss": 1.9118, "step": 1568 }, { "epoch": 0.15322265625, "grad_norm": 0.25117412209510803, "learning_rate": 0.0004821444524455034, "loss": 1.8889, "step": 1569 }, { "epoch": 0.1533203125, "grad_norm": 0.23292681574821472, "learning_rate": 0.00048211667921811084, "loss": 1.9495, "step": 1570 }, { "epoch": 0.15341796875, "grad_norm": 0.24508163332939148, "learning_rate": 0.0004820888853015969, "loss": 1.922, "step": 1571 }, { "epoch": 0.153515625, "grad_norm": 0.26448142528533936, "learning_rate": 0.000482061070698738, "loss": 1.9811, "step": 1572 }, { "epoch": 0.15361328125, "grad_norm": 0.24274107813835144, "learning_rate": 0.0004820332354123125, "loss": 1.9134, "step": 1573 }, { "epoch": 0.1537109375, "grad_norm": 0.2016661912202835, "learning_rate": 0.00048200537944510097, "loss": 1.9151, "step": 1574 }, { "epoch": 0.15380859375, "grad_norm": 0.21885213255882263, "learning_rate": 0.00048197750279988595, "loss": 1.9613, "step": 1575 }, { "epoch": 0.15390625, "grad_norm": 0.2705588936805725, "learning_rate": 0.000481949605479452, "loss": 1.9535, "step": 1576 }, { "epoch": 0.15400390625, "grad_norm": 0.31828904151916504, "learning_rate": 0.00048192168748658595, "loss": 1.9592, "step": 1577 }, { "epoch": 0.1541015625, "grad_norm": 0.38848698139190674, "learning_rate": 0.0004818937488240764, "loss": 1.9027, "step": 1578 }, { "epoch": 0.15419921875, "grad_norm": 0.2857527732849121, "learning_rate": 0.00048186578949471434, "loss": 1.9888, "step": 1579 }, { "epoch": 0.154296875, "grad_norm": 0.30096855759620667, "learning_rate": 0.0004818378095012925, "loss": 1.9199, "step": 1580 }, { "epoch": 0.15439453125, "grad_norm": 0.3158320486545563, "learning_rate": 0.00048180980884660594, "loss": 1.9081, "step": 1581 }, { "epoch": 0.1544921875, "grad_norm": 0.22765566408634186, "learning_rate": 0.00048178178753345157, "loss": 1.9181, "step": 1582 }, { "epoch": 0.15458984375, "grad_norm": 0.28936147689819336, "learning_rate": 0.0004817537455646286, "loss": 1.9424, "step": 1583 }, { "epoch": 0.1546875, "grad_norm": 0.3088884651660919, "learning_rate": 0.00048172568294293816, "loss": 1.9088, "step": 1584 }, { "epoch": 0.15478515625, "grad_norm": 0.2758021950721741, "learning_rate": 0.00048169759967118343, "loss": 1.9298, "step": 1585 }, { "epoch": 0.1548828125, "grad_norm": 0.2999780774116516, "learning_rate": 0.0004816694957521696, "loss": 1.9335, "step": 1586 }, { "epoch": 0.15498046875, "grad_norm": 0.22048184275627136, "learning_rate": 0.0004816413711887041, "loss": 1.8843, "step": 1587 }, { "epoch": 0.155078125, "grad_norm": 0.26630088686943054, "learning_rate": 0.0004816132259835963, "loss": 1.9073, "step": 1588 }, { "epoch": 0.15517578125, "grad_norm": 0.2615686058998108, "learning_rate": 0.0004815850601396577, "loss": 1.9313, "step": 1589 }, { "epoch": 0.1552734375, "grad_norm": 0.2473047822713852, "learning_rate": 0.00048155687365970167, "loss": 1.8878, "step": 1590 }, { "epoch": 0.15537109375, "grad_norm": 0.3328106999397278, "learning_rate": 0.000481528666546544, "loss": 1.9424, "step": 1591 }, { "epoch": 0.15546875, "grad_norm": 0.26094701886177063, "learning_rate": 0.0004815004388030022, "loss": 1.9623, "step": 1592 }, { "epoch": 0.15556640625, "grad_norm": 0.3564736247062683, "learning_rate": 0.0004814721904318959, "loss": 1.8966, "step": 1593 }, { "epoch": 0.1556640625, "grad_norm": 0.3765439987182617, "learning_rate": 0.000481443921436047, "loss": 1.9189, "step": 1594 }, { "epoch": 0.15576171875, "grad_norm": 0.36775973439216614, "learning_rate": 0.0004814156318182793, "loss": 1.9427, "step": 1595 }, { "epoch": 0.155859375, "grad_norm": 0.3822444975376129, "learning_rate": 0.00048138732158141867, "loss": 1.9329, "step": 1596 }, { "epoch": 0.15595703125, "grad_norm": 0.26008525490760803, "learning_rate": 0.00048135899072829295, "loss": 1.928, "step": 1597 }, { "epoch": 0.1560546875, "grad_norm": 0.32837969064712524, "learning_rate": 0.0004813306392617324, "loss": 1.9409, "step": 1598 }, { "epoch": 0.15615234375, "grad_norm": 0.30874279141426086, "learning_rate": 0.0004813022671845687, "loss": 1.9378, "step": 1599 }, { "epoch": 0.15625, "grad_norm": 0.2988525629043579, "learning_rate": 0.00048127387449963633, "loss": 1.9091, "step": 1600 }, { "epoch": 0.15634765625, "grad_norm": 0.4251779615879059, "learning_rate": 0.0004812454612097713, "loss": 1.9693, "step": 1601 }, { "epoch": 0.1564453125, "grad_norm": 0.2827276885509491, "learning_rate": 0.00048121702731781184, "loss": 1.9177, "step": 1602 }, { "epoch": 0.15654296875, "grad_norm": 0.24983742833137512, "learning_rate": 0.00048118857282659834, "loss": 1.9235, "step": 1603 }, { "epoch": 0.156640625, "grad_norm": 0.25737494230270386, "learning_rate": 0.000481160097738973, "loss": 1.9678, "step": 1604 }, { "epoch": 0.15673828125, "grad_norm": 0.38471952080726624, "learning_rate": 0.00048113160205778035, "loss": 1.9403, "step": 1605 }, { "epoch": 0.1568359375, "grad_norm": 0.30553486943244934, "learning_rate": 0.0004811030857858667, "loss": 1.939, "step": 1606 }, { "epoch": 0.15693359375, "grad_norm": 0.3170030117034912, "learning_rate": 0.0004810745489260808, "loss": 1.9331, "step": 1607 }, { "epoch": 0.15703125, "grad_norm": 0.3244478404521942, "learning_rate": 0.0004810459914812731, "loss": 1.8986, "step": 1608 }, { "epoch": 0.15712890625, "grad_norm": 0.24035608768463135, "learning_rate": 0.0004810174134542963, "loss": 1.9542, "step": 1609 }, { "epoch": 0.1572265625, "grad_norm": 0.29910752177238464, "learning_rate": 0.00048098881484800505, "loss": 1.8521, "step": 1610 }, { "epoch": 0.15732421875, "grad_norm": 0.2832294702529907, "learning_rate": 0.000480960195665256, "loss": 1.9567, "step": 1611 }, { "epoch": 0.157421875, "grad_norm": 0.3093940317630768, "learning_rate": 0.0004809315559089081, "loss": 1.9332, "step": 1612 }, { "epoch": 0.15751953125, "grad_norm": 0.2190847098827362, "learning_rate": 0.00048090289558182215, "loss": 1.8979, "step": 1613 }, { "epoch": 0.1576171875, "grad_norm": 0.2461078017950058, "learning_rate": 0.00048087421468686107, "loss": 1.9032, "step": 1614 }, { "epoch": 0.15771484375, "grad_norm": 0.2852519750595093, "learning_rate": 0.0004808455132268898, "loss": 1.9457, "step": 1615 }, { "epoch": 0.1578125, "grad_norm": 0.23334845900535583, "learning_rate": 0.00048081679120477543, "loss": 1.8718, "step": 1616 }, { "epoch": 0.15791015625, "grad_norm": 0.2714453637599945, "learning_rate": 0.000480788048623387, "loss": 1.9212, "step": 1617 }, { "epoch": 0.1580078125, "grad_norm": 0.2445923238992691, "learning_rate": 0.0004807592854855955, "loss": 1.9322, "step": 1618 }, { "epoch": 0.15810546875, "grad_norm": 0.2522878050804138, "learning_rate": 0.0004807305017942744, "loss": 1.939, "step": 1619 }, { "epoch": 0.158203125, "grad_norm": 0.28525444865226746, "learning_rate": 0.0004807016975522987, "loss": 1.929, "step": 1620 }, { "epoch": 0.15830078125, "grad_norm": 0.22004222869873047, "learning_rate": 0.0004806728727625458, "loss": 1.9431, "step": 1621 }, { "epoch": 0.1583984375, "grad_norm": 0.20370328426361084, "learning_rate": 0.000480644027427895, "loss": 1.9063, "step": 1622 }, { "epoch": 0.15849609375, "grad_norm": 0.21263106167316437, "learning_rate": 0.00048061516155122763, "loss": 1.9098, "step": 1623 }, { "epoch": 0.15859375, "grad_norm": 0.25402888655662537, "learning_rate": 0.0004805862751354273, "loss": 1.8932, "step": 1624 }, { "epoch": 0.15869140625, "grad_norm": 0.21452878415584564, "learning_rate": 0.0004805573681833794, "loss": 1.9674, "step": 1625 }, { "epoch": 0.1587890625, "grad_norm": 0.2378520965576172, "learning_rate": 0.0004805284406979715, "loss": 1.8841, "step": 1626 }, { "epoch": 0.15888671875, "grad_norm": 0.20481383800506592, "learning_rate": 0.0004804994926820932, "loss": 1.8836, "step": 1627 }, { "epoch": 0.158984375, "grad_norm": 0.19195064902305603, "learning_rate": 0.00048047052413863607, "loss": 1.9694, "step": 1628 }, { "epoch": 0.15908203125, "grad_norm": 0.228655144572258, "learning_rate": 0.0004804415350704939, "loss": 1.9323, "step": 1629 }, { "epoch": 0.1591796875, "grad_norm": 0.2156544327735901, "learning_rate": 0.0004804125254805625, "loss": 1.9236, "step": 1630 }, { "epoch": 0.15927734375, "grad_norm": 0.24159860610961914, "learning_rate": 0.00048038349537173953, "loss": 1.9611, "step": 1631 }, { "epoch": 0.159375, "grad_norm": 0.21641957759857178, "learning_rate": 0.0004803544447469249, "loss": 1.9237, "step": 1632 }, { "epoch": 0.15947265625, "grad_norm": 0.21140074729919434, "learning_rate": 0.0004803253736090205, "loss": 1.9297, "step": 1633 }, { "epoch": 0.1595703125, "grad_norm": 0.23588809370994568, "learning_rate": 0.00048029628196093047, "loss": 1.8995, "step": 1634 }, { "epoch": 0.15966796875, "grad_norm": 0.2083786576986313, "learning_rate": 0.00048026716980556044, "loss": 1.9154, "step": 1635 }, { "epoch": 0.159765625, "grad_norm": 0.21067486703395844, "learning_rate": 0.00048023803714581867, "loss": 1.9032, "step": 1636 }, { "epoch": 0.15986328125, "grad_norm": 0.19560249149799347, "learning_rate": 0.0004802088839846153, "loss": 1.9352, "step": 1637 }, { "epoch": 0.1599609375, "grad_norm": 0.21477989852428436, "learning_rate": 0.0004801797103248624, "loss": 1.9192, "step": 1638 }, { "epoch": 0.16005859375, "grad_norm": 0.27302122116088867, "learning_rate": 0.0004801505161694741, "loss": 1.9177, "step": 1639 }, { "epoch": 0.16015625, "grad_norm": 0.251449316740036, "learning_rate": 0.00048012130152136676, "loss": 1.9383, "step": 1640 }, { "epoch": 0.16025390625, "grad_norm": 0.29725855588912964, "learning_rate": 0.00048009206638345857, "loss": 1.9027, "step": 1641 }, { "epoch": 0.1603515625, "grad_norm": 0.3364620804786682, "learning_rate": 0.00048006281075866996, "loss": 1.8999, "step": 1642 }, { "epoch": 0.16044921875, "grad_norm": 0.33240416646003723, "learning_rate": 0.00048003353464992323, "loss": 1.9023, "step": 1643 }, { "epoch": 0.160546875, "grad_norm": 0.29456380009651184, "learning_rate": 0.0004800042380601428, "loss": 1.8874, "step": 1644 }, { "epoch": 0.16064453125, "grad_norm": 0.20352846384048462, "learning_rate": 0.00047997492099225515, "loss": 1.8813, "step": 1645 }, { "epoch": 0.1607421875, "grad_norm": 0.2666397988796234, "learning_rate": 0.0004799455834491889, "loss": 1.9582, "step": 1646 }, { "epoch": 0.16083984375, "grad_norm": 0.34426215291023254, "learning_rate": 0.0004799162254338744, "loss": 1.9372, "step": 1647 }, { "epoch": 0.1609375, "grad_norm": 0.3887472450733185, "learning_rate": 0.00047988684694924447, "loss": 1.9381, "step": 1648 }, { "epoch": 0.16103515625, "grad_norm": 0.3413529694080353, "learning_rate": 0.0004798574479982336, "loss": 1.8966, "step": 1649 }, { "epoch": 0.1611328125, "grad_norm": 0.2743453085422516, "learning_rate": 0.00047982802858377853, "loss": 1.9228, "step": 1650 }, { "epoch": 0.16123046875, "grad_norm": 0.29950737953186035, "learning_rate": 0.00047979858870881805, "loss": 1.9204, "step": 1651 }, { "epoch": 0.161328125, "grad_norm": 0.273075133562088, "learning_rate": 0.00047976912837629295, "loss": 1.8683, "step": 1652 }, { "epoch": 0.16142578125, "grad_norm": 0.2547522485256195, "learning_rate": 0.00047973964758914596, "loss": 1.9275, "step": 1653 }, { "epoch": 0.1615234375, "grad_norm": 0.2716279923915863, "learning_rate": 0.00047971014635032205, "loss": 1.8917, "step": 1654 }, { "epoch": 0.16162109375, "grad_norm": 0.28852027654647827, "learning_rate": 0.00047968062466276803, "loss": 1.9418, "step": 1655 }, { "epoch": 0.16171875, "grad_norm": 0.19788789749145508, "learning_rate": 0.00047965108252943293, "loss": 1.922, "step": 1656 }, { "epoch": 0.16181640625, "grad_norm": 0.33596405386924744, "learning_rate": 0.0004796215199532678, "loss": 1.9291, "step": 1657 }, { "epoch": 0.1619140625, "grad_norm": 0.32864540815353394, "learning_rate": 0.00047959193693722553, "loss": 1.9793, "step": 1658 }, { "epoch": 0.16201171875, "grad_norm": 0.25629329681396484, "learning_rate": 0.0004795623334842613, "loss": 1.9453, "step": 1659 }, { "epoch": 0.162109375, "grad_norm": 0.2877620458602905, "learning_rate": 0.0004795327095973322, "loss": 1.9225, "step": 1660 }, { "epoch": 0.16220703125, "grad_norm": 0.24507132172584534, "learning_rate": 0.00047950306527939744, "loss": 1.8917, "step": 1661 }, { "epoch": 0.1623046875, "grad_norm": 0.2875724732875824, "learning_rate": 0.0004794734005334182, "loss": 1.8819, "step": 1662 }, { "epoch": 0.16240234375, "grad_norm": 0.2515362501144409, "learning_rate": 0.0004794437153623577, "loss": 1.9232, "step": 1663 }, { "epoch": 0.1625, "grad_norm": 0.27037444710731506, "learning_rate": 0.00047941400976918123, "loss": 1.9567, "step": 1664 }, { "epoch": 0.16259765625, "grad_norm": 0.29190242290496826, "learning_rate": 0.0004793842837568562, "loss": 1.9266, "step": 1665 }, { "epoch": 0.1626953125, "grad_norm": 0.26770535111427307, "learning_rate": 0.0004793545373283518, "loss": 1.9163, "step": 1666 }, { "epoch": 0.16279296875, "grad_norm": 0.25330448150634766, "learning_rate": 0.0004793247704866396, "loss": 1.9182, "step": 1667 }, { "epoch": 0.162890625, "grad_norm": 0.27300313115119934, "learning_rate": 0.00047929498323469295, "loss": 1.9592, "step": 1668 }, { "epoch": 0.16298828125, "grad_norm": 0.21413999795913696, "learning_rate": 0.00047926517557548746, "loss": 1.919, "step": 1669 }, { "epoch": 0.1630859375, "grad_norm": 0.27660825848579407, "learning_rate": 0.0004792353475120004, "loss": 1.9176, "step": 1670 }, { "epoch": 0.16318359375, "grad_norm": 0.2682628631591797, "learning_rate": 0.00047920549904721156, "loss": 1.9381, "step": 1671 }, { "epoch": 0.16328125, "grad_norm": 0.2719988226890564, "learning_rate": 0.00047917563018410247, "loss": 1.9137, "step": 1672 }, { "epoch": 0.16337890625, "grad_norm": 0.23097728192806244, "learning_rate": 0.0004791457409256568, "loss": 1.9128, "step": 1673 }, { "epoch": 0.1634765625, "grad_norm": 0.27505284547805786, "learning_rate": 0.00047911583127486015, "loss": 1.9026, "step": 1674 }, { "epoch": 0.16357421875, "grad_norm": 0.31115567684173584, "learning_rate": 0.0004790859012347002, "loss": 1.9438, "step": 1675 }, { "epoch": 0.163671875, "grad_norm": 0.2766979932785034, "learning_rate": 0.0004790559508081668, "loss": 1.9389, "step": 1676 }, { "epoch": 0.16376953125, "grad_norm": 0.25364476442337036, "learning_rate": 0.0004790259799982517, "loss": 1.9621, "step": 1677 }, { "epoch": 0.1638671875, "grad_norm": 0.2380513697862625, "learning_rate": 0.0004789959888079487, "loss": 1.8891, "step": 1678 }, { "epoch": 0.16396484375, "grad_norm": 0.21375243365764618, "learning_rate": 0.0004789659772402536, "loss": 1.8864, "step": 1679 }, { "epoch": 0.1640625, "grad_norm": 0.2307070791721344, "learning_rate": 0.00047893594529816445, "loss": 1.8975, "step": 1680 }, { "epoch": 0.16416015625, "grad_norm": 0.20362551510334015, "learning_rate": 0.000478905892984681, "loss": 1.9489, "step": 1681 }, { "epoch": 0.1642578125, "grad_norm": 0.23438306152820587, "learning_rate": 0.00047887582030280536, "loss": 1.9874, "step": 1682 }, { "epoch": 0.16435546875, "grad_norm": 0.19733543694019318, "learning_rate": 0.00047884572725554137, "loss": 1.9147, "step": 1683 }, { "epoch": 0.164453125, "grad_norm": 0.2748172879219055, "learning_rate": 0.0004788156138458952, "loss": 1.863, "step": 1684 }, { "epoch": 0.16455078125, "grad_norm": 0.26911303400993347, "learning_rate": 0.00047878548007687476, "loss": 1.8852, "step": 1685 }, { "epoch": 0.1646484375, "grad_norm": 0.20826198160648346, "learning_rate": 0.0004787553259514903, "loss": 1.933, "step": 1686 }, { "epoch": 0.16474609375, "grad_norm": 0.3136061429977417, "learning_rate": 0.00047872515147275393, "loss": 1.9264, "step": 1687 }, { "epoch": 0.16484375, "grad_norm": 0.22931191325187683, "learning_rate": 0.0004786949566436797, "loss": 1.9175, "step": 1688 }, { "epoch": 0.16494140625, "grad_norm": 0.24129685759544373, "learning_rate": 0.0004786647414672839, "loss": 1.9114, "step": 1689 }, { "epoch": 0.1650390625, "grad_norm": 0.2148384302854538, "learning_rate": 0.00047863450594658473, "loss": 1.9189, "step": 1690 }, { "epoch": 0.16513671875, "grad_norm": 0.24560751020908356, "learning_rate": 0.0004786042500846025, "loss": 1.9459, "step": 1691 }, { "epoch": 0.165234375, "grad_norm": 0.2328975796699524, "learning_rate": 0.0004785739738843594, "loss": 1.9037, "step": 1692 }, { "epoch": 0.16533203125, "grad_norm": 0.19054313004016876, "learning_rate": 0.00047854367734887984, "loss": 1.9035, "step": 1693 }, { "epoch": 0.1654296875, "grad_norm": 0.19610174000263214, "learning_rate": 0.00047851336048119016, "loss": 1.8709, "step": 1694 }, { "epoch": 0.16552734375, "grad_norm": 0.22446346282958984, "learning_rate": 0.0004784830232843186, "loss": 1.8866, "step": 1695 }, { "epoch": 0.165625, "grad_norm": 0.22055386006832123, "learning_rate": 0.0004784526657612959, "loss": 1.9104, "step": 1696 }, { "epoch": 0.16572265625, "grad_norm": 0.19776830077171326, "learning_rate": 0.0004784222879151542, "loss": 1.9302, "step": 1697 }, { "epoch": 0.1658203125, "grad_norm": 0.19348515570163727, "learning_rate": 0.0004783918897489282, "loss": 1.944, "step": 1698 }, { "epoch": 0.16591796875, "grad_norm": 0.21035566926002502, "learning_rate": 0.0004783614712656542, "loss": 1.9697, "step": 1699 }, { "epoch": 0.166015625, "grad_norm": 0.20167845487594604, "learning_rate": 0.0004783310324683709, "loss": 1.9087, "step": 1700 }, { "epoch": 0.16611328125, "grad_norm": 0.24482861161231995, "learning_rate": 0.0004783005733601188, "loss": 1.9316, "step": 1701 }, { "epoch": 0.1662109375, "grad_norm": 0.2270205020904541, "learning_rate": 0.00047827009394394045, "loss": 1.885, "step": 1702 }, { "epoch": 0.16630859375, "grad_norm": 0.2608849108219147, "learning_rate": 0.0004782395942228806, "loss": 1.9639, "step": 1703 }, { "epoch": 0.16640625, "grad_norm": 0.29610663652420044, "learning_rate": 0.0004782090741999858, "loss": 1.9248, "step": 1704 }, { "epoch": 0.16650390625, "grad_norm": 0.44481170177459717, "learning_rate": 0.0004781785338783047, "loss": 1.9352, "step": 1705 }, { "epoch": 0.1666015625, "grad_norm": 0.4372797906398773, "learning_rate": 0.0004781479732608881, "loss": 1.9557, "step": 1706 }, { "epoch": 0.16669921875, "grad_norm": 0.2859649956226349, "learning_rate": 0.00047811739235078876, "loss": 1.9753, "step": 1707 }, { "epoch": 0.166796875, "grad_norm": 0.26134198904037476, "learning_rate": 0.00047808679115106135, "loss": 1.9513, "step": 1708 }, { "epoch": 0.16689453125, "grad_norm": 0.30859240889549255, "learning_rate": 0.0004780561696647627, "loss": 1.9084, "step": 1709 }, { "epoch": 0.1669921875, "grad_norm": 0.33629563450813293, "learning_rate": 0.0004780255278949516, "loss": 1.928, "step": 1710 }, { "epoch": 0.16708984375, "grad_norm": 0.2672228217124939, "learning_rate": 0.0004779948658446889, "loss": 1.9033, "step": 1711 }, { "epoch": 0.1671875, "grad_norm": 0.2845448851585388, "learning_rate": 0.00047796418351703744, "loss": 1.8892, "step": 1712 }, { "epoch": 0.16728515625, "grad_norm": 0.236465185880661, "learning_rate": 0.0004779334809150622, "loss": 1.9363, "step": 1713 }, { "epoch": 0.1673828125, "grad_norm": 0.2997010052204132, "learning_rate": 0.00047790275804183004, "loss": 1.8917, "step": 1714 }, { "epoch": 0.16748046875, "grad_norm": 0.3848778307437897, "learning_rate": 0.00047787201490040985, "loss": 1.959, "step": 1715 }, { "epoch": 0.167578125, "grad_norm": 0.2715388834476471, "learning_rate": 0.00047784125149387277, "loss": 1.9463, "step": 1716 }, { "epoch": 0.16767578125, "grad_norm": 0.3423638343811035, "learning_rate": 0.0004778104678252916, "loss": 1.9393, "step": 1717 }, { "epoch": 0.1677734375, "grad_norm": 0.3216153383255005, "learning_rate": 0.0004777796638977414, "loss": 1.9688, "step": 1718 }, { "epoch": 0.16787109375, "grad_norm": 0.24108828604221344, "learning_rate": 0.00047774883971429935, "loss": 1.8822, "step": 1719 }, { "epoch": 0.16796875, "grad_norm": 0.3516473174095154, "learning_rate": 0.0004777179952780443, "loss": 1.912, "step": 1720 }, { "epoch": 0.16806640625, "grad_norm": 0.2251645028591156, "learning_rate": 0.0004776871305920575, "loss": 1.8989, "step": 1721 }, { "epoch": 0.1681640625, "grad_norm": 0.282797634601593, "learning_rate": 0.000477656245659422, "loss": 1.8483, "step": 1722 }, { "epoch": 0.16826171875, "grad_norm": 0.3045579195022583, "learning_rate": 0.0004776253404832229, "loss": 1.8934, "step": 1723 }, { "epoch": 0.168359375, "grad_norm": 0.27416738867759705, "learning_rate": 0.00047759441506654745, "loss": 1.924, "step": 1724 }, { "epoch": 0.16845703125, "grad_norm": 0.2600431442260742, "learning_rate": 0.0004775634694124847, "loss": 1.9507, "step": 1725 }, { "epoch": 0.1685546875, "grad_norm": 0.22953875362873077, "learning_rate": 0.0004775325035241259, "loss": 1.9062, "step": 1726 }, { "epoch": 0.16865234375, "grad_norm": 0.20656803250312805, "learning_rate": 0.00047750151740456426, "loss": 1.9312, "step": 1727 }, { "epoch": 0.16875, "grad_norm": 0.21857194602489471, "learning_rate": 0.0004774705110568951, "loss": 1.931, "step": 1728 }, { "epoch": 0.16884765625, "grad_norm": 0.2585659325122833, "learning_rate": 0.0004774394844842155, "loss": 1.9149, "step": 1729 }, { "epoch": 0.1689453125, "grad_norm": 0.23686473071575165, "learning_rate": 0.000477408437689625, "loss": 1.9316, "step": 1730 }, { "epoch": 0.16904296875, "grad_norm": 0.21281464397907257, "learning_rate": 0.0004773773706762247, "loss": 1.9278, "step": 1731 }, { "epoch": 0.169140625, "grad_norm": 0.2731045186519623, "learning_rate": 0.00047734628344711796, "loss": 1.9088, "step": 1732 }, { "epoch": 0.16923828125, "grad_norm": 0.21235254406929016, "learning_rate": 0.00047731517600541014, "loss": 1.9418, "step": 1733 }, { "epoch": 0.1693359375, "grad_norm": 0.2536609172821045, "learning_rate": 0.0004772840483542085, "loss": 1.9651, "step": 1734 }, { "epoch": 0.16943359375, "grad_norm": 0.22243930399417877, "learning_rate": 0.00047725290049662264, "loss": 1.9158, "step": 1735 }, { "epoch": 0.16953125, "grad_norm": 0.24414226412773132, "learning_rate": 0.00047722173243576374, "loss": 1.9167, "step": 1736 }, { "epoch": 0.16962890625, "grad_norm": 0.25493788719177246, "learning_rate": 0.0004771905441747453, "loss": 1.9244, "step": 1737 }, { "epoch": 0.1697265625, "grad_norm": 0.21703946590423584, "learning_rate": 0.0004771593357166827, "loss": 1.8882, "step": 1738 }, { "epoch": 0.16982421875, "grad_norm": 0.27353623509407043, "learning_rate": 0.00047712810706469354, "loss": 1.9134, "step": 1739 }, { "epoch": 0.169921875, "grad_norm": 0.2163049578666687, "learning_rate": 0.0004770968582218971, "loss": 1.9226, "step": 1740 }, { "epoch": 0.17001953125, "grad_norm": 0.1948620080947876, "learning_rate": 0.000477065589191415, "loss": 1.8663, "step": 1741 }, { "epoch": 0.1701171875, "grad_norm": 0.22654829919338226, "learning_rate": 0.0004770342999763706, "loss": 1.9327, "step": 1742 }, { "epoch": 0.17021484375, "grad_norm": 0.24024905264377594, "learning_rate": 0.00047700299057988957, "loss": 1.9581, "step": 1743 }, { "epoch": 0.1703125, "grad_norm": 0.2727830111980438, "learning_rate": 0.0004769716610050993, "loss": 1.9158, "step": 1744 }, { "epoch": 0.17041015625, "grad_norm": 0.2848643362522125, "learning_rate": 0.0004769403112551295, "loss": 1.8943, "step": 1745 }, { "epoch": 0.1705078125, "grad_norm": 0.29621654748916626, "learning_rate": 0.00047690894133311157, "loss": 1.9182, "step": 1746 }, { "epoch": 0.17060546875, "grad_norm": 0.317327618598938, "learning_rate": 0.00047687755124217914, "loss": 1.9624, "step": 1747 }, { "epoch": 0.170703125, "grad_norm": 0.3084838390350342, "learning_rate": 0.00047684614098546783, "loss": 1.919, "step": 1748 }, { "epoch": 0.17080078125, "grad_norm": 0.26487037539482117, "learning_rate": 0.00047681471056611527, "loss": 1.9005, "step": 1749 }, { "epoch": 0.1708984375, "grad_norm": 0.2641773223876953, "learning_rate": 0.000476783259987261, "loss": 1.9085, "step": 1750 }, { "epoch": 0.17099609375, "grad_norm": 0.30503541231155396, "learning_rate": 0.00047675178925204667, "loss": 1.9148, "step": 1751 }, { "epoch": 0.17109375, "grad_norm": 0.2899084985256195, "learning_rate": 0.00047672029836361596, "loss": 1.9369, "step": 1752 }, { "epoch": 0.17119140625, "grad_norm": 0.3330906629562378, "learning_rate": 0.00047668878732511457, "loss": 1.8529, "step": 1753 }, { "epoch": 0.1712890625, "grad_norm": 0.32898756861686707, "learning_rate": 0.00047665725613969015, "loss": 1.8918, "step": 1754 }, { "epoch": 0.17138671875, "grad_norm": 0.21951165795326233, "learning_rate": 0.0004766257048104923, "loss": 1.8868, "step": 1755 }, { "epoch": 0.171484375, "grad_norm": 0.20533274114131927, "learning_rate": 0.00047659413334067276, "loss": 1.8993, "step": 1756 }, { "epoch": 0.17158203125, "grad_norm": 0.19753561913967133, "learning_rate": 0.0004765625417333853, "loss": 1.913, "step": 1757 }, { "epoch": 0.1716796875, "grad_norm": 0.20068642497062683, "learning_rate": 0.0004765309299917856, "loss": 1.8639, "step": 1758 }, { "epoch": 0.17177734375, "grad_norm": 0.22603540122509003, "learning_rate": 0.00047649929811903143, "loss": 1.9014, "step": 1759 }, { "epoch": 0.171875, "grad_norm": 0.18157346546649933, "learning_rate": 0.00047646764611828247, "loss": 1.8956, "step": 1760 }, { "epoch": 0.17197265625, "grad_norm": 0.21239453554153442, "learning_rate": 0.0004764359739927005, "loss": 1.9636, "step": 1761 }, { "epoch": 0.1720703125, "grad_norm": 0.25655651092529297, "learning_rate": 0.0004764042817454494, "loss": 1.8901, "step": 1762 }, { "epoch": 0.17216796875, "grad_norm": 0.20679731667041779, "learning_rate": 0.0004763725693796947, "loss": 1.8908, "step": 1763 }, { "epoch": 0.172265625, "grad_norm": 0.20754674077033997, "learning_rate": 0.0004763408368986043, "loss": 1.9445, "step": 1764 }, { "epoch": 0.17236328125, "grad_norm": 0.18912455439567566, "learning_rate": 0.0004763090843053481, "loss": 1.9176, "step": 1765 }, { "epoch": 0.1724609375, "grad_norm": 0.20957696437835693, "learning_rate": 0.0004762773116030978, "loss": 1.9206, "step": 1766 }, { "epoch": 0.17255859375, "grad_norm": 0.1931106299161911, "learning_rate": 0.0004762455187950272, "loss": 1.9051, "step": 1767 }, { "epoch": 0.17265625, "grad_norm": 0.218103289604187, "learning_rate": 0.00047621370588431217, "loss": 1.9326, "step": 1768 }, { "epoch": 0.17275390625, "grad_norm": 0.20649409294128418, "learning_rate": 0.0004761818728741305, "loss": 1.9156, "step": 1769 }, { "epoch": 0.1728515625, "grad_norm": 0.2224937528371811, "learning_rate": 0.000476150019767662, "loss": 1.9181, "step": 1770 }, { "epoch": 0.17294921875, "grad_norm": 0.30575814843177795, "learning_rate": 0.00047611814656808865, "loss": 1.9276, "step": 1771 }, { "epoch": 0.173046875, "grad_norm": 0.34954264760017395, "learning_rate": 0.0004760862532785942, "loss": 1.8818, "step": 1772 }, { "epoch": 0.17314453125, "grad_norm": 0.2726197838783264, "learning_rate": 0.0004760543399023644, "loss": 1.9279, "step": 1773 }, { "epoch": 0.1732421875, "grad_norm": 0.2664092779159546, "learning_rate": 0.00047602240644258726, "loss": 1.9055, "step": 1774 }, { "epoch": 0.17333984375, "grad_norm": 0.22661474347114563, "learning_rate": 0.00047599045290245277, "loss": 1.9406, "step": 1775 }, { "epoch": 0.1734375, "grad_norm": 0.20825809240341187, "learning_rate": 0.00047595847928515243, "loss": 1.9189, "step": 1776 }, { "epoch": 0.17353515625, "grad_norm": 0.200863778591156, "learning_rate": 0.0004759264855938805, "loss": 1.8985, "step": 1777 }, { "epoch": 0.1736328125, "grad_norm": 0.18549345433712006, "learning_rate": 0.00047589447183183254, "loss": 1.8855, "step": 1778 }, { "epoch": 0.17373046875, "grad_norm": 0.19504952430725098, "learning_rate": 0.00047586243800220673, "loss": 1.9494, "step": 1779 }, { "epoch": 0.173828125, "grad_norm": 0.2109365612268448, "learning_rate": 0.0004758303841082029, "loss": 1.9223, "step": 1780 }, { "epoch": 0.17392578125, "grad_norm": 0.20636534690856934, "learning_rate": 0.00047579831015302277, "loss": 1.8939, "step": 1781 }, { "epoch": 0.1740234375, "grad_norm": 0.2035193145275116, "learning_rate": 0.0004757662161398704, "loss": 1.933, "step": 1782 }, { "epoch": 0.17412109375, "grad_norm": 0.20754744112491608, "learning_rate": 0.0004757341020719517, "loss": 1.8626, "step": 1783 }, { "epoch": 0.17421875, "grad_norm": 0.29153457283973694, "learning_rate": 0.00047570196795247455, "loss": 1.8666, "step": 1784 }, { "epoch": 0.17431640625, "grad_norm": 0.36487746238708496, "learning_rate": 0.00047566981378464884, "loss": 1.8917, "step": 1785 }, { "epoch": 0.1744140625, "grad_norm": 0.32802101969718933, "learning_rate": 0.0004756376395716865, "loss": 1.8797, "step": 1786 }, { "epoch": 0.17451171875, "grad_norm": 0.2676317095756531, "learning_rate": 0.0004756054453168015, "loss": 1.9127, "step": 1787 }, { "epoch": 0.174609375, "grad_norm": 0.21236379444599152, "learning_rate": 0.0004755732310232097, "loss": 1.8628, "step": 1788 }, { "epoch": 0.17470703125, "grad_norm": 0.2659597396850586, "learning_rate": 0.000475540996694129, "loss": 1.8725, "step": 1789 }, { "epoch": 0.1748046875, "grad_norm": 0.2773102819919586, "learning_rate": 0.00047550874233277937, "loss": 1.9394, "step": 1790 }, { "epoch": 0.17490234375, "grad_norm": 0.2603963613510132, "learning_rate": 0.00047547646794238277, "loss": 1.9164, "step": 1791 }, { "epoch": 0.175, "grad_norm": 0.21353478729724884, "learning_rate": 0.00047544417352616305, "loss": 1.9091, "step": 1792 }, { "epoch": 0.17509765625, "grad_norm": 0.29758280515670776, "learning_rate": 0.0004754118590873462, "loss": 1.9137, "step": 1793 }, { "epoch": 0.1751953125, "grad_norm": 0.27226370573043823, "learning_rate": 0.0004753795246291601, "loss": 1.944, "step": 1794 }, { "epoch": 0.17529296875, "grad_norm": 0.3030240833759308, "learning_rate": 0.0004753471701548347, "loss": 1.8922, "step": 1795 }, { "epoch": 0.175390625, "grad_norm": 0.2778298258781433, "learning_rate": 0.0004753147956676019, "loss": 1.9321, "step": 1796 }, { "epoch": 0.17548828125, "grad_norm": 0.2760380506515503, "learning_rate": 0.0004752824011706956, "loss": 1.9063, "step": 1797 }, { "epoch": 0.1755859375, "grad_norm": 0.2793836295604706, "learning_rate": 0.00047524998666735184, "loss": 1.9186, "step": 1798 }, { "epoch": 0.17568359375, "grad_norm": 0.25465574860572815, "learning_rate": 0.00047521755216080845, "loss": 1.9247, "step": 1799 }, { "epoch": 0.17578125, "grad_norm": 0.23424822092056274, "learning_rate": 0.00047518509765430536, "loss": 1.924, "step": 1800 } ], "logging_steps": 1.0, "max_steps": 10240, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.102051070139433e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }