diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,164331 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 23470, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 8.521516829995739e-05, + "grad_norm": 18.5, + "learning_rate": 2.5000000000000002e-08, + "loss": 1.21, + "step": 1 + }, + { + "epoch": 0.00017043033659991478, + "grad_norm": 34.25, + "learning_rate": 5.0000000000000004e-08, + "loss": 1.5951, + "step": 2 + }, + { + "epoch": 0.00025564550489987217, + "grad_norm": 32.25, + "learning_rate": 7.500000000000001e-08, + "loss": 1.4702, + "step": 3 + }, + { + "epoch": 0.00034086067319982955, + "grad_norm": 27.5, + "learning_rate": 1.0000000000000001e-07, + "loss": 1.3454, + "step": 4 + }, + { + "epoch": 0.00042607584149978694, + "grad_norm": 28.125, + "learning_rate": 1.2500000000000002e-07, + "loss": 1.16, + "step": 5 + }, + { + "epoch": 0.0005112910097997443, + "grad_norm": 27.75, + "learning_rate": 1.5000000000000002e-07, + "loss": 1.6235, + "step": 6 + }, + { + "epoch": 0.0005965061780997018, + "grad_norm": 27.0, + "learning_rate": 1.7500000000000002e-07, + "loss": 1.3127, + "step": 7 + }, + { + "epoch": 0.0006817213463996591, + "grad_norm": 35.0, + "learning_rate": 2.0000000000000002e-07, + "loss": 1.4014, + "step": 8 + }, + { + "epoch": 0.0007669365146996165, + "grad_norm": 19.375, + "learning_rate": 2.2500000000000002e-07, + "loss": 1.1322, + "step": 9 + }, + { + "epoch": 0.0008521516829995739, + "grad_norm": 28.25, + "learning_rate": 2.5000000000000004e-07, + "loss": 1.2969, + "step": 10 + }, + { + "epoch": 0.0009373668512995313, + "grad_norm": 43.0, + "learning_rate": 2.75e-07, + "loss": 1.4659, + "step": 11 + }, + { + "epoch": 0.0010225820195994887, + "grad_norm": 14.875, + "learning_rate": 3.0000000000000004e-07, + "loss": 0.8019, + "step": 12 + }, + { + "epoch": 0.001107797187899446, + "grad_norm": 45.75, + "learning_rate": 3.25e-07, + "loss": 1.4666, + "step": 13 + }, + { + "epoch": 0.0011930123561994035, + "grad_norm": 52.0, + "learning_rate": 3.5000000000000004e-07, + "loss": 2.0725, + "step": 14 + }, + { + "epoch": 0.0012782275244993609, + "grad_norm": 91.5, + "learning_rate": 3.75e-07, + "loss": 1.517, + "step": 15 + }, + { + "epoch": 0.0013634426927993182, + "grad_norm": 53.5, + "learning_rate": 4.0000000000000003e-07, + "loss": 1.6968, + "step": 16 + }, + { + "epoch": 0.0014486578610992758, + "grad_norm": 68.0, + "learning_rate": 4.2500000000000006e-07, + "loss": 1.3316, + "step": 17 + }, + { + "epoch": 0.001533873029399233, + "grad_norm": 17.875, + "learning_rate": 4.5000000000000003e-07, + "loss": 1.1111, + "step": 18 + }, + { + "epoch": 0.0016190881976991904, + "grad_norm": 36.75, + "learning_rate": 4.7500000000000006e-07, + "loss": 1.5615, + "step": 19 + }, + { + "epoch": 0.0017043033659991478, + "grad_norm": 55.75, + "learning_rate": 5.000000000000001e-07, + "loss": 1.5168, + "step": 20 + }, + { + "epoch": 0.0017895185342991053, + "grad_norm": 23.625, + "learning_rate": 5.250000000000001e-07, + "loss": 1.3197, + "step": 21 + }, + { + "epoch": 0.0018747337025990627, + "grad_norm": 31.5, + "learning_rate": 5.5e-07, + "loss": 1.3818, + "step": 22 + }, + { + "epoch": 0.00195994887089902, + "grad_norm": 24.75, + "learning_rate": 5.750000000000001e-07, + "loss": 1.4715, + "step": 23 + }, + { + "epoch": 0.0020451640391989773, + "grad_norm": 26.25, + "learning_rate": 6.000000000000001e-07, + "loss": 1.2183, + "step": 24 + }, + { + "epoch": 0.0021303792074989347, + "grad_norm": 52.5, + "learning_rate": 6.25e-07, + "loss": 1.5819, + "step": 25 + }, + { + "epoch": 0.002215594375798892, + "grad_norm": 18.875, + "learning_rate": 6.5e-07, + "loss": 1.224, + "step": 26 + }, + { + "epoch": 0.0023008095440988498, + "grad_norm": 47.75, + "learning_rate": 6.750000000000001e-07, + "loss": 1.0704, + "step": 27 + }, + { + "epoch": 0.002386024712398807, + "grad_norm": 174.0, + "learning_rate": 7.000000000000001e-07, + "loss": 1.2073, + "step": 28 + }, + { + "epoch": 0.0024712398806987644, + "grad_norm": 37.75, + "learning_rate": 7.25e-07, + "loss": 1.6614, + "step": 29 + }, + { + "epoch": 0.0025564550489987218, + "grad_norm": 47.25, + "learning_rate": 7.5e-07, + "loss": 1.1309, + "step": 30 + }, + { + "epoch": 0.002641670217298679, + "grad_norm": 24.625, + "learning_rate": 7.750000000000001e-07, + "loss": 1.4704, + "step": 31 + }, + { + "epoch": 0.0027268853855986364, + "grad_norm": 53.25, + "learning_rate": 8.000000000000001e-07, + "loss": 0.93, + "step": 32 + }, + { + "epoch": 0.0028121005538985938, + "grad_norm": 33.25, + "learning_rate": 8.250000000000001e-07, + "loss": 1.2446, + "step": 33 + }, + { + "epoch": 0.0028973157221985515, + "grad_norm": 25.75, + "learning_rate": 8.500000000000001e-07, + "loss": 1.4553, + "step": 34 + }, + { + "epoch": 0.002982530890498509, + "grad_norm": 17.0, + "learning_rate": 8.75e-07, + "loss": 1.0626, + "step": 35 + }, + { + "epoch": 0.003067746058798466, + "grad_norm": 39.75, + "learning_rate": 9.000000000000001e-07, + "loss": 1.3777, + "step": 36 + }, + { + "epoch": 0.0031529612270984235, + "grad_norm": 22.875, + "learning_rate": 9.25e-07, + "loss": 1.2808, + "step": 37 + }, + { + "epoch": 0.003238176395398381, + "grad_norm": 15.125, + "learning_rate": 9.500000000000001e-07, + "loss": 1.0063, + "step": 38 + }, + { + "epoch": 0.003323391563698338, + "grad_norm": 15.5625, + "learning_rate": 9.750000000000002e-07, + "loss": 1.2981, + "step": 39 + }, + { + "epoch": 0.0034086067319982955, + "grad_norm": 15.75, + "learning_rate": 1.0000000000000002e-06, + "loss": 0.8237, + "step": 40 + }, + { + "epoch": 0.0034938219002982533, + "grad_norm": 21.375, + "learning_rate": 1.025e-06, + "loss": 1.3627, + "step": 41 + }, + { + "epoch": 0.0035790370685982106, + "grad_norm": 24.125, + "learning_rate": 1.0500000000000001e-06, + "loss": 1.4088, + "step": 42 + }, + { + "epoch": 0.003664252236898168, + "grad_norm": 18.75, + "learning_rate": 1.075e-06, + "loss": 1.1094, + "step": 43 + }, + { + "epoch": 0.0037494674051981253, + "grad_norm": 20.875, + "learning_rate": 1.1e-06, + "loss": 1.3761, + "step": 44 + }, + { + "epoch": 0.0038346825734980826, + "grad_norm": 30.5, + "learning_rate": 1.125e-06, + "loss": 1.1231, + "step": 45 + }, + { + "epoch": 0.00391989774179804, + "grad_norm": 18.0, + "learning_rate": 1.1500000000000002e-06, + "loss": 1.0932, + "step": 46 + }, + { + "epoch": 0.004005112910097998, + "grad_norm": 34.0, + "learning_rate": 1.175e-06, + "loss": 1.5352, + "step": 47 + }, + { + "epoch": 0.004090328078397955, + "grad_norm": 34.25, + "learning_rate": 1.2000000000000002e-06, + "loss": 1.4602, + "step": 48 + }, + { + "epoch": 0.004175543246697912, + "grad_norm": 27.25, + "learning_rate": 1.2250000000000001e-06, + "loss": 1.0401, + "step": 49 + }, + { + "epoch": 0.004260758414997869, + "grad_norm": 22.625, + "learning_rate": 1.25e-06, + "loss": 1.3274, + "step": 50 + }, + { + "epoch": 0.004345973583297827, + "grad_norm": 27.0, + "learning_rate": 1.275e-06, + "loss": 1.5387, + "step": 51 + }, + { + "epoch": 0.004431188751597784, + "grad_norm": 26.0, + "learning_rate": 1.3e-06, + "loss": 1.1358, + "step": 52 + }, + { + "epoch": 0.004516403919897742, + "grad_norm": 27.375, + "learning_rate": 1.3250000000000002e-06, + "loss": 1.2844, + "step": 53 + }, + { + "epoch": 0.0046016190881976995, + "grad_norm": 23.25, + "learning_rate": 1.3500000000000002e-06, + "loss": 1.1243, + "step": 54 + }, + { + "epoch": 0.004686834256497656, + "grad_norm": 49.25, + "learning_rate": 1.3750000000000002e-06, + "loss": 1.3856, + "step": 55 + }, + { + "epoch": 0.004772049424797614, + "grad_norm": 44.5, + "learning_rate": 1.4000000000000001e-06, + "loss": 1.4186, + "step": 56 + }, + { + "epoch": 0.004857264593097571, + "grad_norm": 26.75, + "learning_rate": 1.425e-06, + "loss": 1.0897, + "step": 57 + }, + { + "epoch": 0.004942479761397529, + "grad_norm": 17.0, + "learning_rate": 1.45e-06, + "loss": 1.0632, + "step": 58 + }, + { + "epoch": 0.005027694929697486, + "grad_norm": 98.5, + "learning_rate": 1.475e-06, + "loss": 1.0083, + "step": 59 + }, + { + "epoch": 0.0051129100979974435, + "grad_norm": 62.25, + "learning_rate": 1.5e-06, + "loss": 0.9524, + "step": 60 + }, + { + "epoch": 0.005198125266297401, + "grad_norm": 74.0, + "learning_rate": 1.525e-06, + "loss": 1.3847, + "step": 61 + }, + { + "epoch": 0.005283340434597358, + "grad_norm": 21.0, + "learning_rate": 1.5500000000000002e-06, + "loss": 0.7665, + "step": 62 + }, + { + "epoch": 0.005368555602897316, + "grad_norm": 27.625, + "learning_rate": 1.5750000000000002e-06, + "loss": 0.9051, + "step": 63 + }, + { + "epoch": 0.005453770771197273, + "grad_norm": 35.75, + "learning_rate": 1.6000000000000001e-06, + "loss": 1.1105, + "step": 64 + }, + { + "epoch": 0.005538985939497231, + "grad_norm": 17.375, + "learning_rate": 1.6250000000000001e-06, + "loss": 0.9974, + "step": 65 + }, + { + "epoch": 0.0056242011077971875, + "grad_norm": 13.1875, + "learning_rate": 1.6500000000000003e-06, + "loss": 0.9864, + "step": 66 + }, + { + "epoch": 0.005709416276097145, + "grad_norm": 10.375, + "learning_rate": 1.6750000000000003e-06, + "loss": 0.5889, + "step": 67 + }, + { + "epoch": 0.005794631444397103, + "grad_norm": 23.875, + "learning_rate": 1.7000000000000002e-06, + "loss": 1.0158, + "step": 68 + }, + { + "epoch": 0.00587984661269706, + "grad_norm": 15.875, + "learning_rate": 1.725e-06, + "loss": 0.6812, + "step": 69 + }, + { + "epoch": 0.005965061780997018, + "grad_norm": 35.75, + "learning_rate": 1.75e-06, + "loss": 1.4199, + "step": 70 + }, + { + "epoch": 0.006050276949296975, + "grad_norm": 15.4375, + "learning_rate": 1.7750000000000002e-06, + "loss": 1.2153, + "step": 71 + }, + { + "epoch": 0.006135492117596932, + "grad_norm": 13.75, + "learning_rate": 1.8000000000000001e-06, + "loss": 0.8568, + "step": 72 + }, + { + "epoch": 0.006220707285896889, + "grad_norm": 15.8125, + "learning_rate": 1.825e-06, + "loss": 1.2163, + "step": 73 + }, + { + "epoch": 0.006305922454196847, + "grad_norm": 22.5, + "learning_rate": 1.85e-06, + "loss": 1.0211, + "step": 74 + }, + { + "epoch": 0.006391137622496805, + "grad_norm": 41.25, + "learning_rate": 1.8750000000000003e-06, + "loss": 1.7824, + "step": 75 + }, + { + "epoch": 0.006476352790796762, + "grad_norm": 18.5, + "learning_rate": 1.9000000000000002e-06, + "loss": 0.9872, + "step": 76 + }, + { + "epoch": 0.0065615679590967195, + "grad_norm": 14.75, + "learning_rate": 1.925e-06, + "loss": 1.1802, + "step": 77 + }, + { + "epoch": 0.006646783127396676, + "grad_norm": 17.75, + "learning_rate": 1.9500000000000004e-06, + "loss": 1.0842, + "step": 78 + }, + { + "epoch": 0.006731998295696634, + "grad_norm": 17.75, + "learning_rate": 1.975e-06, + "loss": 0.6949, + "step": 79 + }, + { + "epoch": 0.006817213463996591, + "grad_norm": 15.6875, + "learning_rate": 2.0000000000000003e-06, + "loss": 1.0766, + "step": 80 + }, + { + "epoch": 0.006902428632296549, + "grad_norm": 23.0, + "learning_rate": 2.025e-06, + "loss": 1.2883, + "step": 81 + }, + { + "epoch": 0.006987643800596507, + "grad_norm": 14.0, + "learning_rate": 2.05e-06, + "loss": 0.9718, + "step": 82 + }, + { + "epoch": 0.0070728589688964635, + "grad_norm": 13.1875, + "learning_rate": 2.075e-06, + "loss": 1.0055, + "step": 83 + }, + { + "epoch": 0.007158074137196421, + "grad_norm": 24.25, + "learning_rate": 2.1000000000000002e-06, + "loss": 1.0122, + "step": 84 + }, + { + "epoch": 0.007243289305496378, + "grad_norm": 13.375, + "learning_rate": 2.125e-06, + "loss": 0.6408, + "step": 85 + }, + { + "epoch": 0.007328504473796336, + "grad_norm": 12.25, + "learning_rate": 2.15e-06, + "loss": 0.9325, + "step": 86 + }, + { + "epoch": 0.007413719642096293, + "grad_norm": 16.75, + "learning_rate": 2.1750000000000004e-06, + "loss": 1.1798, + "step": 87 + }, + { + "epoch": 0.007498934810396251, + "grad_norm": 14.875, + "learning_rate": 2.2e-06, + "loss": 0.9925, + "step": 88 + }, + { + "epoch": 0.0075841499786962075, + "grad_norm": 16.25, + "learning_rate": 2.2250000000000003e-06, + "loss": 0.8689, + "step": 89 + }, + { + "epoch": 0.007669365146996165, + "grad_norm": 19.25, + "learning_rate": 2.25e-06, + "loss": 1.1733, + "step": 90 + }, + { + "epoch": 0.007754580315296123, + "grad_norm": 14.3125, + "learning_rate": 2.2750000000000002e-06, + "loss": 0.8938, + "step": 91 + }, + { + "epoch": 0.00783979548359608, + "grad_norm": 15.25, + "learning_rate": 2.3000000000000004e-06, + "loss": 0.7673, + "step": 92 + }, + { + "epoch": 0.007925010651896037, + "grad_norm": 17.5, + "learning_rate": 2.325e-06, + "loss": 0.8134, + "step": 93 + }, + { + "epoch": 0.008010225820195995, + "grad_norm": 11.625, + "learning_rate": 2.35e-06, + "loss": 0.7227, + "step": 94 + }, + { + "epoch": 0.008095440988495952, + "grad_norm": 20.375, + "learning_rate": 2.375e-06, + "loss": 0.8756, + "step": 95 + }, + { + "epoch": 0.00818065615679591, + "grad_norm": 16.875, + "learning_rate": 2.4000000000000003e-06, + "loss": 1.0683, + "step": 96 + }, + { + "epoch": 0.008265871325095868, + "grad_norm": 20.125, + "learning_rate": 2.425e-06, + "loss": 1.3241, + "step": 97 + }, + { + "epoch": 0.008351086493395825, + "grad_norm": 13.375, + "learning_rate": 2.4500000000000003e-06, + "loss": 0.7591, + "step": 98 + }, + { + "epoch": 0.008436301661695782, + "grad_norm": 12.9375, + "learning_rate": 2.475e-06, + "loss": 0.8464, + "step": 99 + }, + { + "epoch": 0.008521516829995739, + "grad_norm": 20.75, + "learning_rate": 2.5e-06, + "loss": 1.0074, + "step": 100 + }, + { + "epoch": 0.008606731998295697, + "grad_norm": 25.125, + "learning_rate": 2.499999988705622e-06, + "loss": 1.1472, + "step": 101 + }, + { + "epoch": 0.008691947166595654, + "grad_norm": 21.375, + "learning_rate": 2.4999999548224866e-06, + "loss": 1.5212, + "step": 102 + }, + { + "epoch": 0.008777162334895611, + "grad_norm": 13.25, + "learning_rate": 2.499999898350595e-06, + "loss": 0.6892, + "step": 103 + }, + { + "epoch": 0.008862377503195568, + "grad_norm": 22.25, + "learning_rate": 2.4999998192899483e-06, + "loss": 1.219, + "step": 104 + }, + { + "epoch": 0.008947592671495527, + "grad_norm": 19.875, + "learning_rate": 2.4999997176405483e-06, + "loss": 1.0446, + "step": 105 + }, + { + "epoch": 0.009032807839795483, + "grad_norm": 12.5, + "learning_rate": 2.4999995934023964e-06, + "loss": 0.8379, + "step": 106 + }, + { + "epoch": 0.00911802300809544, + "grad_norm": 19.25, + "learning_rate": 2.4999994465754946e-06, + "loss": 0.9465, + "step": 107 + }, + { + "epoch": 0.009203238176395399, + "grad_norm": 19.5, + "learning_rate": 2.4999992771598457e-06, + "loss": 0.952, + "step": 108 + }, + { + "epoch": 0.009288453344695356, + "grad_norm": 15.0, + "learning_rate": 2.4999990851554527e-06, + "loss": 0.8594, + "step": 109 + }, + { + "epoch": 0.009373668512995313, + "grad_norm": 18.75, + "learning_rate": 2.4999988705623203e-06, + "loss": 0.9269, + "step": 110 + }, + { + "epoch": 0.00945888368129527, + "grad_norm": 13.5, + "learning_rate": 2.4999986333804506e-06, + "loss": 0.7822, + "step": 111 + }, + { + "epoch": 0.009544098849595228, + "grad_norm": 10.625, + "learning_rate": 2.4999983736098486e-06, + "loss": 0.5126, + "step": 112 + }, + { + "epoch": 0.009629314017895185, + "grad_norm": 15.9375, + "learning_rate": 2.499998091250519e-06, + "loss": 0.9473, + "step": 113 + }, + { + "epoch": 0.009714529186195142, + "grad_norm": 13.4375, + "learning_rate": 2.499997786302467e-06, + "loss": 0.7369, + "step": 114 + }, + { + "epoch": 0.0097997443544951, + "grad_norm": 10.6875, + "learning_rate": 2.499997458765698e-06, + "loss": 0.5523, + "step": 115 + }, + { + "epoch": 0.009884959522795058, + "grad_norm": 15.875, + "learning_rate": 2.499997108640218e-06, + "loss": 0.8407, + "step": 116 + }, + { + "epoch": 0.009970174691095015, + "grad_norm": 18.875, + "learning_rate": 2.4999967359260335e-06, + "loss": 0.8085, + "step": 117 + }, + { + "epoch": 0.010055389859394971, + "grad_norm": 15.3125, + "learning_rate": 2.4999963406231508e-06, + "loss": 0.923, + "step": 118 + }, + { + "epoch": 0.01014060502769493, + "grad_norm": 15.4375, + "learning_rate": 2.4999959227315773e-06, + "loss": 1.1125, + "step": 119 + }, + { + "epoch": 0.010225820195994887, + "grad_norm": 14.4375, + "learning_rate": 2.4999954822513203e-06, + "loss": 1.0749, + "step": 120 + }, + { + "epoch": 0.010311035364294844, + "grad_norm": 13.4375, + "learning_rate": 2.499995019182388e-06, + "loss": 0.9686, + "step": 121 + }, + { + "epoch": 0.010396250532594803, + "grad_norm": 21.125, + "learning_rate": 2.4999945335247888e-06, + "loss": 1.3533, + "step": 122 + }, + { + "epoch": 0.01048146570089476, + "grad_norm": 14.625, + "learning_rate": 2.4999940252785316e-06, + "loss": 0.8111, + "step": 123 + }, + { + "epoch": 0.010566680869194716, + "grad_norm": 13.0625, + "learning_rate": 2.4999934944436256e-06, + "loss": 0.6665, + "step": 124 + }, + { + "epoch": 0.010651896037494673, + "grad_norm": 18.125, + "learning_rate": 2.4999929410200797e-06, + "loss": 0.7747, + "step": 125 + }, + { + "epoch": 0.010737111205794632, + "grad_norm": 22.25, + "learning_rate": 2.4999923650079044e-06, + "loss": 1.0984, + "step": 126 + }, + { + "epoch": 0.010822326374094589, + "grad_norm": 19.0, + "learning_rate": 2.4999917664071105e-06, + "loss": 1.0494, + "step": 127 + }, + { + "epoch": 0.010907541542394546, + "grad_norm": 14.1875, + "learning_rate": 2.499991145217708e-06, + "loss": 0.9469, + "step": 128 + }, + { + "epoch": 0.010992756710694504, + "grad_norm": 10.75, + "learning_rate": 2.4999905014397087e-06, + "loss": 0.5802, + "step": 129 + }, + { + "epoch": 0.011077971878994461, + "grad_norm": 14.25, + "learning_rate": 2.4999898350731242e-06, + "loss": 0.9716, + "step": 130 + }, + { + "epoch": 0.011163187047294418, + "grad_norm": 15.4375, + "learning_rate": 2.4999891461179667e-06, + "loss": 1.0489, + "step": 131 + }, + { + "epoch": 0.011248402215594375, + "grad_norm": 21.0, + "learning_rate": 2.4999884345742476e-06, + "loss": 1.3434, + "step": 132 + }, + { + "epoch": 0.011333617383894334, + "grad_norm": 14.8125, + "learning_rate": 2.499987700441981e-06, + "loss": 0.6333, + "step": 133 + }, + { + "epoch": 0.01141883255219429, + "grad_norm": 15.375, + "learning_rate": 2.4999869437211795e-06, + "loss": 0.8497, + "step": 134 + }, + { + "epoch": 0.011504047720494247, + "grad_norm": 22.875, + "learning_rate": 2.499986164411857e-06, + "loss": 1.2663, + "step": 135 + }, + { + "epoch": 0.011589262888794206, + "grad_norm": 17.375, + "learning_rate": 2.4999853625140277e-06, + "loss": 1.0296, + "step": 136 + }, + { + "epoch": 0.011674478057094163, + "grad_norm": 18.25, + "learning_rate": 2.499984538027706e-06, + "loss": 0.7126, + "step": 137 + }, + { + "epoch": 0.01175969322539412, + "grad_norm": 25.375, + "learning_rate": 2.4999836909529067e-06, + "loss": 1.3332, + "step": 138 + }, + { + "epoch": 0.011844908393694077, + "grad_norm": 12.8125, + "learning_rate": 2.499982821289645e-06, + "loss": 0.7095, + "step": 139 + }, + { + "epoch": 0.011930123561994035, + "grad_norm": 15.3125, + "learning_rate": 2.4999819290379367e-06, + "loss": 1.0405, + "step": 140 + }, + { + "epoch": 0.012015338730293992, + "grad_norm": 13.6875, + "learning_rate": 2.499981014197798e-06, + "loss": 0.537, + "step": 141 + }, + { + "epoch": 0.01210055389859395, + "grad_norm": 12.125, + "learning_rate": 2.4999800767692456e-06, + "loss": 0.6729, + "step": 142 + }, + { + "epoch": 0.012185769066893908, + "grad_norm": 20.125, + "learning_rate": 2.4999791167522964e-06, + "loss": 1.2783, + "step": 143 + }, + { + "epoch": 0.012270984235193865, + "grad_norm": 16.125, + "learning_rate": 2.4999781341469673e-06, + "loss": 1.2955, + "step": 144 + }, + { + "epoch": 0.012356199403493822, + "grad_norm": 11.1875, + "learning_rate": 2.4999771289532765e-06, + "loss": 0.6357, + "step": 145 + }, + { + "epoch": 0.012441414571793779, + "grad_norm": 15.375, + "learning_rate": 2.499976101171242e-06, + "loss": 1.0036, + "step": 146 + }, + { + "epoch": 0.012526629740093737, + "grad_norm": 24.125, + "learning_rate": 2.4999750508008825e-06, + "loss": 1.1779, + "step": 147 + }, + { + "epoch": 0.012611844908393694, + "grad_norm": 13.0625, + "learning_rate": 2.4999739778422164e-06, + "loss": 0.8735, + "step": 148 + }, + { + "epoch": 0.012697060076693651, + "grad_norm": 14.6875, + "learning_rate": 2.499972882295264e-06, + "loss": 0.7951, + "step": 149 + }, + { + "epoch": 0.01278227524499361, + "grad_norm": 16.25, + "learning_rate": 2.4999717641600446e-06, + "loss": 0.8998, + "step": 150 + }, + { + "epoch": 0.012867490413293567, + "grad_norm": 20.0, + "learning_rate": 2.4999706234365786e-06, + "loss": 1.4487, + "step": 151 + }, + { + "epoch": 0.012952705581593523, + "grad_norm": 16.125, + "learning_rate": 2.4999694601248865e-06, + "loss": 1.0404, + "step": 152 + }, + { + "epoch": 0.01303792074989348, + "grad_norm": 15.3125, + "learning_rate": 2.4999682742249894e-06, + "loss": 0.8277, + "step": 153 + }, + { + "epoch": 0.013123135918193439, + "grad_norm": 18.875, + "learning_rate": 2.499967065736908e-06, + "loss": 0.9591, + "step": 154 + }, + { + "epoch": 0.013208351086493396, + "grad_norm": 13.6875, + "learning_rate": 2.4999658346606655e-06, + "loss": 0.7346, + "step": 155 + }, + { + "epoch": 0.013293566254793353, + "grad_norm": 19.25, + "learning_rate": 2.499964580996283e-06, + "loss": 0.8684, + "step": 156 + }, + { + "epoch": 0.013378781423093311, + "grad_norm": 14.0625, + "learning_rate": 2.499963304743784e-06, + "loss": 0.8248, + "step": 157 + }, + { + "epoch": 0.013463996591393268, + "grad_norm": 15.375, + "learning_rate": 2.499962005903191e-06, + "loss": 1.0007, + "step": 158 + }, + { + "epoch": 0.013549211759693225, + "grad_norm": 12.9375, + "learning_rate": 2.4999606844745272e-06, + "loss": 0.8773, + "step": 159 + }, + { + "epoch": 0.013634426927993182, + "grad_norm": 12.625, + "learning_rate": 2.499959340457817e-06, + "loss": 0.8744, + "step": 160 + }, + { + "epoch": 0.01371964209629314, + "grad_norm": 20.25, + "learning_rate": 2.499957973853085e-06, + "loss": 1.0151, + "step": 161 + }, + { + "epoch": 0.013804857264593098, + "grad_norm": 21.75, + "learning_rate": 2.4999565846603556e-06, + "loss": 1.0634, + "step": 162 + }, + { + "epoch": 0.013890072432893055, + "grad_norm": 17.375, + "learning_rate": 2.4999551728796533e-06, + "loss": 0.9191, + "step": 163 + }, + { + "epoch": 0.013975287601193013, + "grad_norm": 14.125, + "learning_rate": 2.4999537385110047e-06, + "loss": 0.7709, + "step": 164 + }, + { + "epoch": 0.01406050276949297, + "grad_norm": 15.3125, + "learning_rate": 2.4999522815544345e-06, + "loss": 0.6242, + "step": 165 + }, + { + "epoch": 0.014145717937792927, + "grad_norm": 12.375, + "learning_rate": 2.49995080200997e-06, + "loss": 0.6428, + "step": 166 + }, + { + "epoch": 0.014230933106092884, + "grad_norm": 18.125, + "learning_rate": 2.499949299877637e-06, + "loss": 0.9351, + "step": 167 + }, + { + "epoch": 0.014316148274392843, + "grad_norm": 27.75, + "learning_rate": 2.499947775157464e-06, + "loss": 1.3185, + "step": 168 + }, + { + "epoch": 0.0144013634426928, + "grad_norm": 19.375, + "learning_rate": 2.4999462278494776e-06, + "loss": 0.847, + "step": 169 + }, + { + "epoch": 0.014486578610992756, + "grad_norm": 16.25, + "learning_rate": 2.4999446579537054e-06, + "loss": 1.1577, + "step": 170 + }, + { + "epoch": 0.014571793779292713, + "grad_norm": 10.25, + "learning_rate": 2.499943065470177e-06, + "loss": 0.4071, + "step": 171 + }, + { + "epoch": 0.014657008947592672, + "grad_norm": 11.25, + "learning_rate": 2.49994145039892e-06, + "loss": 0.5795, + "step": 172 + }, + { + "epoch": 0.014742224115892629, + "grad_norm": 14.875, + "learning_rate": 2.499939812739964e-06, + "loss": 0.7708, + "step": 173 + }, + { + "epoch": 0.014827439284192586, + "grad_norm": 13.8125, + "learning_rate": 2.499938152493339e-06, + "loss": 0.9726, + "step": 174 + }, + { + "epoch": 0.014912654452492544, + "grad_norm": 16.375, + "learning_rate": 2.4999364696590747e-06, + "loss": 0.9742, + "step": 175 + }, + { + "epoch": 0.014997869620792501, + "grad_norm": 19.125, + "learning_rate": 2.4999347642372017e-06, + "loss": 0.7613, + "step": 176 + }, + { + "epoch": 0.015083084789092458, + "grad_norm": 18.375, + "learning_rate": 2.49993303622775e-06, + "loss": 1.0638, + "step": 177 + }, + { + "epoch": 0.015168299957392415, + "grad_norm": 16.25, + "learning_rate": 2.499931285630752e-06, + "loss": 0.7163, + "step": 178 + }, + { + "epoch": 0.015253515125692374, + "grad_norm": 15.75, + "learning_rate": 2.4999295124462384e-06, + "loss": 0.958, + "step": 179 + }, + { + "epoch": 0.01533873029399233, + "grad_norm": 13.9375, + "learning_rate": 2.4999277166742418e-06, + "loss": 0.8928, + "step": 180 + }, + { + "epoch": 0.015423945462292287, + "grad_norm": 14.8125, + "learning_rate": 2.4999258983147944e-06, + "loss": 0.836, + "step": 181 + }, + { + "epoch": 0.015509160630592246, + "grad_norm": 14.0, + "learning_rate": 2.499924057367929e-06, + "loss": 0.883, + "step": 182 + }, + { + "epoch": 0.015594375798892203, + "grad_norm": 15.375, + "learning_rate": 2.499922193833679e-06, + "loss": 0.9203, + "step": 183 + }, + { + "epoch": 0.01567959096719216, + "grad_norm": 10.375, + "learning_rate": 2.4999203077120782e-06, + "loss": 0.4823, + "step": 184 + }, + { + "epoch": 0.015764806135492117, + "grad_norm": 20.125, + "learning_rate": 2.4999183990031606e-06, + "loss": 1.1846, + "step": 185 + }, + { + "epoch": 0.015850021303792074, + "grad_norm": 14.125, + "learning_rate": 2.4999164677069605e-06, + "loss": 0.7984, + "step": 186 + }, + { + "epoch": 0.015935236472092034, + "grad_norm": 15.375, + "learning_rate": 2.499914513823513e-06, + "loss": 0.7548, + "step": 187 + }, + { + "epoch": 0.01602045164039199, + "grad_norm": 14.0, + "learning_rate": 2.4999125373528532e-06, + "loss": 0.8622, + "step": 188 + }, + { + "epoch": 0.016105666808691948, + "grad_norm": 11.8125, + "learning_rate": 2.499910538295017e-06, + "loss": 0.4636, + "step": 189 + }, + { + "epoch": 0.016190881976991905, + "grad_norm": 10.9375, + "learning_rate": 2.4999085166500406e-06, + "loss": 0.5639, + "step": 190 + }, + { + "epoch": 0.01627609714529186, + "grad_norm": 13.75, + "learning_rate": 2.4999064724179605e-06, + "loss": 0.7793, + "step": 191 + }, + { + "epoch": 0.01636131231359182, + "grad_norm": 13.4375, + "learning_rate": 2.4999044055988135e-06, + "loss": 0.8171, + "step": 192 + }, + { + "epoch": 0.016446527481891775, + "grad_norm": 13.4375, + "learning_rate": 2.499902316192637e-06, + "loss": 0.6888, + "step": 193 + }, + { + "epoch": 0.016531742650191736, + "grad_norm": 14.0625, + "learning_rate": 2.4999002041994685e-06, + "loss": 0.7249, + "step": 194 + }, + { + "epoch": 0.016616957818491693, + "grad_norm": 13.1875, + "learning_rate": 2.499898069619347e-06, + "loss": 0.7304, + "step": 195 + }, + { + "epoch": 0.01670217298679165, + "grad_norm": 12.6875, + "learning_rate": 2.49989591245231e-06, + "loss": 0.863, + "step": 196 + }, + { + "epoch": 0.016787388155091607, + "grad_norm": 14.375, + "learning_rate": 2.4998937326983974e-06, + "loss": 0.7757, + "step": 197 + }, + { + "epoch": 0.016872603323391563, + "grad_norm": 17.0, + "learning_rate": 2.4998915303576478e-06, + "loss": 0.8415, + "step": 198 + }, + { + "epoch": 0.01695781849169152, + "grad_norm": 18.875, + "learning_rate": 2.4998893054301014e-06, + "loss": 1.3295, + "step": 199 + }, + { + "epoch": 0.017043033659991477, + "grad_norm": 15.4375, + "learning_rate": 2.4998870579157984e-06, + "loss": 0.9835, + "step": 200 + }, + { + "epoch": 0.017128248828291438, + "grad_norm": 21.75, + "learning_rate": 2.4998847878147796e-06, + "loss": 1.1561, + "step": 201 + }, + { + "epoch": 0.017213463996591395, + "grad_norm": 10.4375, + "learning_rate": 2.499882495127086e-06, + "loss": 0.5615, + "step": 202 + }, + { + "epoch": 0.01729867916489135, + "grad_norm": 14.3125, + "learning_rate": 2.499880179852759e-06, + "loss": 0.8957, + "step": 203 + }, + { + "epoch": 0.01738389433319131, + "grad_norm": 17.25, + "learning_rate": 2.4998778419918395e-06, + "loss": 0.6773, + "step": 204 + }, + { + "epoch": 0.017469109501491265, + "grad_norm": 12.25, + "learning_rate": 2.4998754815443714e-06, + "loss": 0.9909, + "step": 205 + }, + { + "epoch": 0.017554324669791222, + "grad_norm": 13.1875, + "learning_rate": 2.499873098510396e-06, + "loss": 0.8327, + "step": 206 + }, + { + "epoch": 0.01763953983809118, + "grad_norm": 15.5, + "learning_rate": 2.4998706928899567e-06, + "loss": 0.7118, + "step": 207 + }, + { + "epoch": 0.017724755006391136, + "grad_norm": 13.0, + "learning_rate": 2.4998682646830974e-06, + "loss": 0.6673, + "step": 208 + }, + { + "epoch": 0.017809970174691096, + "grad_norm": 19.75, + "learning_rate": 2.499865813889862e-06, + "loss": 0.6474, + "step": 209 + }, + { + "epoch": 0.017895185342991053, + "grad_norm": 12.3125, + "learning_rate": 2.499863340510294e-06, + "loss": 0.6667, + "step": 210 + }, + { + "epoch": 0.01798040051129101, + "grad_norm": 35.5, + "learning_rate": 2.4998608445444387e-06, + "loss": 0.7577, + "step": 211 + }, + { + "epoch": 0.018065615679590967, + "grad_norm": 18.25, + "learning_rate": 2.4998583259923413e-06, + "loss": 0.5801, + "step": 212 + }, + { + "epoch": 0.018150830847890924, + "grad_norm": 14.375, + "learning_rate": 2.4998557848540473e-06, + "loss": 0.6952, + "step": 213 + }, + { + "epoch": 0.01823604601619088, + "grad_norm": 17.5, + "learning_rate": 2.4998532211296024e-06, + "loss": 0.9168, + "step": 214 + }, + { + "epoch": 0.018321261184490838, + "grad_norm": 15.0, + "learning_rate": 2.4998506348190523e-06, + "loss": 0.7822, + "step": 215 + }, + { + "epoch": 0.018406476352790798, + "grad_norm": 19.25, + "learning_rate": 2.499848025922445e-06, + "loss": 0.7619, + "step": 216 + }, + { + "epoch": 0.018491691521090755, + "grad_norm": 16.25, + "learning_rate": 2.4998453944398273e-06, + "loss": 0.83, + "step": 217 + }, + { + "epoch": 0.018576906689390712, + "grad_norm": 22.375, + "learning_rate": 2.499842740371246e-06, + "loss": 1.004, + "step": 218 + }, + { + "epoch": 0.01866212185769067, + "grad_norm": 14.8125, + "learning_rate": 2.4998400637167495e-06, + "loss": 0.6864, + "step": 219 + }, + { + "epoch": 0.018747337025990626, + "grad_norm": 13.25, + "learning_rate": 2.4998373644763864e-06, + "loss": 0.765, + "step": 220 + }, + { + "epoch": 0.018832552194290583, + "grad_norm": 23.5, + "learning_rate": 2.4998346426502052e-06, + "loss": 1.162, + "step": 221 + }, + { + "epoch": 0.01891776736259054, + "grad_norm": 12.375, + "learning_rate": 2.4998318982382556e-06, + "loss": 0.6497, + "step": 222 + }, + { + "epoch": 0.0190029825308905, + "grad_norm": 13.0, + "learning_rate": 2.4998291312405867e-06, + "loss": 0.556, + "step": 223 + }, + { + "epoch": 0.019088197699190457, + "grad_norm": 20.0, + "learning_rate": 2.4998263416572484e-06, + "loss": 0.8032, + "step": 224 + }, + { + "epoch": 0.019173412867490414, + "grad_norm": 18.75, + "learning_rate": 2.499823529488291e-06, + "loss": 0.7872, + "step": 225 + }, + { + "epoch": 0.01925862803579037, + "grad_norm": 16.75, + "learning_rate": 2.499820694733766e-06, + "loss": 0.7131, + "step": 226 + }, + { + "epoch": 0.019343843204090327, + "grad_norm": 13.3125, + "learning_rate": 2.499817837393724e-06, + "loss": 0.7218, + "step": 227 + }, + { + "epoch": 0.019429058372390284, + "grad_norm": 13.3125, + "learning_rate": 2.4998149574682175e-06, + "loss": 0.8036, + "step": 228 + }, + { + "epoch": 0.01951427354069024, + "grad_norm": 14.9375, + "learning_rate": 2.4998120549572972e-06, + "loss": 0.9664, + "step": 229 + }, + { + "epoch": 0.0195994887089902, + "grad_norm": 13.75, + "learning_rate": 2.4998091298610168e-06, + "loss": 0.8357, + "step": 230 + }, + { + "epoch": 0.01968470387729016, + "grad_norm": 11.875, + "learning_rate": 2.4998061821794282e-06, + "loss": 0.7416, + "step": 231 + }, + { + "epoch": 0.019769919045590115, + "grad_norm": 13.5, + "learning_rate": 2.4998032119125858e-06, + "loss": 0.5895, + "step": 232 + }, + { + "epoch": 0.019855134213890072, + "grad_norm": 18.875, + "learning_rate": 2.4998002190605423e-06, + "loss": 0.8778, + "step": 233 + }, + { + "epoch": 0.01994034938219003, + "grad_norm": 18.625, + "learning_rate": 2.499797203623352e-06, + "loss": 0.5428, + "step": 234 + }, + { + "epoch": 0.020025564550489986, + "grad_norm": 16.75, + "learning_rate": 2.4997941656010695e-06, + "loss": 1.2933, + "step": 235 + }, + { + "epoch": 0.020110779718789943, + "grad_norm": 16.25, + "learning_rate": 2.4997911049937493e-06, + "loss": 0.7574, + "step": 236 + }, + { + "epoch": 0.020195994887089903, + "grad_norm": 18.75, + "learning_rate": 2.4997880218014476e-06, + "loss": 1.1304, + "step": 237 + }, + { + "epoch": 0.02028121005538986, + "grad_norm": 14.4375, + "learning_rate": 2.499784916024219e-06, + "loss": 0.7651, + "step": 238 + }, + { + "epoch": 0.020366425223689817, + "grad_norm": 16.625, + "learning_rate": 2.4997817876621204e-06, + "loss": 0.8226, + "step": 239 + }, + { + "epoch": 0.020451640391989774, + "grad_norm": 14.5, + "learning_rate": 2.4997786367152086e-06, + "loss": 0.9215, + "step": 240 + }, + { + "epoch": 0.02053685556028973, + "grad_norm": 17.25, + "learning_rate": 2.4997754631835396e-06, + "loss": 1.1801, + "step": 241 + }, + { + "epoch": 0.020622070728589688, + "grad_norm": 13.5625, + "learning_rate": 2.4997722670671715e-06, + "loss": 0.8299, + "step": 242 + }, + { + "epoch": 0.020707285896889645, + "grad_norm": 14.4375, + "learning_rate": 2.4997690483661615e-06, + "loss": 0.9563, + "step": 243 + }, + { + "epoch": 0.020792501065189605, + "grad_norm": 11.4375, + "learning_rate": 2.4997658070805684e-06, + "loss": 0.5569, + "step": 244 + }, + { + "epoch": 0.020877716233489562, + "grad_norm": 14.75, + "learning_rate": 2.4997625432104504e-06, + "loss": 0.9559, + "step": 245 + }, + { + "epoch": 0.02096293140178952, + "grad_norm": 45.75, + "learning_rate": 2.499759256755866e-06, + "loss": 0.9006, + "step": 246 + }, + { + "epoch": 0.021048146570089476, + "grad_norm": 18.875, + "learning_rate": 2.499755947716876e-06, + "loss": 1.1924, + "step": 247 + }, + { + "epoch": 0.021133361738389433, + "grad_norm": 15.75, + "learning_rate": 2.4997526160935385e-06, + "loss": 0.8332, + "step": 248 + }, + { + "epoch": 0.02121857690668939, + "grad_norm": 19.375, + "learning_rate": 2.499749261885915e-06, + "loss": 1.0387, + "step": 249 + }, + { + "epoch": 0.021303792074989347, + "grad_norm": 18.625, + "learning_rate": 2.4997458850940654e-06, + "loss": 1.1405, + "step": 250 + }, + { + "epoch": 0.021389007243289307, + "grad_norm": 10.3125, + "learning_rate": 2.499742485718051e-06, + "loss": 0.5173, + "step": 251 + }, + { + "epoch": 0.021474222411589264, + "grad_norm": 13.625, + "learning_rate": 2.4997390637579326e-06, + "loss": 0.9637, + "step": 252 + }, + { + "epoch": 0.02155943757988922, + "grad_norm": 21.75, + "learning_rate": 2.4997356192137734e-06, + "loss": 1.2523, + "step": 253 + }, + { + "epoch": 0.021644652748189178, + "grad_norm": 17.875, + "learning_rate": 2.4997321520856347e-06, + "loss": 1.0835, + "step": 254 + }, + { + "epoch": 0.021729867916489135, + "grad_norm": 14.6875, + "learning_rate": 2.4997286623735788e-06, + "loss": 0.8121, + "step": 255 + }, + { + "epoch": 0.02181508308478909, + "grad_norm": 17.125, + "learning_rate": 2.49972515007767e-06, + "loss": 1.1225, + "step": 256 + }, + { + "epoch": 0.02190029825308905, + "grad_norm": 9.75, + "learning_rate": 2.499721615197971e-06, + "loss": 0.3624, + "step": 257 + }, + { + "epoch": 0.02198551342138901, + "grad_norm": 22.125, + "learning_rate": 2.499718057734545e-06, + "loss": 1.3517, + "step": 258 + }, + { + "epoch": 0.022070728589688966, + "grad_norm": 23.625, + "learning_rate": 2.4997144776874577e-06, + "loss": 0.867, + "step": 259 + }, + { + "epoch": 0.022155943757988922, + "grad_norm": 21.75, + "learning_rate": 2.499710875056773e-06, + "loss": 1.0509, + "step": 260 + }, + { + "epoch": 0.02224115892628888, + "grad_norm": 39.75, + "learning_rate": 2.4997072498425554e-06, + "loss": 1.6689, + "step": 261 + }, + { + "epoch": 0.022326374094588836, + "grad_norm": 36.0, + "learning_rate": 2.4997036020448717e-06, + "loss": 1.2132, + "step": 262 + }, + { + "epoch": 0.022411589262888793, + "grad_norm": 16.25, + "learning_rate": 2.4996999316637874e-06, + "loss": 0.7928, + "step": 263 + }, + { + "epoch": 0.02249680443118875, + "grad_norm": 12.9375, + "learning_rate": 2.4996962386993685e-06, + "loss": 0.8249, + "step": 264 + }, + { + "epoch": 0.02258201959948871, + "grad_norm": 16.625, + "learning_rate": 2.4996925231516816e-06, + "loss": 0.744, + "step": 265 + }, + { + "epoch": 0.022667234767788667, + "grad_norm": 11.25, + "learning_rate": 2.4996887850207945e-06, + "loss": 0.7313, + "step": 266 + }, + { + "epoch": 0.022752449936088624, + "grad_norm": 9.0625, + "learning_rate": 2.4996850243067744e-06, + "loss": 0.3756, + "step": 267 + }, + { + "epoch": 0.02283766510438858, + "grad_norm": 12.1875, + "learning_rate": 2.4996812410096892e-06, + "loss": 0.5983, + "step": 268 + }, + { + "epoch": 0.022922880272688538, + "grad_norm": 11.875, + "learning_rate": 2.499677435129607e-06, + "loss": 0.4525, + "step": 269 + }, + { + "epoch": 0.023008095440988495, + "grad_norm": 43.5, + "learning_rate": 2.4996736066665973e-06, + "loss": 1.4105, + "step": 270 + }, + { + "epoch": 0.023093310609288452, + "grad_norm": 12.375, + "learning_rate": 2.499669755620729e-06, + "loss": 1.0042, + "step": 271 + }, + { + "epoch": 0.023178525777588412, + "grad_norm": 14.3125, + "learning_rate": 2.499665881992071e-06, + "loss": 1.0983, + "step": 272 + }, + { + "epoch": 0.02326374094588837, + "grad_norm": 17.875, + "learning_rate": 2.4996619857806946e-06, + "loss": 1.0489, + "step": 273 + }, + { + "epoch": 0.023348956114188326, + "grad_norm": 14.25, + "learning_rate": 2.499658066986669e-06, + "loss": 0.9086, + "step": 274 + }, + { + "epoch": 0.023434171282488283, + "grad_norm": 29.5, + "learning_rate": 2.499654125610065e-06, + "loss": 1.3386, + "step": 275 + }, + { + "epoch": 0.02351938645078824, + "grad_norm": 16.75, + "learning_rate": 2.4996501616509553e-06, + "loss": 1.0422, + "step": 276 + }, + { + "epoch": 0.023604601619088197, + "grad_norm": 13.375, + "learning_rate": 2.4996461751094105e-06, + "loss": 0.8694, + "step": 277 + }, + { + "epoch": 0.023689816787388154, + "grad_norm": 16.125, + "learning_rate": 2.4996421659855022e-06, + "loss": 0.9185, + "step": 278 + }, + { + "epoch": 0.023775031955688114, + "grad_norm": 16.25, + "learning_rate": 2.4996381342793034e-06, + "loss": 0.7571, + "step": 279 + }, + { + "epoch": 0.02386024712398807, + "grad_norm": 16.75, + "learning_rate": 2.4996340799908873e-06, + "loss": 0.8722, + "step": 280 + }, + { + "epoch": 0.023945462292288028, + "grad_norm": 12.0, + "learning_rate": 2.4996300031203264e-06, + "loss": 0.6721, + "step": 281 + }, + { + "epoch": 0.024030677460587985, + "grad_norm": 21.25, + "learning_rate": 2.499625903667695e-06, + "loss": 0.7462, + "step": 282 + }, + { + "epoch": 0.02411589262888794, + "grad_norm": 11.5, + "learning_rate": 2.499621781633067e-06, + "loss": 0.5317, + "step": 283 + }, + { + "epoch": 0.0242011077971879, + "grad_norm": 13.9375, + "learning_rate": 2.4996176370165164e-06, + "loss": 0.7925, + "step": 284 + }, + { + "epoch": 0.024286322965487855, + "grad_norm": 17.125, + "learning_rate": 2.4996134698181184e-06, + "loss": 1.1314, + "step": 285 + }, + { + "epoch": 0.024371538133787816, + "grad_norm": 14.3125, + "learning_rate": 2.499609280037949e-06, + "loss": 0.8216, + "step": 286 + }, + { + "epoch": 0.024456753302087773, + "grad_norm": 19.75, + "learning_rate": 2.499605067676083e-06, + "loss": 1.0857, + "step": 287 + }, + { + "epoch": 0.02454196847038773, + "grad_norm": 12.0625, + "learning_rate": 2.499600832732597e-06, + "loss": 0.6202, + "step": 288 + }, + { + "epoch": 0.024627183638687686, + "grad_norm": 18.375, + "learning_rate": 2.4995965752075673e-06, + "loss": 1.033, + "step": 289 + }, + { + "epoch": 0.024712398806987643, + "grad_norm": 13.4375, + "learning_rate": 2.499592295101071e-06, + "loss": 0.648, + "step": 290 + }, + { + "epoch": 0.0247976139752876, + "grad_norm": 18.75, + "learning_rate": 2.499587992413185e-06, + "loss": 1.016, + "step": 291 + }, + { + "epoch": 0.024882829143587557, + "grad_norm": 14.8125, + "learning_rate": 2.499583667143988e-06, + "loss": 0.5957, + "step": 292 + }, + { + "epoch": 0.024968044311887518, + "grad_norm": 21.875, + "learning_rate": 2.499579319293557e-06, + "loss": 0.9271, + "step": 293 + }, + { + "epoch": 0.025053259480187474, + "grad_norm": 13.5625, + "learning_rate": 2.4995749488619716e-06, + "loss": 0.7341, + "step": 294 + }, + { + "epoch": 0.02513847464848743, + "grad_norm": 14.375, + "learning_rate": 2.49957055584931e-06, + "loss": 0.8839, + "step": 295 + }, + { + "epoch": 0.025223689816787388, + "grad_norm": 22.25, + "learning_rate": 2.499566140255652e-06, + "loss": 1.6852, + "step": 296 + }, + { + "epoch": 0.025308904985087345, + "grad_norm": 26.5, + "learning_rate": 2.4995617020810775e-06, + "loss": 0.7103, + "step": 297 + }, + { + "epoch": 0.025394120153387302, + "grad_norm": 16.125, + "learning_rate": 2.4995572413256663e-06, + "loss": 0.8754, + "step": 298 + }, + { + "epoch": 0.02547933532168726, + "grad_norm": 17.125, + "learning_rate": 2.4995527579894994e-06, + "loss": 0.9294, + "step": 299 + }, + { + "epoch": 0.02556455048998722, + "grad_norm": 14.0625, + "learning_rate": 2.4995482520726573e-06, + "loss": 0.6675, + "step": 300 + }, + { + "epoch": 0.025649765658287176, + "grad_norm": 18.375, + "learning_rate": 2.499543723575222e-06, + "loss": 0.9184, + "step": 301 + }, + { + "epoch": 0.025734980826587133, + "grad_norm": 21.125, + "learning_rate": 2.499539172497275e-06, + "loss": 0.9095, + "step": 302 + }, + { + "epoch": 0.02582019599488709, + "grad_norm": 13.625, + "learning_rate": 2.499534598838899e-06, + "loss": 0.9908, + "step": 303 + }, + { + "epoch": 0.025905411163187047, + "grad_norm": 17.5, + "learning_rate": 2.4995300026001764e-06, + "loss": 1.0894, + "step": 304 + }, + { + "epoch": 0.025990626331487004, + "grad_norm": 12.125, + "learning_rate": 2.4995253837811895e-06, + "loss": 0.7298, + "step": 305 + }, + { + "epoch": 0.02607584149978696, + "grad_norm": 16.5, + "learning_rate": 2.499520742382023e-06, + "loss": 0.8227, + "step": 306 + }, + { + "epoch": 0.02616105666808692, + "grad_norm": 22.375, + "learning_rate": 2.4995160784027596e-06, + "loss": 1.1159, + "step": 307 + }, + { + "epoch": 0.026246271836386878, + "grad_norm": 12.75, + "learning_rate": 2.499511391843484e-06, + "loss": 0.5054, + "step": 308 + }, + { + "epoch": 0.026331487004686835, + "grad_norm": 13.6875, + "learning_rate": 2.4995066827042823e-06, + "loss": 0.8779, + "step": 309 + }, + { + "epoch": 0.026416702172986792, + "grad_norm": 11.0625, + "learning_rate": 2.4995019509852374e-06, + "loss": 0.5208, + "step": 310 + }, + { + "epoch": 0.02650191734128675, + "grad_norm": 15.75, + "learning_rate": 2.4994971966864363e-06, + "loss": 0.9356, + "step": 311 + }, + { + "epoch": 0.026587132509586706, + "grad_norm": 21.125, + "learning_rate": 2.499492419807964e-06, + "loss": 0.8451, + "step": 312 + }, + { + "epoch": 0.026672347677886662, + "grad_norm": 15.375, + "learning_rate": 2.499487620349907e-06, + "loss": 0.9195, + "step": 313 + }, + { + "epoch": 0.026757562846186623, + "grad_norm": 15.5625, + "learning_rate": 2.4994827983123527e-06, + "loss": 0.8463, + "step": 314 + }, + { + "epoch": 0.02684277801448658, + "grad_norm": 15.9375, + "learning_rate": 2.4994779536953877e-06, + "loss": 0.8846, + "step": 315 + }, + { + "epoch": 0.026927993182786537, + "grad_norm": 23.625, + "learning_rate": 2.4994730864990997e-06, + "loss": 1.1352, + "step": 316 + }, + { + "epoch": 0.027013208351086494, + "grad_norm": 27.5, + "learning_rate": 2.4994681967235764e-06, + "loss": 1.2751, + "step": 317 + }, + { + "epoch": 0.02709842351938645, + "grad_norm": 13.25, + "learning_rate": 2.4994632843689063e-06, + "loss": 0.6847, + "step": 318 + }, + { + "epoch": 0.027183638687686407, + "grad_norm": 29.25, + "learning_rate": 2.4994583494351784e-06, + "loss": 1.1252, + "step": 319 + }, + { + "epoch": 0.027268853855986364, + "grad_norm": 24.625, + "learning_rate": 2.4994533919224815e-06, + "loss": 0.9692, + "step": 320 + }, + { + "epoch": 0.027354069024286325, + "grad_norm": 17.25, + "learning_rate": 2.4994484118309056e-06, + "loss": 0.6664, + "step": 321 + }, + { + "epoch": 0.02743928419258628, + "grad_norm": 20.375, + "learning_rate": 2.4994434091605402e-06, + "loss": 0.9731, + "step": 322 + }, + { + "epoch": 0.02752449936088624, + "grad_norm": 20.0, + "learning_rate": 2.4994383839114762e-06, + "loss": 0.8914, + "step": 323 + }, + { + "epoch": 0.027609714529186195, + "grad_norm": 13.9375, + "learning_rate": 2.499433336083804e-06, + "loss": 0.7852, + "step": 324 + }, + { + "epoch": 0.027694929697486152, + "grad_norm": 13.6875, + "learning_rate": 2.4994282656776154e-06, + "loss": 0.5759, + "step": 325 + }, + { + "epoch": 0.02778014486578611, + "grad_norm": 21.5, + "learning_rate": 2.4994231726930012e-06, + "loss": 0.8376, + "step": 326 + }, + { + "epoch": 0.027865360034086066, + "grad_norm": 21.125, + "learning_rate": 2.4994180571300546e-06, + "loss": 1.1522, + "step": 327 + }, + { + "epoch": 0.027950575202386026, + "grad_norm": 17.75, + "learning_rate": 2.4994129189888667e-06, + "loss": 0.6931, + "step": 328 + }, + { + "epoch": 0.028035790370685983, + "grad_norm": 15.5625, + "learning_rate": 2.4994077582695307e-06, + "loss": 0.9413, + "step": 329 + }, + { + "epoch": 0.02812100553898594, + "grad_norm": 11.9375, + "learning_rate": 2.4994025749721405e-06, + "loss": 0.7517, + "step": 330 + }, + { + "epoch": 0.028206220707285897, + "grad_norm": 15.0625, + "learning_rate": 2.49939736909679e-06, + "loss": 0.622, + "step": 331 + }, + { + "epoch": 0.028291435875585854, + "grad_norm": 20.625, + "learning_rate": 2.499392140643572e-06, + "loss": 1.185, + "step": 332 + }, + { + "epoch": 0.02837665104388581, + "grad_norm": 22.5, + "learning_rate": 2.4993868896125818e-06, + "loss": 1.1348, + "step": 333 + }, + { + "epoch": 0.028461866212185768, + "grad_norm": 13.0625, + "learning_rate": 2.4993816160039137e-06, + "loss": 0.5254, + "step": 334 + }, + { + "epoch": 0.028547081380485728, + "grad_norm": 15.6875, + "learning_rate": 2.499376319817664e-06, + "loss": 1.1253, + "step": 335 + }, + { + "epoch": 0.028632296548785685, + "grad_norm": 17.625, + "learning_rate": 2.499371001053928e-06, + "loss": 1.0347, + "step": 336 + }, + { + "epoch": 0.028717511717085642, + "grad_norm": 11.25, + "learning_rate": 2.499365659712801e-06, + "loss": 0.5701, + "step": 337 + }, + { + "epoch": 0.0288027268853856, + "grad_norm": 24.875, + "learning_rate": 2.4993602957943808e-06, + "loss": 1.1916, + "step": 338 + }, + { + "epoch": 0.028887942053685556, + "grad_norm": 12.0625, + "learning_rate": 2.4993549092987633e-06, + "loss": 0.6434, + "step": 339 + }, + { + "epoch": 0.028973157221985513, + "grad_norm": 14.625, + "learning_rate": 2.4993495002260463e-06, + "loss": 0.8653, + "step": 340 + }, + { + "epoch": 0.02905837239028547, + "grad_norm": 15.375, + "learning_rate": 2.4993440685763277e-06, + "loss": 0.5834, + "step": 341 + }, + { + "epoch": 0.029143587558585426, + "grad_norm": 11.9375, + "learning_rate": 2.4993386143497056e-06, + "loss": 0.5358, + "step": 342 + }, + { + "epoch": 0.029228802726885387, + "grad_norm": 19.75, + "learning_rate": 2.499333137546278e-06, + "loss": 0.9414, + "step": 343 + }, + { + "epoch": 0.029314017895185344, + "grad_norm": 21.75, + "learning_rate": 2.499327638166145e-06, + "loss": 1.1704, + "step": 344 + }, + { + "epoch": 0.0293992330634853, + "grad_norm": 22.875, + "learning_rate": 2.499322116209405e-06, + "loss": 1.2015, + "step": 345 + }, + { + "epoch": 0.029484448231785258, + "grad_norm": 14.5, + "learning_rate": 2.4993165716761576e-06, + "loss": 0.8069, + "step": 346 + }, + { + "epoch": 0.029569663400085214, + "grad_norm": 11.9375, + "learning_rate": 2.4993110045665037e-06, + "loss": 0.4925, + "step": 347 + }, + { + "epoch": 0.02965487856838517, + "grad_norm": 17.0, + "learning_rate": 2.499305414880544e-06, + "loss": 0.8036, + "step": 348 + }, + { + "epoch": 0.029740093736685128, + "grad_norm": 13.125, + "learning_rate": 2.499299802618379e-06, + "loss": 0.6162, + "step": 349 + }, + { + "epoch": 0.02982530890498509, + "grad_norm": 16.5, + "learning_rate": 2.49929416778011e-06, + "loss": 0.7732, + "step": 350 + }, + { + "epoch": 0.029910524073285046, + "grad_norm": 23.0, + "learning_rate": 2.4992885103658394e-06, + "loss": 1.1847, + "step": 351 + }, + { + "epoch": 0.029995739241585002, + "grad_norm": 13.25, + "learning_rate": 2.4992828303756694e-06, + "loss": 0.6363, + "step": 352 + }, + { + "epoch": 0.03008095440988496, + "grad_norm": 11.25, + "learning_rate": 2.4992771278097017e-06, + "loss": 0.486, + "step": 353 + }, + { + "epoch": 0.030166169578184916, + "grad_norm": 14.125, + "learning_rate": 2.499271402668041e-06, + "loss": 1.2398, + "step": 354 + }, + { + "epoch": 0.030251384746484873, + "grad_norm": 13.4375, + "learning_rate": 2.499265654950789e-06, + "loss": 0.8022, + "step": 355 + }, + { + "epoch": 0.03033659991478483, + "grad_norm": 19.75, + "learning_rate": 2.4992598846580507e-06, + "loss": 0.9641, + "step": 356 + }, + { + "epoch": 0.03042181508308479, + "grad_norm": 16.125, + "learning_rate": 2.4992540917899298e-06, + "loss": 0.7756, + "step": 357 + }, + { + "epoch": 0.030507030251384747, + "grad_norm": 16.125, + "learning_rate": 2.499248276346532e-06, + "loss": 1.0406, + "step": 358 + }, + { + "epoch": 0.030592245419684704, + "grad_norm": 14.0625, + "learning_rate": 2.499242438327961e-06, + "loss": 0.6911, + "step": 359 + }, + { + "epoch": 0.03067746058798466, + "grad_norm": 18.125, + "learning_rate": 2.499236577734323e-06, + "loss": 0.5916, + "step": 360 + }, + { + "epoch": 0.030762675756284618, + "grad_norm": 16.0, + "learning_rate": 2.499230694565724e-06, + "loss": 0.9698, + "step": 361 + }, + { + "epoch": 0.030847890924584575, + "grad_norm": 15.5, + "learning_rate": 2.49922478882227e-06, + "loss": 1.1002, + "step": 362 + }, + { + "epoch": 0.030933106092884532, + "grad_norm": 15.3125, + "learning_rate": 2.499218860504068e-06, + "loss": 0.8531, + "step": 363 + }, + { + "epoch": 0.031018321261184492, + "grad_norm": 22.5, + "learning_rate": 2.4992129096112248e-06, + "loss": 1.0199, + "step": 364 + }, + { + "epoch": 0.03110353642948445, + "grad_norm": 15.375, + "learning_rate": 2.499206936143848e-06, + "loss": 0.65, + "step": 365 + }, + { + "epoch": 0.031188751597784406, + "grad_norm": 19.5, + "learning_rate": 2.499200940102046e-06, + "loss": 0.8134, + "step": 366 + }, + { + "epoch": 0.03127396676608436, + "grad_norm": 14.0625, + "learning_rate": 2.499194921485927e-06, + "loss": 0.9027, + "step": 367 + }, + { + "epoch": 0.03135918193438432, + "grad_norm": 13.0625, + "learning_rate": 2.4991888802955993e-06, + "loss": 0.5978, + "step": 368 + }, + { + "epoch": 0.03144439710268428, + "grad_norm": 11.25, + "learning_rate": 2.4991828165311723e-06, + "loss": 0.5651, + "step": 369 + }, + { + "epoch": 0.031529612270984234, + "grad_norm": 20.625, + "learning_rate": 2.499176730192756e-06, + "loss": 1.0047, + "step": 370 + }, + { + "epoch": 0.03161482743928419, + "grad_norm": 19.25, + "learning_rate": 2.4991706212804596e-06, + "loss": 0.9238, + "step": 371 + }, + { + "epoch": 0.03170004260758415, + "grad_norm": 12.5625, + "learning_rate": 2.4991644897943945e-06, + "loss": 0.6923, + "step": 372 + }, + { + "epoch": 0.031785257775884104, + "grad_norm": 13.4375, + "learning_rate": 2.4991583357346705e-06, + "loss": 0.7106, + "step": 373 + }, + { + "epoch": 0.03187047294418407, + "grad_norm": 11.625, + "learning_rate": 2.4991521591013995e-06, + "loss": 0.4967, + "step": 374 + }, + { + "epoch": 0.031955688112484025, + "grad_norm": 15.8125, + "learning_rate": 2.499145959894693e-06, + "loss": 0.9125, + "step": 375 + }, + { + "epoch": 0.03204090328078398, + "grad_norm": 17.0, + "learning_rate": 2.499139738114662e-06, + "loss": 0.9366, + "step": 376 + }, + { + "epoch": 0.03212611844908394, + "grad_norm": 17.0, + "learning_rate": 2.4991334937614206e-06, + "loss": 1.1503, + "step": 377 + }, + { + "epoch": 0.032211333617383896, + "grad_norm": 13.5625, + "learning_rate": 2.4991272268350807e-06, + "loss": 0.7203, + "step": 378 + }, + { + "epoch": 0.03229654878568385, + "grad_norm": 19.125, + "learning_rate": 2.4991209373357558e-06, + "loss": 0.6034, + "step": 379 + }, + { + "epoch": 0.03238176395398381, + "grad_norm": 11.375, + "learning_rate": 2.4991146252635593e-06, + "loss": 0.5002, + "step": 380 + }, + { + "epoch": 0.032466979122283766, + "grad_norm": 32.25, + "learning_rate": 2.4991082906186057e-06, + "loss": 1.7323, + "step": 381 + }, + { + "epoch": 0.03255219429058372, + "grad_norm": 25.25, + "learning_rate": 2.499101933401009e-06, + "loss": 1.4725, + "step": 382 + }, + { + "epoch": 0.03263740945888368, + "grad_norm": 14.4375, + "learning_rate": 2.4990955536108846e-06, + "loss": 0.7112, + "step": 383 + }, + { + "epoch": 0.03272262462718364, + "grad_norm": 20.75, + "learning_rate": 2.499089151248347e-06, + "loss": 0.9326, + "step": 384 + }, + { + "epoch": 0.032807839795483594, + "grad_norm": 19.625, + "learning_rate": 2.4990827263135125e-06, + "loss": 1.218, + "step": 385 + }, + { + "epoch": 0.03289305496378355, + "grad_norm": 14.0625, + "learning_rate": 2.499076278806498e-06, + "loss": 0.8987, + "step": 386 + }, + { + "epoch": 0.03297827013208351, + "grad_norm": 13.375, + "learning_rate": 2.499069808727418e-06, + "loss": 0.5982, + "step": 387 + }, + { + "epoch": 0.03306348530038347, + "grad_norm": 21.375, + "learning_rate": 2.4990633160763904e-06, + "loss": 1.0252, + "step": 388 + }, + { + "epoch": 0.03314870046868343, + "grad_norm": 13.8125, + "learning_rate": 2.499056800853533e-06, + "loss": 0.869, + "step": 389 + }, + { + "epoch": 0.033233915636983385, + "grad_norm": 20.75, + "learning_rate": 2.4990502630589637e-06, + "loss": 0.5836, + "step": 390 + }, + { + "epoch": 0.03331913080528334, + "grad_norm": 17.625, + "learning_rate": 2.4990437026927993e-06, + "loss": 1.0526, + "step": 391 + }, + { + "epoch": 0.0334043459735833, + "grad_norm": 11.5625, + "learning_rate": 2.49903711975516e-06, + "loss": 0.5121, + "step": 392 + }, + { + "epoch": 0.033489561141883256, + "grad_norm": 17.125, + "learning_rate": 2.4990305142461635e-06, + "loss": 0.7733, + "step": 393 + }, + { + "epoch": 0.03357477631018321, + "grad_norm": 12.875, + "learning_rate": 2.4990238861659293e-06, + "loss": 0.6252, + "step": 394 + }, + { + "epoch": 0.03365999147848317, + "grad_norm": 13.5, + "learning_rate": 2.499017235514578e-06, + "loss": 0.7226, + "step": 395 + }, + { + "epoch": 0.03374520664678313, + "grad_norm": 20.875, + "learning_rate": 2.499010562292229e-06, + "loss": 1.1063, + "step": 396 + }, + { + "epoch": 0.033830421815083084, + "grad_norm": 11.125, + "learning_rate": 2.4990038664990036e-06, + "loss": 0.523, + "step": 397 + }, + { + "epoch": 0.03391563698338304, + "grad_norm": 20.25, + "learning_rate": 2.4989971481350217e-06, + "loss": 1.1988, + "step": 398 + }, + { + "epoch": 0.034000852151683, + "grad_norm": 12.875, + "learning_rate": 2.4989904072004056e-06, + "loss": 0.6412, + "step": 399 + }, + { + "epoch": 0.034086067319982954, + "grad_norm": 21.5, + "learning_rate": 2.498983643695277e-06, + "loss": 1.3073, + "step": 400 + }, + { + "epoch": 0.03417128248828291, + "grad_norm": 14.6875, + "learning_rate": 2.4989768576197583e-06, + "loss": 0.8451, + "step": 401 + }, + { + "epoch": 0.034256497656582875, + "grad_norm": 13.875, + "learning_rate": 2.4989700489739713e-06, + "loss": 0.6877, + "step": 402 + }, + { + "epoch": 0.03434171282488283, + "grad_norm": 10.75, + "learning_rate": 2.49896321775804e-06, + "loss": 0.4754, + "step": 403 + }, + { + "epoch": 0.03442692799318279, + "grad_norm": 17.875, + "learning_rate": 2.498956363972087e-06, + "loss": 0.9302, + "step": 404 + }, + { + "epoch": 0.034512143161482746, + "grad_norm": 31.25, + "learning_rate": 2.4989494876162368e-06, + "loss": 1.2954, + "step": 405 + }, + { + "epoch": 0.0345973583297827, + "grad_norm": 12.125, + "learning_rate": 2.4989425886906133e-06, + "loss": 0.5317, + "step": 406 + }, + { + "epoch": 0.03468257349808266, + "grad_norm": 12.3125, + "learning_rate": 2.4989356671953417e-06, + "loss": 0.4213, + "step": 407 + }, + { + "epoch": 0.03476778866638262, + "grad_norm": 20.375, + "learning_rate": 2.4989287231305464e-06, + "loss": 0.9238, + "step": 408 + }, + { + "epoch": 0.034853003834682574, + "grad_norm": 13.6875, + "learning_rate": 2.498921756496353e-06, + "loss": 0.5157, + "step": 409 + }, + { + "epoch": 0.03493821900298253, + "grad_norm": 17.5, + "learning_rate": 2.4989147672928877e-06, + "loss": 1.048, + "step": 410 + }, + { + "epoch": 0.03502343417128249, + "grad_norm": 18.625, + "learning_rate": 2.4989077555202766e-06, + "loss": 0.9328, + "step": 411 + }, + { + "epoch": 0.035108649339582444, + "grad_norm": 17.75, + "learning_rate": 2.498900721178647e-06, + "loss": 0.8179, + "step": 412 + }, + { + "epoch": 0.0351938645078824, + "grad_norm": 18.75, + "learning_rate": 2.498893664268125e-06, + "loss": 1.1035, + "step": 413 + }, + { + "epoch": 0.03527907967618236, + "grad_norm": 13.25, + "learning_rate": 2.4988865847888388e-06, + "loss": 0.7153, + "step": 414 + }, + { + "epoch": 0.035364294844482315, + "grad_norm": 18.375, + "learning_rate": 2.4988794827409165e-06, + "loss": 0.9215, + "step": 415 + }, + { + "epoch": 0.03544951001278227, + "grad_norm": 13.0, + "learning_rate": 2.4988723581244857e-06, + "loss": 0.5237, + "step": 416 + }, + { + "epoch": 0.035534725181082236, + "grad_norm": 13.125, + "learning_rate": 2.4988652109396756e-06, + "loss": 0.8331, + "step": 417 + }, + { + "epoch": 0.03561994034938219, + "grad_norm": 16.0, + "learning_rate": 2.498858041186615e-06, + "loss": 1.0079, + "step": 418 + }, + { + "epoch": 0.03570515551768215, + "grad_norm": 15.1875, + "learning_rate": 2.4988508488654342e-06, + "loss": 0.6557, + "step": 419 + }, + { + "epoch": 0.035790370685982106, + "grad_norm": 16.625, + "learning_rate": 2.4988436339762625e-06, + "loss": 1.0191, + "step": 420 + }, + { + "epoch": 0.03587558585428206, + "grad_norm": 13.1875, + "learning_rate": 2.4988363965192308e-06, + "loss": 0.5937, + "step": 421 + }, + { + "epoch": 0.03596080102258202, + "grad_norm": 18.0, + "learning_rate": 2.4988291364944694e-06, + "loss": 1.2424, + "step": 422 + }, + { + "epoch": 0.03604601619088198, + "grad_norm": 21.625, + "learning_rate": 2.4988218539021097e-06, + "loss": 0.9282, + "step": 423 + }, + { + "epoch": 0.036131231359181934, + "grad_norm": 13.0625, + "learning_rate": 2.4988145487422833e-06, + "loss": 0.4837, + "step": 424 + }, + { + "epoch": 0.03621644652748189, + "grad_norm": 16.5, + "learning_rate": 2.4988072210151224e-06, + "loss": 0.8311, + "step": 425 + }, + { + "epoch": 0.03630166169578185, + "grad_norm": 12.6875, + "learning_rate": 2.4987998707207592e-06, + "loss": 0.9997, + "step": 426 + }, + { + "epoch": 0.036386876864081805, + "grad_norm": 13.5625, + "learning_rate": 2.4987924978593263e-06, + "loss": 0.8165, + "step": 427 + }, + { + "epoch": 0.03647209203238176, + "grad_norm": 12.0625, + "learning_rate": 2.4987851024309577e-06, + "loss": 0.5936, + "step": 428 + }, + { + "epoch": 0.03655730720068172, + "grad_norm": 14.375, + "learning_rate": 2.4987776844357862e-06, + "loss": 0.7518, + "step": 429 + }, + { + "epoch": 0.036642522368981675, + "grad_norm": 16.125, + "learning_rate": 2.4987702438739464e-06, + "loss": 0.8977, + "step": 430 + }, + { + "epoch": 0.03672773753728164, + "grad_norm": 38.75, + "learning_rate": 2.498762780745572e-06, + "loss": 1.6056, + "step": 431 + }, + { + "epoch": 0.036812952705581596, + "grad_norm": 20.625, + "learning_rate": 2.4987552950507994e-06, + "loss": 1.0696, + "step": 432 + }, + { + "epoch": 0.03689816787388155, + "grad_norm": 17.75, + "learning_rate": 2.498747786789762e-06, + "loss": 1.1169, + "step": 433 + }, + { + "epoch": 0.03698338304218151, + "grad_norm": 10.375, + "learning_rate": 2.4987402559625965e-06, + "loss": 0.4596, + "step": 434 + }, + { + "epoch": 0.03706859821048147, + "grad_norm": 14.25, + "learning_rate": 2.4987327025694392e-06, + "loss": 0.9313, + "step": 435 + }, + { + "epoch": 0.037153813378781424, + "grad_norm": 12.6875, + "learning_rate": 2.4987251266104263e-06, + "loss": 0.7626, + "step": 436 + }, + { + "epoch": 0.03723902854708138, + "grad_norm": 13.8125, + "learning_rate": 2.4987175280856947e-06, + "loss": 0.5739, + "step": 437 + }, + { + "epoch": 0.03732424371538134, + "grad_norm": 30.625, + "learning_rate": 2.4987099069953815e-06, + "loss": 0.6925, + "step": 438 + }, + { + "epoch": 0.037409458883681294, + "grad_norm": 13.6875, + "learning_rate": 2.498702263339625e-06, + "loss": 0.3508, + "step": 439 + }, + { + "epoch": 0.03749467405198125, + "grad_norm": 15.4375, + "learning_rate": 2.4986945971185623e-06, + "loss": 0.9168, + "step": 440 + }, + { + "epoch": 0.03757988922028121, + "grad_norm": 27.125, + "learning_rate": 2.4986869083323324e-06, + "loss": 0.8307, + "step": 441 + }, + { + "epoch": 0.037665104388581165, + "grad_norm": 13.375, + "learning_rate": 2.498679196981075e-06, + "loss": 0.7669, + "step": 442 + }, + { + "epoch": 0.03775031955688112, + "grad_norm": 19.25, + "learning_rate": 2.4986714630649287e-06, + "loss": 1.4351, + "step": 443 + }, + { + "epoch": 0.03783553472518108, + "grad_norm": 16.375, + "learning_rate": 2.4986637065840336e-06, + "loss": 1.2245, + "step": 444 + }, + { + "epoch": 0.03792074989348104, + "grad_norm": 13.3125, + "learning_rate": 2.4986559275385294e-06, + "loss": 0.7966, + "step": 445 + }, + { + "epoch": 0.038005965061781, + "grad_norm": 12.8125, + "learning_rate": 2.4986481259285574e-06, + "loss": 0.6735, + "step": 446 + }, + { + "epoch": 0.03809118023008096, + "grad_norm": 13.1875, + "learning_rate": 2.4986403017542575e-06, + "loss": 0.8942, + "step": 447 + }, + { + "epoch": 0.03817639539838091, + "grad_norm": 14.75, + "learning_rate": 2.498632455015772e-06, + "loss": 0.8208, + "step": 448 + }, + { + "epoch": 0.03826161056668087, + "grad_norm": 14.0625, + "learning_rate": 2.4986245857132426e-06, + "loss": 0.6493, + "step": 449 + }, + { + "epoch": 0.03834682573498083, + "grad_norm": 15.1875, + "learning_rate": 2.4986166938468114e-06, + "loss": 0.9837, + "step": 450 + }, + { + "epoch": 0.038432040903280784, + "grad_norm": 16.0, + "learning_rate": 2.4986087794166206e-06, + "loss": 0.8642, + "step": 451 + }, + { + "epoch": 0.03851725607158074, + "grad_norm": 15.375, + "learning_rate": 2.4986008424228143e-06, + "loss": 0.9541, + "step": 452 + }, + { + "epoch": 0.0386024712398807, + "grad_norm": 13.125, + "learning_rate": 2.4985928828655346e-06, + "loss": 0.7419, + "step": 453 + }, + { + "epoch": 0.038687686408180655, + "grad_norm": 17.375, + "learning_rate": 2.498584900744926e-06, + "loss": 0.8758, + "step": 454 + }, + { + "epoch": 0.03877290157648061, + "grad_norm": 13.875, + "learning_rate": 2.498576896061133e-06, + "loss": 0.7497, + "step": 455 + }, + { + "epoch": 0.03885811674478057, + "grad_norm": 15.5625, + "learning_rate": 2.4985688688142996e-06, + "loss": 0.9892, + "step": 456 + }, + { + "epoch": 0.038943331913080526, + "grad_norm": 12.0625, + "learning_rate": 2.4985608190045714e-06, + "loss": 0.6137, + "step": 457 + }, + { + "epoch": 0.03902854708138048, + "grad_norm": 10.625, + "learning_rate": 2.4985527466320937e-06, + "loss": 0.4528, + "step": 458 + }, + { + "epoch": 0.039113762249680446, + "grad_norm": 10.6875, + "learning_rate": 2.4985446516970117e-06, + "loss": 0.5635, + "step": 459 + }, + { + "epoch": 0.0391989774179804, + "grad_norm": 16.625, + "learning_rate": 2.4985365341994734e-06, + "loss": 0.9851, + "step": 460 + }, + { + "epoch": 0.03928419258628036, + "grad_norm": 24.625, + "learning_rate": 2.498528394139624e-06, + "loss": 1.373, + "step": 461 + }, + { + "epoch": 0.03936940775458032, + "grad_norm": 17.875, + "learning_rate": 2.498520231517611e-06, + "loss": 0.8514, + "step": 462 + }, + { + "epoch": 0.039454622922880274, + "grad_norm": 15.5, + "learning_rate": 2.4985120463335816e-06, + "loss": 1.1372, + "step": 463 + }, + { + "epoch": 0.03953983809118023, + "grad_norm": 11.1875, + "learning_rate": 2.498503838587684e-06, + "loss": 0.4535, + "step": 464 + }, + { + "epoch": 0.03962505325948019, + "grad_norm": 17.875, + "learning_rate": 2.4984956082800673e-06, + "loss": 0.8392, + "step": 465 + }, + { + "epoch": 0.039710268427780145, + "grad_norm": 18.75, + "learning_rate": 2.498487355410879e-06, + "loss": 1.1626, + "step": 466 + }, + { + "epoch": 0.0397954835960801, + "grad_norm": 16.5, + "learning_rate": 2.498479079980269e-06, + "loss": 1.2061, + "step": 467 + }, + { + "epoch": 0.03988069876438006, + "grad_norm": 11.75, + "learning_rate": 2.4984707819883863e-06, + "loss": 0.6979, + "step": 468 + }, + { + "epoch": 0.039965913932680015, + "grad_norm": 14.9375, + "learning_rate": 2.4984624614353813e-06, + "loss": 0.9957, + "step": 469 + }, + { + "epoch": 0.04005112910097997, + "grad_norm": 17.25, + "learning_rate": 2.4984541183214038e-06, + "loss": 0.6148, + "step": 470 + }, + { + "epoch": 0.04013634426927993, + "grad_norm": 13.8125, + "learning_rate": 2.4984457526466055e-06, + "loss": 0.614, + "step": 471 + }, + { + "epoch": 0.040221559437579886, + "grad_norm": 13.5, + "learning_rate": 2.498437364411137e-06, + "loss": 0.7254, + "step": 472 + }, + { + "epoch": 0.04030677460587985, + "grad_norm": 13.0, + "learning_rate": 2.4984289536151496e-06, + "loss": 0.7426, + "step": 473 + }, + { + "epoch": 0.04039198977417981, + "grad_norm": 14.3125, + "learning_rate": 2.4984205202587957e-06, + "loss": 0.6586, + "step": 474 + }, + { + "epoch": 0.040477204942479764, + "grad_norm": 16.375, + "learning_rate": 2.498412064342228e-06, + "loss": 0.7571, + "step": 475 + }, + { + "epoch": 0.04056242011077972, + "grad_norm": 14.125, + "learning_rate": 2.4984035858655985e-06, + "loss": 0.8438, + "step": 476 + }, + { + "epoch": 0.04064763527907968, + "grad_norm": 14.5, + "learning_rate": 2.498395084829061e-06, + "loss": 0.7932, + "step": 477 + }, + { + "epoch": 0.040732850447379634, + "grad_norm": 10.8125, + "learning_rate": 2.498386561232769e-06, + "loss": 0.5459, + "step": 478 + }, + { + "epoch": 0.04081806561567959, + "grad_norm": 12.5625, + "learning_rate": 2.4983780150768763e-06, + "loss": 0.529, + "step": 479 + }, + { + "epoch": 0.04090328078397955, + "grad_norm": 13.625, + "learning_rate": 2.498369446361538e-06, + "loss": 0.6605, + "step": 480 + }, + { + "epoch": 0.040988495952279505, + "grad_norm": 10.125, + "learning_rate": 2.4983608550869078e-06, + "loss": 0.4761, + "step": 481 + }, + { + "epoch": 0.04107371112057946, + "grad_norm": 11.375, + "learning_rate": 2.498352241253142e-06, + "loss": 0.607, + "step": 482 + }, + { + "epoch": 0.04115892628887942, + "grad_norm": 21.75, + "learning_rate": 2.498343604860396e-06, + "loss": 0.7981, + "step": 483 + }, + { + "epoch": 0.041244141457179376, + "grad_norm": 16.0, + "learning_rate": 2.498334945908825e-06, + "loss": 1.15, + "step": 484 + }, + { + "epoch": 0.04132935662547933, + "grad_norm": 16.875, + "learning_rate": 2.4983262643985873e-06, + "loss": 1.2257, + "step": 485 + }, + { + "epoch": 0.04141457179377929, + "grad_norm": 32.5, + "learning_rate": 2.498317560329838e-06, + "loss": 1.0988, + "step": 486 + }, + { + "epoch": 0.04149978696207925, + "grad_norm": 33.75, + "learning_rate": 2.4983088337027354e-06, + "loss": 1.1718, + "step": 487 + }, + { + "epoch": 0.04158500213037921, + "grad_norm": 14.0, + "learning_rate": 2.4983000845174365e-06, + "loss": 0.9357, + "step": 488 + }, + { + "epoch": 0.04167021729867917, + "grad_norm": 20.0, + "learning_rate": 2.4982913127741006e-06, + "loss": 1.2299, + "step": 489 + }, + { + "epoch": 0.041755432466979124, + "grad_norm": 11.8125, + "learning_rate": 2.4982825184728847e-06, + "loss": 0.6561, + "step": 490 + }, + { + "epoch": 0.04184064763527908, + "grad_norm": 17.125, + "learning_rate": 2.4982737016139486e-06, + "loss": 1.0078, + "step": 491 + }, + { + "epoch": 0.04192586280357904, + "grad_norm": 13.5625, + "learning_rate": 2.498264862197452e-06, + "loss": 0.7485, + "step": 492 + }, + { + "epoch": 0.042011077971878995, + "grad_norm": 15.375, + "learning_rate": 2.4982560002235535e-06, + "loss": 1.1153, + "step": 493 + }, + { + "epoch": 0.04209629314017895, + "grad_norm": 19.25, + "learning_rate": 2.498247115692414e-06, + "loss": 1.0774, + "step": 494 + }, + { + "epoch": 0.04218150830847891, + "grad_norm": 14.5, + "learning_rate": 2.498238208604194e-06, + "loss": 0.6382, + "step": 495 + }, + { + "epoch": 0.042266723476778865, + "grad_norm": 12.625, + "learning_rate": 2.498229278959054e-06, + "loss": 0.72, + "step": 496 + }, + { + "epoch": 0.04235193864507882, + "grad_norm": 13.3125, + "learning_rate": 2.4982203267571563e-06, + "loss": 0.8994, + "step": 497 + }, + { + "epoch": 0.04243715381337878, + "grad_norm": 15.0, + "learning_rate": 2.498211351998662e-06, + "loss": 0.7694, + "step": 498 + }, + { + "epoch": 0.042522368981678736, + "grad_norm": 13.125, + "learning_rate": 2.498202354683733e-06, + "loss": 0.7086, + "step": 499 + }, + { + "epoch": 0.04260758414997869, + "grad_norm": 18.25, + "learning_rate": 2.4981933348125327e-06, + "loss": 1.0981, + "step": 500 + }, + { + "epoch": 0.04269279931827866, + "grad_norm": 15.1875, + "learning_rate": 2.4981842923852232e-06, + "loss": 0.716, + "step": 501 + }, + { + "epoch": 0.042778014486578614, + "grad_norm": 15.5625, + "learning_rate": 2.498175227401969e-06, + "loss": 0.8802, + "step": 502 + }, + { + "epoch": 0.04286322965487857, + "grad_norm": 13.1875, + "learning_rate": 2.498166139862933e-06, + "loss": 0.613, + "step": 503 + }, + { + "epoch": 0.04294844482317853, + "grad_norm": 10.875, + "learning_rate": 2.49815702976828e-06, + "loss": 0.4915, + "step": 504 + }, + { + "epoch": 0.043033659991478485, + "grad_norm": 14.125, + "learning_rate": 2.498147897118174e-06, + "loss": 0.9676, + "step": 505 + }, + { + "epoch": 0.04311887515977844, + "grad_norm": 17.125, + "learning_rate": 2.4981387419127806e-06, + "loss": 0.8427, + "step": 506 + }, + { + "epoch": 0.0432040903280784, + "grad_norm": 15.375, + "learning_rate": 2.4981295641522647e-06, + "loss": 1.1045, + "step": 507 + }, + { + "epoch": 0.043289305496378355, + "grad_norm": 17.25, + "learning_rate": 2.498120363836793e-06, + "loss": 1.0553, + "step": 508 + }, + { + "epoch": 0.04337452066467831, + "grad_norm": 17.0, + "learning_rate": 2.498111140966531e-06, + "loss": 1.1626, + "step": 509 + }, + { + "epoch": 0.04345973583297827, + "grad_norm": 17.0, + "learning_rate": 2.4981018955416454e-06, + "loss": 0.7456, + "step": 510 + }, + { + "epoch": 0.043544951001278226, + "grad_norm": 20.875, + "learning_rate": 2.498092627562304e-06, + "loss": 0.9689, + "step": 511 + }, + { + "epoch": 0.04363016616957818, + "grad_norm": 17.625, + "learning_rate": 2.498083337028673e-06, + "loss": 0.8882, + "step": 512 + }, + { + "epoch": 0.04371538133787814, + "grad_norm": 10.5, + "learning_rate": 2.4980740239409216e-06, + "loss": 0.5084, + "step": 513 + }, + { + "epoch": 0.0438005965061781, + "grad_norm": 13.875, + "learning_rate": 2.498064688299218e-06, + "loss": 0.8218, + "step": 514 + }, + { + "epoch": 0.04388581167447806, + "grad_norm": 12.4375, + "learning_rate": 2.4980553301037296e-06, + "loss": 0.4602, + "step": 515 + }, + { + "epoch": 0.04397102684277802, + "grad_norm": 17.375, + "learning_rate": 2.4980459493546266e-06, + "loss": 1.0326, + "step": 516 + }, + { + "epoch": 0.044056242011077974, + "grad_norm": 31.25, + "learning_rate": 2.4980365460520784e-06, + "loss": 1.1643, + "step": 517 + }, + { + "epoch": 0.04414145717937793, + "grad_norm": 13.5625, + "learning_rate": 2.4980271201962547e-06, + "loss": 0.7769, + "step": 518 + }, + { + "epoch": 0.04422667234767789, + "grad_norm": 20.375, + "learning_rate": 2.498017671787326e-06, + "loss": 1.2848, + "step": 519 + }, + { + "epoch": 0.044311887515977845, + "grad_norm": 17.125, + "learning_rate": 2.498008200825463e-06, + "loss": 0.9698, + "step": 520 + }, + { + "epoch": 0.0443971026842778, + "grad_norm": 12.625, + "learning_rate": 2.497998707310837e-06, + "loss": 0.4529, + "step": 521 + }, + { + "epoch": 0.04448231785257776, + "grad_norm": 16.125, + "learning_rate": 2.497989191243619e-06, + "loss": 1.1708, + "step": 522 + }, + { + "epoch": 0.044567533020877716, + "grad_norm": 15.375, + "learning_rate": 2.4979796526239816e-06, + "loss": 0.9436, + "step": 523 + }, + { + "epoch": 0.04465274818917767, + "grad_norm": 17.125, + "learning_rate": 2.497970091452097e-06, + "loss": 0.7137, + "step": 524 + }, + { + "epoch": 0.04473796335747763, + "grad_norm": 16.875, + "learning_rate": 2.4979605077281375e-06, + "loss": 0.9704, + "step": 525 + }, + { + "epoch": 0.044823178525777586, + "grad_norm": 12.5625, + "learning_rate": 2.497950901452277e-06, + "loss": 0.4724, + "step": 526 + }, + { + "epoch": 0.04490839369407754, + "grad_norm": 19.25, + "learning_rate": 2.4979412726246894e-06, + "loss": 1.0186, + "step": 527 + }, + { + "epoch": 0.0449936088623775, + "grad_norm": 22.25, + "learning_rate": 2.4979316212455473e-06, + "loss": 1.0909, + "step": 528 + }, + { + "epoch": 0.045078824030677464, + "grad_norm": 13.5625, + "learning_rate": 2.4979219473150263e-06, + "loss": 0.5592, + "step": 529 + }, + { + "epoch": 0.04516403919897742, + "grad_norm": 11.5, + "learning_rate": 2.4979122508333004e-06, + "loss": 0.678, + "step": 530 + }, + { + "epoch": 0.04524925436727738, + "grad_norm": 15.8125, + "learning_rate": 2.497902531800546e-06, + "loss": 1.113, + "step": 531 + }, + { + "epoch": 0.045334469535577335, + "grad_norm": 10.0625, + "learning_rate": 2.4978927902169376e-06, + "loss": 0.477, + "step": 532 + }, + { + "epoch": 0.04541968470387729, + "grad_norm": 13.0625, + "learning_rate": 2.4978830260826512e-06, + "loss": 1.0077, + "step": 533 + }, + { + "epoch": 0.04550489987217725, + "grad_norm": 14.125, + "learning_rate": 2.4978732393978644e-06, + "loss": 1.0396, + "step": 534 + }, + { + "epoch": 0.045590115040477205, + "grad_norm": 15.875, + "learning_rate": 2.497863430162753e-06, + "loss": 1.0904, + "step": 535 + }, + { + "epoch": 0.04567533020877716, + "grad_norm": 12.5, + "learning_rate": 2.4978535983774945e-06, + "loss": 0.5543, + "step": 536 + }, + { + "epoch": 0.04576054537707712, + "grad_norm": 13.8125, + "learning_rate": 2.4978437440422664e-06, + "loss": 0.5372, + "step": 537 + }, + { + "epoch": 0.045845760545377076, + "grad_norm": 11.5, + "learning_rate": 2.4978338671572476e-06, + "loss": 0.6376, + "step": 538 + }, + { + "epoch": 0.04593097571367703, + "grad_norm": 15.25, + "learning_rate": 2.4978239677226155e-06, + "loss": 0.9596, + "step": 539 + }, + { + "epoch": 0.04601619088197699, + "grad_norm": 28.875, + "learning_rate": 2.49781404573855e-06, + "loss": 1.2914, + "step": 540 + }, + { + "epoch": 0.04610140605027695, + "grad_norm": 16.0, + "learning_rate": 2.49780410120523e-06, + "loss": 0.8702, + "step": 541 + }, + { + "epoch": 0.046186621218576904, + "grad_norm": 14.9375, + "learning_rate": 2.4977941341228347e-06, + "loss": 0.6548, + "step": 542 + }, + { + "epoch": 0.04627183638687686, + "grad_norm": 13.6875, + "learning_rate": 2.4977841444915448e-06, + "loss": 0.8117, + "step": 543 + }, + { + "epoch": 0.046357051555176824, + "grad_norm": 12.625, + "learning_rate": 2.4977741323115406e-06, + "loss": 0.4431, + "step": 544 + }, + { + "epoch": 0.04644226672347678, + "grad_norm": 14.4375, + "learning_rate": 2.497764097583003e-06, + "loss": 1.0855, + "step": 545 + }, + { + "epoch": 0.04652748189177674, + "grad_norm": 14.3125, + "learning_rate": 2.497754040306114e-06, + "loss": 0.892, + "step": 546 + }, + { + "epoch": 0.046612697060076695, + "grad_norm": 12.375, + "learning_rate": 2.497743960481054e-06, + "loss": 0.6311, + "step": 547 + }, + { + "epoch": 0.04669791222837665, + "grad_norm": 14.6875, + "learning_rate": 2.497733858108007e-06, + "loss": 0.9285, + "step": 548 + }, + { + "epoch": 0.04678312739667661, + "grad_norm": 21.25, + "learning_rate": 2.4977237331871535e-06, + "loss": 0.9391, + "step": 549 + }, + { + "epoch": 0.046868342564976566, + "grad_norm": 15.375, + "learning_rate": 2.497713585718678e-06, + "loss": 0.8202, + "step": 550 + }, + { + "epoch": 0.04695355773327652, + "grad_norm": 15.9375, + "learning_rate": 2.497703415702763e-06, + "loss": 0.9054, + "step": 551 + }, + { + "epoch": 0.04703877290157648, + "grad_norm": 13.5625, + "learning_rate": 2.4976932231395926e-06, + "loss": 0.5876, + "step": 552 + }, + { + "epoch": 0.04712398806987644, + "grad_norm": 12.375, + "learning_rate": 2.4976830080293514e-06, + "loss": 0.5749, + "step": 553 + }, + { + "epoch": 0.04720920323817639, + "grad_norm": 16.75, + "learning_rate": 2.497672770372223e-06, + "loss": 1.0638, + "step": 554 + }, + { + "epoch": 0.04729441840647635, + "grad_norm": 19.0, + "learning_rate": 2.497662510168393e-06, + "loss": 1.0446, + "step": 555 + }, + { + "epoch": 0.04737963357477631, + "grad_norm": 28.25, + "learning_rate": 2.4976522274180475e-06, + "loss": 1.5739, + "step": 556 + }, + { + "epoch": 0.047464848743076264, + "grad_norm": 22.125, + "learning_rate": 2.497641922121371e-06, + "loss": 1.1316, + "step": 557 + }, + { + "epoch": 0.04755006391137623, + "grad_norm": 14.75, + "learning_rate": 2.497631594278551e-06, + "loss": 1.0194, + "step": 558 + }, + { + "epoch": 0.047635279079676185, + "grad_norm": 12.25, + "learning_rate": 2.497621243889773e-06, + "loss": 0.5275, + "step": 559 + }, + { + "epoch": 0.04772049424797614, + "grad_norm": 14.1875, + "learning_rate": 2.4976108709552247e-06, + "loss": 0.5929, + "step": 560 + }, + { + "epoch": 0.0478057094162761, + "grad_norm": 18.25, + "learning_rate": 2.4976004754750933e-06, + "loss": 0.9813, + "step": 561 + }, + { + "epoch": 0.047890924584576056, + "grad_norm": 11.1875, + "learning_rate": 2.4975900574495672e-06, + "loss": 0.3754, + "step": 562 + }, + { + "epoch": 0.04797613975287601, + "grad_norm": 14.5, + "learning_rate": 2.4975796168788337e-06, + "loss": 0.7562, + "step": 563 + }, + { + "epoch": 0.04806135492117597, + "grad_norm": 18.0, + "learning_rate": 2.497569153763082e-06, + "loss": 1.0602, + "step": 564 + }, + { + "epoch": 0.048146570089475926, + "grad_norm": 21.375, + "learning_rate": 2.4975586681025016e-06, + "loss": 0.6363, + "step": 565 + }, + { + "epoch": 0.04823178525777588, + "grad_norm": 23.625, + "learning_rate": 2.497548159897281e-06, + "loss": 0.9576, + "step": 566 + }, + { + "epoch": 0.04831700042607584, + "grad_norm": 14.3125, + "learning_rate": 2.497537629147611e-06, + "loss": 0.901, + "step": 567 + }, + { + "epoch": 0.0484022155943758, + "grad_norm": 13.4375, + "learning_rate": 2.497527075853681e-06, + "loss": 0.7384, + "step": 568 + }, + { + "epoch": 0.048487430762675754, + "grad_norm": 15.875, + "learning_rate": 2.497516500015683e-06, + "loss": 0.8411, + "step": 569 + }, + { + "epoch": 0.04857264593097571, + "grad_norm": 16.625, + "learning_rate": 2.497505901633807e-06, + "loss": 0.9076, + "step": 570 + }, + { + "epoch": 0.04865786109927567, + "grad_norm": 14.6875, + "learning_rate": 2.4974952807082454e-06, + "loss": 0.6311, + "step": 571 + }, + { + "epoch": 0.04874307626757563, + "grad_norm": 15.25, + "learning_rate": 2.497484637239189e-06, + "loss": 0.5999, + "step": 572 + }, + { + "epoch": 0.04882829143587559, + "grad_norm": 15.5, + "learning_rate": 2.4974739712268307e-06, + "loss": 0.6612, + "step": 573 + }, + { + "epoch": 0.048913506604175545, + "grad_norm": 15.625, + "learning_rate": 2.4974632826713636e-06, + "loss": 0.6295, + "step": 574 + }, + { + "epoch": 0.0489987217724755, + "grad_norm": 14.1875, + "learning_rate": 2.4974525715729808e-06, + "loss": 0.4852, + "step": 575 + }, + { + "epoch": 0.04908393694077546, + "grad_norm": 13.3125, + "learning_rate": 2.497441837931875e-06, + "loss": 0.5265, + "step": 576 + }, + { + "epoch": 0.049169152109075416, + "grad_norm": 17.25, + "learning_rate": 2.4974310817482414e-06, + "loss": 0.891, + "step": 577 + }, + { + "epoch": 0.04925436727737537, + "grad_norm": 21.125, + "learning_rate": 2.4974203030222732e-06, + "loss": 1.0509, + "step": 578 + }, + { + "epoch": 0.04933958244567533, + "grad_norm": 15.0625, + "learning_rate": 2.497409501754166e-06, + "loss": 0.8276, + "step": 579 + }, + { + "epoch": 0.04942479761397529, + "grad_norm": 17.0, + "learning_rate": 2.497398677944115e-06, + "loss": 0.9166, + "step": 580 + }, + { + "epoch": 0.049510012782275244, + "grad_norm": 17.75, + "learning_rate": 2.4973878315923153e-06, + "loss": 1.0019, + "step": 581 + }, + { + "epoch": 0.0495952279505752, + "grad_norm": 15.9375, + "learning_rate": 2.497376962698963e-06, + "loss": 1.0199, + "step": 582 + }, + { + "epoch": 0.04968044311887516, + "grad_norm": 14.625, + "learning_rate": 2.497366071264255e-06, + "loss": 1.015, + "step": 583 + }, + { + "epoch": 0.049765658287175114, + "grad_norm": 14.25, + "learning_rate": 2.4973551572883875e-06, + "loss": 0.9148, + "step": 584 + }, + { + "epoch": 0.04985087345547507, + "grad_norm": 15.4375, + "learning_rate": 2.497344220771558e-06, + "loss": 1.0273, + "step": 585 + }, + { + "epoch": 0.049936088623775035, + "grad_norm": 23.125, + "learning_rate": 2.497333261713964e-06, + "loss": 1.0809, + "step": 586 + }, + { + "epoch": 0.05002130379207499, + "grad_norm": 18.375, + "learning_rate": 2.4973222801158044e-06, + "loss": 1.0111, + "step": 587 + }, + { + "epoch": 0.05010651896037495, + "grad_norm": 10.5, + "learning_rate": 2.4973112759772763e-06, + "loss": 0.5785, + "step": 588 + }, + { + "epoch": 0.050191734128674906, + "grad_norm": 15.8125, + "learning_rate": 2.497300249298579e-06, + "loss": 0.9292, + "step": 589 + }, + { + "epoch": 0.05027694929697486, + "grad_norm": 18.625, + "learning_rate": 2.497289200079912e-06, + "loss": 1.1965, + "step": 590 + }, + { + "epoch": 0.05036216446527482, + "grad_norm": 12.4375, + "learning_rate": 2.4972781283214753e-06, + "loss": 0.7628, + "step": 591 + }, + { + "epoch": 0.050447379633574777, + "grad_norm": 18.625, + "learning_rate": 2.497267034023468e-06, + "loss": 1.1404, + "step": 592 + }, + { + "epoch": 0.05053259480187473, + "grad_norm": 23.0, + "learning_rate": 2.4972559171860917e-06, + "loss": 0.9371, + "step": 593 + }, + { + "epoch": 0.05061780997017469, + "grad_norm": 20.875, + "learning_rate": 2.4972447778095468e-06, + "loss": 1.2122, + "step": 594 + }, + { + "epoch": 0.05070302513847465, + "grad_norm": 20.625, + "learning_rate": 2.4972336158940344e-06, + "loss": 1.4592, + "step": 595 + }, + { + "epoch": 0.050788240306774604, + "grad_norm": 17.25, + "learning_rate": 2.4972224314397564e-06, + "loss": 0.9688, + "step": 596 + }, + { + "epoch": 0.05087345547507456, + "grad_norm": 13.6875, + "learning_rate": 2.497211224446915e-06, + "loss": 0.7272, + "step": 597 + }, + { + "epoch": 0.05095867064337452, + "grad_norm": 21.375, + "learning_rate": 2.4971999949157123e-06, + "loss": 0.7135, + "step": 598 + }, + { + "epoch": 0.051043885811674475, + "grad_norm": 21.0, + "learning_rate": 2.4971887428463515e-06, + "loss": 1.1397, + "step": 599 + }, + { + "epoch": 0.05112910097997444, + "grad_norm": 15.4375, + "learning_rate": 2.4971774682390365e-06, + "loss": 0.8959, + "step": 600 + }, + { + "epoch": 0.051214316148274396, + "grad_norm": 25.625, + "learning_rate": 2.49716617109397e-06, + "loss": 1.153, + "step": 601 + }, + { + "epoch": 0.05129953131657435, + "grad_norm": 14.3125, + "learning_rate": 2.497154851411357e-06, + "loss": 0.4679, + "step": 602 + }, + { + "epoch": 0.05138474648487431, + "grad_norm": 15.125, + "learning_rate": 2.4971435091914014e-06, + "loss": 0.9237, + "step": 603 + }, + { + "epoch": 0.051469961653174266, + "grad_norm": 12.3125, + "learning_rate": 2.4971321444343084e-06, + "loss": 0.6365, + "step": 604 + }, + { + "epoch": 0.05155517682147422, + "grad_norm": 17.75, + "learning_rate": 2.497120757140284e-06, + "loss": 1.0036, + "step": 605 + }, + { + "epoch": 0.05164039198977418, + "grad_norm": 12.625, + "learning_rate": 2.4971093473095333e-06, + "loss": 0.5024, + "step": 606 + }, + { + "epoch": 0.05172560715807414, + "grad_norm": 15.3125, + "learning_rate": 2.4970979149422624e-06, + "loss": 0.5912, + "step": 607 + }, + { + "epoch": 0.051810822326374094, + "grad_norm": 40.5, + "learning_rate": 2.497086460038678e-06, + "loss": 0.3844, + "step": 608 + }, + { + "epoch": 0.05189603749467405, + "grad_norm": 14.0625, + "learning_rate": 2.4970749825989877e-06, + "loss": 0.9913, + "step": 609 + }, + { + "epoch": 0.05198125266297401, + "grad_norm": 12.1875, + "learning_rate": 2.497063482623398e-06, + "loss": 0.56, + "step": 610 + }, + { + "epoch": 0.052066467831273965, + "grad_norm": 27.875, + "learning_rate": 2.4970519601121175e-06, + "loss": 0.9598, + "step": 611 + }, + { + "epoch": 0.05215168299957392, + "grad_norm": 12.25, + "learning_rate": 2.4970404150653534e-06, + "loss": 0.6372, + "step": 612 + }, + { + "epoch": 0.05223689816787388, + "grad_norm": 16.375, + "learning_rate": 2.497028847483316e-06, + "loss": 1.0319, + "step": 613 + }, + { + "epoch": 0.05232211333617384, + "grad_norm": 17.625, + "learning_rate": 2.4970172573662124e-06, + "loss": 1.1488, + "step": 614 + }, + { + "epoch": 0.0524073285044738, + "grad_norm": 17.0, + "learning_rate": 2.4970056447142535e-06, + "loss": 0.7887, + "step": 615 + }, + { + "epoch": 0.052492543672773756, + "grad_norm": 25.5, + "learning_rate": 2.4969940095276483e-06, + "loss": 1.1172, + "step": 616 + }, + { + "epoch": 0.05257775884107371, + "grad_norm": 17.75, + "learning_rate": 2.496982351806608e-06, + "loss": 0.7054, + "step": 617 + }, + { + "epoch": 0.05266297400937367, + "grad_norm": 18.125, + "learning_rate": 2.496970671551342e-06, + "loss": 1.325, + "step": 618 + }, + { + "epoch": 0.05274818917767363, + "grad_norm": 20.25, + "learning_rate": 2.4969589687620626e-06, + "loss": 0.9802, + "step": 619 + }, + { + "epoch": 0.052833404345973584, + "grad_norm": 15.25, + "learning_rate": 2.49694724343898e-06, + "loss": 1.0539, + "step": 620 + }, + { + "epoch": 0.05291861951427354, + "grad_norm": 17.125, + "learning_rate": 2.496935495582308e-06, + "loss": 0.5991, + "step": 621 + }, + { + "epoch": 0.0530038346825735, + "grad_norm": 12.875, + "learning_rate": 2.4969237251922567e-06, + "loss": 0.5742, + "step": 622 + }, + { + "epoch": 0.053089049850873454, + "grad_norm": 20.0, + "learning_rate": 2.49691193226904e-06, + "loss": 1.1053, + "step": 623 + }, + { + "epoch": 0.05317426501917341, + "grad_norm": 17.375, + "learning_rate": 2.496900116812871e-06, + "loss": 0.9491, + "step": 624 + }, + { + "epoch": 0.05325948018747337, + "grad_norm": 20.625, + "learning_rate": 2.496888278823963e-06, + "loss": 1.0348, + "step": 625 + }, + { + "epoch": 0.053344695355773325, + "grad_norm": 17.25, + "learning_rate": 2.4968764183025297e-06, + "loss": 0.7143, + "step": 626 + }, + { + "epoch": 0.05342991052407328, + "grad_norm": 14.1875, + "learning_rate": 2.496864535248786e-06, + "loss": 0.8898, + "step": 627 + }, + { + "epoch": 0.053515125692373246, + "grad_norm": 15.8125, + "learning_rate": 2.496852629662946e-06, + "loss": 1.0079, + "step": 628 + }, + { + "epoch": 0.0536003408606732, + "grad_norm": 17.375, + "learning_rate": 2.4968407015452255e-06, + "loss": 0.9257, + "step": 629 + }, + { + "epoch": 0.05368555602897316, + "grad_norm": 14.0625, + "learning_rate": 2.4968287508958397e-06, + "loss": 0.8194, + "step": 630 + }, + { + "epoch": 0.053770771197273116, + "grad_norm": 15.9375, + "learning_rate": 2.4968167777150044e-06, + "loss": 0.7869, + "step": 631 + }, + { + "epoch": 0.05385598636557307, + "grad_norm": 13.5, + "learning_rate": 2.4968047820029364e-06, + "loss": 0.4579, + "step": 632 + }, + { + "epoch": 0.05394120153387303, + "grad_norm": 21.125, + "learning_rate": 2.496792763759852e-06, + "loss": 0.8788, + "step": 633 + }, + { + "epoch": 0.05402641670217299, + "grad_norm": 16.0, + "learning_rate": 2.4967807229859685e-06, + "loss": 1.048, + "step": 634 + }, + { + "epoch": 0.054111631870472944, + "grad_norm": 13.5, + "learning_rate": 2.496768659681504e-06, + "loss": 0.7781, + "step": 635 + }, + { + "epoch": 0.0541968470387729, + "grad_norm": 12.0625, + "learning_rate": 2.496756573846676e-06, + "loss": 0.7897, + "step": 636 + }, + { + "epoch": 0.05428206220707286, + "grad_norm": 14.25, + "learning_rate": 2.496744465481703e-06, + "loss": 0.7913, + "step": 637 + }, + { + "epoch": 0.054367277375372815, + "grad_norm": 11.0625, + "learning_rate": 2.4967323345868036e-06, + "loss": 0.3933, + "step": 638 + }, + { + "epoch": 0.05445249254367277, + "grad_norm": 13.5, + "learning_rate": 2.496720181162197e-06, + "loss": 0.8064, + "step": 639 + }, + { + "epoch": 0.05453770771197273, + "grad_norm": 14.875, + "learning_rate": 2.496708005208104e-06, + "loss": 0.5314, + "step": 640 + }, + { + "epoch": 0.054622922880272685, + "grad_norm": 20.0, + "learning_rate": 2.4966958067247426e-06, + "loss": 0.864, + "step": 641 + }, + { + "epoch": 0.05470813804857265, + "grad_norm": 13.25, + "learning_rate": 2.496683585712335e-06, + "loss": 0.8099, + "step": 642 + }, + { + "epoch": 0.054793353216872606, + "grad_norm": 12.0625, + "learning_rate": 2.4966713421711013e-06, + "loss": 0.3473, + "step": 643 + }, + { + "epoch": 0.05487856838517256, + "grad_norm": 15.375, + "learning_rate": 2.4966590761012627e-06, + "loss": 0.7167, + "step": 644 + }, + { + "epoch": 0.05496378355347252, + "grad_norm": 12.5625, + "learning_rate": 2.496646787503041e-06, + "loss": 0.4618, + "step": 645 + }, + { + "epoch": 0.05504899872177248, + "grad_norm": 14.875, + "learning_rate": 2.4966344763766583e-06, + "loss": 0.8458, + "step": 646 + }, + { + "epoch": 0.055134213890072434, + "grad_norm": 15.5, + "learning_rate": 2.496622142722337e-06, + "loss": 1.1295, + "step": 647 + }, + { + "epoch": 0.05521942905837239, + "grad_norm": 12.6875, + "learning_rate": 2.4966097865402995e-06, + "loss": 0.6643, + "step": 648 + }, + { + "epoch": 0.05530464422667235, + "grad_norm": 13.6875, + "learning_rate": 2.4965974078307703e-06, + "loss": 0.6109, + "step": 649 + }, + { + "epoch": 0.055389859394972304, + "grad_norm": 13.6875, + "learning_rate": 2.496585006593972e-06, + "loss": 0.6909, + "step": 650 + }, + { + "epoch": 0.05547507456327226, + "grad_norm": 20.375, + "learning_rate": 2.4965725828301294e-06, + "loss": 0.9731, + "step": 651 + }, + { + "epoch": 0.05556028973157222, + "grad_norm": 14.0625, + "learning_rate": 2.4965601365394666e-06, + "loss": 0.9019, + "step": 652 + }, + { + "epoch": 0.055645504899872175, + "grad_norm": 19.75, + "learning_rate": 2.496547667722208e-06, + "loss": 0.9211, + "step": 653 + }, + { + "epoch": 0.05573072006817213, + "grad_norm": 14.625, + "learning_rate": 2.4965351763785804e-06, + "loss": 0.7265, + "step": 654 + }, + { + "epoch": 0.05581593523647209, + "grad_norm": 17.75, + "learning_rate": 2.4965226625088084e-06, + "loss": 0.6484, + "step": 655 + }, + { + "epoch": 0.05590115040477205, + "grad_norm": 16.25, + "learning_rate": 2.4965101261131185e-06, + "loss": 1.1539, + "step": 656 + }, + { + "epoch": 0.05598636557307201, + "grad_norm": 12.6875, + "learning_rate": 2.496497567191737e-06, + "loss": 0.7442, + "step": 657 + }, + { + "epoch": 0.05607158074137197, + "grad_norm": 45.0, + "learning_rate": 2.4964849857448907e-06, + "loss": 0.923, + "step": 658 + }, + { + "epoch": 0.056156795909671924, + "grad_norm": 13.9375, + "learning_rate": 2.496472381772808e-06, + "loss": 0.5966, + "step": 659 + }, + { + "epoch": 0.05624201107797188, + "grad_norm": 13.375, + "learning_rate": 2.4964597552757155e-06, + "loss": 0.6612, + "step": 660 + }, + { + "epoch": 0.05632722624627184, + "grad_norm": 19.75, + "learning_rate": 2.496447106253842e-06, + "loss": 1.2901, + "step": 661 + }, + { + "epoch": 0.056412441414571794, + "grad_norm": 21.125, + "learning_rate": 2.4964344347074158e-06, + "loss": 0.8304, + "step": 662 + }, + { + "epoch": 0.05649765658287175, + "grad_norm": 16.125, + "learning_rate": 2.496421740636666e-06, + "loss": 1.0131, + "step": 663 + }, + { + "epoch": 0.05658287175117171, + "grad_norm": 17.75, + "learning_rate": 2.4964090240418215e-06, + "loss": 0.9323, + "step": 664 + }, + { + "epoch": 0.056668086919471665, + "grad_norm": 15.25, + "learning_rate": 2.4963962849231135e-06, + "loss": 0.7873, + "step": 665 + }, + { + "epoch": 0.05675330208777162, + "grad_norm": 14.8125, + "learning_rate": 2.4963835232807702e-06, + "loss": 0.7644, + "step": 666 + }, + { + "epoch": 0.05683851725607158, + "grad_norm": 16.875, + "learning_rate": 2.496370739115024e-06, + "loss": 1.2728, + "step": 667 + }, + { + "epoch": 0.056923732424371536, + "grad_norm": 14.125, + "learning_rate": 2.4963579324261052e-06, + "loss": 0.6924, + "step": 668 + }, + { + "epoch": 0.05700894759267149, + "grad_norm": 18.125, + "learning_rate": 2.496345103214245e-06, + "loss": 0.9334, + "step": 669 + }, + { + "epoch": 0.057094162760971456, + "grad_norm": 14.8125, + "learning_rate": 2.4963322514796755e-06, + "loss": 0.6154, + "step": 670 + }, + { + "epoch": 0.05717937792927141, + "grad_norm": 15.125, + "learning_rate": 2.4963193772226295e-06, + "loss": 1.0111, + "step": 671 + }, + { + "epoch": 0.05726459309757137, + "grad_norm": 13.8125, + "learning_rate": 2.4963064804433382e-06, + "loss": 0.755, + "step": 672 + }, + { + "epoch": 0.05734980826587133, + "grad_norm": 17.25, + "learning_rate": 2.4962935611420364e-06, + "loss": 1.1122, + "step": 673 + }, + { + "epoch": 0.057435023434171284, + "grad_norm": 17.625, + "learning_rate": 2.496280619318956e-06, + "loss": 0.6942, + "step": 674 + }, + { + "epoch": 0.05752023860247124, + "grad_norm": 20.25, + "learning_rate": 2.496267654974332e-06, + "loss": 0.9186, + "step": 675 + }, + { + "epoch": 0.0576054537707712, + "grad_norm": 14.875, + "learning_rate": 2.4962546681083982e-06, + "loss": 0.7839, + "step": 676 + }, + { + "epoch": 0.057690668939071155, + "grad_norm": 14.375, + "learning_rate": 2.496241658721389e-06, + "loss": 0.769, + "step": 677 + }, + { + "epoch": 0.05777588410737111, + "grad_norm": 14.5625, + "learning_rate": 2.4962286268135406e-06, + "loss": 0.8106, + "step": 678 + }, + { + "epoch": 0.05786109927567107, + "grad_norm": 15.5, + "learning_rate": 2.4962155723850867e-06, + "loss": 0.7286, + "step": 679 + }, + { + "epoch": 0.057946314443971025, + "grad_norm": 18.5, + "learning_rate": 2.4962024954362653e-06, + "loss": 0.6561, + "step": 680 + }, + { + "epoch": 0.05803152961227098, + "grad_norm": 19.875, + "learning_rate": 2.496189395967311e-06, + "loss": 0.9052, + "step": 681 + }, + { + "epoch": 0.05811674478057094, + "grad_norm": 17.125, + "learning_rate": 2.4961762739784613e-06, + "loss": 1.5005, + "step": 682 + }, + { + "epoch": 0.058201959948870896, + "grad_norm": 11.6875, + "learning_rate": 2.496163129469953e-06, + "loss": 0.4962, + "step": 683 + }, + { + "epoch": 0.05828717511717085, + "grad_norm": 16.125, + "learning_rate": 2.496149962442024e-06, + "loss": 0.8615, + "step": 684 + }, + { + "epoch": 0.05837239028547082, + "grad_norm": 11.0625, + "learning_rate": 2.496136772894912e-06, + "loss": 0.6755, + "step": 685 + }, + { + "epoch": 0.058457605453770774, + "grad_norm": 20.125, + "learning_rate": 2.496123560828856e-06, + "loss": 1.1005, + "step": 686 + }, + { + "epoch": 0.05854282062207073, + "grad_norm": 15.0625, + "learning_rate": 2.4961103262440934e-06, + "loss": 1.0075, + "step": 687 + }, + { + "epoch": 0.05862803579037069, + "grad_norm": 15.1875, + "learning_rate": 2.4960970691408645e-06, + "loss": 0.6836, + "step": 688 + }, + { + "epoch": 0.058713250958670644, + "grad_norm": 11.0, + "learning_rate": 2.4960837895194083e-06, + "loss": 0.7389, + "step": 689 + }, + { + "epoch": 0.0587984661269706, + "grad_norm": 13.6875, + "learning_rate": 2.4960704873799653e-06, + "loss": 0.7842, + "step": 690 + }, + { + "epoch": 0.05888368129527056, + "grad_norm": 15.4375, + "learning_rate": 2.4960571627227753e-06, + "loss": 0.2997, + "step": 691 + }, + { + "epoch": 0.058968896463570515, + "grad_norm": 12.625, + "learning_rate": 2.49604381554808e-06, + "loss": 0.5738, + "step": 692 + }, + { + "epoch": 0.05905411163187047, + "grad_norm": 11.5, + "learning_rate": 2.496030445856119e-06, + "loss": 0.4964, + "step": 693 + }, + { + "epoch": 0.05913932680017043, + "grad_norm": 23.375, + "learning_rate": 2.4960170536471355e-06, + "loss": 0.8448, + "step": 694 + }, + { + "epoch": 0.059224541968470386, + "grad_norm": 17.25, + "learning_rate": 2.496003638921371e-06, + "loss": 0.7069, + "step": 695 + }, + { + "epoch": 0.05930975713677034, + "grad_norm": 17.375, + "learning_rate": 2.495990201679067e-06, + "loss": 0.7296, + "step": 696 + }, + { + "epoch": 0.0593949723050703, + "grad_norm": 14.625, + "learning_rate": 2.495976741920468e-06, + "loss": 1.0706, + "step": 697 + }, + { + "epoch": 0.059480187473370257, + "grad_norm": 16.375, + "learning_rate": 2.4959632596458153e-06, + "loss": 0.6849, + "step": 698 + }, + { + "epoch": 0.05956540264167022, + "grad_norm": 14.25, + "learning_rate": 2.4959497548553546e-06, + "loss": 0.9897, + "step": 699 + }, + { + "epoch": 0.05965061780997018, + "grad_norm": 16.25, + "learning_rate": 2.4959362275493287e-06, + "loss": 1.1302, + "step": 700 + }, + { + "epoch": 0.059735832978270134, + "grad_norm": 14.0625, + "learning_rate": 2.495922677727982e-06, + "loss": 0.8003, + "step": 701 + }, + { + "epoch": 0.05982104814657009, + "grad_norm": 11.3125, + "learning_rate": 2.49590910539156e-06, + "loss": 0.4348, + "step": 702 + }, + { + "epoch": 0.05990626331487005, + "grad_norm": 19.125, + "learning_rate": 2.495895510540307e-06, + "loss": 0.8835, + "step": 703 + }, + { + "epoch": 0.059991478483170005, + "grad_norm": 20.25, + "learning_rate": 2.49588189317447e-06, + "loss": 1.1262, + "step": 704 + }, + { + "epoch": 0.06007669365146996, + "grad_norm": 13.25, + "learning_rate": 2.495868253294294e-06, + "loss": 0.6083, + "step": 705 + }, + { + "epoch": 0.06016190881976992, + "grad_norm": 18.25, + "learning_rate": 2.495854590900026e-06, + "loss": 1.07, + "step": 706 + }, + { + "epoch": 0.060247123988069876, + "grad_norm": 18.75, + "learning_rate": 2.4958409059919124e-06, + "loss": 1.0939, + "step": 707 + }, + { + "epoch": 0.06033233915636983, + "grad_norm": 14.1875, + "learning_rate": 2.495827198570201e-06, + "loss": 1.0319, + "step": 708 + }, + { + "epoch": 0.06041755432466979, + "grad_norm": 16.875, + "learning_rate": 2.4958134686351396e-06, + "loss": 0.9677, + "step": 709 + }, + { + "epoch": 0.060502769492969746, + "grad_norm": 9.5, + "learning_rate": 2.495799716186976e-06, + "loss": 0.2821, + "step": 710 + }, + { + "epoch": 0.0605879846612697, + "grad_norm": 16.75, + "learning_rate": 2.4957859412259584e-06, + "loss": 0.4358, + "step": 711 + }, + { + "epoch": 0.06067319982956966, + "grad_norm": 14.5625, + "learning_rate": 2.4957721437523365e-06, + "loss": 1.0668, + "step": 712 + }, + { + "epoch": 0.060758414997869624, + "grad_norm": 15.5625, + "learning_rate": 2.495758323766359e-06, + "loss": 0.9974, + "step": 713 + }, + { + "epoch": 0.06084363016616958, + "grad_norm": 17.375, + "learning_rate": 2.495744481268276e-06, + "loss": 1.0701, + "step": 714 + }, + { + "epoch": 0.06092884533446954, + "grad_norm": 10.375, + "learning_rate": 2.4957306162583376e-06, + "loss": 0.3589, + "step": 715 + }, + { + "epoch": 0.061014060502769495, + "grad_norm": 13.375, + "learning_rate": 2.495716728736794e-06, + "loss": 0.6527, + "step": 716 + }, + { + "epoch": 0.06109927567106945, + "grad_norm": 15.1875, + "learning_rate": 2.4957028187038966e-06, + "loss": 0.8977, + "step": 717 + }, + { + "epoch": 0.06118449083936941, + "grad_norm": 16.5, + "learning_rate": 2.4956888861598968e-06, + "loss": 0.5937, + "step": 718 + }, + { + "epoch": 0.061269706007669365, + "grad_norm": 30.625, + "learning_rate": 2.495674931105046e-06, + "loss": 1.0829, + "step": 719 + }, + { + "epoch": 0.06135492117596932, + "grad_norm": 17.875, + "learning_rate": 2.4956609535395966e-06, + "loss": 1.0039, + "step": 720 + }, + { + "epoch": 0.06144013634426928, + "grad_norm": 18.0, + "learning_rate": 2.495646953463801e-06, + "loss": 1.0935, + "step": 721 + }, + { + "epoch": 0.061525351512569236, + "grad_norm": 18.875, + "learning_rate": 2.4956329308779125e-06, + "loss": 1.0674, + "step": 722 + }, + { + "epoch": 0.06161056668086919, + "grad_norm": 19.375, + "learning_rate": 2.4956188857821846e-06, + "loss": 0.6899, + "step": 723 + }, + { + "epoch": 0.06169578184916915, + "grad_norm": 16.75, + "learning_rate": 2.495604818176871e-06, + "loss": 0.7459, + "step": 724 + }, + { + "epoch": 0.06178099701746911, + "grad_norm": 13.125, + "learning_rate": 2.4955907280622253e-06, + "loss": 0.6505, + "step": 725 + }, + { + "epoch": 0.061866212185769064, + "grad_norm": 20.125, + "learning_rate": 2.4955766154385025e-06, + "loss": 1.2594, + "step": 726 + }, + { + "epoch": 0.06195142735406903, + "grad_norm": 15.8125, + "learning_rate": 2.495562480305958e-06, + "loss": 0.8341, + "step": 727 + }, + { + "epoch": 0.062036642522368984, + "grad_norm": 16.25, + "learning_rate": 2.495548322664847e-06, + "loss": 0.582, + "step": 728 + }, + { + "epoch": 0.06212185769066894, + "grad_norm": 17.625, + "learning_rate": 2.495534142515425e-06, + "loss": 0.8811, + "step": 729 + }, + { + "epoch": 0.0622070728589689, + "grad_norm": 16.75, + "learning_rate": 2.495519939857949e-06, + "loss": 0.984, + "step": 730 + }, + { + "epoch": 0.062292288027268855, + "grad_norm": 23.75, + "learning_rate": 2.4955057146926754e-06, + "loss": 1.0588, + "step": 731 + }, + { + "epoch": 0.06237750319556881, + "grad_norm": 32.75, + "learning_rate": 2.4954914670198605e-06, + "loss": 1.0906, + "step": 732 + }, + { + "epoch": 0.06246271836386877, + "grad_norm": 39.5, + "learning_rate": 2.4954771968397624e-06, + "loss": 1.5367, + "step": 733 + }, + { + "epoch": 0.06254793353216873, + "grad_norm": 16.0, + "learning_rate": 2.4954629041526393e-06, + "loss": 0.6287, + "step": 734 + }, + { + "epoch": 0.06263314870046868, + "grad_norm": 10.375, + "learning_rate": 2.495448588958749e-06, + "loss": 0.3451, + "step": 735 + }, + { + "epoch": 0.06271836386876864, + "grad_norm": 13.6875, + "learning_rate": 2.4954342512583503e-06, + "loss": 0.789, + "step": 736 + }, + { + "epoch": 0.0628035790370686, + "grad_norm": 13.8125, + "learning_rate": 2.495419891051702e-06, + "loss": 0.4972, + "step": 737 + }, + { + "epoch": 0.06288879420536855, + "grad_norm": 17.125, + "learning_rate": 2.4954055083390643e-06, + "loss": 0.7407, + "step": 738 + }, + { + "epoch": 0.06297400937366851, + "grad_norm": 18.125, + "learning_rate": 2.4953911031206966e-06, + "loss": 1.0121, + "step": 739 + }, + { + "epoch": 0.06305922454196847, + "grad_norm": 17.375, + "learning_rate": 2.495376675396859e-06, + "loss": 0.5802, + "step": 740 + }, + { + "epoch": 0.06314443971026842, + "grad_norm": 11.875, + "learning_rate": 2.4953622251678133e-06, + "loss": 0.5409, + "step": 741 + }, + { + "epoch": 0.06322965487856838, + "grad_norm": 12.5625, + "learning_rate": 2.4953477524338194e-06, + "loss": 0.6038, + "step": 742 + }, + { + "epoch": 0.06331487004686834, + "grad_norm": 13.0, + "learning_rate": 2.495333257195139e-06, + "loss": 0.5836, + "step": 743 + }, + { + "epoch": 0.0634000852151683, + "grad_norm": 17.5, + "learning_rate": 2.4953187394520348e-06, + "loss": 0.9286, + "step": 744 + }, + { + "epoch": 0.06348530038346825, + "grad_norm": 18.875, + "learning_rate": 2.4953041992047687e-06, + "loss": 0.8986, + "step": 745 + }, + { + "epoch": 0.06357051555176821, + "grad_norm": 11.125, + "learning_rate": 2.4952896364536032e-06, + "loss": 0.4778, + "step": 746 + }, + { + "epoch": 0.06365573072006817, + "grad_norm": 20.25, + "learning_rate": 2.495275051198802e-06, + "loss": 1.0113, + "step": 747 + }, + { + "epoch": 0.06374094588836814, + "grad_norm": 13.875, + "learning_rate": 2.4952604434406285e-06, + "loss": 0.9292, + "step": 748 + }, + { + "epoch": 0.0638261610566681, + "grad_norm": 11.9375, + "learning_rate": 2.495245813179346e-06, + "loss": 0.8457, + "step": 749 + }, + { + "epoch": 0.06391137622496805, + "grad_norm": 15.75, + "learning_rate": 2.49523116041522e-06, + "loss": 0.9048, + "step": 750 + }, + { + "epoch": 0.06399659139326801, + "grad_norm": 11.25, + "learning_rate": 2.4952164851485143e-06, + "loss": 0.5674, + "step": 751 + }, + { + "epoch": 0.06408180656156796, + "grad_norm": 14.9375, + "learning_rate": 2.495201787379495e-06, + "loss": 0.9348, + "step": 752 + }, + { + "epoch": 0.06416702172986792, + "grad_norm": 15.125, + "learning_rate": 2.495187067108427e-06, + "loss": 1.1715, + "step": 753 + }, + { + "epoch": 0.06425223689816788, + "grad_norm": 14.5, + "learning_rate": 2.495172324335577e-06, + "loss": 0.5497, + "step": 754 + }, + { + "epoch": 0.06433745206646783, + "grad_norm": 13.375, + "learning_rate": 2.4951575590612106e-06, + "loss": 0.503, + "step": 755 + }, + { + "epoch": 0.06442266723476779, + "grad_norm": 15.375, + "learning_rate": 2.4951427712855952e-06, + "loss": 1.2421, + "step": 756 + }, + { + "epoch": 0.06450788240306775, + "grad_norm": 13.25, + "learning_rate": 2.4951279610089978e-06, + "loss": 0.5308, + "step": 757 + }, + { + "epoch": 0.0645930975713677, + "grad_norm": 19.625, + "learning_rate": 2.4951131282316858e-06, + "loss": 1.1092, + "step": 758 + }, + { + "epoch": 0.06467831273966766, + "grad_norm": 13.3125, + "learning_rate": 2.4950982729539274e-06, + "loss": 0.8727, + "step": 759 + }, + { + "epoch": 0.06476352790796762, + "grad_norm": 13.6875, + "learning_rate": 2.495083395175992e-06, + "loss": 0.8149, + "step": 760 + }, + { + "epoch": 0.06484874307626758, + "grad_norm": 13.875, + "learning_rate": 2.495068494898147e-06, + "loss": 0.6574, + "step": 761 + }, + { + "epoch": 0.06493395824456753, + "grad_norm": 16.0, + "learning_rate": 2.495053572120662e-06, + "loss": 0.8122, + "step": 762 + }, + { + "epoch": 0.06501917341286749, + "grad_norm": 12.1875, + "learning_rate": 2.4950386268438074e-06, + "loss": 0.6436, + "step": 763 + }, + { + "epoch": 0.06510438858116745, + "grad_norm": 10.4375, + "learning_rate": 2.4950236590678526e-06, + "loss": 0.3646, + "step": 764 + }, + { + "epoch": 0.0651896037494674, + "grad_norm": 20.875, + "learning_rate": 2.4950086687930684e-06, + "loss": 0.7299, + "step": 765 + }, + { + "epoch": 0.06527481891776736, + "grad_norm": 14.5, + "learning_rate": 2.494993656019726e-06, + "loss": 0.8499, + "step": 766 + }, + { + "epoch": 0.06536003408606732, + "grad_norm": 18.375, + "learning_rate": 2.494978620748096e-06, + "loss": 1.2044, + "step": 767 + }, + { + "epoch": 0.06544524925436727, + "grad_norm": 13.8125, + "learning_rate": 2.49496356297845e-06, + "loss": 0.4475, + "step": 768 + }, + { + "epoch": 0.06553046442266723, + "grad_norm": 11.25, + "learning_rate": 2.494948482711061e-06, + "loss": 0.5462, + "step": 769 + }, + { + "epoch": 0.06561567959096719, + "grad_norm": 15.125, + "learning_rate": 2.494933379946201e-06, + "loss": 0.9137, + "step": 770 + }, + { + "epoch": 0.06570089475926714, + "grad_norm": 15.5, + "learning_rate": 2.4949182546841426e-06, + "loss": 1.0257, + "step": 771 + }, + { + "epoch": 0.0657861099275671, + "grad_norm": 13.6875, + "learning_rate": 2.49490310692516e-06, + "loss": 0.9288, + "step": 772 + }, + { + "epoch": 0.06587132509586706, + "grad_norm": 13.625, + "learning_rate": 2.494887936669526e-06, + "loss": 0.7343, + "step": 773 + }, + { + "epoch": 0.06595654026416702, + "grad_norm": 19.0, + "learning_rate": 2.4948727439175153e-06, + "loss": 0.9281, + "step": 774 + }, + { + "epoch": 0.06604175543246697, + "grad_norm": 19.0, + "learning_rate": 2.4948575286694022e-06, + "loss": 0.632, + "step": 775 + }, + { + "epoch": 0.06612697060076694, + "grad_norm": 17.25, + "learning_rate": 2.494842290925462e-06, + "loss": 0.7674, + "step": 776 + }, + { + "epoch": 0.0662121857690669, + "grad_norm": 11.625, + "learning_rate": 2.4948270306859695e-06, + "loss": 0.5834, + "step": 777 + }, + { + "epoch": 0.06629740093736686, + "grad_norm": 16.5, + "learning_rate": 2.494811747951201e-06, + "loss": 0.7233, + "step": 778 + }, + { + "epoch": 0.06638261610566681, + "grad_norm": 12.3125, + "learning_rate": 2.4947964427214326e-06, + "loss": 0.75, + "step": 779 + }, + { + "epoch": 0.06646783127396677, + "grad_norm": 14.625, + "learning_rate": 2.49478111499694e-06, + "loss": 0.591, + "step": 780 + }, + { + "epoch": 0.06655304644226673, + "grad_norm": 16.5, + "learning_rate": 2.494765764778002e-06, + "loss": 0.8904, + "step": 781 + }, + { + "epoch": 0.06663826161056668, + "grad_norm": 16.75, + "learning_rate": 2.494750392064894e-06, + "loss": 0.9807, + "step": 782 + }, + { + "epoch": 0.06672347677886664, + "grad_norm": 34.5, + "learning_rate": 2.4947349968578956e-06, + "loss": 0.9434, + "step": 783 + }, + { + "epoch": 0.0668086919471666, + "grad_norm": 10.5625, + "learning_rate": 2.494719579157283e-06, + "loss": 0.5366, + "step": 784 + }, + { + "epoch": 0.06689390711546656, + "grad_norm": 13.1875, + "learning_rate": 2.494704138963337e-06, + "loss": 0.648, + "step": 785 + }, + { + "epoch": 0.06697912228376651, + "grad_norm": 14.9375, + "learning_rate": 2.494688676276335e-06, + "loss": 1.1552, + "step": 786 + }, + { + "epoch": 0.06706433745206647, + "grad_norm": 14.4375, + "learning_rate": 2.494673191096557e-06, + "loss": 1.0506, + "step": 787 + }, + { + "epoch": 0.06714955262036643, + "grad_norm": 11.0, + "learning_rate": 2.494657683424283e-06, + "loss": 0.4094, + "step": 788 + }, + { + "epoch": 0.06723476778866638, + "grad_norm": 17.0, + "learning_rate": 2.494642153259793e-06, + "loss": 0.9257, + "step": 789 + }, + { + "epoch": 0.06731998295696634, + "grad_norm": 24.375, + "learning_rate": 2.494626600603368e-06, + "loss": 0.6784, + "step": 790 + }, + { + "epoch": 0.0674051981252663, + "grad_norm": 14.125, + "learning_rate": 2.4946110254552886e-06, + "loss": 0.6307, + "step": 791 + }, + { + "epoch": 0.06749041329356625, + "grad_norm": 18.75, + "learning_rate": 2.494595427815836e-06, + "loss": 1.0362, + "step": 792 + }, + { + "epoch": 0.06757562846186621, + "grad_norm": 13.0625, + "learning_rate": 2.4945798076852928e-06, + "loss": 0.6927, + "step": 793 + }, + { + "epoch": 0.06766084363016617, + "grad_norm": 15.0, + "learning_rate": 2.494564165063941e-06, + "loss": 1.057, + "step": 794 + }, + { + "epoch": 0.06774605879846612, + "grad_norm": 16.125, + "learning_rate": 2.494548499952063e-06, + "loss": 0.5934, + "step": 795 + }, + { + "epoch": 0.06783127396676608, + "grad_norm": 13.75, + "learning_rate": 2.4945328123499424e-06, + "loss": 0.7026, + "step": 796 + }, + { + "epoch": 0.06791648913506604, + "grad_norm": 15.4375, + "learning_rate": 2.4945171022578622e-06, + "loss": 0.9534, + "step": 797 + }, + { + "epoch": 0.068001704303366, + "grad_norm": 12.0625, + "learning_rate": 2.4945013696761066e-06, + "loss": 0.6558, + "step": 798 + }, + { + "epoch": 0.06808691947166595, + "grad_norm": 35.25, + "learning_rate": 2.49448561460496e-06, + "loss": 1.0942, + "step": 799 + }, + { + "epoch": 0.06817213463996591, + "grad_norm": 18.625, + "learning_rate": 2.4944698370447066e-06, + "loss": 0.7719, + "step": 800 + }, + { + "epoch": 0.06825734980826587, + "grad_norm": 11.0625, + "learning_rate": 2.494454036995632e-06, + "loss": 0.7603, + "step": 801 + }, + { + "epoch": 0.06834256497656582, + "grad_norm": 13.125, + "learning_rate": 2.4944382144580213e-06, + "loss": 0.7476, + "step": 802 + }, + { + "epoch": 0.06842778014486578, + "grad_norm": 13.1875, + "learning_rate": 2.494422369432161e-06, + "loss": 0.3921, + "step": 803 + }, + { + "epoch": 0.06851299531316575, + "grad_norm": 15.75, + "learning_rate": 2.494406501918337e-06, + "loss": 1.0187, + "step": 804 + }, + { + "epoch": 0.06859821048146571, + "grad_norm": 13.625, + "learning_rate": 2.494390611916836e-06, + "loss": 0.6273, + "step": 805 + }, + { + "epoch": 0.06868342564976566, + "grad_norm": 13.5, + "learning_rate": 2.4943746994279456e-06, + "loss": 0.7789, + "step": 806 + }, + { + "epoch": 0.06876864081806562, + "grad_norm": 23.75, + "learning_rate": 2.4943587644519528e-06, + "loss": 1.0404, + "step": 807 + }, + { + "epoch": 0.06885385598636558, + "grad_norm": 9.25, + "learning_rate": 2.494342806989146e-06, + "loss": 0.3221, + "step": 808 + }, + { + "epoch": 0.06893907115466553, + "grad_norm": 21.75, + "learning_rate": 2.4943268270398135e-06, + "loss": 1.2056, + "step": 809 + }, + { + "epoch": 0.06902428632296549, + "grad_norm": 14.3125, + "learning_rate": 2.494310824604244e-06, + "loss": 0.5997, + "step": 810 + }, + { + "epoch": 0.06910950149126545, + "grad_norm": 13.6875, + "learning_rate": 2.494294799682726e-06, + "loss": 0.7444, + "step": 811 + }, + { + "epoch": 0.0691947166595654, + "grad_norm": 24.25, + "learning_rate": 2.494278752275551e-06, + "loss": 0.7059, + "step": 812 + }, + { + "epoch": 0.06927993182786536, + "grad_norm": 13.6875, + "learning_rate": 2.494262682383006e-06, + "loss": 0.9566, + "step": 813 + }, + { + "epoch": 0.06936514699616532, + "grad_norm": 10.0625, + "learning_rate": 2.4942465900053846e-06, + "loss": 0.5052, + "step": 814 + }, + { + "epoch": 0.06945036216446528, + "grad_norm": 6.25, + "learning_rate": 2.4942304751429757e-06, + "loss": 0.1778, + "step": 815 + }, + { + "epoch": 0.06953557733276523, + "grad_norm": 13.1875, + "learning_rate": 2.4942143377960708e-06, + "loss": 0.6495, + "step": 816 + }, + { + "epoch": 0.06962079250106519, + "grad_norm": 9.5625, + "learning_rate": 2.4941981779649614e-06, + "loss": 0.3381, + "step": 817 + }, + { + "epoch": 0.06970600766936515, + "grad_norm": 13.0625, + "learning_rate": 2.4941819956499404e-06, + "loss": 0.8703, + "step": 818 + }, + { + "epoch": 0.0697912228376651, + "grad_norm": 13.6875, + "learning_rate": 2.4941657908512994e-06, + "loss": 0.6306, + "step": 819 + }, + { + "epoch": 0.06987643800596506, + "grad_norm": 12.375, + "learning_rate": 2.494149563569331e-06, + "loss": 0.4808, + "step": 820 + }, + { + "epoch": 0.06996165317426502, + "grad_norm": 15.5, + "learning_rate": 2.4941333138043294e-06, + "loss": 0.7736, + "step": 821 + }, + { + "epoch": 0.07004686834256497, + "grad_norm": 13.6875, + "learning_rate": 2.4941170415565873e-06, + "loss": 0.6722, + "step": 822 + }, + { + "epoch": 0.07013208351086493, + "grad_norm": 11.6875, + "learning_rate": 2.4941007468264e-06, + "loss": 0.5138, + "step": 823 + }, + { + "epoch": 0.07021729867916489, + "grad_norm": 14.25, + "learning_rate": 2.4940844296140605e-06, + "loss": 0.6698, + "step": 824 + }, + { + "epoch": 0.07030251384746485, + "grad_norm": 15.1875, + "learning_rate": 2.4940680899198644e-06, + "loss": 0.8847, + "step": 825 + }, + { + "epoch": 0.0703877290157648, + "grad_norm": 14.625, + "learning_rate": 2.4940517277441073e-06, + "loss": 0.841, + "step": 826 + }, + { + "epoch": 0.07047294418406476, + "grad_norm": 12.25, + "learning_rate": 2.494035343087084e-06, + "loss": 0.4975, + "step": 827 + }, + { + "epoch": 0.07055815935236472, + "grad_norm": 14.25, + "learning_rate": 2.4940189359490915e-06, + "loss": 0.6772, + "step": 828 + }, + { + "epoch": 0.07064337452066467, + "grad_norm": 14.0625, + "learning_rate": 2.4940025063304256e-06, + "loss": 0.7616, + "step": 829 + }, + { + "epoch": 0.07072858968896463, + "grad_norm": 35.25, + "learning_rate": 2.4939860542313833e-06, + "loss": 1.4091, + "step": 830 + }, + { + "epoch": 0.07081380485726459, + "grad_norm": 11.8125, + "learning_rate": 2.4939695796522624e-06, + "loss": 0.5019, + "step": 831 + }, + { + "epoch": 0.07089902002556454, + "grad_norm": 15.4375, + "learning_rate": 2.49395308259336e-06, + "loss": 0.656, + "step": 832 + }, + { + "epoch": 0.07098423519386451, + "grad_norm": 18.375, + "learning_rate": 2.493936563054975e-06, + "loss": 0.8033, + "step": 833 + }, + { + "epoch": 0.07106945036216447, + "grad_norm": 16.5, + "learning_rate": 2.493920021037405e-06, + "loss": 1.0798, + "step": 834 + }, + { + "epoch": 0.07115466553046443, + "grad_norm": 12.375, + "learning_rate": 2.4939034565409493e-06, + "loss": 0.5937, + "step": 835 + }, + { + "epoch": 0.07123988069876439, + "grad_norm": 11.9375, + "learning_rate": 2.493886869565908e-06, + "loss": 0.661, + "step": 836 + }, + { + "epoch": 0.07132509586706434, + "grad_norm": 27.25, + "learning_rate": 2.4938702601125792e-06, + "loss": 1.5583, + "step": 837 + }, + { + "epoch": 0.0714103110353643, + "grad_norm": 12.3125, + "learning_rate": 2.4938536281812645e-06, + "loss": 0.5201, + "step": 838 + }, + { + "epoch": 0.07149552620366426, + "grad_norm": 15.4375, + "learning_rate": 2.493836973772264e-06, + "loss": 1.0613, + "step": 839 + }, + { + "epoch": 0.07158074137196421, + "grad_norm": 12.4375, + "learning_rate": 2.4938202968858784e-06, + "loss": 0.6031, + "step": 840 + }, + { + "epoch": 0.07166595654026417, + "grad_norm": 14.6875, + "learning_rate": 2.4938035975224092e-06, + "loss": 0.6495, + "step": 841 + }, + { + "epoch": 0.07175117170856413, + "grad_norm": 20.125, + "learning_rate": 2.4937868756821585e-06, + "loss": 0.77, + "step": 842 + }, + { + "epoch": 0.07183638687686408, + "grad_norm": 11.875, + "learning_rate": 2.493770131365428e-06, + "loss": 0.5273, + "step": 843 + }, + { + "epoch": 0.07192160204516404, + "grad_norm": 27.0, + "learning_rate": 2.493753364572521e-06, + "loss": 1.2015, + "step": 844 + }, + { + "epoch": 0.072006817213464, + "grad_norm": 13.4375, + "learning_rate": 2.4937365753037394e-06, + "loss": 0.7516, + "step": 845 + }, + { + "epoch": 0.07209203238176395, + "grad_norm": 15.3125, + "learning_rate": 2.4937197635593875e-06, + "loss": 1.0393, + "step": 846 + }, + { + "epoch": 0.07217724755006391, + "grad_norm": 16.625, + "learning_rate": 2.4937029293397685e-06, + "loss": 1.0639, + "step": 847 + }, + { + "epoch": 0.07226246271836387, + "grad_norm": 14.375, + "learning_rate": 2.4936860726451874e-06, + "loss": 0.6474, + "step": 848 + }, + { + "epoch": 0.07234767788666382, + "grad_norm": 13.3125, + "learning_rate": 2.493669193475948e-06, + "loss": 0.8118, + "step": 849 + }, + { + "epoch": 0.07243289305496378, + "grad_norm": 13.5625, + "learning_rate": 2.493652291832356e-06, + "loss": 0.8965, + "step": 850 + }, + { + "epoch": 0.07251810822326374, + "grad_norm": 9.8125, + "learning_rate": 2.4936353677147165e-06, + "loss": 0.3174, + "step": 851 + }, + { + "epoch": 0.0726033233915637, + "grad_norm": 14.3125, + "learning_rate": 2.493618421123335e-06, + "loss": 1.0894, + "step": 852 + }, + { + "epoch": 0.07268853855986365, + "grad_norm": 13.0625, + "learning_rate": 2.4936014520585185e-06, + "loss": 0.7397, + "step": 853 + }, + { + "epoch": 0.07277375372816361, + "grad_norm": 12.9375, + "learning_rate": 2.493584460520573e-06, + "loss": 0.711, + "step": 854 + }, + { + "epoch": 0.07285896889646357, + "grad_norm": 13.8125, + "learning_rate": 2.4935674465098054e-06, + "loss": 0.7559, + "step": 855 + }, + { + "epoch": 0.07294418406476352, + "grad_norm": 16.5, + "learning_rate": 2.493550410026524e-06, + "loss": 0.669, + "step": 856 + }, + { + "epoch": 0.07302939923306348, + "grad_norm": 16.5, + "learning_rate": 2.493533351071036e-06, + "loss": 1.0556, + "step": 857 + }, + { + "epoch": 0.07311461440136344, + "grad_norm": 12.1875, + "learning_rate": 2.4935162696436497e-06, + "loss": 0.7872, + "step": 858 + }, + { + "epoch": 0.0731998295696634, + "grad_norm": 17.625, + "learning_rate": 2.4934991657446743e-06, + "loss": 0.8275, + "step": 859 + }, + { + "epoch": 0.07328504473796335, + "grad_norm": 11.3125, + "learning_rate": 2.4934820393744187e-06, + "loss": 0.6826, + "step": 860 + }, + { + "epoch": 0.07337025990626332, + "grad_norm": 13.4375, + "learning_rate": 2.4934648905331923e-06, + "loss": 0.8516, + "step": 861 + }, + { + "epoch": 0.07345547507456328, + "grad_norm": 14.0625, + "learning_rate": 2.493447719221304e-06, + "loss": 0.9051, + "step": 862 + }, + { + "epoch": 0.07354069024286324, + "grad_norm": 19.25, + "learning_rate": 2.4934305254390657e-06, + "loss": 0.9234, + "step": 863 + }, + { + "epoch": 0.07362590541116319, + "grad_norm": 13.8125, + "learning_rate": 2.4934133091867873e-06, + "loss": 0.6714, + "step": 864 + }, + { + "epoch": 0.07371112057946315, + "grad_norm": 13.75, + "learning_rate": 2.49339607046478e-06, + "loss": 0.6635, + "step": 865 + }, + { + "epoch": 0.0737963357477631, + "grad_norm": 15.3125, + "learning_rate": 2.4933788092733556e-06, + "loss": 0.9683, + "step": 866 + }, + { + "epoch": 0.07388155091606306, + "grad_norm": 17.375, + "learning_rate": 2.4933615256128256e-06, + "loss": 1.2642, + "step": 867 + }, + { + "epoch": 0.07396676608436302, + "grad_norm": 15.125, + "learning_rate": 2.493344219483503e-06, + "loss": 1.0422, + "step": 868 + }, + { + "epoch": 0.07405198125266298, + "grad_norm": 17.5, + "learning_rate": 2.4933268908857e-06, + "loss": 0.8265, + "step": 869 + }, + { + "epoch": 0.07413719642096293, + "grad_norm": 13.25, + "learning_rate": 2.493309539819729e-06, + "loss": 0.8263, + "step": 870 + }, + { + "epoch": 0.07422241158926289, + "grad_norm": 15.75, + "learning_rate": 2.493292166285905e-06, + "loss": 0.7975, + "step": 871 + }, + { + "epoch": 0.07430762675756285, + "grad_norm": 12.375, + "learning_rate": 2.4932747702845413e-06, + "loss": 0.5741, + "step": 872 + }, + { + "epoch": 0.0743928419258628, + "grad_norm": 17.0, + "learning_rate": 2.4932573518159522e-06, + "loss": 0.9232, + "step": 873 + }, + { + "epoch": 0.07447805709416276, + "grad_norm": 12.5, + "learning_rate": 2.493239910880453e-06, + "loss": 0.5227, + "step": 874 + }, + { + "epoch": 0.07456327226246272, + "grad_norm": 14.125, + "learning_rate": 2.493222447478358e-06, + "loss": 0.6978, + "step": 875 + }, + { + "epoch": 0.07464848743076268, + "grad_norm": 15.875, + "learning_rate": 2.493204961609983e-06, + "loss": 1.008, + "step": 876 + }, + { + "epoch": 0.07473370259906263, + "grad_norm": 19.25, + "learning_rate": 2.4931874532756447e-06, + "loss": 0.8061, + "step": 877 + }, + { + "epoch": 0.07481891776736259, + "grad_norm": 12.9375, + "learning_rate": 2.493169922475658e-06, + "loss": 0.7792, + "step": 878 + }, + { + "epoch": 0.07490413293566255, + "grad_norm": 17.75, + "learning_rate": 2.493152369210342e-06, + "loss": 1.2115, + "step": 879 + }, + { + "epoch": 0.0749893481039625, + "grad_norm": 15.625, + "learning_rate": 2.493134793480012e-06, + "loss": 0.8823, + "step": 880 + }, + { + "epoch": 0.07507456327226246, + "grad_norm": 24.875, + "learning_rate": 2.493117195284986e-06, + "loss": 1.1535, + "step": 881 + }, + { + "epoch": 0.07515977844056242, + "grad_norm": 22.25, + "learning_rate": 2.493099574625583e-06, + "loss": 1.3755, + "step": 882 + }, + { + "epoch": 0.07524499360886237, + "grad_norm": 13.75, + "learning_rate": 2.49308193150212e-06, + "loss": 0.935, + "step": 883 + }, + { + "epoch": 0.07533020877716233, + "grad_norm": 15.8125, + "learning_rate": 2.493064265914916e-06, + "loss": 1.0063, + "step": 884 + }, + { + "epoch": 0.07541542394546229, + "grad_norm": 12.0, + "learning_rate": 2.4930465778642916e-06, + "loss": 0.5326, + "step": 885 + }, + { + "epoch": 0.07550063911376224, + "grad_norm": 12.75, + "learning_rate": 2.4930288673505652e-06, + "loss": 0.6159, + "step": 886 + }, + { + "epoch": 0.0755858542820622, + "grad_norm": 21.5, + "learning_rate": 2.4930111343740574e-06, + "loss": 0.887, + "step": 887 + }, + { + "epoch": 0.07567106945036216, + "grad_norm": 13.1875, + "learning_rate": 2.4929933789350884e-06, + "loss": 0.9039, + "step": 888 + }, + { + "epoch": 0.07575628461866213, + "grad_norm": 18.625, + "learning_rate": 2.492975601033979e-06, + "loss": 0.9532, + "step": 889 + }, + { + "epoch": 0.07584149978696209, + "grad_norm": 17.0, + "learning_rate": 2.4929578006710508e-06, + "loss": 0.6734, + "step": 890 + }, + { + "epoch": 0.07592671495526204, + "grad_norm": 18.25, + "learning_rate": 2.492939977846625e-06, + "loss": 0.8263, + "step": 891 + }, + { + "epoch": 0.076011930123562, + "grad_norm": 21.875, + "learning_rate": 2.4929221325610243e-06, + "loss": 1.0999, + "step": 892 + }, + { + "epoch": 0.07609714529186196, + "grad_norm": 17.0, + "learning_rate": 2.4929042648145705e-06, + "loss": 0.7115, + "step": 893 + }, + { + "epoch": 0.07618236046016191, + "grad_norm": 14.3125, + "learning_rate": 2.4928863746075867e-06, + "loss": 0.6361, + "step": 894 + }, + { + "epoch": 0.07626757562846187, + "grad_norm": 13.75, + "learning_rate": 2.4928684619403964e-06, + "loss": 1.11, + "step": 895 + }, + { + "epoch": 0.07635279079676183, + "grad_norm": 17.125, + "learning_rate": 2.4928505268133234e-06, + "loss": 0.8011, + "step": 896 + }, + { + "epoch": 0.07643800596506178, + "grad_norm": 12.9375, + "learning_rate": 2.4928325692266914e-06, + "loss": 0.3879, + "step": 897 + }, + { + "epoch": 0.07652322113336174, + "grad_norm": 18.75, + "learning_rate": 2.492814589180825e-06, + "loss": 1.1808, + "step": 898 + }, + { + "epoch": 0.0766084363016617, + "grad_norm": 15.875, + "learning_rate": 2.4927965866760494e-06, + "loss": 1.0175, + "step": 899 + }, + { + "epoch": 0.07669365146996165, + "grad_norm": 18.25, + "learning_rate": 2.49277856171269e-06, + "loss": 1.067, + "step": 900 + }, + { + "epoch": 0.07677886663826161, + "grad_norm": 13.4375, + "learning_rate": 2.4927605142910716e-06, + "loss": 0.6143, + "step": 901 + }, + { + "epoch": 0.07686408180656157, + "grad_norm": 14.25, + "learning_rate": 2.4927424444115217e-06, + "loss": 0.7555, + "step": 902 + }, + { + "epoch": 0.07694929697486153, + "grad_norm": 18.375, + "learning_rate": 2.492724352074366e-06, + "loss": 0.511, + "step": 903 + }, + { + "epoch": 0.07703451214316148, + "grad_norm": 12.25, + "learning_rate": 2.492706237279931e-06, + "loss": 0.3614, + "step": 904 + }, + { + "epoch": 0.07711972731146144, + "grad_norm": 16.875, + "learning_rate": 2.4926881000285453e-06, + "loss": 0.9427, + "step": 905 + }, + { + "epoch": 0.0772049424797614, + "grad_norm": 13.625, + "learning_rate": 2.4926699403205356e-06, + "loss": 0.8772, + "step": 906 + }, + { + "epoch": 0.07729015764806135, + "grad_norm": 13.5, + "learning_rate": 2.492651758156231e-06, + "loss": 0.6097, + "step": 907 + }, + { + "epoch": 0.07737537281636131, + "grad_norm": 15.3125, + "learning_rate": 2.492633553535959e-06, + "loss": 1.0343, + "step": 908 + }, + { + "epoch": 0.07746058798466127, + "grad_norm": 25.375, + "learning_rate": 2.4926153264600494e-06, + "loss": 0.4935, + "step": 909 + }, + { + "epoch": 0.07754580315296122, + "grad_norm": 12.5625, + "learning_rate": 2.4925970769288314e-06, + "loss": 0.5237, + "step": 910 + }, + { + "epoch": 0.07763101832126118, + "grad_norm": 15.6875, + "learning_rate": 2.492578804942635e-06, + "loss": 0.8587, + "step": 911 + }, + { + "epoch": 0.07771623348956114, + "grad_norm": 16.0, + "learning_rate": 2.4925605105017893e-06, + "loss": 0.9437, + "step": 912 + }, + { + "epoch": 0.0778014486578611, + "grad_norm": 15.625, + "learning_rate": 2.492542193606626e-06, + "loss": 0.8953, + "step": 913 + }, + { + "epoch": 0.07788666382616105, + "grad_norm": 20.875, + "learning_rate": 2.4925238542574758e-06, + "loss": 0.9435, + "step": 914 + }, + { + "epoch": 0.07797187899446101, + "grad_norm": 17.0, + "learning_rate": 2.4925054924546702e-06, + "loss": 1.2316, + "step": 915 + }, + { + "epoch": 0.07805709416276096, + "grad_norm": 23.875, + "learning_rate": 2.492487108198541e-06, + "loss": 1.1427, + "step": 916 + }, + { + "epoch": 0.07814230933106094, + "grad_norm": 12.125, + "learning_rate": 2.49246870148942e-06, + "loss": 0.6263, + "step": 917 + }, + { + "epoch": 0.07822752449936089, + "grad_norm": 10.5, + "learning_rate": 2.4924502723276403e-06, + "loss": 0.4651, + "step": 918 + }, + { + "epoch": 0.07831273966766085, + "grad_norm": 13.25, + "learning_rate": 2.492431820713535e-06, + "loss": 0.3829, + "step": 919 + }, + { + "epoch": 0.0783979548359608, + "grad_norm": 11.5, + "learning_rate": 2.4924133466474372e-06, + "loss": 0.4553, + "step": 920 + }, + { + "epoch": 0.07848317000426076, + "grad_norm": 17.125, + "learning_rate": 2.4923948501296805e-06, + "loss": 0.7863, + "step": 921 + }, + { + "epoch": 0.07856838517256072, + "grad_norm": 16.25, + "learning_rate": 2.4923763311606003e-06, + "loss": 0.8902, + "step": 922 + }, + { + "epoch": 0.07865360034086068, + "grad_norm": 12.875, + "learning_rate": 2.4923577897405297e-06, + "loss": 0.8082, + "step": 923 + }, + { + "epoch": 0.07873881550916063, + "grad_norm": 16.875, + "learning_rate": 2.492339225869805e-06, + "loss": 0.9792, + "step": 924 + }, + { + "epoch": 0.07882403067746059, + "grad_norm": 16.125, + "learning_rate": 2.4923206395487616e-06, + "loss": 1.0222, + "step": 925 + }, + { + "epoch": 0.07890924584576055, + "grad_norm": 14.8125, + "learning_rate": 2.4923020307777344e-06, + "loss": 0.6075, + "step": 926 + }, + { + "epoch": 0.0789944610140605, + "grad_norm": 12.625, + "learning_rate": 2.4922833995570604e-06, + "loss": 0.7074, + "step": 927 + }, + { + "epoch": 0.07907967618236046, + "grad_norm": 14.5625, + "learning_rate": 2.492264745887076e-06, + "loss": 0.783, + "step": 928 + }, + { + "epoch": 0.07916489135066042, + "grad_norm": 16.875, + "learning_rate": 2.492246069768119e-06, + "loss": 0.7242, + "step": 929 + }, + { + "epoch": 0.07925010651896038, + "grad_norm": 11.625, + "learning_rate": 2.4922273712005256e-06, + "loss": 0.4453, + "step": 930 + }, + { + "epoch": 0.07933532168726033, + "grad_norm": 22.625, + "learning_rate": 2.492208650184635e-06, + "loss": 1.0361, + "step": 931 + }, + { + "epoch": 0.07942053685556029, + "grad_norm": 14.875, + "learning_rate": 2.492189906720785e-06, + "loss": 0.9549, + "step": 932 + }, + { + "epoch": 0.07950575202386025, + "grad_norm": 10.625, + "learning_rate": 2.4921711408093137e-06, + "loss": 0.4851, + "step": 933 + }, + { + "epoch": 0.0795909671921602, + "grad_norm": 10.875, + "learning_rate": 2.492152352450561e-06, + "loss": 0.3842, + "step": 934 + }, + { + "epoch": 0.07967618236046016, + "grad_norm": 18.5, + "learning_rate": 2.4921335416448668e-06, + "loss": 1.4054, + "step": 935 + }, + { + "epoch": 0.07976139752876012, + "grad_norm": 13.5625, + "learning_rate": 2.4921147083925695e-06, + "loss": 0.7233, + "step": 936 + }, + { + "epoch": 0.07984661269706007, + "grad_norm": 11.375, + "learning_rate": 2.492095852694011e-06, + "loss": 0.7253, + "step": 937 + }, + { + "epoch": 0.07993182786536003, + "grad_norm": 18.875, + "learning_rate": 2.492076974549531e-06, + "loss": 0.893, + "step": 938 + }, + { + "epoch": 0.08001704303365999, + "grad_norm": 18.75, + "learning_rate": 2.492058073959472e-06, + "loss": 0.863, + "step": 939 + }, + { + "epoch": 0.08010225820195994, + "grad_norm": 21.25, + "learning_rate": 2.492039150924174e-06, + "loss": 1.0865, + "step": 940 + }, + { + "epoch": 0.0801874733702599, + "grad_norm": 15.8125, + "learning_rate": 2.492020205443979e-06, + "loss": 1.0112, + "step": 941 + }, + { + "epoch": 0.08027268853855986, + "grad_norm": 11.3125, + "learning_rate": 2.4920012375192305e-06, + "loss": 0.53, + "step": 942 + }, + { + "epoch": 0.08035790370685982, + "grad_norm": 11.9375, + "learning_rate": 2.491982247150271e-06, + "loss": 0.679, + "step": 943 + }, + { + "epoch": 0.08044311887515977, + "grad_norm": 16.375, + "learning_rate": 2.491963234337443e-06, + "loss": 0.9403, + "step": 944 + }, + { + "epoch": 0.08052833404345973, + "grad_norm": 14.1875, + "learning_rate": 2.4919441990810904e-06, + "loss": 0.8987, + "step": 945 + }, + { + "epoch": 0.0806135492117597, + "grad_norm": 14.8125, + "learning_rate": 2.4919251413815577e-06, + "loss": 0.9483, + "step": 946 + }, + { + "epoch": 0.08069876438005966, + "grad_norm": 20.75, + "learning_rate": 2.4919060612391882e-06, + "loss": 1.0219, + "step": 947 + }, + { + "epoch": 0.08078397954835961, + "grad_norm": 14.0625, + "learning_rate": 2.4918869586543276e-06, + "loss": 0.8508, + "step": 948 + }, + { + "epoch": 0.08086919471665957, + "grad_norm": 13.5625, + "learning_rate": 2.491867833627321e-06, + "loss": 0.8654, + "step": 949 + }, + { + "epoch": 0.08095440988495953, + "grad_norm": 11.375, + "learning_rate": 2.491848686158514e-06, + "loss": 0.5309, + "step": 950 + }, + { + "epoch": 0.08103962505325948, + "grad_norm": 17.875, + "learning_rate": 2.491829516248252e-06, + "loss": 0.9581, + "step": 951 + }, + { + "epoch": 0.08112484022155944, + "grad_norm": 15.3125, + "learning_rate": 2.491810323896882e-06, + "loss": 0.9936, + "step": 952 + }, + { + "epoch": 0.0812100553898594, + "grad_norm": 16.5, + "learning_rate": 2.4917911091047507e-06, + "loss": 0.9531, + "step": 953 + }, + { + "epoch": 0.08129527055815935, + "grad_norm": 9.875, + "learning_rate": 2.4917718718722056e-06, + "loss": 0.3583, + "step": 954 + }, + { + "epoch": 0.08138048572645931, + "grad_norm": 11.75, + "learning_rate": 2.491752612199594e-06, + "loss": 0.6067, + "step": 955 + }, + { + "epoch": 0.08146570089475927, + "grad_norm": 15.1875, + "learning_rate": 2.491733330087264e-06, + "loss": 0.9601, + "step": 956 + }, + { + "epoch": 0.08155091606305923, + "grad_norm": 13.8125, + "learning_rate": 2.491714025535564e-06, + "loss": 0.8267, + "step": 957 + }, + { + "epoch": 0.08163613123135918, + "grad_norm": 11.0, + "learning_rate": 2.491694698544843e-06, + "loss": 0.5244, + "step": 958 + }, + { + "epoch": 0.08172134639965914, + "grad_norm": 12.375, + "learning_rate": 2.49167534911545e-06, + "loss": 0.5638, + "step": 959 + }, + { + "epoch": 0.0818065615679591, + "grad_norm": 16.625, + "learning_rate": 2.491655977247735e-06, + "loss": 0.8481, + "step": 960 + }, + { + "epoch": 0.08189177673625905, + "grad_norm": 21.125, + "learning_rate": 2.491636582942048e-06, + "loss": 0.7118, + "step": 961 + }, + { + "epoch": 0.08197699190455901, + "grad_norm": 24.625, + "learning_rate": 2.491617166198739e-06, + "loss": 1.2096, + "step": 962 + }, + { + "epoch": 0.08206220707285897, + "grad_norm": 19.25, + "learning_rate": 2.4915977270181594e-06, + "loss": 1.4455, + "step": 963 + }, + { + "epoch": 0.08214742224115892, + "grad_norm": 19.75, + "learning_rate": 2.491578265400661e-06, + "loss": 1.0058, + "step": 964 + }, + { + "epoch": 0.08223263740945888, + "grad_norm": 25.5, + "learning_rate": 2.4915587813465943e-06, + "loss": 0.9515, + "step": 965 + }, + { + "epoch": 0.08231785257775884, + "grad_norm": 14.125, + "learning_rate": 2.4915392748563115e-06, + "loss": 0.8341, + "step": 966 + }, + { + "epoch": 0.0824030677460588, + "grad_norm": 18.625, + "learning_rate": 2.4915197459301664e-06, + "loss": 0.7962, + "step": 967 + }, + { + "epoch": 0.08248828291435875, + "grad_norm": 14.0625, + "learning_rate": 2.4915001945685106e-06, + "loss": 0.7048, + "step": 968 + }, + { + "epoch": 0.08257349808265871, + "grad_norm": 14.9375, + "learning_rate": 2.491480620771698e-06, + "loss": 0.5696, + "step": 969 + }, + { + "epoch": 0.08265871325095867, + "grad_norm": 14.25, + "learning_rate": 2.4914610245400816e-06, + "loss": 0.7926, + "step": 970 + }, + { + "epoch": 0.08274392841925862, + "grad_norm": 10.8125, + "learning_rate": 2.4914414058740166e-06, + "loss": 0.5785, + "step": 971 + }, + { + "epoch": 0.08282914358755858, + "grad_norm": 20.0, + "learning_rate": 2.491421764773857e-06, + "loss": 0.9236, + "step": 972 + }, + { + "epoch": 0.08291435875585854, + "grad_norm": 16.625, + "learning_rate": 2.491402101239958e-06, + "loss": 0.998, + "step": 973 + }, + { + "epoch": 0.0829995739241585, + "grad_norm": 11.0625, + "learning_rate": 2.4913824152726746e-06, + "loss": 0.3975, + "step": 974 + }, + { + "epoch": 0.08308478909245846, + "grad_norm": 19.75, + "learning_rate": 2.491362706872362e-06, + "loss": 0.9719, + "step": 975 + }, + { + "epoch": 0.08317000426075842, + "grad_norm": 13.5625, + "learning_rate": 2.491342976039378e-06, + "loss": 0.7411, + "step": 976 + }, + { + "epoch": 0.08325521942905838, + "grad_norm": 15.4375, + "learning_rate": 2.4913232227740775e-06, + "loss": 0.7216, + "step": 977 + }, + { + "epoch": 0.08334043459735833, + "grad_norm": 18.875, + "learning_rate": 2.4913034470768184e-06, + "loss": 1.3156, + "step": 978 + }, + { + "epoch": 0.08342564976565829, + "grad_norm": 11.875, + "learning_rate": 2.4912836489479575e-06, + "loss": 0.4551, + "step": 979 + }, + { + "epoch": 0.08351086493395825, + "grad_norm": 15.75, + "learning_rate": 2.4912638283878533e-06, + "loss": 1.2054, + "step": 980 + }, + { + "epoch": 0.0835960801022582, + "grad_norm": 16.625, + "learning_rate": 2.491243985396863e-06, + "loss": 1.133, + "step": 981 + }, + { + "epoch": 0.08368129527055816, + "grad_norm": 12.25, + "learning_rate": 2.491224119975346e-06, + "loss": 0.8804, + "step": 982 + }, + { + "epoch": 0.08376651043885812, + "grad_norm": 17.75, + "learning_rate": 2.4912042321236614e-06, + "loss": 0.7901, + "step": 983 + }, + { + "epoch": 0.08385172560715808, + "grad_norm": 13.125, + "learning_rate": 2.4911843218421673e-06, + "loss": 0.7225, + "step": 984 + }, + { + "epoch": 0.08393694077545803, + "grad_norm": 12.1875, + "learning_rate": 2.4911643891312247e-06, + "loss": 0.3106, + "step": 985 + }, + { + "epoch": 0.08402215594375799, + "grad_norm": 11.8125, + "learning_rate": 2.491144433991194e-06, + "loss": 1.1128, + "step": 986 + }, + { + "epoch": 0.08410737111205795, + "grad_norm": 11.9375, + "learning_rate": 2.491124456422435e-06, + "loss": 0.7801, + "step": 987 + }, + { + "epoch": 0.0841925862803579, + "grad_norm": 14.25, + "learning_rate": 2.491104456425309e-06, + "loss": 0.9965, + "step": 988 + }, + { + "epoch": 0.08427780144865786, + "grad_norm": 11.0625, + "learning_rate": 2.4910844340001773e-06, + "loss": 0.5957, + "step": 989 + }, + { + "epoch": 0.08436301661695782, + "grad_norm": 15.875, + "learning_rate": 2.4910643891474024e-06, + "loss": 0.7323, + "step": 990 + }, + { + "epoch": 0.08444823178525777, + "grad_norm": 16.125, + "learning_rate": 2.4910443218673453e-06, + "loss": 0.9596, + "step": 991 + }, + { + "epoch": 0.08453344695355773, + "grad_norm": 16.25, + "learning_rate": 2.49102423216037e-06, + "loss": 0.764, + "step": 992 + }, + { + "epoch": 0.08461866212185769, + "grad_norm": 14.375, + "learning_rate": 2.4910041200268383e-06, + "loss": 0.7823, + "step": 993 + }, + { + "epoch": 0.08470387729015764, + "grad_norm": 14.8125, + "learning_rate": 2.4909839854671147e-06, + "loss": 0.7277, + "step": 994 + }, + { + "epoch": 0.0847890924584576, + "grad_norm": 11.5, + "learning_rate": 2.4909638284815624e-06, + "loss": 0.5851, + "step": 995 + }, + { + "epoch": 0.08487430762675756, + "grad_norm": 19.875, + "learning_rate": 2.4909436490705456e-06, + "loss": 1.0849, + "step": 996 + }, + { + "epoch": 0.08495952279505752, + "grad_norm": 20.125, + "learning_rate": 2.4909234472344297e-06, + "loss": 1.1363, + "step": 997 + }, + { + "epoch": 0.08504473796335747, + "grad_norm": 30.375, + "learning_rate": 2.4909032229735786e-06, + "loss": 1.4282, + "step": 998 + }, + { + "epoch": 0.08512995313165743, + "grad_norm": 11.1875, + "learning_rate": 2.4908829762883586e-06, + "loss": 0.7139, + "step": 999 + }, + { + "epoch": 0.08521516829995739, + "grad_norm": 13.0625, + "learning_rate": 2.4908627071791357e-06, + "loss": 0.5121, + "step": 1000 + }, + { + "epoch": 0.08530038346825734, + "grad_norm": 15.8125, + "learning_rate": 2.4908424156462757e-06, + "loss": 0.9741, + "step": 1001 + }, + { + "epoch": 0.08538559863655731, + "grad_norm": 21.25, + "learning_rate": 2.490822101690146e-06, + "loss": 1.3311, + "step": 1002 + }, + { + "epoch": 0.08547081380485727, + "grad_norm": 15.5, + "learning_rate": 2.4908017653111125e-06, + "loss": 0.6252, + "step": 1003 + }, + { + "epoch": 0.08555602897315723, + "grad_norm": 15.875, + "learning_rate": 2.4907814065095437e-06, + "loss": 1.0719, + "step": 1004 + }, + { + "epoch": 0.08564124414145718, + "grad_norm": 13.1875, + "learning_rate": 2.490761025285807e-06, + "loss": 0.6015, + "step": 1005 + }, + { + "epoch": 0.08572645930975714, + "grad_norm": 15.3125, + "learning_rate": 2.4907406216402714e-06, + "loss": 0.9579, + "step": 1006 + }, + { + "epoch": 0.0858116744780571, + "grad_norm": 15.0625, + "learning_rate": 2.4907201955733047e-06, + "loss": 0.9232, + "step": 1007 + }, + { + "epoch": 0.08589688964635706, + "grad_norm": 12.6875, + "learning_rate": 2.490699747085276e-06, + "loss": 0.7117, + "step": 1008 + }, + { + "epoch": 0.08598210481465701, + "grad_norm": 17.75, + "learning_rate": 2.490679276176556e-06, + "loss": 0.7037, + "step": 1009 + }, + { + "epoch": 0.08606731998295697, + "grad_norm": 17.75, + "learning_rate": 2.4906587828475136e-06, + "loss": 0.9069, + "step": 1010 + }, + { + "epoch": 0.08615253515125693, + "grad_norm": 15.625, + "learning_rate": 2.490638267098519e-06, + "loss": 0.8875, + "step": 1011 + }, + { + "epoch": 0.08623775031955688, + "grad_norm": 15.375, + "learning_rate": 2.4906177289299438e-06, + "loss": 0.9406, + "step": 1012 + }, + { + "epoch": 0.08632296548785684, + "grad_norm": 11.625, + "learning_rate": 2.4905971683421588e-06, + "loss": 0.6685, + "step": 1013 + }, + { + "epoch": 0.0864081806561568, + "grad_norm": 14.375, + "learning_rate": 2.490576585335535e-06, + "loss": 0.8152, + "step": 1014 + }, + { + "epoch": 0.08649339582445675, + "grad_norm": 12.0625, + "learning_rate": 2.4905559799104457e-06, + "loss": 0.5096, + "step": 1015 + }, + { + "epoch": 0.08657861099275671, + "grad_norm": 12.5625, + "learning_rate": 2.4905353520672617e-06, + "loss": 0.8894, + "step": 1016 + }, + { + "epoch": 0.08666382616105667, + "grad_norm": 11.375, + "learning_rate": 2.4905147018063563e-06, + "loss": 0.5815, + "step": 1017 + }, + { + "epoch": 0.08674904132935662, + "grad_norm": 22.875, + "learning_rate": 2.490494029128103e-06, + "loss": 1.2529, + "step": 1018 + }, + { + "epoch": 0.08683425649765658, + "grad_norm": 14.3125, + "learning_rate": 2.490473334032875e-06, + "loss": 0.8467, + "step": 1019 + }, + { + "epoch": 0.08691947166595654, + "grad_norm": 12.125, + "learning_rate": 2.490452616521047e-06, + "loss": 0.7415, + "step": 1020 + }, + { + "epoch": 0.0870046868342565, + "grad_norm": 20.5, + "learning_rate": 2.4904318765929924e-06, + "loss": 1.1804, + "step": 1021 + }, + { + "epoch": 0.08708990200255645, + "grad_norm": 12.875, + "learning_rate": 2.4904111142490868e-06, + "loss": 0.4877, + "step": 1022 + }, + { + "epoch": 0.08717511717085641, + "grad_norm": 19.75, + "learning_rate": 2.4903903294897043e-06, + "loss": 1.1736, + "step": 1023 + }, + { + "epoch": 0.08726033233915637, + "grad_norm": 17.125, + "learning_rate": 2.4903695223152217e-06, + "loss": 1.029, + "step": 1024 + }, + { + "epoch": 0.08734554750745632, + "grad_norm": 13.125, + "learning_rate": 2.4903486927260145e-06, + "loss": 0.6916, + "step": 1025 + }, + { + "epoch": 0.08743076267575628, + "grad_norm": 17.875, + "learning_rate": 2.4903278407224597e-06, + "loss": 0.8215, + "step": 1026 + }, + { + "epoch": 0.08751597784405624, + "grad_norm": 10.625, + "learning_rate": 2.490306966304933e-06, + "loss": 0.7558, + "step": 1027 + }, + { + "epoch": 0.0876011930123562, + "grad_norm": 22.0, + "learning_rate": 2.490286069473812e-06, + "loss": 1.0058, + "step": 1028 + }, + { + "epoch": 0.08768640818065615, + "grad_norm": 10.625, + "learning_rate": 2.490265150229475e-06, + "loss": 0.4715, + "step": 1029 + }, + { + "epoch": 0.08777162334895612, + "grad_norm": 10.875, + "learning_rate": 2.4902442085722992e-06, + "loss": 0.4575, + "step": 1030 + }, + { + "epoch": 0.08785683851725608, + "grad_norm": 13.25, + "learning_rate": 2.4902232445026636e-06, + "loss": 0.5331, + "step": 1031 + }, + { + "epoch": 0.08794205368555603, + "grad_norm": 15.5625, + "learning_rate": 2.4902022580209465e-06, + "loss": 0.9019, + "step": 1032 + }, + { + "epoch": 0.08802726885385599, + "grad_norm": 14.625, + "learning_rate": 2.490181249127528e-06, + "loss": 0.768, + "step": 1033 + }, + { + "epoch": 0.08811248402215595, + "grad_norm": 14.5, + "learning_rate": 2.490160217822787e-06, + "loss": 0.8467, + "step": 1034 + }, + { + "epoch": 0.0881976991904559, + "grad_norm": 12.125, + "learning_rate": 2.490139164107104e-06, + "loss": 0.5999, + "step": 1035 + }, + { + "epoch": 0.08828291435875586, + "grad_norm": 14.0625, + "learning_rate": 2.490118087980859e-06, + "loss": 1.0654, + "step": 1036 + }, + { + "epoch": 0.08836812952705582, + "grad_norm": 14.3125, + "learning_rate": 2.4900969894444328e-06, + "loss": 0.7843, + "step": 1037 + }, + { + "epoch": 0.08845334469535578, + "grad_norm": 14.25, + "learning_rate": 2.4900758684982072e-06, + "loss": 0.6245, + "step": 1038 + }, + { + "epoch": 0.08853855986365573, + "grad_norm": 16.125, + "learning_rate": 2.4900547251425638e-06, + "loss": 0.3501, + "step": 1039 + }, + { + "epoch": 0.08862377503195569, + "grad_norm": 11.5625, + "learning_rate": 2.4900335593778844e-06, + "loss": 0.9506, + "step": 1040 + }, + { + "epoch": 0.08870899020025565, + "grad_norm": 13.625, + "learning_rate": 2.490012371204552e-06, + "loss": 1.1008, + "step": 1041 + }, + { + "epoch": 0.0887942053685556, + "grad_norm": 17.5, + "learning_rate": 2.4899911606229487e-06, + "loss": 0.8739, + "step": 1042 + }, + { + "epoch": 0.08887942053685556, + "grad_norm": 15.0625, + "learning_rate": 2.489969927633459e-06, + "loss": 0.4483, + "step": 1043 + }, + { + "epoch": 0.08896463570515552, + "grad_norm": 27.125, + "learning_rate": 2.4899486722364645e-06, + "loss": 1.1257, + "step": 1044 + }, + { + "epoch": 0.08904985087345547, + "grad_norm": 17.75, + "learning_rate": 2.4899273944323516e-06, + "loss": 1.1577, + "step": 1045 + }, + { + "epoch": 0.08913506604175543, + "grad_norm": 14.3125, + "learning_rate": 2.4899060942215037e-06, + "loss": 0.8145, + "step": 1046 + }, + { + "epoch": 0.08922028121005539, + "grad_norm": 13.4375, + "learning_rate": 2.4898847716043055e-06, + "loss": 0.5171, + "step": 1047 + }, + { + "epoch": 0.08930549637835535, + "grad_norm": 21.0, + "learning_rate": 2.489863426581143e-06, + "loss": 1.287, + "step": 1048 + }, + { + "epoch": 0.0893907115466553, + "grad_norm": 18.125, + "learning_rate": 2.4898420591524016e-06, + "loss": 0.7502, + "step": 1049 + }, + { + "epoch": 0.08947592671495526, + "grad_norm": 24.625, + "learning_rate": 2.4898206693184672e-06, + "loss": 1.3815, + "step": 1050 + }, + { + "epoch": 0.08956114188325522, + "grad_norm": 13.875, + "learning_rate": 2.489799257079727e-06, + "loss": 0.6419, + "step": 1051 + }, + { + "epoch": 0.08964635705155517, + "grad_norm": 19.25, + "learning_rate": 2.4897778224365672e-06, + "loss": 1.0908, + "step": 1052 + }, + { + "epoch": 0.08973157221985513, + "grad_norm": 24.5, + "learning_rate": 2.4897563653893754e-06, + "loss": 1.2304, + "step": 1053 + }, + { + "epoch": 0.08981678738815509, + "grad_norm": 19.5, + "learning_rate": 2.4897348859385395e-06, + "loss": 1.2056, + "step": 1054 + }, + { + "epoch": 0.08990200255645504, + "grad_norm": 14.6875, + "learning_rate": 2.4897133840844474e-06, + "loss": 0.6569, + "step": 1055 + }, + { + "epoch": 0.089987217724755, + "grad_norm": 13.875, + "learning_rate": 2.4896918598274884e-06, + "loss": 0.5882, + "step": 1056 + }, + { + "epoch": 0.09007243289305496, + "grad_norm": 16.75, + "learning_rate": 2.4896703131680503e-06, + "loss": 1.1163, + "step": 1057 + }, + { + "epoch": 0.09015764806135493, + "grad_norm": 15.375, + "learning_rate": 2.489648744106523e-06, + "loss": 0.7096, + "step": 1058 + }, + { + "epoch": 0.09024286322965488, + "grad_norm": 14.625, + "learning_rate": 2.4896271526432966e-06, + "loss": 0.7612, + "step": 1059 + }, + { + "epoch": 0.09032807839795484, + "grad_norm": 13.0625, + "learning_rate": 2.489605538778761e-06, + "loss": 0.6753, + "step": 1060 + }, + { + "epoch": 0.0904132935662548, + "grad_norm": 12.875, + "learning_rate": 2.4895839025133067e-06, + "loss": 0.5948, + "step": 1061 + }, + { + "epoch": 0.09049850873455476, + "grad_norm": 14.75, + "learning_rate": 2.489562243847325e-06, + "loss": 0.6922, + "step": 1062 + }, + { + "epoch": 0.09058372390285471, + "grad_norm": 18.375, + "learning_rate": 2.489540562781207e-06, + "loss": 0.8443, + "step": 1063 + }, + { + "epoch": 0.09066893907115467, + "grad_norm": 12.375, + "learning_rate": 2.4895188593153444e-06, + "loss": 0.6662, + "step": 1064 + }, + { + "epoch": 0.09075415423945463, + "grad_norm": 16.375, + "learning_rate": 2.4894971334501295e-06, + "loss": 1.2845, + "step": 1065 + }, + { + "epoch": 0.09083936940775458, + "grad_norm": 14.8125, + "learning_rate": 2.4894753851859555e-06, + "loss": 1.016, + "step": 1066 + }, + { + "epoch": 0.09092458457605454, + "grad_norm": 15.0, + "learning_rate": 2.489453614523214e-06, + "loss": 0.7773, + "step": 1067 + }, + { + "epoch": 0.0910097997443545, + "grad_norm": 13.8125, + "learning_rate": 2.4894318214623004e-06, + "loss": 0.7754, + "step": 1068 + }, + { + "epoch": 0.09109501491265445, + "grad_norm": 20.5, + "learning_rate": 2.4894100060036066e-06, + "loss": 1.1329, + "step": 1069 + }, + { + "epoch": 0.09118023008095441, + "grad_norm": 11.75, + "learning_rate": 2.489388168147528e-06, + "loss": 0.4827, + "step": 1070 + }, + { + "epoch": 0.09126544524925437, + "grad_norm": 18.0, + "learning_rate": 2.4893663078944593e-06, + "loss": 0.8682, + "step": 1071 + }, + { + "epoch": 0.09135066041755432, + "grad_norm": 19.75, + "learning_rate": 2.4893444252447948e-06, + "loss": 0.9551, + "step": 1072 + }, + { + "epoch": 0.09143587558585428, + "grad_norm": 16.75, + "learning_rate": 2.4893225201989295e-06, + "loss": 0.7951, + "step": 1073 + }, + { + "epoch": 0.09152109075415424, + "grad_norm": 14.8125, + "learning_rate": 2.489300592757261e-06, + "loss": 0.995, + "step": 1074 + }, + { + "epoch": 0.0916063059224542, + "grad_norm": 20.875, + "learning_rate": 2.4892786429201843e-06, + "loss": 1.0128, + "step": 1075 + }, + { + "epoch": 0.09169152109075415, + "grad_norm": 17.25, + "learning_rate": 2.4892566706880963e-06, + "loss": 0.8333, + "step": 1076 + }, + { + "epoch": 0.09177673625905411, + "grad_norm": 13.4375, + "learning_rate": 2.489234676061394e-06, + "loss": 0.9616, + "step": 1077 + }, + { + "epoch": 0.09186195142735407, + "grad_norm": 21.0, + "learning_rate": 2.489212659040475e-06, + "loss": 1.1042, + "step": 1078 + }, + { + "epoch": 0.09194716659565402, + "grad_norm": 14.875, + "learning_rate": 2.4891906196257377e-06, + "loss": 0.821, + "step": 1079 + }, + { + "epoch": 0.09203238176395398, + "grad_norm": 21.625, + "learning_rate": 2.489168557817579e-06, + "loss": 1.1489, + "step": 1080 + }, + { + "epoch": 0.09211759693225394, + "grad_norm": 11.6875, + "learning_rate": 2.489146473616399e-06, + "loss": 0.5219, + "step": 1081 + }, + { + "epoch": 0.0922028121005539, + "grad_norm": 16.0, + "learning_rate": 2.4891243670225955e-06, + "loss": 0.7303, + "step": 1082 + }, + { + "epoch": 0.09228802726885385, + "grad_norm": 19.125, + "learning_rate": 2.4891022380365693e-06, + "loss": 0.7661, + "step": 1083 + }, + { + "epoch": 0.09237324243715381, + "grad_norm": 17.625, + "learning_rate": 2.4890800866587193e-06, + "loss": 1.1133, + "step": 1084 + }, + { + "epoch": 0.09245845760545376, + "grad_norm": 17.875, + "learning_rate": 2.489057912889446e-06, + "loss": 0.7884, + "step": 1085 + }, + { + "epoch": 0.09254367277375372, + "grad_norm": 12.4375, + "learning_rate": 2.4890357167291507e-06, + "loss": 0.6475, + "step": 1086 + }, + { + "epoch": 0.09262888794205369, + "grad_norm": 18.375, + "learning_rate": 2.489013498178234e-06, + "loss": 0.9722, + "step": 1087 + }, + { + "epoch": 0.09271410311035365, + "grad_norm": 17.375, + "learning_rate": 2.488991257237097e-06, + "loss": 0.8523, + "step": 1088 + }, + { + "epoch": 0.0927993182786536, + "grad_norm": 18.875, + "learning_rate": 2.488968993906142e-06, + "loss": 0.7299, + "step": 1089 + }, + { + "epoch": 0.09288453344695356, + "grad_norm": 18.375, + "learning_rate": 2.488946708185772e-06, + "loss": 1.2717, + "step": 1090 + }, + { + "epoch": 0.09296974861525352, + "grad_norm": 17.375, + "learning_rate": 2.488924400076389e-06, + "loss": 0.8368, + "step": 1091 + }, + { + "epoch": 0.09305496378355348, + "grad_norm": 12.5625, + "learning_rate": 2.4889020695783962e-06, + "loss": 0.7039, + "step": 1092 + }, + { + "epoch": 0.09314017895185343, + "grad_norm": 18.625, + "learning_rate": 2.4888797166921967e-06, + "loss": 1.0184, + "step": 1093 + }, + { + "epoch": 0.09322539412015339, + "grad_norm": 15.125, + "learning_rate": 2.4888573414181956e-06, + "loss": 0.8016, + "step": 1094 + }, + { + "epoch": 0.09331060928845335, + "grad_norm": 13.75, + "learning_rate": 2.488834943756796e-06, + "loss": 0.608, + "step": 1095 + }, + { + "epoch": 0.0933958244567533, + "grad_norm": 19.875, + "learning_rate": 2.4888125237084037e-06, + "loss": 1.0198, + "step": 1096 + }, + { + "epoch": 0.09348103962505326, + "grad_norm": 15.875, + "learning_rate": 2.488790081273423e-06, + "loss": 1.0101, + "step": 1097 + }, + { + "epoch": 0.09356625479335322, + "grad_norm": 15.6875, + "learning_rate": 2.48876761645226e-06, + "loss": 0.9163, + "step": 1098 + }, + { + "epoch": 0.09365146996165317, + "grad_norm": 15.6875, + "learning_rate": 2.4887451292453208e-06, + "loss": 1.0795, + "step": 1099 + }, + { + "epoch": 0.09373668512995313, + "grad_norm": 19.125, + "learning_rate": 2.488722619653011e-06, + "loss": 0.7447, + "step": 1100 + }, + { + "epoch": 0.09382190029825309, + "grad_norm": 10.25, + "learning_rate": 2.4887000876757383e-06, + "loss": 0.4926, + "step": 1101 + }, + { + "epoch": 0.09390711546655305, + "grad_norm": 18.125, + "learning_rate": 2.488677533313909e-06, + "loss": 1.1527, + "step": 1102 + }, + { + "epoch": 0.093992330634853, + "grad_norm": 16.375, + "learning_rate": 2.4886549565679314e-06, + "loss": 1.2142, + "step": 1103 + }, + { + "epoch": 0.09407754580315296, + "grad_norm": 16.125, + "learning_rate": 2.488632357438213e-06, + "loss": 1.0449, + "step": 1104 + }, + { + "epoch": 0.09416276097145292, + "grad_norm": 26.625, + "learning_rate": 2.488609735925162e-06, + "loss": 1.2415, + "step": 1105 + }, + { + "epoch": 0.09424797613975287, + "grad_norm": 11.9375, + "learning_rate": 2.4885870920291883e-06, + "loss": 0.5278, + "step": 1106 + }, + { + "epoch": 0.09433319130805283, + "grad_norm": 13.75, + "learning_rate": 2.4885644257506997e-06, + "loss": 0.5945, + "step": 1107 + }, + { + "epoch": 0.09441840647635279, + "grad_norm": 18.25, + "learning_rate": 2.4885417370901066e-06, + "loss": 1.0857, + "step": 1108 + }, + { + "epoch": 0.09450362164465274, + "grad_norm": 18.125, + "learning_rate": 2.4885190260478194e-06, + "loss": 1.0826, + "step": 1109 + }, + { + "epoch": 0.0945888368129527, + "grad_norm": 12.0625, + "learning_rate": 2.4884962926242473e-06, + "loss": 0.5277, + "step": 1110 + }, + { + "epoch": 0.09467405198125266, + "grad_norm": 13.4375, + "learning_rate": 2.4884735368198022e-06, + "loss": 0.8315, + "step": 1111 + }, + { + "epoch": 0.09475926714955261, + "grad_norm": 12.8125, + "learning_rate": 2.4884507586348943e-06, + "loss": 0.7635, + "step": 1112 + }, + { + "epoch": 0.09484448231785257, + "grad_norm": 17.75, + "learning_rate": 2.4884279580699367e-06, + "loss": 0.8582, + "step": 1113 + }, + { + "epoch": 0.09492969748615253, + "grad_norm": 21.0, + "learning_rate": 2.48840513512534e-06, + "loss": 1.1656, + "step": 1114 + }, + { + "epoch": 0.0950149126544525, + "grad_norm": 13.1875, + "learning_rate": 2.4883822898015174e-06, + "loss": 0.7417, + "step": 1115 + }, + { + "epoch": 0.09510012782275246, + "grad_norm": 18.375, + "learning_rate": 2.4883594220988815e-06, + "loss": 1.1427, + "step": 1116 + }, + { + "epoch": 0.09518534299105241, + "grad_norm": 19.5, + "learning_rate": 2.488336532017845e-06, + "loss": 0.9757, + "step": 1117 + }, + { + "epoch": 0.09527055815935237, + "grad_norm": 21.75, + "learning_rate": 2.488313619558823e-06, + "loss": 1.0315, + "step": 1118 + }, + { + "epoch": 0.09535577332765233, + "grad_norm": 12.8125, + "learning_rate": 2.488290684722228e-06, + "loss": 0.6196, + "step": 1119 + }, + { + "epoch": 0.09544098849595228, + "grad_norm": 17.0, + "learning_rate": 2.488267727508476e-06, + "loss": 1.1974, + "step": 1120 + }, + { + "epoch": 0.09552620366425224, + "grad_norm": 12.875, + "learning_rate": 2.4882447479179804e-06, + "loss": 0.638, + "step": 1121 + }, + { + "epoch": 0.0956114188325522, + "grad_norm": 11.375, + "learning_rate": 2.488221745951157e-06, + "loss": 0.6605, + "step": 1122 + }, + { + "epoch": 0.09569663400085215, + "grad_norm": 14.0625, + "learning_rate": 2.4881987216084217e-06, + "loss": 0.5806, + "step": 1123 + }, + { + "epoch": 0.09578184916915211, + "grad_norm": 19.625, + "learning_rate": 2.4881756748901903e-06, + "loss": 1.0507, + "step": 1124 + }, + { + "epoch": 0.09586706433745207, + "grad_norm": 13.6875, + "learning_rate": 2.488152605796879e-06, + "loss": 1.028, + "step": 1125 + }, + { + "epoch": 0.09595227950575203, + "grad_norm": 14.1875, + "learning_rate": 2.488129514328906e-06, + "loss": 0.5205, + "step": 1126 + }, + { + "epoch": 0.09603749467405198, + "grad_norm": 16.0, + "learning_rate": 2.488106400486687e-06, + "loss": 0.8333, + "step": 1127 + }, + { + "epoch": 0.09612270984235194, + "grad_norm": 16.25, + "learning_rate": 2.4880832642706404e-06, + "loss": 0.7568, + "step": 1128 + }, + { + "epoch": 0.0962079250106519, + "grad_norm": 26.125, + "learning_rate": 2.4880601056811843e-06, + "loss": 1.2304, + "step": 1129 + }, + { + "epoch": 0.09629314017895185, + "grad_norm": 20.5, + "learning_rate": 2.488036924718737e-06, + "loss": 1.0342, + "step": 1130 + }, + { + "epoch": 0.09637835534725181, + "grad_norm": 18.375, + "learning_rate": 2.4880137213837176e-06, + "loss": 1.3621, + "step": 1131 + }, + { + "epoch": 0.09646357051555177, + "grad_norm": 27.75, + "learning_rate": 2.487990495676545e-06, + "loss": 1.5038, + "step": 1132 + }, + { + "epoch": 0.09654878568385172, + "grad_norm": 14.8125, + "learning_rate": 2.48796724759764e-06, + "loss": 0.9412, + "step": 1133 + }, + { + "epoch": 0.09663400085215168, + "grad_norm": 14.0625, + "learning_rate": 2.487943977147421e-06, + "loss": 0.467, + "step": 1134 + }, + { + "epoch": 0.09671921602045164, + "grad_norm": 14.1875, + "learning_rate": 2.4879206843263105e-06, + "loss": 0.7452, + "step": 1135 + }, + { + "epoch": 0.0968044311887516, + "grad_norm": 12.6875, + "learning_rate": 2.487897369134728e-06, + "loss": 0.6697, + "step": 1136 + }, + { + "epoch": 0.09688964635705155, + "grad_norm": 15.1875, + "learning_rate": 2.4878740315730947e-06, + "loss": 1.3572, + "step": 1137 + }, + { + "epoch": 0.09697486152535151, + "grad_norm": 15.0625, + "learning_rate": 2.487850671641834e-06, + "loss": 0.9798, + "step": 1138 + }, + { + "epoch": 0.09706007669365146, + "grad_norm": 14.0, + "learning_rate": 2.4878272893413656e-06, + "loss": 0.5652, + "step": 1139 + }, + { + "epoch": 0.09714529186195142, + "grad_norm": 14.625, + "learning_rate": 2.487803884672114e-06, + "loss": 0.912, + "step": 1140 + }, + { + "epoch": 0.09723050703025138, + "grad_norm": 12.75, + "learning_rate": 2.4877804576345017e-06, + "loss": 0.3384, + "step": 1141 + }, + { + "epoch": 0.09731572219855134, + "grad_norm": 19.5, + "learning_rate": 2.4877570082289518e-06, + "loss": 1.1014, + "step": 1142 + }, + { + "epoch": 0.0974009373668513, + "grad_norm": 13.1875, + "learning_rate": 2.487733536455888e-06, + "loss": 0.5668, + "step": 1143 + }, + { + "epoch": 0.09748615253515126, + "grad_norm": 11.625, + "learning_rate": 2.487710042315734e-06, + "loss": 0.6697, + "step": 1144 + }, + { + "epoch": 0.09757136770345122, + "grad_norm": 12.25, + "learning_rate": 2.4876865258089157e-06, + "loss": 0.7408, + "step": 1145 + }, + { + "epoch": 0.09765658287175118, + "grad_norm": 11.25, + "learning_rate": 2.487662986935857e-06, + "loss": 0.5894, + "step": 1146 + }, + { + "epoch": 0.09774179804005113, + "grad_norm": 12.3125, + "learning_rate": 2.487639425696984e-06, + "loss": 0.4808, + "step": 1147 + }, + { + "epoch": 0.09782701320835109, + "grad_norm": 15.6875, + "learning_rate": 2.487615842092721e-06, + "loss": 0.9066, + "step": 1148 + }, + { + "epoch": 0.09791222837665105, + "grad_norm": 14.4375, + "learning_rate": 2.487592236123496e-06, + "loss": 0.5614, + "step": 1149 + }, + { + "epoch": 0.097997443544951, + "grad_norm": 10.75, + "learning_rate": 2.487568607789735e-06, + "loss": 0.3832, + "step": 1150 + }, + { + "epoch": 0.09808265871325096, + "grad_norm": 11.125, + "learning_rate": 2.4875449570918644e-06, + "loss": 0.3155, + "step": 1151 + }, + { + "epoch": 0.09816787388155092, + "grad_norm": 18.25, + "learning_rate": 2.4875212840303115e-06, + "loss": 0.9819, + "step": 1152 + }, + { + "epoch": 0.09825308904985088, + "grad_norm": 24.125, + "learning_rate": 2.4874975886055054e-06, + "loss": 1.5273, + "step": 1153 + }, + { + "epoch": 0.09833830421815083, + "grad_norm": 15.75, + "learning_rate": 2.487473870817873e-06, + "loss": 0.7491, + "step": 1154 + }, + { + "epoch": 0.09842351938645079, + "grad_norm": 14.9375, + "learning_rate": 2.4874501306678436e-06, + "loss": 0.9202, + "step": 1155 + }, + { + "epoch": 0.09850873455475075, + "grad_norm": 17.5, + "learning_rate": 2.4874263681558456e-06, + "loss": 0.8345, + "step": 1156 + }, + { + "epoch": 0.0985939497230507, + "grad_norm": 14.0625, + "learning_rate": 2.4874025832823096e-06, + "loss": 0.8706, + "step": 1157 + }, + { + "epoch": 0.09867916489135066, + "grad_norm": 26.0, + "learning_rate": 2.487378776047664e-06, + "loss": 0.9641, + "step": 1158 + }, + { + "epoch": 0.09876438005965062, + "grad_norm": 14.625, + "learning_rate": 2.4873549464523392e-06, + "loss": 0.3667, + "step": 1159 + }, + { + "epoch": 0.09884959522795057, + "grad_norm": 13.8125, + "learning_rate": 2.4873310944967673e-06, + "loss": 0.8958, + "step": 1160 + }, + { + "epoch": 0.09893481039625053, + "grad_norm": 22.625, + "learning_rate": 2.487307220181378e-06, + "loss": 0.8326, + "step": 1161 + }, + { + "epoch": 0.09902002556455049, + "grad_norm": 17.125, + "learning_rate": 2.4872833235066024e-06, + "loss": 0.745, + "step": 1162 + }, + { + "epoch": 0.09910524073285044, + "grad_norm": 13.125, + "learning_rate": 2.487259404472874e-06, + "loss": 0.6665, + "step": 1163 + }, + { + "epoch": 0.0991904559011504, + "grad_norm": 12.4375, + "learning_rate": 2.487235463080623e-06, + "loss": 0.6353, + "step": 1164 + }, + { + "epoch": 0.09927567106945036, + "grad_norm": 16.0, + "learning_rate": 2.4872114993302835e-06, + "loss": 1.1939, + "step": 1165 + }, + { + "epoch": 0.09936088623775031, + "grad_norm": 15.3125, + "learning_rate": 2.4871875132222877e-06, + "loss": 1.0054, + "step": 1166 + }, + { + "epoch": 0.09944610140605027, + "grad_norm": 12.1875, + "learning_rate": 2.48716350475707e-06, + "loss": 0.5284, + "step": 1167 + }, + { + "epoch": 0.09953131657435023, + "grad_norm": 13.8125, + "learning_rate": 2.4871394739350636e-06, + "loss": 0.7509, + "step": 1168 + }, + { + "epoch": 0.09961653174265019, + "grad_norm": 15.375, + "learning_rate": 2.4871154207567027e-06, + "loss": 0.8657, + "step": 1169 + }, + { + "epoch": 0.09970174691095014, + "grad_norm": 14.1875, + "learning_rate": 2.487091345222422e-06, + "loss": 0.6932, + "step": 1170 + }, + { + "epoch": 0.09978696207925011, + "grad_norm": 30.75, + "learning_rate": 2.4870672473326567e-06, + "loss": 0.9452, + "step": 1171 + }, + { + "epoch": 0.09987217724755007, + "grad_norm": 12.625, + "learning_rate": 2.4870431270878427e-06, + "loss": 0.6746, + "step": 1172 + }, + { + "epoch": 0.09995739241585003, + "grad_norm": 13.6875, + "learning_rate": 2.487018984488415e-06, + "loss": 0.677, + "step": 1173 + }, + { + "epoch": 0.10004260758414998, + "grad_norm": 12.25, + "learning_rate": 2.486994819534811e-06, + "loss": 0.6327, + "step": 1174 + }, + { + "epoch": 0.10012782275244994, + "grad_norm": 15.375, + "learning_rate": 2.4869706322274663e-06, + "loss": 0.8116, + "step": 1175 + }, + { + "epoch": 0.1002130379207499, + "grad_norm": 16.0, + "learning_rate": 2.486946422566818e-06, + "loss": 0.6631, + "step": 1176 + }, + { + "epoch": 0.10029825308904985, + "grad_norm": 10.9375, + "learning_rate": 2.486922190553304e-06, + "loss": 0.5065, + "step": 1177 + }, + { + "epoch": 0.10038346825734981, + "grad_norm": 15.4375, + "learning_rate": 2.486897936187363e-06, + "loss": 1.0097, + "step": 1178 + }, + { + "epoch": 0.10046868342564977, + "grad_norm": 17.75, + "learning_rate": 2.486873659469432e-06, + "loss": 0.7886, + "step": 1179 + }, + { + "epoch": 0.10055389859394973, + "grad_norm": 14.625, + "learning_rate": 2.4868493603999503e-06, + "loss": 0.8535, + "step": 1180 + }, + { + "epoch": 0.10063911376224968, + "grad_norm": 16.25, + "learning_rate": 2.486825038979357e-06, + "loss": 0.7982, + "step": 1181 + }, + { + "epoch": 0.10072432893054964, + "grad_norm": 12.3125, + "learning_rate": 2.486800695208091e-06, + "loss": 0.5671, + "step": 1182 + }, + { + "epoch": 0.1008095440988496, + "grad_norm": 16.125, + "learning_rate": 2.486776329086593e-06, + "loss": 0.8319, + "step": 1183 + }, + { + "epoch": 0.10089475926714955, + "grad_norm": 20.625, + "learning_rate": 2.4867519406153036e-06, + "loss": 1.1765, + "step": 1184 + }, + { + "epoch": 0.10097997443544951, + "grad_norm": 22.125, + "learning_rate": 2.4867275297946627e-06, + "loss": 1.0212, + "step": 1185 + }, + { + "epoch": 0.10106518960374947, + "grad_norm": 14.0, + "learning_rate": 2.4867030966251112e-06, + "loss": 0.9917, + "step": 1186 + }, + { + "epoch": 0.10115040477204942, + "grad_norm": 18.25, + "learning_rate": 2.4866786411070914e-06, + "loss": 0.8056, + "step": 1187 + }, + { + "epoch": 0.10123561994034938, + "grad_norm": 14.1875, + "learning_rate": 2.486654163241045e-06, + "loss": 0.8442, + "step": 1188 + }, + { + "epoch": 0.10132083510864934, + "grad_norm": 15.8125, + "learning_rate": 2.4866296630274144e-06, + "loss": 1.195, + "step": 1189 + }, + { + "epoch": 0.1014060502769493, + "grad_norm": 16.125, + "learning_rate": 2.4866051404666424e-06, + "loss": 0.9384, + "step": 1190 + }, + { + "epoch": 0.10149126544524925, + "grad_norm": 16.25, + "learning_rate": 2.4865805955591723e-06, + "loss": 0.6707, + "step": 1191 + }, + { + "epoch": 0.10157648061354921, + "grad_norm": 13.6875, + "learning_rate": 2.4865560283054466e-06, + "loss": 0.805, + "step": 1192 + }, + { + "epoch": 0.10166169578184917, + "grad_norm": 14.8125, + "learning_rate": 2.48653143870591e-06, + "loss": 0.4752, + "step": 1193 + }, + { + "epoch": 0.10174691095014912, + "grad_norm": 19.25, + "learning_rate": 2.4865068267610077e-06, + "loss": 0.9507, + "step": 1194 + }, + { + "epoch": 0.10183212611844908, + "grad_norm": 22.625, + "learning_rate": 2.486482192471183e-06, + "loss": 0.9932, + "step": 1195 + }, + { + "epoch": 0.10191734128674904, + "grad_norm": 23.0, + "learning_rate": 2.486457535836882e-06, + "loss": 0.9625, + "step": 1196 + }, + { + "epoch": 0.10200255645504899, + "grad_norm": 21.125, + "learning_rate": 2.48643285685855e-06, + "loss": 1.0072, + "step": 1197 + }, + { + "epoch": 0.10208777162334895, + "grad_norm": 21.5, + "learning_rate": 2.4864081555366327e-06, + "loss": 1.1831, + "step": 1198 + }, + { + "epoch": 0.10217298679164892, + "grad_norm": 12.625, + "learning_rate": 2.4863834318715767e-06, + "loss": 0.7271, + "step": 1199 + }, + { + "epoch": 0.10225820195994888, + "grad_norm": 12.25, + "learning_rate": 2.4863586858638287e-06, + "loss": 0.5084, + "step": 1200 + }, + { + "epoch": 0.10234341712824883, + "grad_norm": 16.125, + "learning_rate": 2.4863339175138364e-06, + "loss": 0.9377, + "step": 1201 + }, + { + "epoch": 0.10242863229654879, + "grad_norm": 14.0, + "learning_rate": 2.486309126822047e-06, + "loss": 0.7694, + "step": 1202 + }, + { + "epoch": 0.10251384746484875, + "grad_norm": 10.3125, + "learning_rate": 2.4862843137889083e-06, + "loss": 0.4483, + "step": 1203 + }, + { + "epoch": 0.1025990626331487, + "grad_norm": 22.375, + "learning_rate": 2.4862594784148687e-06, + "loss": 1.1014, + "step": 1204 + }, + { + "epoch": 0.10268427780144866, + "grad_norm": 11.6875, + "learning_rate": 2.4862346207003775e-06, + "loss": 0.4566, + "step": 1205 + }, + { + "epoch": 0.10276949296974862, + "grad_norm": 15.9375, + "learning_rate": 2.486209740645883e-06, + "loss": 0.8798, + "step": 1206 + }, + { + "epoch": 0.10285470813804858, + "grad_norm": 10.5625, + "learning_rate": 2.486184838251836e-06, + "loss": 0.4533, + "step": 1207 + }, + { + "epoch": 0.10293992330634853, + "grad_norm": 12.375, + "learning_rate": 2.4861599135186855e-06, + "loss": 0.4927, + "step": 1208 + }, + { + "epoch": 0.10302513847464849, + "grad_norm": 13.8125, + "learning_rate": 2.4861349664468825e-06, + "loss": 0.7832, + "step": 1209 + }, + { + "epoch": 0.10311035364294845, + "grad_norm": 15.375, + "learning_rate": 2.4861099970368772e-06, + "loss": 0.8285, + "step": 1210 + }, + { + "epoch": 0.1031955688112484, + "grad_norm": 13.125, + "learning_rate": 2.486085005289122e-06, + "loss": 0.5292, + "step": 1211 + }, + { + "epoch": 0.10328078397954836, + "grad_norm": 12.5625, + "learning_rate": 2.4860599912040672e-06, + "loss": 0.7756, + "step": 1212 + }, + { + "epoch": 0.10336599914784832, + "grad_norm": 12.8125, + "learning_rate": 2.486034954782166e-06, + "loss": 0.8251, + "step": 1213 + }, + { + "epoch": 0.10345121431614827, + "grad_norm": 14.1875, + "learning_rate": 2.4860098960238697e-06, + "loss": 0.8704, + "step": 1214 + }, + { + "epoch": 0.10353642948444823, + "grad_norm": 14.1875, + "learning_rate": 2.4859848149296315e-06, + "loss": 0.94, + "step": 1215 + }, + { + "epoch": 0.10362164465274819, + "grad_norm": 13.6875, + "learning_rate": 2.4859597114999054e-06, + "loss": 0.6716, + "step": 1216 + }, + { + "epoch": 0.10370685982104814, + "grad_norm": 29.375, + "learning_rate": 2.4859345857351443e-06, + "loss": 1.5182, + "step": 1217 + }, + { + "epoch": 0.1037920749893481, + "grad_norm": 21.875, + "learning_rate": 2.4859094376358024e-06, + "loss": 1.1244, + "step": 1218 + }, + { + "epoch": 0.10387729015764806, + "grad_norm": 13.9375, + "learning_rate": 2.4858842672023337e-06, + "loss": 0.8572, + "step": 1219 + }, + { + "epoch": 0.10396250532594802, + "grad_norm": 13.8125, + "learning_rate": 2.485859074435194e-06, + "loss": 0.6713, + "step": 1220 + }, + { + "epoch": 0.10404772049424797, + "grad_norm": 21.375, + "learning_rate": 2.485833859334838e-06, + "loss": 0.9288, + "step": 1221 + }, + { + "epoch": 0.10413293566254793, + "grad_norm": 13.0, + "learning_rate": 2.4858086219017214e-06, + "loss": 0.5945, + "step": 1222 + }, + { + "epoch": 0.10421815083084789, + "grad_norm": 17.5, + "learning_rate": 2.4857833621363003e-06, + "loss": 0.7457, + "step": 1223 + }, + { + "epoch": 0.10430336599914784, + "grad_norm": 16.625, + "learning_rate": 2.4857580800390312e-06, + "loss": 0.7915, + "step": 1224 + }, + { + "epoch": 0.1043885811674478, + "grad_norm": 14.5625, + "learning_rate": 2.485732775610371e-06, + "loss": 0.9525, + "step": 1225 + }, + { + "epoch": 0.10447379633574776, + "grad_norm": 14.0, + "learning_rate": 2.4857074488507765e-06, + "loss": 0.4845, + "step": 1226 + }, + { + "epoch": 0.10455901150404771, + "grad_norm": 16.75, + "learning_rate": 2.485682099760706e-06, + "loss": 0.889, + "step": 1227 + }, + { + "epoch": 0.10464422667234768, + "grad_norm": 15.375, + "learning_rate": 2.4856567283406177e-06, + "loss": 0.7665, + "step": 1228 + }, + { + "epoch": 0.10472944184064764, + "grad_norm": 13.875, + "learning_rate": 2.4856313345909697e-06, + "loss": 0.5197, + "step": 1229 + }, + { + "epoch": 0.1048146570089476, + "grad_norm": 20.625, + "learning_rate": 2.4856059185122208e-06, + "loss": 1.014, + "step": 1230 + }, + { + "epoch": 0.10489987217724756, + "grad_norm": 10.875, + "learning_rate": 2.4855804801048304e-06, + "loss": 0.389, + "step": 1231 + }, + { + "epoch": 0.10498508734554751, + "grad_norm": 22.75, + "learning_rate": 2.4855550193692585e-06, + "loss": 1.3122, + "step": 1232 + }, + { + "epoch": 0.10507030251384747, + "grad_norm": 12.8125, + "learning_rate": 2.4855295363059646e-06, + "loss": 0.7156, + "step": 1233 + }, + { + "epoch": 0.10515551768214743, + "grad_norm": 13.6875, + "learning_rate": 2.48550403091541e-06, + "loss": 0.6026, + "step": 1234 + }, + { + "epoch": 0.10524073285044738, + "grad_norm": 14.1875, + "learning_rate": 2.4854785031980547e-06, + "loss": 0.7546, + "step": 1235 + }, + { + "epoch": 0.10532594801874734, + "grad_norm": 11.0625, + "learning_rate": 2.485452953154361e-06, + "loss": 0.4887, + "step": 1236 + }, + { + "epoch": 0.1054111631870473, + "grad_norm": 13.1875, + "learning_rate": 2.4854273807847893e-06, + "loss": 0.7915, + "step": 1237 + }, + { + "epoch": 0.10549637835534725, + "grad_norm": 16.5, + "learning_rate": 2.485401786089803e-06, + "loss": 1.1728, + "step": 1238 + }, + { + "epoch": 0.10558159352364721, + "grad_norm": 20.625, + "learning_rate": 2.4853761690698643e-06, + "loss": 1.2652, + "step": 1239 + }, + { + "epoch": 0.10566680869194717, + "grad_norm": 24.125, + "learning_rate": 2.485350529725436e-06, + "loss": 1.2301, + "step": 1240 + }, + { + "epoch": 0.10575202386024712, + "grad_norm": 17.375, + "learning_rate": 2.485324868056981e-06, + "loss": 0.9542, + "step": 1241 + }, + { + "epoch": 0.10583723902854708, + "grad_norm": 20.125, + "learning_rate": 2.4852991840649638e-06, + "loss": 0.576, + "step": 1242 + }, + { + "epoch": 0.10592245419684704, + "grad_norm": 17.5, + "learning_rate": 2.4852734777498478e-06, + "loss": 0.972, + "step": 1243 + }, + { + "epoch": 0.106007669365147, + "grad_norm": 12.3125, + "learning_rate": 2.485247749112098e-06, + "loss": 0.6378, + "step": 1244 + }, + { + "epoch": 0.10609288453344695, + "grad_norm": 13.875, + "learning_rate": 2.485221998152179e-06, + "loss": 0.3143, + "step": 1245 + }, + { + "epoch": 0.10617809970174691, + "grad_norm": 11.375, + "learning_rate": 2.485196224870557e-06, + "loss": 0.6087, + "step": 1246 + }, + { + "epoch": 0.10626331487004687, + "grad_norm": 12.875, + "learning_rate": 2.485170429267697e-06, + "loss": 0.6459, + "step": 1247 + }, + { + "epoch": 0.10634853003834682, + "grad_norm": 24.75, + "learning_rate": 2.4851446113440647e-06, + "loss": 0.9692, + "step": 1248 + }, + { + "epoch": 0.10643374520664678, + "grad_norm": 15.125, + "learning_rate": 2.485118771100128e-06, + "loss": 0.7668, + "step": 1249 + }, + { + "epoch": 0.10651896037494674, + "grad_norm": 16.0, + "learning_rate": 2.4850929085363525e-06, + "loss": 0.6654, + "step": 1250 + }, + { + "epoch": 0.1066041755432467, + "grad_norm": 12.1875, + "learning_rate": 2.4850670236532065e-06, + "loss": 0.6495, + "step": 1251 + }, + { + "epoch": 0.10668939071154665, + "grad_norm": 14.0625, + "learning_rate": 2.4850411164511573e-06, + "loss": 0.7962, + "step": 1252 + }, + { + "epoch": 0.1067746058798466, + "grad_norm": 13.4375, + "learning_rate": 2.4850151869306734e-06, + "loss": 0.7298, + "step": 1253 + }, + { + "epoch": 0.10685982104814656, + "grad_norm": 18.375, + "learning_rate": 2.4849892350922228e-06, + "loss": 0.8274, + "step": 1254 + }, + { + "epoch": 0.10694503621644652, + "grad_norm": 22.625, + "learning_rate": 2.4849632609362747e-06, + "loss": 1.147, + "step": 1255 + }, + { + "epoch": 0.10703025138474649, + "grad_norm": 17.125, + "learning_rate": 2.4849372644632993e-06, + "loss": 0.8296, + "step": 1256 + }, + { + "epoch": 0.10711546655304645, + "grad_norm": 14.875, + "learning_rate": 2.484911245673765e-06, + "loss": 0.73, + "step": 1257 + }, + { + "epoch": 0.1072006817213464, + "grad_norm": 12.125, + "learning_rate": 2.4848852045681434e-06, + "loss": 0.6973, + "step": 1258 + }, + { + "epoch": 0.10728589688964636, + "grad_norm": 18.625, + "learning_rate": 2.4848591411469035e-06, + "loss": 1.1448, + "step": 1259 + }, + { + "epoch": 0.10737111205794632, + "grad_norm": 14.0, + "learning_rate": 2.484833055410518e-06, + "loss": 0.7084, + "step": 1260 + }, + { + "epoch": 0.10745632722624628, + "grad_norm": 14.125, + "learning_rate": 2.4848069473594573e-06, + "loss": 0.8765, + "step": 1261 + }, + { + "epoch": 0.10754154239454623, + "grad_norm": 13.5625, + "learning_rate": 2.484780816994193e-06, + "loss": 0.5018, + "step": 1262 + }, + { + "epoch": 0.10762675756284619, + "grad_norm": 10.0, + "learning_rate": 2.484754664315198e-06, + "loss": 0.4434, + "step": 1263 + }, + { + "epoch": 0.10771197273114615, + "grad_norm": 13.875, + "learning_rate": 2.4847284893229445e-06, + "loss": 0.7946, + "step": 1264 + }, + { + "epoch": 0.1077971878994461, + "grad_norm": 16.75, + "learning_rate": 2.4847022920179054e-06, + "loss": 0.9332, + "step": 1265 + }, + { + "epoch": 0.10788240306774606, + "grad_norm": 20.75, + "learning_rate": 2.4846760724005544e-06, + "loss": 1.1173, + "step": 1266 + }, + { + "epoch": 0.10796761823604602, + "grad_norm": 11.9375, + "learning_rate": 2.484649830471365e-06, + "loss": 0.4956, + "step": 1267 + }, + { + "epoch": 0.10805283340434597, + "grad_norm": 12.75, + "learning_rate": 2.4846235662308123e-06, + "loss": 0.4356, + "step": 1268 + }, + { + "epoch": 0.10813804857264593, + "grad_norm": 17.0, + "learning_rate": 2.48459727967937e-06, + "loss": 1.0364, + "step": 1269 + }, + { + "epoch": 0.10822326374094589, + "grad_norm": 16.0, + "learning_rate": 2.484570970817513e-06, + "loss": 0.8467, + "step": 1270 + }, + { + "epoch": 0.10830847890924584, + "grad_norm": 13.375, + "learning_rate": 2.4845446396457173e-06, + "loss": 0.591, + "step": 1271 + }, + { + "epoch": 0.1083936940775458, + "grad_norm": 12.5625, + "learning_rate": 2.4845182861644584e-06, + "loss": 0.4932, + "step": 1272 + }, + { + "epoch": 0.10847890924584576, + "grad_norm": 14.9375, + "learning_rate": 2.4844919103742126e-06, + "loss": 0.8037, + "step": 1273 + }, + { + "epoch": 0.10856412441414572, + "grad_norm": 11.6875, + "learning_rate": 2.4844655122754567e-06, + "loss": 0.5426, + "step": 1274 + }, + { + "epoch": 0.10864933958244567, + "grad_norm": 13.25, + "learning_rate": 2.4844390918686678e-06, + "loss": 0.7429, + "step": 1275 + }, + { + "epoch": 0.10873455475074563, + "grad_norm": 33.25, + "learning_rate": 2.4844126491543234e-06, + "loss": 1.3516, + "step": 1276 + }, + { + "epoch": 0.10881976991904559, + "grad_norm": 15.375, + "learning_rate": 2.484386184132901e-06, + "loss": 0.9762, + "step": 1277 + }, + { + "epoch": 0.10890498508734554, + "grad_norm": 19.625, + "learning_rate": 2.4843596968048784e-06, + "loss": 1.1892, + "step": 1278 + }, + { + "epoch": 0.1089902002556455, + "grad_norm": 12.4375, + "learning_rate": 2.4843331871707355e-06, + "loss": 0.6167, + "step": 1279 + }, + { + "epoch": 0.10907541542394546, + "grad_norm": 13.5, + "learning_rate": 2.48430665523095e-06, + "loss": 0.6848, + "step": 1280 + }, + { + "epoch": 0.10916063059224541, + "grad_norm": 17.125, + "learning_rate": 2.484280100986003e-06, + "loss": 1.2219, + "step": 1281 + }, + { + "epoch": 0.10924584576054537, + "grad_norm": 12.0, + "learning_rate": 2.484253524436373e-06, + "loss": 0.5148, + "step": 1282 + }, + { + "epoch": 0.10933106092884533, + "grad_norm": 20.5, + "learning_rate": 2.48422692558254e-06, + "loss": 1.3084, + "step": 1283 + }, + { + "epoch": 0.1094162760971453, + "grad_norm": 14.5, + "learning_rate": 2.484200304424986e-06, + "loss": 0.6643, + "step": 1284 + }, + { + "epoch": 0.10950149126544526, + "grad_norm": 14.4375, + "learning_rate": 2.4841736609641916e-06, + "loss": 1.0793, + "step": 1285 + }, + { + "epoch": 0.10958670643374521, + "grad_norm": 15.75, + "learning_rate": 2.484146995200638e-06, + "loss": 1.0071, + "step": 1286 + }, + { + "epoch": 0.10967192160204517, + "grad_norm": 15.0625, + "learning_rate": 2.4841203071348067e-06, + "loss": 0.6509, + "step": 1287 + }, + { + "epoch": 0.10975713677034513, + "grad_norm": 15.3125, + "learning_rate": 2.4840935967671807e-06, + "loss": 0.6941, + "step": 1288 + }, + { + "epoch": 0.10984235193864508, + "grad_norm": 11.375, + "learning_rate": 2.484066864098243e-06, + "loss": 0.6386, + "step": 1289 + }, + { + "epoch": 0.10992756710694504, + "grad_norm": 10.875, + "learning_rate": 2.4840401091284754e-06, + "loss": 0.5552, + "step": 1290 + }, + { + "epoch": 0.110012782275245, + "grad_norm": 10.25, + "learning_rate": 2.4840133318583625e-06, + "loss": 0.541, + "step": 1291 + }, + { + "epoch": 0.11009799744354495, + "grad_norm": 17.25, + "learning_rate": 2.4839865322883873e-06, + "loss": 1.0938, + "step": 1292 + }, + { + "epoch": 0.11018321261184491, + "grad_norm": 16.875, + "learning_rate": 2.483959710419035e-06, + "loss": 0.8905, + "step": 1293 + }, + { + "epoch": 0.11026842778014487, + "grad_norm": 23.375, + "learning_rate": 2.4839328662507896e-06, + "loss": 0.9578, + "step": 1294 + }, + { + "epoch": 0.11035364294844482, + "grad_norm": 18.75, + "learning_rate": 2.483905999784137e-06, + "loss": 0.7415, + "step": 1295 + }, + { + "epoch": 0.11043885811674478, + "grad_norm": 12.4375, + "learning_rate": 2.4838791110195616e-06, + "loss": 0.6517, + "step": 1296 + }, + { + "epoch": 0.11052407328504474, + "grad_norm": 17.375, + "learning_rate": 2.48385219995755e-06, + "loss": 0.7366, + "step": 1297 + }, + { + "epoch": 0.1106092884533447, + "grad_norm": 17.875, + "learning_rate": 2.4838252665985887e-06, + "loss": 0.6285, + "step": 1298 + }, + { + "epoch": 0.11069450362164465, + "grad_norm": 14.5, + "learning_rate": 2.483798310943164e-06, + "loss": 0.6958, + "step": 1299 + }, + { + "epoch": 0.11077971878994461, + "grad_norm": 15.3125, + "learning_rate": 2.4837713329917634e-06, + "loss": 0.5696, + "step": 1300 + }, + { + "epoch": 0.11086493395824457, + "grad_norm": 22.875, + "learning_rate": 2.483744332744874e-06, + "loss": 1.282, + "step": 1301 + }, + { + "epoch": 0.11095014912654452, + "grad_norm": 20.25, + "learning_rate": 2.4837173102029833e-06, + "loss": 1.2556, + "step": 1302 + }, + { + "epoch": 0.11103536429484448, + "grad_norm": 15.5625, + "learning_rate": 2.483690265366581e-06, + "loss": 0.9616, + "step": 1303 + }, + { + "epoch": 0.11112057946314444, + "grad_norm": 13.8125, + "learning_rate": 2.4836631982361546e-06, + "loss": 0.8363, + "step": 1304 + }, + { + "epoch": 0.1112057946314444, + "grad_norm": 14.0, + "learning_rate": 2.483636108812194e-06, + "loss": 0.5904, + "step": 1305 + }, + { + "epoch": 0.11129100979974435, + "grad_norm": 20.25, + "learning_rate": 2.483608997095188e-06, + "loss": 1.1782, + "step": 1306 + }, + { + "epoch": 0.11137622496804431, + "grad_norm": 13.875, + "learning_rate": 2.483581863085627e-06, + "loss": 0.5778, + "step": 1307 + }, + { + "epoch": 0.11146144013634426, + "grad_norm": 18.5, + "learning_rate": 2.4835547067840015e-06, + "loss": 0.9187, + "step": 1308 + }, + { + "epoch": 0.11154665530464422, + "grad_norm": 22.375, + "learning_rate": 2.483527528190802e-06, + "loss": 1.2057, + "step": 1309 + }, + { + "epoch": 0.11163187047294418, + "grad_norm": 13.0, + "learning_rate": 2.4835003273065197e-06, + "loss": 0.8024, + "step": 1310 + }, + { + "epoch": 0.11171708564124413, + "grad_norm": 11.1875, + "learning_rate": 2.4834731041316457e-06, + "loss": 0.5612, + "step": 1311 + }, + { + "epoch": 0.1118023008095441, + "grad_norm": 19.5, + "learning_rate": 2.4834458586666725e-06, + "loss": 0.5, + "step": 1312 + }, + { + "epoch": 0.11188751597784406, + "grad_norm": 11.8125, + "learning_rate": 2.483418590912093e-06, + "loss": 0.5764, + "step": 1313 + }, + { + "epoch": 0.11197273114614402, + "grad_norm": 11.0, + "learning_rate": 2.4833913008683985e-06, + "loss": 0.5075, + "step": 1314 + }, + { + "epoch": 0.11205794631444398, + "grad_norm": 10.5, + "learning_rate": 2.483363988536083e-06, + "loss": 0.3628, + "step": 1315 + }, + { + "epoch": 0.11214316148274393, + "grad_norm": 19.375, + "learning_rate": 2.48333665391564e-06, + "loss": 0.8608, + "step": 1316 + }, + { + "epoch": 0.11222837665104389, + "grad_norm": 17.0, + "learning_rate": 2.4833092970075638e-06, + "loss": 1.0525, + "step": 1317 + }, + { + "epoch": 0.11231359181934385, + "grad_norm": 13.5, + "learning_rate": 2.4832819178123477e-06, + "loss": 0.9775, + "step": 1318 + }, + { + "epoch": 0.1123988069876438, + "grad_norm": 15.3125, + "learning_rate": 2.483254516330488e-06, + "loss": 0.6232, + "step": 1319 + }, + { + "epoch": 0.11248402215594376, + "grad_norm": 24.125, + "learning_rate": 2.4832270925624784e-06, + "loss": 0.9508, + "step": 1320 + }, + { + "epoch": 0.11256923732424372, + "grad_norm": 22.0, + "learning_rate": 2.4831996465088156e-06, + "loss": 0.8841, + "step": 1321 + }, + { + "epoch": 0.11265445249254367, + "grad_norm": 12.3125, + "learning_rate": 2.483172178169995e-06, + "loss": 0.6611, + "step": 1322 + }, + { + "epoch": 0.11273966766084363, + "grad_norm": 15.625, + "learning_rate": 2.4831446875465133e-06, + "loss": 1.064, + "step": 1323 + }, + { + "epoch": 0.11282488282914359, + "grad_norm": 20.75, + "learning_rate": 2.483117174638867e-06, + "loss": 0.9941, + "step": 1324 + }, + { + "epoch": 0.11291009799744355, + "grad_norm": 17.0, + "learning_rate": 2.483089639447553e-06, + "loss": 0.9082, + "step": 1325 + }, + { + "epoch": 0.1129953131657435, + "grad_norm": 20.75, + "learning_rate": 2.4830620819730693e-06, + "loss": 0.8576, + "step": 1326 + }, + { + "epoch": 0.11308052833404346, + "grad_norm": 18.375, + "learning_rate": 2.483034502215914e-06, + "loss": 1.2188, + "step": 1327 + }, + { + "epoch": 0.11316574350234342, + "grad_norm": 11.1875, + "learning_rate": 2.483006900176585e-06, + "loss": 0.4908, + "step": 1328 + }, + { + "epoch": 0.11325095867064337, + "grad_norm": 13.25, + "learning_rate": 2.4829792758555817e-06, + "loss": 0.8232, + "step": 1329 + }, + { + "epoch": 0.11333617383894333, + "grad_norm": 16.125, + "learning_rate": 2.482951629253403e-06, + "loss": 1.0177, + "step": 1330 + }, + { + "epoch": 0.11342138900724329, + "grad_norm": 13.9375, + "learning_rate": 2.4829239603705486e-06, + "loss": 0.7121, + "step": 1331 + }, + { + "epoch": 0.11350660417554324, + "grad_norm": 15.0625, + "learning_rate": 2.4828962692075187e-06, + "loss": 0.6933, + "step": 1332 + }, + { + "epoch": 0.1135918193438432, + "grad_norm": 16.25, + "learning_rate": 2.482868555764813e-06, + "loss": 0.9299, + "step": 1333 + }, + { + "epoch": 0.11367703451214316, + "grad_norm": 12.6875, + "learning_rate": 2.482840820042933e-06, + "loss": 0.5928, + "step": 1334 + }, + { + "epoch": 0.11376224968044311, + "grad_norm": 10.875, + "learning_rate": 2.4828130620423795e-06, + "loss": 0.4289, + "step": 1335 + }, + { + "epoch": 0.11384746484874307, + "grad_norm": 15.5, + "learning_rate": 2.4827852817636545e-06, + "loss": 0.7182, + "step": 1336 + }, + { + "epoch": 0.11393268001704303, + "grad_norm": 11.0, + "learning_rate": 2.4827574792072594e-06, + "loss": 0.4529, + "step": 1337 + }, + { + "epoch": 0.11401789518534299, + "grad_norm": 21.0, + "learning_rate": 2.482729654373697e-06, + "loss": 0.8668, + "step": 1338 + }, + { + "epoch": 0.11410311035364294, + "grad_norm": 17.625, + "learning_rate": 2.4827018072634707e-06, + "loss": 1.015, + "step": 1339 + }, + { + "epoch": 0.11418832552194291, + "grad_norm": 14.0625, + "learning_rate": 2.4826739378770833e-06, + "loss": 0.7146, + "step": 1340 + }, + { + "epoch": 0.11427354069024287, + "grad_norm": 10.9375, + "learning_rate": 2.4826460462150376e-06, + "loss": 0.3754, + "step": 1341 + }, + { + "epoch": 0.11435875585854283, + "grad_norm": 16.875, + "learning_rate": 2.4826181322778385e-06, + "loss": 0.8204, + "step": 1342 + }, + { + "epoch": 0.11444397102684278, + "grad_norm": 15.5625, + "learning_rate": 2.4825901960659904e-06, + "loss": 0.947, + "step": 1343 + }, + { + "epoch": 0.11452918619514274, + "grad_norm": 11.4375, + "learning_rate": 2.482562237579998e-06, + "loss": 0.415, + "step": 1344 + }, + { + "epoch": 0.1146144013634427, + "grad_norm": 18.75, + "learning_rate": 2.4825342568203664e-06, + "loss": 0.9712, + "step": 1345 + }, + { + "epoch": 0.11469961653174265, + "grad_norm": 15.75, + "learning_rate": 2.4825062537876014e-06, + "loss": 0.8323, + "step": 1346 + }, + { + "epoch": 0.11478483170004261, + "grad_norm": 15.0625, + "learning_rate": 2.4824782284822092e-06, + "loss": 0.9422, + "step": 1347 + }, + { + "epoch": 0.11487004686834257, + "grad_norm": 14.8125, + "learning_rate": 2.4824501809046956e-06, + "loss": 0.4773, + "step": 1348 + }, + { + "epoch": 0.11495526203664252, + "grad_norm": 13.3125, + "learning_rate": 2.4824221110555683e-06, + "loss": 0.7092, + "step": 1349 + }, + { + "epoch": 0.11504047720494248, + "grad_norm": 15.5625, + "learning_rate": 2.482394018935334e-06, + "loss": 0.8438, + "step": 1350 + }, + { + "epoch": 0.11512569237324244, + "grad_norm": 14.3125, + "learning_rate": 2.482365904544501e-06, + "loss": 0.7924, + "step": 1351 + }, + { + "epoch": 0.1152109075415424, + "grad_norm": 16.125, + "learning_rate": 2.4823377678835763e-06, + "loss": 0.9336, + "step": 1352 + }, + { + "epoch": 0.11529612270984235, + "grad_norm": 18.0, + "learning_rate": 2.4823096089530692e-06, + "loss": 0.926, + "step": 1353 + }, + { + "epoch": 0.11538133787814231, + "grad_norm": 12.125, + "learning_rate": 2.482281427753488e-06, + "loss": 0.6261, + "step": 1354 + }, + { + "epoch": 0.11546655304644227, + "grad_norm": 19.75, + "learning_rate": 2.482253224285342e-06, + "loss": 0.4516, + "step": 1355 + }, + { + "epoch": 0.11555176821474222, + "grad_norm": 10.9375, + "learning_rate": 2.4822249985491416e-06, + "loss": 0.5531, + "step": 1356 + }, + { + "epoch": 0.11563698338304218, + "grad_norm": 13.625, + "learning_rate": 2.4821967505453965e-06, + "loss": 0.7452, + "step": 1357 + }, + { + "epoch": 0.11572219855134214, + "grad_norm": 13.8125, + "learning_rate": 2.4821684802746167e-06, + "loss": 0.7824, + "step": 1358 + }, + { + "epoch": 0.1158074137196421, + "grad_norm": 15.6875, + "learning_rate": 2.4821401877373135e-06, + "loss": 0.6703, + "step": 1359 + }, + { + "epoch": 0.11589262888794205, + "grad_norm": 18.875, + "learning_rate": 2.4821118729339985e-06, + "loss": 0.873, + "step": 1360 + }, + { + "epoch": 0.11597784405624201, + "grad_norm": 27.0, + "learning_rate": 2.4820835358651823e-06, + "loss": 0.9078, + "step": 1361 + }, + { + "epoch": 0.11606305922454196, + "grad_norm": 13.5625, + "learning_rate": 2.4820551765313784e-06, + "loss": 0.6403, + "step": 1362 + }, + { + "epoch": 0.11614827439284192, + "grad_norm": 18.0, + "learning_rate": 2.482026794933098e-06, + "loss": 1.0153, + "step": 1363 + }, + { + "epoch": 0.11623348956114188, + "grad_norm": 17.5, + "learning_rate": 2.4819983910708546e-06, + "loss": 0.8461, + "step": 1364 + }, + { + "epoch": 0.11631870472944184, + "grad_norm": 14.625, + "learning_rate": 2.4819699649451613e-06, + "loss": 0.9227, + "step": 1365 + }, + { + "epoch": 0.11640391989774179, + "grad_norm": 15.5, + "learning_rate": 2.4819415165565317e-06, + "loss": 0.6959, + "step": 1366 + }, + { + "epoch": 0.11648913506604175, + "grad_norm": 13.25, + "learning_rate": 2.481913045905481e-06, + "loss": 0.6191, + "step": 1367 + }, + { + "epoch": 0.1165743502343417, + "grad_norm": 14.625, + "learning_rate": 2.4818845529925216e-06, + "loss": 0.3106, + "step": 1368 + }, + { + "epoch": 0.11665956540264168, + "grad_norm": 24.25, + "learning_rate": 2.4818560378181707e-06, + "loss": 1.1964, + "step": 1369 + }, + { + "epoch": 0.11674478057094163, + "grad_norm": 19.375, + "learning_rate": 2.481827500382942e-06, + "loss": 0.9775, + "step": 1370 + }, + { + "epoch": 0.11682999573924159, + "grad_norm": 12.875, + "learning_rate": 2.4817989406873512e-06, + "loss": 0.7623, + "step": 1371 + }, + { + "epoch": 0.11691521090754155, + "grad_norm": 28.375, + "learning_rate": 2.481770358731915e-06, + "loss": 0.5829, + "step": 1372 + }, + { + "epoch": 0.1170004260758415, + "grad_norm": 11.4375, + "learning_rate": 2.48174175451715e-06, + "loss": 0.5868, + "step": 1373 + }, + { + "epoch": 0.11708564124414146, + "grad_norm": 12.75, + "learning_rate": 2.481713128043573e-06, + "loss": 0.5655, + "step": 1374 + }, + { + "epoch": 0.11717085641244142, + "grad_norm": 16.75, + "learning_rate": 2.4816844793117014e-06, + "loss": 0.9376, + "step": 1375 + }, + { + "epoch": 0.11725607158074138, + "grad_norm": 14.25, + "learning_rate": 2.481655808322052e-06, + "loss": 0.5245, + "step": 1376 + }, + { + "epoch": 0.11734128674904133, + "grad_norm": 13.9375, + "learning_rate": 2.4816271150751444e-06, + "loss": 0.5217, + "step": 1377 + }, + { + "epoch": 0.11742650191734129, + "grad_norm": 14.75, + "learning_rate": 2.481598399571496e-06, + "loss": 0.8422, + "step": 1378 + }, + { + "epoch": 0.11751171708564125, + "grad_norm": 15.625, + "learning_rate": 2.4815696618116257e-06, + "loss": 1.1816, + "step": 1379 + }, + { + "epoch": 0.1175969322539412, + "grad_norm": 13.5625, + "learning_rate": 2.4815409017960538e-06, + "loss": 0.8437, + "step": 1380 + }, + { + "epoch": 0.11768214742224116, + "grad_norm": 13.3125, + "learning_rate": 2.481512119525299e-06, + "loss": 0.8476, + "step": 1381 + }, + { + "epoch": 0.11776736259054112, + "grad_norm": 13.5, + "learning_rate": 2.481483314999882e-06, + "loss": 0.8877, + "step": 1382 + }, + { + "epoch": 0.11785257775884107, + "grad_norm": 12.875, + "learning_rate": 2.481454488220323e-06, + "loss": 0.7549, + "step": 1383 + }, + { + "epoch": 0.11793779292714103, + "grad_norm": 19.375, + "learning_rate": 2.4814256391871435e-06, + "loss": 0.7129, + "step": 1384 + }, + { + "epoch": 0.11802300809544099, + "grad_norm": 11.1875, + "learning_rate": 2.481396767900864e-06, + "loss": 0.5135, + "step": 1385 + }, + { + "epoch": 0.11810822326374094, + "grad_norm": 14.625, + "learning_rate": 2.481367874362006e-06, + "loss": 0.7443, + "step": 1386 + }, + { + "epoch": 0.1181934384320409, + "grad_norm": 9.1875, + "learning_rate": 2.4813389585710933e-06, + "loss": 0.3536, + "step": 1387 + }, + { + "epoch": 0.11827865360034086, + "grad_norm": 14.75, + "learning_rate": 2.481310020528647e-06, + "loss": 0.6719, + "step": 1388 + }, + { + "epoch": 0.11836386876864081, + "grad_norm": 10.9375, + "learning_rate": 2.4812810602351907e-06, + "loss": 0.5734, + "step": 1389 + }, + { + "epoch": 0.11844908393694077, + "grad_norm": 19.375, + "learning_rate": 2.481252077691247e-06, + "loss": 1.0354, + "step": 1390 + }, + { + "epoch": 0.11853429910524073, + "grad_norm": 16.75, + "learning_rate": 2.4812230728973407e-06, + "loss": 0.8676, + "step": 1391 + }, + { + "epoch": 0.11861951427354069, + "grad_norm": 13.5, + "learning_rate": 2.481194045853995e-06, + "loss": 0.6957, + "step": 1392 + }, + { + "epoch": 0.11870472944184064, + "grad_norm": 16.0, + "learning_rate": 2.481164996561735e-06, + "loss": 0.9095, + "step": 1393 + }, + { + "epoch": 0.1187899446101406, + "grad_norm": 10.5625, + "learning_rate": 2.4811359250210852e-06, + "loss": 0.4006, + "step": 1394 + }, + { + "epoch": 0.11887515977844056, + "grad_norm": 12.3125, + "learning_rate": 2.4811068312325713e-06, + "loss": 0.6876, + "step": 1395 + }, + { + "epoch": 0.11896037494674051, + "grad_norm": 12.6875, + "learning_rate": 2.4810777151967196e-06, + "loss": 0.7951, + "step": 1396 + }, + { + "epoch": 0.11904559011504048, + "grad_norm": 15.4375, + "learning_rate": 2.4810485769140554e-06, + "loss": 0.6443, + "step": 1397 + }, + { + "epoch": 0.11913080528334044, + "grad_norm": 14.5625, + "learning_rate": 2.481019416385105e-06, + "loss": 0.9024, + "step": 1398 + }, + { + "epoch": 0.1192160204516404, + "grad_norm": 13.375, + "learning_rate": 2.480990233610396e-06, + "loss": 0.6823, + "step": 1399 + }, + { + "epoch": 0.11930123561994035, + "grad_norm": 24.125, + "learning_rate": 2.4809610285904558e-06, + "loss": 0.7423, + "step": 1400 + }, + { + "epoch": 0.11938645078824031, + "grad_norm": 19.0, + "learning_rate": 2.480931801325812e-06, + "loss": 0.9985, + "step": 1401 + }, + { + "epoch": 0.11947166595654027, + "grad_norm": 16.25, + "learning_rate": 2.480902551816993e-06, + "loss": 0.7579, + "step": 1402 + }, + { + "epoch": 0.11955688112484023, + "grad_norm": 16.75, + "learning_rate": 2.4808732800645264e-06, + "loss": 1.2039, + "step": 1403 + }, + { + "epoch": 0.11964209629314018, + "grad_norm": 15.4375, + "learning_rate": 2.4808439860689427e-06, + "loss": 0.6401, + "step": 1404 + }, + { + "epoch": 0.11972731146144014, + "grad_norm": 25.125, + "learning_rate": 2.48081466983077e-06, + "loss": 0.9703, + "step": 1405 + }, + { + "epoch": 0.1198125266297401, + "grad_norm": 15.5, + "learning_rate": 2.4807853313505385e-06, + "loss": 0.8871, + "step": 1406 + }, + { + "epoch": 0.11989774179804005, + "grad_norm": 25.875, + "learning_rate": 2.4807559706287786e-06, + "loss": 0.7812, + "step": 1407 + }, + { + "epoch": 0.11998295696634001, + "grad_norm": 13.8125, + "learning_rate": 2.4807265876660205e-06, + "loss": 0.8993, + "step": 1408 + }, + { + "epoch": 0.12006817213463997, + "grad_norm": 22.375, + "learning_rate": 2.4806971824627955e-06, + "loss": 1.2011, + "step": 1409 + }, + { + "epoch": 0.12015338730293992, + "grad_norm": 14.375, + "learning_rate": 2.480667755019635e-06, + "loss": 0.6522, + "step": 1410 + }, + { + "epoch": 0.12023860247123988, + "grad_norm": 16.375, + "learning_rate": 2.4806383053370704e-06, + "loss": 0.703, + "step": 1411 + }, + { + "epoch": 0.12032381763953984, + "grad_norm": 16.25, + "learning_rate": 2.480608833415634e-06, + "loss": 1.1941, + "step": 1412 + }, + { + "epoch": 0.1204090328078398, + "grad_norm": 16.375, + "learning_rate": 2.4805793392558587e-06, + "loss": 0.6089, + "step": 1413 + }, + { + "epoch": 0.12049424797613975, + "grad_norm": 14.25, + "learning_rate": 2.4805498228582776e-06, + "loss": 0.8614, + "step": 1414 + }, + { + "epoch": 0.12057946314443971, + "grad_norm": 13.5, + "learning_rate": 2.4805202842234233e-06, + "loss": 0.7704, + "step": 1415 + }, + { + "epoch": 0.12066467831273966, + "grad_norm": 12.875, + "learning_rate": 2.4804907233518305e-06, + "loss": 0.827, + "step": 1416 + }, + { + "epoch": 0.12074989348103962, + "grad_norm": 12.5, + "learning_rate": 2.4804611402440326e-06, + "loss": 0.6887, + "step": 1417 + }, + { + "epoch": 0.12083510864933958, + "grad_norm": 15.5, + "learning_rate": 2.480431534900565e-06, + "loss": 0.703, + "step": 1418 + }, + { + "epoch": 0.12092032381763954, + "grad_norm": 15.5625, + "learning_rate": 2.4804019073219617e-06, + "loss": 0.7831, + "step": 1419 + }, + { + "epoch": 0.12100553898593949, + "grad_norm": 16.375, + "learning_rate": 2.480372257508759e-06, + "loss": 0.946, + "step": 1420 + }, + { + "epoch": 0.12109075415423945, + "grad_norm": 15.4375, + "learning_rate": 2.480342585461492e-06, + "loss": 0.7256, + "step": 1421 + }, + { + "epoch": 0.1211759693225394, + "grad_norm": 31.75, + "learning_rate": 2.480312891180698e-06, + "loss": 1.0698, + "step": 1422 + }, + { + "epoch": 0.12126118449083936, + "grad_norm": 17.625, + "learning_rate": 2.480283174666912e-06, + "loss": 1.0257, + "step": 1423 + }, + { + "epoch": 0.12134639965913932, + "grad_norm": 12.375, + "learning_rate": 2.4802534359206727e-06, + "loss": 0.5837, + "step": 1424 + }, + { + "epoch": 0.12143161482743929, + "grad_norm": 19.625, + "learning_rate": 2.480223674942516e-06, + "loss": 1.1336, + "step": 1425 + }, + { + "epoch": 0.12151682999573925, + "grad_norm": 14.875, + "learning_rate": 2.4801938917329808e-06, + "loss": 0.6785, + "step": 1426 + }, + { + "epoch": 0.1216020451640392, + "grad_norm": 17.375, + "learning_rate": 2.4801640862926054e-06, + "loss": 1.2337, + "step": 1427 + }, + { + "epoch": 0.12168726033233916, + "grad_norm": 14.75, + "learning_rate": 2.4801342586219273e-06, + "loss": 0.8515, + "step": 1428 + }, + { + "epoch": 0.12177247550063912, + "grad_norm": 12.5625, + "learning_rate": 2.4801044087214867e-06, + "loss": 0.6902, + "step": 1429 + }, + { + "epoch": 0.12185769066893908, + "grad_norm": 17.0, + "learning_rate": 2.4800745365918216e-06, + "loss": 0.6936, + "step": 1430 + }, + { + "epoch": 0.12194290583723903, + "grad_norm": 20.5, + "learning_rate": 2.480044642233473e-06, + "loss": 1.3433, + "step": 1431 + }, + { + "epoch": 0.12202812100553899, + "grad_norm": 14.3125, + "learning_rate": 2.480014725646981e-06, + "loss": 0.793, + "step": 1432 + }, + { + "epoch": 0.12211333617383895, + "grad_norm": 11.25, + "learning_rate": 2.4799847868328863e-06, + "loss": 0.622, + "step": 1433 + }, + { + "epoch": 0.1221985513421389, + "grad_norm": 12.5, + "learning_rate": 2.4799548257917292e-06, + "loss": 0.741, + "step": 1434 + }, + { + "epoch": 0.12228376651043886, + "grad_norm": 14.25, + "learning_rate": 2.479924842524052e-06, + "loss": 0.8168, + "step": 1435 + }, + { + "epoch": 0.12236898167873882, + "grad_norm": 15.1875, + "learning_rate": 2.4798948370303953e-06, + "loss": 0.8108, + "step": 1436 + }, + { + "epoch": 0.12245419684703877, + "grad_norm": 19.25, + "learning_rate": 2.4798648093113027e-06, + "loss": 0.8577, + "step": 1437 + }, + { + "epoch": 0.12253941201533873, + "grad_norm": 11.125, + "learning_rate": 2.479834759367316e-06, + "loss": 0.4022, + "step": 1438 + }, + { + "epoch": 0.12262462718363869, + "grad_norm": 13.8125, + "learning_rate": 2.479804687198979e-06, + "loss": 1.0363, + "step": 1439 + }, + { + "epoch": 0.12270984235193864, + "grad_norm": 15.0, + "learning_rate": 2.4797745928068346e-06, + "loss": 0.9002, + "step": 1440 + }, + { + "epoch": 0.1227950575202386, + "grad_norm": 13.6875, + "learning_rate": 2.479744476191426e-06, + "loss": 0.661, + "step": 1441 + }, + { + "epoch": 0.12288027268853856, + "grad_norm": 18.0, + "learning_rate": 2.479714337353299e-06, + "loss": 0.9553, + "step": 1442 + }, + { + "epoch": 0.12296548785683852, + "grad_norm": 15.1875, + "learning_rate": 2.4796841762929967e-06, + "loss": 0.9295, + "step": 1443 + }, + { + "epoch": 0.12305070302513847, + "grad_norm": 20.0, + "learning_rate": 2.4796539930110652e-06, + "loss": 0.8901, + "step": 1444 + }, + { + "epoch": 0.12313591819343843, + "grad_norm": 13.9375, + "learning_rate": 2.479623787508049e-06, + "loss": 0.6673, + "step": 1445 + }, + { + "epoch": 0.12322113336173839, + "grad_norm": 17.25, + "learning_rate": 2.479593559784495e-06, + "loss": 1.2743, + "step": 1446 + }, + { + "epoch": 0.12330634853003834, + "grad_norm": 14.875, + "learning_rate": 2.479563309840949e-06, + "loss": 0.8796, + "step": 1447 + }, + { + "epoch": 0.1233915636983383, + "grad_norm": 13.9375, + "learning_rate": 2.479533037677957e-06, + "loss": 0.6922, + "step": 1448 + }, + { + "epoch": 0.12347677886663826, + "grad_norm": 11.125, + "learning_rate": 2.4795027432960673e-06, + "loss": 0.5181, + "step": 1449 + }, + { + "epoch": 0.12356199403493821, + "grad_norm": 12.3125, + "learning_rate": 2.479472426695826e-06, + "loss": 0.5184, + "step": 1450 + }, + { + "epoch": 0.12364720920323817, + "grad_norm": 17.5, + "learning_rate": 2.4794420878777818e-06, + "loss": 1.2067, + "step": 1451 + }, + { + "epoch": 0.12373242437153813, + "grad_norm": 17.625, + "learning_rate": 2.4794117268424834e-06, + "loss": 0.9037, + "step": 1452 + }, + { + "epoch": 0.1238176395398381, + "grad_norm": 18.625, + "learning_rate": 2.4793813435904782e-06, + "loss": 1.0361, + "step": 1453 + }, + { + "epoch": 0.12390285470813805, + "grad_norm": 13.75, + "learning_rate": 2.479350938122316e-06, + "loss": 0.8246, + "step": 1454 + }, + { + "epoch": 0.12398806987643801, + "grad_norm": 15.1875, + "learning_rate": 2.479320510438546e-06, + "loss": 0.7256, + "step": 1455 + }, + { + "epoch": 0.12407328504473797, + "grad_norm": 24.5, + "learning_rate": 2.4792900605397185e-06, + "loss": 1.6031, + "step": 1456 + }, + { + "epoch": 0.12415850021303793, + "grad_norm": 17.75, + "learning_rate": 2.4792595884263833e-06, + "loss": 1.1003, + "step": 1457 + }, + { + "epoch": 0.12424371538133788, + "grad_norm": 16.875, + "learning_rate": 2.479229094099091e-06, + "loss": 0.7001, + "step": 1458 + }, + { + "epoch": 0.12432893054963784, + "grad_norm": 20.0, + "learning_rate": 2.479198577558393e-06, + "loss": 0.9321, + "step": 1459 + }, + { + "epoch": 0.1244141457179378, + "grad_norm": 16.125, + "learning_rate": 2.479168038804841e-06, + "loss": 0.9677, + "step": 1460 + }, + { + "epoch": 0.12449936088623775, + "grad_norm": 13.875, + "learning_rate": 2.4791374778389866e-06, + "loss": 0.7319, + "step": 1461 + }, + { + "epoch": 0.12458457605453771, + "grad_norm": 14.875, + "learning_rate": 2.4791068946613815e-06, + "loss": 0.8257, + "step": 1462 + }, + { + "epoch": 0.12466979122283767, + "grad_norm": 13.375, + "learning_rate": 2.479076289272579e-06, + "loss": 0.7955, + "step": 1463 + }, + { + "epoch": 0.12475500639113762, + "grad_norm": 15.1875, + "learning_rate": 2.479045661673132e-06, + "loss": 0.4965, + "step": 1464 + }, + { + "epoch": 0.12484022155943758, + "grad_norm": 16.5, + "learning_rate": 2.479015011863594e-06, + "loss": 0.6017, + "step": 1465 + }, + { + "epoch": 0.12492543672773754, + "grad_norm": 15.875, + "learning_rate": 2.478984339844519e-06, + "loss": 0.7101, + "step": 1466 + }, + { + "epoch": 0.1250106518960375, + "grad_norm": 13.625, + "learning_rate": 2.478953645616461e-06, + "loss": 0.9422, + "step": 1467 + }, + { + "epoch": 0.12509586706433745, + "grad_norm": 14.3125, + "learning_rate": 2.4789229291799748e-06, + "loss": 0.8229, + "step": 1468 + }, + { + "epoch": 0.1251810822326374, + "grad_norm": 10.1875, + "learning_rate": 2.4788921905356157e-06, + "loss": 0.3218, + "step": 1469 + }, + { + "epoch": 0.12526629740093737, + "grad_norm": 20.125, + "learning_rate": 2.4788614296839387e-06, + "loss": 0.7974, + "step": 1470 + }, + { + "epoch": 0.12535151256923732, + "grad_norm": 13.375, + "learning_rate": 2.4788306466255e-06, + "loss": 0.5381, + "step": 1471 + }, + { + "epoch": 0.12543672773753728, + "grad_norm": 13.0, + "learning_rate": 2.4787998413608556e-06, + "loss": 0.6639, + "step": 1472 + }, + { + "epoch": 0.12552194290583724, + "grad_norm": 23.75, + "learning_rate": 2.478769013890563e-06, + "loss": 1.1722, + "step": 1473 + }, + { + "epoch": 0.1256071580741372, + "grad_norm": 14.0, + "learning_rate": 2.478738164215178e-06, + "loss": 0.7418, + "step": 1474 + }, + { + "epoch": 0.12569237324243715, + "grad_norm": 15.3125, + "learning_rate": 2.4787072923352597e-06, + "loss": 1.033, + "step": 1475 + }, + { + "epoch": 0.1257775884107371, + "grad_norm": 14.25, + "learning_rate": 2.478676398251364e-06, + "loss": 0.6169, + "step": 1476 + }, + { + "epoch": 0.12586280357903706, + "grad_norm": 16.125, + "learning_rate": 2.478645481964051e-06, + "loss": 1.163, + "step": 1477 + }, + { + "epoch": 0.12594801874733702, + "grad_norm": 25.375, + "learning_rate": 2.4786145434738783e-06, + "loss": 1.371, + "step": 1478 + }, + { + "epoch": 0.12603323391563698, + "grad_norm": 15.125, + "learning_rate": 2.4785835827814055e-06, + "loss": 0.672, + "step": 1479 + }, + { + "epoch": 0.12611844908393693, + "grad_norm": 14.0, + "learning_rate": 2.478552599887192e-06, + "loss": 0.8676, + "step": 1480 + }, + { + "epoch": 0.1262036642522369, + "grad_norm": 22.25, + "learning_rate": 2.4785215947917973e-06, + "loss": 1.2902, + "step": 1481 + }, + { + "epoch": 0.12628887942053685, + "grad_norm": 16.125, + "learning_rate": 2.478490567495782e-06, + "loss": 0.8189, + "step": 1482 + }, + { + "epoch": 0.1263740945888368, + "grad_norm": 18.25, + "learning_rate": 2.4784595179997073e-06, + "loss": 0.9453, + "step": 1483 + }, + { + "epoch": 0.12645930975713676, + "grad_norm": 11.0625, + "learning_rate": 2.4784284463041334e-06, + "loss": 0.6488, + "step": 1484 + }, + { + "epoch": 0.12654452492543672, + "grad_norm": 12.0625, + "learning_rate": 2.4783973524096226e-06, + "loss": 0.5557, + "step": 1485 + }, + { + "epoch": 0.12662974009373668, + "grad_norm": 14.5, + "learning_rate": 2.478366236316736e-06, + "loss": 0.7248, + "step": 1486 + }, + { + "epoch": 0.12671495526203663, + "grad_norm": 20.75, + "learning_rate": 2.4783350980260366e-06, + "loss": 0.633, + "step": 1487 + }, + { + "epoch": 0.1268001704303366, + "grad_norm": 13.375, + "learning_rate": 2.4783039375380864e-06, + "loss": 0.5183, + "step": 1488 + }, + { + "epoch": 0.12688538559863655, + "grad_norm": 14.5625, + "learning_rate": 2.478272754853449e-06, + "loss": 0.8545, + "step": 1489 + }, + { + "epoch": 0.1269706007669365, + "grad_norm": 14.5625, + "learning_rate": 2.4782415499726883e-06, + "loss": 0.6475, + "step": 1490 + }, + { + "epoch": 0.12705581593523646, + "grad_norm": 9.6875, + "learning_rate": 2.478210322896367e-06, + "loss": 0.3222, + "step": 1491 + }, + { + "epoch": 0.12714103110353642, + "grad_norm": 23.0, + "learning_rate": 2.4781790736250507e-06, + "loss": 1.0186, + "step": 1492 + }, + { + "epoch": 0.12722624627183637, + "grad_norm": 15.75, + "learning_rate": 2.478147802159303e-06, + "loss": 0.5955, + "step": 1493 + }, + { + "epoch": 0.12731146144013633, + "grad_norm": 12.375, + "learning_rate": 2.4781165084996895e-06, + "loss": 0.6227, + "step": 1494 + }, + { + "epoch": 0.1273966766084363, + "grad_norm": 27.125, + "learning_rate": 2.478085192646776e-06, + "loss": 1.0827, + "step": 1495 + }, + { + "epoch": 0.12748189177673627, + "grad_norm": 12.875, + "learning_rate": 2.478053854601128e-06, + "loss": 0.492, + "step": 1496 + }, + { + "epoch": 0.12756710694503623, + "grad_norm": 14.625, + "learning_rate": 2.478022494363312e-06, + "loss": 0.5216, + "step": 1497 + }, + { + "epoch": 0.1276523221133362, + "grad_norm": 15.375, + "learning_rate": 2.4779911119338948e-06, + "loss": 0.7451, + "step": 1498 + }, + { + "epoch": 0.12773753728163614, + "grad_norm": 12.125, + "learning_rate": 2.477959707313443e-06, + "loss": 0.6604, + "step": 1499 + }, + { + "epoch": 0.1278227524499361, + "grad_norm": 14.875, + "learning_rate": 2.477928280502525e-06, + "loss": 0.8966, + "step": 1500 + }, + { + "epoch": 0.12790796761823606, + "grad_norm": 19.5, + "learning_rate": 2.477896831501708e-06, + "loss": 1.037, + "step": 1501 + }, + { + "epoch": 0.12799318278653601, + "grad_norm": 12.875, + "learning_rate": 2.47786536031156e-06, + "loss": 0.5206, + "step": 1502 + }, + { + "epoch": 0.12807839795483597, + "grad_norm": 17.0, + "learning_rate": 2.4778338669326507e-06, + "loss": 0.9273, + "step": 1503 + }, + { + "epoch": 0.12816361312313593, + "grad_norm": 14.4375, + "learning_rate": 2.477802351365549e-06, + "loss": 0.7273, + "step": 1504 + }, + { + "epoch": 0.12824882829143588, + "grad_norm": 15.125, + "learning_rate": 2.4777708136108235e-06, + "loss": 1.0315, + "step": 1505 + }, + { + "epoch": 0.12833404345973584, + "grad_norm": 16.125, + "learning_rate": 2.4777392536690455e-06, + "loss": 0.8453, + "step": 1506 + }, + { + "epoch": 0.1284192586280358, + "grad_norm": 15.25, + "learning_rate": 2.477707671540784e-06, + "loss": 1.0709, + "step": 1507 + }, + { + "epoch": 0.12850447379633576, + "grad_norm": 14.125, + "learning_rate": 2.4776760672266102e-06, + "loss": 0.4945, + "step": 1508 + }, + { + "epoch": 0.1285896889646357, + "grad_norm": 12.3125, + "learning_rate": 2.477644440727096e-06, + "loss": 0.7331, + "step": 1509 + }, + { + "epoch": 0.12867490413293567, + "grad_norm": 13.125, + "learning_rate": 2.477612792042812e-06, + "loss": 0.772, + "step": 1510 + }, + { + "epoch": 0.12876011930123563, + "grad_norm": 13.9375, + "learning_rate": 2.47758112117433e-06, + "loss": 1.0932, + "step": 1511 + }, + { + "epoch": 0.12884533446953558, + "grad_norm": 15.0625, + "learning_rate": 2.4775494281222236e-06, + "loss": 0.5275, + "step": 1512 + }, + { + "epoch": 0.12893054963783554, + "grad_norm": 16.75, + "learning_rate": 2.4775177128870636e-06, + "loss": 1.2726, + "step": 1513 + }, + { + "epoch": 0.1290157648061355, + "grad_norm": 14.5, + "learning_rate": 2.477485975469425e-06, + "loss": 1.0652, + "step": 1514 + }, + { + "epoch": 0.12910097997443545, + "grad_norm": 14.8125, + "learning_rate": 2.47745421586988e-06, + "loss": 0.765, + "step": 1515 + }, + { + "epoch": 0.1291861951427354, + "grad_norm": 16.125, + "learning_rate": 2.477422434089003e-06, + "loss": 0.6144, + "step": 1516 + }, + { + "epoch": 0.12927141031103537, + "grad_norm": 12.125, + "learning_rate": 2.4773906301273686e-06, + "loss": 0.3885, + "step": 1517 + }, + { + "epoch": 0.12935662547933532, + "grad_norm": 14.375, + "learning_rate": 2.477358803985551e-06, + "loss": 0.2456, + "step": 1518 + }, + { + "epoch": 0.12944184064763528, + "grad_norm": 11.4375, + "learning_rate": 2.477326955664126e-06, + "loss": 0.4207, + "step": 1519 + }, + { + "epoch": 0.12952705581593524, + "grad_norm": 12.6875, + "learning_rate": 2.4772950851636686e-06, + "loss": 0.7109, + "step": 1520 + }, + { + "epoch": 0.1296122709842352, + "grad_norm": 13.9375, + "learning_rate": 2.4772631924847547e-06, + "loss": 0.747, + "step": 1521 + }, + { + "epoch": 0.12969748615253515, + "grad_norm": 18.5, + "learning_rate": 2.477231277627961e-06, + "loss": 0.8909, + "step": 1522 + }, + { + "epoch": 0.1297827013208351, + "grad_norm": 13.75, + "learning_rate": 2.4771993405938636e-06, + "loss": 0.7773, + "step": 1523 + }, + { + "epoch": 0.12986791648913507, + "grad_norm": 18.25, + "learning_rate": 2.4771673813830405e-06, + "loss": 0.8698, + "step": 1524 + }, + { + "epoch": 0.12995313165743502, + "grad_norm": 18.0, + "learning_rate": 2.477135399996069e-06, + "loss": 0.9722, + "step": 1525 + }, + { + "epoch": 0.13003834682573498, + "grad_norm": 15.0625, + "learning_rate": 2.4771033964335263e-06, + "loss": 0.7828, + "step": 1526 + }, + { + "epoch": 0.13012356199403494, + "grad_norm": 18.875, + "learning_rate": 2.4770713706959914e-06, + "loss": 0.7726, + "step": 1527 + }, + { + "epoch": 0.1302087771623349, + "grad_norm": 17.625, + "learning_rate": 2.4770393227840432e-06, + "loss": 0.7019, + "step": 1528 + }, + { + "epoch": 0.13029399233063485, + "grad_norm": 16.375, + "learning_rate": 2.4770072526982607e-06, + "loss": 0.9926, + "step": 1529 + }, + { + "epoch": 0.1303792074989348, + "grad_norm": 17.75, + "learning_rate": 2.476975160439223e-06, + "loss": 0.8013, + "step": 1530 + }, + { + "epoch": 0.13046442266723476, + "grad_norm": 23.125, + "learning_rate": 2.47694304600751e-06, + "loss": 1.0229, + "step": 1531 + }, + { + "epoch": 0.13054963783553472, + "grad_norm": 15.125, + "learning_rate": 2.4769109094037027e-06, + "loss": 0.69, + "step": 1532 + }, + { + "epoch": 0.13063485300383468, + "grad_norm": 13.625, + "learning_rate": 2.4768787506283813e-06, + "loss": 0.808, + "step": 1533 + }, + { + "epoch": 0.13072006817213463, + "grad_norm": 13.25, + "learning_rate": 2.4768465696821275e-06, + "loss": 0.7263, + "step": 1534 + }, + { + "epoch": 0.1308052833404346, + "grad_norm": 15.4375, + "learning_rate": 2.4768143665655224e-06, + "loss": 0.5559, + "step": 1535 + }, + { + "epoch": 0.13089049850873455, + "grad_norm": 11.3125, + "learning_rate": 2.4767821412791482e-06, + "loss": 0.5647, + "step": 1536 + }, + { + "epoch": 0.1309757136770345, + "grad_norm": 15.75, + "learning_rate": 2.4767498938235866e-06, + "loss": 1.0662, + "step": 1537 + }, + { + "epoch": 0.13106092884533446, + "grad_norm": 19.875, + "learning_rate": 2.476717624199421e-06, + "loss": 0.9566, + "step": 1538 + }, + { + "epoch": 0.13114614401363442, + "grad_norm": 11.125, + "learning_rate": 2.476685332407235e-06, + "loss": 0.4021, + "step": 1539 + }, + { + "epoch": 0.13123135918193438, + "grad_norm": 17.625, + "learning_rate": 2.4766530184476107e-06, + "loss": 1.0331, + "step": 1540 + }, + { + "epoch": 0.13131657435023433, + "grad_norm": 14.0625, + "learning_rate": 2.4766206823211335e-06, + "loss": 1.2049, + "step": 1541 + }, + { + "epoch": 0.1314017895185343, + "grad_norm": 13.4375, + "learning_rate": 2.4765883240283868e-06, + "loss": 0.5505, + "step": 1542 + }, + { + "epoch": 0.13148700468683425, + "grad_norm": 12.75, + "learning_rate": 2.4765559435699554e-06, + "loss": 0.5995, + "step": 1543 + }, + { + "epoch": 0.1315722198551342, + "grad_norm": 13.3125, + "learning_rate": 2.476523540946425e-06, + "loss": 0.8021, + "step": 1544 + }, + { + "epoch": 0.13165743502343416, + "grad_norm": 12.125, + "learning_rate": 2.4764911161583806e-06, + "loss": 0.5415, + "step": 1545 + }, + { + "epoch": 0.13174265019173412, + "grad_norm": 17.375, + "learning_rate": 2.4764586692064086e-06, + "loss": 0.9332, + "step": 1546 + }, + { + "epoch": 0.13182786536003407, + "grad_norm": 15.875, + "learning_rate": 2.4764262000910953e-06, + "loss": 0.5249, + "step": 1547 + }, + { + "epoch": 0.13191308052833403, + "grad_norm": 12.0, + "learning_rate": 2.476393708813027e-06, + "loss": 0.448, + "step": 1548 + }, + { + "epoch": 0.131998295696634, + "grad_norm": 13.75, + "learning_rate": 2.476361195372791e-06, + "loss": 0.4997, + "step": 1549 + }, + { + "epoch": 0.13208351086493395, + "grad_norm": 20.125, + "learning_rate": 2.4763286597709755e-06, + "loss": 1.3269, + "step": 1550 + }, + { + "epoch": 0.1321687260332339, + "grad_norm": 13.25, + "learning_rate": 2.476296102008168e-06, + "loss": 0.6287, + "step": 1551 + }, + { + "epoch": 0.1322539412015339, + "grad_norm": 11.125, + "learning_rate": 2.476263522084956e-06, + "loss": 0.4037, + "step": 1552 + }, + { + "epoch": 0.13233915636983384, + "grad_norm": 18.375, + "learning_rate": 2.4762309200019296e-06, + "loss": 0.5243, + "step": 1553 + }, + { + "epoch": 0.1324243715381338, + "grad_norm": 19.5, + "learning_rate": 2.4761982957596776e-06, + "loss": 0.793, + "step": 1554 + }, + { + "epoch": 0.13250958670643376, + "grad_norm": 12.75, + "learning_rate": 2.476165649358789e-06, + "loss": 0.655, + "step": 1555 + }, + { + "epoch": 0.13259480187473371, + "grad_norm": 11.375, + "learning_rate": 2.4761329807998546e-06, + "loss": 0.6106, + "step": 1556 + }, + { + "epoch": 0.13268001704303367, + "grad_norm": 22.5, + "learning_rate": 2.4761002900834635e-06, + "loss": 0.905, + "step": 1557 + }, + { + "epoch": 0.13276523221133363, + "grad_norm": 22.125, + "learning_rate": 2.4760675772102077e-06, + "loss": 0.7622, + "step": 1558 + }, + { + "epoch": 0.13285044737963358, + "grad_norm": 11.6875, + "learning_rate": 2.476034842180678e-06, + "loss": 0.3594, + "step": 1559 + }, + { + "epoch": 0.13293566254793354, + "grad_norm": 12.375, + "learning_rate": 2.4760020849954653e-06, + "loss": 0.7484, + "step": 1560 + }, + { + "epoch": 0.1330208777162335, + "grad_norm": 13.0, + "learning_rate": 2.4759693056551626e-06, + "loss": 0.5805, + "step": 1561 + }, + { + "epoch": 0.13310609288453346, + "grad_norm": 12.375, + "learning_rate": 2.4759365041603617e-06, + "loss": 0.5127, + "step": 1562 + }, + { + "epoch": 0.1331913080528334, + "grad_norm": 15.6875, + "learning_rate": 2.475903680511655e-06, + "loss": 0.9681, + "step": 1563 + }, + { + "epoch": 0.13327652322113337, + "grad_norm": 20.125, + "learning_rate": 2.4758708347096363e-06, + "loss": 1.1556, + "step": 1564 + }, + { + "epoch": 0.13336173838943333, + "grad_norm": 14.625, + "learning_rate": 2.475837966754899e-06, + "loss": 0.7612, + "step": 1565 + }, + { + "epoch": 0.13344695355773328, + "grad_norm": 17.75, + "learning_rate": 2.475805076648037e-06, + "loss": 1.1628, + "step": 1566 + }, + { + "epoch": 0.13353216872603324, + "grad_norm": 17.0, + "learning_rate": 2.4757721643896444e-06, + "loss": 0.9244, + "step": 1567 + }, + { + "epoch": 0.1336173838943332, + "grad_norm": 13.8125, + "learning_rate": 2.4757392299803163e-06, + "loss": 0.6452, + "step": 1568 + }, + { + "epoch": 0.13370259906263315, + "grad_norm": 12.5625, + "learning_rate": 2.4757062734206476e-06, + "loss": 0.61, + "step": 1569 + }, + { + "epoch": 0.1337878142309331, + "grad_norm": 16.75, + "learning_rate": 2.4756732947112343e-06, + "loss": 1.1258, + "step": 1570 + }, + { + "epoch": 0.13387302939923307, + "grad_norm": 18.125, + "learning_rate": 2.475640293852672e-06, + "loss": 0.9646, + "step": 1571 + }, + { + "epoch": 0.13395824456753302, + "grad_norm": 13.9375, + "learning_rate": 2.4756072708455565e-06, + "loss": 0.7332, + "step": 1572 + }, + { + "epoch": 0.13404345973583298, + "grad_norm": 19.125, + "learning_rate": 2.4755742256904856e-06, + "loss": 1.1055, + "step": 1573 + }, + { + "epoch": 0.13412867490413294, + "grad_norm": 13.9375, + "learning_rate": 2.475541158388056e-06, + "loss": 0.9313, + "step": 1574 + }, + { + "epoch": 0.1342138900724329, + "grad_norm": 15.0, + "learning_rate": 2.4755080689388656e-06, + "loss": 1.0465, + "step": 1575 + }, + { + "epoch": 0.13429910524073285, + "grad_norm": 14.4375, + "learning_rate": 2.475474957343512e-06, + "loss": 0.8469, + "step": 1576 + }, + { + "epoch": 0.1343843204090328, + "grad_norm": 14.875, + "learning_rate": 2.4754418236025935e-06, + "loss": 0.6181, + "step": 1577 + }, + { + "epoch": 0.13446953557733277, + "grad_norm": 15.4375, + "learning_rate": 2.4754086677167087e-06, + "loss": 0.7345, + "step": 1578 + }, + { + "epoch": 0.13455475074563272, + "grad_norm": 17.75, + "learning_rate": 2.4753754896864575e-06, + "loss": 0.9474, + "step": 1579 + }, + { + "epoch": 0.13463996591393268, + "grad_norm": 14.125, + "learning_rate": 2.4753422895124387e-06, + "loss": 0.7746, + "step": 1580 + }, + { + "epoch": 0.13472518108223264, + "grad_norm": 13.4375, + "learning_rate": 2.4753090671952523e-06, + "loss": 0.6436, + "step": 1581 + }, + { + "epoch": 0.1348103962505326, + "grad_norm": 10.6875, + "learning_rate": 2.475275822735499e-06, + "loss": 0.4949, + "step": 1582 + }, + { + "epoch": 0.13489561141883255, + "grad_norm": 23.75, + "learning_rate": 2.47524255613378e-06, + "loss": 0.9854, + "step": 1583 + }, + { + "epoch": 0.1349808265871325, + "grad_norm": 17.25, + "learning_rate": 2.4752092673906955e-06, + "loss": 1.0981, + "step": 1584 + }, + { + "epoch": 0.13506604175543246, + "grad_norm": 13.3125, + "learning_rate": 2.4751759565068476e-06, + "loss": 0.5711, + "step": 1585 + }, + { + "epoch": 0.13515125692373242, + "grad_norm": 11.9375, + "learning_rate": 2.475142623482838e-06, + "loss": 0.7141, + "step": 1586 + }, + { + "epoch": 0.13523647209203238, + "grad_norm": 15.125, + "learning_rate": 2.4751092683192696e-06, + "loss": 0.6999, + "step": 1587 + }, + { + "epoch": 0.13532168726033234, + "grad_norm": 11.3125, + "learning_rate": 2.4750758910167443e-06, + "loss": 0.4597, + "step": 1588 + }, + { + "epoch": 0.1354069024286323, + "grad_norm": 34.25, + "learning_rate": 2.475042491575866e-06, + "loss": 1.183, + "step": 1589 + }, + { + "epoch": 0.13549211759693225, + "grad_norm": 12.0, + "learning_rate": 2.4750090699972383e-06, + "loss": 0.673, + "step": 1590 + }, + { + "epoch": 0.1355773327652322, + "grad_norm": 12.6875, + "learning_rate": 2.4749756262814643e-06, + "loss": 0.5467, + "step": 1591 + }, + { + "epoch": 0.13566254793353216, + "grad_norm": 13.8125, + "learning_rate": 2.474942160429149e-06, + "loss": 0.5855, + "step": 1592 + }, + { + "epoch": 0.13574776310183212, + "grad_norm": 15.0, + "learning_rate": 2.4749086724408975e-06, + "loss": 0.7791, + "step": 1593 + }, + { + "epoch": 0.13583297827013208, + "grad_norm": 15.5, + "learning_rate": 2.4748751623173143e-06, + "loss": 0.8921, + "step": 1594 + }, + { + "epoch": 0.13591819343843203, + "grad_norm": 15.25, + "learning_rate": 2.474841630059005e-06, + "loss": 0.661, + "step": 1595 + }, + { + "epoch": 0.136003408606732, + "grad_norm": 13.875, + "learning_rate": 2.474808075666576e-06, + "loss": 0.8196, + "step": 1596 + }, + { + "epoch": 0.13608862377503195, + "grad_norm": 14.8125, + "learning_rate": 2.4747744991406337e-06, + "loss": 0.9434, + "step": 1597 + }, + { + "epoch": 0.1361738389433319, + "grad_norm": 13.0, + "learning_rate": 2.4747409004817842e-06, + "loss": 0.6281, + "step": 1598 + }, + { + "epoch": 0.13625905411163186, + "grad_norm": 14.25, + "learning_rate": 2.474707279690635e-06, + "loss": 0.7573, + "step": 1599 + }, + { + "epoch": 0.13634426927993182, + "grad_norm": 23.0, + "learning_rate": 2.474673636767794e-06, + "loss": 1.0137, + "step": 1600 + }, + { + "epoch": 0.13642948444823177, + "grad_norm": 14.0625, + "learning_rate": 2.474639971713869e-06, + "loss": 0.6528, + "step": 1601 + }, + { + "epoch": 0.13651469961653173, + "grad_norm": 15.625, + "learning_rate": 2.474606284529468e-06, + "loss": 1.0082, + "step": 1602 + }, + { + "epoch": 0.1365999147848317, + "grad_norm": 13.0, + "learning_rate": 2.4745725752152e-06, + "loss": 0.6146, + "step": 1603 + }, + { + "epoch": 0.13668512995313165, + "grad_norm": 14.6875, + "learning_rate": 2.474538843771674e-06, + "loss": 1.1543, + "step": 1604 + }, + { + "epoch": 0.1367703451214316, + "grad_norm": 18.75, + "learning_rate": 2.4745050901994998e-06, + "loss": 0.6986, + "step": 1605 + }, + { + "epoch": 0.13685556028973156, + "grad_norm": 13.0, + "learning_rate": 2.4744713144992875e-06, + "loss": 0.6663, + "step": 1606 + }, + { + "epoch": 0.13694077545803152, + "grad_norm": 14.375, + "learning_rate": 2.4744375166716473e-06, + "loss": 0.9152, + "step": 1607 + }, + { + "epoch": 0.1370259906263315, + "grad_norm": 16.125, + "learning_rate": 2.47440369671719e-06, + "loss": 0.894, + "step": 1608 + }, + { + "epoch": 0.13711120579463146, + "grad_norm": 13.5, + "learning_rate": 2.474369854636526e-06, + "loss": 0.9786, + "step": 1609 + }, + { + "epoch": 0.13719642096293141, + "grad_norm": 16.375, + "learning_rate": 2.474335990430268e-06, + "loss": 0.4418, + "step": 1610 + }, + { + "epoch": 0.13728163613123137, + "grad_norm": 10.875, + "learning_rate": 2.4743021040990272e-06, + "loss": 0.5231, + "step": 1611 + }, + { + "epoch": 0.13736685129953133, + "grad_norm": 11.625, + "learning_rate": 2.4742681956434165e-06, + "loss": 0.4635, + "step": 1612 + }, + { + "epoch": 0.13745206646783129, + "grad_norm": 14.6875, + "learning_rate": 2.4742342650640485e-06, + "loss": 0.649, + "step": 1613 + }, + { + "epoch": 0.13753728163613124, + "grad_norm": 19.5, + "learning_rate": 2.474200312361536e-06, + "loss": 0.8201, + "step": 1614 + }, + { + "epoch": 0.1376224968044312, + "grad_norm": 15.625, + "learning_rate": 2.4741663375364934e-06, + "loss": 0.7977, + "step": 1615 + }, + { + "epoch": 0.13770771197273116, + "grad_norm": 16.0, + "learning_rate": 2.4741323405895334e-06, + "loss": 0.9941, + "step": 1616 + }, + { + "epoch": 0.1377929271410311, + "grad_norm": 21.5, + "learning_rate": 2.474098321521271e-06, + "loss": 0.8637, + "step": 1617 + }, + { + "epoch": 0.13787814230933107, + "grad_norm": 9.8125, + "learning_rate": 2.4740642803323217e-06, + "loss": 0.3601, + "step": 1618 + }, + { + "epoch": 0.13796335747763103, + "grad_norm": 12.875, + "learning_rate": 2.4740302170232995e-06, + "loss": 0.6797, + "step": 1619 + }, + { + "epoch": 0.13804857264593098, + "grad_norm": 14.125, + "learning_rate": 2.4739961315948204e-06, + "loss": 0.8079, + "step": 1620 + }, + { + "epoch": 0.13813378781423094, + "grad_norm": 18.125, + "learning_rate": 2.4739620240475006e-06, + "loss": 0.9553, + "step": 1621 + }, + { + "epoch": 0.1382190029825309, + "grad_norm": 20.875, + "learning_rate": 2.4739278943819563e-06, + "loss": 1.2068, + "step": 1622 + }, + { + "epoch": 0.13830421815083085, + "grad_norm": 16.0, + "learning_rate": 2.4738937425988037e-06, + "loss": 0.7125, + "step": 1623 + }, + { + "epoch": 0.1383894333191308, + "grad_norm": 12.0625, + "learning_rate": 2.4738595686986603e-06, + "loss": 0.7978, + "step": 1624 + }, + { + "epoch": 0.13847464848743077, + "grad_norm": 20.875, + "learning_rate": 2.473825372682145e-06, + "loss": 1.0319, + "step": 1625 + }, + { + "epoch": 0.13855986365573073, + "grad_norm": 13.875, + "learning_rate": 2.4737911545498733e-06, + "loss": 0.7193, + "step": 1626 + }, + { + "epoch": 0.13864507882403068, + "grad_norm": 12.5, + "learning_rate": 2.4737569143024654e-06, + "loss": 0.5906, + "step": 1627 + }, + { + "epoch": 0.13873029399233064, + "grad_norm": 19.0, + "learning_rate": 2.4737226519405393e-06, + "loss": 0.8338, + "step": 1628 + }, + { + "epoch": 0.1388155091606306, + "grad_norm": 18.5, + "learning_rate": 2.473688367464715e-06, + "loss": 1.0743, + "step": 1629 + }, + { + "epoch": 0.13890072432893055, + "grad_norm": 20.375, + "learning_rate": 2.4736540608756105e-06, + "loss": 0.6711, + "step": 1630 + }, + { + "epoch": 0.1389859394972305, + "grad_norm": 27.125, + "learning_rate": 2.473619732173847e-06, + "loss": 1.1368, + "step": 1631 + }, + { + "epoch": 0.13907115466553047, + "grad_norm": 18.125, + "learning_rate": 2.473585381360044e-06, + "loss": 1.0869, + "step": 1632 + }, + { + "epoch": 0.13915636983383042, + "grad_norm": 17.5, + "learning_rate": 2.4735510084348236e-06, + "loss": 0.8473, + "step": 1633 + }, + { + "epoch": 0.13924158500213038, + "grad_norm": 12.4375, + "learning_rate": 2.4735166133988057e-06, + "loss": 0.5834, + "step": 1634 + }, + { + "epoch": 0.13932680017043034, + "grad_norm": 17.125, + "learning_rate": 2.4734821962526122e-06, + "loss": 0.8453, + "step": 1635 + }, + { + "epoch": 0.1394120153387303, + "grad_norm": 22.875, + "learning_rate": 2.473447756996865e-06, + "loss": 0.7648, + "step": 1636 + }, + { + "epoch": 0.13949723050703025, + "grad_norm": 11.6875, + "learning_rate": 2.4734132956321872e-06, + "loss": 0.4743, + "step": 1637 + }, + { + "epoch": 0.1395824456753302, + "grad_norm": 15.5625, + "learning_rate": 2.4733788121592004e-06, + "loss": 1.1727, + "step": 1638 + }, + { + "epoch": 0.13966766084363016, + "grad_norm": 16.0, + "learning_rate": 2.4733443065785283e-06, + "loss": 1.2217, + "step": 1639 + }, + { + "epoch": 0.13975287601193012, + "grad_norm": 16.25, + "learning_rate": 2.4733097788907947e-06, + "loss": 0.8836, + "step": 1640 + }, + { + "epoch": 0.13983809118023008, + "grad_norm": 20.125, + "learning_rate": 2.4732752290966227e-06, + "loss": 1.039, + "step": 1641 + }, + { + "epoch": 0.13992330634853004, + "grad_norm": 20.125, + "learning_rate": 2.4732406571966373e-06, + "loss": 0.864, + "step": 1642 + }, + { + "epoch": 0.14000852151683, + "grad_norm": 18.25, + "learning_rate": 2.4732060631914633e-06, + "loss": 0.8466, + "step": 1643 + }, + { + "epoch": 0.14009373668512995, + "grad_norm": 17.5, + "learning_rate": 2.4731714470817257e-06, + "loss": 0.9993, + "step": 1644 + }, + { + "epoch": 0.1401789518534299, + "grad_norm": 19.625, + "learning_rate": 2.47313680886805e-06, + "loss": 0.9561, + "step": 1645 + }, + { + "epoch": 0.14026416702172986, + "grad_norm": 18.25, + "learning_rate": 2.4731021485510625e-06, + "loss": 0.8679, + "step": 1646 + }, + { + "epoch": 0.14034938219002982, + "grad_norm": 11.375, + "learning_rate": 2.4730674661313892e-06, + "loss": 0.5669, + "step": 1647 + }, + { + "epoch": 0.14043459735832978, + "grad_norm": 16.5, + "learning_rate": 2.4730327616096566e-06, + "loss": 1.0393, + "step": 1648 + }, + { + "epoch": 0.14051981252662973, + "grad_norm": 19.5, + "learning_rate": 2.472998034986493e-06, + "loss": 0.9216, + "step": 1649 + }, + { + "epoch": 0.1406050276949297, + "grad_norm": 16.5, + "learning_rate": 2.472963286262524e-06, + "loss": 0.9496, + "step": 1650 + }, + { + "epoch": 0.14069024286322965, + "grad_norm": 15.5, + "learning_rate": 2.4729285154383792e-06, + "loss": 0.8578, + "step": 1651 + }, + { + "epoch": 0.1407754580315296, + "grad_norm": 11.875, + "learning_rate": 2.472893722514686e-06, + "loss": 0.4865, + "step": 1652 + }, + { + "epoch": 0.14086067319982956, + "grad_norm": 16.375, + "learning_rate": 2.472858907492074e-06, + "loss": 0.9434, + "step": 1653 + }, + { + "epoch": 0.14094588836812952, + "grad_norm": 11.75, + "learning_rate": 2.4728240703711712e-06, + "loss": 0.5191, + "step": 1654 + }, + { + "epoch": 0.14103110353642948, + "grad_norm": 15.125, + "learning_rate": 2.4727892111526084e-06, + "loss": 0.8929, + "step": 1655 + }, + { + "epoch": 0.14111631870472943, + "grad_norm": 11.0625, + "learning_rate": 2.4727543298370148e-06, + "loss": 0.4518, + "step": 1656 + }, + { + "epoch": 0.1412015338730294, + "grad_norm": 12.4375, + "learning_rate": 2.472719426425021e-06, + "loss": 0.6647, + "step": 1657 + }, + { + "epoch": 0.14128674904132935, + "grad_norm": 17.0, + "learning_rate": 2.4726845009172572e-06, + "loss": 0.847, + "step": 1658 + }, + { + "epoch": 0.1413719642096293, + "grad_norm": 15.4375, + "learning_rate": 2.472649553314355e-06, + "loss": 1.0141, + "step": 1659 + }, + { + "epoch": 0.14145717937792926, + "grad_norm": 19.25, + "learning_rate": 2.472614583616946e-06, + "loss": 0.6686, + "step": 1660 + }, + { + "epoch": 0.14154239454622922, + "grad_norm": 12.5625, + "learning_rate": 2.4725795918256628e-06, + "loss": 0.7507, + "step": 1661 + }, + { + "epoch": 0.14162760971452917, + "grad_norm": 17.0, + "learning_rate": 2.472544577941136e-06, + "loss": 0.6057, + "step": 1662 + }, + { + "epoch": 0.14171282488282913, + "grad_norm": 12.4375, + "learning_rate": 2.472509541964e-06, + "loss": 0.5135, + "step": 1663 + }, + { + "epoch": 0.1417980400511291, + "grad_norm": 16.25, + "learning_rate": 2.472474483894887e-06, + "loss": 0.7836, + "step": 1664 + }, + { + "epoch": 0.14188325521942907, + "grad_norm": 12.375, + "learning_rate": 2.4724394037344302e-06, + "loss": 0.4222, + "step": 1665 + }, + { + "epoch": 0.14196847038772903, + "grad_norm": 11.125, + "learning_rate": 2.472404301483265e-06, + "loss": 0.4369, + "step": 1666 + }, + { + "epoch": 0.14205368555602899, + "grad_norm": 14.3125, + "learning_rate": 2.4723691771420243e-06, + "loss": 0.8266, + "step": 1667 + }, + { + "epoch": 0.14213890072432894, + "grad_norm": 18.5, + "learning_rate": 2.472334030711344e-06, + "loss": 0.9728, + "step": 1668 + }, + { + "epoch": 0.1422241158926289, + "grad_norm": 13.8125, + "learning_rate": 2.4722988621918575e-06, + "loss": 0.6679, + "step": 1669 + }, + { + "epoch": 0.14230933106092886, + "grad_norm": 42.25, + "learning_rate": 2.472263671584202e-06, + "loss": 1.1168, + "step": 1670 + }, + { + "epoch": 0.1423945462292288, + "grad_norm": 14.5, + "learning_rate": 2.4722284588890135e-06, + "loss": 0.603, + "step": 1671 + }, + { + "epoch": 0.14247976139752877, + "grad_norm": 14.0, + "learning_rate": 2.4721932241069267e-06, + "loss": 0.8431, + "step": 1672 + }, + { + "epoch": 0.14256497656582873, + "grad_norm": 18.25, + "learning_rate": 2.47215796723858e-06, + "loss": 1.4129, + "step": 1673 + }, + { + "epoch": 0.14265019173412868, + "grad_norm": 19.375, + "learning_rate": 2.4721226882846096e-06, + "loss": 1.0845, + "step": 1674 + }, + { + "epoch": 0.14273540690242864, + "grad_norm": 39.0, + "learning_rate": 2.4720873872456532e-06, + "loss": 1.4384, + "step": 1675 + }, + { + "epoch": 0.1428206220707286, + "grad_norm": 16.0, + "learning_rate": 2.472052064122349e-06, + "loss": 0.4952, + "step": 1676 + }, + { + "epoch": 0.14290583723902855, + "grad_norm": 10.5, + "learning_rate": 2.4720167189153352e-06, + "loss": 0.4848, + "step": 1677 + }, + { + "epoch": 0.1429910524073285, + "grad_norm": 16.25, + "learning_rate": 2.47198135162525e-06, + "loss": 0.5999, + "step": 1678 + }, + { + "epoch": 0.14307626757562847, + "grad_norm": 18.875, + "learning_rate": 2.4719459622527335e-06, + "loss": 1.2016, + "step": 1679 + }, + { + "epoch": 0.14316148274392843, + "grad_norm": 18.625, + "learning_rate": 2.4719105507984246e-06, + "loss": 0.6198, + "step": 1680 + }, + { + "epoch": 0.14324669791222838, + "grad_norm": 12.875, + "learning_rate": 2.4718751172629633e-06, + "loss": 0.6453, + "step": 1681 + }, + { + "epoch": 0.14333191308052834, + "grad_norm": 12.1875, + "learning_rate": 2.4718396616469896e-06, + "loss": 0.7325, + "step": 1682 + }, + { + "epoch": 0.1434171282488283, + "grad_norm": 13.8125, + "learning_rate": 2.471804183951145e-06, + "loss": 0.6485, + "step": 1683 + }, + { + "epoch": 0.14350234341712825, + "grad_norm": 13.0625, + "learning_rate": 2.4717686841760703e-06, + "loss": 0.7609, + "step": 1684 + }, + { + "epoch": 0.1435875585854282, + "grad_norm": 16.0, + "learning_rate": 2.4717331623224064e-06, + "loss": 0.4967, + "step": 1685 + }, + { + "epoch": 0.14367277375372817, + "grad_norm": 12.6875, + "learning_rate": 2.471697618390796e-06, + "loss": 0.6717, + "step": 1686 + }, + { + "epoch": 0.14375798892202812, + "grad_norm": 12.1875, + "learning_rate": 2.4716620523818805e-06, + "loss": 0.4425, + "step": 1687 + }, + { + "epoch": 0.14384320409032808, + "grad_norm": 16.625, + "learning_rate": 2.471626464296304e-06, + "loss": 0.8764, + "step": 1688 + }, + { + "epoch": 0.14392841925862804, + "grad_norm": 18.0, + "learning_rate": 2.471590854134709e-06, + "loss": 1.065, + "step": 1689 + }, + { + "epoch": 0.144013634426928, + "grad_norm": 15.0, + "learning_rate": 2.4715552218977383e-06, + "loss": 0.647, + "step": 1690 + }, + { + "epoch": 0.14409884959522795, + "grad_norm": 12.25, + "learning_rate": 2.4715195675860366e-06, + "loss": 0.7388, + "step": 1691 + }, + { + "epoch": 0.1441840647635279, + "grad_norm": 17.625, + "learning_rate": 2.471483891200248e-06, + "loss": 0.9272, + "step": 1692 + }, + { + "epoch": 0.14426927993182787, + "grad_norm": 13.0, + "learning_rate": 2.471448192741017e-06, + "loss": 0.6184, + "step": 1693 + }, + { + "epoch": 0.14435449510012782, + "grad_norm": 13.4375, + "learning_rate": 2.471412472208989e-06, + "loss": 0.7274, + "step": 1694 + }, + { + "epoch": 0.14443971026842778, + "grad_norm": 16.0, + "learning_rate": 2.47137672960481e-06, + "loss": 0.5611, + "step": 1695 + }, + { + "epoch": 0.14452492543672774, + "grad_norm": 16.75, + "learning_rate": 2.4713409649291243e-06, + "loss": 1.1467, + "step": 1696 + }, + { + "epoch": 0.1446101406050277, + "grad_norm": 15.625, + "learning_rate": 2.47130517818258e-06, + "loss": 0.7028, + "step": 1697 + }, + { + "epoch": 0.14469535577332765, + "grad_norm": 18.0, + "learning_rate": 2.4712693693658222e-06, + "loss": 0.9602, + "step": 1698 + }, + { + "epoch": 0.1447805709416276, + "grad_norm": 14.625, + "learning_rate": 2.4712335384794995e-06, + "loss": 0.6669, + "step": 1699 + }, + { + "epoch": 0.14486578610992756, + "grad_norm": 16.375, + "learning_rate": 2.471197685524258e-06, + "loss": 0.9796, + "step": 1700 + }, + { + "epoch": 0.14495100127822752, + "grad_norm": 18.875, + "learning_rate": 2.471161810500747e-06, + "loss": 1.2473, + "step": 1701 + }, + { + "epoch": 0.14503621644652748, + "grad_norm": 17.0, + "learning_rate": 2.471125913409614e-06, + "loss": 0.9406, + "step": 1702 + }, + { + "epoch": 0.14512143161482743, + "grad_norm": 9.9375, + "learning_rate": 2.4710899942515077e-06, + "loss": 0.2845, + "step": 1703 + }, + { + "epoch": 0.1452066467831274, + "grad_norm": 16.875, + "learning_rate": 2.471054053027077e-06, + "loss": 0.9779, + "step": 1704 + }, + { + "epoch": 0.14529186195142735, + "grad_norm": 12.875, + "learning_rate": 2.4710180897369725e-06, + "loss": 0.6043, + "step": 1705 + }, + { + "epoch": 0.1453770771197273, + "grad_norm": 15.8125, + "learning_rate": 2.470982104381843e-06, + "loss": 1.1665, + "step": 1706 + }, + { + "epoch": 0.14546229228802726, + "grad_norm": 16.25, + "learning_rate": 2.4709460969623385e-06, + "loss": 0.7901, + "step": 1707 + }, + { + "epoch": 0.14554750745632722, + "grad_norm": 19.125, + "learning_rate": 2.4709100674791108e-06, + "loss": 1.1028, + "step": 1708 + }, + { + "epoch": 0.14563272262462718, + "grad_norm": 11.75, + "learning_rate": 2.4708740159328105e-06, + "loss": 0.591, + "step": 1709 + }, + { + "epoch": 0.14571793779292713, + "grad_norm": 17.5, + "learning_rate": 2.4708379423240887e-06, + "loss": 0.8908, + "step": 1710 + }, + { + "epoch": 0.1458031529612271, + "grad_norm": 16.625, + "learning_rate": 2.4708018466535976e-06, + "loss": 0.9323, + "step": 1711 + }, + { + "epoch": 0.14588836812952705, + "grad_norm": 13.625, + "learning_rate": 2.4707657289219898e-06, + "loss": 0.6885, + "step": 1712 + }, + { + "epoch": 0.145973583297827, + "grad_norm": 19.0, + "learning_rate": 2.4707295891299175e-06, + "loss": 0.7523, + "step": 1713 + }, + { + "epoch": 0.14605879846612696, + "grad_norm": 14.5, + "learning_rate": 2.4706934272780342e-06, + "loss": 0.7756, + "step": 1714 + }, + { + "epoch": 0.14614401363442692, + "grad_norm": 13.75, + "learning_rate": 2.470657243366993e-06, + "loss": 0.759, + "step": 1715 + }, + { + "epoch": 0.14622922880272687, + "grad_norm": 13.75, + "learning_rate": 2.4706210373974483e-06, + "loss": 0.834, + "step": 1716 + }, + { + "epoch": 0.14631444397102683, + "grad_norm": 14.75, + "learning_rate": 2.4705848093700534e-06, + "loss": 0.9064, + "step": 1717 + }, + { + "epoch": 0.1463996591393268, + "grad_norm": 17.125, + "learning_rate": 2.470548559285464e-06, + "loss": 1.0722, + "step": 1718 + }, + { + "epoch": 0.14648487430762674, + "grad_norm": 16.25, + "learning_rate": 2.470512287144335e-06, + "loss": 0.8775, + "step": 1719 + }, + { + "epoch": 0.1465700894759267, + "grad_norm": 33.0, + "learning_rate": 2.4704759929473213e-06, + "loss": 1.0421, + "step": 1720 + }, + { + "epoch": 0.14665530464422669, + "grad_norm": 11.125, + "learning_rate": 2.470439676695079e-06, + "loss": 0.4325, + "step": 1721 + }, + { + "epoch": 0.14674051981252664, + "grad_norm": 13.625, + "learning_rate": 2.470403338388265e-06, + "loss": 0.5844, + "step": 1722 + }, + { + "epoch": 0.1468257349808266, + "grad_norm": 11.1875, + "learning_rate": 2.4703669780275353e-06, + "loss": 0.4866, + "step": 1723 + }, + { + "epoch": 0.14691095014912656, + "grad_norm": 19.75, + "learning_rate": 2.4703305956135466e-06, + "loss": 0.9827, + "step": 1724 + }, + { + "epoch": 0.1469961653174265, + "grad_norm": 18.5, + "learning_rate": 2.4702941911469573e-06, + "loss": 1.1083, + "step": 1725 + }, + { + "epoch": 0.14708138048572647, + "grad_norm": 17.625, + "learning_rate": 2.470257764628425e-06, + "loss": 1.0822, + "step": 1726 + }, + { + "epoch": 0.14716659565402643, + "grad_norm": 16.125, + "learning_rate": 2.470221316058608e-06, + "loss": 0.3616, + "step": 1727 + }, + { + "epoch": 0.14725181082232638, + "grad_norm": 11.8125, + "learning_rate": 2.4701848454381643e-06, + "loss": 0.4721, + "step": 1728 + }, + { + "epoch": 0.14733702599062634, + "grad_norm": 14.5, + "learning_rate": 2.4701483527677536e-06, + "loss": 0.7631, + "step": 1729 + }, + { + "epoch": 0.1474222411589263, + "grad_norm": 15.0, + "learning_rate": 2.470111838048035e-06, + "loss": 0.7506, + "step": 1730 + }, + { + "epoch": 0.14750745632722626, + "grad_norm": 21.25, + "learning_rate": 2.470075301279669e-06, + "loss": 0.8739, + "step": 1731 + }, + { + "epoch": 0.1475926714955262, + "grad_norm": 15.625, + "learning_rate": 2.4700387424633154e-06, + "loss": 0.8766, + "step": 1732 + }, + { + "epoch": 0.14767788666382617, + "grad_norm": 17.25, + "learning_rate": 2.4700021615996343e-06, + "loss": 1.1099, + "step": 1733 + }, + { + "epoch": 0.14776310183212613, + "grad_norm": 30.25, + "learning_rate": 2.4699655586892877e-06, + "loss": 1.156, + "step": 1734 + }, + { + "epoch": 0.14784831700042608, + "grad_norm": 24.625, + "learning_rate": 2.4699289337329373e-06, + "loss": 0.9198, + "step": 1735 + }, + { + "epoch": 0.14793353216872604, + "grad_norm": 11.1875, + "learning_rate": 2.4698922867312436e-06, + "loss": 0.342, + "step": 1736 + }, + { + "epoch": 0.148018747337026, + "grad_norm": 12.6875, + "learning_rate": 2.46985561768487e-06, + "loss": 0.7033, + "step": 1737 + }, + { + "epoch": 0.14810396250532595, + "grad_norm": 20.625, + "learning_rate": 2.469818926594478e-06, + "loss": 1.0989, + "step": 1738 + }, + { + "epoch": 0.1481891776736259, + "grad_norm": 16.375, + "learning_rate": 2.469782213460732e-06, + "loss": 0.9749, + "step": 1739 + }, + { + "epoch": 0.14827439284192587, + "grad_norm": 9.75, + "learning_rate": 2.4697454782842946e-06, + "loss": 0.4032, + "step": 1740 + }, + { + "epoch": 0.14835960801022582, + "grad_norm": 12.0625, + "learning_rate": 2.4697087210658297e-06, + "loss": 0.4214, + "step": 1741 + }, + { + "epoch": 0.14844482317852578, + "grad_norm": 11.5625, + "learning_rate": 2.4696719418060023e-06, + "loss": 0.5647, + "step": 1742 + }, + { + "epoch": 0.14853003834682574, + "grad_norm": 14.1875, + "learning_rate": 2.4696351405054763e-06, + "loss": 0.9841, + "step": 1743 + }, + { + "epoch": 0.1486152535151257, + "grad_norm": 15.1875, + "learning_rate": 2.4695983171649162e-06, + "loss": 0.804, + "step": 1744 + }, + { + "epoch": 0.14870046868342565, + "grad_norm": 13.8125, + "learning_rate": 2.4695614717849888e-06, + "loss": 0.9698, + "step": 1745 + }, + { + "epoch": 0.1487856838517256, + "grad_norm": 16.875, + "learning_rate": 2.469524604366359e-06, + "loss": 0.7278, + "step": 1746 + }, + { + "epoch": 0.14887089902002557, + "grad_norm": 12.6875, + "learning_rate": 2.4694877149096934e-06, + "loss": 0.5082, + "step": 1747 + }, + { + "epoch": 0.14895611418832552, + "grad_norm": 18.625, + "learning_rate": 2.4694508034156583e-06, + "loss": 0.9462, + "step": 1748 + }, + { + "epoch": 0.14904132935662548, + "grad_norm": 11.75, + "learning_rate": 2.4694138698849206e-06, + "loss": 0.4739, + "step": 1749 + }, + { + "epoch": 0.14912654452492544, + "grad_norm": 13.8125, + "learning_rate": 2.469376914318149e-06, + "loss": 0.5964, + "step": 1750 + }, + { + "epoch": 0.1492117596932254, + "grad_norm": 20.25, + "learning_rate": 2.4693399367160095e-06, + "loss": 0.9735, + "step": 1751 + }, + { + "epoch": 0.14929697486152535, + "grad_norm": 16.25, + "learning_rate": 2.4693029370791715e-06, + "loss": 1.0902, + "step": 1752 + }, + { + "epoch": 0.1493821900298253, + "grad_norm": 11.8125, + "learning_rate": 2.4692659154083033e-06, + "loss": 0.72, + "step": 1753 + }, + { + "epoch": 0.14946740519812526, + "grad_norm": 13.0, + "learning_rate": 2.469228871704074e-06, + "loss": 0.6453, + "step": 1754 + }, + { + "epoch": 0.14955262036642522, + "grad_norm": 11.5625, + "learning_rate": 2.469191805967153e-06, + "loss": 0.5758, + "step": 1755 + }, + { + "epoch": 0.14963783553472518, + "grad_norm": 12.1875, + "learning_rate": 2.4691547181982102e-06, + "loss": 0.5256, + "step": 1756 + }, + { + "epoch": 0.14972305070302513, + "grad_norm": 16.75, + "learning_rate": 2.4691176083979155e-06, + "loss": 0.9599, + "step": 1757 + }, + { + "epoch": 0.1498082658713251, + "grad_norm": 19.25, + "learning_rate": 2.4690804765669396e-06, + "loss": 0.9446, + "step": 1758 + }, + { + "epoch": 0.14989348103962505, + "grad_norm": 19.0, + "learning_rate": 2.469043322705954e-06, + "loss": 1.1701, + "step": 1759 + }, + { + "epoch": 0.149978696207925, + "grad_norm": 13.9375, + "learning_rate": 2.4690061468156293e-06, + "loss": 0.7094, + "step": 1760 + }, + { + "epoch": 0.15006391137622496, + "grad_norm": 10.0625, + "learning_rate": 2.4689689488966377e-06, + "loss": 0.292, + "step": 1761 + }, + { + "epoch": 0.15014912654452492, + "grad_norm": 11.8125, + "learning_rate": 2.468931728949652e-06, + "loss": 0.5294, + "step": 1762 + }, + { + "epoch": 0.15023434171282488, + "grad_norm": 14.6875, + "learning_rate": 2.4688944869753435e-06, + "loss": 0.9632, + "step": 1763 + }, + { + "epoch": 0.15031955688112483, + "grad_norm": 12.1875, + "learning_rate": 2.4688572229743863e-06, + "loss": 0.3175, + "step": 1764 + }, + { + "epoch": 0.1504047720494248, + "grad_norm": 30.125, + "learning_rate": 2.4688199369474532e-06, + "loss": 1.283, + "step": 1765 + }, + { + "epoch": 0.15048998721772475, + "grad_norm": 15.375, + "learning_rate": 2.4687826288952186e-06, + "loss": 0.9624, + "step": 1766 + }, + { + "epoch": 0.1505752023860247, + "grad_norm": 16.5, + "learning_rate": 2.468745298818356e-06, + "loss": 0.8091, + "step": 1767 + }, + { + "epoch": 0.15066041755432466, + "grad_norm": 15.25, + "learning_rate": 2.4687079467175402e-06, + "loss": 0.6241, + "step": 1768 + }, + { + "epoch": 0.15074563272262462, + "grad_norm": 13.0625, + "learning_rate": 2.4686705725934467e-06, + "loss": 0.6405, + "step": 1769 + }, + { + "epoch": 0.15083084789092457, + "grad_norm": 15.0, + "learning_rate": 2.46863317644675e-06, + "loss": 0.8592, + "step": 1770 + }, + { + "epoch": 0.15091606305922453, + "grad_norm": 19.125, + "learning_rate": 2.468595758278127e-06, + "loss": 0.775, + "step": 1771 + }, + { + "epoch": 0.1510012782275245, + "grad_norm": 10.6875, + "learning_rate": 2.468558318088253e-06, + "loss": 0.4237, + "step": 1772 + }, + { + "epoch": 0.15108649339582444, + "grad_norm": 26.375, + "learning_rate": 2.4685208558778045e-06, + "loss": 0.6044, + "step": 1773 + }, + { + "epoch": 0.1511717085641244, + "grad_norm": 22.125, + "learning_rate": 2.4684833716474594e-06, + "loss": 1.1508, + "step": 1774 + }, + { + "epoch": 0.15125692373242436, + "grad_norm": 13.5625, + "learning_rate": 2.468445865397894e-06, + "loss": 0.6571, + "step": 1775 + }, + { + "epoch": 0.15134213890072432, + "grad_norm": 11.75, + "learning_rate": 2.4684083371297866e-06, + "loss": 0.4636, + "step": 1776 + }, + { + "epoch": 0.15142735406902427, + "grad_norm": 14.0625, + "learning_rate": 2.4683707868438156e-06, + "loss": 0.8693, + "step": 1777 + }, + { + "epoch": 0.15151256923732426, + "grad_norm": 13.4375, + "learning_rate": 2.468333214540659e-06, + "loss": 0.7335, + "step": 1778 + }, + { + "epoch": 0.15159778440562421, + "grad_norm": 20.125, + "learning_rate": 2.4682956202209963e-06, + "loss": 0.9285, + "step": 1779 + }, + { + "epoch": 0.15168299957392417, + "grad_norm": 16.125, + "learning_rate": 2.468258003885507e-06, + "loss": 1.2837, + "step": 1780 + }, + { + "epoch": 0.15176821474222413, + "grad_norm": 37.0, + "learning_rate": 2.46822036553487e-06, + "loss": 1.6967, + "step": 1781 + }, + { + "epoch": 0.15185342991052408, + "grad_norm": 14.75, + "learning_rate": 2.468182705169766e-06, + "loss": 1.1296, + "step": 1782 + }, + { + "epoch": 0.15193864507882404, + "grad_norm": 15.0, + "learning_rate": 2.468145022790876e-06, + "loss": 0.8443, + "step": 1783 + }, + { + "epoch": 0.152023860247124, + "grad_norm": 18.375, + "learning_rate": 2.46810731839888e-06, + "loss": 1.1151, + "step": 1784 + }, + { + "epoch": 0.15210907541542396, + "grad_norm": 13.5625, + "learning_rate": 2.46806959199446e-06, + "loss": 0.3695, + "step": 1785 + }, + { + "epoch": 0.1521942905837239, + "grad_norm": 18.625, + "learning_rate": 2.4680318435782977e-06, + "loss": 0.9289, + "step": 1786 + }, + { + "epoch": 0.15227950575202387, + "grad_norm": 13.6875, + "learning_rate": 2.467994073151075e-06, + "loss": 0.7206, + "step": 1787 + }, + { + "epoch": 0.15236472092032383, + "grad_norm": 13.375, + "learning_rate": 2.4679562807134746e-06, + "loss": 0.5083, + "step": 1788 + }, + { + "epoch": 0.15244993608862378, + "grad_norm": 19.75, + "learning_rate": 2.4679184662661797e-06, + "loss": 1.1709, + "step": 1789 + }, + { + "epoch": 0.15253515125692374, + "grad_norm": 19.625, + "learning_rate": 2.4678806298098732e-06, + "loss": 1.1609, + "step": 1790 + }, + { + "epoch": 0.1526203664252237, + "grad_norm": 11.875, + "learning_rate": 2.467842771345239e-06, + "loss": 0.7869, + "step": 1791 + }, + { + "epoch": 0.15270558159352365, + "grad_norm": 12.3125, + "learning_rate": 2.4678048908729615e-06, + "loss": 0.7349, + "step": 1792 + }, + { + "epoch": 0.1527907967618236, + "grad_norm": 15.5625, + "learning_rate": 2.467766988393725e-06, + "loss": 0.8078, + "step": 1793 + }, + { + "epoch": 0.15287601193012357, + "grad_norm": 16.875, + "learning_rate": 2.4677290639082143e-06, + "loss": 0.7204, + "step": 1794 + }, + { + "epoch": 0.15296122709842352, + "grad_norm": 13.125, + "learning_rate": 2.467691117417115e-06, + "loss": 0.6347, + "step": 1795 + }, + { + "epoch": 0.15304644226672348, + "grad_norm": 15.9375, + "learning_rate": 2.4676531489211127e-06, + "loss": 0.9882, + "step": 1796 + }, + { + "epoch": 0.15313165743502344, + "grad_norm": 13.75, + "learning_rate": 2.4676151584208937e-06, + "loss": 0.638, + "step": 1797 + }, + { + "epoch": 0.1532168726033234, + "grad_norm": 17.25, + "learning_rate": 2.4675771459171443e-06, + "loss": 0.7384, + "step": 1798 + }, + { + "epoch": 0.15330208777162335, + "grad_norm": 11.75, + "learning_rate": 2.467539111410551e-06, + "loss": 0.4922, + "step": 1799 + }, + { + "epoch": 0.1533873029399233, + "grad_norm": 19.75, + "learning_rate": 2.4675010549018024e-06, + "loss": 1.1412, + "step": 1800 + }, + { + "epoch": 0.15347251810822327, + "grad_norm": 11.375, + "learning_rate": 2.467462976391585e-06, + "loss": 0.4821, + "step": 1801 + }, + { + "epoch": 0.15355773327652322, + "grad_norm": 22.625, + "learning_rate": 2.4674248758805873e-06, + "loss": 1.1391, + "step": 1802 + }, + { + "epoch": 0.15364294844482318, + "grad_norm": 15.25, + "learning_rate": 2.4673867533694982e-06, + "loss": 1.0296, + "step": 1803 + }, + { + "epoch": 0.15372816361312314, + "grad_norm": 15.6875, + "learning_rate": 2.4673486088590064e-06, + "loss": 0.8798, + "step": 1804 + }, + { + "epoch": 0.1538133787814231, + "grad_norm": 15.9375, + "learning_rate": 2.4673104423498005e-06, + "loss": 1.0988, + "step": 1805 + }, + { + "epoch": 0.15389859394972305, + "grad_norm": 16.625, + "learning_rate": 2.4672722538425713e-06, + "loss": 1.1313, + "step": 1806 + }, + { + "epoch": 0.153983809118023, + "grad_norm": 13.75, + "learning_rate": 2.4672340433380083e-06, + "loss": 0.7441, + "step": 1807 + }, + { + "epoch": 0.15406902428632296, + "grad_norm": 20.375, + "learning_rate": 2.467195810836802e-06, + "loss": 0.914, + "step": 1808 + }, + { + "epoch": 0.15415423945462292, + "grad_norm": 12.0625, + "learning_rate": 2.4671575563396437e-06, + "loss": 0.3241, + "step": 1809 + }, + { + "epoch": 0.15423945462292288, + "grad_norm": 21.875, + "learning_rate": 2.467119279847224e-06, + "loss": 1.3322, + "step": 1810 + }, + { + "epoch": 0.15432466979122283, + "grad_norm": 9.8125, + "learning_rate": 2.4670809813602355e-06, + "loss": 0.3766, + "step": 1811 + }, + { + "epoch": 0.1544098849595228, + "grad_norm": 14.9375, + "learning_rate": 2.467042660879369e-06, + "loss": 0.6851, + "step": 1812 + }, + { + "epoch": 0.15449510012782275, + "grad_norm": 14.375, + "learning_rate": 2.4670043184053184e-06, + "loss": 0.9174, + "step": 1813 + }, + { + "epoch": 0.1545803152961227, + "grad_norm": 18.875, + "learning_rate": 2.466965953938776e-06, + "loss": 1.0705, + "step": 1814 + }, + { + "epoch": 0.15466553046442266, + "grad_norm": 20.25, + "learning_rate": 2.4669275674804347e-06, + "loss": 1.5079, + "step": 1815 + }, + { + "epoch": 0.15475074563272262, + "grad_norm": 19.375, + "learning_rate": 2.466889159030989e-06, + "loss": 0.9318, + "step": 1816 + }, + { + "epoch": 0.15483596080102258, + "grad_norm": 18.625, + "learning_rate": 2.466850728591132e-06, + "loss": 0.8093, + "step": 1817 + }, + { + "epoch": 0.15492117596932253, + "grad_norm": 12.125, + "learning_rate": 2.466812276161559e-06, + "loss": 0.4672, + "step": 1818 + }, + { + "epoch": 0.1550063911376225, + "grad_norm": 14.3125, + "learning_rate": 2.4667738017429648e-06, + "loss": 0.6757, + "step": 1819 + }, + { + "epoch": 0.15509160630592245, + "grad_norm": 11.375, + "learning_rate": 2.466735305336044e-06, + "loss": 0.5264, + "step": 1820 + }, + { + "epoch": 0.1551768214742224, + "grad_norm": 14.3125, + "learning_rate": 2.4666967869414932e-06, + "loss": 0.7891, + "step": 1821 + }, + { + "epoch": 0.15526203664252236, + "grad_norm": 21.125, + "learning_rate": 2.4666582465600073e-06, + "loss": 0.8732, + "step": 1822 + }, + { + "epoch": 0.15534725181082232, + "grad_norm": 12.1875, + "learning_rate": 2.466619684192284e-06, + "loss": 0.5744, + "step": 1823 + }, + { + "epoch": 0.15543246697912227, + "grad_norm": 11.625, + "learning_rate": 2.4665810998390196e-06, + "loss": 0.5755, + "step": 1824 + }, + { + "epoch": 0.15551768214742223, + "grad_norm": 13.8125, + "learning_rate": 2.4665424935009106e-06, + "loss": 0.746, + "step": 1825 + }, + { + "epoch": 0.1556028973157222, + "grad_norm": 15.125, + "learning_rate": 2.466503865178656e-06, + "loss": 0.7084, + "step": 1826 + }, + { + "epoch": 0.15568811248402215, + "grad_norm": 16.25, + "learning_rate": 2.4664652148729533e-06, + "loss": 0.6737, + "step": 1827 + }, + { + "epoch": 0.1557733276523221, + "grad_norm": 12.0625, + "learning_rate": 2.466426542584501e-06, + "loss": 0.6415, + "step": 1828 + }, + { + "epoch": 0.15585854282062206, + "grad_norm": 12.125, + "learning_rate": 2.466387848313998e-06, + "loss": 0.8392, + "step": 1829 + }, + { + "epoch": 0.15594375798892202, + "grad_norm": 14.8125, + "learning_rate": 2.466349132062143e-06, + "loss": 0.7633, + "step": 1830 + }, + { + "epoch": 0.15602897315722197, + "grad_norm": 22.875, + "learning_rate": 2.466310393829636e-06, + "loss": 1.0498, + "step": 1831 + }, + { + "epoch": 0.15611418832552193, + "grad_norm": 12.0625, + "learning_rate": 2.4662716336171778e-06, + "loss": 0.4587, + "step": 1832 + }, + { + "epoch": 0.1561994034938219, + "grad_norm": 15.6875, + "learning_rate": 2.4662328514254672e-06, + "loss": 0.6356, + "step": 1833 + }, + { + "epoch": 0.15628461866212187, + "grad_norm": 17.5, + "learning_rate": 2.4661940472552067e-06, + "loss": 1.142, + "step": 1834 + }, + { + "epoch": 0.15636983383042183, + "grad_norm": 13.25, + "learning_rate": 2.466155221107096e-06, + "loss": 0.7789, + "step": 1835 + }, + { + "epoch": 0.15645504899872179, + "grad_norm": 16.125, + "learning_rate": 2.4661163729818383e-06, + "loss": 0.9163, + "step": 1836 + }, + { + "epoch": 0.15654026416702174, + "grad_norm": 15.1875, + "learning_rate": 2.4660775028801345e-06, + "loss": 0.5911, + "step": 1837 + }, + { + "epoch": 0.1566254793353217, + "grad_norm": 14.5, + "learning_rate": 2.4660386108026872e-06, + "loss": 0.8785, + "step": 1838 + }, + { + "epoch": 0.15671069450362166, + "grad_norm": 23.75, + "learning_rate": 2.4659996967501997e-06, + "loss": 0.7775, + "step": 1839 + }, + { + "epoch": 0.1567959096719216, + "grad_norm": 12.0, + "learning_rate": 2.4659607607233748e-06, + "loss": 0.4976, + "step": 1840 + }, + { + "epoch": 0.15688112484022157, + "grad_norm": 19.5, + "learning_rate": 2.465921802722916e-06, + "loss": 0.9605, + "step": 1841 + }, + { + "epoch": 0.15696634000852153, + "grad_norm": 16.125, + "learning_rate": 2.465882822749528e-06, + "loss": 0.8789, + "step": 1842 + }, + { + "epoch": 0.15705155517682148, + "grad_norm": 15.4375, + "learning_rate": 2.465843820803914e-06, + "loss": 1.0392, + "step": 1843 + }, + { + "epoch": 0.15713677034512144, + "grad_norm": 13.75, + "learning_rate": 2.46580479688678e-06, + "loss": 0.5005, + "step": 1844 + }, + { + "epoch": 0.1572219855134214, + "grad_norm": 20.25, + "learning_rate": 2.465765750998831e-06, + "loss": 1.0982, + "step": 1845 + }, + { + "epoch": 0.15730720068172135, + "grad_norm": 13.125, + "learning_rate": 2.4657266831407717e-06, + "loss": 0.7649, + "step": 1846 + }, + { + "epoch": 0.1573924158500213, + "grad_norm": 30.625, + "learning_rate": 2.465687593313309e-06, + "loss": 0.9712, + "step": 1847 + }, + { + "epoch": 0.15747763101832127, + "grad_norm": 11.9375, + "learning_rate": 2.4656484815171494e-06, + "loss": 0.7227, + "step": 1848 + }, + { + "epoch": 0.15756284618662122, + "grad_norm": 13.625, + "learning_rate": 2.4656093477529986e-06, + "loss": 0.7647, + "step": 1849 + }, + { + "epoch": 0.15764806135492118, + "grad_norm": 13.6875, + "learning_rate": 2.465570192021565e-06, + "loss": 0.6293, + "step": 1850 + }, + { + "epoch": 0.15773327652322114, + "grad_norm": 17.75, + "learning_rate": 2.4655310143235556e-06, + "loss": 0.6958, + "step": 1851 + }, + { + "epoch": 0.1578184916915211, + "grad_norm": 27.0, + "learning_rate": 2.4654918146596783e-06, + "loss": 0.9169, + "step": 1852 + }, + { + "epoch": 0.15790370685982105, + "grad_norm": 19.125, + "learning_rate": 2.465452593030642e-06, + "loss": 0.9063, + "step": 1853 + }, + { + "epoch": 0.157988922028121, + "grad_norm": 13.875, + "learning_rate": 2.4654133494371547e-06, + "loss": 0.5622, + "step": 1854 + }, + { + "epoch": 0.15807413719642097, + "grad_norm": 9.6875, + "learning_rate": 2.465374083879926e-06, + "loss": 0.3138, + "step": 1855 + }, + { + "epoch": 0.15815935236472092, + "grad_norm": 12.0625, + "learning_rate": 2.465334796359666e-06, + "loss": 0.5773, + "step": 1856 + }, + { + "epoch": 0.15824456753302088, + "grad_norm": 21.125, + "learning_rate": 2.4652954868770834e-06, + "loss": 0.8814, + "step": 1857 + }, + { + "epoch": 0.15832978270132084, + "grad_norm": 31.0, + "learning_rate": 2.4652561554328895e-06, + "loss": 0.8206, + "step": 1858 + }, + { + "epoch": 0.1584149978696208, + "grad_norm": 12.5, + "learning_rate": 2.465216802027795e-06, + "loss": 0.6012, + "step": 1859 + }, + { + "epoch": 0.15850021303792075, + "grad_norm": 13.75, + "learning_rate": 2.465177426662511e-06, + "loss": 0.7477, + "step": 1860 + }, + { + "epoch": 0.1585854282062207, + "grad_norm": 15.5, + "learning_rate": 2.4651380293377486e-06, + "loss": 1.0542, + "step": 1861 + }, + { + "epoch": 0.15867064337452066, + "grad_norm": 15.625, + "learning_rate": 2.46509861005422e-06, + "loss": 0.745, + "step": 1862 + }, + { + "epoch": 0.15875585854282062, + "grad_norm": 16.375, + "learning_rate": 2.4650591688126384e-06, + "loss": 0.7779, + "step": 1863 + }, + { + "epoch": 0.15884107371112058, + "grad_norm": 19.0, + "learning_rate": 2.465019705613715e-06, + "loss": 0.9052, + "step": 1864 + }, + { + "epoch": 0.15892628887942054, + "grad_norm": 15.1875, + "learning_rate": 2.4649802204581643e-06, + "loss": 0.7703, + "step": 1865 + }, + { + "epoch": 0.1590115040477205, + "grad_norm": 13.25, + "learning_rate": 2.4649407133466987e-06, + "loss": 0.639, + "step": 1866 + }, + { + "epoch": 0.15909671921602045, + "grad_norm": 17.125, + "learning_rate": 2.464901184280033e-06, + "loss": 0.809, + "step": 1867 + }, + { + "epoch": 0.1591819343843204, + "grad_norm": 17.375, + "learning_rate": 2.4648616332588807e-06, + "loss": 0.6356, + "step": 1868 + }, + { + "epoch": 0.15926714955262036, + "grad_norm": 16.625, + "learning_rate": 2.4648220602839577e-06, + "loss": 0.9886, + "step": 1869 + }, + { + "epoch": 0.15935236472092032, + "grad_norm": 16.125, + "learning_rate": 2.4647824653559786e-06, + "loss": 1.0655, + "step": 1870 + }, + { + "epoch": 0.15943757988922028, + "grad_norm": 14.5625, + "learning_rate": 2.4647428484756584e-06, + "loss": 0.9297, + "step": 1871 + }, + { + "epoch": 0.15952279505752023, + "grad_norm": 14.25, + "learning_rate": 2.4647032096437133e-06, + "loss": 1.1566, + "step": 1872 + }, + { + "epoch": 0.1596080102258202, + "grad_norm": 14.875, + "learning_rate": 2.46466354886086e-06, + "loss": 0.9909, + "step": 1873 + }, + { + "epoch": 0.15969322539412015, + "grad_norm": 18.875, + "learning_rate": 2.4646238661278148e-06, + "loss": 0.9221, + "step": 1874 + }, + { + "epoch": 0.1597784405624201, + "grad_norm": 18.875, + "learning_rate": 2.4645841614452955e-06, + "loss": 0.9497, + "step": 1875 + }, + { + "epoch": 0.15986365573072006, + "grad_norm": 8.8125, + "learning_rate": 2.464544434814018e-06, + "loss": 0.3158, + "step": 1876 + }, + { + "epoch": 0.15994887089902002, + "grad_norm": 22.625, + "learning_rate": 2.4645046862347024e-06, + "loss": 1.2811, + "step": 1877 + }, + { + "epoch": 0.16003408606731997, + "grad_norm": 24.375, + "learning_rate": 2.4644649157080647e-06, + "loss": 0.6415, + "step": 1878 + }, + { + "epoch": 0.16011930123561993, + "grad_norm": 16.125, + "learning_rate": 2.4644251232348253e-06, + "loss": 1.0919, + "step": 1879 + }, + { + "epoch": 0.1602045164039199, + "grad_norm": 16.0, + "learning_rate": 2.4643853088157026e-06, + "loss": 0.6815, + "step": 1880 + }, + { + "epoch": 0.16028973157221985, + "grad_norm": 13.125, + "learning_rate": 2.4643454724514165e-06, + "loss": 0.766, + "step": 1881 + }, + { + "epoch": 0.1603749467405198, + "grad_norm": 15.375, + "learning_rate": 2.464305614142686e-06, + "loss": 0.843, + "step": 1882 + }, + { + "epoch": 0.16046016190881976, + "grad_norm": 14.375, + "learning_rate": 2.4642657338902324e-06, + "loss": 0.5713, + "step": 1883 + }, + { + "epoch": 0.16054537707711972, + "grad_norm": 16.625, + "learning_rate": 2.464225831694776e-06, + "loss": 0.2904, + "step": 1884 + }, + { + "epoch": 0.16063059224541967, + "grad_norm": 35.5, + "learning_rate": 2.464185907557038e-06, + "loss": 0.6779, + "step": 1885 + }, + { + "epoch": 0.16071580741371963, + "grad_norm": 10.625, + "learning_rate": 2.4641459614777393e-06, + "loss": 0.4465, + "step": 1886 + }, + { + "epoch": 0.1608010225820196, + "grad_norm": 12.875, + "learning_rate": 2.4641059934576022e-06, + "loss": 0.3821, + "step": 1887 + }, + { + "epoch": 0.16088623775031954, + "grad_norm": 17.625, + "learning_rate": 2.4640660034973492e-06, + "loss": 0.8156, + "step": 1888 + }, + { + "epoch": 0.1609714529186195, + "grad_norm": 25.0, + "learning_rate": 2.4640259915977024e-06, + "loss": 0.854, + "step": 1889 + }, + { + "epoch": 0.16105666808691946, + "grad_norm": 10.75, + "learning_rate": 2.463985957759385e-06, + "loss": 0.3041, + "step": 1890 + }, + { + "epoch": 0.16114188325521944, + "grad_norm": 15.25, + "learning_rate": 2.463945901983121e-06, + "loss": 0.8962, + "step": 1891 + }, + { + "epoch": 0.1612270984235194, + "grad_norm": 10.3125, + "learning_rate": 2.4639058242696337e-06, + "loss": 0.3486, + "step": 1892 + }, + { + "epoch": 0.16131231359181936, + "grad_norm": 11.8125, + "learning_rate": 2.4638657246196474e-06, + "loss": 0.5969, + "step": 1893 + }, + { + "epoch": 0.1613975287601193, + "grad_norm": 34.5, + "learning_rate": 2.463825603033887e-06, + "loss": 1.2165, + "step": 1894 + }, + { + "epoch": 0.16148274392841927, + "grad_norm": 12.9375, + "learning_rate": 2.4637854595130774e-06, + "loss": 0.6792, + "step": 1895 + }, + { + "epoch": 0.16156795909671923, + "grad_norm": 17.625, + "learning_rate": 2.4637452940579434e-06, + "loss": 0.7919, + "step": 1896 + }, + { + "epoch": 0.16165317426501918, + "grad_norm": 11.6875, + "learning_rate": 2.463705106669212e-06, + "loss": 0.5261, + "step": 1897 + }, + { + "epoch": 0.16173838943331914, + "grad_norm": 30.25, + "learning_rate": 2.4636648973476086e-06, + "loss": 1.0703, + "step": 1898 + }, + { + "epoch": 0.1618236046016191, + "grad_norm": 16.75, + "learning_rate": 2.46362466609386e-06, + "loss": 0.9846, + "step": 1899 + }, + { + "epoch": 0.16190881976991905, + "grad_norm": 16.5, + "learning_rate": 2.4635844129086934e-06, + "loss": 0.5692, + "step": 1900 + }, + { + "epoch": 0.161994034938219, + "grad_norm": 17.75, + "learning_rate": 2.463544137792836e-06, + "loss": 0.9089, + "step": 1901 + }, + { + "epoch": 0.16207925010651897, + "grad_norm": 21.5, + "learning_rate": 2.463503840747016e-06, + "loss": 1.0333, + "step": 1902 + }, + { + "epoch": 0.16216446527481893, + "grad_norm": 10.25, + "learning_rate": 2.463463521771961e-06, + "loss": 0.3731, + "step": 1903 + }, + { + "epoch": 0.16224968044311888, + "grad_norm": 13.375, + "learning_rate": 2.4634231808684e-06, + "loss": 0.5566, + "step": 1904 + }, + { + "epoch": 0.16233489561141884, + "grad_norm": 12.0625, + "learning_rate": 2.463382818037062e-06, + "loss": 0.5679, + "step": 1905 + }, + { + "epoch": 0.1624201107797188, + "grad_norm": 13.5, + "learning_rate": 2.4633424332786767e-06, + "loss": 0.8186, + "step": 1906 + }, + { + "epoch": 0.16250532594801875, + "grad_norm": 15.625, + "learning_rate": 2.463302026593973e-06, + "loss": 0.7229, + "step": 1907 + }, + { + "epoch": 0.1625905411163187, + "grad_norm": 14.5625, + "learning_rate": 2.463261597983682e-06, + "loss": 0.8144, + "step": 1908 + }, + { + "epoch": 0.16267575628461867, + "grad_norm": 16.25, + "learning_rate": 2.4632211474485333e-06, + "loss": 0.9194, + "step": 1909 + }, + { + "epoch": 0.16276097145291862, + "grad_norm": 14.1875, + "learning_rate": 2.463180674989259e-06, + "loss": 0.5902, + "step": 1910 + }, + { + "epoch": 0.16284618662121858, + "grad_norm": 11.1875, + "learning_rate": 2.4631401806065898e-06, + "loss": 0.4058, + "step": 1911 + }, + { + "epoch": 0.16293140178951854, + "grad_norm": 15.1875, + "learning_rate": 2.4630996643012577e-06, + "loss": 1.0518, + "step": 1912 + }, + { + "epoch": 0.1630166169578185, + "grad_norm": 16.125, + "learning_rate": 2.4630591260739946e-06, + "loss": 0.8537, + "step": 1913 + }, + { + "epoch": 0.16310183212611845, + "grad_norm": 22.875, + "learning_rate": 2.4630185659255333e-06, + "loss": 1.2322, + "step": 1914 + }, + { + "epoch": 0.1631870472944184, + "grad_norm": 13.4375, + "learning_rate": 2.462977983856607e-06, + "loss": 0.5618, + "step": 1915 + }, + { + "epoch": 0.16327226246271836, + "grad_norm": 12.0, + "learning_rate": 2.4629373798679487e-06, + "loss": 0.6039, + "step": 1916 + }, + { + "epoch": 0.16335747763101832, + "grad_norm": 13.375, + "learning_rate": 2.462896753960292e-06, + "loss": 0.7162, + "step": 1917 + }, + { + "epoch": 0.16344269279931828, + "grad_norm": 15.9375, + "learning_rate": 2.4628561061343716e-06, + "loss": 1.0867, + "step": 1918 + }, + { + "epoch": 0.16352790796761824, + "grad_norm": 16.25, + "learning_rate": 2.462815436390922e-06, + "loss": 0.9963, + "step": 1919 + }, + { + "epoch": 0.1636131231359182, + "grad_norm": 13.1875, + "learning_rate": 2.462774744730677e-06, + "loss": 0.6273, + "step": 1920 + }, + { + "epoch": 0.16369833830421815, + "grad_norm": 13.3125, + "learning_rate": 2.4627340311543735e-06, + "loss": 0.7721, + "step": 1921 + }, + { + "epoch": 0.1637835534725181, + "grad_norm": 18.5, + "learning_rate": 2.462693295662746e-06, + "loss": 0.6745, + "step": 1922 + }, + { + "epoch": 0.16386876864081806, + "grad_norm": 11.5625, + "learning_rate": 2.462652538256532e-06, + "loss": 0.4459, + "step": 1923 + }, + { + "epoch": 0.16395398380911802, + "grad_norm": 12.4375, + "learning_rate": 2.462611758936467e-06, + "loss": 0.6698, + "step": 1924 + }, + { + "epoch": 0.16403919897741798, + "grad_norm": 13.5625, + "learning_rate": 2.4625709577032876e-06, + "loss": 0.9035, + "step": 1925 + }, + { + "epoch": 0.16412441414571793, + "grad_norm": 14.625, + "learning_rate": 2.4625301345577318e-06, + "loss": 0.8877, + "step": 1926 + }, + { + "epoch": 0.1642096293140179, + "grad_norm": 16.875, + "learning_rate": 2.4624892895005377e-06, + "loss": 0.8654, + "step": 1927 + }, + { + "epoch": 0.16429484448231785, + "grad_norm": 12.875, + "learning_rate": 2.4624484225324424e-06, + "loss": 0.8497, + "step": 1928 + }, + { + "epoch": 0.1643800596506178, + "grad_norm": 17.875, + "learning_rate": 2.4624075336541848e-06, + "loss": 0.8131, + "step": 1929 + }, + { + "epoch": 0.16446527481891776, + "grad_norm": 16.75, + "learning_rate": 2.4623666228665043e-06, + "loss": 0.8727, + "step": 1930 + }, + { + "epoch": 0.16455048998721772, + "grad_norm": 14.875, + "learning_rate": 2.4623256901701396e-06, + "loss": 0.8846, + "step": 1931 + }, + { + "epoch": 0.16463570515551768, + "grad_norm": 12.0, + "learning_rate": 2.4622847355658306e-06, + "loss": 0.7339, + "step": 1932 + }, + { + "epoch": 0.16472092032381763, + "grad_norm": 17.5, + "learning_rate": 2.4622437590543176e-06, + "loss": 0.9891, + "step": 1933 + }, + { + "epoch": 0.1648061354921176, + "grad_norm": 14.125, + "learning_rate": 2.46220276063634e-06, + "loss": 0.5347, + "step": 1934 + }, + { + "epoch": 0.16489135066041755, + "grad_norm": 15.9375, + "learning_rate": 2.4621617403126407e-06, + "loss": 1.1538, + "step": 1935 + }, + { + "epoch": 0.1649765658287175, + "grad_norm": 20.625, + "learning_rate": 2.462120698083959e-06, + "loss": 1.0269, + "step": 1936 + }, + { + "epoch": 0.16506178099701746, + "grad_norm": 12.9375, + "learning_rate": 2.4620796339510382e-06, + "loss": 0.7374, + "step": 1937 + }, + { + "epoch": 0.16514699616531742, + "grad_norm": 11.9375, + "learning_rate": 2.462038547914619e-06, + "loss": 0.4763, + "step": 1938 + }, + { + "epoch": 0.16523221133361737, + "grad_norm": 24.375, + "learning_rate": 2.4619974399754444e-06, + "loss": 1.3747, + "step": 1939 + }, + { + "epoch": 0.16531742650191733, + "grad_norm": 16.125, + "learning_rate": 2.4619563101342577e-06, + "loss": 1.0067, + "step": 1940 + }, + { + "epoch": 0.1654026416702173, + "grad_norm": 15.25, + "learning_rate": 2.4619151583918013e-06, + "loss": 1.0633, + "step": 1941 + }, + { + "epoch": 0.16548785683851724, + "grad_norm": 13.0, + "learning_rate": 2.4618739847488196e-06, + "loss": 0.5575, + "step": 1942 + }, + { + "epoch": 0.1655730720068172, + "grad_norm": 14.25, + "learning_rate": 2.461832789206056e-06, + "loss": 0.5893, + "step": 1943 + }, + { + "epoch": 0.16565828717511716, + "grad_norm": 12.9375, + "learning_rate": 2.461791571764256e-06, + "loss": 0.7454, + "step": 1944 + }, + { + "epoch": 0.16574350234341712, + "grad_norm": 20.5, + "learning_rate": 2.4617503324241633e-06, + "loss": 0.9566, + "step": 1945 + }, + { + "epoch": 0.16582871751171707, + "grad_norm": 12.5625, + "learning_rate": 2.4617090711865235e-06, + "loss": 0.7483, + "step": 1946 + }, + { + "epoch": 0.16591393268001706, + "grad_norm": 22.625, + "learning_rate": 2.4616677880520825e-06, + "loss": 0.6018, + "step": 1947 + }, + { + "epoch": 0.165999147848317, + "grad_norm": 13.5625, + "learning_rate": 2.461626483021586e-06, + "loss": 0.7312, + "step": 1948 + }, + { + "epoch": 0.16608436301661697, + "grad_norm": 17.5, + "learning_rate": 2.461585156095781e-06, + "loss": 0.7235, + "step": 1949 + }, + { + "epoch": 0.16616957818491693, + "grad_norm": 29.125, + "learning_rate": 2.4615438072754134e-06, + "loss": 1.2423, + "step": 1950 + }, + { + "epoch": 0.16625479335321688, + "grad_norm": 14.6875, + "learning_rate": 2.461502436561231e-06, + "loss": 0.7071, + "step": 1951 + }, + { + "epoch": 0.16634000852151684, + "grad_norm": 28.875, + "learning_rate": 2.4614610439539813e-06, + "loss": 0.7138, + "step": 1952 + }, + { + "epoch": 0.1664252236898168, + "grad_norm": 12.375, + "learning_rate": 2.4614196294544122e-06, + "loss": 0.4549, + "step": 1953 + }, + { + "epoch": 0.16651043885811675, + "grad_norm": 15.9375, + "learning_rate": 2.4613781930632723e-06, + "loss": 0.7831, + "step": 1954 + }, + { + "epoch": 0.1665956540264167, + "grad_norm": 22.125, + "learning_rate": 2.4613367347813105e-06, + "loss": 1.1, + "step": 1955 + }, + { + "epoch": 0.16668086919471667, + "grad_norm": 15.3125, + "learning_rate": 2.4612952546092763e-06, + "loss": 0.8827, + "step": 1956 + }, + { + "epoch": 0.16676608436301663, + "grad_norm": 13.9375, + "learning_rate": 2.461253752547918e-06, + "loss": 0.6588, + "step": 1957 + }, + { + "epoch": 0.16685129953131658, + "grad_norm": 16.625, + "learning_rate": 2.4612122285979863e-06, + "loss": 1.0942, + "step": 1958 + }, + { + "epoch": 0.16693651469961654, + "grad_norm": 24.75, + "learning_rate": 2.4611706827602324e-06, + "loss": 1.3617, + "step": 1959 + }, + { + "epoch": 0.1670217298679165, + "grad_norm": 14.1875, + "learning_rate": 2.4611291150354054e-06, + "loss": 0.5499, + "step": 1960 + }, + { + "epoch": 0.16710694503621645, + "grad_norm": 19.0, + "learning_rate": 2.461087525424258e-06, + "loss": 0.4473, + "step": 1961 + }, + { + "epoch": 0.1671921602045164, + "grad_norm": 16.875, + "learning_rate": 2.461045913927541e-06, + "loss": 0.92, + "step": 1962 + }, + { + "epoch": 0.16727737537281637, + "grad_norm": 20.75, + "learning_rate": 2.461004280546007e-06, + "loss": 0.6915, + "step": 1963 + }, + { + "epoch": 0.16736259054111632, + "grad_norm": 13.9375, + "learning_rate": 2.4609626252804074e-06, + "loss": 0.6387, + "step": 1964 + }, + { + "epoch": 0.16744780570941628, + "grad_norm": 13.75, + "learning_rate": 2.4609209481314957e-06, + "loss": 0.7349, + "step": 1965 + }, + { + "epoch": 0.16753302087771624, + "grad_norm": 17.0, + "learning_rate": 2.4608792491000245e-06, + "loss": 0.7464, + "step": 1966 + }, + { + "epoch": 0.1676182360460162, + "grad_norm": 14.6875, + "learning_rate": 2.4608375281867475e-06, + "loss": 0.6818, + "step": 1967 + }, + { + "epoch": 0.16770345121431615, + "grad_norm": 12.25, + "learning_rate": 2.460795785392419e-06, + "loss": 0.6621, + "step": 1968 + }, + { + "epoch": 0.1677886663826161, + "grad_norm": 18.25, + "learning_rate": 2.4607540207177934e-06, + "loss": 0.9225, + "step": 1969 + }, + { + "epoch": 0.16787388155091607, + "grad_norm": 14.3125, + "learning_rate": 2.460712234163625e-06, + "loss": 0.6546, + "step": 1970 + }, + { + "epoch": 0.16795909671921602, + "grad_norm": 16.375, + "learning_rate": 2.460670425730669e-06, + "loss": 0.8066, + "step": 1971 + }, + { + "epoch": 0.16804431188751598, + "grad_norm": 10.75, + "learning_rate": 2.4606285954196805e-06, + "loss": 0.5965, + "step": 1972 + }, + { + "epoch": 0.16812952705581594, + "grad_norm": 27.125, + "learning_rate": 2.4605867432314167e-06, + "loss": 0.9837, + "step": 1973 + }, + { + "epoch": 0.1682147422241159, + "grad_norm": 16.125, + "learning_rate": 2.460544869166633e-06, + "loss": 0.9392, + "step": 1974 + }, + { + "epoch": 0.16829995739241585, + "grad_norm": 15.125, + "learning_rate": 2.460502973226086e-06, + "loss": 0.9626, + "step": 1975 + }, + { + "epoch": 0.1683851725607158, + "grad_norm": 25.0, + "learning_rate": 2.4604610554105326e-06, + "loss": 1.2058, + "step": 1976 + }, + { + "epoch": 0.16847038772901576, + "grad_norm": 17.875, + "learning_rate": 2.460419115720731e-06, + "loss": 1.023, + "step": 1977 + }, + { + "epoch": 0.16855560289731572, + "grad_norm": 14.25, + "learning_rate": 2.460377154157439e-06, + "loss": 0.6717, + "step": 1978 + }, + { + "epoch": 0.16864081806561568, + "grad_norm": 13.0625, + "learning_rate": 2.4603351707214144e-06, + "loss": 0.6409, + "step": 1979 + }, + { + "epoch": 0.16872603323391563, + "grad_norm": 15.125, + "learning_rate": 2.4602931654134165e-06, + "loss": 0.959, + "step": 1980 + }, + { + "epoch": 0.1688112484022156, + "grad_norm": 18.25, + "learning_rate": 2.460251138234204e-06, + "loss": 1.0022, + "step": 1981 + }, + { + "epoch": 0.16889646357051555, + "grad_norm": 14.3125, + "learning_rate": 2.4602090891845363e-06, + "loss": 0.6976, + "step": 1982 + }, + { + "epoch": 0.1689816787388155, + "grad_norm": 13.0, + "learning_rate": 2.4601670182651736e-06, + "loss": 0.5451, + "step": 1983 + }, + { + "epoch": 0.16906689390711546, + "grad_norm": 16.875, + "learning_rate": 2.4601249254768762e-06, + "loss": 1.0195, + "step": 1984 + }, + { + "epoch": 0.16915210907541542, + "grad_norm": 19.0, + "learning_rate": 2.460082810820404e-06, + "loss": 0.8803, + "step": 1985 + }, + { + "epoch": 0.16923732424371538, + "grad_norm": 18.875, + "learning_rate": 2.460040674296519e-06, + "loss": 1.1858, + "step": 1986 + }, + { + "epoch": 0.16932253941201533, + "grad_norm": 18.375, + "learning_rate": 2.4599985159059817e-06, + "loss": 1.0269, + "step": 1987 + }, + { + "epoch": 0.1694077545803153, + "grad_norm": 11.5, + "learning_rate": 2.4599563356495547e-06, + "loss": 0.4623, + "step": 1988 + }, + { + "epoch": 0.16949296974861525, + "grad_norm": 12.875, + "learning_rate": 2.459914133528e-06, + "loss": 0.5695, + "step": 1989 + }, + { + "epoch": 0.1695781849169152, + "grad_norm": 13.6875, + "learning_rate": 2.4598719095420803e-06, + "loss": 0.7257, + "step": 1990 + }, + { + "epoch": 0.16966340008521516, + "grad_norm": 18.875, + "learning_rate": 2.4598296636925586e-06, + "loss": 0.8458, + "step": 1991 + }, + { + "epoch": 0.16974861525351512, + "grad_norm": 15.0625, + "learning_rate": 2.4597873959801982e-06, + "loss": 0.6678, + "step": 1992 + }, + { + "epoch": 0.16983383042181507, + "grad_norm": 13.6875, + "learning_rate": 2.459745106405763e-06, + "loss": 0.4443, + "step": 1993 + }, + { + "epoch": 0.16991904559011503, + "grad_norm": 14.25, + "learning_rate": 2.4597027949700176e-06, + "loss": 0.6835, + "step": 1994 + }, + { + "epoch": 0.170004260758415, + "grad_norm": 17.5, + "learning_rate": 2.4596604616737257e-06, + "loss": 0.7451, + "step": 1995 + }, + { + "epoch": 0.17008947592671494, + "grad_norm": 15.4375, + "learning_rate": 2.459618106517653e-06, + "loss": 0.6533, + "step": 1996 + }, + { + "epoch": 0.1701746910950149, + "grad_norm": 16.875, + "learning_rate": 2.459575729502565e-06, + "loss": 1.0351, + "step": 1997 + }, + { + "epoch": 0.17025990626331486, + "grad_norm": 21.875, + "learning_rate": 2.4595333306292272e-06, + "loss": 0.3285, + "step": 1998 + }, + { + "epoch": 0.17034512143161482, + "grad_norm": 12.5, + "learning_rate": 2.4594909098984058e-06, + "loss": 0.6, + "step": 1999 + }, + { + "epoch": 0.17043033659991477, + "grad_norm": 16.0, + "learning_rate": 2.459448467310867e-06, + "loss": 1.0138, + "step": 2000 + }, + { + "epoch": 0.17051555176821473, + "grad_norm": 11.375, + "learning_rate": 2.4594060028673788e-06, + "loss": 0.5304, + "step": 2001 + }, + { + "epoch": 0.1706007669365147, + "grad_norm": 18.75, + "learning_rate": 2.4593635165687076e-06, + "loss": 0.7781, + "step": 2002 + }, + { + "epoch": 0.17068598210481467, + "grad_norm": 22.625, + "learning_rate": 2.459321008415622e-06, + "loss": 1.0972, + "step": 2003 + }, + { + "epoch": 0.17077119727311463, + "grad_norm": 12.125, + "learning_rate": 2.459278478408889e-06, + "loss": 0.4955, + "step": 2004 + }, + { + "epoch": 0.17085641244141458, + "grad_norm": 11.25, + "learning_rate": 2.4592359265492786e-06, + "loss": 0.472, + "step": 2005 + }, + { + "epoch": 0.17094162760971454, + "grad_norm": 12.5625, + "learning_rate": 2.4591933528375586e-06, + "loss": 0.475, + "step": 2006 + }, + { + "epoch": 0.1710268427780145, + "grad_norm": 14.4375, + "learning_rate": 2.459150757274499e-06, + "loss": 0.9155, + "step": 2007 + }, + { + "epoch": 0.17111205794631446, + "grad_norm": 19.5, + "learning_rate": 2.459108139860869e-06, + "loss": 1.1182, + "step": 2008 + }, + { + "epoch": 0.1711972731146144, + "grad_norm": 13.0625, + "learning_rate": 2.459065500597439e-06, + "loss": 0.619, + "step": 2009 + }, + { + "epoch": 0.17128248828291437, + "grad_norm": 15.6875, + "learning_rate": 2.45902283948498e-06, + "loss": 0.9494, + "step": 2010 + }, + { + "epoch": 0.17136770345121433, + "grad_norm": 12.375, + "learning_rate": 2.4589801565242627e-06, + "loss": 0.5641, + "step": 2011 + }, + { + "epoch": 0.17145291861951428, + "grad_norm": 12.75, + "learning_rate": 2.458937451716058e-06, + "loss": 0.6059, + "step": 2012 + }, + { + "epoch": 0.17153813378781424, + "grad_norm": 14.375, + "learning_rate": 2.4588947250611377e-06, + "loss": 0.8325, + "step": 2013 + }, + { + "epoch": 0.1716233489561142, + "grad_norm": 12.25, + "learning_rate": 2.4588519765602743e-06, + "loss": 0.6977, + "step": 2014 + }, + { + "epoch": 0.17170856412441415, + "grad_norm": 15.625, + "learning_rate": 2.4588092062142394e-06, + "loss": 0.6614, + "step": 2015 + }, + { + "epoch": 0.1717937792927141, + "grad_norm": 18.125, + "learning_rate": 2.4587664140238076e-06, + "loss": 1.1969, + "step": 2016 + }, + { + "epoch": 0.17187899446101407, + "grad_norm": 22.125, + "learning_rate": 2.458723599989751e-06, + "loss": 0.9492, + "step": 2017 + }, + { + "epoch": 0.17196420962931402, + "grad_norm": 13.6875, + "learning_rate": 2.458680764112843e-06, + "loss": 0.6721, + "step": 2018 + }, + { + "epoch": 0.17204942479761398, + "grad_norm": 18.5, + "learning_rate": 2.458637906393858e-06, + "loss": 0.8085, + "step": 2019 + }, + { + "epoch": 0.17213463996591394, + "grad_norm": 11.0, + "learning_rate": 2.458595026833571e-06, + "loss": 0.4131, + "step": 2020 + }, + { + "epoch": 0.1722198551342139, + "grad_norm": 18.625, + "learning_rate": 2.4585521254327567e-06, + "loss": 0.8082, + "step": 2021 + }, + { + "epoch": 0.17230507030251385, + "grad_norm": 12.9375, + "learning_rate": 2.45850920219219e-06, + "loss": 0.8042, + "step": 2022 + }, + { + "epoch": 0.1723902854708138, + "grad_norm": 13.4375, + "learning_rate": 2.458466257112647e-06, + "loss": 0.7371, + "step": 2023 + }, + { + "epoch": 0.17247550063911377, + "grad_norm": 10.1875, + "learning_rate": 2.4584232901949036e-06, + "loss": 0.4148, + "step": 2024 + }, + { + "epoch": 0.17256071580741372, + "grad_norm": 13.5625, + "learning_rate": 2.4583803014397355e-06, + "loss": 0.7513, + "step": 2025 + }, + { + "epoch": 0.17264593097571368, + "grad_norm": 8.125, + "learning_rate": 2.458337290847921e-06, + "loss": 0.2694, + "step": 2026 + }, + { + "epoch": 0.17273114614401364, + "grad_norm": 12.6875, + "learning_rate": 2.4582942584202363e-06, + "loss": 1.0149, + "step": 2027 + }, + { + "epoch": 0.1728163613123136, + "grad_norm": 17.5, + "learning_rate": 2.458251204157459e-06, + "loss": 0.9005, + "step": 2028 + }, + { + "epoch": 0.17290157648061355, + "grad_norm": 12.5625, + "learning_rate": 2.458208128060368e-06, + "loss": 0.6792, + "step": 2029 + }, + { + "epoch": 0.1729867916489135, + "grad_norm": 16.375, + "learning_rate": 2.458165030129741e-06, + "loss": 0.6161, + "step": 2030 + }, + { + "epoch": 0.17307200681721346, + "grad_norm": 15.1875, + "learning_rate": 2.458121910366357e-06, + "loss": 0.4292, + "step": 2031 + }, + { + "epoch": 0.17315722198551342, + "grad_norm": 18.125, + "learning_rate": 2.4580787687709947e-06, + "loss": 1.0835, + "step": 2032 + }, + { + "epoch": 0.17324243715381338, + "grad_norm": 12.875, + "learning_rate": 2.458035605344435e-06, + "loss": 0.7919, + "step": 2033 + }, + { + "epoch": 0.17332765232211333, + "grad_norm": 15.5, + "learning_rate": 2.4579924200874566e-06, + "loss": 0.7203, + "step": 2034 + }, + { + "epoch": 0.1734128674904133, + "grad_norm": 22.0, + "learning_rate": 2.4579492130008406e-06, + "loss": 1.0555, + "step": 2035 + }, + { + "epoch": 0.17349808265871325, + "grad_norm": 16.25, + "learning_rate": 2.4579059840853677e-06, + "loss": 0.8128, + "step": 2036 + }, + { + "epoch": 0.1735832978270132, + "grad_norm": 31.375, + "learning_rate": 2.457862733341819e-06, + "loss": 0.9135, + "step": 2037 + }, + { + "epoch": 0.17366851299531316, + "grad_norm": 14.125, + "learning_rate": 2.457819460770976e-06, + "loss": 0.713, + "step": 2038 + }, + { + "epoch": 0.17375372816361312, + "grad_norm": 13.0625, + "learning_rate": 2.457776166373621e-06, + "loss": 0.6362, + "step": 2039 + }, + { + "epoch": 0.17383894333191308, + "grad_norm": 16.5, + "learning_rate": 2.457732850150536e-06, + "loss": 1.0217, + "step": 2040 + }, + { + "epoch": 0.17392415850021303, + "grad_norm": 17.625, + "learning_rate": 2.457689512102504e-06, + "loss": 0.6053, + "step": 2041 + }, + { + "epoch": 0.174009373668513, + "grad_norm": 14.75, + "learning_rate": 2.457646152230308e-06, + "loss": 0.8687, + "step": 2042 + }, + { + "epoch": 0.17409458883681295, + "grad_norm": 12.25, + "learning_rate": 2.4576027705347316e-06, + "loss": 0.5309, + "step": 2043 + }, + { + "epoch": 0.1741798040051129, + "grad_norm": 21.75, + "learning_rate": 2.457559367016559e-06, + "loss": 0.903, + "step": 2044 + }, + { + "epoch": 0.17426501917341286, + "grad_norm": 15.0, + "learning_rate": 2.4575159416765742e-06, + "loss": 0.6945, + "step": 2045 + }, + { + "epoch": 0.17435023434171282, + "grad_norm": 17.5, + "learning_rate": 2.4574724945155624e-06, + "loss": 0.9111, + "step": 2046 + }, + { + "epoch": 0.17443544951001277, + "grad_norm": 14.875, + "learning_rate": 2.457429025534308e-06, + "loss": 0.8245, + "step": 2047 + }, + { + "epoch": 0.17452066467831273, + "grad_norm": 25.125, + "learning_rate": 2.4573855347335972e-06, + "loss": 1.5115, + "step": 2048 + }, + { + "epoch": 0.1746058798466127, + "grad_norm": 15.9375, + "learning_rate": 2.4573420221142155e-06, + "loss": 0.8505, + "step": 2049 + }, + { + "epoch": 0.17469109501491265, + "grad_norm": 17.0, + "learning_rate": 2.4572984876769495e-06, + "loss": 0.8712, + "step": 2050 + }, + { + "epoch": 0.1747763101832126, + "grad_norm": 13.9375, + "learning_rate": 2.457254931422586e-06, + "loss": 0.6953, + "step": 2051 + }, + { + "epoch": 0.17486152535151256, + "grad_norm": 15.875, + "learning_rate": 2.4572113533519116e-06, + "loss": 0.5377, + "step": 2052 + }, + { + "epoch": 0.17494674051981252, + "grad_norm": 13.1875, + "learning_rate": 2.457167753465714e-06, + "loss": 0.7744, + "step": 2053 + }, + { + "epoch": 0.17503195568811247, + "grad_norm": 21.5, + "learning_rate": 2.457124131764781e-06, + "loss": 0.846, + "step": 2054 + }, + { + "epoch": 0.17511717085641243, + "grad_norm": 14.875, + "learning_rate": 2.4570804882499015e-06, + "loss": 0.9177, + "step": 2055 + }, + { + "epoch": 0.1752023860247124, + "grad_norm": 12.125, + "learning_rate": 2.457036822921864e-06, + "loss": 0.802, + "step": 2056 + }, + { + "epoch": 0.17528760119301234, + "grad_norm": 12.6875, + "learning_rate": 2.456993135781457e-06, + "loss": 0.5264, + "step": 2057 + }, + { + "epoch": 0.1753728163613123, + "grad_norm": 12.25, + "learning_rate": 2.4569494268294702e-06, + "loss": 0.5958, + "step": 2058 + }, + { + "epoch": 0.17545803152961226, + "grad_norm": 19.5, + "learning_rate": 2.4569056960666937e-06, + "loss": 1.366, + "step": 2059 + }, + { + "epoch": 0.17554324669791224, + "grad_norm": 12.3125, + "learning_rate": 2.456861943493918e-06, + "loss": 0.7018, + "step": 2060 + }, + { + "epoch": 0.1756284618662122, + "grad_norm": 11.8125, + "learning_rate": 2.456818169111933e-06, + "loss": 0.7075, + "step": 2061 + }, + { + "epoch": 0.17571367703451216, + "grad_norm": 16.25, + "learning_rate": 2.45677437292153e-06, + "loss": 0.8157, + "step": 2062 + }, + { + "epoch": 0.1757988922028121, + "grad_norm": 18.75, + "learning_rate": 2.4567305549235008e-06, + "loss": 1.0475, + "step": 2063 + }, + { + "epoch": 0.17588410737111207, + "grad_norm": 10.5625, + "learning_rate": 2.456686715118637e-06, + "loss": 0.327, + "step": 2064 + }, + { + "epoch": 0.17596932253941203, + "grad_norm": 19.0, + "learning_rate": 2.456642853507731e-06, + "loss": 1.1008, + "step": 2065 + }, + { + "epoch": 0.17605453770771198, + "grad_norm": 14.5, + "learning_rate": 2.4565989700915746e-06, + "loss": 1.0194, + "step": 2066 + }, + { + "epoch": 0.17613975287601194, + "grad_norm": 26.0, + "learning_rate": 2.456555064870962e-06, + "loss": 0.8765, + "step": 2067 + }, + { + "epoch": 0.1762249680443119, + "grad_norm": 13.5625, + "learning_rate": 2.456511137846686e-06, + "loss": 0.9368, + "step": 2068 + }, + { + "epoch": 0.17631018321261185, + "grad_norm": 16.5, + "learning_rate": 2.4564671890195406e-06, + "loss": 0.9802, + "step": 2069 + }, + { + "epoch": 0.1763953983809118, + "grad_norm": 17.125, + "learning_rate": 2.4564232183903196e-06, + "loss": 1.0024, + "step": 2070 + }, + { + "epoch": 0.17648061354921177, + "grad_norm": 15.125, + "learning_rate": 2.4563792259598186e-06, + "loss": 0.9557, + "step": 2071 + }, + { + "epoch": 0.17656582871751172, + "grad_norm": 17.25, + "learning_rate": 2.4563352117288312e-06, + "loss": 0.5583, + "step": 2072 + }, + { + "epoch": 0.17665104388581168, + "grad_norm": 14.75, + "learning_rate": 2.4562911756981537e-06, + "loss": 1.0684, + "step": 2073 + }, + { + "epoch": 0.17673625905411164, + "grad_norm": 13.625, + "learning_rate": 2.4562471178685815e-06, + "loss": 0.9078, + "step": 2074 + }, + { + "epoch": 0.1768214742224116, + "grad_norm": 14.9375, + "learning_rate": 2.456203038240911e-06, + "loss": 0.9359, + "step": 2075 + }, + { + "epoch": 0.17690668939071155, + "grad_norm": 19.75, + "learning_rate": 2.4561589368159388e-06, + "loss": 1.1046, + "step": 2076 + }, + { + "epoch": 0.1769919045590115, + "grad_norm": 15.9375, + "learning_rate": 2.4561148135944616e-06, + "loss": 0.8891, + "step": 2077 + }, + { + "epoch": 0.17707711972731147, + "grad_norm": 28.25, + "learning_rate": 2.4560706685772767e-06, + "loss": 0.8514, + "step": 2078 + }, + { + "epoch": 0.17716233489561142, + "grad_norm": 28.0, + "learning_rate": 2.4560265017651826e-06, + "loss": 0.5024, + "step": 2079 + }, + { + "epoch": 0.17724755006391138, + "grad_norm": 18.5, + "learning_rate": 2.455982313158976e-06, + "loss": 1.2425, + "step": 2080 + }, + { + "epoch": 0.17733276523221134, + "grad_norm": 13.4375, + "learning_rate": 2.455938102759457e-06, + "loss": 0.7255, + "step": 2081 + }, + { + "epoch": 0.1774179804005113, + "grad_norm": 14.625, + "learning_rate": 2.4558938705674236e-06, + "loss": 1.0185, + "step": 2082 + }, + { + "epoch": 0.17750319556881125, + "grad_norm": 16.25, + "learning_rate": 2.455849616583675e-06, + "loss": 0.8518, + "step": 2083 + }, + { + "epoch": 0.1775884107371112, + "grad_norm": 11.25, + "learning_rate": 2.4558053408090115e-06, + "loss": 0.9188, + "step": 2084 + }, + { + "epoch": 0.17767362590541116, + "grad_norm": 12.75, + "learning_rate": 2.455761043244233e-06, + "loss": 0.8994, + "step": 2085 + }, + { + "epoch": 0.17775884107371112, + "grad_norm": 13.25, + "learning_rate": 2.4557167238901404e-06, + "loss": 0.6106, + "step": 2086 + }, + { + "epoch": 0.17784405624201108, + "grad_norm": 20.875, + "learning_rate": 2.4556723827475333e-06, + "loss": 0.9989, + "step": 2087 + }, + { + "epoch": 0.17792927141031104, + "grad_norm": 14.625, + "learning_rate": 2.4556280198172144e-06, + "loss": 0.6338, + "step": 2088 + }, + { + "epoch": 0.178014486578611, + "grad_norm": 15.75, + "learning_rate": 2.455583635099985e-06, + "loss": 0.8638, + "step": 2089 + }, + { + "epoch": 0.17809970174691095, + "grad_norm": 15.4375, + "learning_rate": 2.4555392285966467e-06, + "loss": 0.8543, + "step": 2090 + }, + { + "epoch": 0.1781849169152109, + "grad_norm": 10.0625, + "learning_rate": 2.455494800308002e-06, + "loss": 0.3567, + "step": 2091 + }, + { + "epoch": 0.17827013208351086, + "grad_norm": 14.25, + "learning_rate": 2.4554503502348544e-06, + "loss": 0.5069, + "step": 2092 + }, + { + "epoch": 0.17835534725181082, + "grad_norm": 16.875, + "learning_rate": 2.4554058783780064e-06, + "loss": 0.6473, + "step": 2093 + }, + { + "epoch": 0.17844056242011078, + "grad_norm": 14.1875, + "learning_rate": 2.455361384738263e-06, + "loss": 0.735, + "step": 2094 + }, + { + "epoch": 0.17852577758841073, + "grad_norm": 23.625, + "learning_rate": 2.4553168693164262e-06, + "loss": 1.4259, + "step": 2095 + }, + { + "epoch": 0.1786109927567107, + "grad_norm": 12.625, + "learning_rate": 2.455272332113302e-06, + "loss": 0.6685, + "step": 2096 + }, + { + "epoch": 0.17869620792501065, + "grad_norm": 15.5, + "learning_rate": 2.4552277731296952e-06, + "loss": 0.7064, + "step": 2097 + }, + { + "epoch": 0.1787814230933106, + "grad_norm": 18.375, + "learning_rate": 2.4551831923664098e-06, + "loss": 0.8756, + "step": 2098 + }, + { + "epoch": 0.17886663826161056, + "grad_norm": 19.0, + "learning_rate": 2.4551385898242525e-06, + "loss": 1.028, + "step": 2099 + }, + { + "epoch": 0.17895185342991052, + "grad_norm": 15.9375, + "learning_rate": 2.455093965504029e-06, + "loss": 0.7538, + "step": 2100 + }, + { + "epoch": 0.17903706859821047, + "grad_norm": 18.375, + "learning_rate": 2.4550493194065462e-06, + "loss": 0.9061, + "step": 2101 + }, + { + "epoch": 0.17912228376651043, + "grad_norm": 19.375, + "learning_rate": 2.45500465153261e-06, + "loss": 0.9957, + "step": 2102 + }, + { + "epoch": 0.1792074989348104, + "grad_norm": 16.375, + "learning_rate": 2.454959961883028e-06, + "loss": 0.7387, + "step": 2103 + }, + { + "epoch": 0.17929271410311035, + "grad_norm": 12.875, + "learning_rate": 2.454915250458608e-06, + "loss": 0.6402, + "step": 2104 + }, + { + "epoch": 0.1793779292714103, + "grad_norm": 11.5, + "learning_rate": 2.4548705172601574e-06, + "loss": 0.4893, + "step": 2105 + }, + { + "epoch": 0.17946314443971026, + "grad_norm": 13.0, + "learning_rate": 2.4548257622884853e-06, + "loss": 0.6715, + "step": 2106 + }, + { + "epoch": 0.17954835960801022, + "grad_norm": 13.5625, + "learning_rate": 2.4547809855444e-06, + "loss": 0.9421, + "step": 2107 + }, + { + "epoch": 0.17963357477631017, + "grad_norm": 13.875, + "learning_rate": 2.454736187028711e-06, + "loss": 0.6871, + "step": 2108 + }, + { + "epoch": 0.17971878994461013, + "grad_norm": 13.125, + "learning_rate": 2.4546913667422272e-06, + "loss": 0.6145, + "step": 2109 + }, + { + "epoch": 0.1798040051129101, + "grad_norm": 15.1875, + "learning_rate": 2.4546465246857595e-06, + "loss": 0.6275, + "step": 2110 + }, + { + "epoch": 0.17988922028121004, + "grad_norm": 17.375, + "learning_rate": 2.4546016608601173e-06, + "loss": 0.9303, + "step": 2111 + }, + { + "epoch": 0.17997443544951, + "grad_norm": 13.125, + "learning_rate": 2.454556775266112e-06, + "loss": 0.6837, + "step": 2112 + }, + { + "epoch": 0.18005965061780996, + "grad_norm": 17.25, + "learning_rate": 2.454511867904555e-06, + "loss": 0.9435, + "step": 2113 + }, + { + "epoch": 0.18014486578610991, + "grad_norm": 15.5, + "learning_rate": 2.4544669387762567e-06, + "loss": 0.7194, + "step": 2114 + }, + { + "epoch": 0.18023008095440987, + "grad_norm": 16.75, + "learning_rate": 2.45442198788203e-06, + "loss": 1.0713, + "step": 2115 + }, + { + "epoch": 0.18031529612270986, + "grad_norm": 24.375, + "learning_rate": 2.4543770152226857e-06, + "loss": 1.5671, + "step": 2116 + }, + { + "epoch": 0.1804005112910098, + "grad_norm": 21.625, + "learning_rate": 2.454332020799039e-06, + "loss": 0.9992, + "step": 2117 + }, + { + "epoch": 0.18048572645930977, + "grad_norm": 13.5, + "learning_rate": 2.454287004611901e-06, + "loss": 0.7161, + "step": 2118 + }, + { + "epoch": 0.18057094162760973, + "grad_norm": 15.4375, + "learning_rate": 2.454241966662086e-06, + "loss": 0.756, + "step": 2119 + }, + { + "epoch": 0.18065615679590968, + "grad_norm": 12.625, + "learning_rate": 2.4541969069504075e-06, + "loss": 0.7193, + "step": 2120 + }, + { + "epoch": 0.18074137196420964, + "grad_norm": 11.4375, + "learning_rate": 2.45415182547768e-06, + "loss": 0.5944, + "step": 2121 + }, + { + "epoch": 0.1808265871325096, + "grad_norm": 18.75, + "learning_rate": 2.4541067222447184e-06, + "loss": 1.2357, + "step": 2122 + }, + { + "epoch": 0.18091180230080955, + "grad_norm": 14.5, + "learning_rate": 2.4540615972523373e-06, + "loss": 0.3164, + "step": 2123 + }, + { + "epoch": 0.1809970174691095, + "grad_norm": 13.25, + "learning_rate": 2.4540164505013523e-06, + "loss": 0.6552, + "step": 2124 + }, + { + "epoch": 0.18108223263740947, + "grad_norm": 19.75, + "learning_rate": 2.45397128199258e-06, + "loss": 1.3354, + "step": 2125 + }, + { + "epoch": 0.18116744780570943, + "grad_norm": 12.625, + "learning_rate": 2.4539260917268354e-06, + "loss": 0.708, + "step": 2126 + }, + { + "epoch": 0.18125266297400938, + "grad_norm": 15.3125, + "learning_rate": 2.4538808797049357e-06, + "loss": 0.7667, + "step": 2127 + }, + { + "epoch": 0.18133787814230934, + "grad_norm": 16.0, + "learning_rate": 2.4538356459276977e-06, + "loss": 0.9075, + "step": 2128 + }, + { + "epoch": 0.1814230933106093, + "grad_norm": 13.6875, + "learning_rate": 2.4537903903959394e-06, + "loss": 0.7951, + "step": 2129 + }, + { + "epoch": 0.18150830847890925, + "grad_norm": 14.125, + "learning_rate": 2.4537451131104776e-06, + "loss": 0.8273, + "step": 2130 + }, + { + "epoch": 0.1815935236472092, + "grad_norm": 14.75, + "learning_rate": 2.453699814072132e-06, + "loss": 0.7433, + "step": 2131 + }, + { + "epoch": 0.18167873881550917, + "grad_norm": 63.5, + "learning_rate": 2.45365449328172e-06, + "loss": 1.5492, + "step": 2132 + }, + { + "epoch": 0.18176395398380912, + "grad_norm": 16.0, + "learning_rate": 2.4536091507400608e-06, + "loss": 0.9812, + "step": 2133 + }, + { + "epoch": 0.18184916915210908, + "grad_norm": 14.125, + "learning_rate": 2.453563786447974e-06, + "loss": 0.7726, + "step": 2134 + }, + { + "epoch": 0.18193438432040904, + "grad_norm": 15.5, + "learning_rate": 2.4535184004062795e-06, + "loss": 1.0586, + "step": 2135 + }, + { + "epoch": 0.182019599488709, + "grad_norm": 14.8125, + "learning_rate": 2.453472992615797e-06, + "loss": 0.5624, + "step": 2136 + }, + { + "epoch": 0.18210481465700895, + "grad_norm": 13.0, + "learning_rate": 2.4534275630773475e-06, + "loss": 0.5086, + "step": 2137 + }, + { + "epoch": 0.1821900298253089, + "grad_norm": 12.75, + "learning_rate": 2.453382111791752e-06, + "loss": 0.5329, + "step": 2138 + }, + { + "epoch": 0.18227524499360886, + "grad_norm": 13.875, + "learning_rate": 2.453336638759831e-06, + "loss": 0.6857, + "step": 2139 + }, + { + "epoch": 0.18236046016190882, + "grad_norm": 13.9375, + "learning_rate": 2.453291143982407e-06, + "loss": 0.8337, + "step": 2140 + }, + { + "epoch": 0.18244567533020878, + "grad_norm": 21.5, + "learning_rate": 2.4532456274603024e-06, + "loss": 1.211, + "step": 2141 + }, + { + "epoch": 0.18253089049850874, + "grad_norm": 14.5625, + "learning_rate": 2.4532000891943393e-06, + "loss": 0.8509, + "step": 2142 + }, + { + "epoch": 0.1826161056668087, + "grad_norm": 17.25, + "learning_rate": 2.4531545291853407e-06, + "loss": 0.831, + "step": 2143 + }, + { + "epoch": 0.18270132083510865, + "grad_norm": 12.0625, + "learning_rate": 2.4531089474341296e-06, + "loss": 0.7519, + "step": 2144 + }, + { + "epoch": 0.1827865360034086, + "grad_norm": 9.5, + "learning_rate": 2.45306334394153e-06, + "loss": 0.2772, + "step": 2145 + }, + { + "epoch": 0.18287175117170856, + "grad_norm": 15.75, + "learning_rate": 2.4530177187083666e-06, + "loss": 0.5885, + "step": 2146 + }, + { + "epoch": 0.18295696634000852, + "grad_norm": 15.9375, + "learning_rate": 2.4529720717354627e-06, + "loss": 1.0621, + "step": 2147 + }, + { + "epoch": 0.18304218150830848, + "grad_norm": 17.375, + "learning_rate": 2.4529264030236445e-06, + "loss": 0.9031, + "step": 2148 + }, + { + "epoch": 0.18312739667660843, + "grad_norm": 16.25, + "learning_rate": 2.452880712573736e-06, + "loss": 0.8009, + "step": 2149 + }, + { + "epoch": 0.1832126118449084, + "grad_norm": 13.5625, + "learning_rate": 2.452835000386563e-06, + "loss": 0.7724, + "step": 2150 + }, + { + "epoch": 0.18329782701320835, + "grad_norm": 21.75, + "learning_rate": 2.4527892664629525e-06, + "loss": 0.8147, + "step": 2151 + }, + { + "epoch": 0.1833830421815083, + "grad_norm": 24.75, + "learning_rate": 2.452743510803731e-06, + "loss": 0.7664, + "step": 2152 + }, + { + "epoch": 0.18346825734980826, + "grad_norm": 20.75, + "learning_rate": 2.452697733409724e-06, + "loss": 1.2781, + "step": 2153 + }, + { + "epoch": 0.18355347251810822, + "grad_norm": 20.625, + "learning_rate": 2.4526519342817593e-06, + "loss": 1.1239, + "step": 2154 + }, + { + "epoch": 0.18363868768640818, + "grad_norm": 14.1875, + "learning_rate": 2.452606113420666e-06, + "loss": 0.7286, + "step": 2155 + }, + { + "epoch": 0.18372390285470813, + "grad_norm": 19.75, + "learning_rate": 2.45256027082727e-06, + "loss": 0.8668, + "step": 2156 + }, + { + "epoch": 0.1838091180230081, + "grad_norm": 17.5, + "learning_rate": 2.452514406502401e-06, + "loss": 0.8513, + "step": 2157 + }, + { + "epoch": 0.18389433319130805, + "grad_norm": 13.8125, + "learning_rate": 2.452468520446887e-06, + "loss": 0.9288, + "step": 2158 + }, + { + "epoch": 0.183979548359608, + "grad_norm": 28.125, + "learning_rate": 2.452422612661558e-06, + "loss": 1.1164, + "step": 2159 + }, + { + "epoch": 0.18406476352790796, + "grad_norm": 14.9375, + "learning_rate": 2.452376683147243e-06, + "loss": 0.846, + "step": 2160 + }, + { + "epoch": 0.18414997869620792, + "grad_norm": 13.5625, + "learning_rate": 2.4523307319047726e-06, + "loss": 0.562, + "step": 2161 + }, + { + "epoch": 0.18423519386450787, + "grad_norm": 10.4375, + "learning_rate": 2.4522847589349768e-06, + "loss": 0.561, + "step": 2162 + }, + { + "epoch": 0.18432040903280783, + "grad_norm": 14.3125, + "learning_rate": 2.4522387642386863e-06, + "loss": 0.7294, + "step": 2163 + }, + { + "epoch": 0.1844056242011078, + "grad_norm": 22.375, + "learning_rate": 2.4521927478167325e-06, + "loss": 1.2182, + "step": 2164 + }, + { + "epoch": 0.18449083936940774, + "grad_norm": 12.125, + "learning_rate": 2.4521467096699464e-06, + "loss": 0.5937, + "step": 2165 + }, + { + "epoch": 0.1845760545377077, + "grad_norm": 13.625, + "learning_rate": 2.452100649799161e-06, + "loss": 0.7398, + "step": 2166 + }, + { + "epoch": 0.18466126970600766, + "grad_norm": 30.75, + "learning_rate": 2.452054568205208e-06, + "loss": 0.9536, + "step": 2167 + }, + { + "epoch": 0.18474648487430761, + "grad_norm": 11.25, + "learning_rate": 2.4520084648889198e-06, + "loss": 0.4099, + "step": 2168 + }, + { + "epoch": 0.18483170004260757, + "grad_norm": 13.0, + "learning_rate": 2.45196233985113e-06, + "loss": 0.5028, + "step": 2169 + }, + { + "epoch": 0.18491691521090753, + "grad_norm": 14.75, + "learning_rate": 2.451916193092672e-06, + "loss": 0.9825, + "step": 2170 + }, + { + "epoch": 0.18500213037920749, + "grad_norm": 11.625, + "learning_rate": 2.45187002461438e-06, + "loss": 0.7238, + "step": 2171 + }, + { + "epoch": 0.18508734554750744, + "grad_norm": 13.0, + "learning_rate": 2.451823834417088e-06, + "loss": 0.6578, + "step": 2172 + }, + { + "epoch": 0.18517256071580743, + "grad_norm": 11.5625, + "learning_rate": 2.451777622501631e-06, + "loss": 0.5017, + "step": 2173 + }, + { + "epoch": 0.18525777588410738, + "grad_norm": 11.625, + "learning_rate": 2.4517313888688437e-06, + "loss": 0.5247, + "step": 2174 + }, + { + "epoch": 0.18534299105240734, + "grad_norm": 20.5, + "learning_rate": 2.451685133519562e-06, + "loss": 1.0488, + "step": 2175 + }, + { + "epoch": 0.1854282062207073, + "grad_norm": 15.625, + "learning_rate": 2.451638856454621e-06, + "loss": 0.972, + "step": 2176 + }, + { + "epoch": 0.18551342138900725, + "grad_norm": 14.875, + "learning_rate": 2.4515925576748578e-06, + "loss": 0.6616, + "step": 2177 + }, + { + "epoch": 0.1855986365573072, + "grad_norm": 13.25, + "learning_rate": 2.451546237181109e-06, + "loss": 0.6429, + "step": 2178 + }, + { + "epoch": 0.18568385172560717, + "grad_norm": 20.125, + "learning_rate": 2.4514998949742115e-06, + "loss": 1.1132, + "step": 2179 + }, + { + "epoch": 0.18576906689390713, + "grad_norm": 17.25, + "learning_rate": 2.4514535310550023e-06, + "loss": 1.0648, + "step": 2180 + }, + { + "epoch": 0.18585428206220708, + "grad_norm": 13.5, + "learning_rate": 2.45140714542432e-06, + "loss": 1.0407, + "step": 2181 + }, + { + "epoch": 0.18593949723050704, + "grad_norm": 20.5, + "learning_rate": 2.451360738083002e-06, + "loss": 0.9394, + "step": 2182 + }, + { + "epoch": 0.186024712398807, + "grad_norm": 19.875, + "learning_rate": 2.4513143090318877e-06, + "loss": 1.2023, + "step": 2183 + }, + { + "epoch": 0.18610992756710695, + "grad_norm": 15.6875, + "learning_rate": 2.4512678582718156e-06, + "loss": 0.638, + "step": 2184 + }, + { + "epoch": 0.1861951427354069, + "grad_norm": 14.5625, + "learning_rate": 2.4512213858036258e-06, + "loss": 0.9932, + "step": 2185 + }, + { + "epoch": 0.18628035790370687, + "grad_norm": 16.875, + "learning_rate": 2.4511748916281575e-06, + "loss": 0.5861, + "step": 2186 + }, + { + "epoch": 0.18636557307200682, + "grad_norm": 12.625, + "learning_rate": 2.4511283757462507e-06, + "loss": 0.57, + "step": 2187 + }, + { + "epoch": 0.18645078824030678, + "grad_norm": 14.875, + "learning_rate": 2.4510818381587464e-06, + "loss": 0.8754, + "step": 2188 + }, + { + "epoch": 0.18653600340860674, + "grad_norm": 16.5, + "learning_rate": 2.451035278866486e-06, + "loss": 0.8999, + "step": 2189 + }, + { + "epoch": 0.1866212185769067, + "grad_norm": 18.25, + "learning_rate": 2.4509886978703097e-06, + "loss": 0.9735, + "step": 2190 + }, + { + "epoch": 0.18670643374520665, + "grad_norm": 13.3125, + "learning_rate": 2.4509420951710605e-06, + "loss": 0.6872, + "step": 2191 + }, + { + "epoch": 0.1867916489135066, + "grad_norm": 9.625, + "learning_rate": 2.45089547076958e-06, + "loss": 0.3071, + "step": 2192 + }, + { + "epoch": 0.18687686408180657, + "grad_norm": 14.125, + "learning_rate": 2.45084882466671e-06, + "loss": 0.5016, + "step": 2193 + }, + { + "epoch": 0.18696207925010652, + "grad_norm": 15.1875, + "learning_rate": 2.4508021568632952e-06, + "loss": 0.5695, + "step": 2194 + }, + { + "epoch": 0.18704729441840648, + "grad_norm": 17.0, + "learning_rate": 2.450755467360177e-06, + "loss": 0.7118, + "step": 2195 + }, + { + "epoch": 0.18713250958670644, + "grad_norm": 12.625, + "learning_rate": 2.4507087561582004e-06, + "loss": 0.5187, + "step": 2196 + }, + { + "epoch": 0.1872177247550064, + "grad_norm": 19.75, + "learning_rate": 2.4506620232582096e-06, + "loss": 0.855, + "step": 2197 + }, + { + "epoch": 0.18730293992330635, + "grad_norm": 21.875, + "learning_rate": 2.450615268661048e-06, + "loss": 0.8528, + "step": 2198 + }, + { + "epoch": 0.1873881550916063, + "grad_norm": 20.25, + "learning_rate": 2.4505684923675614e-06, + "loss": 1.024, + "step": 2199 + }, + { + "epoch": 0.18747337025990626, + "grad_norm": 14.6875, + "learning_rate": 2.450521694378595e-06, + "loss": 0.6702, + "step": 2200 + }, + { + "epoch": 0.18755858542820622, + "grad_norm": 13.9375, + "learning_rate": 2.4504748746949946e-06, + "loss": 0.6394, + "step": 2201 + }, + { + "epoch": 0.18764380059650618, + "grad_norm": 14.75, + "learning_rate": 2.4504280333176054e-06, + "loss": 0.8527, + "step": 2202 + }, + { + "epoch": 0.18772901576480613, + "grad_norm": 13.1875, + "learning_rate": 2.450381170247275e-06, + "loss": 0.6188, + "step": 2203 + }, + { + "epoch": 0.1878142309331061, + "grad_norm": 18.125, + "learning_rate": 2.45033428548485e-06, + "loss": 0.4409, + "step": 2204 + }, + { + "epoch": 0.18789944610140605, + "grad_norm": 17.375, + "learning_rate": 2.450287379031177e-06, + "loss": 1.0834, + "step": 2205 + }, + { + "epoch": 0.187984661269706, + "grad_norm": 15.0, + "learning_rate": 2.4502404508871042e-06, + "loss": 0.7946, + "step": 2206 + }, + { + "epoch": 0.18806987643800596, + "grad_norm": 14.0625, + "learning_rate": 2.4501935010534794e-06, + "loss": 0.8715, + "step": 2207 + }, + { + "epoch": 0.18815509160630592, + "grad_norm": 12.25, + "learning_rate": 2.4501465295311512e-06, + "loss": 0.5447, + "step": 2208 + }, + { + "epoch": 0.18824030677460588, + "grad_norm": 19.75, + "learning_rate": 2.4500995363209684e-06, + "loss": 1.1739, + "step": 2209 + }, + { + "epoch": 0.18832552194290583, + "grad_norm": 12.75, + "learning_rate": 2.4500525214237804e-06, + "loss": 0.6727, + "step": 2210 + }, + { + "epoch": 0.1884107371112058, + "grad_norm": 13.9375, + "learning_rate": 2.4500054848404365e-06, + "loss": 0.6016, + "step": 2211 + }, + { + "epoch": 0.18849595227950575, + "grad_norm": 14.875, + "learning_rate": 2.4499584265717866e-06, + "loss": 0.8378, + "step": 2212 + }, + { + "epoch": 0.1885811674478057, + "grad_norm": 17.625, + "learning_rate": 2.4499113466186812e-06, + "loss": 0.6287, + "step": 2213 + }, + { + "epoch": 0.18866638261610566, + "grad_norm": 16.75, + "learning_rate": 2.4498642449819716e-06, + "loss": 0.9848, + "step": 2214 + }, + { + "epoch": 0.18875159778440562, + "grad_norm": 15.875, + "learning_rate": 2.4498171216625085e-06, + "loss": 1.0045, + "step": 2215 + }, + { + "epoch": 0.18883681295270557, + "grad_norm": 16.25, + "learning_rate": 2.449769976661143e-06, + "loss": 0.9238, + "step": 2216 + }, + { + "epoch": 0.18892202812100553, + "grad_norm": 15.75, + "learning_rate": 2.449722809978728e-06, + "loss": 1.0048, + "step": 2217 + }, + { + "epoch": 0.1890072432893055, + "grad_norm": 13.6875, + "learning_rate": 2.449675621616115e-06, + "loss": 0.5783, + "step": 2218 + }, + { + "epoch": 0.18909245845760544, + "grad_norm": 12.5, + "learning_rate": 2.4496284115741575e-06, + "loss": 0.3754, + "step": 2219 + }, + { + "epoch": 0.1891776736259054, + "grad_norm": 14.1875, + "learning_rate": 2.4495811798537082e-06, + "loss": 0.916, + "step": 2220 + }, + { + "epoch": 0.18926288879420536, + "grad_norm": 17.0, + "learning_rate": 2.4495339264556208e-06, + "loss": 0.7974, + "step": 2221 + }, + { + "epoch": 0.18934810396250532, + "grad_norm": 13.6875, + "learning_rate": 2.449486651380749e-06, + "loss": 0.761, + "step": 2222 + }, + { + "epoch": 0.18943331913080527, + "grad_norm": 15.3125, + "learning_rate": 2.4494393546299473e-06, + "loss": 1.0153, + "step": 2223 + }, + { + "epoch": 0.18951853429910523, + "grad_norm": 14.0, + "learning_rate": 2.4493920362040703e-06, + "loss": 0.6706, + "step": 2224 + }, + { + "epoch": 0.18960374946740519, + "grad_norm": 13.1875, + "learning_rate": 2.449344696103973e-06, + "loss": 0.7151, + "step": 2225 + }, + { + "epoch": 0.18968896463570514, + "grad_norm": 12.75, + "learning_rate": 2.449297334330511e-06, + "loss": 0.4276, + "step": 2226 + }, + { + "epoch": 0.1897741798040051, + "grad_norm": 15.4375, + "learning_rate": 2.4492499508845403e-06, + "loss": 0.7543, + "step": 2227 + }, + { + "epoch": 0.18985939497230506, + "grad_norm": 11.9375, + "learning_rate": 2.449202545766917e-06, + "loss": 0.6175, + "step": 2228 + }, + { + "epoch": 0.18994461014060504, + "grad_norm": 13.5, + "learning_rate": 2.4491551189784978e-06, + "loss": 0.7408, + "step": 2229 + }, + { + "epoch": 0.190029825308905, + "grad_norm": 13.875, + "learning_rate": 2.44910767052014e-06, + "loss": 0.7936, + "step": 2230 + }, + { + "epoch": 0.19011504047720496, + "grad_norm": 17.375, + "learning_rate": 2.4490602003927006e-06, + "loss": 0.5238, + "step": 2231 + }, + { + "epoch": 0.1902002556455049, + "grad_norm": 13.6875, + "learning_rate": 2.4490127085970376e-06, + "loss": 0.8359, + "step": 2232 + }, + { + "epoch": 0.19028547081380487, + "grad_norm": 11.5625, + "learning_rate": 2.4489651951340093e-06, + "loss": 0.5499, + "step": 2233 + }, + { + "epoch": 0.19037068598210483, + "grad_norm": 24.875, + "learning_rate": 2.448917660004474e-06, + "loss": 1.2219, + "step": 2234 + }, + { + "epoch": 0.19045590115040478, + "grad_norm": 14.4375, + "learning_rate": 2.4488701032092917e-06, + "loss": 0.8399, + "step": 2235 + }, + { + "epoch": 0.19054111631870474, + "grad_norm": 19.125, + "learning_rate": 2.44882252474932e-06, + "loss": 0.8813, + "step": 2236 + }, + { + "epoch": 0.1906263314870047, + "grad_norm": 11.6875, + "learning_rate": 2.4487749246254206e-06, + "loss": 0.4664, + "step": 2237 + }, + { + "epoch": 0.19071154665530465, + "grad_norm": 14.375, + "learning_rate": 2.448727302838453e-06, + "loss": 0.8084, + "step": 2238 + }, + { + "epoch": 0.1907967618236046, + "grad_norm": 15.5, + "learning_rate": 2.448679659389277e-06, + "loss": 0.3714, + "step": 2239 + }, + { + "epoch": 0.19088197699190457, + "grad_norm": 13.1875, + "learning_rate": 2.4486319942787547e-06, + "loss": 0.6104, + "step": 2240 + }, + { + "epoch": 0.19096719216020452, + "grad_norm": 21.375, + "learning_rate": 2.4485843075077464e-06, + "loss": 0.9878, + "step": 2241 + }, + { + "epoch": 0.19105240732850448, + "grad_norm": 18.5, + "learning_rate": 2.4485365990771145e-06, + "loss": 1.0321, + "step": 2242 + }, + { + "epoch": 0.19113762249680444, + "grad_norm": 15.0625, + "learning_rate": 2.448488868987721e-06, + "loss": 0.8145, + "step": 2243 + }, + { + "epoch": 0.1912228376651044, + "grad_norm": 18.5, + "learning_rate": 2.448441117240429e-06, + "loss": 1.2219, + "step": 2244 + }, + { + "epoch": 0.19130805283340435, + "grad_norm": 15.0, + "learning_rate": 2.4483933438361005e-06, + "loss": 0.6343, + "step": 2245 + }, + { + "epoch": 0.1913932680017043, + "grad_norm": 15.5625, + "learning_rate": 2.4483455487755992e-06, + "loss": 0.7822, + "step": 2246 + }, + { + "epoch": 0.19147848317000427, + "grad_norm": 15.875, + "learning_rate": 2.448297732059789e-06, + "loss": 0.8574, + "step": 2247 + }, + { + "epoch": 0.19156369833830422, + "grad_norm": 15.25, + "learning_rate": 2.4482498936895337e-06, + "loss": 1.0431, + "step": 2248 + }, + { + "epoch": 0.19164891350660418, + "grad_norm": 15.9375, + "learning_rate": 2.4482020336656978e-06, + "loss": 0.7653, + "step": 2249 + }, + { + "epoch": 0.19173412867490414, + "grad_norm": 16.625, + "learning_rate": 2.4481541519891464e-06, + "loss": 0.869, + "step": 2250 + }, + { + "epoch": 0.1918193438432041, + "grad_norm": 14.125, + "learning_rate": 2.4481062486607447e-06, + "loss": 0.7833, + "step": 2251 + }, + { + "epoch": 0.19190455901150405, + "grad_norm": 16.875, + "learning_rate": 2.448058323681358e-06, + "loss": 0.9384, + "step": 2252 + }, + { + "epoch": 0.191989774179804, + "grad_norm": 25.0, + "learning_rate": 2.4480103770518528e-06, + "loss": 0.7465, + "step": 2253 + }, + { + "epoch": 0.19207498934810396, + "grad_norm": 16.0, + "learning_rate": 2.447962408773095e-06, + "loss": 0.449, + "step": 2254 + }, + { + "epoch": 0.19216020451640392, + "grad_norm": 9.125, + "learning_rate": 2.4479144188459524e-06, + "loss": 0.3205, + "step": 2255 + }, + { + "epoch": 0.19224541968470388, + "grad_norm": 14.6875, + "learning_rate": 2.4478664072712916e-06, + "loss": 0.8656, + "step": 2256 + }, + { + "epoch": 0.19233063485300383, + "grad_norm": 14.9375, + "learning_rate": 2.4478183740499805e-06, + "loss": 0.9468, + "step": 2257 + }, + { + "epoch": 0.1924158500213038, + "grad_norm": 12.375, + "learning_rate": 2.447770319182886e-06, + "loss": 0.5751, + "step": 2258 + }, + { + "epoch": 0.19250106518960375, + "grad_norm": 26.125, + "learning_rate": 2.4477222426708783e-06, + "loss": 1.1603, + "step": 2259 + }, + { + "epoch": 0.1925862803579037, + "grad_norm": 11.125, + "learning_rate": 2.447674144514825e-06, + "loss": 0.4086, + "step": 2260 + }, + { + "epoch": 0.19267149552620366, + "grad_norm": 16.125, + "learning_rate": 2.4476260247155955e-06, + "loss": 0.9462, + "step": 2261 + }, + { + "epoch": 0.19275671069450362, + "grad_norm": 23.625, + "learning_rate": 2.447577883274059e-06, + "loss": 0.9794, + "step": 2262 + }, + { + "epoch": 0.19284192586280358, + "grad_norm": 13.75, + "learning_rate": 2.4475297201910865e-06, + "loss": 0.6012, + "step": 2263 + }, + { + "epoch": 0.19292714103110353, + "grad_norm": 12.9375, + "learning_rate": 2.4474815354675473e-06, + "loss": 0.7466, + "step": 2264 + }, + { + "epoch": 0.1930123561994035, + "grad_norm": 30.375, + "learning_rate": 2.4474333291043127e-06, + "loss": 1.0961, + "step": 2265 + }, + { + "epoch": 0.19309757136770345, + "grad_norm": 19.625, + "learning_rate": 2.4473851011022544e-06, + "loss": 0.5913, + "step": 2266 + }, + { + "epoch": 0.1931827865360034, + "grad_norm": 32.25, + "learning_rate": 2.4473368514622426e-06, + "loss": 1.0623, + "step": 2267 + }, + { + "epoch": 0.19326800170430336, + "grad_norm": 16.375, + "learning_rate": 2.44728858018515e-06, + "loss": 0.8422, + "step": 2268 + }, + { + "epoch": 0.19335321687260332, + "grad_norm": 13.375, + "learning_rate": 2.4472402872718486e-06, + "loss": 0.5212, + "step": 2269 + }, + { + "epoch": 0.19343843204090327, + "grad_norm": 15.6875, + "learning_rate": 2.4471919727232113e-06, + "loss": 0.7781, + "step": 2270 + }, + { + "epoch": 0.19352364720920323, + "grad_norm": 22.625, + "learning_rate": 2.447143636540111e-06, + "loss": 0.7535, + "step": 2271 + }, + { + "epoch": 0.1936088623775032, + "grad_norm": 24.25, + "learning_rate": 2.447095278723422e-06, + "loss": 1.0257, + "step": 2272 + }, + { + "epoch": 0.19369407754580314, + "grad_norm": 12.375, + "learning_rate": 2.4470468992740173e-06, + "loss": 0.6115, + "step": 2273 + }, + { + "epoch": 0.1937792927141031, + "grad_norm": 15.875, + "learning_rate": 2.4469984981927714e-06, + "loss": 0.7066, + "step": 2274 + }, + { + "epoch": 0.19386450788240306, + "grad_norm": 13.0625, + "learning_rate": 2.446950075480559e-06, + "loss": 0.5466, + "step": 2275 + }, + { + "epoch": 0.19394972305070302, + "grad_norm": 14.4375, + "learning_rate": 2.4469016311382547e-06, + "loss": 0.7272, + "step": 2276 + }, + { + "epoch": 0.19403493821900297, + "grad_norm": 18.75, + "learning_rate": 2.4468531651667347e-06, + "loss": 1.1441, + "step": 2277 + }, + { + "epoch": 0.19412015338730293, + "grad_norm": 16.5, + "learning_rate": 2.4468046775668746e-06, + "loss": 0.8277, + "step": 2278 + }, + { + "epoch": 0.1942053685556029, + "grad_norm": 18.875, + "learning_rate": 2.4467561683395504e-06, + "loss": 0.9888, + "step": 2279 + }, + { + "epoch": 0.19429058372390284, + "grad_norm": 13.0625, + "learning_rate": 2.446707637485639e-06, + "loss": 0.6971, + "step": 2280 + }, + { + "epoch": 0.1943757988922028, + "grad_norm": 14.625, + "learning_rate": 2.446659085006017e-06, + "loss": 0.6163, + "step": 2281 + }, + { + "epoch": 0.19446101406050276, + "grad_norm": 10.3125, + "learning_rate": 2.4466105109015618e-06, + "loss": 0.754, + "step": 2282 + }, + { + "epoch": 0.19454622922880271, + "grad_norm": 11.25, + "learning_rate": 2.4465619151731513e-06, + "loss": 0.5766, + "step": 2283 + }, + { + "epoch": 0.19463144439710267, + "grad_norm": 14.0625, + "learning_rate": 2.446513297821664e-06, + "loss": 0.8218, + "step": 2284 + }, + { + "epoch": 0.19471665956540266, + "grad_norm": 14.3125, + "learning_rate": 2.4464646588479784e-06, + "loss": 0.8447, + "step": 2285 + }, + { + "epoch": 0.1948018747337026, + "grad_norm": 15.5, + "learning_rate": 2.446415998252973e-06, + "loss": 0.9637, + "step": 2286 + }, + { + "epoch": 0.19488708990200257, + "grad_norm": 17.125, + "learning_rate": 2.4463673160375274e-06, + "loss": 0.4817, + "step": 2287 + }, + { + "epoch": 0.19497230507030253, + "grad_norm": 15.625, + "learning_rate": 2.4463186122025217e-06, + "loss": 0.6415, + "step": 2288 + }, + { + "epoch": 0.19505752023860248, + "grad_norm": 11.6875, + "learning_rate": 2.4462698867488356e-06, + "loss": 0.6263, + "step": 2289 + }, + { + "epoch": 0.19514273540690244, + "grad_norm": 12.8125, + "learning_rate": 2.4462211396773495e-06, + "loss": 0.9363, + "step": 2290 + }, + { + "epoch": 0.1952279505752024, + "grad_norm": 16.25, + "learning_rate": 2.4461723709889448e-06, + "loss": 0.7639, + "step": 2291 + }, + { + "epoch": 0.19531316574350235, + "grad_norm": 12.0, + "learning_rate": 2.446123580684502e-06, + "loss": 0.6631, + "step": 2292 + }, + { + "epoch": 0.1953983809118023, + "grad_norm": 14.625, + "learning_rate": 2.4460747687649035e-06, + "loss": 0.9672, + "step": 2293 + }, + { + "epoch": 0.19548359608010227, + "grad_norm": 15.8125, + "learning_rate": 2.4460259352310313e-06, + "loss": 0.8038, + "step": 2294 + }, + { + "epoch": 0.19556881124840222, + "grad_norm": 13.0, + "learning_rate": 2.445977080083768e-06, + "loss": 0.4675, + "step": 2295 + }, + { + "epoch": 0.19565402641670218, + "grad_norm": 10.4375, + "learning_rate": 2.4459282033239957e-06, + "loss": 0.3959, + "step": 2296 + }, + { + "epoch": 0.19573924158500214, + "grad_norm": 15.6875, + "learning_rate": 2.4458793049525987e-06, + "loss": 0.9084, + "step": 2297 + }, + { + "epoch": 0.1958244567533021, + "grad_norm": 16.75, + "learning_rate": 2.4458303849704597e-06, + "loss": 0.7232, + "step": 2298 + }, + { + "epoch": 0.19590967192160205, + "grad_norm": 13.875, + "learning_rate": 2.4457814433784633e-06, + "loss": 0.722, + "step": 2299 + }, + { + "epoch": 0.195994887089902, + "grad_norm": 16.125, + "learning_rate": 2.4457324801774934e-06, + "loss": 1.0233, + "step": 2300 + }, + { + "epoch": 0.19608010225820197, + "grad_norm": 15.25, + "learning_rate": 2.4456834953684357e-06, + "loss": 0.5372, + "step": 2301 + }, + { + "epoch": 0.19616531742650192, + "grad_norm": 13.375, + "learning_rate": 2.445634488952175e-06, + "loss": 0.7053, + "step": 2302 + }, + { + "epoch": 0.19625053259480188, + "grad_norm": 22.875, + "learning_rate": 2.4455854609295963e-06, + "loss": 1.2031, + "step": 2303 + }, + { + "epoch": 0.19633574776310184, + "grad_norm": 11.875, + "learning_rate": 2.4455364113015863e-06, + "loss": 0.6793, + "step": 2304 + }, + { + "epoch": 0.1964209629314018, + "grad_norm": 16.5, + "learning_rate": 2.445487340069031e-06, + "loss": 1.0749, + "step": 2305 + }, + { + "epoch": 0.19650617809970175, + "grad_norm": 13.0625, + "learning_rate": 2.4454382472328174e-06, + "loss": 0.8275, + "step": 2306 + }, + { + "epoch": 0.1965913932680017, + "grad_norm": 13.1875, + "learning_rate": 2.4453891327938324e-06, + "loss": 0.8679, + "step": 2307 + }, + { + "epoch": 0.19667660843630166, + "grad_norm": 21.625, + "learning_rate": 2.445339996752964e-06, + "loss": 1.1166, + "step": 2308 + }, + { + "epoch": 0.19676182360460162, + "grad_norm": 12.25, + "learning_rate": 2.4452908391111e-06, + "loss": 0.4847, + "step": 2309 + }, + { + "epoch": 0.19684703877290158, + "grad_norm": 20.375, + "learning_rate": 2.445241659869128e-06, + "loss": 1.0299, + "step": 2310 + }, + { + "epoch": 0.19693225394120153, + "grad_norm": 12.5, + "learning_rate": 2.4451924590279374e-06, + "loss": 0.5539, + "step": 2311 + }, + { + "epoch": 0.1970174691095015, + "grad_norm": 20.5, + "learning_rate": 2.4451432365884175e-06, + "loss": 0.8347, + "step": 2312 + }, + { + "epoch": 0.19710268427780145, + "grad_norm": 16.25, + "learning_rate": 2.4450939925514577e-06, + "loss": 0.7147, + "step": 2313 + }, + { + "epoch": 0.1971878994461014, + "grad_norm": 14.25, + "learning_rate": 2.445044726917947e-06, + "loss": 0.9938, + "step": 2314 + }, + { + "epoch": 0.19727311461440136, + "grad_norm": 16.375, + "learning_rate": 2.444995439688777e-06, + "loss": 0.9131, + "step": 2315 + }, + { + "epoch": 0.19735832978270132, + "grad_norm": 19.5, + "learning_rate": 2.4449461308648375e-06, + "loss": 0.9354, + "step": 2316 + }, + { + "epoch": 0.19744354495100128, + "grad_norm": 12.875, + "learning_rate": 2.44489680044702e-06, + "loss": 0.6274, + "step": 2317 + }, + { + "epoch": 0.19752876011930123, + "grad_norm": 13.0, + "learning_rate": 2.4448474484362157e-06, + "loss": 0.4424, + "step": 2318 + }, + { + "epoch": 0.1976139752876012, + "grad_norm": 17.0, + "learning_rate": 2.444798074833316e-06, + "loss": 0.7284, + "step": 2319 + }, + { + "epoch": 0.19769919045590115, + "grad_norm": 20.75, + "learning_rate": 2.444748679639214e-06, + "loss": 1.13, + "step": 2320 + }, + { + "epoch": 0.1977844056242011, + "grad_norm": 15.375, + "learning_rate": 2.444699262854802e-06, + "loss": 1.0087, + "step": 2321 + }, + { + "epoch": 0.19786962079250106, + "grad_norm": 13.5, + "learning_rate": 2.444649824480973e-06, + "loss": 0.7771, + "step": 2322 + }, + { + "epoch": 0.19795483596080102, + "grad_norm": 22.875, + "learning_rate": 2.4446003645186204e-06, + "loss": 1.2427, + "step": 2323 + }, + { + "epoch": 0.19804005112910097, + "grad_norm": 16.25, + "learning_rate": 2.4445508829686374e-06, + "loss": 0.9601, + "step": 2324 + }, + { + "epoch": 0.19812526629740093, + "grad_norm": 13.1875, + "learning_rate": 2.4445013798319194e-06, + "loss": 0.8291, + "step": 2325 + }, + { + "epoch": 0.1982104814657009, + "grad_norm": 19.0, + "learning_rate": 2.44445185510936e-06, + "loss": 1.0551, + "step": 2326 + }, + { + "epoch": 0.19829569663400085, + "grad_norm": 13.8125, + "learning_rate": 2.444402308801855e-06, + "loss": 0.5958, + "step": 2327 + }, + { + "epoch": 0.1983809118023008, + "grad_norm": 16.625, + "learning_rate": 2.4443527409102986e-06, + "loss": 1.0035, + "step": 2328 + }, + { + "epoch": 0.19846612697060076, + "grad_norm": 12.3125, + "learning_rate": 2.4443031514355873e-06, + "loss": 0.6038, + "step": 2329 + }, + { + "epoch": 0.19855134213890072, + "grad_norm": 15.625, + "learning_rate": 2.444253540378617e-06, + "loss": 1.0203, + "step": 2330 + }, + { + "epoch": 0.19863655730720067, + "grad_norm": 20.0, + "learning_rate": 2.4442039077402845e-06, + "loss": 1.0133, + "step": 2331 + }, + { + "epoch": 0.19872177247550063, + "grad_norm": 17.125, + "learning_rate": 2.444154253521486e-06, + "loss": 0.9023, + "step": 2332 + }, + { + "epoch": 0.1988069876438006, + "grad_norm": 15.3125, + "learning_rate": 2.4441045777231197e-06, + "loss": 1.0655, + "step": 2333 + }, + { + "epoch": 0.19889220281210054, + "grad_norm": 21.75, + "learning_rate": 2.4440548803460835e-06, + "loss": 1.1842, + "step": 2334 + }, + { + "epoch": 0.1989774179804005, + "grad_norm": 18.5, + "learning_rate": 2.444005161391274e-06, + "loss": 1.0103, + "step": 2335 + }, + { + "epoch": 0.19906263314870046, + "grad_norm": 21.75, + "learning_rate": 2.443955420859591e-06, + "loss": 0.8935, + "step": 2336 + }, + { + "epoch": 0.19914784831700041, + "grad_norm": 17.75, + "learning_rate": 2.4439056587519327e-06, + "loss": 1.0824, + "step": 2337 + }, + { + "epoch": 0.19923306348530037, + "grad_norm": 14.0, + "learning_rate": 2.443855875069199e-06, + "loss": 0.8669, + "step": 2338 + }, + { + "epoch": 0.19931827865360033, + "grad_norm": 14.375, + "learning_rate": 2.443806069812289e-06, + "loss": 0.7385, + "step": 2339 + }, + { + "epoch": 0.19940349382190029, + "grad_norm": 13.5, + "learning_rate": 2.4437562429821026e-06, + "loss": 0.7603, + "step": 2340 + }, + { + "epoch": 0.19948870899020024, + "grad_norm": 12.5, + "learning_rate": 2.4437063945795405e-06, + "loss": 0.4115, + "step": 2341 + }, + { + "epoch": 0.19957392415850023, + "grad_norm": 21.875, + "learning_rate": 2.4436565246055035e-06, + "loss": 1.0797, + "step": 2342 + }, + { + "epoch": 0.19965913932680018, + "grad_norm": 13.75, + "learning_rate": 2.443606633060893e-06, + "loss": 0.9298, + "step": 2343 + }, + { + "epoch": 0.19974435449510014, + "grad_norm": 11.4375, + "learning_rate": 2.44355671994661e-06, + "loss": 0.4778, + "step": 2344 + }, + { + "epoch": 0.1998295696634001, + "grad_norm": 18.625, + "learning_rate": 2.443506785263557e-06, + "loss": 0.786, + "step": 2345 + }, + { + "epoch": 0.19991478483170005, + "grad_norm": 15.625, + "learning_rate": 2.4434568290126364e-06, + "loss": 0.7459, + "step": 2346 + }, + { + "epoch": 0.2, + "grad_norm": 20.5, + "learning_rate": 2.443406851194751e-06, + "loss": 0.77, + "step": 2347 + }, + { + "epoch": 0.20008521516829997, + "grad_norm": 19.125, + "learning_rate": 2.443356851810803e-06, + "loss": 0.9424, + "step": 2348 + }, + { + "epoch": 0.20017043033659992, + "grad_norm": 14.25, + "learning_rate": 2.4433068308616976e-06, + "loss": 0.5636, + "step": 2349 + }, + { + "epoch": 0.20025564550489988, + "grad_norm": 15.25, + "learning_rate": 2.443256788348337e-06, + "loss": 0.6556, + "step": 2350 + }, + { + "epoch": 0.20034086067319984, + "grad_norm": 16.375, + "learning_rate": 2.443206724271627e-06, + "loss": 1.2988, + "step": 2351 + }, + { + "epoch": 0.2004260758414998, + "grad_norm": 14.4375, + "learning_rate": 2.443156638632471e-06, + "loss": 0.6152, + "step": 2352 + }, + { + "epoch": 0.20051129100979975, + "grad_norm": 14.5, + "learning_rate": 2.4431065314317747e-06, + "loss": 0.6758, + "step": 2353 + }, + { + "epoch": 0.2005965061780997, + "grad_norm": 17.125, + "learning_rate": 2.4430564026704444e-06, + "loss": 1.0294, + "step": 2354 + }, + { + "epoch": 0.20068172134639967, + "grad_norm": 13.4375, + "learning_rate": 2.4430062523493845e-06, + "loss": 0.7881, + "step": 2355 + }, + { + "epoch": 0.20076693651469962, + "grad_norm": 12.5, + "learning_rate": 2.442956080469502e-06, + "loss": 0.5759, + "step": 2356 + }, + { + "epoch": 0.20085215168299958, + "grad_norm": 13.875, + "learning_rate": 2.442905887031704e-06, + "loss": 0.7606, + "step": 2357 + }, + { + "epoch": 0.20093736685129954, + "grad_norm": 12.75, + "learning_rate": 2.442855672036897e-06, + "loss": 0.6191, + "step": 2358 + }, + { + "epoch": 0.2010225820195995, + "grad_norm": 16.375, + "learning_rate": 2.4428054354859878e-06, + "loss": 0.7308, + "step": 2359 + }, + { + "epoch": 0.20110779718789945, + "grad_norm": 17.25, + "learning_rate": 2.442755177379885e-06, + "loss": 0.999, + "step": 2360 + }, + { + "epoch": 0.2011930123561994, + "grad_norm": 25.875, + "learning_rate": 2.4427048977194975e-06, + "loss": 1.0313, + "step": 2361 + }, + { + "epoch": 0.20127822752449936, + "grad_norm": 17.125, + "learning_rate": 2.442654596505733e-06, + "loss": 1.0347, + "step": 2362 + }, + { + "epoch": 0.20136344269279932, + "grad_norm": 12.125, + "learning_rate": 2.4426042737395e-06, + "loss": 0.6473, + "step": 2363 + }, + { + "epoch": 0.20144865786109928, + "grad_norm": 17.0, + "learning_rate": 2.4425539294217087e-06, + "loss": 0.6536, + "step": 2364 + }, + { + "epoch": 0.20153387302939924, + "grad_norm": 22.5, + "learning_rate": 2.442503563553269e-06, + "loss": 1.0198, + "step": 2365 + }, + { + "epoch": 0.2016190881976992, + "grad_norm": 17.0, + "learning_rate": 2.4424531761350906e-06, + "loss": 0.5529, + "step": 2366 + }, + { + "epoch": 0.20170430336599915, + "grad_norm": 13.75, + "learning_rate": 2.4424027671680845e-06, + "loss": 0.4424, + "step": 2367 + }, + { + "epoch": 0.2017895185342991, + "grad_norm": 21.625, + "learning_rate": 2.442352336653161e-06, + "loss": 1.1686, + "step": 2368 + }, + { + "epoch": 0.20187473370259906, + "grad_norm": 9.75, + "learning_rate": 2.4423018845912315e-06, + "loss": 0.3541, + "step": 2369 + }, + { + "epoch": 0.20195994887089902, + "grad_norm": 36.0, + "learning_rate": 2.4422514109832083e-06, + "loss": 0.951, + "step": 2370 + }, + { + "epoch": 0.20204516403919898, + "grad_norm": 11.5625, + "learning_rate": 2.4422009158300035e-06, + "loss": 0.4967, + "step": 2371 + }, + { + "epoch": 0.20213037920749893, + "grad_norm": 14.25, + "learning_rate": 2.442150399132529e-06, + "loss": 0.7606, + "step": 2372 + }, + { + "epoch": 0.2022155943757989, + "grad_norm": 13.25, + "learning_rate": 2.442099860891698e-06, + "loss": 0.504, + "step": 2373 + }, + { + "epoch": 0.20230080954409885, + "grad_norm": 10.25, + "learning_rate": 2.442049301108424e-06, + "loss": 0.3663, + "step": 2374 + }, + { + "epoch": 0.2023860247123988, + "grad_norm": 9.875, + "learning_rate": 2.4419987197836204e-06, + "loss": 0.4461, + "step": 2375 + }, + { + "epoch": 0.20247123988069876, + "grad_norm": 17.375, + "learning_rate": 2.4419481169182006e-06, + "loss": 0.7311, + "step": 2376 + }, + { + "epoch": 0.20255645504899872, + "grad_norm": 19.125, + "learning_rate": 2.4418974925130805e-06, + "loss": 1.0781, + "step": 2377 + }, + { + "epoch": 0.20264167021729868, + "grad_norm": 16.125, + "learning_rate": 2.4418468465691743e-06, + "loss": 0.7007, + "step": 2378 + }, + { + "epoch": 0.20272688538559863, + "grad_norm": 20.75, + "learning_rate": 2.4417961790873963e-06, + "loss": 0.8812, + "step": 2379 + }, + { + "epoch": 0.2028121005538986, + "grad_norm": 11.875, + "learning_rate": 2.4417454900686634e-06, + "loss": 0.5282, + "step": 2380 + }, + { + "epoch": 0.20289731572219855, + "grad_norm": 12.6875, + "learning_rate": 2.441694779513891e-06, + "loss": 0.6174, + "step": 2381 + }, + { + "epoch": 0.2029825308904985, + "grad_norm": 30.0, + "learning_rate": 2.4416440474239953e-06, + "loss": 1.0134, + "step": 2382 + }, + { + "epoch": 0.20306774605879846, + "grad_norm": 17.25, + "learning_rate": 2.4415932937998936e-06, + "loss": 0.6321, + "step": 2383 + }, + { + "epoch": 0.20315296122709842, + "grad_norm": 10.625, + "learning_rate": 2.441542518642503e-06, + "loss": 0.4032, + "step": 2384 + }, + { + "epoch": 0.20323817639539837, + "grad_norm": 15.25, + "learning_rate": 2.441491721952741e-06, + "loss": 0.8512, + "step": 2385 + }, + { + "epoch": 0.20332339156369833, + "grad_norm": 18.75, + "learning_rate": 2.4414409037315255e-06, + "loss": 0.8377, + "step": 2386 + }, + { + "epoch": 0.2034086067319983, + "grad_norm": 13.1875, + "learning_rate": 2.4413900639797745e-06, + "loss": 0.7815, + "step": 2387 + }, + { + "epoch": 0.20349382190029824, + "grad_norm": 12.5, + "learning_rate": 2.4413392026984066e-06, + "loss": 0.7139, + "step": 2388 + }, + { + "epoch": 0.2035790370685982, + "grad_norm": 14.75, + "learning_rate": 2.441288319888342e-06, + "loss": 0.5836, + "step": 2389 + }, + { + "epoch": 0.20366425223689816, + "grad_norm": 10.0, + "learning_rate": 2.4412374155504993e-06, + "loss": 0.3388, + "step": 2390 + }, + { + "epoch": 0.20374946740519811, + "grad_norm": 19.875, + "learning_rate": 2.4411864896857986e-06, + "loss": 1.0113, + "step": 2391 + }, + { + "epoch": 0.20383468257349807, + "grad_norm": 15.5, + "learning_rate": 2.4411355422951607e-06, + "loss": 0.9502, + "step": 2392 + }, + { + "epoch": 0.20391989774179803, + "grad_norm": 34.5, + "learning_rate": 2.4410845733795047e-06, + "loss": 0.6113, + "step": 2393 + }, + { + "epoch": 0.20400511291009799, + "grad_norm": 13.4375, + "learning_rate": 2.4410335829397535e-06, + "loss": 0.8717, + "step": 2394 + }, + { + "epoch": 0.20409032807839794, + "grad_norm": 13.5, + "learning_rate": 2.4409825709768277e-06, + "loss": 0.9236, + "step": 2395 + }, + { + "epoch": 0.2041755432466979, + "grad_norm": 23.75, + "learning_rate": 2.440931537491649e-06, + "loss": 0.8897, + "step": 2396 + }, + { + "epoch": 0.20426075841499786, + "grad_norm": 11.875, + "learning_rate": 2.44088048248514e-06, + "loss": 0.5197, + "step": 2397 + }, + { + "epoch": 0.20434597358329784, + "grad_norm": 17.25, + "learning_rate": 2.440829405958223e-06, + "loss": 1.1689, + "step": 2398 + }, + { + "epoch": 0.2044311887515978, + "grad_norm": 17.25, + "learning_rate": 2.4407783079118214e-06, + "loss": 1.0093, + "step": 2399 + }, + { + "epoch": 0.20451640391989775, + "grad_norm": 12.75, + "learning_rate": 2.440727188346858e-06, + "loss": 0.4181, + "step": 2400 + }, + { + "epoch": 0.2046016190881977, + "grad_norm": 11.125, + "learning_rate": 2.4406760472642576e-06, + "loss": 0.6745, + "step": 2401 + }, + { + "epoch": 0.20468683425649767, + "grad_norm": 14.5625, + "learning_rate": 2.440624884664943e-06, + "loss": 0.7777, + "step": 2402 + }, + { + "epoch": 0.20477204942479763, + "grad_norm": 12.875, + "learning_rate": 2.44057370054984e-06, + "loss": 0.6037, + "step": 2403 + }, + { + "epoch": 0.20485726459309758, + "grad_norm": 13.5, + "learning_rate": 2.4405224949198725e-06, + "loss": 0.7889, + "step": 2404 + }, + { + "epoch": 0.20494247976139754, + "grad_norm": 18.625, + "learning_rate": 2.4404712677759667e-06, + "loss": 0.6122, + "step": 2405 + }, + { + "epoch": 0.2050276949296975, + "grad_norm": 21.5, + "learning_rate": 2.440420019119048e-06, + "loss": 0.9379, + "step": 2406 + }, + { + "epoch": 0.20511291009799745, + "grad_norm": 19.875, + "learning_rate": 2.440368748950042e-06, + "loss": 1.0805, + "step": 2407 + }, + { + "epoch": 0.2051981252662974, + "grad_norm": 12.75, + "learning_rate": 2.4403174572698764e-06, + "loss": 0.7804, + "step": 2408 + }, + { + "epoch": 0.20528334043459737, + "grad_norm": 13.8125, + "learning_rate": 2.440266144079477e-06, + "loss": 0.8601, + "step": 2409 + }, + { + "epoch": 0.20536855560289732, + "grad_norm": 11.6875, + "learning_rate": 2.4402148093797717e-06, + "loss": 0.5067, + "step": 2410 + }, + { + "epoch": 0.20545377077119728, + "grad_norm": 19.125, + "learning_rate": 2.440163453171688e-06, + "loss": 1.0207, + "step": 2411 + }, + { + "epoch": 0.20553898593949724, + "grad_norm": 19.0, + "learning_rate": 2.4401120754561535e-06, + "loss": 0.9645, + "step": 2412 + }, + { + "epoch": 0.2056242011077972, + "grad_norm": 11.625, + "learning_rate": 2.4400606762340973e-06, + "loss": 0.3906, + "step": 2413 + }, + { + "epoch": 0.20570941627609715, + "grad_norm": 17.375, + "learning_rate": 2.4400092555064475e-06, + "loss": 0.7098, + "step": 2414 + }, + { + "epoch": 0.2057946314443971, + "grad_norm": 11.0625, + "learning_rate": 2.4399578132741343e-06, + "loss": 0.6405, + "step": 2415 + }, + { + "epoch": 0.20587984661269706, + "grad_norm": 14.0625, + "learning_rate": 2.439906349538087e-06, + "loss": 0.6483, + "step": 2416 + }, + { + "epoch": 0.20596506178099702, + "grad_norm": 16.625, + "learning_rate": 2.439854864299235e-06, + "loss": 0.6714, + "step": 2417 + }, + { + "epoch": 0.20605027694929698, + "grad_norm": 17.625, + "learning_rate": 2.4398033575585095e-06, + "loss": 0.7971, + "step": 2418 + }, + { + "epoch": 0.20613549211759694, + "grad_norm": 13.375, + "learning_rate": 2.4397518293168403e-06, + "loss": 0.5174, + "step": 2419 + }, + { + "epoch": 0.2062207072858969, + "grad_norm": 17.0, + "learning_rate": 2.4397002795751595e-06, + "loss": 0.9742, + "step": 2420 + }, + { + "epoch": 0.20630592245419685, + "grad_norm": 20.375, + "learning_rate": 2.4396487083343984e-06, + "loss": 0.95, + "step": 2421 + }, + { + "epoch": 0.2063911376224968, + "grad_norm": 11.9375, + "learning_rate": 2.439597115595489e-06, + "loss": 0.4189, + "step": 2422 + }, + { + "epoch": 0.20647635279079676, + "grad_norm": 10.1875, + "learning_rate": 2.4395455013593634e-06, + "loss": 0.4058, + "step": 2423 + }, + { + "epoch": 0.20656156795909672, + "grad_norm": 10.25, + "learning_rate": 2.439493865626954e-06, + "loss": 0.3387, + "step": 2424 + }, + { + "epoch": 0.20664678312739668, + "grad_norm": 37.25, + "learning_rate": 2.439442208399195e-06, + "loss": 0.9265, + "step": 2425 + }, + { + "epoch": 0.20673199829569663, + "grad_norm": 23.625, + "learning_rate": 2.4393905296770183e-06, + "loss": 0.9737, + "step": 2426 + }, + { + "epoch": 0.2068172134639966, + "grad_norm": 13.625, + "learning_rate": 2.4393388294613596e-06, + "loss": 0.8829, + "step": 2427 + }, + { + "epoch": 0.20690242863229655, + "grad_norm": 14.5, + "learning_rate": 2.4392871077531516e-06, + "loss": 0.6078, + "step": 2428 + }, + { + "epoch": 0.2069876438005965, + "grad_norm": 12.1875, + "learning_rate": 2.43923536455333e-06, + "loss": 0.4589, + "step": 2429 + }, + { + "epoch": 0.20707285896889646, + "grad_norm": 16.375, + "learning_rate": 2.4391835998628295e-06, + "loss": 1.1341, + "step": 2430 + }, + { + "epoch": 0.20715807413719642, + "grad_norm": 16.875, + "learning_rate": 2.4391318136825855e-06, + "loss": 0.7221, + "step": 2431 + }, + { + "epoch": 0.20724328930549638, + "grad_norm": 17.25, + "learning_rate": 2.439080006013534e-06, + "loss": 0.7102, + "step": 2432 + }, + { + "epoch": 0.20732850447379633, + "grad_norm": 15.5625, + "learning_rate": 2.4390281768566108e-06, + "loss": 0.7828, + "step": 2433 + }, + { + "epoch": 0.2074137196420963, + "grad_norm": 13.0625, + "learning_rate": 2.4389763262127532e-06, + "loss": 0.7476, + "step": 2434 + }, + { + "epoch": 0.20749893481039625, + "grad_norm": 11.9375, + "learning_rate": 2.438924454082897e-06, + "loss": 0.5536, + "step": 2435 + }, + { + "epoch": 0.2075841499786962, + "grad_norm": 14.125, + "learning_rate": 2.4388725604679807e-06, + "loss": 0.7197, + "step": 2436 + }, + { + "epoch": 0.20766936514699616, + "grad_norm": 14.3125, + "learning_rate": 2.4388206453689422e-06, + "loss": 0.7193, + "step": 2437 + }, + { + "epoch": 0.20775458031529612, + "grad_norm": 14.9375, + "learning_rate": 2.438768708786719e-06, + "loss": 0.9624, + "step": 2438 + }, + { + "epoch": 0.20783979548359607, + "grad_norm": 10.3125, + "learning_rate": 2.4387167507222494e-06, + "loss": 0.419, + "step": 2439 + }, + { + "epoch": 0.20792501065189603, + "grad_norm": 16.5, + "learning_rate": 2.4386647711764732e-06, + "loss": 0.9562, + "step": 2440 + }, + { + "epoch": 0.208010225820196, + "grad_norm": 15.375, + "learning_rate": 2.438612770150329e-06, + "loss": 0.829, + "step": 2441 + }, + { + "epoch": 0.20809544098849594, + "grad_norm": 12.5625, + "learning_rate": 2.438560747644757e-06, + "loss": 0.5211, + "step": 2442 + }, + { + "epoch": 0.2081806561567959, + "grad_norm": 19.375, + "learning_rate": 2.438508703660697e-06, + "loss": 1.0678, + "step": 2443 + }, + { + "epoch": 0.20826587132509586, + "grad_norm": 11.6875, + "learning_rate": 2.4384566381990897e-06, + "loss": 0.4248, + "step": 2444 + }, + { + "epoch": 0.20835108649339582, + "grad_norm": 17.75, + "learning_rate": 2.438404551260876e-06, + "loss": 0.7505, + "step": 2445 + }, + { + "epoch": 0.20843630166169577, + "grad_norm": 17.5, + "learning_rate": 2.438352442846997e-06, + "loss": 1.0375, + "step": 2446 + }, + { + "epoch": 0.20852151682999573, + "grad_norm": 16.5, + "learning_rate": 2.438300312958394e-06, + "loss": 0.8947, + "step": 2447 + }, + { + "epoch": 0.20860673199829569, + "grad_norm": 11.0, + "learning_rate": 2.4382481615960096e-06, + "loss": 0.498, + "step": 2448 + }, + { + "epoch": 0.20869194716659564, + "grad_norm": 18.125, + "learning_rate": 2.438195988760786e-06, + "loss": 0.7602, + "step": 2449 + }, + { + "epoch": 0.2087771623348956, + "grad_norm": 12.5, + "learning_rate": 2.438143794453666e-06, + "loss": 0.5734, + "step": 2450 + }, + { + "epoch": 0.20886237750319556, + "grad_norm": 17.75, + "learning_rate": 2.438091578675593e-06, + "loss": 1.0439, + "step": 2451 + }, + { + "epoch": 0.2089475926714955, + "grad_norm": 17.25, + "learning_rate": 2.4380393414275107e-06, + "loss": 1.0062, + "step": 2452 + }, + { + "epoch": 0.20903280783979547, + "grad_norm": 15.9375, + "learning_rate": 2.4379870827103626e-06, + "loss": 0.8711, + "step": 2453 + }, + { + "epoch": 0.20911802300809543, + "grad_norm": 11.375, + "learning_rate": 2.4379348025250933e-06, + "loss": 0.5635, + "step": 2454 + }, + { + "epoch": 0.2092032381763954, + "grad_norm": 14.5625, + "learning_rate": 2.4378825008726473e-06, + "loss": 0.8082, + "step": 2455 + }, + { + "epoch": 0.20928845334469537, + "grad_norm": 14.3125, + "learning_rate": 2.43783017775397e-06, + "loss": 0.7393, + "step": 2456 + }, + { + "epoch": 0.20937366851299533, + "grad_norm": 16.25, + "learning_rate": 2.4377778331700077e-06, + "loss": 0.7298, + "step": 2457 + }, + { + "epoch": 0.20945888368129528, + "grad_norm": 24.375, + "learning_rate": 2.437725467121705e-06, + "loss": 1.121, + "step": 2458 + }, + { + "epoch": 0.20954409884959524, + "grad_norm": 14.5, + "learning_rate": 2.4376730796100084e-06, + "loss": 0.6564, + "step": 2459 + }, + { + "epoch": 0.2096293140178952, + "grad_norm": 11.4375, + "learning_rate": 2.437620670635865e-06, + "loss": 0.6601, + "step": 2460 + }, + { + "epoch": 0.20971452918619515, + "grad_norm": 14.6875, + "learning_rate": 2.4375682402002227e-06, + "loss": 0.8472, + "step": 2461 + }, + { + "epoch": 0.2097997443544951, + "grad_norm": 11.8125, + "learning_rate": 2.4375157883040273e-06, + "loss": 0.5964, + "step": 2462 + }, + { + "epoch": 0.20988495952279507, + "grad_norm": 20.625, + "learning_rate": 2.437463314948228e-06, + "loss": 1.0367, + "step": 2463 + }, + { + "epoch": 0.20997017469109502, + "grad_norm": 20.25, + "learning_rate": 2.437410820133772e-06, + "loss": 1.0741, + "step": 2464 + }, + { + "epoch": 0.21005538985939498, + "grad_norm": 21.75, + "learning_rate": 2.437358303861609e-06, + "loss": 1.0369, + "step": 2465 + }, + { + "epoch": 0.21014060502769494, + "grad_norm": 17.875, + "learning_rate": 2.437305766132687e-06, + "loss": 0.7807, + "step": 2466 + }, + { + "epoch": 0.2102258201959949, + "grad_norm": 12.875, + "learning_rate": 2.437253206947956e-06, + "loss": 0.7198, + "step": 2467 + }, + { + "epoch": 0.21031103536429485, + "grad_norm": 17.75, + "learning_rate": 2.437200626308366e-06, + "loss": 0.6326, + "step": 2468 + }, + { + "epoch": 0.2103962505325948, + "grad_norm": 11.5, + "learning_rate": 2.437148024214866e-06, + "loss": 0.5773, + "step": 2469 + }, + { + "epoch": 0.21048146570089477, + "grad_norm": 14.8125, + "learning_rate": 2.4370954006684083e-06, + "loss": 0.8714, + "step": 2470 + }, + { + "epoch": 0.21056668086919472, + "grad_norm": 16.375, + "learning_rate": 2.437042755669943e-06, + "loss": 0.79, + "step": 2471 + }, + { + "epoch": 0.21065189603749468, + "grad_norm": 14.0, + "learning_rate": 2.4369900892204205e-06, + "loss": 0.7128, + "step": 2472 + }, + { + "epoch": 0.21073711120579464, + "grad_norm": 25.75, + "learning_rate": 2.4369374013207946e-06, + "loss": 0.9695, + "step": 2473 + }, + { + "epoch": 0.2108223263740946, + "grad_norm": 13.0, + "learning_rate": 2.4368846919720156e-06, + "loss": 0.6582, + "step": 2474 + }, + { + "epoch": 0.21090754154239455, + "grad_norm": 11.625, + "learning_rate": 2.4368319611750367e-06, + "loss": 0.5139, + "step": 2475 + }, + { + "epoch": 0.2109927567106945, + "grad_norm": 13.5625, + "learning_rate": 2.4367792089308114e-06, + "loss": 0.8152, + "step": 2476 + }, + { + "epoch": 0.21107797187899446, + "grad_norm": 16.0, + "learning_rate": 2.436726435240292e-06, + "loss": 0.6635, + "step": 2477 + }, + { + "epoch": 0.21116318704729442, + "grad_norm": 15.3125, + "learning_rate": 2.436673640104433e-06, + "loss": 0.5012, + "step": 2478 + }, + { + "epoch": 0.21124840221559438, + "grad_norm": 18.375, + "learning_rate": 2.436620823524187e-06, + "loss": 0.9702, + "step": 2479 + }, + { + "epoch": 0.21133361738389433, + "grad_norm": 13.0625, + "learning_rate": 2.4365679855005104e-06, + "loss": 0.5865, + "step": 2480 + }, + { + "epoch": 0.2114188325521943, + "grad_norm": 14.125, + "learning_rate": 2.436515126034357e-06, + "loss": 0.8313, + "step": 2481 + }, + { + "epoch": 0.21150404772049425, + "grad_norm": 14.75, + "learning_rate": 2.436462245126682e-06, + "loss": 0.4818, + "step": 2482 + }, + { + "epoch": 0.2115892628887942, + "grad_norm": 11.1875, + "learning_rate": 2.436409342778441e-06, + "loss": 0.8752, + "step": 2483 + }, + { + "epoch": 0.21167447805709416, + "grad_norm": 11.25, + "learning_rate": 2.4363564189905904e-06, + "loss": 0.3124, + "step": 2484 + }, + { + "epoch": 0.21175969322539412, + "grad_norm": 17.75, + "learning_rate": 2.436303473764086e-06, + "loss": 0.8358, + "step": 2485 + }, + { + "epoch": 0.21184490839369408, + "grad_norm": 9.75, + "learning_rate": 2.4362505070998847e-06, + "loss": 0.2835, + "step": 2486 + }, + { + "epoch": 0.21193012356199403, + "grad_norm": 15.4375, + "learning_rate": 2.4361975189989447e-06, + "loss": 0.9199, + "step": 2487 + }, + { + "epoch": 0.212015338730294, + "grad_norm": 18.125, + "learning_rate": 2.4361445094622223e-06, + "loss": 0.7091, + "step": 2488 + }, + { + "epoch": 0.21210055389859395, + "grad_norm": 13.9375, + "learning_rate": 2.4360914784906757e-06, + "loss": 0.8237, + "step": 2489 + }, + { + "epoch": 0.2121857690668939, + "grad_norm": 15.1875, + "learning_rate": 2.436038426085264e-06, + "loss": 0.9451, + "step": 2490 + }, + { + "epoch": 0.21227098423519386, + "grad_norm": 12.375, + "learning_rate": 2.4359853522469447e-06, + "loss": 0.7429, + "step": 2491 + }, + { + "epoch": 0.21235619940349382, + "grad_norm": 16.125, + "learning_rate": 2.4359322569766776e-06, + "loss": 0.9901, + "step": 2492 + }, + { + "epoch": 0.21244141457179377, + "grad_norm": 16.5, + "learning_rate": 2.4358791402754216e-06, + "loss": 0.5445, + "step": 2493 + }, + { + "epoch": 0.21252662974009373, + "grad_norm": 11.75, + "learning_rate": 2.4358260021441376e-06, + "loss": 0.662, + "step": 2494 + }, + { + "epoch": 0.2126118449083937, + "grad_norm": 13.75, + "learning_rate": 2.4357728425837853e-06, + "loss": 0.4285, + "step": 2495 + }, + { + "epoch": 0.21269706007669364, + "grad_norm": 13.875, + "learning_rate": 2.4357196615953256e-06, + "loss": 0.8422, + "step": 2496 + }, + { + "epoch": 0.2127822752449936, + "grad_norm": 16.25, + "learning_rate": 2.4356664591797186e-06, + "loss": 0.5245, + "step": 2497 + }, + { + "epoch": 0.21286749041329356, + "grad_norm": 13.5625, + "learning_rate": 2.435613235337927e-06, + "loss": 1.0852, + "step": 2498 + }, + { + "epoch": 0.21295270558159352, + "grad_norm": 16.875, + "learning_rate": 2.4355599900709114e-06, + "loss": 0.3149, + "step": 2499 + }, + { + "epoch": 0.21303792074989347, + "grad_norm": 18.75, + "learning_rate": 2.4355067233796352e-06, + "loss": 1.0901, + "step": 2500 + }, + { + "epoch": 0.21312313591819343, + "grad_norm": 14.5, + "learning_rate": 2.43545343526506e-06, + "loss": 0.8986, + "step": 2501 + }, + { + "epoch": 0.2132083510864934, + "grad_norm": 18.125, + "learning_rate": 2.4354001257281497e-06, + "loss": 0.8688, + "step": 2502 + }, + { + "epoch": 0.21329356625479334, + "grad_norm": 19.375, + "learning_rate": 2.4353467947698665e-06, + "loss": 1.086, + "step": 2503 + }, + { + "epoch": 0.2133787814230933, + "grad_norm": 24.375, + "learning_rate": 2.435293442391175e-06, + "loss": 1.2636, + "step": 2504 + }, + { + "epoch": 0.21346399659139326, + "grad_norm": 13.9375, + "learning_rate": 2.4352400685930388e-06, + "loss": 0.5842, + "step": 2505 + }, + { + "epoch": 0.2135492117596932, + "grad_norm": 11.125, + "learning_rate": 2.435186673376423e-06, + "loss": 0.5385, + "step": 2506 + }, + { + "epoch": 0.21363442692799317, + "grad_norm": 17.5, + "learning_rate": 2.4351332567422924e-06, + "loss": 0.6024, + "step": 2507 + }, + { + "epoch": 0.21371964209629313, + "grad_norm": 12.5, + "learning_rate": 2.4350798186916118e-06, + "loss": 0.6682, + "step": 2508 + }, + { + "epoch": 0.21380485726459308, + "grad_norm": 13.8125, + "learning_rate": 2.4350263592253476e-06, + "loss": 1.0078, + "step": 2509 + }, + { + "epoch": 0.21389007243289304, + "grad_norm": 15.1875, + "learning_rate": 2.4349728783444646e-06, + "loss": 0.6571, + "step": 2510 + }, + { + "epoch": 0.21397528760119303, + "grad_norm": 11.9375, + "learning_rate": 2.434919376049931e-06, + "loss": 0.5937, + "step": 2511 + }, + { + "epoch": 0.21406050276949298, + "grad_norm": 13.5625, + "learning_rate": 2.4348658523427125e-06, + "loss": 0.7072, + "step": 2512 + }, + { + "epoch": 0.21414571793779294, + "grad_norm": 18.75, + "learning_rate": 2.4348123072237766e-06, + "loss": 0.8007, + "step": 2513 + }, + { + "epoch": 0.2142309331060929, + "grad_norm": 15.8125, + "learning_rate": 2.4347587406940908e-06, + "loss": 0.8709, + "step": 2514 + }, + { + "epoch": 0.21431614827439285, + "grad_norm": 13.125, + "learning_rate": 2.4347051527546227e-06, + "loss": 0.9103, + "step": 2515 + }, + { + "epoch": 0.2144013634426928, + "grad_norm": 11.625, + "learning_rate": 2.434651543406342e-06, + "loss": 0.6602, + "step": 2516 + }, + { + "epoch": 0.21448657861099277, + "grad_norm": 15.625, + "learning_rate": 2.4345979126502164e-06, + "loss": 0.7888, + "step": 2517 + }, + { + "epoch": 0.21457179377929272, + "grad_norm": 13.125, + "learning_rate": 2.434544260487215e-06, + "loss": 0.6301, + "step": 2518 + }, + { + "epoch": 0.21465700894759268, + "grad_norm": 17.25, + "learning_rate": 2.434490586918308e-06, + "loss": 1.1467, + "step": 2519 + }, + { + "epoch": 0.21474222411589264, + "grad_norm": 20.0, + "learning_rate": 2.4344368919444646e-06, + "loss": 0.903, + "step": 2520 + }, + { + "epoch": 0.2148274392841926, + "grad_norm": 13.625, + "learning_rate": 2.4343831755666557e-06, + "loss": 0.7378, + "step": 2521 + }, + { + "epoch": 0.21491265445249255, + "grad_norm": 13.6875, + "learning_rate": 2.434329437785852e-06, + "loss": 0.7339, + "step": 2522 + }, + { + "epoch": 0.2149978696207925, + "grad_norm": 20.875, + "learning_rate": 2.4342756786030246e-06, + "loss": 1.2273, + "step": 2523 + }, + { + "epoch": 0.21508308478909247, + "grad_norm": 10.0625, + "learning_rate": 2.434221898019144e-06, + "loss": 0.4503, + "step": 2524 + }, + { + "epoch": 0.21516829995739242, + "grad_norm": 13.9375, + "learning_rate": 2.434168096035184e-06, + "loss": 0.5745, + "step": 2525 + }, + { + "epoch": 0.21525351512569238, + "grad_norm": 15.0625, + "learning_rate": 2.434114272652115e-06, + "loss": 1.0181, + "step": 2526 + }, + { + "epoch": 0.21533873029399234, + "grad_norm": 14.1875, + "learning_rate": 2.43406042787091e-06, + "loss": 0.8084, + "step": 2527 + }, + { + "epoch": 0.2154239454622923, + "grad_norm": 14.625, + "learning_rate": 2.4340065616925433e-06, + "loss": 0.8192, + "step": 2528 + }, + { + "epoch": 0.21550916063059225, + "grad_norm": 12.875, + "learning_rate": 2.433952674117987e-06, + "loss": 0.7902, + "step": 2529 + }, + { + "epoch": 0.2155943757988922, + "grad_norm": 15.3125, + "learning_rate": 2.4338987651482153e-06, + "loss": 0.904, + "step": 2530 + }, + { + "epoch": 0.21567959096719216, + "grad_norm": 18.125, + "learning_rate": 2.4338448347842027e-06, + "loss": 0.8997, + "step": 2531 + }, + { + "epoch": 0.21576480613549212, + "grad_norm": 15.5625, + "learning_rate": 2.433790883026923e-06, + "loss": 0.7253, + "step": 2532 + }, + { + "epoch": 0.21585002130379208, + "grad_norm": 15.1875, + "learning_rate": 2.433736909877352e-06, + "loss": 0.9365, + "step": 2533 + }, + { + "epoch": 0.21593523647209203, + "grad_norm": 16.75, + "learning_rate": 2.433682915336465e-06, + "loss": 0.9074, + "step": 2534 + }, + { + "epoch": 0.216020451640392, + "grad_norm": 10.875, + "learning_rate": 2.4336288994052364e-06, + "loss": 0.359, + "step": 2535 + }, + { + "epoch": 0.21610566680869195, + "grad_norm": 14.1875, + "learning_rate": 2.4335748620846444e-06, + "loss": 0.6362, + "step": 2536 + }, + { + "epoch": 0.2161908819769919, + "grad_norm": 11.5625, + "learning_rate": 2.433520803375664e-06, + "loss": 0.418, + "step": 2537 + }, + { + "epoch": 0.21627609714529186, + "grad_norm": 14.0, + "learning_rate": 2.4334667232792724e-06, + "loss": 0.7132, + "step": 2538 + }, + { + "epoch": 0.21636131231359182, + "grad_norm": 13.6875, + "learning_rate": 2.4334126217964475e-06, + "loss": 0.5964, + "step": 2539 + }, + { + "epoch": 0.21644652748189178, + "grad_norm": 26.75, + "learning_rate": 2.4333584989281663e-06, + "loss": 0.6838, + "step": 2540 + }, + { + "epoch": 0.21653174265019173, + "grad_norm": 18.875, + "learning_rate": 2.433304354675407e-06, + "loss": 0.9192, + "step": 2541 + }, + { + "epoch": 0.2166169578184917, + "grad_norm": 19.375, + "learning_rate": 2.433250189039148e-06, + "loss": 0.9965, + "step": 2542 + }, + { + "epoch": 0.21670217298679165, + "grad_norm": 13.625, + "learning_rate": 2.4331960020203686e-06, + "loss": 0.4586, + "step": 2543 + }, + { + "epoch": 0.2167873881550916, + "grad_norm": 21.625, + "learning_rate": 2.4331417936200475e-06, + "loss": 0.6214, + "step": 2544 + }, + { + "epoch": 0.21687260332339156, + "grad_norm": 13.375, + "learning_rate": 2.4330875638391644e-06, + "loss": 0.5138, + "step": 2545 + }, + { + "epoch": 0.21695781849169152, + "grad_norm": 11.0625, + "learning_rate": 2.433033312678699e-06, + "loss": 0.6062, + "step": 2546 + }, + { + "epoch": 0.21704303365999147, + "grad_norm": 28.125, + "learning_rate": 2.4329790401396326e-06, + "loss": 0.8695, + "step": 2547 + }, + { + "epoch": 0.21712824882829143, + "grad_norm": 38.75, + "learning_rate": 2.432924746222945e-06, + "loss": 1.2791, + "step": 2548 + }, + { + "epoch": 0.2172134639965914, + "grad_norm": 16.25, + "learning_rate": 2.432870430929618e-06, + "loss": 0.8814, + "step": 2549 + }, + { + "epoch": 0.21729867916489135, + "grad_norm": 15.375, + "learning_rate": 2.4328160942606324e-06, + "loss": 0.7691, + "step": 2550 + }, + { + "epoch": 0.2173838943331913, + "grad_norm": 28.0, + "learning_rate": 2.432761736216971e-06, + "loss": 0.7908, + "step": 2551 + }, + { + "epoch": 0.21746910950149126, + "grad_norm": 21.375, + "learning_rate": 2.4327073567996152e-06, + "loss": 0.9392, + "step": 2552 + }, + { + "epoch": 0.21755432466979122, + "grad_norm": 11.9375, + "learning_rate": 2.4326529560095486e-06, + "loss": 0.6047, + "step": 2553 + }, + { + "epoch": 0.21763953983809117, + "grad_norm": 11.8125, + "learning_rate": 2.4325985338477537e-06, + "loss": 0.4832, + "step": 2554 + }, + { + "epoch": 0.21772475500639113, + "grad_norm": 11.9375, + "learning_rate": 2.432544090315214e-06, + "loss": 0.4691, + "step": 2555 + }, + { + "epoch": 0.2178099701746911, + "grad_norm": 20.125, + "learning_rate": 2.4324896254129137e-06, + "loss": 0.727, + "step": 2556 + }, + { + "epoch": 0.21789518534299104, + "grad_norm": 16.125, + "learning_rate": 2.4324351391418364e-06, + "loss": 0.7773, + "step": 2557 + }, + { + "epoch": 0.217980400511291, + "grad_norm": 18.5, + "learning_rate": 2.4323806315029673e-06, + "loss": 1.0957, + "step": 2558 + }, + { + "epoch": 0.21806561567959096, + "grad_norm": 19.125, + "learning_rate": 2.432326102497291e-06, + "loss": 1.3076, + "step": 2559 + }, + { + "epoch": 0.21815083084789091, + "grad_norm": 12.4375, + "learning_rate": 2.4322715521257934e-06, + "loss": 0.6208, + "step": 2560 + }, + { + "epoch": 0.21823604601619087, + "grad_norm": 16.0, + "learning_rate": 2.43221698038946e-06, + "loss": 0.8164, + "step": 2561 + }, + { + "epoch": 0.21832126118449083, + "grad_norm": 11.125, + "learning_rate": 2.432162387289277e-06, + "loss": 0.4064, + "step": 2562 + }, + { + "epoch": 0.21840647635279078, + "grad_norm": 12.0625, + "learning_rate": 2.432107772826231e-06, + "loss": 0.3819, + "step": 2563 + }, + { + "epoch": 0.21849169152109074, + "grad_norm": 16.875, + "learning_rate": 2.4320531370013084e-06, + "loss": 0.904, + "step": 2564 + }, + { + "epoch": 0.2185769066893907, + "grad_norm": 14.375, + "learning_rate": 2.4319984798154973e-06, + "loss": 0.8427, + "step": 2565 + }, + { + "epoch": 0.21866212185769066, + "grad_norm": 13.9375, + "learning_rate": 2.431943801269785e-06, + "loss": 0.6886, + "step": 2566 + }, + { + "epoch": 0.2187473370259906, + "grad_norm": 17.0, + "learning_rate": 2.4318891013651594e-06, + "loss": 0.8962, + "step": 2567 + }, + { + "epoch": 0.2188325521942906, + "grad_norm": 9.3125, + "learning_rate": 2.4318343801026095e-06, + "loss": 0.3265, + "step": 2568 + }, + { + "epoch": 0.21891776736259055, + "grad_norm": 18.625, + "learning_rate": 2.4317796374831238e-06, + "loss": 0.957, + "step": 2569 + }, + { + "epoch": 0.2190029825308905, + "grad_norm": 13.9375, + "learning_rate": 2.431724873507692e-06, + "loss": 0.8619, + "step": 2570 + }, + { + "epoch": 0.21908819769919047, + "grad_norm": 11.5625, + "learning_rate": 2.431670088177303e-06, + "loss": 0.507, + "step": 2571 + }, + { + "epoch": 0.21917341286749042, + "grad_norm": 17.25, + "learning_rate": 2.4316152814929478e-06, + "loss": 0.7112, + "step": 2572 + }, + { + "epoch": 0.21925862803579038, + "grad_norm": 13.125, + "learning_rate": 2.431560453455616e-06, + "loss": 0.4879, + "step": 2573 + }, + { + "epoch": 0.21934384320409034, + "grad_norm": 13.6875, + "learning_rate": 2.431505604066298e-06, + "loss": 0.8353, + "step": 2574 + }, + { + "epoch": 0.2194290583723903, + "grad_norm": 16.625, + "learning_rate": 2.4314507333259864e-06, + "loss": 0.9581, + "step": 2575 + }, + { + "epoch": 0.21951427354069025, + "grad_norm": 17.5, + "learning_rate": 2.431395841235672e-06, + "loss": 0.9309, + "step": 2576 + }, + { + "epoch": 0.2195994887089902, + "grad_norm": 19.75, + "learning_rate": 2.4313409277963463e-06, + "loss": 0.7396, + "step": 2577 + }, + { + "epoch": 0.21968470387729017, + "grad_norm": 15.0, + "learning_rate": 2.4312859930090025e-06, + "loss": 0.8397, + "step": 2578 + }, + { + "epoch": 0.21976991904559012, + "grad_norm": 16.25, + "learning_rate": 2.4312310368746327e-06, + "loss": 0.9028, + "step": 2579 + }, + { + "epoch": 0.21985513421389008, + "grad_norm": 13.625, + "learning_rate": 2.431176059394231e-06, + "loss": 0.8091, + "step": 2580 + }, + { + "epoch": 0.21994034938219004, + "grad_norm": 18.0, + "learning_rate": 2.431121060568789e-06, + "loss": 0.96, + "step": 2581 + }, + { + "epoch": 0.22002556455049, + "grad_norm": 15.8125, + "learning_rate": 2.4310660403993023e-06, + "loss": 0.7458, + "step": 2582 + }, + { + "epoch": 0.22011077971878995, + "grad_norm": 11.5625, + "learning_rate": 2.4310109988867647e-06, + "loss": 0.5909, + "step": 2583 + }, + { + "epoch": 0.2201959948870899, + "grad_norm": 13.375, + "learning_rate": 2.4309559360321707e-06, + "loss": 0.7709, + "step": 2584 + }, + { + "epoch": 0.22028121005538986, + "grad_norm": 17.0, + "learning_rate": 2.430900851836515e-06, + "loss": 0.7266, + "step": 2585 + }, + { + "epoch": 0.22036642522368982, + "grad_norm": 14.125, + "learning_rate": 2.430845746300794e-06, + "loss": 0.3861, + "step": 2586 + }, + { + "epoch": 0.22045164039198978, + "grad_norm": 17.375, + "learning_rate": 2.430790619426002e-06, + "loss": 1.3531, + "step": 2587 + }, + { + "epoch": 0.22053685556028974, + "grad_norm": 14.0625, + "learning_rate": 2.430735471213137e-06, + "loss": 0.7736, + "step": 2588 + }, + { + "epoch": 0.2206220707285897, + "grad_norm": 11.625, + "learning_rate": 2.4306803016631945e-06, + "loss": 0.5715, + "step": 2589 + }, + { + "epoch": 0.22070728589688965, + "grad_norm": 18.75, + "learning_rate": 2.4306251107771717e-06, + "loss": 0.9396, + "step": 2590 + }, + { + "epoch": 0.2207925010651896, + "grad_norm": 13.8125, + "learning_rate": 2.430569898556066e-06, + "loss": 0.821, + "step": 2591 + }, + { + "epoch": 0.22087771623348956, + "grad_norm": 16.25, + "learning_rate": 2.430514665000875e-06, + "loss": 0.9442, + "step": 2592 + }, + { + "epoch": 0.22096293140178952, + "grad_norm": 14.5, + "learning_rate": 2.4304594101125967e-06, + "loss": 1.0966, + "step": 2593 + }, + { + "epoch": 0.22104814657008948, + "grad_norm": 13.125, + "learning_rate": 2.4304041338922303e-06, + "loss": 0.6438, + "step": 2594 + }, + { + "epoch": 0.22113336173838943, + "grad_norm": 21.25, + "learning_rate": 2.4303488363407736e-06, + "loss": 1.0892, + "step": 2595 + }, + { + "epoch": 0.2212185769066894, + "grad_norm": 22.5, + "learning_rate": 2.430293517459227e-06, + "loss": 0.9584, + "step": 2596 + }, + { + "epoch": 0.22130379207498935, + "grad_norm": 14.75, + "learning_rate": 2.4302381772485893e-06, + "loss": 1.055, + "step": 2597 + }, + { + "epoch": 0.2213890072432893, + "grad_norm": 15.25, + "learning_rate": 2.430182815709861e-06, + "loss": 1.1028, + "step": 2598 + }, + { + "epoch": 0.22147422241158926, + "grad_norm": 13.4375, + "learning_rate": 2.4301274328440427e-06, + "loss": 0.7913, + "step": 2599 + }, + { + "epoch": 0.22155943757988922, + "grad_norm": 14.25, + "learning_rate": 2.430072028652135e-06, + "loss": 0.5136, + "step": 2600 + }, + { + "epoch": 0.22164465274818917, + "grad_norm": 15.4375, + "learning_rate": 2.4300166031351384e-06, + "loss": 0.7976, + "step": 2601 + }, + { + "epoch": 0.22172986791648913, + "grad_norm": 15.0, + "learning_rate": 2.429961156294056e-06, + "loss": 0.7175, + "step": 2602 + }, + { + "epoch": 0.2218150830847891, + "grad_norm": 23.125, + "learning_rate": 2.429905688129889e-06, + "loss": 1.0837, + "step": 2603 + }, + { + "epoch": 0.22190029825308905, + "grad_norm": 23.25, + "learning_rate": 2.429850198643639e-06, + "loss": 0.8739, + "step": 2604 + }, + { + "epoch": 0.221985513421389, + "grad_norm": 15.9375, + "learning_rate": 2.42979468783631e-06, + "loss": 0.6373, + "step": 2605 + }, + { + "epoch": 0.22207072858968896, + "grad_norm": 15.4375, + "learning_rate": 2.4297391557089044e-06, + "loss": 0.7459, + "step": 2606 + }, + { + "epoch": 0.22215594375798892, + "grad_norm": 16.125, + "learning_rate": 2.429683602262426e-06, + "loss": 1.026, + "step": 2607 + }, + { + "epoch": 0.22224115892628887, + "grad_norm": 14.5625, + "learning_rate": 2.429628027497879e-06, + "loss": 0.6947, + "step": 2608 + }, + { + "epoch": 0.22232637409458883, + "grad_norm": 11.5625, + "learning_rate": 2.4295724314162668e-06, + "loss": 0.637, + "step": 2609 + }, + { + "epoch": 0.2224115892628888, + "grad_norm": 16.125, + "learning_rate": 2.4295168140185945e-06, + "loss": 1.2364, + "step": 2610 + }, + { + "epoch": 0.22249680443118874, + "grad_norm": 14.3125, + "learning_rate": 2.429461175305868e-06, + "loss": 0.5258, + "step": 2611 + }, + { + "epoch": 0.2225820195994887, + "grad_norm": 17.75, + "learning_rate": 2.429405515279091e-06, + "loss": 0.9123, + "step": 2612 + }, + { + "epoch": 0.22266723476778866, + "grad_norm": 18.625, + "learning_rate": 2.4293498339392715e-06, + "loss": 0.5855, + "step": 2613 + }, + { + "epoch": 0.22275244993608861, + "grad_norm": 19.625, + "learning_rate": 2.4292941312874137e-06, + "loss": 0.913, + "step": 2614 + }, + { + "epoch": 0.22283766510438857, + "grad_norm": 12.875, + "learning_rate": 2.4292384073245255e-06, + "loss": 0.5384, + "step": 2615 + }, + { + "epoch": 0.22292288027268853, + "grad_norm": 16.0, + "learning_rate": 2.4291826620516135e-06, + "loss": 0.9442, + "step": 2616 + }, + { + "epoch": 0.22300809544098849, + "grad_norm": 19.75, + "learning_rate": 2.429126895469685e-06, + "loss": 0.6352, + "step": 2617 + }, + { + "epoch": 0.22309331060928844, + "grad_norm": 11.375, + "learning_rate": 2.4290711075797475e-06, + "loss": 0.6309, + "step": 2618 + }, + { + "epoch": 0.2231785257775884, + "grad_norm": 14.75, + "learning_rate": 2.4290152983828095e-06, + "loss": 0.4835, + "step": 2619 + }, + { + "epoch": 0.22326374094588836, + "grad_norm": 15.375, + "learning_rate": 2.4289594678798796e-06, + "loss": 0.6368, + "step": 2620 + }, + { + "epoch": 0.2233489561141883, + "grad_norm": 15.5625, + "learning_rate": 2.4289036160719666e-06, + "loss": 0.9495, + "step": 2621 + }, + { + "epoch": 0.22343417128248827, + "grad_norm": 18.625, + "learning_rate": 2.4288477429600798e-06, + "loss": 0.4465, + "step": 2622 + }, + { + "epoch": 0.22351938645078823, + "grad_norm": 11.0625, + "learning_rate": 2.4287918485452288e-06, + "loss": 0.4655, + "step": 2623 + }, + { + "epoch": 0.2236046016190882, + "grad_norm": 16.125, + "learning_rate": 2.4287359328284237e-06, + "loss": 0.9579, + "step": 2624 + }, + { + "epoch": 0.22368981678738817, + "grad_norm": 19.0, + "learning_rate": 2.428679995810675e-06, + "loss": 0.717, + "step": 2625 + }, + { + "epoch": 0.22377503195568813, + "grad_norm": 23.125, + "learning_rate": 2.4286240374929937e-06, + "loss": 1.1671, + "step": 2626 + }, + { + "epoch": 0.22386024712398808, + "grad_norm": 16.0, + "learning_rate": 2.4285680578763907e-06, + "loss": 0.914, + "step": 2627 + }, + { + "epoch": 0.22394546229228804, + "grad_norm": 25.25, + "learning_rate": 2.4285120569618777e-06, + "loss": 1.0151, + "step": 2628 + }, + { + "epoch": 0.224030677460588, + "grad_norm": 64.5, + "learning_rate": 2.4284560347504666e-06, + "loss": 0.6493, + "step": 2629 + }, + { + "epoch": 0.22411589262888795, + "grad_norm": 17.375, + "learning_rate": 2.4283999912431704e-06, + "loss": 0.8151, + "step": 2630 + }, + { + "epoch": 0.2242011077971879, + "grad_norm": 15.9375, + "learning_rate": 2.428343926441001e-06, + "loss": 0.7265, + "step": 2631 + }, + { + "epoch": 0.22428632296548787, + "grad_norm": 19.0, + "learning_rate": 2.4282878403449724e-06, + "loss": 0.7539, + "step": 2632 + }, + { + "epoch": 0.22437153813378782, + "grad_norm": 15.75, + "learning_rate": 2.428231732956097e-06, + "loss": 0.8051, + "step": 2633 + }, + { + "epoch": 0.22445675330208778, + "grad_norm": 15.3125, + "learning_rate": 2.4281756042753894e-06, + "loss": 1.016, + "step": 2634 + }, + { + "epoch": 0.22454196847038774, + "grad_norm": 12.6875, + "learning_rate": 2.4281194543038643e-06, + "loss": 0.4299, + "step": 2635 + }, + { + "epoch": 0.2246271836386877, + "grad_norm": 14.25, + "learning_rate": 2.428063283042536e-06, + "loss": 0.762, + "step": 2636 + }, + { + "epoch": 0.22471239880698765, + "grad_norm": 17.125, + "learning_rate": 2.428007090492419e-06, + "loss": 0.9848, + "step": 2637 + }, + { + "epoch": 0.2247976139752876, + "grad_norm": 12.6875, + "learning_rate": 2.4279508766545303e-06, + "loss": 0.5421, + "step": 2638 + }, + { + "epoch": 0.22488282914358756, + "grad_norm": 13.8125, + "learning_rate": 2.427894641529884e-06, + "loss": 0.6195, + "step": 2639 + }, + { + "epoch": 0.22496804431188752, + "grad_norm": 12.75, + "learning_rate": 2.427838385119497e-06, + "loss": 0.6019, + "step": 2640 + }, + { + "epoch": 0.22505325948018748, + "grad_norm": 13.0, + "learning_rate": 2.4277821074243862e-06, + "loss": 0.4765, + "step": 2641 + }, + { + "epoch": 0.22513847464848744, + "grad_norm": 16.625, + "learning_rate": 2.4277258084455682e-06, + "loss": 0.8688, + "step": 2642 + }, + { + "epoch": 0.2252236898167874, + "grad_norm": 13.8125, + "learning_rate": 2.4276694881840606e-06, + "loss": 0.6365, + "step": 2643 + }, + { + "epoch": 0.22530890498508735, + "grad_norm": 20.0, + "learning_rate": 2.4276131466408815e-06, + "loss": 1.0313, + "step": 2644 + }, + { + "epoch": 0.2253941201533873, + "grad_norm": 14.4375, + "learning_rate": 2.4275567838170483e-06, + "loss": 0.6653, + "step": 2645 + }, + { + "epoch": 0.22547933532168726, + "grad_norm": 11.9375, + "learning_rate": 2.42750039971358e-06, + "loss": 0.5904, + "step": 2646 + }, + { + "epoch": 0.22556455048998722, + "grad_norm": 14.0, + "learning_rate": 2.427443994331495e-06, + "loss": 0.6693, + "step": 2647 + }, + { + "epoch": 0.22564976565828718, + "grad_norm": 13.875, + "learning_rate": 2.4273875676718133e-06, + "loss": 0.5469, + "step": 2648 + }, + { + "epoch": 0.22573498082658713, + "grad_norm": 16.375, + "learning_rate": 2.427331119735554e-06, + "loss": 1.003, + "step": 2649 + }, + { + "epoch": 0.2258201959948871, + "grad_norm": 12.6875, + "learning_rate": 2.427274650523738e-06, + "loss": 0.4572, + "step": 2650 + }, + { + "epoch": 0.22590541116318705, + "grad_norm": 19.625, + "learning_rate": 2.427218160037385e-06, + "loss": 0.9576, + "step": 2651 + }, + { + "epoch": 0.225990626331487, + "grad_norm": 13.6875, + "learning_rate": 2.4271616482775164e-06, + "loss": 0.6773, + "step": 2652 + }, + { + "epoch": 0.22607584149978696, + "grad_norm": 13.5, + "learning_rate": 2.4271051152451526e-06, + "loss": 0.5406, + "step": 2653 + }, + { + "epoch": 0.22616105666808692, + "grad_norm": 14.25, + "learning_rate": 2.427048560941316e-06, + "loss": 0.622, + "step": 2654 + }, + { + "epoch": 0.22624627183638688, + "grad_norm": 17.875, + "learning_rate": 2.426991985367028e-06, + "loss": 1.1171, + "step": 2655 + }, + { + "epoch": 0.22633148700468683, + "grad_norm": 16.375, + "learning_rate": 2.4269353885233114e-06, + "loss": 1.0306, + "step": 2656 + }, + { + "epoch": 0.2264167021729868, + "grad_norm": 10.375, + "learning_rate": 2.426878770411189e-06, + "loss": 0.4994, + "step": 2657 + }, + { + "epoch": 0.22650191734128675, + "grad_norm": 10.9375, + "learning_rate": 2.426822131031684e-06, + "loss": 0.4089, + "step": 2658 + }, + { + "epoch": 0.2265871325095867, + "grad_norm": 16.625, + "learning_rate": 2.426765470385819e-06, + "loss": 0.9012, + "step": 2659 + }, + { + "epoch": 0.22667234767788666, + "grad_norm": 21.375, + "learning_rate": 2.426708788474619e-06, + "loss": 1.65, + "step": 2660 + }, + { + "epoch": 0.22675756284618662, + "grad_norm": 19.5, + "learning_rate": 2.426652085299108e-06, + "loss": 1.0165, + "step": 2661 + }, + { + "epoch": 0.22684277801448657, + "grad_norm": 13.1875, + "learning_rate": 2.4265953608603105e-06, + "loss": 0.6429, + "step": 2662 + }, + { + "epoch": 0.22692799318278653, + "grad_norm": 12.4375, + "learning_rate": 2.426538615159252e-06, + "loss": 0.5751, + "step": 2663 + }, + { + "epoch": 0.2270132083510865, + "grad_norm": 16.0, + "learning_rate": 2.426481848196957e-06, + "loss": 1.1199, + "step": 2664 + }, + { + "epoch": 0.22709842351938644, + "grad_norm": 15.0625, + "learning_rate": 2.4264250599744522e-06, + "loss": 0.5531, + "step": 2665 + }, + { + "epoch": 0.2271836386876864, + "grad_norm": 13.8125, + "learning_rate": 2.4263682504927635e-06, + "loss": 0.8152, + "step": 2666 + }, + { + "epoch": 0.22726885385598636, + "grad_norm": 43.75, + "learning_rate": 2.426311419752918e-06, + "loss": 1.0729, + "step": 2667 + }, + { + "epoch": 0.22735406902428631, + "grad_norm": 14.9375, + "learning_rate": 2.4262545677559418e-06, + "loss": 0.9786, + "step": 2668 + }, + { + "epoch": 0.22743928419258627, + "grad_norm": 14.875, + "learning_rate": 2.426197694502863e-06, + "loss": 0.7922, + "step": 2669 + }, + { + "epoch": 0.22752449936088623, + "grad_norm": 10.6875, + "learning_rate": 2.426140799994709e-06, + "loss": 0.4565, + "step": 2670 + }, + { + "epoch": 0.22760971452918619, + "grad_norm": 15.1875, + "learning_rate": 2.426083884232508e-06, + "loss": 0.7747, + "step": 2671 + }, + { + "epoch": 0.22769492969748614, + "grad_norm": 17.375, + "learning_rate": 2.4260269472172887e-06, + "loss": 0.5734, + "step": 2672 + }, + { + "epoch": 0.2277801448657861, + "grad_norm": 14.0, + "learning_rate": 2.4259699889500797e-06, + "loss": 0.9756, + "step": 2673 + }, + { + "epoch": 0.22786536003408606, + "grad_norm": 13.8125, + "learning_rate": 2.4259130094319106e-06, + "loss": 0.6701, + "step": 2674 + }, + { + "epoch": 0.227950575202386, + "grad_norm": 13.625, + "learning_rate": 2.4258560086638113e-06, + "loss": 0.7146, + "step": 2675 + }, + { + "epoch": 0.22803579037068597, + "grad_norm": 18.75, + "learning_rate": 2.4257989866468105e-06, + "loss": 1.1415, + "step": 2676 + }, + { + "epoch": 0.22812100553898593, + "grad_norm": 16.625, + "learning_rate": 2.425741943381941e-06, + "loss": 0.8519, + "step": 2677 + }, + { + "epoch": 0.22820622070728588, + "grad_norm": 11.375, + "learning_rate": 2.4256848788702314e-06, + "loss": 0.7111, + "step": 2678 + }, + { + "epoch": 0.22829143587558584, + "grad_norm": 14.6875, + "learning_rate": 2.425627793112714e-06, + "loss": 0.7178, + "step": 2679 + }, + { + "epoch": 0.22837665104388583, + "grad_norm": 15.8125, + "learning_rate": 2.42557068611042e-06, + "loss": 0.8471, + "step": 2680 + }, + { + "epoch": 0.22846186621218578, + "grad_norm": 13.625, + "learning_rate": 2.4255135578643817e-06, + "loss": 0.7265, + "step": 2681 + }, + { + "epoch": 0.22854708138048574, + "grad_norm": 13.4375, + "learning_rate": 2.4254564083756318e-06, + "loss": 0.5264, + "step": 2682 + }, + { + "epoch": 0.2286322965487857, + "grad_norm": 15.4375, + "learning_rate": 2.4253992376452017e-06, + "loss": 0.9509, + "step": 2683 + }, + { + "epoch": 0.22871751171708565, + "grad_norm": 15.5625, + "learning_rate": 2.4253420456741263e-06, + "loss": 0.9853, + "step": 2684 + }, + { + "epoch": 0.2288027268853856, + "grad_norm": 23.375, + "learning_rate": 2.4252848324634375e-06, + "loss": 1.2609, + "step": 2685 + }, + { + "epoch": 0.22888794205368557, + "grad_norm": 13.0625, + "learning_rate": 2.4252275980141705e-06, + "loss": 0.7184, + "step": 2686 + }, + { + "epoch": 0.22897315722198552, + "grad_norm": 15.5625, + "learning_rate": 2.4251703423273586e-06, + "loss": 0.8876, + "step": 2687 + }, + { + "epoch": 0.22905837239028548, + "grad_norm": 12.75, + "learning_rate": 2.4251130654040374e-06, + "loss": 0.7087, + "step": 2688 + }, + { + "epoch": 0.22914358755858544, + "grad_norm": 17.125, + "learning_rate": 2.4250557672452406e-06, + "loss": 0.8459, + "step": 2689 + }, + { + "epoch": 0.2292288027268854, + "grad_norm": 15.0625, + "learning_rate": 2.424998447852005e-06, + "loss": 0.9062, + "step": 2690 + }, + { + "epoch": 0.22931401789518535, + "grad_norm": 13.8125, + "learning_rate": 2.424941107225366e-06, + "loss": 0.6738, + "step": 2691 + }, + { + "epoch": 0.2293992330634853, + "grad_norm": 16.125, + "learning_rate": 2.42488374536636e-06, + "loss": 0.8291, + "step": 2692 + }, + { + "epoch": 0.22948444823178527, + "grad_norm": 14.9375, + "learning_rate": 2.424826362276023e-06, + "loss": 0.5306, + "step": 2693 + }, + { + "epoch": 0.22956966340008522, + "grad_norm": 17.0, + "learning_rate": 2.4247689579553918e-06, + "loss": 0.8042, + "step": 2694 + }, + { + "epoch": 0.22965487856838518, + "grad_norm": 14.0625, + "learning_rate": 2.4247115324055045e-06, + "loss": 1.0936, + "step": 2695 + }, + { + "epoch": 0.22974009373668514, + "grad_norm": 11.5, + "learning_rate": 2.424654085627398e-06, + "loss": 0.5308, + "step": 2696 + }, + { + "epoch": 0.2298253089049851, + "grad_norm": 11.875, + "learning_rate": 2.4245966176221118e-06, + "loss": 0.4608, + "step": 2697 + }, + { + "epoch": 0.22991052407328505, + "grad_norm": 14.875, + "learning_rate": 2.4245391283906837e-06, + "loss": 0.6074, + "step": 2698 + }, + { + "epoch": 0.229995739241585, + "grad_norm": 15.875, + "learning_rate": 2.424481617934152e-06, + "loss": 0.9357, + "step": 2699 + }, + { + "epoch": 0.23008095440988496, + "grad_norm": 14.375, + "learning_rate": 2.4244240862535563e-06, + "loss": 0.9879, + "step": 2700 + }, + { + "epoch": 0.23016616957818492, + "grad_norm": 19.125, + "learning_rate": 2.4243665333499365e-06, + "loss": 1.1863, + "step": 2701 + }, + { + "epoch": 0.23025138474648488, + "grad_norm": 15.3125, + "learning_rate": 2.424308959224333e-06, + "loss": 0.8961, + "step": 2702 + }, + { + "epoch": 0.23033659991478483, + "grad_norm": 12.375, + "learning_rate": 2.4242513638777854e-06, + "loss": 0.5673, + "step": 2703 + }, + { + "epoch": 0.2304218150830848, + "grad_norm": 10.9375, + "learning_rate": 2.424193747311335e-06, + "loss": 0.4202, + "step": 2704 + }, + { + "epoch": 0.23050703025138475, + "grad_norm": 17.375, + "learning_rate": 2.4241361095260225e-06, + "loss": 0.6607, + "step": 2705 + }, + { + "epoch": 0.2305922454196847, + "grad_norm": 13.625, + "learning_rate": 2.4240784505228897e-06, + "loss": 0.6322, + "step": 2706 + }, + { + "epoch": 0.23067746058798466, + "grad_norm": 12.5625, + "learning_rate": 2.4240207703029788e-06, + "loss": 0.6797, + "step": 2707 + }, + { + "epoch": 0.23076267575628462, + "grad_norm": 12.6875, + "learning_rate": 2.4239630688673324e-06, + "loss": 0.4156, + "step": 2708 + }, + { + "epoch": 0.23084789092458458, + "grad_norm": 17.375, + "learning_rate": 2.4239053462169924e-06, + "loss": 1.1465, + "step": 2709 + }, + { + "epoch": 0.23093310609288453, + "grad_norm": 12.3125, + "learning_rate": 2.4238476023530028e-06, + "loss": 0.4007, + "step": 2710 + }, + { + "epoch": 0.2310183212611845, + "grad_norm": 10.625, + "learning_rate": 2.423789837276406e-06, + "loss": 0.3558, + "step": 2711 + }, + { + "epoch": 0.23110353642948445, + "grad_norm": 11.6875, + "learning_rate": 2.423732050988247e-06, + "loss": 0.5028, + "step": 2712 + }, + { + "epoch": 0.2311887515977844, + "grad_norm": 13.25, + "learning_rate": 2.42367424348957e-06, + "loss": 0.7019, + "step": 2713 + }, + { + "epoch": 0.23127396676608436, + "grad_norm": 12.5625, + "learning_rate": 2.4236164147814184e-06, + "loss": 0.47, + "step": 2714 + }, + { + "epoch": 0.23135918193438432, + "grad_norm": 20.0, + "learning_rate": 2.423558564864838e-06, + "loss": 0.9292, + "step": 2715 + }, + { + "epoch": 0.23144439710268427, + "grad_norm": 16.75, + "learning_rate": 2.4235006937408743e-06, + "loss": 1.0831, + "step": 2716 + }, + { + "epoch": 0.23152961227098423, + "grad_norm": 15.3125, + "learning_rate": 2.4234428014105735e-06, + "loss": 0.7353, + "step": 2717 + }, + { + "epoch": 0.2316148274392842, + "grad_norm": 21.25, + "learning_rate": 2.423384887874981e-06, + "loss": 1.2695, + "step": 2718 + }, + { + "epoch": 0.23170004260758414, + "grad_norm": 13.125, + "learning_rate": 2.4233269531351437e-06, + "loss": 0.7943, + "step": 2719 + }, + { + "epoch": 0.2317852577758841, + "grad_norm": 12.1875, + "learning_rate": 2.423268997192108e-06, + "loss": 0.6412, + "step": 2720 + }, + { + "epoch": 0.23187047294418406, + "grad_norm": 14.5, + "learning_rate": 2.4232110200469223e-06, + "loss": 0.6674, + "step": 2721 + }, + { + "epoch": 0.23195568811248402, + "grad_norm": 16.0, + "learning_rate": 2.423153021700634e-06, + "loss": 0.7757, + "step": 2722 + }, + { + "epoch": 0.23204090328078397, + "grad_norm": 13.875, + "learning_rate": 2.42309500215429e-06, + "loss": 1.1354, + "step": 2723 + }, + { + "epoch": 0.23212611844908393, + "grad_norm": 18.875, + "learning_rate": 2.4230369614089404e-06, + "loss": 0.9389, + "step": 2724 + }, + { + "epoch": 0.23221133361738389, + "grad_norm": 15.375, + "learning_rate": 2.422978899465633e-06, + "loss": 0.7042, + "step": 2725 + }, + { + "epoch": 0.23229654878568384, + "grad_norm": 11.8125, + "learning_rate": 2.4229208163254176e-06, + "loss": 0.5903, + "step": 2726 + }, + { + "epoch": 0.2323817639539838, + "grad_norm": 16.25, + "learning_rate": 2.4228627119893434e-06, + "loss": 0.7102, + "step": 2727 + }, + { + "epoch": 0.23246697912228376, + "grad_norm": 15.6875, + "learning_rate": 2.42280458645846e-06, + "loss": 0.9055, + "step": 2728 + }, + { + "epoch": 0.2325521942905837, + "grad_norm": 15.4375, + "learning_rate": 2.4227464397338195e-06, + "loss": 0.8032, + "step": 2729 + }, + { + "epoch": 0.23263740945888367, + "grad_norm": 22.125, + "learning_rate": 2.4226882718164708e-06, + "loss": 1.1022, + "step": 2730 + }, + { + "epoch": 0.23272262462718363, + "grad_norm": 19.0, + "learning_rate": 2.422630082707466e-06, + "loss": 0.8255, + "step": 2731 + }, + { + "epoch": 0.23280783979548358, + "grad_norm": 21.0, + "learning_rate": 2.4225718724078565e-06, + "loss": 0.7442, + "step": 2732 + }, + { + "epoch": 0.23289305496378354, + "grad_norm": 18.375, + "learning_rate": 2.422513640918694e-06, + "loss": 0.688, + "step": 2733 + }, + { + "epoch": 0.2329782701320835, + "grad_norm": 24.5, + "learning_rate": 2.4224553882410313e-06, + "loss": 0.693, + "step": 2734 + }, + { + "epoch": 0.23306348530038346, + "grad_norm": 13.9375, + "learning_rate": 2.42239711437592e-06, + "loss": 1.0697, + "step": 2735 + }, + { + "epoch": 0.2331487004686834, + "grad_norm": 18.125, + "learning_rate": 2.4223388193244144e-06, + "loss": 0.9238, + "step": 2736 + }, + { + "epoch": 0.2332339156369834, + "grad_norm": 17.0, + "learning_rate": 2.4222805030875677e-06, + "loss": 0.892, + "step": 2737 + }, + { + "epoch": 0.23331913080528335, + "grad_norm": 12.125, + "learning_rate": 2.4222221656664334e-06, + "loss": 0.6879, + "step": 2738 + }, + { + "epoch": 0.2334043459735833, + "grad_norm": 13.9375, + "learning_rate": 2.4221638070620655e-06, + "loss": 0.6441, + "step": 2739 + }, + { + "epoch": 0.23348956114188327, + "grad_norm": 16.375, + "learning_rate": 2.422105427275519e-06, + "loss": 0.6626, + "step": 2740 + }, + { + "epoch": 0.23357477631018322, + "grad_norm": 10.1875, + "learning_rate": 2.422047026307849e-06, + "loss": 0.4695, + "step": 2741 + }, + { + "epoch": 0.23365999147848318, + "grad_norm": 11.0, + "learning_rate": 2.4219886041601104e-06, + "loss": 0.3842, + "step": 2742 + }, + { + "epoch": 0.23374520664678314, + "grad_norm": 11.4375, + "learning_rate": 2.4219301608333595e-06, + "loss": 0.4441, + "step": 2743 + }, + { + "epoch": 0.2338304218150831, + "grad_norm": 14.0, + "learning_rate": 2.4218716963286516e-06, + "loss": 0.5696, + "step": 2744 + }, + { + "epoch": 0.23391563698338305, + "grad_norm": 13.375, + "learning_rate": 2.4218132106470443e-06, + "loss": 0.6914, + "step": 2745 + }, + { + "epoch": 0.234000852151683, + "grad_norm": 12.375, + "learning_rate": 2.4217547037895935e-06, + "loss": 0.5383, + "step": 2746 + }, + { + "epoch": 0.23408606731998297, + "grad_norm": 11.9375, + "learning_rate": 2.421696175757357e-06, + "loss": 0.6258, + "step": 2747 + }, + { + "epoch": 0.23417128248828292, + "grad_norm": 22.5, + "learning_rate": 2.4216376265513928e-06, + "loss": 1.0195, + "step": 2748 + }, + { + "epoch": 0.23425649765658288, + "grad_norm": 17.0, + "learning_rate": 2.421579056172758e-06, + "loss": 0.9471, + "step": 2749 + }, + { + "epoch": 0.23434171282488284, + "grad_norm": 35.75, + "learning_rate": 2.4215204646225118e-06, + "loss": 0.9019, + "step": 2750 + }, + { + "epoch": 0.2344269279931828, + "grad_norm": 13.625, + "learning_rate": 2.4214618519017125e-06, + "loss": 0.5979, + "step": 2751 + }, + { + "epoch": 0.23451214316148275, + "grad_norm": 20.5, + "learning_rate": 2.42140321801142e-06, + "loss": 1.1552, + "step": 2752 + }, + { + "epoch": 0.2345973583297827, + "grad_norm": 14.0, + "learning_rate": 2.421344562952693e-06, + "loss": 0.5761, + "step": 2753 + }, + { + "epoch": 0.23468257349808266, + "grad_norm": 13.5, + "learning_rate": 2.421285886726592e-06, + "loss": 0.7395, + "step": 2754 + }, + { + "epoch": 0.23476778866638262, + "grad_norm": 19.125, + "learning_rate": 2.421227189334177e-06, + "loss": 0.9516, + "step": 2755 + }, + { + "epoch": 0.23485300383468258, + "grad_norm": 15.0, + "learning_rate": 2.4211684707765094e-06, + "loss": 1.0189, + "step": 2756 + }, + { + "epoch": 0.23493821900298253, + "grad_norm": 17.125, + "learning_rate": 2.4211097310546495e-06, + "loss": 1.0226, + "step": 2757 + }, + { + "epoch": 0.2350234341712825, + "grad_norm": 26.625, + "learning_rate": 2.4210509701696592e-06, + "loss": 0.9777, + "step": 2758 + }, + { + "epoch": 0.23510864933958245, + "grad_norm": 29.125, + "learning_rate": 2.4209921881226e-06, + "loss": 1.006, + "step": 2759 + }, + { + "epoch": 0.2351938645078824, + "grad_norm": 14.125, + "learning_rate": 2.420933384914535e-06, + "loss": 0.833, + "step": 2760 + }, + { + "epoch": 0.23527907967618236, + "grad_norm": 14.1875, + "learning_rate": 2.420874560546526e-06, + "loss": 0.9195, + "step": 2761 + }, + { + "epoch": 0.23536429484448232, + "grad_norm": 13.375, + "learning_rate": 2.420815715019636e-06, + "loss": 0.84, + "step": 2762 + }, + { + "epoch": 0.23544951001278228, + "grad_norm": 18.0, + "learning_rate": 2.420756848334929e-06, + "loss": 0.5436, + "step": 2763 + }, + { + "epoch": 0.23553472518108223, + "grad_norm": 10.625, + "learning_rate": 2.420697960493469e-06, + "loss": 0.5073, + "step": 2764 + }, + { + "epoch": 0.2356199403493822, + "grad_norm": 15.625, + "learning_rate": 2.4206390514963186e-06, + "loss": 0.5162, + "step": 2765 + }, + { + "epoch": 0.23570515551768215, + "grad_norm": 20.75, + "learning_rate": 2.420580121344544e-06, + "loss": 0.9209, + "step": 2766 + }, + { + "epoch": 0.2357903706859821, + "grad_norm": 12.25, + "learning_rate": 2.420521170039209e-06, + "loss": 0.5864, + "step": 2767 + }, + { + "epoch": 0.23587558585428206, + "grad_norm": 15.4375, + "learning_rate": 2.42046219758138e-06, + "loss": 0.9031, + "step": 2768 + }, + { + "epoch": 0.23596080102258202, + "grad_norm": 13.0, + "learning_rate": 2.420403203972122e-06, + "loss": 0.6023, + "step": 2769 + }, + { + "epoch": 0.23604601619088197, + "grad_norm": 14.5, + "learning_rate": 2.4203441892125007e-06, + "loss": 0.7756, + "step": 2770 + }, + { + "epoch": 0.23613123135918193, + "grad_norm": 16.375, + "learning_rate": 2.420285153303583e-06, + "loss": 0.83, + "step": 2771 + }, + { + "epoch": 0.2362164465274819, + "grad_norm": 16.375, + "learning_rate": 2.4202260962464365e-06, + "loss": 0.6134, + "step": 2772 + }, + { + "epoch": 0.23630166169578184, + "grad_norm": 11.6875, + "learning_rate": 2.4201670180421273e-06, + "loss": 0.5596, + "step": 2773 + }, + { + "epoch": 0.2363868768640818, + "grad_norm": 16.0, + "learning_rate": 2.420107918691723e-06, + "loss": 1.1389, + "step": 2774 + }, + { + "epoch": 0.23647209203238176, + "grad_norm": 17.625, + "learning_rate": 2.4200487981962926e-06, + "loss": 1.1048, + "step": 2775 + }, + { + "epoch": 0.23655730720068172, + "grad_norm": 16.625, + "learning_rate": 2.4199896565569037e-06, + "loss": 0.7548, + "step": 2776 + }, + { + "epoch": 0.23664252236898167, + "grad_norm": 14.625, + "learning_rate": 2.419930493774625e-06, + "loss": 0.6183, + "step": 2777 + }, + { + "epoch": 0.23672773753728163, + "grad_norm": 14.1875, + "learning_rate": 2.419871309850526e-06, + "loss": 1.0604, + "step": 2778 + }, + { + "epoch": 0.2368129527055816, + "grad_norm": 19.625, + "learning_rate": 2.419812104785676e-06, + "loss": 1.0181, + "step": 2779 + }, + { + "epoch": 0.23689816787388154, + "grad_norm": 17.375, + "learning_rate": 2.4197528785811447e-06, + "loss": 0.9889, + "step": 2780 + }, + { + "epoch": 0.2369833830421815, + "grad_norm": 15.5625, + "learning_rate": 2.4196936312380025e-06, + "loss": 0.946, + "step": 2781 + }, + { + "epoch": 0.23706859821048146, + "grad_norm": 11.625, + "learning_rate": 2.4196343627573203e-06, + "loss": 0.7725, + "step": 2782 + }, + { + "epoch": 0.23715381337878141, + "grad_norm": 19.0, + "learning_rate": 2.4195750731401698e-06, + "loss": 0.9169, + "step": 2783 + }, + { + "epoch": 0.23723902854708137, + "grad_norm": 19.625, + "learning_rate": 2.4195157623876206e-06, + "loss": 0.891, + "step": 2784 + }, + { + "epoch": 0.23732424371538133, + "grad_norm": 17.0, + "learning_rate": 2.419456430500746e-06, + "loss": 1.0242, + "step": 2785 + }, + { + "epoch": 0.23740945888368128, + "grad_norm": 12.375, + "learning_rate": 2.4193970774806175e-06, + "loss": 0.7015, + "step": 2786 + }, + { + "epoch": 0.23749467405198124, + "grad_norm": 14.375, + "learning_rate": 2.4193377033283082e-06, + "loss": 0.8451, + "step": 2787 + }, + { + "epoch": 0.2375798892202812, + "grad_norm": 15.5625, + "learning_rate": 2.4192783080448906e-06, + "loss": 1.0183, + "step": 2788 + }, + { + "epoch": 0.23766510438858116, + "grad_norm": 11.6875, + "learning_rate": 2.419218891631438e-06, + "loss": 0.5921, + "step": 2789 + }, + { + "epoch": 0.2377503195568811, + "grad_norm": 19.5, + "learning_rate": 2.4191594540890246e-06, + "loss": 0.7494, + "step": 2790 + }, + { + "epoch": 0.23783553472518107, + "grad_norm": 21.625, + "learning_rate": 2.419099995418724e-06, + "loss": 0.9608, + "step": 2791 + }, + { + "epoch": 0.23792074989348103, + "grad_norm": 12.5, + "learning_rate": 2.4190405156216105e-06, + "loss": 0.5553, + "step": 2792 + }, + { + "epoch": 0.238005965061781, + "grad_norm": 24.0, + "learning_rate": 2.41898101469876e-06, + "loss": 0.9755, + "step": 2793 + }, + { + "epoch": 0.23809118023008097, + "grad_norm": 17.75, + "learning_rate": 2.4189214926512467e-06, + "loss": 0.8864, + "step": 2794 + }, + { + "epoch": 0.23817639539838092, + "grad_norm": 16.75, + "learning_rate": 2.4188619494801463e-06, + "loss": 0.8475, + "step": 2795 + }, + { + "epoch": 0.23826161056668088, + "grad_norm": 19.5, + "learning_rate": 2.4188023851865354e-06, + "loss": 0.9995, + "step": 2796 + }, + { + "epoch": 0.23834682573498084, + "grad_norm": 14.3125, + "learning_rate": 2.41874279977149e-06, + "loss": 0.612, + "step": 2797 + }, + { + "epoch": 0.2384320409032808, + "grad_norm": 13.3125, + "learning_rate": 2.4186831932360873e-06, + "loss": 0.7524, + "step": 2798 + }, + { + "epoch": 0.23851725607158075, + "grad_norm": 19.75, + "learning_rate": 2.4186235655814037e-06, + "loss": 1.0977, + "step": 2799 + }, + { + "epoch": 0.2386024712398807, + "grad_norm": 23.25, + "learning_rate": 2.4185639168085175e-06, + "loss": 1.1972, + "step": 2800 + }, + { + "epoch": 0.23868768640818067, + "grad_norm": 18.625, + "learning_rate": 2.418504246918506e-06, + "loss": 0.735, + "step": 2801 + }, + { + "epoch": 0.23877290157648062, + "grad_norm": 16.75, + "learning_rate": 2.4184445559124476e-06, + "loss": 1.0113, + "step": 2802 + }, + { + "epoch": 0.23885811674478058, + "grad_norm": 13.4375, + "learning_rate": 2.4183848437914216e-06, + "loss": 0.6332, + "step": 2803 + }, + { + "epoch": 0.23894333191308054, + "grad_norm": 16.375, + "learning_rate": 2.4183251105565058e-06, + "loss": 1.0582, + "step": 2804 + }, + { + "epoch": 0.2390285470813805, + "grad_norm": 12.75, + "learning_rate": 2.418265356208781e-06, + "loss": 0.8828, + "step": 2805 + }, + { + "epoch": 0.23911376224968045, + "grad_norm": 12.5, + "learning_rate": 2.4182055807493263e-06, + "loss": 0.4269, + "step": 2806 + }, + { + "epoch": 0.2391989774179804, + "grad_norm": 20.875, + "learning_rate": 2.418145784179222e-06, + "loss": 1.118, + "step": 2807 + }, + { + "epoch": 0.23928419258628036, + "grad_norm": 13.5625, + "learning_rate": 2.4180859664995486e-06, + "loss": 0.8421, + "step": 2808 + }, + { + "epoch": 0.23936940775458032, + "grad_norm": 13.4375, + "learning_rate": 2.4180261277113874e-06, + "loss": 0.633, + "step": 2809 + }, + { + "epoch": 0.23945462292288028, + "grad_norm": 14.0, + "learning_rate": 2.4179662678158195e-06, + "loss": 0.7742, + "step": 2810 + }, + { + "epoch": 0.23953983809118023, + "grad_norm": 16.75, + "learning_rate": 2.4179063868139265e-06, + "loss": 1.0349, + "step": 2811 + }, + { + "epoch": 0.2396250532594802, + "grad_norm": 11.0, + "learning_rate": 2.417846484706791e-06, + "loss": 0.435, + "step": 2812 + }, + { + "epoch": 0.23971026842778015, + "grad_norm": 27.5, + "learning_rate": 2.417786561495495e-06, + "loss": 1.2232, + "step": 2813 + }, + { + "epoch": 0.2397954835960801, + "grad_norm": 30.875, + "learning_rate": 2.417726617181121e-06, + "loss": 0.952, + "step": 2814 + }, + { + "epoch": 0.23988069876438006, + "grad_norm": 15.5, + "learning_rate": 2.4176666517647534e-06, + "loss": 0.7361, + "step": 2815 + }, + { + "epoch": 0.23996591393268002, + "grad_norm": 14.375, + "learning_rate": 2.417606665247475e-06, + "loss": 0.7363, + "step": 2816 + }, + { + "epoch": 0.24005112910097998, + "grad_norm": 14.0, + "learning_rate": 2.4175466576303703e-06, + "loss": 0.7376, + "step": 2817 + }, + { + "epoch": 0.24013634426927993, + "grad_norm": 18.875, + "learning_rate": 2.417486628914523e-06, + "loss": 1.02, + "step": 2818 + }, + { + "epoch": 0.2402215594375799, + "grad_norm": 21.625, + "learning_rate": 2.4174265791010187e-06, + "loss": 0.9254, + "step": 2819 + }, + { + "epoch": 0.24030677460587985, + "grad_norm": 19.5, + "learning_rate": 2.417366508190942e-06, + "loss": 0.6651, + "step": 2820 + }, + { + "epoch": 0.2403919897741798, + "grad_norm": 12.25, + "learning_rate": 2.4173064161853788e-06, + "loss": 0.7512, + "step": 2821 + }, + { + "epoch": 0.24047720494247976, + "grad_norm": 15.0, + "learning_rate": 2.4172463030854146e-06, + "loss": 0.5555, + "step": 2822 + }, + { + "epoch": 0.24056242011077972, + "grad_norm": 13.75, + "learning_rate": 2.417186168892136e-06, + "loss": 0.715, + "step": 2823 + }, + { + "epoch": 0.24064763527907967, + "grad_norm": 20.25, + "learning_rate": 2.41712601360663e-06, + "loss": 1.1555, + "step": 2824 + }, + { + "epoch": 0.24073285044737963, + "grad_norm": 12.25, + "learning_rate": 2.4170658372299827e-06, + "loss": 0.3854, + "step": 2825 + }, + { + "epoch": 0.2408180656156796, + "grad_norm": 10.5625, + "learning_rate": 2.4170056397632823e-06, + "loss": 0.484, + "step": 2826 + }, + { + "epoch": 0.24090328078397955, + "grad_norm": 15.375, + "learning_rate": 2.4169454212076167e-06, + "loss": 1.2465, + "step": 2827 + }, + { + "epoch": 0.2409884959522795, + "grad_norm": 14.5, + "learning_rate": 2.416885181564074e-06, + "loss": 0.839, + "step": 2828 + }, + { + "epoch": 0.24107371112057946, + "grad_norm": 15.375, + "learning_rate": 2.4168249208337422e-06, + "loss": 1.0157, + "step": 2829 + }, + { + "epoch": 0.24115892628887942, + "grad_norm": 21.875, + "learning_rate": 2.4167646390177115e-06, + "loss": 1.0009, + "step": 2830 + }, + { + "epoch": 0.24124414145717937, + "grad_norm": 23.125, + "learning_rate": 2.41670433611707e-06, + "loss": 0.9316, + "step": 2831 + }, + { + "epoch": 0.24132935662547933, + "grad_norm": 12.25, + "learning_rate": 2.416644012132908e-06, + "loss": 0.5364, + "step": 2832 + }, + { + "epoch": 0.2414145717937793, + "grad_norm": 13.0, + "learning_rate": 2.416583667066315e-06, + "loss": 0.6709, + "step": 2833 + }, + { + "epoch": 0.24149978696207924, + "grad_norm": 14.875, + "learning_rate": 2.4165233009183827e-06, + "loss": 0.9483, + "step": 2834 + }, + { + "epoch": 0.2415850021303792, + "grad_norm": 10.5625, + "learning_rate": 2.416462913690201e-06, + "loss": 0.9666, + "step": 2835 + }, + { + "epoch": 0.24167021729867916, + "grad_norm": 12.0, + "learning_rate": 2.416402505382862e-06, + "loss": 0.5579, + "step": 2836 + }, + { + "epoch": 0.24175543246697911, + "grad_norm": 14.0625, + "learning_rate": 2.416342075997456e-06, + "loss": 0.734, + "step": 2837 + }, + { + "epoch": 0.24184064763527907, + "grad_norm": 15.5, + "learning_rate": 2.4162816255350764e-06, + "loss": 0.6405, + "step": 2838 + }, + { + "epoch": 0.24192586280357903, + "grad_norm": 16.125, + "learning_rate": 2.416221153996815e-06, + "loss": 0.6583, + "step": 2839 + }, + { + "epoch": 0.24201107797187899, + "grad_norm": 14.4375, + "learning_rate": 2.4161606613837645e-06, + "loss": 0.8384, + "step": 2840 + }, + { + "epoch": 0.24209629314017894, + "grad_norm": 15.75, + "learning_rate": 2.4161001476970186e-06, + "loss": 0.7441, + "step": 2841 + }, + { + "epoch": 0.2421815083084789, + "grad_norm": 18.25, + "learning_rate": 2.41603961293767e-06, + "loss": 1.0302, + "step": 2842 + }, + { + "epoch": 0.24226672347677886, + "grad_norm": 14.4375, + "learning_rate": 2.415979057106813e-06, + "loss": 0.6088, + "step": 2843 + }, + { + "epoch": 0.2423519386450788, + "grad_norm": 18.625, + "learning_rate": 2.415918480205542e-06, + "loss": 0.98, + "step": 2844 + }, + { + "epoch": 0.24243715381337877, + "grad_norm": 20.625, + "learning_rate": 2.4158578822349516e-06, + "loss": 1.1366, + "step": 2845 + }, + { + "epoch": 0.24252236898167873, + "grad_norm": 13.8125, + "learning_rate": 2.4157972631961374e-06, + "loss": 0.7558, + "step": 2846 + }, + { + "epoch": 0.24260758414997868, + "grad_norm": 26.875, + "learning_rate": 2.415736623090194e-06, + "loss": 0.5989, + "step": 2847 + }, + { + "epoch": 0.24269279931827864, + "grad_norm": 16.125, + "learning_rate": 2.4156759619182175e-06, + "loss": 0.7427, + "step": 2848 + }, + { + "epoch": 0.2427780144865786, + "grad_norm": 12.0625, + "learning_rate": 2.4156152796813044e-06, + "loss": 0.5771, + "step": 2849 + }, + { + "epoch": 0.24286322965487858, + "grad_norm": 15.5625, + "learning_rate": 2.415554576380551e-06, + "loss": 0.6352, + "step": 2850 + }, + { + "epoch": 0.24294844482317854, + "grad_norm": 11.0, + "learning_rate": 2.415493852017054e-06, + "loss": 0.5072, + "step": 2851 + }, + { + "epoch": 0.2430336599914785, + "grad_norm": 13.5625, + "learning_rate": 2.415433106591911e-06, + "loss": 0.7335, + "step": 2852 + }, + { + "epoch": 0.24311887515977845, + "grad_norm": 12.4375, + "learning_rate": 2.4153723401062207e-06, + "loss": 0.4162, + "step": 2853 + }, + { + "epoch": 0.2432040903280784, + "grad_norm": 16.125, + "learning_rate": 2.4153115525610797e-06, + "loss": 0.6565, + "step": 2854 + }, + { + "epoch": 0.24328930549637837, + "grad_norm": 14.25, + "learning_rate": 2.4152507439575874e-06, + "loss": 0.5961, + "step": 2855 + }, + { + "epoch": 0.24337452066467832, + "grad_norm": 14.3125, + "learning_rate": 2.415189914296842e-06, + "loss": 0.6748, + "step": 2856 + }, + { + "epoch": 0.24345973583297828, + "grad_norm": 13.4375, + "learning_rate": 2.4151290635799436e-06, + "loss": 0.6063, + "step": 2857 + }, + { + "epoch": 0.24354495100127824, + "grad_norm": 15.375, + "learning_rate": 2.4150681918079914e-06, + "loss": 0.9039, + "step": 2858 + }, + { + "epoch": 0.2436301661695782, + "grad_norm": 18.875, + "learning_rate": 2.4150072989820853e-06, + "loss": 0.815, + "step": 2859 + }, + { + "epoch": 0.24371538133787815, + "grad_norm": 79.0, + "learning_rate": 2.4149463851033257e-06, + "loss": 0.7938, + "step": 2860 + }, + { + "epoch": 0.2438005965061781, + "grad_norm": 14.5, + "learning_rate": 2.4148854501728137e-06, + "loss": 0.8861, + "step": 2861 + }, + { + "epoch": 0.24388581167447806, + "grad_norm": 13.5625, + "learning_rate": 2.41482449419165e-06, + "loss": 0.7299, + "step": 2862 + }, + { + "epoch": 0.24397102684277802, + "grad_norm": 14.875, + "learning_rate": 2.414763517160936e-06, + "loss": 0.8017, + "step": 2863 + }, + { + "epoch": 0.24405624201107798, + "grad_norm": 12.9375, + "learning_rate": 2.4147025190817745e-06, + "loss": 0.4447, + "step": 2864 + }, + { + "epoch": 0.24414145717937794, + "grad_norm": 14.3125, + "learning_rate": 2.4146414999552677e-06, + "loss": 0.7857, + "step": 2865 + }, + { + "epoch": 0.2442266723476779, + "grad_norm": 18.625, + "learning_rate": 2.414580459782517e-06, + "loss": 0.5316, + "step": 2866 + }, + { + "epoch": 0.24431188751597785, + "grad_norm": 45.0, + "learning_rate": 2.414519398564627e-06, + "loss": 0.7077, + "step": 2867 + }, + { + "epoch": 0.2443971026842778, + "grad_norm": 14.4375, + "learning_rate": 2.4144583163027e-06, + "loss": 0.5152, + "step": 2868 + }, + { + "epoch": 0.24448231785257776, + "grad_norm": 13.8125, + "learning_rate": 2.4143972129978406e-06, + "loss": 0.6902, + "step": 2869 + }, + { + "epoch": 0.24456753302087772, + "grad_norm": 13.25, + "learning_rate": 2.4143360886511523e-06, + "loss": 0.8894, + "step": 2870 + }, + { + "epoch": 0.24465274818917768, + "grad_norm": 16.625, + "learning_rate": 2.4142749432637404e-06, + "loss": 1.2033, + "step": 2871 + }, + { + "epoch": 0.24473796335747763, + "grad_norm": 19.75, + "learning_rate": 2.414213776836709e-06, + "loss": 0.9076, + "step": 2872 + }, + { + "epoch": 0.2448231785257776, + "grad_norm": 17.625, + "learning_rate": 2.4141525893711643e-06, + "loss": 0.72, + "step": 2873 + }, + { + "epoch": 0.24490839369407755, + "grad_norm": 34.75, + "learning_rate": 2.4140913808682117e-06, + "loss": 1.1777, + "step": 2874 + }, + { + "epoch": 0.2449936088623775, + "grad_norm": 12.125, + "learning_rate": 2.414030151328957e-06, + "loss": 0.623, + "step": 2875 + }, + { + "epoch": 0.24507882403067746, + "grad_norm": 15.25, + "learning_rate": 2.4139689007545072e-06, + "loss": 0.9947, + "step": 2876 + }, + { + "epoch": 0.24516403919897742, + "grad_norm": 19.125, + "learning_rate": 2.413907629145969e-06, + "loss": 1.0784, + "step": 2877 + }, + { + "epoch": 0.24524925436727738, + "grad_norm": 14.9375, + "learning_rate": 2.4138463365044492e-06, + "loss": 0.9475, + "step": 2878 + }, + { + "epoch": 0.24533446953557733, + "grad_norm": 16.5, + "learning_rate": 2.413785022831056e-06, + "loss": 0.9571, + "step": 2879 + }, + { + "epoch": 0.2454196847038773, + "grad_norm": 12.8125, + "learning_rate": 2.413723688126897e-06, + "loss": 0.6369, + "step": 2880 + }, + { + "epoch": 0.24550489987217725, + "grad_norm": 13.25, + "learning_rate": 2.413662332393081e-06, + "loss": 0.6627, + "step": 2881 + }, + { + "epoch": 0.2455901150404772, + "grad_norm": 12.875, + "learning_rate": 2.4136009556307162e-06, + "loss": 0.6959, + "step": 2882 + }, + { + "epoch": 0.24567533020877716, + "grad_norm": 13.5, + "learning_rate": 2.413539557840912e-06, + "loss": 0.6722, + "step": 2883 + }, + { + "epoch": 0.24576054537707712, + "grad_norm": 13.8125, + "learning_rate": 2.4134781390247785e-06, + "loss": 0.5039, + "step": 2884 + }, + { + "epoch": 0.24584576054537707, + "grad_norm": 16.5, + "learning_rate": 2.4134166991834247e-06, + "loss": 0.5201, + "step": 2885 + }, + { + "epoch": 0.24593097571367703, + "grad_norm": 16.75, + "learning_rate": 2.413355238317961e-06, + "loss": 0.7207, + "step": 2886 + }, + { + "epoch": 0.246016190881977, + "grad_norm": 15.75, + "learning_rate": 2.413293756429499e-06, + "loss": 1.0481, + "step": 2887 + }, + { + "epoch": 0.24610140605027694, + "grad_norm": 13.8125, + "learning_rate": 2.4132322535191482e-06, + "loss": 0.7666, + "step": 2888 + }, + { + "epoch": 0.2461866212185769, + "grad_norm": 17.0, + "learning_rate": 2.4131707295880213e-06, + "loss": 0.8501, + "step": 2889 + }, + { + "epoch": 0.24627183638687686, + "grad_norm": 14.5625, + "learning_rate": 2.4131091846372293e-06, + "loss": 0.8717, + "step": 2890 + }, + { + "epoch": 0.24635705155517681, + "grad_norm": 13.0625, + "learning_rate": 2.4130476186678852e-06, + "loss": 0.6441, + "step": 2891 + }, + { + "epoch": 0.24644226672347677, + "grad_norm": 15.5, + "learning_rate": 2.412986031681101e-06, + "loss": 0.8361, + "step": 2892 + }, + { + "epoch": 0.24652748189177673, + "grad_norm": 17.375, + "learning_rate": 2.4129244236779897e-06, + "loss": 0.6939, + "step": 2893 + }, + { + "epoch": 0.24661269706007669, + "grad_norm": 19.25, + "learning_rate": 2.4128627946596648e-06, + "loss": 0.8974, + "step": 2894 + }, + { + "epoch": 0.24669791222837664, + "grad_norm": 13.125, + "learning_rate": 2.4128011446272397e-06, + "loss": 0.571, + "step": 2895 + }, + { + "epoch": 0.2467831273966766, + "grad_norm": 16.0, + "learning_rate": 2.4127394735818283e-06, + "loss": 1.073, + "step": 2896 + }, + { + "epoch": 0.24686834256497656, + "grad_norm": 18.625, + "learning_rate": 2.412677781524546e-06, + "loss": 0.8276, + "step": 2897 + }, + { + "epoch": 0.2469535577332765, + "grad_norm": 13.875, + "learning_rate": 2.4126160684565065e-06, + "loss": 0.9196, + "step": 2898 + }, + { + "epoch": 0.24703877290157647, + "grad_norm": 21.375, + "learning_rate": 2.412554334378826e-06, + "loss": 0.8342, + "step": 2899 + }, + { + "epoch": 0.24712398806987643, + "grad_norm": 11.5625, + "learning_rate": 2.4124925792926195e-06, + "loss": 0.4633, + "step": 2900 + }, + { + "epoch": 0.24720920323817638, + "grad_norm": 15.5625, + "learning_rate": 2.4124308031990028e-06, + "loss": 0.8989, + "step": 2901 + }, + { + "epoch": 0.24729441840647634, + "grad_norm": 11.0, + "learning_rate": 2.4123690060990927e-06, + "loss": 0.3603, + "step": 2902 + }, + { + "epoch": 0.2473796335747763, + "grad_norm": 13.0, + "learning_rate": 2.412307187994006e-06, + "loss": 0.5763, + "step": 2903 + }, + { + "epoch": 0.24746484874307625, + "grad_norm": 16.375, + "learning_rate": 2.4122453488848595e-06, + "loss": 0.6921, + "step": 2904 + }, + { + "epoch": 0.2475500639113762, + "grad_norm": 33.25, + "learning_rate": 2.412183488772771e-06, + "loss": 0.9493, + "step": 2905 + }, + { + "epoch": 0.2476352790796762, + "grad_norm": 14.0, + "learning_rate": 2.4121216076588577e-06, + "loss": 0.7555, + "step": 2906 + }, + { + "epoch": 0.24772049424797615, + "grad_norm": 17.5, + "learning_rate": 2.412059705544239e-06, + "loss": 0.5751, + "step": 2907 + }, + { + "epoch": 0.2478057094162761, + "grad_norm": 12.4375, + "learning_rate": 2.4119977824300324e-06, + "loss": 0.5633, + "step": 2908 + }, + { + "epoch": 0.24789092458457607, + "grad_norm": 17.875, + "learning_rate": 2.4119358383173573e-06, + "loss": 0.7717, + "step": 2909 + }, + { + "epoch": 0.24797613975287602, + "grad_norm": 14.0625, + "learning_rate": 2.411873873207334e-06, + "loss": 0.8224, + "step": 2910 + }, + { + "epoch": 0.24806135492117598, + "grad_norm": 15.1875, + "learning_rate": 2.41181188710108e-06, + "loss": 0.7632, + "step": 2911 + }, + { + "epoch": 0.24814657008947594, + "grad_norm": 12.4375, + "learning_rate": 2.411749879999718e-06, + "loss": 0.6182, + "step": 2912 + }, + { + "epoch": 0.2482317852577759, + "grad_norm": 11.6875, + "learning_rate": 2.4116878519043676e-06, + "loss": 0.529, + "step": 2913 + }, + { + "epoch": 0.24831700042607585, + "grad_norm": 14.0, + "learning_rate": 2.411625802816149e-06, + "loss": 0.6263, + "step": 2914 + }, + { + "epoch": 0.2484022155943758, + "grad_norm": 15.8125, + "learning_rate": 2.4115637327361844e-06, + "loss": 0.9153, + "step": 2915 + }, + { + "epoch": 0.24848743076267577, + "grad_norm": 13.5, + "learning_rate": 2.4115016416655947e-06, + "loss": 0.8225, + "step": 2916 + }, + { + "epoch": 0.24857264593097572, + "grad_norm": 13.5, + "learning_rate": 2.411439529605503e-06, + "loss": 0.8688, + "step": 2917 + }, + { + "epoch": 0.24865786109927568, + "grad_norm": 12.125, + "learning_rate": 2.41137739655703e-06, + "loss": 0.5164, + "step": 2918 + }, + { + "epoch": 0.24874307626757564, + "grad_norm": 12.75, + "learning_rate": 2.4113152425213006e-06, + "loss": 0.6873, + "step": 2919 + }, + { + "epoch": 0.2488282914358756, + "grad_norm": 12.375, + "learning_rate": 2.4112530674994367e-06, + "loss": 0.4927, + "step": 2920 + }, + { + "epoch": 0.24891350660417555, + "grad_norm": 14.75, + "learning_rate": 2.411190871492562e-06, + "loss": 0.4649, + "step": 2921 + }, + { + "epoch": 0.2489987217724755, + "grad_norm": 14.0625, + "learning_rate": 2.4111286545018006e-06, + "loss": 0.8986, + "step": 2922 + }, + { + "epoch": 0.24908393694077546, + "grad_norm": 9.8125, + "learning_rate": 2.411066416528277e-06, + "loss": 0.321, + "step": 2923 + }, + { + "epoch": 0.24916915210907542, + "grad_norm": 16.0, + "learning_rate": 2.4110041575731156e-06, + "loss": 1.0307, + "step": 2924 + }, + { + "epoch": 0.24925436727737538, + "grad_norm": 15.4375, + "learning_rate": 2.4109418776374417e-06, + "loss": 0.7543, + "step": 2925 + }, + { + "epoch": 0.24933958244567533, + "grad_norm": 31.0, + "learning_rate": 2.4108795767223803e-06, + "loss": 1.2791, + "step": 2926 + }, + { + "epoch": 0.2494247976139753, + "grad_norm": 13.3125, + "learning_rate": 2.4108172548290576e-06, + "loss": 0.8779, + "step": 2927 + }, + { + "epoch": 0.24951001278227525, + "grad_norm": 20.625, + "learning_rate": 2.4107549119586e-06, + "loss": 1.0068, + "step": 2928 + }, + { + "epoch": 0.2495952279505752, + "grad_norm": 13.6875, + "learning_rate": 2.410692548112134e-06, + "loss": 0.7455, + "step": 2929 + }, + { + "epoch": 0.24968044311887516, + "grad_norm": 13.9375, + "learning_rate": 2.4106301632907865e-06, + "loss": 0.8551, + "step": 2930 + }, + { + "epoch": 0.24976565828717512, + "grad_norm": 22.125, + "learning_rate": 2.410567757495685e-06, + "loss": 1.1662, + "step": 2931 + }, + { + "epoch": 0.24985087345547508, + "grad_norm": 17.625, + "learning_rate": 2.4105053307279567e-06, + "loss": 1.0824, + "step": 2932 + }, + { + "epoch": 0.24993608862377503, + "grad_norm": 9.6875, + "learning_rate": 2.41044288298873e-06, + "loss": 0.367, + "step": 2933 + }, + { + "epoch": 0.250021303792075, + "grad_norm": 11.8125, + "learning_rate": 2.410380414279134e-06, + "loss": 0.5967, + "step": 2934 + }, + { + "epoch": 0.25010651896037495, + "grad_norm": 15.0625, + "learning_rate": 2.4103179246002967e-06, + "loss": 0.6324, + "step": 2935 + }, + { + "epoch": 0.2501917341286749, + "grad_norm": 15.375, + "learning_rate": 2.410255413953348e-06, + "loss": 0.8187, + "step": 2936 + }, + { + "epoch": 0.25027694929697486, + "grad_norm": 21.0, + "learning_rate": 2.4101928823394167e-06, + "loss": 0.6911, + "step": 2937 + }, + { + "epoch": 0.2503621644652748, + "grad_norm": 19.625, + "learning_rate": 2.410130329759634e-06, + "loss": 1.0849, + "step": 2938 + }, + { + "epoch": 0.2504473796335748, + "grad_norm": 21.875, + "learning_rate": 2.4100677562151294e-06, + "loss": 1.3756, + "step": 2939 + }, + { + "epoch": 0.25053259480187473, + "grad_norm": 17.25, + "learning_rate": 2.4100051617070334e-06, + "loss": 1.3955, + "step": 2940 + }, + { + "epoch": 0.2506178099701747, + "grad_norm": 38.0, + "learning_rate": 2.4099425462364784e-06, + "loss": 1.5249, + "step": 2941 + }, + { + "epoch": 0.25070302513847464, + "grad_norm": 33.25, + "learning_rate": 2.4098799098045948e-06, + "loss": 1.2549, + "step": 2942 + }, + { + "epoch": 0.2507882403067746, + "grad_norm": 12.0, + "learning_rate": 2.4098172524125153e-06, + "loss": 0.566, + "step": 2943 + }, + { + "epoch": 0.25087345547507456, + "grad_norm": 13.625, + "learning_rate": 2.409754574061371e-06, + "loss": 0.7195, + "step": 2944 + }, + { + "epoch": 0.2509586706433745, + "grad_norm": 11.5625, + "learning_rate": 2.4096918747522956e-06, + "loss": 0.5083, + "step": 2945 + }, + { + "epoch": 0.25104388581167447, + "grad_norm": 11.125, + "learning_rate": 2.409629154486422e-06, + "loss": 0.4639, + "step": 2946 + }, + { + "epoch": 0.25112910097997443, + "grad_norm": 13.75, + "learning_rate": 2.409566413264884e-06, + "loss": 0.6159, + "step": 2947 + }, + { + "epoch": 0.2512143161482744, + "grad_norm": 18.125, + "learning_rate": 2.409503651088814e-06, + "loss": 0.7635, + "step": 2948 + }, + { + "epoch": 0.25129953131657434, + "grad_norm": 14.75, + "learning_rate": 2.4094408679593477e-06, + "loss": 0.8745, + "step": 2949 + }, + { + "epoch": 0.2513847464848743, + "grad_norm": 13.0625, + "learning_rate": 2.4093780638776188e-06, + "loss": 1.0396, + "step": 2950 + }, + { + "epoch": 0.25146996165317426, + "grad_norm": 14.6875, + "learning_rate": 2.409315238844762e-06, + "loss": 0.5989, + "step": 2951 + }, + { + "epoch": 0.2515551768214742, + "grad_norm": 14.75, + "learning_rate": 2.4092523928619137e-06, + "loss": 1.1529, + "step": 2952 + }, + { + "epoch": 0.25164039198977417, + "grad_norm": 18.25, + "learning_rate": 2.4091895259302084e-06, + "loss": 1.2555, + "step": 2953 + }, + { + "epoch": 0.2517256071580741, + "grad_norm": 13.9375, + "learning_rate": 2.409126638050783e-06, + "loss": 0.6195, + "step": 2954 + }, + { + "epoch": 0.2518108223263741, + "grad_norm": 16.5, + "learning_rate": 2.4090637292247737e-06, + "loss": 0.9051, + "step": 2955 + }, + { + "epoch": 0.25189603749467404, + "grad_norm": 13.5625, + "learning_rate": 2.409000799453317e-06, + "loss": 0.8554, + "step": 2956 + }, + { + "epoch": 0.251981252662974, + "grad_norm": 13.6875, + "learning_rate": 2.408937848737551e-06, + "loss": 0.9069, + "step": 2957 + }, + { + "epoch": 0.25206646783127395, + "grad_norm": 16.375, + "learning_rate": 2.408874877078612e-06, + "loss": 1.0019, + "step": 2958 + }, + { + "epoch": 0.2521516829995739, + "grad_norm": 12.5625, + "learning_rate": 2.408811884477639e-06, + "loss": 0.7156, + "step": 2959 + }, + { + "epoch": 0.25223689816787387, + "grad_norm": 18.25, + "learning_rate": 2.4087488709357698e-06, + "loss": 0.95, + "step": 2960 + }, + { + "epoch": 0.2523221133361738, + "grad_norm": 19.0, + "learning_rate": 2.4086858364541437e-06, + "loss": 1.1111, + "step": 2961 + }, + { + "epoch": 0.2524073285044738, + "grad_norm": 13.9375, + "learning_rate": 2.4086227810338987e-06, + "loss": 0.7465, + "step": 2962 + }, + { + "epoch": 0.25249254367277374, + "grad_norm": 12.25, + "learning_rate": 2.4085597046761754e-06, + "loss": 0.5839, + "step": 2963 + }, + { + "epoch": 0.2525777588410737, + "grad_norm": 17.5, + "learning_rate": 2.408496607382113e-06, + "loss": 0.939, + "step": 2964 + }, + { + "epoch": 0.25266297400937365, + "grad_norm": 17.0, + "learning_rate": 2.408433489152852e-06, + "loss": 1.0018, + "step": 2965 + }, + { + "epoch": 0.2527481891776736, + "grad_norm": 17.625, + "learning_rate": 2.4083703499895335e-06, + "loss": 1.0089, + "step": 2966 + }, + { + "epoch": 0.25283340434597357, + "grad_norm": 14.375, + "learning_rate": 2.4083071898932975e-06, + "loss": 1.0389, + "step": 2967 + }, + { + "epoch": 0.2529186195142735, + "grad_norm": 15.625, + "learning_rate": 2.4082440088652857e-06, + "loss": 0.7878, + "step": 2968 + }, + { + "epoch": 0.2530038346825735, + "grad_norm": 14.0625, + "learning_rate": 2.4081808069066396e-06, + "loss": 1.1512, + "step": 2969 + }, + { + "epoch": 0.25308904985087344, + "grad_norm": 17.5, + "learning_rate": 2.4081175840185023e-06, + "loss": 0.9583, + "step": 2970 + }, + { + "epoch": 0.2531742650191734, + "grad_norm": 16.625, + "learning_rate": 2.4080543402020155e-06, + "loss": 0.4247, + "step": 2971 + }, + { + "epoch": 0.25325948018747335, + "grad_norm": 11.125, + "learning_rate": 2.4079910754583223e-06, + "loss": 0.4451, + "step": 2972 + }, + { + "epoch": 0.2533446953557733, + "grad_norm": 17.625, + "learning_rate": 2.4079277897885662e-06, + "loss": 0.6546, + "step": 2973 + }, + { + "epoch": 0.25342991052407327, + "grad_norm": 12.625, + "learning_rate": 2.40786448319389e-06, + "loss": 0.5652, + "step": 2974 + }, + { + "epoch": 0.2535151256923732, + "grad_norm": 19.25, + "learning_rate": 2.4078011556754386e-06, + "loss": 0.7326, + "step": 2975 + }, + { + "epoch": 0.2536003408606732, + "grad_norm": 13.75, + "learning_rate": 2.407737807234356e-06, + "loss": 0.5198, + "step": 2976 + }, + { + "epoch": 0.25368555602897314, + "grad_norm": 14.9375, + "learning_rate": 2.407674437871787e-06, + "loss": 1.0169, + "step": 2977 + }, + { + "epoch": 0.2537707711972731, + "grad_norm": 11.1875, + "learning_rate": 2.4076110475888763e-06, + "loss": 0.3128, + "step": 2978 + }, + { + "epoch": 0.25385598636557305, + "grad_norm": 15.8125, + "learning_rate": 2.4075476363867707e-06, + "loss": 0.8734, + "step": 2979 + }, + { + "epoch": 0.253941201533873, + "grad_norm": 14.1875, + "learning_rate": 2.4074842042666147e-06, + "loss": 0.8517, + "step": 2980 + }, + { + "epoch": 0.25402641670217296, + "grad_norm": 16.375, + "learning_rate": 2.4074207512295557e-06, + "loss": 0.8789, + "step": 2981 + }, + { + "epoch": 0.2541116318704729, + "grad_norm": 14.0625, + "learning_rate": 2.4073572772767396e-06, + "loss": 0.604, + "step": 2982 + }, + { + "epoch": 0.2541968470387729, + "grad_norm": 33.75, + "learning_rate": 2.4072937824093138e-06, + "loss": 0.9227, + "step": 2983 + }, + { + "epoch": 0.25428206220707283, + "grad_norm": 18.625, + "learning_rate": 2.4072302666284255e-06, + "loss": 0.8889, + "step": 2984 + }, + { + "epoch": 0.2543672773753728, + "grad_norm": 16.875, + "learning_rate": 2.4071667299352225e-06, + "loss": 0.7892, + "step": 2985 + }, + { + "epoch": 0.25445249254367275, + "grad_norm": 16.125, + "learning_rate": 2.407103172330853e-06, + "loss": 0.647, + "step": 2986 + }, + { + "epoch": 0.2545377077119727, + "grad_norm": 19.625, + "learning_rate": 2.407039593816466e-06, + "loss": 1.0276, + "step": 2987 + }, + { + "epoch": 0.25462292288027266, + "grad_norm": 18.125, + "learning_rate": 2.4069759943932096e-06, + "loss": 0.5737, + "step": 2988 + }, + { + "epoch": 0.2547081380485726, + "grad_norm": 11.5, + "learning_rate": 2.406912374062234e-06, + "loss": 0.8007, + "step": 2989 + }, + { + "epoch": 0.2547933532168726, + "grad_norm": 11.625, + "learning_rate": 2.406848732824688e-06, + "loss": 0.4117, + "step": 2990 + }, + { + "epoch": 0.2548785683851726, + "grad_norm": 14.25, + "learning_rate": 2.4067850706817224e-06, + "loss": 0.6301, + "step": 2991 + }, + { + "epoch": 0.25496378355347254, + "grad_norm": 15.1875, + "learning_rate": 2.4067213876344873e-06, + "loss": 0.7766, + "step": 2992 + }, + { + "epoch": 0.2550489987217725, + "grad_norm": 16.75, + "learning_rate": 2.4066576836841334e-06, + "loss": 0.9498, + "step": 2993 + }, + { + "epoch": 0.25513421389007246, + "grad_norm": 17.75, + "learning_rate": 2.406593958831812e-06, + "loss": 0.9111, + "step": 2994 + }, + { + "epoch": 0.2552194290583724, + "grad_norm": 20.625, + "learning_rate": 2.406530213078675e-06, + "loss": 0.7957, + "step": 2995 + }, + { + "epoch": 0.2553046442266724, + "grad_norm": 14.4375, + "learning_rate": 2.406466446425874e-06, + "loss": 0.618, + "step": 2996 + }, + { + "epoch": 0.25538985939497233, + "grad_norm": 14.1875, + "learning_rate": 2.4064026588745616e-06, + "loss": 0.6432, + "step": 2997 + }, + { + "epoch": 0.2554750745632723, + "grad_norm": 13.4375, + "learning_rate": 2.4063388504258897e-06, + "loss": 0.7925, + "step": 2998 + }, + { + "epoch": 0.25556028973157224, + "grad_norm": 18.125, + "learning_rate": 2.4062750210810124e-06, + "loss": 1.2936, + "step": 2999 + }, + { + "epoch": 0.2556455048998722, + "grad_norm": 14.875, + "learning_rate": 2.4062111708410825e-06, + "loss": 0.8411, + "step": 3000 + }, + { + "epoch": 0.25573072006817216, + "grad_norm": 10.875, + "learning_rate": 2.406147299707254e-06, + "loss": 0.4682, + "step": 3001 + }, + { + "epoch": 0.2558159352364721, + "grad_norm": 11.8125, + "learning_rate": 2.406083407680682e-06, + "loss": 0.6219, + "step": 3002 + }, + { + "epoch": 0.25590115040477207, + "grad_norm": 17.0, + "learning_rate": 2.4060194947625195e-06, + "loss": 0.9872, + "step": 3003 + }, + { + "epoch": 0.25598636557307203, + "grad_norm": 13.5625, + "learning_rate": 2.4059555609539226e-06, + "loss": 0.6533, + "step": 3004 + }, + { + "epoch": 0.256071580741372, + "grad_norm": 17.875, + "learning_rate": 2.405891606256046e-06, + "loss": 1.0268, + "step": 3005 + }, + { + "epoch": 0.25615679590967194, + "grad_norm": 16.625, + "learning_rate": 2.4058276306700462e-06, + "loss": 0.8031, + "step": 3006 + }, + { + "epoch": 0.2562420110779719, + "grad_norm": 19.625, + "learning_rate": 2.4057636341970786e-06, + "loss": 0.8053, + "step": 3007 + }, + { + "epoch": 0.25632722624627186, + "grad_norm": 71.5, + "learning_rate": 2.4056996168382996e-06, + "loss": 0.8514, + "step": 3008 + }, + { + "epoch": 0.2564124414145718, + "grad_norm": 17.75, + "learning_rate": 2.405635578594867e-06, + "loss": 1.1079, + "step": 3009 + }, + { + "epoch": 0.25649765658287177, + "grad_norm": 15.0, + "learning_rate": 2.405571519467937e-06, + "loss": 0.8001, + "step": 3010 + }, + { + "epoch": 0.2565828717511717, + "grad_norm": 16.75, + "learning_rate": 2.4055074394586676e-06, + "loss": 0.8068, + "step": 3011 + }, + { + "epoch": 0.2566680869194717, + "grad_norm": 12.0625, + "learning_rate": 2.405443338568217e-06, + "loss": 0.5894, + "step": 3012 + }, + { + "epoch": 0.25675330208777164, + "grad_norm": 16.5, + "learning_rate": 2.405379216797743e-06, + "loss": 0.7341, + "step": 3013 + }, + { + "epoch": 0.2568385172560716, + "grad_norm": 14.5, + "learning_rate": 2.4053150741484056e-06, + "loss": 0.6545, + "step": 3014 + }, + { + "epoch": 0.25692373242437155, + "grad_norm": 14.5625, + "learning_rate": 2.4052509106213623e-06, + "loss": 0.7748, + "step": 3015 + }, + { + "epoch": 0.2570089475926715, + "grad_norm": 28.125, + "learning_rate": 2.4051867262177736e-06, + "loss": 1.1048, + "step": 3016 + }, + { + "epoch": 0.25709416276097147, + "grad_norm": 11.3125, + "learning_rate": 2.4051225209387992e-06, + "loss": 0.6033, + "step": 3017 + }, + { + "epoch": 0.2571793779292714, + "grad_norm": 23.375, + "learning_rate": 2.4050582947855987e-06, + "loss": 1.2305, + "step": 3018 + }, + { + "epoch": 0.2572645930975714, + "grad_norm": 22.75, + "learning_rate": 2.404994047759334e-06, + "loss": 0.9808, + "step": 3019 + }, + { + "epoch": 0.25734980826587134, + "grad_norm": 14.4375, + "learning_rate": 2.4049297798611654e-06, + "loss": 0.8482, + "step": 3020 + }, + { + "epoch": 0.2574350234341713, + "grad_norm": 14.8125, + "learning_rate": 2.4048654910922537e-06, + "loss": 0.5807, + "step": 3021 + }, + { + "epoch": 0.25752023860247125, + "grad_norm": 12.0625, + "learning_rate": 2.404801181453762e-06, + "loss": 0.7013, + "step": 3022 + }, + { + "epoch": 0.2576054537707712, + "grad_norm": 17.375, + "learning_rate": 2.404736850946851e-06, + "loss": 0.7474, + "step": 3023 + }, + { + "epoch": 0.25769066893907117, + "grad_norm": 15.125, + "learning_rate": 2.4046724995726844e-06, + "loss": 0.6789, + "step": 3024 + }, + { + "epoch": 0.2577758841073711, + "grad_norm": 10.875, + "learning_rate": 2.4046081273324246e-06, + "loss": 0.394, + "step": 3025 + }, + { + "epoch": 0.2578610992756711, + "grad_norm": 13.0, + "learning_rate": 2.404543734227235e-06, + "loss": 0.8614, + "step": 3026 + }, + { + "epoch": 0.25794631444397104, + "grad_norm": 13.125, + "learning_rate": 2.404479320258279e-06, + "loss": 0.4764, + "step": 3027 + }, + { + "epoch": 0.258031529612271, + "grad_norm": 13.375, + "learning_rate": 2.4044148854267203e-06, + "loss": 0.7893, + "step": 3028 + }, + { + "epoch": 0.25811674478057095, + "grad_norm": 14.625, + "learning_rate": 2.4043504297337244e-06, + "loss": 0.4692, + "step": 3029 + }, + { + "epoch": 0.2582019599488709, + "grad_norm": 14.8125, + "learning_rate": 2.404285953180455e-06, + "loss": 0.8938, + "step": 3030 + }, + { + "epoch": 0.25828717511717086, + "grad_norm": 12.5, + "learning_rate": 2.4042214557680783e-06, + "loss": 0.6956, + "step": 3031 + }, + { + "epoch": 0.2583723902854708, + "grad_norm": 21.5, + "learning_rate": 2.4041569374977587e-06, + "loss": 1.0409, + "step": 3032 + }, + { + "epoch": 0.2584576054537708, + "grad_norm": 18.25, + "learning_rate": 2.4040923983706626e-06, + "loss": 1.004, + "step": 3033 + }, + { + "epoch": 0.25854282062207073, + "grad_norm": 19.375, + "learning_rate": 2.404027838387957e-06, + "loss": 1.1083, + "step": 3034 + }, + { + "epoch": 0.2586280357903707, + "grad_norm": 17.125, + "learning_rate": 2.403963257550807e-06, + "loss": 0.7026, + "step": 3035 + }, + { + "epoch": 0.25871325095867065, + "grad_norm": 10.9375, + "learning_rate": 2.403898655860381e-06, + "loss": 0.4245, + "step": 3036 + }, + { + "epoch": 0.2587984661269706, + "grad_norm": 15.6875, + "learning_rate": 2.403834033317846e-06, + "loss": 0.6608, + "step": 3037 + }, + { + "epoch": 0.25888368129527056, + "grad_norm": 12.9375, + "learning_rate": 2.4037693899243696e-06, + "loss": 0.6097, + "step": 3038 + }, + { + "epoch": 0.2589688964635705, + "grad_norm": 21.0, + "learning_rate": 2.4037047256811204e-06, + "loss": 1.1294, + "step": 3039 + }, + { + "epoch": 0.2590541116318705, + "grad_norm": 12.8125, + "learning_rate": 2.403640040589266e-06, + "loss": 0.6969, + "step": 3040 + }, + { + "epoch": 0.25913932680017043, + "grad_norm": 12.5625, + "learning_rate": 2.403575334649977e-06, + "loss": 0.7207, + "step": 3041 + }, + { + "epoch": 0.2592245419684704, + "grad_norm": 14.625, + "learning_rate": 2.403510607864421e-06, + "loss": 0.8406, + "step": 3042 + }, + { + "epoch": 0.25930975713677035, + "grad_norm": 18.75, + "learning_rate": 2.4034458602337684e-06, + "loss": 1.0368, + "step": 3043 + }, + { + "epoch": 0.2593949723050703, + "grad_norm": 18.75, + "learning_rate": 2.403381091759189e-06, + "loss": 1.1469, + "step": 3044 + }, + { + "epoch": 0.25948018747337026, + "grad_norm": 15.5625, + "learning_rate": 2.4033163024418543e-06, + "loss": 1.0348, + "step": 3045 + }, + { + "epoch": 0.2595654026416702, + "grad_norm": 13.8125, + "learning_rate": 2.4032514922829337e-06, + "loss": 0.8978, + "step": 3046 + }, + { + "epoch": 0.2596506178099702, + "grad_norm": 14.5625, + "learning_rate": 2.4031866612835993e-06, + "loss": 0.1999, + "step": 3047 + }, + { + "epoch": 0.25973583297827013, + "grad_norm": 11.125, + "learning_rate": 2.4031218094450215e-06, + "loss": 0.4699, + "step": 3048 + }, + { + "epoch": 0.2598210481465701, + "grad_norm": 20.125, + "learning_rate": 2.4030569367683736e-06, + "loss": 0.6024, + "step": 3049 + }, + { + "epoch": 0.25990626331487005, + "grad_norm": 33.75, + "learning_rate": 2.4029920432548275e-06, + "loss": 1.1232, + "step": 3050 + }, + { + "epoch": 0.25999147848317, + "grad_norm": 14.1875, + "learning_rate": 2.4029271289055558e-06, + "loss": 0.6519, + "step": 3051 + }, + { + "epoch": 0.26007669365146996, + "grad_norm": 14.25, + "learning_rate": 2.4028621937217312e-06, + "loss": 0.5323, + "step": 3052 + }, + { + "epoch": 0.2601619088197699, + "grad_norm": 12.8125, + "learning_rate": 2.4027972377045274e-06, + "loss": 0.5999, + "step": 3053 + }, + { + "epoch": 0.2602471239880699, + "grad_norm": 16.125, + "learning_rate": 2.4027322608551183e-06, + "loss": 0.6853, + "step": 3054 + }, + { + "epoch": 0.26033233915636983, + "grad_norm": 14.5625, + "learning_rate": 2.4026672631746783e-06, + "loss": 0.8008, + "step": 3055 + }, + { + "epoch": 0.2604175543246698, + "grad_norm": 17.375, + "learning_rate": 2.402602244664382e-06, + "loss": 0.6959, + "step": 3056 + }, + { + "epoch": 0.26050276949296974, + "grad_norm": 14.5625, + "learning_rate": 2.402537205325404e-06, + "loss": 0.8843, + "step": 3057 + }, + { + "epoch": 0.2605879846612697, + "grad_norm": 11.8125, + "learning_rate": 2.4024721451589194e-06, + "loss": 0.4014, + "step": 3058 + }, + { + "epoch": 0.26067319982956966, + "grad_norm": 15.25, + "learning_rate": 2.4024070641661043e-06, + "loss": 0.9268, + "step": 3059 + }, + { + "epoch": 0.2607584149978696, + "grad_norm": 13.125, + "learning_rate": 2.4023419623481346e-06, + "loss": 0.3544, + "step": 3060 + }, + { + "epoch": 0.26084363016616957, + "grad_norm": 14.6875, + "learning_rate": 2.4022768397061873e-06, + "loss": 0.7147, + "step": 3061 + }, + { + "epoch": 0.26092884533446953, + "grad_norm": 17.875, + "learning_rate": 2.4022116962414384e-06, + "loss": 0.7709, + "step": 3062 + }, + { + "epoch": 0.2610140605027695, + "grad_norm": 13.375, + "learning_rate": 2.402146531955066e-06, + "loss": 0.3639, + "step": 3063 + }, + { + "epoch": 0.26109927567106944, + "grad_norm": 17.5, + "learning_rate": 2.402081346848247e-06, + "loss": 0.7288, + "step": 3064 + }, + { + "epoch": 0.2611844908393694, + "grad_norm": 26.5, + "learning_rate": 2.4020161409221594e-06, + "loss": 1.3285, + "step": 3065 + }, + { + "epoch": 0.26126970600766936, + "grad_norm": 20.75, + "learning_rate": 2.4019509141779816e-06, + "loss": 1.0347, + "step": 3066 + }, + { + "epoch": 0.2613549211759693, + "grad_norm": 13.8125, + "learning_rate": 2.4018856666168925e-06, + "loss": 0.8393, + "step": 3067 + }, + { + "epoch": 0.26144013634426927, + "grad_norm": 16.125, + "learning_rate": 2.4018203982400713e-06, + "loss": 0.7065, + "step": 3068 + }, + { + "epoch": 0.2615253515125692, + "grad_norm": 14.0625, + "learning_rate": 2.401755109048697e-06, + "loss": 0.6089, + "step": 3069 + }, + { + "epoch": 0.2616105666808692, + "grad_norm": 14.4375, + "learning_rate": 2.4016897990439498e-06, + "loss": 0.587, + "step": 3070 + }, + { + "epoch": 0.26169578184916914, + "grad_norm": 16.25, + "learning_rate": 2.4016244682270097e-06, + "loss": 0.8635, + "step": 3071 + }, + { + "epoch": 0.2617809970174691, + "grad_norm": 30.0, + "learning_rate": 2.4015591165990576e-06, + "loss": 0.9933, + "step": 3072 + }, + { + "epoch": 0.26186621218576905, + "grad_norm": 12.6875, + "learning_rate": 2.401493744161274e-06, + "loss": 0.4795, + "step": 3073 + }, + { + "epoch": 0.261951427354069, + "grad_norm": 13.5, + "learning_rate": 2.4014283509148406e-06, + "loss": 0.5747, + "step": 3074 + }, + { + "epoch": 0.26203664252236897, + "grad_norm": 15.75, + "learning_rate": 2.4013629368609393e-06, + "loss": 0.5161, + "step": 3075 + }, + { + "epoch": 0.2621218576906689, + "grad_norm": 16.375, + "learning_rate": 2.4012975020007516e-06, + "loss": 0.6793, + "step": 3076 + }, + { + "epoch": 0.2622070728589689, + "grad_norm": 12.0, + "learning_rate": 2.4012320463354606e-06, + "loss": 0.4181, + "step": 3077 + }, + { + "epoch": 0.26229228802726884, + "grad_norm": 18.0, + "learning_rate": 2.401166569866249e-06, + "loss": 0.6657, + "step": 3078 + }, + { + "epoch": 0.2623775031955688, + "grad_norm": 26.25, + "learning_rate": 2.4011010725942998e-06, + "loss": 1.0912, + "step": 3079 + }, + { + "epoch": 0.26246271836386875, + "grad_norm": 19.25, + "learning_rate": 2.4010355545207965e-06, + "loss": 0.8171, + "step": 3080 + }, + { + "epoch": 0.2625479335321687, + "grad_norm": 15.0625, + "learning_rate": 2.4009700156469233e-06, + "loss": 0.6955, + "step": 3081 + }, + { + "epoch": 0.26263314870046867, + "grad_norm": 15.5, + "learning_rate": 2.4009044559738646e-06, + "loss": 0.688, + "step": 3082 + }, + { + "epoch": 0.2627183638687686, + "grad_norm": 13.0, + "learning_rate": 2.400838875502805e-06, + "loss": 0.6327, + "step": 3083 + }, + { + "epoch": 0.2628035790370686, + "grad_norm": 16.375, + "learning_rate": 2.40077327423493e-06, + "loss": 1.2231, + "step": 3084 + }, + { + "epoch": 0.26288879420536854, + "grad_norm": 16.75, + "learning_rate": 2.400707652171424e-06, + "loss": 1.0987, + "step": 3085 + }, + { + "epoch": 0.2629740093736685, + "grad_norm": 15.6875, + "learning_rate": 2.4006420093134744e-06, + "loss": 0.5817, + "step": 3086 + }, + { + "epoch": 0.26305922454196845, + "grad_norm": 17.625, + "learning_rate": 2.4005763456622663e-06, + "loss": 0.8865, + "step": 3087 + }, + { + "epoch": 0.2631444397102684, + "grad_norm": 12.75, + "learning_rate": 2.4005106612189866e-06, + "loss": 0.5989, + "step": 3088 + }, + { + "epoch": 0.26322965487856836, + "grad_norm": 16.5, + "learning_rate": 2.4004449559848224e-06, + "loss": 1.0143, + "step": 3089 + }, + { + "epoch": 0.2633148700468683, + "grad_norm": 11.9375, + "learning_rate": 2.400379229960961e-06, + "loss": 0.4729, + "step": 3090 + }, + { + "epoch": 0.2634000852151683, + "grad_norm": 17.0, + "learning_rate": 2.40031348314859e-06, + "loss": 0.7569, + "step": 3091 + }, + { + "epoch": 0.26348530038346824, + "grad_norm": 24.75, + "learning_rate": 2.4002477155488976e-06, + "loss": 0.9389, + "step": 3092 + }, + { + "epoch": 0.2635705155517682, + "grad_norm": 13.0625, + "learning_rate": 2.4001819271630726e-06, + "loss": 0.7724, + "step": 3093 + }, + { + "epoch": 0.26365573072006815, + "grad_norm": 12.125, + "learning_rate": 2.4001161179923034e-06, + "loss": 0.6507, + "step": 3094 + }, + { + "epoch": 0.2637409458883681, + "grad_norm": 11.3125, + "learning_rate": 2.400050288037779e-06, + "loss": 0.506, + "step": 3095 + }, + { + "epoch": 0.26382616105666806, + "grad_norm": 13.6875, + "learning_rate": 2.3999844373006904e-06, + "loss": 0.5859, + "step": 3096 + }, + { + "epoch": 0.263911376224968, + "grad_norm": 12.9375, + "learning_rate": 2.3999185657822256e-06, + "loss": 0.6689, + "step": 3097 + }, + { + "epoch": 0.263996591393268, + "grad_norm": 12.25, + "learning_rate": 2.3998526734835765e-06, + "loss": 0.7123, + "step": 3098 + }, + { + "epoch": 0.26408180656156793, + "grad_norm": 16.875, + "learning_rate": 2.3997867604059335e-06, + "loss": 0.7335, + "step": 3099 + }, + { + "epoch": 0.2641670217298679, + "grad_norm": 17.25, + "learning_rate": 2.399720826550487e-06, + "loss": 0.9156, + "step": 3100 + }, + { + "epoch": 0.26425223689816785, + "grad_norm": 17.75, + "learning_rate": 2.3996548719184293e-06, + "loss": 0.7847, + "step": 3101 + }, + { + "epoch": 0.2643374520664678, + "grad_norm": 14.4375, + "learning_rate": 2.399588896510952e-06, + "loss": 0.7428, + "step": 3102 + }, + { + "epoch": 0.26442266723476776, + "grad_norm": 12.1875, + "learning_rate": 2.3995229003292473e-06, + "loss": 0.6125, + "step": 3103 + }, + { + "epoch": 0.2645078824030678, + "grad_norm": 12.75, + "learning_rate": 2.3994568833745085e-06, + "loss": 0.5944, + "step": 3104 + }, + { + "epoch": 0.26459309757136773, + "grad_norm": 15.25, + "learning_rate": 2.399390845647927e-06, + "loss": 0.7549, + "step": 3105 + }, + { + "epoch": 0.2646783127396677, + "grad_norm": 36.25, + "learning_rate": 2.3993247871506976e-06, + "loss": 1.4105, + "step": 3106 + }, + { + "epoch": 0.26476352790796764, + "grad_norm": 13.0, + "learning_rate": 2.3992587078840134e-06, + "loss": 0.7736, + "step": 3107 + }, + { + "epoch": 0.2648487430762676, + "grad_norm": 17.625, + "learning_rate": 2.3991926078490686e-06, + "loss": 0.9964, + "step": 3108 + }, + { + "epoch": 0.26493395824456756, + "grad_norm": 11.3125, + "learning_rate": 2.3991264870470583e-06, + "loss": 0.5037, + "step": 3109 + }, + { + "epoch": 0.2650191734128675, + "grad_norm": 15.375, + "learning_rate": 2.399060345479176e-06, + "loss": 0.4565, + "step": 3110 + }, + { + "epoch": 0.26510438858116747, + "grad_norm": 10.3125, + "learning_rate": 2.3989941831466184e-06, + "loss": 0.4318, + "step": 3111 + }, + { + "epoch": 0.26518960374946743, + "grad_norm": 10.5625, + "learning_rate": 2.3989280000505805e-06, + "loss": 0.3038, + "step": 3112 + }, + { + "epoch": 0.2652748189177674, + "grad_norm": 11.8125, + "learning_rate": 2.3988617961922578e-06, + "loss": 0.5742, + "step": 3113 + }, + { + "epoch": 0.26536003408606734, + "grad_norm": 14.0, + "learning_rate": 2.398795571572848e-06, + "loss": 0.8746, + "step": 3114 + }, + { + "epoch": 0.2654452492543673, + "grad_norm": 17.125, + "learning_rate": 2.398729326193546e-06, + "loss": 1.365, + "step": 3115 + }, + { + "epoch": 0.26553046442266726, + "grad_norm": 22.375, + "learning_rate": 2.398663060055551e-06, + "loss": 1.0579, + "step": 3116 + }, + { + "epoch": 0.2656156795909672, + "grad_norm": 32.0, + "learning_rate": 2.3985967731600586e-06, + "loss": 0.9658, + "step": 3117 + }, + { + "epoch": 0.26570089475926717, + "grad_norm": 21.5, + "learning_rate": 2.398530465508268e-06, + "loss": 0.747, + "step": 3118 + }, + { + "epoch": 0.2657861099275671, + "grad_norm": 17.625, + "learning_rate": 2.3984641371013765e-06, + "loss": 0.9848, + "step": 3119 + }, + { + "epoch": 0.2658713250958671, + "grad_norm": 20.875, + "learning_rate": 2.3983977879405834e-06, + "loss": 0.879, + "step": 3120 + }, + { + "epoch": 0.26595654026416704, + "grad_norm": 13.0, + "learning_rate": 2.398331418027088e-06, + "loss": 0.375, + "step": 3121 + }, + { + "epoch": 0.266041755432467, + "grad_norm": 11.3125, + "learning_rate": 2.3982650273620882e-06, + "loss": 0.5694, + "step": 3122 + }, + { + "epoch": 0.26612697060076695, + "grad_norm": 14.3125, + "learning_rate": 2.398198615946785e-06, + "loss": 0.6548, + "step": 3123 + }, + { + "epoch": 0.2662121857690669, + "grad_norm": 16.25, + "learning_rate": 2.3981321837823784e-06, + "loss": 0.7613, + "step": 3124 + }, + { + "epoch": 0.26629740093736687, + "grad_norm": 22.625, + "learning_rate": 2.398065730870069e-06, + "loss": 0.8212, + "step": 3125 + }, + { + "epoch": 0.2663826161056668, + "grad_norm": 21.75, + "learning_rate": 2.397999257211057e-06, + "loss": 0.922, + "step": 3126 + }, + { + "epoch": 0.2664678312739668, + "grad_norm": 12.1875, + "learning_rate": 2.397932762806544e-06, + "loss": 0.5572, + "step": 3127 + }, + { + "epoch": 0.26655304644226674, + "grad_norm": 24.75, + "learning_rate": 2.3978662476577313e-06, + "loss": 0.9428, + "step": 3128 + }, + { + "epoch": 0.2666382616105667, + "grad_norm": 15.5625, + "learning_rate": 2.397799711765822e-06, + "loss": 0.9027, + "step": 3129 + }, + { + "epoch": 0.26672347677886665, + "grad_norm": 14.8125, + "learning_rate": 2.397733155132017e-06, + "loss": 0.6059, + "step": 3130 + }, + { + "epoch": 0.2668086919471666, + "grad_norm": 13.8125, + "learning_rate": 2.3976665777575203e-06, + "loss": 0.9804, + "step": 3131 + }, + { + "epoch": 0.26689390711546657, + "grad_norm": 10.75, + "learning_rate": 2.3975999796435343e-06, + "loss": 0.3274, + "step": 3132 + }, + { + "epoch": 0.2669791222837665, + "grad_norm": 12.625, + "learning_rate": 2.397533360791263e-06, + "loss": 0.5653, + "step": 3133 + }, + { + "epoch": 0.2670643374520665, + "grad_norm": 15.8125, + "learning_rate": 2.397466721201909e-06, + "loss": 0.5945, + "step": 3134 + }, + { + "epoch": 0.26714955262036644, + "grad_norm": 10.5, + "learning_rate": 2.3974000608766777e-06, + "loss": 0.5088, + "step": 3135 + }, + { + "epoch": 0.2672347677886664, + "grad_norm": 12.8125, + "learning_rate": 2.3973333798167744e-06, + "loss": 0.5068, + "step": 3136 + }, + { + "epoch": 0.26731998295696635, + "grad_norm": 14.625, + "learning_rate": 2.3972666780234026e-06, + "loss": 0.905, + "step": 3137 + }, + { + "epoch": 0.2674051981252663, + "grad_norm": 13.125, + "learning_rate": 2.3971999554977683e-06, + "loss": 0.713, + "step": 3138 + }, + { + "epoch": 0.26749041329356626, + "grad_norm": 11.4375, + "learning_rate": 2.397133212241077e-06, + "loss": 0.3716, + "step": 3139 + }, + { + "epoch": 0.2675756284618662, + "grad_norm": 21.375, + "learning_rate": 2.3970664482545352e-06, + "loss": 0.9597, + "step": 3140 + }, + { + "epoch": 0.2676608436301662, + "grad_norm": 11.25, + "learning_rate": 2.3969996635393493e-06, + "loss": 0.4993, + "step": 3141 + }, + { + "epoch": 0.26774605879846614, + "grad_norm": 13.125, + "learning_rate": 2.396932858096726e-06, + "loss": 0.7593, + "step": 3142 + }, + { + "epoch": 0.2678312739667661, + "grad_norm": 15.875, + "learning_rate": 2.3968660319278725e-06, + "loss": 0.7841, + "step": 3143 + }, + { + "epoch": 0.26791648913506605, + "grad_norm": 16.25, + "learning_rate": 2.3967991850339965e-06, + "loss": 0.7095, + "step": 3144 + }, + { + "epoch": 0.268001704303366, + "grad_norm": 17.625, + "learning_rate": 2.3967323174163066e-06, + "loss": 1.0058, + "step": 3145 + }, + { + "epoch": 0.26808691947166596, + "grad_norm": 10.8125, + "learning_rate": 2.39666542907601e-06, + "loss": 0.3677, + "step": 3146 + }, + { + "epoch": 0.2681721346399659, + "grad_norm": 13.5, + "learning_rate": 2.3965985200143163e-06, + "loss": 0.4982, + "step": 3147 + }, + { + "epoch": 0.2682573498082659, + "grad_norm": 22.75, + "learning_rate": 2.396531590232434e-06, + "loss": 0.8559, + "step": 3148 + }, + { + "epoch": 0.26834256497656583, + "grad_norm": 17.375, + "learning_rate": 2.3964646397315735e-06, + "loss": 1.0388, + "step": 3149 + }, + { + "epoch": 0.2684277801448658, + "grad_norm": 13.0625, + "learning_rate": 2.3963976685129437e-06, + "loss": 0.6354, + "step": 3150 + }, + { + "epoch": 0.26851299531316575, + "grad_norm": 13.625, + "learning_rate": 2.3963306765777554e-06, + "loss": 0.4669, + "step": 3151 + }, + { + "epoch": 0.2685982104814657, + "grad_norm": 15.3125, + "learning_rate": 2.3962636639272193e-06, + "loss": 0.9461, + "step": 3152 + }, + { + "epoch": 0.26868342564976566, + "grad_norm": 13.9375, + "learning_rate": 2.396196630562546e-06, + "loss": 0.6262, + "step": 3153 + }, + { + "epoch": 0.2687686408180656, + "grad_norm": 10.9375, + "learning_rate": 2.3961295764849462e-06, + "loss": 0.374, + "step": 3154 + }, + { + "epoch": 0.2688538559863656, + "grad_norm": 14.5, + "learning_rate": 2.3960625016956335e-06, + "loss": 0.9857, + "step": 3155 + }, + { + "epoch": 0.26893907115466553, + "grad_norm": 11.375, + "learning_rate": 2.3959954061958184e-06, + "loss": 0.5476, + "step": 3156 + }, + { + "epoch": 0.2690242863229655, + "grad_norm": 11.75, + "learning_rate": 2.395928289986714e-06, + "loss": 0.741, + "step": 3157 + }, + { + "epoch": 0.26910950149126545, + "grad_norm": 13.8125, + "learning_rate": 2.395861153069533e-06, + "loss": 0.7711, + "step": 3158 + }, + { + "epoch": 0.2691947166595654, + "grad_norm": 11.375, + "learning_rate": 2.395793995445489e-06, + "loss": 0.6454, + "step": 3159 + }, + { + "epoch": 0.26927993182786536, + "grad_norm": 13.5625, + "learning_rate": 2.395726817115795e-06, + "loss": 0.6697, + "step": 3160 + }, + { + "epoch": 0.2693651469961653, + "grad_norm": 15.3125, + "learning_rate": 2.3956596180816655e-06, + "loss": 0.7545, + "step": 3161 + }, + { + "epoch": 0.2694503621644653, + "grad_norm": 12.625, + "learning_rate": 2.395592398344315e-06, + "loss": 0.6582, + "step": 3162 + }, + { + "epoch": 0.26953557733276523, + "grad_norm": 12.8125, + "learning_rate": 2.395525157904957e-06, + "loss": 0.6127, + "step": 3163 + }, + { + "epoch": 0.2696207925010652, + "grad_norm": 16.75, + "learning_rate": 2.395457896764808e-06, + "loss": 0.9324, + "step": 3164 + }, + { + "epoch": 0.26970600766936514, + "grad_norm": 12.1875, + "learning_rate": 2.395390614925083e-06, + "loss": 0.546, + "step": 3165 + }, + { + "epoch": 0.2697912228376651, + "grad_norm": 12.125, + "learning_rate": 2.3953233123869975e-06, + "loss": 0.648, + "step": 3166 + }, + { + "epoch": 0.26987643800596506, + "grad_norm": 10.6875, + "learning_rate": 2.3952559891517685e-06, + "loss": 0.4571, + "step": 3167 + }, + { + "epoch": 0.269961653174265, + "grad_norm": 11.9375, + "learning_rate": 2.395188645220612e-06, + "loss": 0.3664, + "step": 3168 + }, + { + "epoch": 0.27004686834256497, + "grad_norm": 22.125, + "learning_rate": 2.395121280594745e-06, + "loss": 1.4256, + "step": 3169 + }, + { + "epoch": 0.27013208351086493, + "grad_norm": 24.125, + "learning_rate": 2.3950538952753847e-06, + "loss": 1.0846, + "step": 3170 + }, + { + "epoch": 0.2702172986791649, + "grad_norm": 17.125, + "learning_rate": 2.394986489263749e-06, + "loss": 1.0623, + "step": 3171 + }, + { + "epoch": 0.27030251384746484, + "grad_norm": 13.9375, + "learning_rate": 2.394919062561057e-06, + "loss": 0.7516, + "step": 3172 + }, + { + "epoch": 0.2703877290157648, + "grad_norm": 14.4375, + "learning_rate": 2.3948516151685253e-06, + "loss": 0.8134, + "step": 3173 + }, + { + "epoch": 0.27047294418406476, + "grad_norm": 16.0, + "learning_rate": 2.394784147087374e-06, + "loss": 0.8237, + "step": 3174 + }, + { + "epoch": 0.2705581593523647, + "grad_norm": 13.375, + "learning_rate": 2.394716658318822e-06, + "loss": 0.8622, + "step": 3175 + }, + { + "epoch": 0.27064337452066467, + "grad_norm": 12.0625, + "learning_rate": 2.3946491488640887e-06, + "loss": 0.5916, + "step": 3176 + }, + { + "epoch": 0.2707285896889646, + "grad_norm": 14.625, + "learning_rate": 2.394581618724395e-06, + "loss": 0.5921, + "step": 3177 + }, + { + "epoch": 0.2708138048572646, + "grad_norm": 17.625, + "learning_rate": 2.3945140679009593e-06, + "loss": 0.7363, + "step": 3178 + }, + { + "epoch": 0.27089902002556454, + "grad_norm": 19.375, + "learning_rate": 2.394446496395004e-06, + "loss": 0.9161, + "step": 3179 + }, + { + "epoch": 0.2709842351938645, + "grad_norm": 32.0, + "learning_rate": 2.39437890420775e-06, + "loss": 1.0106, + "step": 3180 + }, + { + "epoch": 0.27106945036216445, + "grad_norm": 21.25, + "learning_rate": 2.394311291340418e-06, + "loss": 0.6123, + "step": 3181 + }, + { + "epoch": 0.2711546655304644, + "grad_norm": 14.9375, + "learning_rate": 2.3942436577942308e-06, + "loss": 0.6568, + "step": 3182 + }, + { + "epoch": 0.27123988069876437, + "grad_norm": 20.375, + "learning_rate": 2.39417600357041e-06, + "loss": 0.804, + "step": 3183 + }, + { + "epoch": 0.2713250958670643, + "grad_norm": 12.6875, + "learning_rate": 2.3941083286701778e-06, + "loss": 0.4972, + "step": 3184 + }, + { + "epoch": 0.2714103110353643, + "grad_norm": 20.625, + "learning_rate": 2.3940406330947578e-06, + "loss": 0.8979, + "step": 3185 + }, + { + "epoch": 0.27149552620366424, + "grad_norm": 8.8125, + "learning_rate": 2.393972916845373e-06, + "loss": 0.5903, + "step": 3186 + }, + { + "epoch": 0.2715807413719642, + "grad_norm": 11.0625, + "learning_rate": 2.3939051799232476e-06, + "loss": 0.4858, + "step": 3187 + }, + { + "epoch": 0.27166595654026415, + "grad_norm": 22.625, + "learning_rate": 2.393837422329605e-06, + "loss": 0.8938, + "step": 3188 + }, + { + "epoch": 0.2717511717085641, + "grad_norm": 31.0, + "learning_rate": 2.3937696440656703e-06, + "loss": 1.5027, + "step": 3189 + }, + { + "epoch": 0.27183638687686407, + "grad_norm": 11.0625, + "learning_rate": 2.3937018451326677e-06, + "loss": 0.4057, + "step": 3190 + }, + { + "epoch": 0.271921602045164, + "grad_norm": 15.6875, + "learning_rate": 2.3936340255318223e-06, + "loss": 0.8049, + "step": 3191 + }, + { + "epoch": 0.272006817213464, + "grad_norm": 10.75, + "learning_rate": 2.3935661852643606e-06, + "loss": 0.4454, + "step": 3192 + }, + { + "epoch": 0.27209203238176394, + "grad_norm": 15.3125, + "learning_rate": 2.393498324331508e-06, + "loss": 0.6401, + "step": 3193 + }, + { + "epoch": 0.2721772475500639, + "grad_norm": 23.0, + "learning_rate": 2.3934304427344906e-06, + "loss": 1.0184, + "step": 3194 + }, + { + "epoch": 0.27226246271836385, + "grad_norm": 10.875, + "learning_rate": 2.3933625404745357e-06, + "loss": 0.4099, + "step": 3195 + }, + { + "epoch": 0.2723476778866638, + "grad_norm": 20.875, + "learning_rate": 2.3932946175528693e-06, + "loss": 1.134, + "step": 3196 + }, + { + "epoch": 0.27243289305496377, + "grad_norm": 29.125, + "learning_rate": 2.39322667397072e-06, + "loss": 0.8978, + "step": 3197 + }, + { + "epoch": 0.2725181082232637, + "grad_norm": 15.4375, + "learning_rate": 2.393158709729315e-06, + "loss": 1.0133, + "step": 3198 + }, + { + "epoch": 0.2726033233915637, + "grad_norm": 18.75, + "learning_rate": 2.3930907248298825e-06, + "loss": 1.0383, + "step": 3199 + }, + { + "epoch": 0.27268853855986364, + "grad_norm": 14.625, + "learning_rate": 2.393022719273651e-06, + "loss": 0.4778, + "step": 3200 + }, + { + "epoch": 0.2727737537281636, + "grad_norm": 16.25, + "learning_rate": 2.3929546930618496e-06, + "loss": 0.9558, + "step": 3201 + }, + { + "epoch": 0.27285896889646355, + "grad_norm": 17.75, + "learning_rate": 2.392886646195708e-06, + "loss": 0.7627, + "step": 3202 + }, + { + "epoch": 0.2729441840647635, + "grad_norm": 19.375, + "learning_rate": 2.392818578676455e-06, + "loss": 1.3218, + "step": 3203 + }, + { + "epoch": 0.27302939923306346, + "grad_norm": 13.5, + "learning_rate": 2.392750490505321e-06, + "loss": 0.7897, + "step": 3204 + }, + { + "epoch": 0.2731146144013634, + "grad_norm": 13.625, + "learning_rate": 2.392682381683537e-06, + "loss": 0.7711, + "step": 3205 + }, + { + "epoch": 0.2731998295696634, + "grad_norm": 13.125, + "learning_rate": 2.3926142522123327e-06, + "loss": 0.6208, + "step": 3206 + }, + { + "epoch": 0.27328504473796333, + "grad_norm": 19.375, + "learning_rate": 2.3925461020929404e-06, + "loss": 0.5715, + "step": 3207 + }, + { + "epoch": 0.2733702599062633, + "grad_norm": 13.5625, + "learning_rate": 2.3924779313265906e-06, + "loss": 0.5413, + "step": 3208 + }, + { + "epoch": 0.27345547507456325, + "grad_norm": 15.1875, + "learning_rate": 2.3924097399145165e-06, + "loss": 0.5909, + "step": 3209 + }, + { + "epoch": 0.2735406902428632, + "grad_norm": 14.5, + "learning_rate": 2.3923415278579486e-06, + "loss": 0.4819, + "step": 3210 + }, + { + "epoch": 0.27362590541116316, + "grad_norm": 18.875, + "learning_rate": 2.3922732951581213e-06, + "loss": 1.0415, + "step": 3211 + }, + { + "epoch": 0.2737111205794631, + "grad_norm": 13.75, + "learning_rate": 2.3922050418162667e-06, + "loss": 0.741, + "step": 3212 + }, + { + "epoch": 0.2737963357477631, + "grad_norm": 10.875, + "learning_rate": 2.392136767833618e-06, + "loss": 0.381, + "step": 3213 + }, + { + "epoch": 0.27388155091606303, + "grad_norm": 14.6875, + "learning_rate": 2.3920684732114103e-06, + "loss": 0.9179, + "step": 3214 + }, + { + "epoch": 0.273966766084363, + "grad_norm": 13.125, + "learning_rate": 2.3920001579508764e-06, + "loss": 0.5845, + "step": 3215 + }, + { + "epoch": 0.274051981252663, + "grad_norm": 14.0625, + "learning_rate": 2.391931822053251e-06, + "loss": 0.6268, + "step": 3216 + }, + { + "epoch": 0.27413719642096296, + "grad_norm": 21.375, + "learning_rate": 2.3918634655197698e-06, + "loss": 1.2118, + "step": 3217 + }, + { + "epoch": 0.2742224115892629, + "grad_norm": 14.5625, + "learning_rate": 2.3917950883516674e-06, + "loss": 0.8733, + "step": 3218 + }, + { + "epoch": 0.2743076267575629, + "grad_norm": 27.5, + "learning_rate": 2.3917266905501796e-06, + "loss": 1.2663, + "step": 3219 + }, + { + "epoch": 0.27439284192586283, + "grad_norm": 13.75, + "learning_rate": 2.391658272116542e-06, + "loss": 0.7653, + "step": 3220 + }, + { + "epoch": 0.2744780570941628, + "grad_norm": 15.0625, + "learning_rate": 2.3915898330519917e-06, + "loss": 0.7122, + "step": 3221 + }, + { + "epoch": 0.27456327226246274, + "grad_norm": 15.6875, + "learning_rate": 2.3915213733577654e-06, + "loss": 0.9794, + "step": 3222 + }, + { + "epoch": 0.2746484874307627, + "grad_norm": 13.6875, + "learning_rate": 2.3914528930350998e-06, + "loss": 0.7808, + "step": 3223 + }, + { + "epoch": 0.27473370259906266, + "grad_norm": 19.5, + "learning_rate": 2.3913843920852325e-06, + "loss": 0.9757, + "step": 3224 + }, + { + "epoch": 0.2748189177673626, + "grad_norm": 18.125, + "learning_rate": 2.3913158705094014e-06, + "loss": 0.7532, + "step": 3225 + }, + { + "epoch": 0.27490413293566257, + "grad_norm": 9.9375, + "learning_rate": 2.391247328308845e-06, + "loss": 0.4005, + "step": 3226 + }, + { + "epoch": 0.2749893481039625, + "grad_norm": 22.375, + "learning_rate": 2.391178765484802e-06, + "loss": 1.2088, + "step": 3227 + }, + { + "epoch": 0.2750745632722625, + "grad_norm": 21.125, + "learning_rate": 2.391110182038511e-06, + "loss": 0.7223, + "step": 3228 + }, + { + "epoch": 0.27515977844056244, + "grad_norm": 16.375, + "learning_rate": 2.391041577971211e-06, + "loss": 1.0136, + "step": 3229 + }, + { + "epoch": 0.2752449936088624, + "grad_norm": 15.75, + "learning_rate": 2.390972953284143e-06, + "loss": 0.9312, + "step": 3230 + }, + { + "epoch": 0.27533020877716236, + "grad_norm": 15.5, + "learning_rate": 2.390904307978546e-06, + "loss": 1.1171, + "step": 3231 + }, + { + "epoch": 0.2754154239454623, + "grad_norm": 15.9375, + "learning_rate": 2.390835642055661e-06, + "loss": 0.7454, + "step": 3232 + }, + { + "epoch": 0.27550063911376227, + "grad_norm": 16.875, + "learning_rate": 2.3907669555167288e-06, + "loss": 0.8645, + "step": 3233 + }, + { + "epoch": 0.2755858542820622, + "grad_norm": 13.8125, + "learning_rate": 2.3906982483629903e-06, + "loss": 0.5586, + "step": 3234 + }, + { + "epoch": 0.2756710694503622, + "grad_norm": 13.875, + "learning_rate": 2.3906295205956877e-06, + "loss": 1.0113, + "step": 3235 + }, + { + "epoch": 0.27575628461866214, + "grad_norm": 12.25, + "learning_rate": 2.3905607722160625e-06, + "loss": 0.6721, + "step": 3236 + }, + { + "epoch": 0.2758414997869621, + "grad_norm": 17.125, + "learning_rate": 2.3904920032253575e-06, + "loss": 1.0627, + "step": 3237 + }, + { + "epoch": 0.27592671495526205, + "grad_norm": 13.5625, + "learning_rate": 2.390423213624815e-06, + "loss": 0.7827, + "step": 3238 + }, + { + "epoch": 0.276011930123562, + "grad_norm": 10.0, + "learning_rate": 2.390354403415678e-06, + "loss": 0.3368, + "step": 3239 + }, + { + "epoch": 0.27609714529186197, + "grad_norm": 15.125, + "learning_rate": 2.3902855725991907e-06, + "loss": 0.5098, + "step": 3240 + }, + { + "epoch": 0.2761823604601619, + "grad_norm": 14.375, + "learning_rate": 2.3902167211765958e-06, + "loss": 0.649, + "step": 3241 + }, + { + "epoch": 0.2762675756284619, + "grad_norm": 15.625, + "learning_rate": 2.390147849149139e-06, + "loss": 0.6001, + "step": 3242 + }, + { + "epoch": 0.27635279079676184, + "grad_norm": 14.375, + "learning_rate": 2.3900789565180632e-06, + "loss": 0.8388, + "step": 3243 + }, + { + "epoch": 0.2764380059650618, + "grad_norm": 13.1875, + "learning_rate": 2.390010043284615e-06, + "loss": 0.587, + "step": 3244 + }, + { + "epoch": 0.27652322113336175, + "grad_norm": 15.3125, + "learning_rate": 2.3899411094500385e-06, + "loss": 0.5722, + "step": 3245 + }, + { + "epoch": 0.2766084363016617, + "grad_norm": 15.6875, + "learning_rate": 2.38987215501558e-06, + "loss": 0.9183, + "step": 3246 + }, + { + "epoch": 0.27669365146996167, + "grad_norm": 13.0, + "learning_rate": 2.389803179982485e-06, + "loss": 0.6159, + "step": 3247 + }, + { + "epoch": 0.2767788666382616, + "grad_norm": 16.5, + "learning_rate": 2.389734184352001e-06, + "loss": 0.7564, + "step": 3248 + }, + { + "epoch": 0.2768640818065616, + "grad_norm": 12.9375, + "learning_rate": 2.389665168125374e-06, + "loss": 0.7234, + "step": 3249 + }, + { + "epoch": 0.27694929697486154, + "grad_norm": 11.4375, + "learning_rate": 2.3895961313038515e-06, + "loss": 0.4677, + "step": 3250 + }, + { + "epoch": 0.2770345121431615, + "grad_norm": 16.625, + "learning_rate": 2.3895270738886807e-06, + "loss": 1.1617, + "step": 3251 + }, + { + "epoch": 0.27711972731146145, + "grad_norm": 17.625, + "learning_rate": 2.38945799588111e-06, + "loss": 0.8164, + "step": 3252 + }, + { + "epoch": 0.2772049424797614, + "grad_norm": 13.4375, + "learning_rate": 2.389388897282387e-06, + "loss": 0.5967, + "step": 3253 + }, + { + "epoch": 0.27729015764806136, + "grad_norm": 14.5625, + "learning_rate": 2.389319778093761e-06, + "loss": 0.8305, + "step": 3254 + }, + { + "epoch": 0.2773753728163613, + "grad_norm": 19.375, + "learning_rate": 2.3892506383164816e-06, + "loss": 0.7634, + "step": 3255 + }, + { + "epoch": 0.2774605879846613, + "grad_norm": 23.25, + "learning_rate": 2.389181477951797e-06, + "loss": 1.2268, + "step": 3256 + }, + { + "epoch": 0.27754580315296123, + "grad_norm": 23.75, + "learning_rate": 2.3891122970009577e-06, + "loss": 1.534, + "step": 3257 + }, + { + "epoch": 0.2776310183212612, + "grad_norm": 15.75, + "learning_rate": 2.389043095465214e-06, + "loss": 0.8984, + "step": 3258 + }, + { + "epoch": 0.27771623348956115, + "grad_norm": 27.125, + "learning_rate": 2.3889738733458155e-06, + "loss": 0.8182, + "step": 3259 + }, + { + "epoch": 0.2778014486578611, + "grad_norm": 14.0, + "learning_rate": 2.3889046306440146e-06, + "loss": 0.8357, + "step": 3260 + }, + { + "epoch": 0.27788666382616106, + "grad_norm": 14.1875, + "learning_rate": 2.3888353673610613e-06, + "loss": 0.8822, + "step": 3261 + }, + { + "epoch": 0.277971878994461, + "grad_norm": 14.0, + "learning_rate": 2.3887660834982076e-06, + "loss": 0.6939, + "step": 3262 + }, + { + "epoch": 0.278057094162761, + "grad_norm": 11.375, + "learning_rate": 2.388696779056706e-06, + "loss": 0.6559, + "step": 3263 + }, + { + "epoch": 0.27814230933106093, + "grad_norm": 18.375, + "learning_rate": 2.3886274540378083e-06, + "loss": 0.5942, + "step": 3264 + }, + { + "epoch": 0.2782275244993609, + "grad_norm": 13.4375, + "learning_rate": 2.388558108442768e-06, + "loss": 0.6218, + "step": 3265 + }, + { + "epoch": 0.27831273966766085, + "grad_norm": 13.125, + "learning_rate": 2.388488742272837e-06, + "loss": 0.5284, + "step": 3266 + }, + { + "epoch": 0.2783979548359608, + "grad_norm": 11.25, + "learning_rate": 2.38841935552927e-06, + "loss": 0.6733, + "step": 3267 + }, + { + "epoch": 0.27848317000426076, + "grad_norm": 15.5, + "learning_rate": 2.3883499482133206e-06, + "loss": 0.9241, + "step": 3268 + }, + { + "epoch": 0.2785683851725607, + "grad_norm": 14.375, + "learning_rate": 2.3882805203262425e-06, + "loss": 0.7892, + "step": 3269 + }, + { + "epoch": 0.2786536003408607, + "grad_norm": 13.125, + "learning_rate": 2.388211071869291e-06, + "loss": 0.7357, + "step": 3270 + }, + { + "epoch": 0.27873881550916063, + "grad_norm": 15.5, + "learning_rate": 2.388141602843721e-06, + "loss": 0.8892, + "step": 3271 + }, + { + "epoch": 0.2788240306774606, + "grad_norm": 12.5, + "learning_rate": 2.388072113250788e-06, + "loss": 0.6221, + "step": 3272 + }, + { + "epoch": 0.27890924584576055, + "grad_norm": 12.6875, + "learning_rate": 2.388002603091747e-06, + "loss": 0.6881, + "step": 3273 + }, + { + "epoch": 0.2789944610140605, + "grad_norm": 15.4375, + "learning_rate": 2.3879330723678547e-06, + "loss": 0.9123, + "step": 3274 + }, + { + "epoch": 0.27907967618236046, + "grad_norm": 13.5625, + "learning_rate": 2.3878635210803675e-06, + "loss": 0.5552, + "step": 3275 + }, + { + "epoch": 0.2791648913506604, + "grad_norm": 13.8125, + "learning_rate": 2.3877939492305423e-06, + "loss": 0.7999, + "step": 3276 + }, + { + "epoch": 0.2792501065189604, + "grad_norm": 14.25, + "learning_rate": 2.3877243568196364e-06, + "loss": 0.5214, + "step": 3277 + }, + { + "epoch": 0.27933532168726033, + "grad_norm": 17.25, + "learning_rate": 2.3876547438489073e-06, + "loss": 0.9477, + "step": 3278 + }, + { + "epoch": 0.2794205368555603, + "grad_norm": 11.6875, + "learning_rate": 2.387585110319613e-06, + "loss": 0.5248, + "step": 3279 + }, + { + "epoch": 0.27950575202386024, + "grad_norm": 14.75, + "learning_rate": 2.3875154562330117e-06, + "loss": 0.5613, + "step": 3280 + }, + { + "epoch": 0.2795909671921602, + "grad_norm": 15.625, + "learning_rate": 2.3874457815903622e-06, + "loss": 0.664, + "step": 3281 + }, + { + "epoch": 0.27967618236046016, + "grad_norm": 10.875, + "learning_rate": 2.387376086392924e-06, + "loss": 0.3871, + "step": 3282 + }, + { + "epoch": 0.2797613975287601, + "grad_norm": 24.375, + "learning_rate": 2.3873063706419557e-06, + "loss": 1.2625, + "step": 3283 + }, + { + "epoch": 0.27984661269706007, + "grad_norm": 16.25, + "learning_rate": 2.3872366343387178e-06, + "loss": 1.1833, + "step": 3284 + }, + { + "epoch": 0.27993182786536003, + "grad_norm": 12.3125, + "learning_rate": 2.3871668774844705e-06, + "loss": 0.769, + "step": 3285 + }, + { + "epoch": 0.28001704303366, + "grad_norm": 12.0, + "learning_rate": 2.3870971000804745e-06, + "loss": 0.5318, + "step": 3286 + }, + { + "epoch": 0.28010225820195994, + "grad_norm": 14.5, + "learning_rate": 2.3870273021279896e-06, + "loss": 0.8197, + "step": 3287 + }, + { + "epoch": 0.2801874733702599, + "grad_norm": 15.4375, + "learning_rate": 2.3869574836282784e-06, + "loss": 0.8311, + "step": 3288 + }, + { + "epoch": 0.28027268853855986, + "grad_norm": 20.25, + "learning_rate": 2.3868876445826024e-06, + "loss": 1.1167, + "step": 3289 + }, + { + "epoch": 0.2803579037068598, + "grad_norm": 15.3125, + "learning_rate": 2.386817784992223e-06, + "loss": 0.6433, + "step": 3290 + }, + { + "epoch": 0.28044311887515977, + "grad_norm": 23.25, + "learning_rate": 2.3867479048584036e-06, + "loss": 1.0603, + "step": 3291 + }, + { + "epoch": 0.2805283340434597, + "grad_norm": 13.8125, + "learning_rate": 2.386678004182406e-06, + "loss": 0.3652, + "step": 3292 + }, + { + "epoch": 0.2806135492117597, + "grad_norm": 13.1875, + "learning_rate": 2.3866080829654944e-06, + "loss": 0.5017, + "step": 3293 + }, + { + "epoch": 0.28069876438005964, + "grad_norm": 19.125, + "learning_rate": 2.386538141208931e-06, + "loss": 0.6801, + "step": 3294 + }, + { + "epoch": 0.2807839795483596, + "grad_norm": 13.0, + "learning_rate": 2.386468178913981e-06, + "loss": 0.7981, + "step": 3295 + }, + { + "epoch": 0.28086919471665955, + "grad_norm": 14.375, + "learning_rate": 2.3863981960819085e-06, + "loss": 0.7691, + "step": 3296 + }, + { + "epoch": 0.2809544098849595, + "grad_norm": 13.625, + "learning_rate": 2.3863281927139773e-06, + "loss": 0.8362, + "step": 3297 + }, + { + "epoch": 0.28103962505325947, + "grad_norm": 18.5, + "learning_rate": 2.3862581688114535e-06, + "loss": 0.592, + "step": 3298 + }, + { + "epoch": 0.2811248402215594, + "grad_norm": 18.875, + "learning_rate": 2.3861881243756018e-06, + "loss": 0.9038, + "step": 3299 + }, + { + "epoch": 0.2812100553898594, + "grad_norm": 11.9375, + "learning_rate": 2.386118059407688e-06, + "loss": 0.5841, + "step": 3300 + }, + { + "epoch": 0.28129527055815934, + "grad_norm": 15.9375, + "learning_rate": 2.386047973908978e-06, + "loss": 0.7435, + "step": 3301 + }, + { + "epoch": 0.2813804857264593, + "grad_norm": 13.0625, + "learning_rate": 2.3859778678807393e-06, + "loss": 0.7852, + "step": 3302 + }, + { + "epoch": 0.28146570089475925, + "grad_norm": 16.5, + "learning_rate": 2.385907741324238e-06, + "loss": 0.7224, + "step": 3303 + }, + { + "epoch": 0.2815509160630592, + "grad_norm": 16.25, + "learning_rate": 2.385837594240742e-06, + "loss": 0.8005, + "step": 3304 + }, + { + "epoch": 0.28163613123135917, + "grad_norm": 13.75, + "learning_rate": 2.385767426631518e-06, + "loss": 0.884, + "step": 3305 + }, + { + "epoch": 0.2817213463996591, + "grad_norm": 12.6875, + "learning_rate": 2.385697238497835e-06, + "loss": 0.5361, + "step": 3306 + }, + { + "epoch": 0.2818065615679591, + "grad_norm": 13.8125, + "learning_rate": 2.3856270298409603e-06, + "loss": 0.6809, + "step": 3307 + }, + { + "epoch": 0.28189177673625904, + "grad_norm": 12.6875, + "learning_rate": 2.3855568006621634e-06, + "loss": 0.4543, + "step": 3308 + }, + { + "epoch": 0.281976991904559, + "grad_norm": 20.875, + "learning_rate": 2.3854865509627135e-06, + "loss": 1.0276, + "step": 3309 + }, + { + "epoch": 0.28206220707285895, + "grad_norm": 12.4375, + "learning_rate": 2.38541628074388e-06, + "loss": 0.5687, + "step": 3310 + }, + { + "epoch": 0.2821474222411589, + "grad_norm": 11.3125, + "learning_rate": 2.3853459900069317e-06, + "loss": 0.6423, + "step": 3311 + }, + { + "epoch": 0.28223263740945886, + "grad_norm": 14.9375, + "learning_rate": 2.38527567875314e-06, + "loss": 0.7432, + "step": 3312 + }, + { + "epoch": 0.2823178525777588, + "grad_norm": 11.25, + "learning_rate": 2.3852053469837756e-06, + "loss": 0.6378, + "step": 3313 + }, + { + "epoch": 0.2824030677460588, + "grad_norm": 25.625, + "learning_rate": 2.3851349947001087e-06, + "loss": 1.0169, + "step": 3314 + }, + { + "epoch": 0.28248828291435873, + "grad_norm": 11.0625, + "learning_rate": 2.3850646219034107e-06, + "loss": 0.5053, + "step": 3315 + }, + { + "epoch": 0.2825734980826587, + "grad_norm": 21.5, + "learning_rate": 2.384994228594954e-06, + "loss": 0.5648, + "step": 3316 + }, + { + "epoch": 0.28265871325095865, + "grad_norm": 23.5, + "learning_rate": 2.38492381477601e-06, + "loss": 1.022, + "step": 3317 + }, + { + "epoch": 0.2827439284192586, + "grad_norm": 13.875, + "learning_rate": 2.384853380447852e-06, + "loss": 0.677, + "step": 3318 + }, + { + "epoch": 0.28282914358755856, + "grad_norm": 11.8125, + "learning_rate": 2.3847829256117516e-06, + "loss": 0.489, + "step": 3319 + }, + { + "epoch": 0.2829143587558585, + "grad_norm": 13.625, + "learning_rate": 2.384712450268983e-06, + "loss": 0.7879, + "step": 3320 + }, + { + "epoch": 0.2829995739241585, + "grad_norm": 18.25, + "learning_rate": 2.384641954420819e-06, + "loss": 1.1059, + "step": 3321 + }, + { + "epoch": 0.28308478909245843, + "grad_norm": 11.5625, + "learning_rate": 2.3845714380685342e-06, + "loss": 0.2821, + "step": 3322 + }, + { + "epoch": 0.2831700042607584, + "grad_norm": 14.0625, + "learning_rate": 2.3845009012134027e-06, + "loss": 0.4754, + "step": 3323 + }, + { + "epoch": 0.28325521942905835, + "grad_norm": 32.5, + "learning_rate": 2.384430343856699e-06, + "loss": 1.4646, + "step": 3324 + }, + { + "epoch": 0.2833404345973583, + "grad_norm": 13.25, + "learning_rate": 2.384359765999698e-06, + "loss": 0.7753, + "step": 3325 + }, + { + "epoch": 0.28342564976565826, + "grad_norm": 17.5, + "learning_rate": 2.3842891676436756e-06, + "loss": 0.2746, + "step": 3326 + }, + { + "epoch": 0.2835108649339582, + "grad_norm": 11.625, + "learning_rate": 2.3842185487899073e-06, + "loss": 0.4905, + "step": 3327 + }, + { + "epoch": 0.2835960801022582, + "grad_norm": 15.375, + "learning_rate": 2.384147909439669e-06, + "loss": 0.6193, + "step": 3328 + }, + { + "epoch": 0.2836812952705582, + "grad_norm": 12.8125, + "learning_rate": 2.3840772495942378e-06, + "loss": 0.7505, + "step": 3329 + }, + { + "epoch": 0.28376651043885814, + "grad_norm": 20.375, + "learning_rate": 2.3840065692548904e-06, + "loss": 0.8517, + "step": 3330 + }, + { + "epoch": 0.2838517256071581, + "grad_norm": 22.5, + "learning_rate": 2.3839358684229035e-06, + "loss": 0.904, + "step": 3331 + }, + { + "epoch": 0.28393694077545806, + "grad_norm": 14.0, + "learning_rate": 2.3838651470995556e-06, + "loss": 0.5554, + "step": 3332 + }, + { + "epoch": 0.284022155943758, + "grad_norm": 14.625, + "learning_rate": 2.383794405286124e-06, + "loss": 0.6426, + "step": 3333 + }, + { + "epoch": 0.28410737111205797, + "grad_norm": 27.125, + "learning_rate": 2.3837236429838878e-06, + "loss": 0.8508, + "step": 3334 + }, + { + "epoch": 0.28419258628035793, + "grad_norm": 23.5, + "learning_rate": 2.383652860194125e-06, + "loss": 0.3894, + "step": 3335 + }, + { + "epoch": 0.2842778014486579, + "grad_norm": 13.0, + "learning_rate": 2.383582056918115e-06, + "loss": 0.5137, + "step": 3336 + }, + { + "epoch": 0.28436301661695784, + "grad_norm": 16.5, + "learning_rate": 2.3835112331571374e-06, + "loss": 0.7774, + "step": 3337 + }, + { + "epoch": 0.2844482317852578, + "grad_norm": 23.625, + "learning_rate": 2.383440388912472e-06, + "loss": 1.0577, + "step": 3338 + }, + { + "epoch": 0.28453344695355776, + "grad_norm": 22.0, + "learning_rate": 2.3833695241853993e-06, + "loss": 1.1389, + "step": 3339 + }, + { + "epoch": 0.2846186621218577, + "grad_norm": 14.875, + "learning_rate": 2.383298638977199e-06, + "loss": 1.037, + "step": 3340 + }, + { + "epoch": 0.28470387729015767, + "grad_norm": 14.4375, + "learning_rate": 2.3832277332891534e-06, + "loss": 0.8874, + "step": 3341 + }, + { + "epoch": 0.2847890924584576, + "grad_norm": 13.5, + "learning_rate": 2.3831568071225424e-06, + "loss": 0.7628, + "step": 3342 + }, + { + "epoch": 0.2848743076267576, + "grad_norm": 27.25, + "learning_rate": 2.383085860478649e-06, + "loss": 0.9163, + "step": 3343 + }, + { + "epoch": 0.28495952279505754, + "grad_norm": 13.1875, + "learning_rate": 2.3830148933587545e-06, + "loss": 0.469, + "step": 3344 + }, + { + "epoch": 0.2850447379633575, + "grad_norm": 13.875, + "learning_rate": 2.3829439057641415e-06, + "loss": 0.4983, + "step": 3345 + }, + { + "epoch": 0.28512995313165745, + "grad_norm": 16.25, + "learning_rate": 2.382872897696093e-06, + "loss": 0.5694, + "step": 3346 + }, + { + "epoch": 0.2852151682999574, + "grad_norm": 12.875, + "learning_rate": 2.3828018691558915e-06, + "loss": 0.589, + "step": 3347 + }, + { + "epoch": 0.28530038346825737, + "grad_norm": 15.875, + "learning_rate": 2.3827308201448215e-06, + "loss": 1.1654, + "step": 3348 + }, + { + "epoch": 0.2853855986365573, + "grad_norm": 12.75, + "learning_rate": 2.382659750664166e-06, + "loss": 0.5626, + "step": 3349 + }, + { + "epoch": 0.2854708138048573, + "grad_norm": 15.0, + "learning_rate": 2.382588660715211e-06, + "loss": 0.9217, + "step": 3350 + }, + { + "epoch": 0.28555602897315724, + "grad_norm": 20.875, + "learning_rate": 2.3825175502992394e-06, + "loss": 0.9546, + "step": 3351 + }, + { + "epoch": 0.2856412441414572, + "grad_norm": 11.5625, + "learning_rate": 2.3824464194175366e-06, + "loss": 0.3705, + "step": 3352 + }, + { + "epoch": 0.28572645930975715, + "grad_norm": 13.1875, + "learning_rate": 2.3823752680713887e-06, + "loss": 0.4902, + "step": 3353 + }, + { + "epoch": 0.2858116744780571, + "grad_norm": 16.75, + "learning_rate": 2.3823040962620805e-06, + "loss": 1.0502, + "step": 3354 + }, + { + "epoch": 0.28589688964635707, + "grad_norm": 13.625, + "learning_rate": 2.382232903990899e-06, + "loss": 0.547, + "step": 3355 + }, + { + "epoch": 0.285982104814657, + "grad_norm": 19.0, + "learning_rate": 2.38216169125913e-06, + "loss": 1.1682, + "step": 3356 + }, + { + "epoch": 0.286067319982957, + "grad_norm": 13.9375, + "learning_rate": 2.382090458068061e-06, + "loss": 0.8007, + "step": 3357 + }, + { + "epoch": 0.28615253515125694, + "grad_norm": 16.0, + "learning_rate": 2.3820192044189794e-06, + "loss": 0.952, + "step": 3358 + }, + { + "epoch": 0.2862377503195569, + "grad_norm": 15.1875, + "learning_rate": 2.3819479303131722e-06, + "loss": 0.6862, + "step": 3359 + }, + { + "epoch": 0.28632296548785685, + "grad_norm": 30.875, + "learning_rate": 2.3818766357519273e-06, + "loss": 0.8938, + "step": 3360 + }, + { + "epoch": 0.2864081806561568, + "grad_norm": 17.5, + "learning_rate": 2.3818053207365337e-06, + "loss": 0.5579, + "step": 3361 + }, + { + "epoch": 0.28649339582445676, + "grad_norm": 11.6875, + "learning_rate": 2.38173398526828e-06, + "loss": 0.4603, + "step": 3362 + }, + { + "epoch": 0.2865786109927567, + "grad_norm": 15.4375, + "learning_rate": 2.381662629348455e-06, + "loss": 1.006, + "step": 3363 + }, + { + "epoch": 0.2866638261610567, + "grad_norm": 14.5625, + "learning_rate": 2.381591252978348e-06, + "loss": 0.8126, + "step": 3364 + }, + { + "epoch": 0.28674904132935664, + "grad_norm": 16.875, + "learning_rate": 2.381519856159249e-06, + "loss": 0.8469, + "step": 3365 + }, + { + "epoch": 0.2868342564976566, + "grad_norm": 12.6875, + "learning_rate": 2.3814484388924487e-06, + "loss": 0.4767, + "step": 3366 + }, + { + "epoch": 0.28691947166595655, + "grad_norm": 18.125, + "learning_rate": 2.3813770011792373e-06, + "loss": 0.7269, + "step": 3367 + }, + { + "epoch": 0.2870046868342565, + "grad_norm": 14.6875, + "learning_rate": 2.3813055430209063e-06, + "loss": 0.5826, + "step": 3368 + }, + { + "epoch": 0.28708990200255646, + "grad_norm": 21.875, + "learning_rate": 2.381234064418746e-06, + "loss": 1.1915, + "step": 3369 + }, + { + "epoch": 0.2871751171708564, + "grad_norm": 11.5625, + "learning_rate": 2.3811625653740487e-06, + "loss": 0.4676, + "step": 3370 + }, + { + "epoch": 0.2872603323391564, + "grad_norm": 14.375, + "learning_rate": 2.3810910458881064e-06, + "loss": 0.6785, + "step": 3371 + }, + { + "epoch": 0.28734554750745633, + "grad_norm": 15.3125, + "learning_rate": 2.3810195059622117e-06, + "loss": 0.6019, + "step": 3372 + }, + { + "epoch": 0.2874307626757563, + "grad_norm": 13.9375, + "learning_rate": 2.380947945597657e-06, + "loss": 0.7969, + "step": 3373 + }, + { + "epoch": 0.28751597784405625, + "grad_norm": 12.3125, + "learning_rate": 2.3808763647957355e-06, + "loss": 0.4861, + "step": 3374 + }, + { + "epoch": 0.2876011930123562, + "grad_norm": 14.125, + "learning_rate": 2.380804763557741e-06, + "loss": 0.87, + "step": 3375 + }, + { + "epoch": 0.28768640818065616, + "grad_norm": 11.5625, + "learning_rate": 2.3807331418849675e-06, + "loss": 0.4479, + "step": 3376 + }, + { + "epoch": 0.2877716233489561, + "grad_norm": 23.625, + "learning_rate": 2.3806614997787084e-06, + "loss": 0.8922, + "step": 3377 + }, + { + "epoch": 0.2878568385172561, + "grad_norm": 20.625, + "learning_rate": 2.3805898372402605e-06, + "loss": 0.7641, + "step": 3378 + }, + { + "epoch": 0.28794205368555603, + "grad_norm": 12.8125, + "learning_rate": 2.380518154270916e-06, + "loss": 0.5588, + "step": 3379 + }, + { + "epoch": 0.288027268853856, + "grad_norm": 13.375, + "learning_rate": 2.380446450871972e-06, + "loss": 0.6506, + "step": 3380 + }, + { + "epoch": 0.28811248402215595, + "grad_norm": 12.1875, + "learning_rate": 2.380374727044724e-06, + "loss": 0.6244, + "step": 3381 + }, + { + "epoch": 0.2881976991904559, + "grad_norm": 14.3125, + "learning_rate": 2.380302982790468e-06, + "loss": 1.1131, + "step": 3382 + }, + { + "epoch": 0.28828291435875586, + "grad_norm": 18.0, + "learning_rate": 2.3802312181105e-06, + "loss": 1.1048, + "step": 3383 + }, + { + "epoch": 0.2883681295270558, + "grad_norm": 19.25, + "learning_rate": 2.3801594330061183e-06, + "loss": 0.6149, + "step": 3384 + }, + { + "epoch": 0.2884533446953558, + "grad_norm": 28.375, + "learning_rate": 2.3800876274786185e-06, + "loss": 0.8047, + "step": 3385 + }, + { + "epoch": 0.28853855986365573, + "grad_norm": 8.6875, + "learning_rate": 2.380015801529299e-06, + "loss": 0.4653, + "step": 3386 + }, + { + "epoch": 0.2886237750319557, + "grad_norm": 15.3125, + "learning_rate": 2.379943955159458e-06, + "loss": 0.7847, + "step": 3387 + }, + { + "epoch": 0.28870899020025564, + "grad_norm": 13.5625, + "learning_rate": 2.379872088370393e-06, + "loss": 0.5854, + "step": 3388 + }, + { + "epoch": 0.2887942053685556, + "grad_norm": 15.875, + "learning_rate": 2.3798002011634035e-06, + "loss": 0.7127, + "step": 3389 + }, + { + "epoch": 0.28887942053685556, + "grad_norm": 24.25, + "learning_rate": 2.379728293539788e-06, + "loss": 1.0345, + "step": 3390 + }, + { + "epoch": 0.2889646357051555, + "grad_norm": 11.875, + "learning_rate": 2.3796563655008466e-06, + "loss": 0.4436, + "step": 3391 + }, + { + "epoch": 0.28904985087345547, + "grad_norm": 23.75, + "learning_rate": 2.3795844170478783e-06, + "loss": 1.2227, + "step": 3392 + }, + { + "epoch": 0.28913506604175543, + "grad_norm": 17.375, + "learning_rate": 2.379512448182184e-06, + "loss": 0.9114, + "step": 3393 + }, + { + "epoch": 0.2892202812100554, + "grad_norm": 14.4375, + "learning_rate": 2.3794404589050636e-06, + "loss": 0.8593, + "step": 3394 + }, + { + "epoch": 0.28930549637835534, + "grad_norm": 14.75, + "learning_rate": 2.3793684492178186e-06, + "loss": 0.6833, + "step": 3395 + }, + { + "epoch": 0.2893907115466553, + "grad_norm": 13.875, + "learning_rate": 2.37929641912175e-06, + "loss": 0.9602, + "step": 3396 + }, + { + "epoch": 0.28947592671495526, + "grad_norm": 16.75, + "learning_rate": 2.37922436861816e-06, + "loss": 0.7795, + "step": 3397 + }, + { + "epoch": 0.2895611418832552, + "grad_norm": 15.4375, + "learning_rate": 2.379152297708349e-06, + "loss": 0.998, + "step": 3398 + }, + { + "epoch": 0.28964635705155517, + "grad_norm": 13.125, + "learning_rate": 2.3790802063936215e-06, + "loss": 0.496, + "step": 3399 + }, + { + "epoch": 0.2897315722198551, + "grad_norm": 10.4375, + "learning_rate": 2.379008094675279e-06, + "loss": 0.448, + "step": 3400 + }, + { + "epoch": 0.2898167873881551, + "grad_norm": 12.3125, + "learning_rate": 2.378935962554625e-06, + "loss": 0.5801, + "step": 3401 + }, + { + "epoch": 0.28990200255645504, + "grad_norm": 15.375, + "learning_rate": 2.378863810032963e-06, + "loss": 0.763, + "step": 3402 + }, + { + "epoch": 0.289987217724755, + "grad_norm": 16.5, + "learning_rate": 2.3787916371115966e-06, + "loss": 1.0671, + "step": 3403 + }, + { + "epoch": 0.29007243289305495, + "grad_norm": 14.1875, + "learning_rate": 2.3787194437918302e-06, + "loss": 0.6688, + "step": 3404 + }, + { + "epoch": 0.2901576480613549, + "grad_norm": 14.1875, + "learning_rate": 2.3786472300749688e-06, + "loss": 0.5871, + "step": 3405 + }, + { + "epoch": 0.29024286322965487, + "grad_norm": 12.3125, + "learning_rate": 2.3785749959623167e-06, + "loss": 0.5202, + "step": 3406 + }, + { + "epoch": 0.2903280783979548, + "grad_norm": 21.625, + "learning_rate": 2.3785027414551793e-06, + "loss": 1.0641, + "step": 3407 + }, + { + "epoch": 0.2904132935662548, + "grad_norm": 11.0625, + "learning_rate": 2.378430466554863e-06, + "loss": 0.3786, + "step": 3408 + }, + { + "epoch": 0.29049850873455474, + "grad_norm": 13.0, + "learning_rate": 2.3783581712626733e-06, + "loss": 0.6681, + "step": 3409 + }, + { + "epoch": 0.2905837239028547, + "grad_norm": 14.8125, + "learning_rate": 2.3782858555799167e-06, + "loss": 0.834, + "step": 3410 + }, + { + "epoch": 0.29066893907115465, + "grad_norm": 14.4375, + "learning_rate": 2.3782135195079004e-06, + "loss": 0.6867, + "step": 3411 + }, + { + "epoch": 0.2907541542394546, + "grad_norm": 14.0, + "learning_rate": 2.3781411630479307e-06, + "loss": 0.8765, + "step": 3412 + }, + { + "epoch": 0.29083936940775457, + "grad_norm": 14.5, + "learning_rate": 2.3780687862013162e-06, + "loss": 0.7197, + "step": 3413 + }, + { + "epoch": 0.2909245845760545, + "grad_norm": 16.0, + "learning_rate": 2.3779963889693644e-06, + "loss": 0.7068, + "step": 3414 + }, + { + "epoch": 0.2910097997443545, + "grad_norm": 13.9375, + "learning_rate": 2.377923971353384e-06, + "loss": 0.619, + "step": 3415 + }, + { + "epoch": 0.29109501491265444, + "grad_norm": 13.375, + "learning_rate": 2.377851533354682e-06, + "loss": 0.4641, + "step": 3416 + }, + { + "epoch": 0.2911802300809544, + "grad_norm": 13.1875, + "learning_rate": 2.3777790749745695e-06, + "loss": 0.5483, + "step": 3417 + }, + { + "epoch": 0.29126544524925435, + "grad_norm": 12.125, + "learning_rate": 2.3777065962143547e-06, + "loss": 0.617, + "step": 3418 + }, + { + "epoch": 0.2913506604175543, + "grad_norm": 15.125, + "learning_rate": 2.377634097075348e-06, + "loss": 0.8582, + "step": 3419 + }, + { + "epoch": 0.29143587558585426, + "grad_norm": 13.125, + "learning_rate": 2.3775615775588594e-06, + "loss": 0.4053, + "step": 3420 + }, + { + "epoch": 0.2915210907541542, + "grad_norm": 20.25, + "learning_rate": 2.377489037666199e-06, + "loss": 1.0821, + "step": 3421 + }, + { + "epoch": 0.2916063059224542, + "grad_norm": 19.0, + "learning_rate": 2.3774164773986775e-06, + "loss": 1.035, + "step": 3422 + }, + { + "epoch": 0.29169152109075414, + "grad_norm": 13.625, + "learning_rate": 2.3773438967576075e-06, + "loss": 0.4144, + "step": 3423 + }, + { + "epoch": 0.2917767362590541, + "grad_norm": 13.375, + "learning_rate": 2.377271295744299e-06, + "loss": 0.6282, + "step": 3424 + }, + { + "epoch": 0.29186195142735405, + "grad_norm": 20.25, + "learning_rate": 2.3771986743600643e-06, + "loss": 1.0134, + "step": 3425 + }, + { + "epoch": 0.291947166595654, + "grad_norm": 14.5, + "learning_rate": 2.377126032606217e-06, + "loss": 0.753, + "step": 3426 + }, + { + "epoch": 0.29203238176395396, + "grad_norm": 21.875, + "learning_rate": 2.377053370484068e-06, + "loss": 1.1791, + "step": 3427 + }, + { + "epoch": 0.2921175969322539, + "grad_norm": 19.75, + "learning_rate": 2.376980687994932e-06, + "loss": 0.9957, + "step": 3428 + }, + { + "epoch": 0.2922028121005539, + "grad_norm": 16.5, + "learning_rate": 2.376907985140121e-06, + "loss": 1.0694, + "step": 3429 + }, + { + "epoch": 0.29228802726885383, + "grad_norm": 16.25, + "learning_rate": 2.37683526192095e-06, + "loss": 1.0766, + "step": 3430 + }, + { + "epoch": 0.2923732424371538, + "grad_norm": 14.0, + "learning_rate": 2.3767625183387323e-06, + "loss": 0.79, + "step": 3431 + }, + { + "epoch": 0.29245845760545375, + "grad_norm": 13.5625, + "learning_rate": 2.376689754394783e-06, + "loss": 0.7916, + "step": 3432 + }, + { + "epoch": 0.2925436727737537, + "grad_norm": 15.875, + "learning_rate": 2.3766169700904176e-06, + "loss": 0.888, + "step": 3433 + }, + { + "epoch": 0.29262888794205366, + "grad_norm": 13.8125, + "learning_rate": 2.3765441654269495e-06, + "loss": 0.8489, + "step": 3434 + }, + { + "epoch": 0.2927141031103536, + "grad_norm": 15.9375, + "learning_rate": 2.376471340405696e-06, + "loss": 0.7272, + "step": 3435 + }, + { + "epoch": 0.2927993182786536, + "grad_norm": 21.0, + "learning_rate": 2.376398495027973e-06, + "loss": 0.9728, + "step": 3436 + }, + { + "epoch": 0.29288453344695353, + "grad_norm": 11.9375, + "learning_rate": 2.376325629295096e-06, + "loss": 0.3709, + "step": 3437 + }, + { + "epoch": 0.2929697486152535, + "grad_norm": 16.625, + "learning_rate": 2.376252743208382e-06, + "loss": 0.8203, + "step": 3438 + }, + { + "epoch": 0.29305496378355345, + "grad_norm": 14.0, + "learning_rate": 2.376179836769149e-06, + "loss": 0.8175, + "step": 3439 + }, + { + "epoch": 0.2931401789518534, + "grad_norm": 13.8125, + "learning_rate": 2.376106909978714e-06, + "loss": 0.7277, + "step": 3440 + }, + { + "epoch": 0.29322539412015336, + "grad_norm": 16.125, + "learning_rate": 2.3760339628383944e-06, + "loss": 0.7932, + "step": 3441 + }, + { + "epoch": 0.29331060928845337, + "grad_norm": 19.375, + "learning_rate": 2.375960995349509e-06, + "loss": 0.3535, + "step": 3442 + }, + { + "epoch": 0.29339582445675333, + "grad_norm": 14.4375, + "learning_rate": 2.3758880075133754e-06, + "loss": 0.7799, + "step": 3443 + }, + { + "epoch": 0.2934810396250533, + "grad_norm": 13.375, + "learning_rate": 2.3758149993313145e-06, + "loss": 0.6711, + "step": 3444 + }, + { + "epoch": 0.29356625479335324, + "grad_norm": 12.9375, + "learning_rate": 2.3757419708046436e-06, + "loss": 0.5204, + "step": 3445 + }, + { + "epoch": 0.2936514699616532, + "grad_norm": 18.375, + "learning_rate": 2.3756689219346833e-06, + "loss": 0.8879, + "step": 3446 + }, + { + "epoch": 0.29373668512995316, + "grad_norm": 13.5, + "learning_rate": 2.375595852722754e-06, + "loss": 0.7165, + "step": 3447 + }, + { + "epoch": 0.2938219002982531, + "grad_norm": 11.125, + "learning_rate": 2.375522763170176e-06, + "loss": 0.4395, + "step": 3448 + }, + { + "epoch": 0.29390711546655307, + "grad_norm": 15.3125, + "learning_rate": 2.3754496532782696e-06, + "loss": 0.9505, + "step": 3449 + }, + { + "epoch": 0.293992330634853, + "grad_norm": 12.6875, + "learning_rate": 2.375376523048356e-06, + "loss": 0.635, + "step": 3450 + }, + { + "epoch": 0.294077545803153, + "grad_norm": 13.5, + "learning_rate": 2.375303372481757e-06, + "loss": 0.6056, + "step": 3451 + }, + { + "epoch": 0.29416276097145294, + "grad_norm": 15.75, + "learning_rate": 2.3752302015797944e-06, + "loss": 0.7424, + "step": 3452 + }, + { + "epoch": 0.2942479761397529, + "grad_norm": 16.375, + "learning_rate": 2.375157010343791e-06, + "loss": 1.001, + "step": 3453 + }, + { + "epoch": 0.29433319130805286, + "grad_norm": 13.375, + "learning_rate": 2.375083798775069e-06, + "loss": 0.3734, + "step": 3454 + }, + { + "epoch": 0.2944184064763528, + "grad_norm": 17.125, + "learning_rate": 2.3750105668749513e-06, + "loss": 0.956, + "step": 3455 + }, + { + "epoch": 0.29450362164465277, + "grad_norm": 12.5625, + "learning_rate": 2.374937314644761e-06, + "loss": 0.4849, + "step": 3456 + }, + { + "epoch": 0.2945888368129527, + "grad_norm": 13.125, + "learning_rate": 2.3748640420858228e-06, + "loss": 0.5476, + "step": 3457 + }, + { + "epoch": 0.2946740519812527, + "grad_norm": 18.5, + "learning_rate": 2.3747907491994597e-06, + "loss": 0.8018, + "step": 3458 + }, + { + "epoch": 0.29475926714955264, + "grad_norm": 13.5625, + "learning_rate": 2.374717435986997e-06, + "loss": 0.6866, + "step": 3459 + }, + { + "epoch": 0.2948444823178526, + "grad_norm": 11.5, + "learning_rate": 2.3746441024497586e-06, + "loss": 0.5413, + "step": 3460 + }, + { + "epoch": 0.29492969748615255, + "grad_norm": 14.4375, + "learning_rate": 2.374570748589071e-06, + "loss": 0.6204, + "step": 3461 + }, + { + "epoch": 0.2950149126544525, + "grad_norm": 12.6875, + "learning_rate": 2.374497374406259e-06, + "loss": 0.2954, + "step": 3462 + }, + { + "epoch": 0.29510012782275247, + "grad_norm": 13.4375, + "learning_rate": 2.3744239799026486e-06, + "loss": 0.8429, + "step": 3463 + }, + { + "epoch": 0.2951853429910524, + "grad_norm": 15.125, + "learning_rate": 2.3743505650795663e-06, + "loss": 0.8303, + "step": 3464 + }, + { + "epoch": 0.2952705581593524, + "grad_norm": 17.75, + "learning_rate": 2.3742771299383387e-06, + "loss": 1.0318, + "step": 3465 + }, + { + "epoch": 0.29535577332765234, + "grad_norm": 13.6875, + "learning_rate": 2.3742036744802927e-06, + "loss": 0.5929, + "step": 3466 + }, + { + "epoch": 0.2954409884959523, + "grad_norm": 17.25, + "learning_rate": 2.3741301987067557e-06, + "loss": 0.8577, + "step": 3467 + }, + { + "epoch": 0.29552620366425225, + "grad_norm": 17.625, + "learning_rate": 2.374056702619056e-06, + "loss": 1.2268, + "step": 3468 + }, + { + "epoch": 0.2956114188325522, + "grad_norm": 14.3125, + "learning_rate": 2.373983186218521e-06, + "loss": 1.0293, + "step": 3469 + }, + { + "epoch": 0.29569663400085217, + "grad_norm": 13.6875, + "learning_rate": 2.37390964950648e-06, + "loss": 0.6229, + "step": 3470 + }, + { + "epoch": 0.2957818491691521, + "grad_norm": 12.5625, + "learning_rate": 2.373836092484261e-06, + "loss": 0.5964, + "step": 3471 + }, + { + "epoch": 0.2958670643374521, + "grad_norm": 11.375, + "learning_rate": 2.3737625151531937e-06, + "loss": 0.4489, + "step": 3472 + }, + { + "epoch": 0.29595227950575204, + "grad_norm": 14.1875, + "learning_rate": 2.373688917514608e-06, + "loss": 0.6087, + "step": 3473 + }, + { + "epoch": 0.296037494674052, + "grad_norm": 14.6875, + "learning_rate": 2.3736152995698335e-06, + "loss": 0.8158, + "step": 3474 + }, + { + "epoch": 0.29612270984235195, + "grad_norm": 12.125, + "learning_rate": 2.3735416613202005e-06, + "loss": 0.8275, + "step": 3475 + }, + { + "epoch": 0.2962079250106519, + "grad_norm": 13.3125, + "learning_rate": 2.37346800276704e-06, + "loss": 0.6294, + "step": 3476 + }, + { + "epoch": 0.29629314017895186, + "grad_norm": 15.0625, + "learning_rate": 2.373394323911683e-06, + "loss": 0.8348, + "step": 3477 + }, + { + "epoch": 0.2963783553472518, + "grad_norm": 14.375, + "learning_rate": 2.3733206247554607e-06, + "loss": 0.8657, + "step": 3478 + }, + { + "epoch": 0.2964635705155518, + "grad_norm": 13.8125, + "learning_rate": 2.3732469052997055e-06, + "loss": 1.0255, + "step": 3479 + }, + { + "epoch": 0.29654878568385173, + "grad_norm": 17.125, + "learning_rate": 2.3731731655457492e-06, + "loss": 0.646, + "step": 3480 + }, + { + "epoch": 0.2966340008521517, + "grad_norm": 20.875, + "learning_rate": 2.373099405494924e-06, + "loss": 1.4531, + "step": 3481 + }, + { + "epoch": 0.29671921602045165, + "grad_norm": 23.25, + "learning_rate": 2.3730256251485633e-06, + "loss": 1.1542, + "step": 3482 + }, + { + "epoch": 0.2968044311887516, + "grad_norm": 19.0, + "learning_rate": 2.372951824508e-06, + "loss": 1.0287, + "step": 3483 + }, + { + "epoch": 0.29688964635705156, + "grad_norm": 15.875, + "learning_rate": 2.3728780035745684e-06, + "loss": 0.9181, + "step": 3484 + }, + { + "epoch": 0.2969748615253515, + "grad_norm": 13.5625, + "learning_rate": 2.3728041623496023e-06, + "loss": 0.7606, + "step": 3485 + }, + { + "epoch": 0.2970600766936515, + "grad_norm": 19.0, + "learning_rate": 2.372730300834436e-06, + "loss": 0.9599, + "step": 3486 + }, + { + "epoch": 0.29714529186195143, + "grad_norm": 13.0625, + "learning_rate": 2.3726564190304036e-06, + "loss": 0.6571, + "step": 3487 + }, + { + "epoch": 0.2972305070302514, + "grad_norm": 17.125, + "learning_rate": 2.372582516938841e-06, + "loss": 0.8738, + "step": 3488 + }, + { + "epoch": 0.29731572219855135, + "grad_norm": 12.25, + "learning_rate": 2.3725085945610833e-06, + "loss": 0.5681, + "step": 3489 + }, + { + "epoch": 0.2974009373668513, + "grad_norm": 16.75, + "learning_rate": 2.372434651898467e-06, + "loss": 0.9971, + "step": 3490 + }, + { + "epoch": 0.29748615253515126, + "grad_norm": 11.0, + "learning_rate": 2.3723606889523277e-06, + "loss": 0.5215, + "step": 3491 + }, + { + "epoch": 0.2975713677034512, + "grad_norm": 11.75, + "learning_rate": 2.372286705724002e-06, + "loss": 0.428, + "step": 3492 + }, + { + "epoch": 0.2976565828717512, + "grad_norm": 16.625, + "learning_rate": 2.3722127022148267e-06, + "loss": 0.9708, + "step": 3493 + }, + { + "epoch": 0.29774179804005113, + "grad_norm": 21.0, + "learning_rate": 2.3721386784261398e-06, + "loss": 0.6979, + "step": 3494 + }, + { + "epoch": 0.2978270132083511, + "grad_norm": 11.75, + "learning_rate": 2.3720646343592783e-06, + "loss": 0.5514, + "step": 3495 + }, + { + "epoch": 0.29791222837665104, + "grad_norm": 15.9375, + "learning_rate": 2.371990570015581e-06, + "loss": 1.0081, + "step": 3496 + }, + { + "epoch": 0.297997443544951, + "grad_norm": 14.6875, + "learning_rate": 2.3719164853963857e-06, + "loss": 0.5804, + "step": 3497 + }, + { + "epoch": 0.29808265871325096, + "grad_norm": 21.375, + "learning_rate": 2.371842380503031e-06, + "loss": 0.7738, + "step": 3498 + }, + { + "epoch": 0.2981678738815509, + "grad_norm": 15.75, + "learning_rate": 2.3717682553368566e-06, + "loss": 0.4976, + "step": 3499 + }, + { + "epoch": 0.2982530890498509, + "grad_norm": 14.5625, + "learning_rate": 2.371694109899202e-06, + "loss": 0.9709, + "step": 3500 + }, + { + "epoch": 0.29833830421815083, + "grad_norm": 14.375, + "learning_rate": 2.3716199441914064e-06, + "loss": 0.9336, + "step": 3501 + }, + { + "epoch": 0.2984235193864508, + "grad_norm": 15.125, + "learning_rate": 2.371545758214811e-06, + "loss": 0.7563, + "step": 3502 + }, + { + "epoch": 0.29850873455475074, + "grad_norm": 13.75, + "learning_rate": 2.3714715519707556e-06, + "loss": 0.4922, + "step": 3503 + }, + { + "epoch": 0.2985939497230507, + "grad_norm": 11.5625, + "learning_rate": 2.3713973254605816e-06, + "loss": 0.3999, + "step": 3504 + }, + { + "epoch": 0.29867916489135066, + "grad_norm": 15.0, + "learning_rate": 2.3713230786856305e-06, + "loss": 0.845, + "step": 3505 + }, + { + "epoch": 0.2987643800596506, + "grad_norm": 13.75, + "learning_rate": 2.3712488116472437e-06, + "loss": 0.7819, + "step": 3506 + }, + { + "epoch": 0.29884959522795057, + "grad_norm": 17.875, + "learning_rate": 2.371174524346763e-06, + "loss": 0.9408, + "step": 3507 + }, + { + "epoch": 0.2989348103962505, + "grad_norm": 17.875, + "learning_rate": 2.371100216785532e-06, + "loss": 1.0446, + "step": 3508 + }, + { + "epoch": 0.2990200255645505, + "grad_norm": 12.0625, + "learning_rate": 2.371025888964892e-06, + "loss": 0.6994, + "step": 3509 + }, + { + "epoch": 0.29910524073285044, + "grad_norm": 20.75, + "learning_rate": 2.3709515408861873e-06, + "loss": 1.1759, + "step": 3510 + }, + { + "epoch": 0.2991904559011504, + "grad_norm": 13.6875, + "learning_rate": 2.3708771725507605e-06, + "loss": 0.7445, + "step": 3511 + }, + { + "epoch": 0.29927567106945036, + "grad_norm": 17.75, + "learning_rate": 2.3708027839599567e-06, + "loss": 0.8531, + "step": 3512 + }, + { + "epoch": 0.2993608862377503, + "grad_norm": 12.6875, + "learning_rate": 2.3707283751151195e-06, + "loss": 0.45, + "step": 3513 + }, + { + "epoch": 0.29944610140605027, + "grad_norm": 9.75, + "learning_rate": 2.3706539460175933e-06, + "loss": 0.373, + "step": 3514 + }, + { + "epoch": 0.2995313165743502, + "grad_norm": 11.1875, + "learning_rate": 2.3705794966687236e-06, + "loss": 0.4872, + "step": 3515 + }, + { + "epoch": 0.2996165317426502, + "grad_norm": 15.625, + "learning_rate": 2.3705050270698554e-06, + "loss": 0.8585, + "step": 3516 + }, + { + "epoch": 0.29970174691095014, + "grad_norm": 12.625, + "learning_rate": 2.3704305372223346e-06, + "loss": 0.5092, + "step": 3517 + }, + { + "epoch": 0.2997869620792501, + "grad_norm": 18.625, + "learning_rate": 2.3703560271275073e-06, + "loss": 1.1137, + "step": 3518 + }, + { + "epoch": 0.29987217724755005, + "grad_norm": 13.3125, + "learning_rate": 2.37028149678672e-06, + "loss": 0.4338, + "step": 3519 + }, + { + "epoch": 0.29995739241585, + "grad_norm": 12.125, + "learning_rate": 2.3702069462013196e-06, + "loss": 0.4063, + "step": 3520 + }, + { + "epoch": 0.30004260758414997, + "grad_norm": 17.875, + "learning_rate": 2.3701323753726536e-06, + "loss": 0.8559, + "step": 3521 + }, + { + "epoch": 0.3001278227524499, + "grad_norm": 19.0, + "learning_rate": 2.3700577843020686e-06, + "loss": 1.0492, + "step": 3522 + }, + { + "epoch": 0.3002130379207499, + "grad_norm": 14.9375, + "learning_rate": 2.3699831729909135e-06, + "loss": 0.8995, + "step": 3523 + }, + { + "epoch": 0.30029825308904984, + "grad_norm": 11.125, + "learning_rate": 2.369908541440536e-06, + "loss": 0.5174, + "step": 3524 + }, + { + "epoch": 0.3003834682573498, + "grad_norm": 11.625, + "learning_rate": 2.369833889652285e-06, + "loss": 0.4916, + "step": 3525 + }, + { + "epoch": 0.30046868342564975, + "grad_norm": 14.1875, + "learning_rate": 2.36975921762751e-06, + "loss": 0.8715, + "step": 3526 + }, + { + "epoch": 0.3005538985939497, + "grad_norm": 9.6875, + "learning_rate": 2.369684525367559e-06, + "loss": 0.441, + "step": 3527 + }, + { + "epoch": 0.30063911376224967, + "grad_norm": 16.125, + "learning_rate": 2.3696098128737833e-06, + "loss": 0.4105, + "step": 3528 + }, + { + "epoch": 0.3007243289305496, + "grad_norm": 13.125, + "learning_rate": 2.3695350801475325e-06, + "loss": 0.6402, + "step": 3529 + }, + { + "epoch": 0.3008095440988496, + "grad_norm": 14.875, + "learning_rate": 2.369460327190157e-06, + "loss": 0.8228, + "step": 3530 + }, + { + "epoch": 0.30089475926714954, + "grad_norm": 17.0, + "learning_rate": 2.3693855540030074e-06, + "loss": 0.6175, + "step": 3531 + }, + { + "epoch": 0.3009799744354495, + "grad_norm": 46.5, + "learning_rate": 2.3693107605874355e-06, + "loss": 0.6063, + "step": 3532 + }, + { + "epoch": 0.30106518960374945, + "grad_norm": 21.125, + "learning_rate": 2.3692359469447923e-06, + "loss": 1.1743, + "step": 3533 + }, + { + "epoch": 0.3011504047720494, + "grad_norm": 14.1875, + "learning_rate": 2.3691611130764303e-06, + "loss": 0.9115, + "step": 3534 + }, + { + "epoch": 0.30123561994034936, + "grad_norm": 12.875, + "learning_rate": 2.3690862589837015e-06, + "loss": 0.5827, + "step": 3535 + }, + { + "epoch": 0.3013208351086493, + "grad_norm": 17.125, + "learning_rate": 2.3690113846679586e-06, + "loss": 0.9201, + "step": 3536 + }, + { + "epoch": 0.3014060502769493, + "grad_norm": 21.125, + "learning_rate": 2.368936490130555e-06, + "loss": 1.3266, + "step": 3537 + }, + { + "epoch": 0.30149126544524923, + "grad_norm": 17.75, + "learning_rate": 2.3688615753728435e-06, + "loss": 0.9453, + "step": 3538 + }, + { + "epoch": 0.3015764806135492, + "grad_norm": 12.3125, + "learning_rate": 2.3687866403961784e-06, + "loss": 0.6202, + "step": 3539 + }, + { + "epoch": 0.30166169578184915, + "grad_norm": 13.25, + "learning_rate": 2.368711685201914e-06, + "loss": 0.553, + "step": 3540 + }, + { + "epoch": 0.3017469109501491, + "grad_norm": 16.625, + "learning_rate": 2.3686367097914038e-06, + "loss": 0.5821, + "step": 3541 + }, + { + "epoch": 0.30183212611844906, + "grad_norm": 13.8125, + "learning_rate": 2.368561714166004e-06, + "loss": 0.5321, + "step": 3542 + }, + { + "epoch": 0.301917341286749, + "grad_norm": 12.8125, + "learning_rate": 2.3684866983270687e-06, + "loss": 0.4729, + "step": 3543 + }, + { + "epoch": 0.302002556455049, + "grad_norm": 13.5, + "learning_rate": 2.3684116622759547e-06, + "loss": 0.6648, + "step": 3544 + }, + { + "epoch": 0.30208777162334893, + "grad_norm": 21.125, + "learning_rate": 2.3683366060140166e-06, + "loss": 0.8463, + "step": 3545 + }, + { + "epoch": 0.3021729867916489, + "grad_norm": 14.25, + "learning_rate": 2.368261529542612e-06, + "loss": 0.7165, + "step": 3546 + }, + { + "epoch": 0.30225820195994885, + "grad_norm": 14.1875, + "learning_rate": 2.368186432863097e-06, + "loss": 0.597, + "step": 3547 + }, + { + "epoch": 0.3023434171282488, + "grad_norm": 15.1875, + "learning_rate": 2.3681113159768284e-06, + "loss": 0.7239, + "step": 3548 + }, + { + "epoch": 0.30242863229654876, + "grad_norm": 27.875, + "learning_rate": 2.368036178885164e-06, + "loss": 1.1407, + "step": 3549 + }, + { + "epoch": 0.3025138474648487, + "grad_norm": 12.1875, + "learning_rate": 2.367961021589462e-06, + "loss": 0.4478, + "step": 3550 + }, + { + "epoch": 0.3025990626331487, + "grad_norm": 17.875, + "learning_rate": 2.3678858440910795e-06, + "loss": 1.024, + "step": 3551 + }, + { + "epoch": 0.30268427780144863, + "grad_norm": 11.4375, + "learning_rate": 2.3678106463913765e-06, + "loss": 0.3483, + "step": 3552 + }, + { + "epoch": 0.3027694929697486, + "grad_norm": 14.75, + "learning_rate": 2.3677354284917102e-06, + "loss": 0.7777, + "step": 3553 + }, + { + "epoch": 0.30285470813804855, + "grad_norm": 16.125, + "learning_rate": 2.3676601903934417e-06, + "loss": 1.0319, + "step": 3554 + }, + { + "epoch": 0.30293992330634856, + "grad_norm": 13.4375, + "learning_rate": 2.367584932097929e-06, + "loss": 0.8394, + "step": 3555 + }, + { + "epoch": 0.3030251384746485, + "grad_norm": 16.625, + "learning_rate": 2.3675096536065327e-06, + "loss": 0.7785, + "step": 3556 + }, + { + "epoch": 0.30311035364294847, + "grad_norm": 13.0, + "learning_rate": 2.3674343549206137e-06, + "loss": 0.7054, + "step": 3557 + }, + { + "epoch": 0.30319556881124843, + "grad_norm": 12.625, + "learning_rate": 2.3673590360415318e-06, + "loss": 0.5428, + "step": 3558 + }, + { + "epoch": 0.3032807839795484, + "grad_norm": 18.125, + "learning_rate": 2.3672836969706484e-06, + "loss": 1.0269, + "step": 3559 + }, + { + "epoch": 0.30336599914784834, + "grad_norm": 13.5625, + "learning_rate": 2.3672083377093253e-06, + "loss": 0.7792, + "step": 3560 + }, + { + "epoch": 0.3034512143161483, + "grad_norm": 16.25, + "learning_rate": 2.367132958258924e-06, + "loss": 0.8149, + "step": 3561 + }, + { + "epoch": 0.30353642948444826, + "grad_norm": 15.8125, + "learning_rate": 2.367057558620807e-06, + "loss": 1.015, + "step": 3562 + }, + { + "epoch": 0.3036216446527482, + "grad_norm": 14.625, + "learning_rate": 2.366982138796337e-06, + "loss": 0.9353, + "step": 3563 + }, + { + "epoch": 0.30370685982104817, + "grad_norm": 21.125, + "learning_rate": 2.366906698786876e-06, + "loss": 0.5762, + "step": 3564 + }, + { + "epoch": 0.3037920749893481, + "grad_norm": 14.1875, + "learning_rate": 2.366831238593788e-06, + "loss": 1.0013, + "step": 3565 + }, + { + "epoch": 0.3038772901576481, + "grad_norm": 11.0625, + "learning_rate": 2.366755758218436e-06, + "loss": 0.3632, + "step": 3566 + }, + { + "epoch": 0.30396250532594804, + "grad_norm": 15.875, + "learning_rate": 2.366680257662185e-06, + "loss": 0.9475, + "step": 3567 + }, + { + "epoch": 0.304047720494248, + "grad_norm": 13.0, + "learning_rate": 2.366604736926399e-06, + "loss": 0.5595, + "step": 3568 + }, + { + "epoch": 0.30413293566254795, + "grad_norm": 12.125, + "learning_rate": 2.3665291960124422e-06, + "loss": 0.5741, + "step": 3569 + }, + { + "epoch": 0.3042181508308479, + "grad_norm": 12.1875, + "learning_rate": 2.3664536349216797e-06, + "loss": 0.4291, + "step": 3570 + }, + { + "epoch": 0.30430336599914787, + "grad_norm": 13.6875, + "learning_rate": 2.366378053655478e-06, + "loss": 0.7145, + "step": 3571 + }, + { + "epoch": 0.3043885811674478, + "grad_norm": 15.8125, + "learning_rate": 2.3663024522152015e-06, + "loss": 0.8785, + "step": 3572 + }, + { + "epoch": 0.3044737963357478, + "grad_norm": 10.125, + "learning_rate": 2.366226830602218e-06, + "loss": 0.3718, + "step": 3573 + }, + { + "epoch": 0.30455901150404774, + "grad_norm": 15.6875, + "learning_rate": 2.3661511888178928e-06, + "loss": 0.5046, + "step": 3574 + }, + { + "epoch": 0.3046442266723477, + "grad_norm": 21.875, + "learning_rate": 2.3660755268635933e-06, + "loss": 0.8949, + "step": 3575 + }, + { + "epoch": 0.30472944184064765, + "grad_norm": 17.875, + "learning_rate": 2.365999844740687e-06, + "loss": 1.0063, + "step": 3576 + }, + { + "epoch": 0.3048146570089476, + "grad_norm": 14.0, + "learning_rate": 2.365924142450541e-06, + "loss": 0.7833, + "step": 3577 + }, + { + "epoch": 0.30489987217724757, + "grad_norm": 14.0, + "learning_rate": 2.3658484199945237e-06, + "loss": 0.6318, + "step": 3578 + }, + { + "epoch": 0.3049850873455475, + "grad_norm": 16.625, + "learning_rate": 2.365772677374003e-06, + "loss": 0.8016, + "step": 3579 + }, + { + "epoch": 0.3050703025138475, + "grad_norm": 17.125, + "learning_rate": 2.3656969145903486e-06, + "loss": 0.9323, + "step": 3580 + }, + { + "epoch": 0.30515551768214744, + "grad_norm": 16.875, + "learning_rate": 2.3656211316449287e-06, + "loss": 1.0339, + "step": 3581 + }, + { + "epoch": 0.3052407328504474, + "grad_norm": 16.75, + "learning_rate": 2.3655453285391133e-06, + "loss": 0.6346, + "step": 3582 + }, + { + "epoch": 0.30532594801874735, + "grad_norm": 21.125, + "learning_rate": 2.3654695052742717e-06, + "loss": 0.9334, + "step": 3583 + }, + { + "epoch": 0.3054111631870473, + "grad_norm": 17.0, + "learning_rate": 2.3653936618517747e-06, + "loss": 0.9575, + "step": 3584 + }, + { + "epoch": 0.30549637835534726, + "grad_norm": 13.625, + "learning_rate": 2.365317798272993e-06, + "loss": 0.615, + "step": 3585 + }, + { + "epoch": 0.3055815935236472, + "grad_norm": 12.75, + "learning_rate": 2.365241914539296e-06, + "loss": 0.6607, + "step": 3586 + }, + { + "epoch": 0.3056668086919472, + "grad_norm": 15.375, + "learning_rate": 2.3651660106520574e-06, + "loss": 0.6869, + "step": 3587 + }, + { + "epoch": 0.30575202386024714, + "grad_norm": 15.75, + "learning_rate": 2.365090086612647e-06, + "loss": 0.7421, + "step": 3588 + }, + { + "epoch": 0.3058372390285471, + "grad_norm": 12.6875, + "learning_rate": 2.365014142422438e-06, + "loss": 0.6018, + "step": 3589 + }, + { + "epoch": 0.30592245419684705, + "grad_norm": 15.75, + "learning_rate": 2.364938178082801e-06, + "loss": 0.9748, + "step": 3590 + }, + { + "epoch": 0.306007669365147, + "grad_norm": 16.75, + "learning_rate": 2.364862193595111e-06, + "loss": 0.93, + "step": 3591 + }, + { + "epoch": 0.30609288453344696, + "grad_norm": 12.5, + "learning_rate": 2.36478618896074e-06, + "loss": 0.6869, + "step": 3592 + }, + { + "epoch": 0.3061780997017469, + "grad_norm": 15.8125, + "learning_rate": 2.364710164181061e-06, + "loss": 0.691, + "step": 3593 + }, + { + "epoch": 0.3062633148700469, + "grad_norm": 25.75, + "learning_rate": 2.364634119257449e-06, + "loss": 1.1043, + "step": 3594 + }, + { + "epoch": 0.30634853003834683, + "grad_norm": 14.9375, + "learning_rate": 2.3645580541912773e-06, + "loss": 0.6804, + "step": 3595 + }, + { + "epoch": 0.3064337452066468, + "grad_norm": 16.875, + "learning_rate": 2.3644819689839205e-06, + "loss": 0.9131, + "step": 3596 + }, + { + "epoch": 0.30651896037494675, + "grad_norm": 11.375, + "learning_rate": 2.3644058636367544e-06, + "loss": 0.5796, + "step": 3597 + }, + { + "epoch": 0.3066041755432467, + "grad_norm": 11.5, + "learning_rate": 2.364329738151153e-06, + "loss": 0.5598, + "step": 3598 + }, + { + "epoch": 0.30668939071154666, + "grad_norm": 13.1875, + "learning_rate": 2.3642535925284937e-06, + "loss": 0.78, + "step": 3599 + }, + { + "epoch": 0.3067746058798466, + "grad_norm": 10.125, + "learning_rate": 2.3641774267701505e-06, + "loss": 0.4551, + "step": 3600 + }, + { + "epoch": 0.3068598210481466, + "grad_norm": 12.375, + "learning_rate": 2.3641012408775018e-06, + "loss": 0.4729, + "step": 3601 + }, + { + "epoch": 0.30694503621644653, + "grad_norm": 15.125, + "learning_rate": 2.364025034851923e-06, + "loss": 0.8796, + "step": 3602 + }, + { + "epoch": 0.3070302513847465, + "grad_norm": 17.625, + "learning_rate": 2.363948808694791e-06, + "loss": 0.7959, + "step": 3603 + }, + { + "epoch": 0.30711546655304645, + "grad_norm": 10.875, + "learning_rate": 2.3638725624074844e-06, + "loss": 0.9934, + "step": 3604 + }, + { + "epoch": 0.3072006817213464, + "grad_norm": 21.875, + "learning_rate": 2.36379629599138e-06, + "loss": 0.9451, + "step": 3605 + }, + { + "epoch": 0.30728589688964636, + "grad_norm": 20.875, + "learning_rate": 2.363720009447857e-06, + "loss": 1.09, + "step": 3606 + }, + { + "epoch": 0.3073711120579463, + "grad_norm": 12.5, + "learning_rate": 2.3636437027782937e-06, + "loss": 0.6467, + "step": 3607 + }, + { + "epoch": 0.3074563272262463, + "grad_norm": 12.4375, + "learning_rate": 2.3635673759840685e-06, + "loss": 0.6728, + "step": 3608 + }, + { + "epoch": 0.30754154239454623, + "grad_norm": 23.5, + "learning_rate": 2.3634910290665614e-06, + "loss": 1.1796, + "step": 3609 + }, + { + "epoch": 0.3076267575628462, + "grad_norm": 15.9375, + "learning_rate": 2.363414662027151e-06, + "loss": 1.0791, + "step": 3610 + }, + { + "epoch": 0.30771197273114614, + "grad_norm": 12.125, + "learning_rate": 2.3633382748672186e-06, + "loss": 0.5334, + "step": 3611 + }, + { + "epoch": 0.3077971878994461, + "grad_norm": 13.0625, + "learning_rate": 2.363261867588144e-06, + "loss": 0.7565, + "step": 3612 + }, + { + "epoch": 0.30788240306774606, + "grad_norm": 15.375, + "learning_rate": 2.3631854401913077e-06, + "loss": 0.8815, + "step": 3613 + }, + { + "epoch": 0.307967618236046, + "grad_norm": 17.25, + "learning_rate": 2.3631089926780913e-06, + "loss": 1.0213, + "step": 3614 + }, + { + "epoch": 0.30805283340434597, + "grad_norm": 18.125, + "learning_rate": 2.363032525049876e-06, + "loss": 0.7114, + "step": 3615 + }, + { + "epoch": 0.30813804857264593, + "grad_norm": 17.25, + "learning_rate": 2.3629560373080436e-06, + "loss": 0.8506, + "step": 3616 + }, + { + "epoch": 0.3082232637409459, + "grad_norm": 29.625, + "learning_rate": 2.3628795294539767e-06, + "loss": 1.0163, + "step": 3617 + }, + { + "epoch": 0.30830847890924584, + "grad_norm": 11.1875, + "learning_rate": 2.3628030014890576e-06, + "loss": 0.3581, + "step": 3618 + }, + { + "epoch": 0.3083936940775458, + "grad_norm": 14.8125, + "learning_rate": 2.362726453414669e-06, + "loss": 0.5291, + "step": 3619 + }, + { + "epoch": 0.30847890924584576, + "grad_norm": 40.25, + "learning_rate": 2.3626498852321947e-06, + "loss": 0.9466, + "step": 3620 + }, + { + "epoch": 0.3085641244141457, + "grad_norm": 12.0, + "learning_rate": 2.362573296943018e-06, + "loss": 0.5573, + "step": 3621 + }, + { + "epoch": 0.30864933958244567, + "grad_norm": 11.4375, + "learning_rate": 2.362496688548523e-06, + "loss": 0.4074, + "step": 3622 + }, + { + "epoch": 0.3087345547507456, + "grad_norm": 15.25, + "learning_rate": 2.3624200600500943e-06, + "loss": 1.2185, + "step": 3623 + }, + { + "epoch": 0.3088197699190456, + "grad_norm": 14.625, + "learning_rate": 2.3623434114491166e-06, + "loss": 0.8924, + "step": 3624 + }, + { + "epoch": 0.30890498508734554, + "grad_norm": 14.5625, + "learning_rate": 2.3622667427469743e-06, + "loss": 0.7894, + "step": 3625 + }, + { + "epoch": 0.3089902002556455, + "grad_norm": 16.875, + "learning_rate": 2.3621900539450544e-06, + "loss": 1.0706, + "step": 3626 + }, + { + "epoch": 0.30907541542394545, + "grad_norm": 23.875, + "learning_rate": 2.362113345044741e-06, + "loss": 1.1409, + "step": 3627 + }, + { + "epoch": 0.3091606305922454, + "grad_norm": 12.125, + "learning_rate": 2.362036616047422e-06, + "loss": 0.7046, + "step": 3628 + }, + { + "epoch": 0.30924584576054537, + "grad_norm": 13.25, + "learning_rate": 2.3619598669544824e-06, + "loss": 0.6346, + "step": 3629 + }, + { + "epoch": 0.3093310609288453, + "grad_norm": 13.5625, + "learning_rate": 2.3618830977673096e-06, + "loss": 0.5055, + "step": 3630 + }, + { + "epoch": 0.3094162760971453, + "grad_norm": 19.0, + "learning_rate": 2.3618063084872917e-06, + "loss": 0.9511, + "step": 3631 + }, + { + "epoch": 0.30950149126544524, + "grad_norm": 13.9375, + "learning_rate": 2.361729499115816e-06, + "loss": 0.8389, + "step": 3632 + }, + { + "epoch": 0.3095867064337452, + "grad_norm": 15.875, + "learning_rate": 2.3616526696542698e-06, + "loss": 0.8227, + "step": 3633 + }, + { + "epoch": 0.30967192160204515, + "grad_norm": 15.9375, + "learning_rate": 2.361575820104042e-06, + "loss": 0.9879, + "step": 3634 + }, + { + "epoch": 0.3097571367703451, + "grad_norm": 17.5, + "learning_rate": 2.3614989504665216e-06, + "loss": 0.9241, + "step": 3635 + }, + { + "epoch": 0.30984235193864507, + "grad_norm": 14.625, + "learning_rate": 2.3614220607430973e-06, + "loss": 0.8256, + "step": 3636 + }, + { + "epoch": 0.309927567106945, + "grad_norm": 12.4375, + "learning_rate": 2.361345150935159e-06, + "loss": 0.5973, + "step": 3637 + }, + { + "epoch": 0.310012782275245, + "grad_norm": 12.75, + "learning_rate": 2.361268221044096e-06, + "loss": 0.5119, + "step": 3638 + }, + { + "epoch": 0.31009799744354494, + "grad_norm": 18.375, + "learning_rate": 2.3611912710712987e-06, + "loss": 0.7814, + "step": 3639 + }, + { + "epoch": 0.3101832126118449, + "grad_norm": 10.5, + "learning_rate": 2.361114301018158e-06, + "loss": 0.4291, + "step": 3640 + }, + { + "epoch": 0.31026842778014485, + "grad_norm": 17.875, + "learning_rate": 2.3610373108860646e-06, + "loss": 0.7523, + "step": 3641 + }, + { + "epoch": 0.3103536429484448, + "grad_norm": 15.3125, + "learning_rate": 2.36096030067641e-06, + "loss": 0.7863, + "step": 3642 + }, + { + "epoch": 0.31043885811674476, + "grad_norm": 14.8125, + "learning_rate": 2.3608832703905853e-06, + "loss": 0.7521, + "step": 3643 + }, + { + "epoch": 0.3105240732850447, + "grad_norm": 14.0625, + "learning_rate": 2.3608062200299825e-06, + "loss": 1.0106, + "step": 3644 + }, + { + "epoch": 0.3106092884533447, + "grad_norm": 16.375, + "learning_rate": 2.3607291495959947e-06, + "loss": 0.7794, + "step": 3645 + }, + { + "epoch": 0.31069450362164464, + "grad_norm": 13.3125, + "learning_rate": 2.360652059090014e-06, + "loss": 0.7421, + "step": 3646 + }, + { + "epoch": 0.3107797187899446, + "grad_norm": 15.0, + "learning_rate": 2.360574948513434e-06, + "loss": 0.78, + "step": 3647 + }, + { + "epoch": 0.31086493395824455, + "grad_norm": 12.6875, + "learning_rate": 2.360497817867648e-06, + "loss": 0.5243, + "step": 3648 + }, + { + "epoch": 0.3109501491265445, + "grad_norm": 12.125, + "learning_rate": 2.3604206671540495e-06, + "loss": 0.5473, + "step": 3649 + }, + { + "epoch": 0.31103536429484446, + "grad_norm": 14.25, + "learning_rate": 2.360343496374033e-06, + "loss": 0.9088, + "step": 3650 + }, + { + "epoch": 0.3111205794631444, + "grad_norm": 17.25, + "learning_rate": 2.360266305528993e-06, + "loss": 0.6185, + "step": 3651 + }, + { + "epoch": 0.3112057946314444, + "grad_norm": 13.375, + "learning_rate": 2.3601890946203245e-06, + "loss": 0.641, + "step": 3652 + }, + { + "epoch": 0.31129100979974433, + "grad_norm": 18.25, + "learning_rate": 2.3601118636494223e-06, + "loss": 0.6078, + "step": 3653 + }, + { + "epoch": 0.3113762249680443, + "grad_norm": 16.625, + "learning_rate": 2.3600346126176827e-06, + "loss": 0.7837, + "step": 3654 + }, + { + "epoch": 0.31146144013634425, + "grad_norm": 10.8125, + "learning_rate": 2.3599573415265013e-06, + "loss": 0.4894, + "step": 3655 + }, + { + "epoch": 0.3115466553046442, + "grad_norm": 14.3125, + "learning_rate": 2.3598800503772747e-06, + "loss": 0.8935, + "step": 3656 + }, + { + "epoch": 0.31163187047294416, + "grad_norm": 11.3125, + "learning_rate": 2.3598027391713994e-06, + "loss": 0.5816, + "step": 3657 + }, + { + "epoch": 0.3117170856412441, + "grad_norm": 18.375, + "learning_rate": 2.359725407910273e-06, + "loss": 0.977, + "step": 3658 + }, + { + "epoch": 0.3118023008095441, + "grad_norm": 14.125, + "learning_rate": 2.3596480565952923e-06, + "loss": 0.5791, + "step": 3659 + }, + { + "epoch": 0.31188751597784403, + "grad_norm": 14.625, + "learning_rate": 2.3595706852278554e-06, + "loss": 0.8229, + "step": 3660 + }, + { + "epoch": 0.311972731146144, + "grad_norm": 17.0, + "learning_rate": 2.35949329380936e-06, + "loss": 0.7805, + "step": 3661 + }, + { + "epoch": 0.31205794631444395, + "grad_norm": 12.5625, + "learning_rate": 2.3594158823412057e-06, + "loss": 0.5918, + "step": 3662 + }, + { + "epoch": 0.3121431614827439, + "grad_norm": 11.5625, + "learning_rate": 2.3593384508247906e-06, + "loss": 0.552, + "step": 3663 + }, + { + "epoch": 0.31222837665104386, + "grad_norm": 17.75, + "learning_rate": 2.359260999261514e-06, + "loss": 1.0187, + "step": 3664 + }, + { + "epoch": 0.3123135918193438, + "grad_norm": 22.75, + "learning_rate": 2.359183527652776e-06, + "loss": 0.9452, + "step": 3665 + }, + { + "epoch": 0.3123988069876438, + "grad_norm": 17.125, + "learning_rate": 2.359106035999976e-06, + "loss": 0.9554, + "step": 3666 + }, + { + "epoch": 0.31248402215594373, + "grad_norm": 13.8125, + "learning_rate": 2.359028524304515e-06, + "loss": 0.4401, + "step": 3667 + }, + { + "epoch": 0.31256923732424374, + "grad_norm": 19.125, + "learning_rate": 2.358950992567793e-06, + "loss": 1.0082, + "step": 3668 + }, + { + "epoch": 0.3126544524925437, + "grad_norm": 15.4375, + "learning_rate": 2.3588734407912116e-06, + "loss": 0.6147, + "step": 3669 + }, + { + "epoch": 0.31273966766084366, + "grad_norm": 15.5625, + "learning_rate": 2.358795868976172e-06, + "loss": 0.7906, + "step": 3670 + }, + { + "epoch": 0.3128248828291436, + "grad_norm": 16.5, + "learning_rate": 2.3587182771240764e-06, + "loss": 0.7904, + "step": 3671 + }, + { + "epoch": 0.31291009799744357, + "grad_norm": 17.0, + "learning_rate": 2.358640665236326e-06, + "loss": 0.8578, + "step": 3672 + }, + { + "epoch": 0.3129953131657435, + "grad_norm": 19.0, + "learning_rate": 2.3585630333143244e-06, + "loss": 1.0, + "step": 3673 + }, + { + "epoch": 0.3130805283340435, + "grad_norm": 12.375, + "learning_rate": 2.358485381359474e-06, + "loss": 0.4281, + "step": 3674 + }, + { + "epoch": 0.31316574350234344, + "grad_norm": 14.375, + "learning_rate": 2.3584077093731782e-06, + "loss": 0.855, + "step": 3675 + }, + { + "epoch": 0.3132509586706434, + "grad_norm": 12.6875, + "learning_rate": 2.35833001735684e-06, + "loss": 0.5559, + "step": 3676 + }, + { + "epoch": 0.31333617383894335, + "grad_norm": 18.875, + "learning_rate": 2.3582523053118644e-06, + "loss": 0.8259, + "step": 3677 + }, + { + "epoch": 0.3134213890072433, + "grad_norm": 30.0, + "learning_rate": 2.358174573239655e-06, + "loss": 1.0447, + "step": 3678 + }, + { + "epoch": 0.31350660417554327, + "grad_norm": 11.5, + "learning_rate": 2.358096821141617e-06, + "loss": 0.349, + "step": 3679 + }, + { + "epoch": 0.3135918193438432, + "grad_norm": 12.25, + "learning_rate": 2.3580190490191545e-06, + "loss": 0.4728, + "step": 3680 + }, + { + "epoch": 0.3136770345121432, + "grad_norm": 14.5625, + "learning_rate": 2.3579412568736743e-06, + "loss": 1.0822, + "step": 3681 + }, + { + "epoch": 0.31376224968044314, + "grad_norm": 15.4375, + "learning_rate": 2.3578634447065808e-06, + "loss": 0.6368, + "step": 3682 + }, + { + "epoch": 0.3138474648487431, + "grad_norm": 14.5, + "learning_rate": 2.357785612519281e-06, + "loss": 0.5831, + "step": 3683 + }, + { + "epoch": 0.31393268001704305, + "grad_norm": 18.125, + "learning_rate": 2.3577077603131816e-06, + "loss": 1.168, + "step": 3684 + }, + { + "epoch": 0.314017895185343, + "grad_norm": 15.3125, + "learning_rate": 2.3576298880896886e-06, + "loss": 0.9287, + "step": 3685 + }, + { + "epoch": 0.31410311035364297, + "grad_norm": 21.75, + "learning_rate": 2.3575519958502103e-06, + "loss": 0.7875, + "step": 3686 + }, + { + "epoch": 0.3141883255219429, + "grad_norm": 20.625, + "learning_rate": 2.357474083596153e-06, + "loss": 0.7885, + "step": 3687 + }, + { + "epoch": 0.3142735406902429, + "grad_norm": 19.5, + "learning_rate": 2.357396151328926e-06, + "loss": 0.9576, + "step": 3688 + }, + { + "epoch": 0.31435875585854284, + "grad_norm": 14.5625, + "learning_rate": 2.3573181990499367e-06, + "loss": 0.5863, + "step": 3689 + }, + { + "epoch": 0.3144439710268428, + "grad_norm": 21.75, + "learning_rate": 2.357240226760594e-06, + "loss": 0.7171, + "step": 3690 + }, + { + "epoch": 0.31452918619514275, + "grad_norm": 11.625, + "learning_rate": 2.357162234462307e-06, + "loss": 0.6142, + "step": 3691 + }, + { + "epoch": 0.3146144013634427, + "grad_norm": 12.5625, + "learning_rate": 2.357084222156485e-06, + "loss": 0.6429, + "step": 3692 + }, + { + "epoch": 0.31469961653174267, + "grad_norm": 16.875, + "learning_rate": 2.357006189844538e-06, + "loss": 1.093, + "step": 3693 + }, + { + "epoch": 0.3147848317000426, + "grad_norm": 15.0625, + "learning_rate": 2.356928137527876e-06, + "loss": 0.5054, + "step": 3694 + }, + { + "epoch": 0.3148700468683426, + "grad_norm": 12.8125, + "learning_rate": 2.3568500652079096e-06, + "loss": 0.6249, + "step": 3695 + }, + { + "epoch": 0.31495526203664254, + "grad_norm": 14.8125, + "learning_rate": 2.356771972886049e-06, + "loss": 0.7313, + "step": 3696 + }, + { + "epoch": 0.3150404772049425, + "grad_norm": 14.375, + "learning_rate": 2.3566938605637063e-06, + "loss": 0.7174, + "step": 3697 + }, + { + "epoch": 0.31512569237324245, + "grad_norm": 12.375, + "learning_rate": 2.3566157282422924e-06, + "loss": 0.4473, + "step": 3698 + }, + { + "epoch": 0.3152109075415424, + "grad_norm": 16.375, + "learning_rate": 2.35653757592322e-06, + "loss": 1.2287, + "step": 3699 + }, + { + "epoch": 0.31529612270984236, + "grad_norm": 11.9375, + "learning_rate": 2.3564594036079003e-06, + "loss": 0.5213, + "step": 3700 + }, + { + "epoch": 0.3153813378781423, + "grad_norm": 15.125, + "learning_rate": 2.356381211297747e-06, + "loss": 0.6685, + "step": 3701 + }, + { + "epoch": 0.3154665530464423, + "grad_norm": 18.25, + "learning_rate": 2.356302998994172e-06, + "loss": 1.0843, + "step": 3702 + }, + { + "epoch": 0.31555176821474223, + "grad_norm": 12.0625, + "learning_rate": 2.3562247666985903e-06, + "loss": 0.5893, + "step": 3703 + }, + { + "epoch": 0.3156369833830422, + "grad_norm": 18.25, + "learning_rate": 2.356146514412414e-06, + "loss": 0.9412, + "step": 3704 + }, + { + "epoch": 0.31572219855134215, + "grad_norm": 11.9375, + "learning_rate": 2.356068242137058e-06, + "loss": 0.4863, + "step": 3705 + }, + { + "epoch": 0.3158074137196421, + "grad_norm": 13.75, + "learning_rate": 2.3559899498739367e-06, + "loss": 1.0099, + "step": 3706 + }, + { + "epoch": 0.31589262888794206, + "grad_norm": 19.875, + "learning_rate": 2.355911637624465e-06, + "loss": 0.7066, + "step": 3707 + }, + { + "epoch": 0.315977844056242, + "grad_norm": 14.3125, + "learning_rate": 2.3558333053900574e-06, + "loss": 0.6905, + "step": 3708 + }, + { + "epoch": 0.316063059224542, + "grad_norm": 12.25, + "learning_rate": 2.3557549531721306e-06, + "loss": 0.6421, + "step": 3709 + }, + { + "epoch": 0.31614827439284193, + "grad_norm": 12.125, + "learning_rate": 2.3556765809720994e-06, + "loss": 0.6885, + "step": 3710 + }, + { + "epoch": 0.3162334895611419, + "grad_norm": 15.0, + "learning_rate": 2.355598188791381e-06, + "loss": 0.7999, + "step": 3711 + }, + { + "epoch": 0.31631870472944185, + "grad_norm": 28.875, + "learning_rate": 2.355519776631391e-06, + "loss": 1.2331, + "step": 3712 + }, + { + "epoch": 0.3164039198977418, + "grad_norm": 9.9375, + "learning_rate": 2.3554413444935474e-06, + "loss": 0.4101, + "step": 3713 + }, + { + "epoch": 0.31648913506604176, + "grad_norm": 12.25, + "learning_rate": 2.3553628923792666e-06, + "loss": 0.5618, + "step": 3714 + }, + { + "epoch": 0.3165743502343417, + "grad_norm": 19.25, + "learning_rate": 2.3552844202899673e-06, + "loss": 0.9511, + "step": 3715 + }, + { + "epoch": 0.3166595654026417, + "grad_norm": 29.375, + "learning_rate": 2.355205928227067e-06, + "loss": 1.1118, + "step": 3716 + }, + { + "epoch": 0.31674478057094163, + "grad_norm": 14.375, + "learning_rate": 2.355127416191984e-06, + "loss": 0.8699, + "step": 3717 + }, + { + "epoch": 0.3168299957392416, + "grad_norm": 13.625, + "learning_rate": 2.3550488841861374e-06, + "loss": 0.5536, + "step": 3718 + }, + { + "epoch": 0.31691521090754154, + "grad_norm": 14.9375, + "learning_rate": 2.354970332210946e-06, + "loss": 0.7641, + "step": 3719 + }, + { + "epoch": 0.3170004260758415, + "grad_norm": 15.75, + "learning_rate": 2.35489176026783e-06, + "loss": 0.5738, + "step": 3720 + }, + { + "epoch": 0.31708564124414146, + "grad_norm": 11.8125, + "learning_rate": 2.3548131683582088e-06, + "loss": 0.497, + "step": 3721 + }, + { + "epoch": 0.3171708564124414, + "grad_norm": 14.8125, + "learning_rate": 2.354734556483503e-06, + "loss": 1.0743, + "step": 3722 + }, + { + "epoch": 0.3172560715807414, + "grad_norm": 11.9375, + "learning_rate": 2.354655924645132e-06, + "loss": 0.7099, + "step": 3723 + }, + { + "epoch": 0.31734128674904133, + "grad_norm": 12.3125, + "learning_rate": 2.354577272844518e-06, + "loss": 0.6531, + "step": 3724 + }, + { + "epoch": 0.3174265019173413, + "grad_norm": 19.25, + "learning_rate": 2.354498601083082e-06, + "loss": 0.8511, + "step": 3725 + }, + { + "epoch": 0.31751171708564124, + "grad_norm": 22.5, + "learning_rate": 2.3544199093622455e-06, + "loss": 1.1105, + "step": 3726 + }, + { + "epoch": 0.3175969322539412, + "grad_norm": 14.125, + "learning_rate": 2.3543411976834305e-06, + "loss": 0.6921, + "step": 3727 + }, + { + "epoch": 0.31768214742224116, + "grad_norm": 13.125, + "learning_rate": 2.3542624660480603e-06, + "loss": 0.4625, + "step": 3728 + }, + { + "epoch": 0.3177673625905411, + "grad_norm": 14.75, + "learning_rate": 2.3541837144575565e-06, + "loss": 1.1254, + "step": 3729 + }, + { + "epoch": 0.31785257775884107, + "grad_norm": 17.375, + "learning_rate": 2.354104942913342e-06, + "loss": 1.0532, + "step": 3730 + }, + { + "epoch": 0.317937792927141, + "grad_norm": 17.625, + "learning_rate": 2.3540261514168418e-06, + "loss": 0.6351, + "step": 3731 + }, + { + "epoch": 0.318023008095441, + "grad_norm": 252.0, + "learning_rate": 2.3539473399694782e-06, + "loss": 1.0827, + "step": 3732 + }, + { + "epoch": 0.31810822326374094, + "grad_norm": 10.5625, + "learning_rate": 2.3538685085726766e-06, + "loss": 0.4128, + "step": 3733 + }, + { + "epoch": 0.3181934384320409, + "grad_norm": 20.125, + "learning_rate": 2.353789657227861e-06, + "loss": 0.355, + "step": 3734 + }, + { + "epoch": 0.31827865360034086, + "grad_norm": 9.9375, + "learning_rate": 2.353710785936456e-06, + "loss": 0.444, + "step": 3735 + }, + { + "epoch": 0.3183638687686408, + "grad_norm": 17.0, + "learning_rate": 2.3536318946998877e-06, + "loss": 0.5497, + "step": 3736 + }, + { + "epoch": 0.31844908393694077, + "grad_norm": 13.0, + "learning_rate": 2.353552983519581e-06, + "loss": 0.6666, + "step": 3737 + }, + { + "epoch": 0.3185342991052407, + "grad_norm": 11.6875, + "learning_rate": 2.353474052396962e-06, + "loss": 0.4177, + "step": 3738 + }, + { + "epoch": 0.3186195142735407, + "grad_norm": 16.875, + "learning_rate": 2.3533951013334575e-06, + "loss": 0.6591, + "step": 3739 + }, + { + "epoch": 0.31870472944184064, + "grad_norm": 14.125, + "learning_rate": 2.353316130330494e-06, + "loss": 0.8239, + "step": 3740 + }, + { + "epoch": 0.3187899446101406, + "grad_norm": 11.4375, + "learning_rate": 2.3532371393894983e-06, + "loss": 0.3792, + "step": 3741 + }, + { + "epoch": 0.31887515977844055, + "grad_norm": 14.5, + "learning_rate": 2.3531581285118983e-06, + "loss": 0.7582, + "step": 3742 + }, + { + "epoch": 0.3189603749467405, + "grad_norm": 18.75, + "learning_rate": 2.3530790976991214e-06, + "loss": 1.1234, + "step": 3743 + }, + { + "epoch": 0.31904559011504047, + "grad_norm": 13.8125, + "learning_rate": 2.353000046952596e-06, + "loss": 0.7819, + "step": 3744 + }, + { + "epoch": 0.3191308052833404, + "grad_norm": 12.9375, + "learning_rate": 2.352920976273751e-06, + "loss": 0.6734, + "step": 3745 + }, + { + "epoch": 0.3192160204516404, + "grad_norm": 16.125, + "learning_rate": 2.3528418856640145e-06, + "loss": 0.542, + "step": 3746 + }, + { + "epoch": 0.31930123561994034, + "grad_norm": 17.0, + "learning_rate": 2.352762775124816e-06, + "loss": 1.0065, + "step": 3747 + }, + { + "epoch": 0.3193864507882403, + "grad_norm": 34.5, + "learning_rate": 2.352683644657586e-06, + "loss": 0.9629, + "step": 3748 + }, + { + "epoch": 0.31947166595654025, + "grad_norm": 12.25, + "learning_rate": 2.3526044942637526e-06, + "loss": 0.7039, + "step": 3749 + }, + { + "epoch": 0.3195568811248402, + "grad_norm": 13.8125, + "learning_rate": 2.352525323944748e-06, + "loss": 0.7703, + "step": 3750 + }, + { + "epoch": 0.31964209629314017, + "grad_norm": 15.5625, + "learning_rate": 2.352446133702002e-06, + "loss": 0.7529, + "step": 3751 + }, + { + "epoch": 0.3197273114614401, + "grad_norm": 13.4375, + "learning_rate": 2.3523669235369457e-06, + "loss": 0.4563, + "step": 3752 + }, + { + "epoch": 0.3198125266297401, + "grad_norm": 12.3125, + "learning_rate": 2.3522876934510104e-06, + "loss": 0.5116, + "step": 3753 + }, + { + "epoch": 0.31989774179804004, + "grad_norm": 15.25, + "learning_rate": 2.352208443445628e-06, + "loss": 0.9719, + "step": 3754 + }, + { + "epoch": 0.31998295696634, + "grad_norm": 15.0, + "learning_rate": 2.352129173522231e-06, + "loss": 0.6953, + "step": 3755 + }, + { + "epoch": 0.32006817213463995, + "grad_norm": 9.5625, + "learning_rate": 2.3520498836822515e-06, + "loss": 0.288, + "step": 3756 + }, + { + "epoch": 0.3201533873029399, + "grad_norm": 12.5625, + "learning_rate": 2.3519705739271224e-06, + "loss": 0.5002, + "step": 3757 + }, + { + "epoch": 0.32023860247123986, + "grad_norm": 14.625, + "learning_rate": 2.351891244258277e-06, + "loss": 0.7877, + "step": 3758 + }, + { + "epoch": 0.3203238176395398, + "grad_norm": 9.5625, + "learning_rate": 2.3518118946771484e-06, + "loss": 0.3697, + "step": 3759 + }, + { + "epoch": 0.3204090328078398, + "grad_norm": 15.25, + "learning_rate": 2.3517325251851716e-06, + "loss": 0.6415, + "step": 3760 + }, + { + "epoch": 0.32049424797613973, + "grad_norm": 17.625, + "learning_rate": 2.3516531357837793e-06, + "loss": 0.9096, + "step": 3761 + }, + { + "epoch": 0.3205794631444397, + "grad_norm": 15.375, + "learning_rate": 2.3515737264744077e-06, + "loss": 1.1126, + "step": 3762 + }, + { + "epoch": 0.32066467831273965, + "grad_norm": 12.5625, + "learning_rate": 2.351494297258491e-06, + "loss": 0.4516, + "step": 3763 + }, + { + "epoch": 0.3207498934810396, + "grad_norm": 17.75, + "learning_rate": 2.3514148481374643e-06, + "loss": 0.9108, + "step": 3764 + }, + { + "epoch": 0.32083510864933956, + "grad_norm": 14.9375, + "learning_rate": 2.351335379112764e-06, + "loss": 0.8066, + "step": 3765 + }, + { + "epoch": 0.3209203238176395, + "grad_norm": 13.9375, + "learning_rate": 2.351255890185826e-06, + "loss": 0.6022, + "step": 3766 + }, + { + "epoch": 0.3210055389859395, + "grad_norm": 15.75, + "learning_rate": 2.351176381358087e-06, + "loss": 1.2152, + "step": 3767 + }, + { + "epoch": 0.32109075415423943, + "grad_norm": 14.0625, + "learning_rate": 2.351096852630983e-06, + "loss": 0.8962, + "step": 3768 + }, + { + "epoch": 0.3211759693225394, + "grad_norm": 15.125, + "learning_rate": 2.351017304005952e-06, + "loss": 0.7988, + "step": 3769 + }, + { + "epoch": 0.32126118449083935, + "grad_norm": 11.625, + "learning_rate": 2.350937735484431e-06, + "loss": 0.3469, + "step": 3770 + }, + { + "epoch": 0.3213463996591393, + "grad_norm": 17.375, + "learning_rate": 2.3508581470678576e-06, + "loss": 1.0225, + "step": 3771 + }, + { + "epoch": 0.32143161482743926, + "grad_norm": 16.625, + "learning_rate": 2.350778538757671e-06, + "loss": 0.7168, + "step": 3772 + }, + { + "epoch": 0.3215168299957392, + "grad_norm": 17.625, + "learning_rate": 2.350698910555309e-06, + "loss": 0.7265, + "step": 3773 + }, + { + "epoch": 0.3216020451640392, + "grad_norm": 19.625, + "learning_rate": 2.350619262462211e-06, + "loss": 0.8753, + "step": 3774 + }, + { + "epoch": 0.32168726033233913, + "grad_norm": 15.4375, + "learning_rate": 2.350539594479816e-06, + "loss": 0.6382, + "step": 3775 + }, + { + "epoch": 0.3217724755006391, + "grad_norm": 12.25, + "learning_rate": 2.350459906609564e-06, + "loss": 0.5823, + "step": 3776 + }, + { + "epoch": 0.32185769066893904, + "grad_norm": 16.875, + "learning_rate": 2.350380198852895e-06, + "loss": 0.6452, + "step": 3777 + }, + { + "epoch": 0.321942905837239, + "grad_norm": 22.375, + "learning_rate": 2.3503004712112495e-06, + "loss": 0.9861, + "step": 3778 + }, + { + "epoch": 0.32202812100553896, + "grad_norm": 29.375, + "learning_rate": 2.3502207236860673e-06, + "loss": 1.3897, + "step": 3779 + }, + { + "epoch": 0.3221133361738389, + "grad_norm": 11.5625, + "learning_rate": 2.350140956278791e-06, + "loss": 0.4045, + "step": 3780 + }, + { + "epoch": 0.32219855134213893, + "grad_norm": 12.5, + "learning_rate": 2.3500611689908607e-06, + "loss": 0.6438, + "step": 3781 + }, + { + "epoch": 0.3222837665104389, + "grad_norm": 14.5625, + "learning_rate": 2.349981361823719e-06, + "loss": 0.8372, + "step": 3782 + }, + { + "epoch": 0.32236898167873884, + "grad_norm": 11.0, + "learning_rate": 2.3499015347788083e-06, + "loss": 0.6228, + "step": 3783 + }, + { + "epoch": 0.3224541968470388, + "grad_norm": 10.5625, + "learning_rate": 2.349821687857571e-06, + "loss": 0.4362, + "step": 3784 + }, + { + "epoch": 0.32253941201533876, + "grad_norm": 12.375, + "learning_rate": 2.3497418210614494e-06, + "loss": 0.6263, + "step": 3785 + }, + { + "epoch": 0.3226246271836387, + "grad_norm": 14.0625, + "learning_rate": 2.3496619343918873e-06, + "loss": 0.5625, + "step": 3786 + }, + { + "epoch": 0.32270984235193867, + "grad_norm": 14.75, + "learning_rate": 2.3495820278503284e-06, + "loss": 0.7221, + "step": 3787 + }, + { + "epoch": 0.3227950575202386, + "grad_norm": 16.0, + "learning_rate": 2.3495021014382163e-06, + "loss": 0.8427, + "step": 3788 + }, + { + "epoch": 0.3228802726885386, + "grad_norm": 19.0, + "learning_rate": 2.3494221551569958e-06, + "loss": 1.0603, + "step": 3789 + }, + { + "epoch": 0.32296548785683854, + "grad_norm": 12.8125, + "learning_rate": 2.3493421890081113e-06, + "loss": 0.6749, + "step": 3790 + }, + { + "epoch": 0.3230507030251385, + "grad_norm": 13.4375, + "learning_rate": 2.3492622029930085e-06, + "loss": 0.5493, + "step": 3791 + }, + { + "epoch": 0.32313591819343845, + "grad_norm": 12.25, + "learning_rate": 2.349182197113132e-06, + "loss": 0.6549, + "step": 3792 + }, + { + "epoch": 0.3232211333617384, + "grad_norm": 17.125, + "learning_rate": 2.3491021713699273e-06, + "loss": 1.0181, + "step": 3793 + }, + { + "epoch": 0.32330634853003837, + "grad_norm": 19.5, + "learning_rate": 2.3490221257648415e-06, + "loss": 0.5747, + "step": 3794 + }, + { + "epoch": 0.3233915636983383, + "grad_norm": 27.375, + "learning_rate": 2.348942060299321e-06, + "loss": 0.975, + "step": 3795 + }, + { + "epoch": 0.3234767788666383, + "grad_norm": 11.5, + "learning_rate": 2.348861974974812e-06, + "loss": 0.4122, + "step": 3796 + }, + { + "epoch": 0.32356199403493824, + "grad_norm": 15.5625, + "learning_rate": 2.3487818697927623e-06, + "loss": 1.05, + "step": 3797 + }, + { + "epoch": 0.3236472092032382, + "grad_norm": 13.3125, + "learning_rate": 2.3487017447546196e-06, + "loss": 0.8526, + "step": 3798 + }, + { + "epoch": 0.32373242437153815, + "grad_norm": 17.625, + "learning_rate": 2.348621599861831e-06, + "loss": 1.1766, + "step": 3799 + }, + { + "epoch": 0.3238176395398381, + "grad_norm": 15.875, + "learning_rate": 2.3485414351158454e-06, + "loss": 0.5928, + "step": 3800 + }, + { + "epoch": 0.32390285470813807, + "grad_norm": 14.5625, + "learning_rate": 2.3484612505181117e-06, + "loss": 0.888, + "step": 3801 + }, + { + "epoch": 0.323988069876438, + "grad_norm": 14.0, + "learning_rate": 2.3483810460700783e-06, + "loss": 0.5663, + "step": 3802 + }, + { + "epoch": 0.324073285044738, + "grad_norm": 13.1875, + "learning_rate": 2.3483008217731953e-06, + "loss": 0.4324, + "step": 3803 + }, + { + "epoch": 0.32415850021303794, + "grad_norm": 12.25, + "learning_rate": 2.348220577628912e-06, + "loss": 0.5498, + "step": 3804 + }, + { + "epoch": 0.3242437153813379, + "grad_norm": 28.25, + "learning_rate": 2.348140313638678e-06, + "loss": 1.1247, + "step": 3805 + }, + { + "epoch": 0.32432893054963785, + "grad_norm": 17.875, + "learning_rate": 2.3480600298039446e-06, + "loss": 0.9968, + "step": 3806 + }, + { + "epoch": 0.3244141457179378, + "grad_norm": 14.375, + "learning_rate": 2.3479797261261623e-06, + "loss": 0.5682, + "step": 3807 + }, + { + "epoch": 0.32449936088623776, + "grad_norm": 19.5, + "learning_rate": 2.3478994026067824e-06, + "loss": 0.9215, + "step": 3808 + }, + { + "epoch": 0.3245845760545377, + "grad_norm": 15.3125, + "learning_rate": 2.347819059247256e-06, + "loss": 0.9493, + "step": 3809 + }, + { + "epoch": 0.3246697912228377, + "grad_norm": 13.6875, + "learning_rate": 2.3477386960490358e-06, + "loss": 0.4373, + "step": 3810 + }, + { + "epoch": 0.32475500639113764, + "grad_norm": 20.375, + "learning_rate": 2.3476583130135727e-06, + "loss": 0.9943, + "step": 3811 + }, + { + "epoch": 0.3248402215594376, + "grad_norm": 13.4375, + "learning_rate": 2.3475779101423205e-06, + "loss": 0.6435, + "step": 3812 + }, + { + "epoch": 0.32492543672773755, + "grad_norm": 13.125, + "learning_rate": 2.3474974874367317e-06, + "loss": 0.7465, + "step": 3813 + }, + { + "epoch": 0.3250106518960375, + "grad_norm": 14.0, + "learning_rate": 2.3474170448982596e-06, + "loss": 0.8447, + "step": 3814 + }, + { + "epoch": 0.32509586706433746, + "grad_norm": 16.125, + "learning_rate": 2.3473365825283583e-06, + "loss": 1.0348, + "step": 3815 + }, + { + "epoch": 0.3251810822326374, + "grad_norm": 11.5625, + "learning_rate": 2.347256100328481e-06, + "loss": 0.4889, + "step": 3816 + }, + { + "epoch": 0.3252662974009374, + "grad_norm": 12.1875, + "learning_rate": 2.3471755983000825e-06, + "loss": 0.6483, + "step": 3817 + }, + { + "epoch": 0.32535151256923733, + "grad_norm": 15.8125, + "learning_rate": 2.347095076444618e-06, + "loss": 0.5665, + "step": 3818 + }, + { + "epoch": 0.3254367277375373, + "grad_norm": 16.75, + "learning_rate": 2.3470145347635427e-06, + "loss": 0.8475, + "step": 3819 + }, + { + "epoch": 0.32552194290583725, + "grad_norm": 13.8125, + "learning_rate": 2.346933973258311e-06, + "loss": 0.908, + "step": 3820 + }, + { + "epoch": 0.3256071580741372, + "grad_norm": 14.6875, + "learning_rate": 2.3468533919303797e-06, + "loss": 0.6939, + "step": 3821 + }, + { + "epoch": 0.32569237324243716, + "grad_norm": 16.25, + "learning_rate": 2.3467727907812044e-06, + "loss": 1.059, + "step": 3822 + }, + { + "epoch": 0.3257775884107371, + "grad_norm": 12.875, + "learning_rate": 2.346692169812242e-06, + "loss": 0.7456, + "step": 3823 + }, + { + "epoch": 0.3258628035790371, + "grad_norm": 13.25, + "learning_rate": 2.3466115290249493e-06, + "loss": 0.6671, + "step": 3824 + }, + { + "epoch": 0.32594801874733703, + "grad_norm": 14.0625, + "learning_rate": 2.346530868420783e-06, + "loss": 0.7496, + "step": 3825 + }, + { + "epoch": 0.326033233915637, + "grad_norm": 14.75, + "learning_rate": 2.346450188001202e-06, + "loss": 0.7022, + "step": 3826 + }, + { + "epoch": 0.32611844908393695, + "grad_norm": 16.75, + "learning_rate": 2.346369487767663e-06, + "loss": 0.7458, + "step": 3827 + }, + { + "epoch": 0.3262036642522369, + "grad_norm": 18.0, + "learning_rate": 2.3462887677216254e-06, + "loss": 0.9712, + "step": 3828 + }, + { + "epoch": 0.32628887942053686, + "grad_norm": 13.625, + "learning_rate": 2.346208027864547e-06, + "loss": 0.667, + "step": 3829 + }, + { + "epoch": 0.3263740945888368, + "grad_norm": 15.6875, + "learning_rate": 2.3461272681978873e-06, + "loss": 0.9095, + "step": 3830 + }, + { + "epoch": 0.3264593097571368, + "grad_norm": 14.6875, + "learning_rate": 2.346046488723106e-06, + "loss": 0.7896, + "step": 3831 + }, + { + "epoch": 0.32654452492543673, + "grad_norm": 12.5, + "learning_rate": 2.3459656894416616e-06, + "loss": 0.5909, + "step": 3832 + }, + { + "epoch": 0.3266297400937367, + "grad_norm": 15.75, + "learning_rate": 2.345884870355016e-06, + "loss": 0.6627, + "step": 3833 + }, + { + "epoch": 0.32671495526203664, + "grad_norm": 11.9375, + "learning_rate": 2.345804031464628e-06, + "loss": 0.4878, + "step": 3834 + }, + { + "epoch": 0.3268001704303366, + "grad_norm": 17.75, + "learning_rate": 2.34572317277196e-06, + "loss": 1.0159, + "step": 3835 + }, + { + "epoch": 0.32688538559863656, + "grad_norm": 18.125, + "learning_rate": 2.3456422942784718e-06, + "loss": 0.7609, + "step": 3836 + }, + { + "epoch": 0.3269706007669365, + "grad_norm": 10.6875, + "learning_rate": 2.345561395985626e-06, + "loss": 0.5056, + "step": 3837 + }, + { + "epoch": 0.32705581593523647, + "grad_norm": 11.4375, + "learning_rate": 2.345480477894884e-06, + "loss": 0.5038, + "step": 3838 + }, + { + "epoch": 0.32714103110353643, + "grad_norm": 10.25, + "learning_rate": 2.3453995400077085e-06, + "loss": 0.4173, + "step": 3839 + }, + { + "epoch": 0.3272262462718364, + "grad_norm": 27.25, + "learning_rate": 2.345318582325561e-06, + "loss": 1.0164, + "step": 3840 + }, + { + "epoch": 0.32731146144013634, + "grad_norm": 12.125, + "learning_rate": 2.345237604849906e-06, + "loss": 0.5116, + "step": 3841 + }, + { + "epoch": 0.3273966766084363, + "grad_norm": 11.375, + "learning_rate": 2.345156607582206e-06, + "loss": 0.3717, + "step": 3842 + }, + { + "epoch": 0.32748189177673626, + "grad_norm": 13.0625, + "learning_rate": 2.3450755905239246e-06, + "loss": 0.555, + "step": 3843 + }, + { + "epoch": 0.3275671069450362, + "grad_norm": 13.625, + "learning_rate": 2.344994553676526e-06, + "loss": 0.5815, + "step": 3844 + }, + { + "epoch": 0.32765232211333617, + "grad_norm": 13.3125, + "learning_rate": 2.344913497041475e-06, + "loss": 0.5783, + "step": 3845 + }, + { + "epoch": 0.3277375372816361, + "grad_norm": 12.5, + "learning_rate": 2.344832420620236e-06, + "loss": 0.6687, + "step": 3846 + }, + { + "epoch": 0.3278227524499361, + "grad_norm": 19.5, + "learning_rate": 2.3447513244142745e-06, + "loss": 0.8924, + "step": 3847 + }, + { + "epoch": 0.32790796761823604, + "grad_norm": 14.0625, + "learning_rate": 2.3446702084250552e-06, + "loss": 0.9233, + "step": 3848 + }, + { + "epoch": 0.327993182786536, + "grad_norm": 19.125, + "learning_rate": 2.344589072654045e-06, + "loss": 0.9593, + "step": 3849 + }, + { + "epoch": 0.32807839795483595, + "grad_norm": 12.1875, + "learning_rate": 2.344507917102709e-06, + "loss": 0.5729, + "step": 3850 + }, + { + "epoch": 0.3281636131231359, + "grad_norm": 17.25, + "learning_rate": 2.3444267417725147e-06, + "loss": 1.1154, + "step": 3851 + }, + { + "epoch": 0.32824882829143587, + "grad_norm": 20.75, + "learning_rate": 2.3443455466649285e-06, + "loss": 0.7493, + "step": 3852 + }, + { + "epoch": 0.3283340434597358, + "grad_norm": 25.875, + "learning_rate": 2.344264331781418e-06, + "loss": 0.8772, + "step": 3853 + }, + { + "epoch": 0.3284192586280358, + "grad_norm": 29.125, + "learning_rate": 2.3441830971234503e-06, + "loss": 1.1599, + "step": 3854 + }, + { + "epoch": 0.32850447379633574, + "grad_norm": 12.875, + "learning_rate": 2.3441018426924937e-06, + "loss": 0.5125, + "step": 3855 + }, + { + "epoch": 0.3285896889646357, + "grad_norm": 19.375, + "learning_rate": 2.344020568490017e-06, + "loss": 1.3051, + "step": 3856 + }, + { + "epoch": 0.32867490413293565, + "grad_norm": 18.5, + "learning_rate": 2.3439392745174885e-06, + "loss": 0.6417, + "step": 3857 + }, + { + "epoch": 0.3287601193012356, + "grad_norm": 17.125, + "learning_rate": 2.3438579607763774e-06, + "loss": 1.1153, + "step": 3858 + }, + { + "epoch": 0.32884533446953557, + "grad_norm": 19.25, + "learning_rate": 2.3437766272681523e-06, + "loss": 1.2917, + "step": 3859 + }, + { + "epoch": 0.3289305496378355, + "grad_norm": 14.1875, + "learning_rate": 2.3436952739942843e-06, + "loss": 0.7366, + "step": 3860 + }, + { + "epoch": 0.3290157648061355, + "grad_norm": 17.0, + "learning_rate": 2.3436139009562425e-06, + "loss": 0.9502, + "step": 3861 + }, + { + "epoch": 0.32910097997443544, + "grad_norm": 19.25, + "learning_rate": 2.343532508155498e-06, + "loss": 0.8141, + "step": 3862 + }, + { + "epoch": 0.3291861951427354, + "grad_norm": 12.4375, + "learning_rate": 2.3434510955935217e-06, + "loss": 0.5179, + "step": 3863 + }, + { + "epoch": 0.32927141031103535, + "grad_norm": 11.0625, + "learning_rate": 2.343369663271784e-06, + "loss": 0.5175, + "step": 3864 + }, + { + "epoch": 0.3293566254793353, + "grad_norm": 21.5, + "learning_rate": 2.3432882111917574e-06, + "loss": 0.9953, + "step": 3865 + }, + { + "epoch": 0.32944184064763526, + "grad_norm": 12.6875, + "learning_rate": 2.3432067393549134e-06, + "loss": 0.5388, + "step": 3866 + }, + { + "epoch": 0.3295270558159352, + "grad_norm": 14.3125, + "learning_rate": 2.3431252477627242e-06, + "loss": 0.7496, + "step": 3867 + }, + { + "epoch": 0.3296122709842352, + "grad_norm": 13.0, + "learning_rate": 2.3430437364166624e-06, + "loss": 0.4234, + "step": 3868 + }, + { + "epoch": 0.32969748615253514, + "grad_norm": 12.5625, + "learning_rate": 2.3429622053182018e-06, + "loss": 0.6817, + "step": 3869 + }, + { + "epoch": 0.3297827013208351, + "grad_norm": 17.0, + "learning_rate": 2.3428806544688145e-06, + "loss": 0.8044, + "step": 3870 + }, + { + "epoch": 0.32986791648913505, + "grad_norm": 15.0625, + "learning_rate": 2.342799083869975e-06, + "loss": 0.8683, + "step": 3871 + }, + { + "epoch": 0.329953131657435, + "grad_norm": 15.625, + "learning_rate": 2.342717493523157e-06, + "loss": 1.0609, + "step": 3872 + }, + { + "epoch": 0.33003834682573496, + "grad_norm": 14.375, + "learning_rate": 2.342635883429835e-06, + "loss": 0.5665, + "step": 3873 + }, + { + "epoch": 0.3301235619940349, + "grad_norm": 11.5625, + "learning_rate": 2.342554253591484e-06, + "loss": 0.5236, + "step": 3874 + }, + { + "epoch": 0.3302087771623349, + "grad_norm": 15.75, + "learning_rate": 2.3424726040095793e-06, + "loss": 0.7493, + "step": 3875 + }, + { + "epoch": 0.33029399233063483, + "grad_norm": 24.625, + "learning_rate": 2.342390934685596e-06, + "loss": 0.9125, + "step": 3876 + }, + { + "epoch": 0.3303792074989348, + "grad_norm": 10.5625, + "learning_rate": 2.34230924562101e-06, + "loss": 0.368, + "step": 3877 + }, + { + "epoch": 0.33046442266723475, + "grad_norm": 11.9375, + "learning_rate": 2.342227536817297e-06, + "loss": 0.5587, + "step": 3878 + }, + { + "epoch": 0.3305496378355347, + "grad_norm": 15.1875, + "learning_rate": 2.3421458082759347e-06, + "loss": 0.4354, + "step": 3879 + }, + { + "epoch": 0.33063485300383466, + "grad_norm": 18.125, + "learning_rate": 2.342064059998399e-06, + "loss": 0.7428, + "step": 3880 + }, + { + "epoch": 0.3307200681721346, + "grad_norm": 13.875, + "learning_rate": 2.341982291986168e-06, + "loss": 0.6956, + "step": 3881 + }, + { + "epoch": 0.3308052833404346, + "grad_norm": 16.75, + "learning_rate": 2.3419005042407184e-06, + "loss": 1.0907, + "step": 3882 + }, + { + "epoch": 0.33089049850873453, + "grad_norm": 14.1875, + "learning_rate": 2.341818696763529e-06, + "loss": 0.8007, + "step": 3883 + }, + { + "epoch": 0.3309757136770345, + "grad_norm": 19.0, + "learning_rate": 2.3417368695560777e-06, + "loss": 1.1786, + "step": 3884 + }, + { + "epoch": 0.33106092884533445, + "grad_norm": 16.875, + "learning_rate": 2.3416550226198435e-06, + "loss": 0.6097, + "step": 3885 + }, + { + "epoch": 0.3311461440136344, + "grad_norm": 25.375, + "learning_rate": 2.341573155956305e-06, + "loss": 0.7952, + "step": 3886 + }, + { + "epoch": 0.33123135918193436, + "grad_norm": 13.3125, + "learning_rate": 2.3414912695669424e-06, + "loss": 0.6954, + "step": 3887 + }, + { + "epoch": 0.3313165743502343, + "grad_norm": 12.9375, + "learning_rate": 2.3414093634532345e-06, + "loss": 0.619, + "step": 3888 + }, + { + "epoch": 0.3314017895185343, + "grad_norm": 19.875, + "learning_rate": 2.3413274376166623e-06, + "loss": 0.7765, + "step": 3889 + }, + { + "epoch": 0.33148700468683423, + "grad_norm": 16.25, + "learning_rate": 2.3412454920587057e-06, + "loss": 1.0166, + "step": 3890 + }, + { + "epoch": 0.3315722198551342, + "grad_norm": 15.0625, + "learning_rate": 2.341163526780846e-06, + "loss": 0.9749, + "step": 3891 + }, + { + "epoch": 0.33165743502343414, + "grad_norm": 13.6875, + "learning_rate": 2.341081541784563e-06, + "loss": 0.8066, + "step": 3892 + }, + { + "epoch": 0.33174265019173416, + "grad_norm": 15.8125, + "learning_rate": 2.3409995370713405e-06, + "loss": 0.5688, + "step": 3893 + }, + { + "epoch": 0.3318278653600341, + "grad_norm": 12.375, + "learning_rate": 2.340917512642659e-06, + "loss": 0.8135, + "step": 3894 + }, + { + "epoch": 0.33191308052833407, + "grad_norm": 13.5625, + "learning_rate": 2.340835468500001e-06, + "loss": 0.6521, + "step": 3895 + }, + { + "epoch": 0.331998295696634, + "grad_norm": 17.75, + "learning_rate": 2.340753404644849e-06, + "loss": 0.8416, + "step": 3896 + }, + { + "epoch": 0.332083510864934, + "grad_norm": 10.375, + "learning_rate": 2.3406713210786867e-06, + "loss": 0.3349, + "step": 3897 + }, + { + "epoch": 0.33216872603323394, + "grad_norm": 17.375, + "learning_rate": 2.3405892178029964e-06, + "loss": 1.0132, + "step": 3898 + }, + { + "epoch": 0.3322539412015339, + "grad_norm": 18.625, + "learning_rate": 2.3405070948192625e-06, + "loss": 1.0266, + "step": 3899 + }, + { + "epoch": 0.33233915636983385, + "grad_norm": 15.875, + "learning_rate": 2.340424952128969e-06, + "loss": 1.0015, + "step": 3900 + }, + { + "epoch": 0.3324243715381338, + "grad_norm": 16.25, + "learning_rate": 2.3403427897335995e-06, + "loss": 0.7218, + "step": 3901 + }, + { + "epoch": 0.33250958670643377, + "grad_norm": 16.75, + "learning_rate": 2.34026060763464e-06, + "loss": 0.7724, + "step": 3902 + }, + { + "epoch": 0.3325948018747337, + "grad_norm": 16.75, + "learning_rate": 2.3401784058335744e-06, + "loss": 0.7607, + "step": 3903 + }, + { + "epoch": 0.3326800170430337, + "grad_norm": 12.375, + "learning_rate": 2.340096184331889e-06, + "loss": 0.7528, + "step": 3904 + }, + { + "epoch": 0.33276523221133364, + "grad_norm": 21.0, + "learning_rate": 2.3400139431310694e-06, + "loss": 0.6997, + "step": 3905 + }, + { + "epoch": 0.3328504473796336, + "grad_norm": 11.25, + "learning_rate": 2.3399316822326018e-06, + "loss": 0.6647, + "step": 3906 + }, + { + "epoch": 0.33293566254793355, + "grad_norm": 13.3125, + "learning_rate": 2.3398494016379725e-06, + "loss": 0.4979, + "step": 3907 + }, + { + "epoch": 0.3330208777162335, + "grad_norm": 18.125, + "learning_rate": 2.339767101348669e-06, + "loss": 0.8077, + "step": 3908 + }, + { + "epoch": 0.33310609288453347, + "grad_norm": 19.5, + "learning_rate": 2.339684781366178e-06, + "loss": 0.9445, + "step": 3909 + }, + { + "epoch": 0.3331913080528334, + "grad_norm": 14.3125, + "learning_rate": 2.3396024416919864e-06, + "loss": 0.8323, + "step": 3910 + }, + { + "epoch": 0.3332765232211334, + "grad_norm": 17.625, + "learning_rate": 2.339520082327584e-06, + "loss": 0.9099, + "step": 3911 + }, + { + "epoch": 0.33336173838943334, + "grad_norm": 17.25, + "learning_rate": 2.339437703274458e-06, + "loss": 0.6971, + "step": 3912 + }, + { + "epoch": 0.3334469535577333, + "grad_norm": 16.25, + "learning_rate": 2.3393553045340967e-06, + "loss": 0.9556, + "step": 3913 + }, + { + "epoch": 0.33353216872603325, + "grad_norm": 12.0, + "learning_rate": 2.33927288610799e-06, + "loss": 0.5052, + "step": 3914 + }, + { + "epoch": 0.3336173838943332, + "grad_norm": 14.4375, + "learning_rate": 2.3391904479976267e-06, + "loss": 0.8101, + "step": 3915 + }, + { + "epoch": 0.33370259906263317, + "grad_norm": 13.5, + "learning_rate": 2.3391079902044968e-06, + "loss": 0.6602, + "step": 3916 + }, + { + "epoch": 0.3337878142309331, + "grad_norm": 17.125, + "learning_rate": 2.3390255127300905e-06, + "loss": 0.96, + "step": 3917 + }, + { + "epoch": 0.3338730293992331, + "grad_norm": 14.5, + "learning_rate": 2.338943015575898e-06, + "loss": 0.9262, + "step": 3918 + }, + { + "epoch": 0.33395824456753304, + "grad_norm": 14.4375, + "learning_rate": 2.3388604987434097e-06, + "loss": 0.6437, + "step": 3919 + }, + { + "epoch": 0.334043459735833, + "grad_norm": 13.0, + "learning_rate": 2.338777962234118e-06, + "loss": 0.8867, + "step": 3920 + }, + { + "epoch": 0.33412867490413295, + "grad_norm": 12.625, + "learning_rate": 2.338695406049513e-06, + "loss": 0.4617, + "step": 3921 + }, + { + "epoch": 0.3342138900724329, + "grad_norm": 12.9375, + "learning_rate": 2.338612830191088e-06, + "loss": 0.5451, + "step": 3922 + }, + { + "epoch": 0.33429910524073286, + "grad_norm": 13.625, + "learning_rate": 2.338530234660334e-06, + "loss": 0.2578, + "step": 3923 + }, + { + "epoch": 0.3343843204090328, + "grad_norm": 28.25, + "learning_rate": 2.338447619458744e-06, + "loss": 1.023, + "step": 3924 + }, + { + "epoch": 0.3344695355773328, + "grad_norm": 22.5, + "learning_rate": 2.3383649845878116e-06, + "loss": 0.9503, + "step": 3925 + }, + { + "epoch": 0.33455475074563273, + "grad_norm": 25.0, + "learning_rate": 2.3382823300490286e-06, + "loss": 1.301, + "step": 3926 + }, + { + "epoch": 0.3346399659139327, + "grad_norm": 13.9375, + "learning_rate": 2.33819965584389e-06, + "loss": 0.5245, + "step": 3927 + }, + { + "epoch": 0.33472518108223265, + "grad_norm": 24.375, + "learning_rate": 2.33811696197389e-06, + "loss": 0.4129, + "step": 3928 + }, + { + "epoch": 0.3348103962505326, + "grad_norm": 13.0625, + "learning_rate": 2.338034248440521e-06, + "loss": 0.4974, + "step": 3929 + }, + { + "epoch": 0.33489561141883256, + "grad_norm": 15.75, + "learning_rate": 2.33795151524528e-06, + "loss": 0.7743, + "step": 3930 + }, + { + "epoch": 0.3349808265871325, + "grad_norm": 21.25, + "learning_rate": 2.3378687623896606e-06, + "loss": 0.9439, + "step": 3931 + }, + { + "epoch": 0.3350660417554325, + "grad_norm": 11.75, + "learning_rate": 2.337785989875159e-06, + "loss": 0.4533, + "step": 3932 + }, + { + "epoch": 0.33515125692373243, + "grad_norm": 15.5, + "learning_rate": 2.3377031977032704e-06, + "loss": 0.994, + "step": 3933 + }, + { + "epoch": 0.3352364720920324, + "grad_norm": 12.25, + "learning_rate": 2.3376203858754916e-06, + "loss": 0.507, + "step": 3934 + }, + { + "epoch": 0.33532168726033235, + "grad_norm": 17.875, + "learning_rate": 2.3375375543933186e-06, + "loss": 0.4147, + "step": 3935 + }, + { + "epoch": 0.3354069024286323, + "grad_norm": 26.875, + "learning_rate": 2.3374547032582482e-06, + "loss": 1.3156, + "step": 3936 + }, + { + "epoch": 0.33549211759693226, + "grad_norm": 12.125, + "learning_rate": 2.337371832471778e-06, + "loss": 0.6413, + "step": 3937 + }, + { + "epoch": 0.3355773327652322, + "grad_norm": 13.1875, + "learning_rate": 2.3372889420354055e-06, + "loss": 0.6281, + "step": 3938 + }, + { + "epoch": 0.3356625479335322, + "grad_norm": 17.25, + "learning_rate": 2.3372060319506283e-06, + "loss": 0.8928, + "step": 3939 + }, + { + "epoch": 0.33574776310183213, + "grad_norm": 12.625, + "learning_rate": 2.3371231022189447e-06, + "loss": 0.7688, + "step": 3940 + }, + { + "epoch": 0.3358329782701321, + "grad_norm": 14.6875, + "learning_rate": 2.3370401528418535e-06, + "loss": 0.7262, + "step": 3941 + }, + { + "epoch": 0.33591819343843204, + "grad_norm": 17.125, + "learning_rate": 2.3369571838208537e-06, + "loss": 0.5853, + "step": 3942 + }, + { + "epoch": 0.336003408606732, + "grad_norm": 16.0, + "learning_rate": 2.336874195157445e-06, + "loss": 0.5609, + "step": 3943 + }, + { + "epoch": 0.33608862377503196, + "grad_norm": 10.4375, + "learning_rate": 2.336791186853126e-06, + "loss": 0.3731, + "step": 3944 + }, + { + "epoch": 0.3361738389433319, + "grad_norm": 12.75, + "learning_rate": 2.3367081589093976e-06, + "loss": 0.8264, + "step": 3945 + }, + { + "epoch": 0.33625905411163187, + "grad_norm": 10.8125, + "learning_rate": 2.3366251113277597e-06, + "loss": 0.4323, + "step": 3946 + }, + { + "epoch": 0.33634426927993183, + "grad_norm": 16.0, + "learning_rate": 2.336542044109714e-06, + "loss": 0.5537, + "step": 3947 + }, + { + "epoch": 0.3364294844482318, + "grad_norm": 20.125, + "learning_rate": 2.3364589572567608e-06, + "loss": 1.0855, + "step": 3948 + }, + { + "epoch": 0.33651469961653174, + "grad_norm": 11.875, + "learning_rate": 2.3363758507704015e-06, + "loss": 0.679, + "step": 3949 + }, + { + "epoch": 0.3365999147848317, + "grad_norm": 18.0, + "learning_rate": 2.336292724652138e-06, + "loss": 0.8954, + "step": 3950 + }, + { + "epoch": 0.33668512995313166, + "grad_norm": 15.0, + "learning_rate": 2.336209578903473e-06, + "loss": 0.9358, + "step": 3951 + }, + { + "epoch": 0.3367703451214316, + "grad_norm": 10.25, + "learning_rate": 2.3361264135259083e-06, + "loss": 0.3877, + "step": 3952 + }, + { + "epoch": 0.33685556028973157, + "grad_norm": 16.25, + "learning_rate": 2.336043228520947e-06, + "loss": 0.6617, + "step": 3953 + }, + { + "epoch": 0.3369407754580315, + "grad_norm": 17.75, + "learning_rate": 2.3359600238900932e-06, + "loss": 1.2946, + "step": 3954 + }, + { + "epoch": 0.3370259906263315, + "grad_norm": 12.5625, + "learning_rate": 2.3358767996348492e-06, + "loss": 0.2643, + "step": 3955 + }, + { + "epoch": 0.33711120579463144, + "grad_norm": 14.6875, + "learning_rate": 2.3357935557567195e-06, + "loss": 0.8753, + "step": 3956 + }, + { + "epoch": 0.3371964209629314, + "grad_norm": 20.875, + "learning_rate": 2.3357102922572083e-06, + "loss": 0.9863, + "step": 3957 + }, + { + "epoch": 0.33728163613123135, + "grad_norm": 11.3125, + "learning_rate": 2.335627009137821e-06, + "loss": 0.4275, + "step": 3958 + }, + { + "epoch": 0.3373668512995313, + "grad_norm": 15.125, + "learning_rate": 2.335543706400061e-06, + "loss": 0.7731, + "step": 3959 + }, + { + "epoch": 0.33745206646783127, + "grad_norm": 12.8125, + "learning_rate": 2.3354603840454355e-06, + "loss": 0.6382, + "step": 3960 + }, + { + "epoch": 0.3375372816361312, + "grad_norm": 20.25, + "learning_rate": 2.335377042075449e-06, + "loss": 1.0024, + "step": 3961 + }, + { + "epoch": 0.3376224968044312, + "grad_norm": 12.875, + "learning_rate": 2.3352936804916076e-06, + "loss": 0.6543, + "step": 3962 + }, + { + "epoch": 0.33770771197273114, + "grad_norm": 14.625, + "learning_rate": 2.3352102992954183e-06, + "loss": 0.9571, + "step": 3963 + }, + { + "epoch": 0.3377929271410311, + "grad_norm": 16.125, + "learning_rate": 2.335126898488388e-06, + "loss": 1.1574, + "step": 3964 + }, + { + "epoch": 0.33787814230933105, + "grad_norm": 17.875, + "learning_rate": 2.335043478072023e-06, + "loss": 0.7375, + "step": 3965 + }, + { + "epoch": 0.337963357477631, + "grad_norm": 12.6875, + "learning_rate": 2.3349600380478317e-06, + "loss": 0.486, + "step": 3966 + }, + { + "epoch": 0.33804857264593097, + "grad_norm": 11.4375, + "learning_rate": 2.334876578417321e-06, + "loss": 0.3549, + "step": 3967 + }, + { + "epoch": 0.3381337878142309, + "grad_norm": 17.5, + "learning_rate": 2.3347930991819996e-06, + "loss": 0.7313, + "step": 3968 + }, + { + "epoch": 0.3382190029825309, + "grad_norm": 16.125, + "learning_rate": 2.334709600343376e-06, + "loss": 0.911, + "step": 3969 + }, + { + "epoch": 0.33830421815083084, + "grad_norm": 16.75, + "learning_rate": 2.3346260819029597e-06, + "loss": 0.797, + "step": 3970 + }, + { + "epoch": 0.3383894333191308, + "grad_norm": 8.6875, + "learning_rate": 2.3345425438622593e-06, + "loss": 0.2527, + "step": 3971 + }, + { + "epoch": 0.33847464848743075, + "grad_norm": 13.0625, + "learning_rate": 2.3344589862227844e-06, + "loss": 0.5059, + "step": 3972 + }, + { + "epoch": 0.3385598636557307, + "grad_norm": 13.625, + "learning_rate": 2.334375408986045e-06, + "loss": 0.8461, + "step": 3973 + }, + { + "epoch": 0.33864507882403067, + "grad_norm": 25.5, + "learning_rate": 2.3342918121535516e-06, + "loss": 1.3825, + "step": 3974 + }, + { + "epoch": 0.3387302939923306, + "grad_norm": 15.4375, + "learning_rate": 2.3342081957268146e-06, + "loss": 0.9739, + "step": 3975 + }, + { + "epoch": 0.3388155091606306, + "grad_norm": 22.125, + "learning_rate": 2.334124559707346e-06, + "loss": 0.7812, + "step": 3976 + }, + { + "epoch": 0.33890072432893054, + "grad_norm": 14.1875, + "learning_rate": 2.3340409040966555e-06, + "loss": 0.8652, + "step": 3977 + }, + { + "epoch": 0.3389859394972305, + "grad_norm": 30.875, + "learning_rate": 2.3339572288962564e-06, + "loss": 1.031, + "step": 3978 + }, + { + "epoch": 0.33907115466553045, + "grad_norm": 12.375, + "learning_rate": 2.3338735341076597e-06, + "loss": 0.7422, + "step": 3979 + }, + { + "epoch": 0.3391563698338304, + "grad_norm": 18.75, + "learning_rate": 2.333789819732379e-06, + "loss": 0.711, + "step": 3980 + }, + { + "epoch": 0.33924158500213036, + "grad_norm": 24.25, + "learning_rate": 2.333706085771926e-06, + "loss": 1.0343, + "step": 3981 + }, + { + "epoch": 0.3393268001704303, + "grad_norm": 11.4375, + "learning_rate": 2.333622332227814e-06, + "loss": 0.6232, + "step": 3982 + }, + { + "epoch": 0.3394120153387303, + "grad_norm": 18.25, + "learning_rate": 2.333538559101557e-06, + "loss": 0.7608, + "step": 3983 + }, + { + "epoch": 0.33949723050703023, + "grad_norm": 20.0, + "learning_rate": 2.333454766394669e-06, + "loss": 0.8179, + "step": 3984 + }, + { + "epoch": 0.3395824456753302, + "grad_norm": 21.875, + "learning_rate": 2.3333709541086636e-06, + "loss": 1.131, + "step": 3985 + }, + { + "epoch": 0.33966766084363015, + "grad_norm": 18.25, + "learning_rate": 2.3332871222450554e-06, + "loss": 0.9852, + "step": 3986 + }, + { + "epoch": 0.3397528760119301, + "grad_norm": 19.75, + "learning_rate": 2.33320327080536e-06, + "loss": 0.8217, + "step": 3987 + }, + { + "epoch": 0.33983809118023006, + "grad_norm": 14.375, + "learning_rate": 2.333119399791092e-06, + "loss": 0.4423, + "step": 3988 + }, + { + "epoch": 0.33992330634853, + "grad_norm": 14.125, + "learning_rate": 2.3330355092037672e-06, + "loss": 0.6775, + "step": 3989 + }, + { + "epoch": 0.34000852151683, + "grad_norm": 18.25, + "learning_rate": 2.332951599044902e-06, + "loss": 0.6429, + "step": 3990 + }, + { + "epoch": 0.34009373668512993, + "grad_norm": 13.625, + "learning_rate": 2.3328676693160124e-06, + "loss": 0.7431, + "step": 3991 + }, + { + "epoch": 0.3401789518534299, + "grad_norm": 15.75, + "learning_rate": 2.332783720018615e-06, + "loss": 0.6359, + "step": 3992 + }, + { + "epoch": 0.34026416702172985, + "grad_norm": 15.3125, + "learning_rate": 2.3326997511542267e-06, + "loss": 1.0031, + "step": 3993 + }, + { + "epoch": 0.3403493821900298, + "grad_norm": 14.1875, + "learning_rate": 2.3326157627243653e-06, + "loss": 0.5978, + "step": 3994 + }, + { + "epoch": 0.34043459735832976, + "grad_norm": 12.9375, + "learning_rate": 2.332531754730549e-06, + "loss": 0.5315, + "step": 3995 + }, + { + "epoch": 0.3405198125266297, + "grad_norm": 13.875, + "learning_rate": 2.3324477271742943e-06, + "loss": 0.6428, + "step": 3996 + }, + { + "epoch": 0.3406050276949297, + "grad_norm": 17.125, + "learning_rate": 2.3323636800571212e-06, + "loss": 0.6487, + "step": 3997 + }, + { + "epoch": 0.34069024286322963, + "grad_norm": 12.8125, + "learning_rate": 2.332279613380548e-06, + "loss": 0.8349, + "step": 3998 + }, + { + "epoch": 0.3407754580315296, + "grad_norm": 15.8125, + "learning_rate": 2.332195527146094e-06, + "loss": 0.641, + "step": 3999 + }, + { + "epoch": 0.34086067319982954, + "grad_norm": 13.4375, + "learning_rate": 2.332111421355278e-06, + "loss": 1.024, + "step": 4000 + }, + { + "epoch": 0.3409458883681295, + "grad_norm": 19.625, + "learning_rate": 2.3320272960096212e-06, + "loss": 1.2106, + "step": 4001 + }, + { + "epoch": 0.34103110353642946, + "grad_norm": 16.875, + "learning_rate": 2.331943151110643e-06, + "loss": 1.1761, + "step": 4002 + }, + { + "epoch": 0.3411163187047294, + "grad_norm": 17.875, + "learning_rate": 2.331858986659864e-06, + "loss": 0.8235, + "step": 4003 + }, + { + "epoch": 0.3412015338730294, + "grad_norm": 11.9375, + "learning_rate": 2.331774802658805e-06, + "loss": 0.4329, + "step": 4004 + }, + { + "epoch": 0.34128674904132933, + "grad_norm": 13.625, + "learning_rate": 2.3316905991089877e-06, + "loss": 0.5224, + "step": 4005 + }, + { + "epoch": 0.34137196420962934, + "grad_norm": 21.0, + "learning_rate": 2.331606376011933e-06, + "loss": 1.1005, + "step": 4006 + }, + { + "epoch": 0.3414571793779293, + "grad_norm": 15.25, + "learning_rate": 2.331522133369164e-06, + "loss": 1.0837, + "step": 4007 + }, + { + "epoch": 0.34154239454622926, + "grad_norm": 15.125, + "learning_rate": 2.3314378711822024e-06, + "loss": 0.9544, + "step": 4008 + }, + { + "epoch": 0.3416276097145292, + "grad_norm": 16.0, + "learning_rate": 2.3313535894525713e-06, + "loss": 0.815, + "step": 4009 + }, + { + "epoch": 0.34171282488282917, + "grad_norm": 13.0625, + "learning_rate": 2.331269288181793e-06, + "loss": 0.4748, + "step": 4010 + }, + { + "epoch": 0.3417980400511291, + "grad_norm": 18.75, + "learning_rate": 2.3311849673713913e-06, + "loss": 1.3333, + "step": 4011 + }, + { + "epoch": 0.3418832552194291, + "grad_norm": 14.0, + "learning_rate": 2.3311006270228905e-06, + "loss": 0.4996, + "step": 4012 + }, + { + "epoch": 0.34196847038772904, + "grad_norm": 12.4375, + "learning_rate": 2.331016267137814e-06, + "loss": 0.5874, + "step": 4013 + }, + { + "epoch": 0.342053685556029, + "grad_norm": 16.125, + "learning_rate": 2.3309318877176865e-06, + "loss": 0.9413, + "step": 4014 + }, + { + "epoch": 0.34213890072432895, + "grad_norm": 25.375, + "learning_rate": 2.3308474887640327e-06, + "loss": 1.0336, + "step": 4015 + }, + { + "epoch": 0.3422241158926289, + "grad_norm": 12.1875, + "learning_rate": 2.330763070278378e-06, + "loss": 0.5361, + "step": 4016 + }, + { + "epoch": 0.34230933106092887, + "grad_norm": 17.375, + "learning_rate": 2.330678632262248e-06, + "loss": 0.3863, + "step": 4017 + }, + { + "epoch": 0.3423945462292288, + "grad_norm": 12.0625, + "learning_rate": 2.3305941747171684e-06, + "loss": 0.7383, + "step": 4018 + }, + { + "epoch": 0.3424797613975288, + "grad_norm": 18.0, + "learning_rate": 2.3305096976446652e-06, + "loss": 0.9696, + "step": 4019 + }, + { + "epoch": 0.34256497656582874, + "grad_norm": 12.5, + "learning_rate": 2.3304252010462654e-06, + "loss": 0.5396, + "step": 4020 + }, + { + "epoch": 0.3426501917341287, + "grad_norm": 14.3125, + "learning_rate": 2.3303406849234956e-06, + "loss": 0.6174, + "step": 4021 + }, + { + "epoch": 0.34273540690242865, + "grad_norm": 20.875, + "learning_rate": 2.330256149277883e-06, + "loss": 1.2139, + "step": 4022 + }, + { + "epoch": 0.3428206220707286, + "grad_norm": 20.25, + "learning_rate": 2.330171594110956e-06, + "loss": 0.904, + "step": 4023 + }, + { + "epoch": 0.34290583723902857, + "grad_norm": 33.25, + "learning_rate": 2.330087019424242e-06, + "loss": 0.9877, + "step": 4024 + }, + { + "epoch": 0.3429910524073285, + "grad_norm": 14.625, + "learning_rate": 2.3300024252192692e-06, + "loss": 0.6767, + "step": 4025 + }, + { + "epoch": 0.3430762675756285, + "grad_norm": 14.4375, + "learning_rate": 2.329917811497567e-06, + "loss": 0.6389, + "step": 4026 + }, + { + "epoch": 0.34316148274392844, + "grad_norm": 18.375, + "learning_rate": 2.3298331782606638e-06, + "loss": 0.7641, + "step": 4027 + }, + { + "epoch": 0.3432466979122284, + "grad_norm": 17.0, + "learning_rate": 2.3297485255100892e-06, + "loss": 0.6028, + "step": 4028 + }, + { + "epoch": 0.34333191308052835, + "grad_norm": 14.3125, + "learning_rate": 2.329663853247373e-06, + "loss": 0.4676, + "step": 4029 + }, + { + "epoch": 0.3434171282488283, + "grad_norm": 24.25, + "learning_rate": 2.3295791614740456e-06, + "loss": 0.8223, + "step": 4030 + }, + { + "epoch": 0.34350234341712826, + "grad_norm": 11.75, + "learning_rate": 2.3294944501916364e-06, + "loss": 0.4249, + "step": 4031 + }, + { + "epoch": 0.3435875585854282, + "grad_norm": 14.5, + "learning_rate": 2.3294097194016773e-06, + "loss": 0.5445, + "step": 4032 + }, + { + "epoch": 0.3436727737537282, + "grad_norm": 22.0, + "learning_rate": 2.3293249691056993e-06, + "loss": 1.0898, + "step": 4033 + }, + { + "epoch": 0.34375798892202813, + "grad_norm": 23.625, + "learning_rate": 2.3292401993052337e-06, + "loss": 0.9155, + "step": 4034 + }, + { + "epoch": 0.3438432040903281, + "grad_norm": 13.5, + "learning_rate": 2.329155410001812e-06, + "loss": 0.6164, + "step": 4035 + }, + { + "epoch": 0.34392841925862805, + "grad_norm": 12.1875, + "learning_rate": 2.3290706011969676e-06, + "loss": 0.6982, + "step": 4036 + }, + { + "epoch": 0.344013634426928, + "grad_norm": 14.8125, + "learning_rate": 2.328985772892232e-06, + "loss": 0.5167, + "step": 4037 + }, + { + "epoch": 0.34409884959522796, + "grad_norm": 14.5, + "learning_rate": 2.3289009250891385e-06, + "loss": 1.0156, + "step": 4038 + }, + { + "epoch": 0.3441840647635279, + "grad_norm": 16.0, + "learning_rate": 2.32881605778922e-06, + "loss": 0.8291, + "step": 4039 + }, + { + "epoch": 0.3442692799318279, + "grad_norm": 13.8125, + "learning_rate": 2.328731170994011e-06, + "loss": 0.749, + "step": 4040 + }, + { + "epoch": 0.34435449510012783, + "grad_norm": 13.6875, + "learning_rate": 2.328646264705045e-06, + "loss": 0.8931, + "step": 4041 + }, + { + "epoch": 0.3444397102684278, + "grad_norm": 13.0, + "learning_rate": 2.328561338923856e-06, + "loss": 0.7366, + "step": 4042 + }, + { + "epoch": 0.34452492543672775, + "grad_norm": 26.25, + "learning_rate": 2.3284763936519795e-06, + "loss": 0.9489, + "step": 4043 + }, + { + "epoch": 0.3446101406050277, + "grad_norm": 11.3125, + "learning_rate": 2.3283914288909498e-06, + "loss": 0.4795, + "step": 4044 + }, + { + "epoch": 0.34469535577332766, + "grad_norm": 14.25, + "learning_rate": 2.3283064446423024e-06, + "loss": 0.7264, + "step": 4045 + }, + { + "epoch": 0.3447805709416276, + "grad_norm": 14.8125, + "learning_rate": 2.328221440907573e-06, + "loss": 0.7318, + "step": 4046 + }, + { + "epoch": 0.3448657861099276, + "grad_norm": 16.375, + "learning_rate": 2.3281364176882986e-06, + "loss": 0.992, + "step": 4047 + }, + { + "epoch": 0.34495100127822753, + "grad_norm": 24.25, + "learning_rate": 2.3280513749860147e-06, + "loss": 0.7895, + "step": 4048 + }, + { + "epoch": 0.3450362164465275, + "grad_norm": 13.75, + "learning_rate": 2.3279663128022585e-06, + "loss": 0.7572, + "step": 4049 + }, + { + "epoch": 0.34512143161482745, + "grad_norm": 17.375, + "learning_rate": 2.327881231138567e-06, + "loss": 1.0086, + "step": 4050 + }, + { + "epoch": 0.3452066467831274, + "grad_norm": 14.9375, + "learning_rate": 2.3277961299964773e-06, + "loss": 1.0919, + "step": 4051 + }, + { + "epoch": 0.34529186195142736, + "grad_norm": 14.1875, + "learning_rate": 2.3277110093775283e-06, + "loss": 0.9007, + "step": 4052 + }, + { + "epoch": 0.3453770771197273, + "grad_norm": 13.4375, + "learning_rate": 2.3276258692832577e-06, + "loss": 0.8951, + "step": 4053 + }, + { + "epoch": 0.3454622922880273, + "grad_norm": 25.75, + "learning_rate": 2.3275407097152036e-06, + "loss": 1.2109, + "step": 4054 + }, + { + "epoch": 0.34554750745632723, + "grad_norm": 17.5, + "learning_rate": 2.3274555306749053e-06, + "loss": 0.9206, + "step": 4055 + }, + { + "epoch": 0.3456327226246272, + "grad_norm": 14.4375, + "learning_rate": 2.3273703321639025e-06, + "loss": 0.7756, + "step": 4056 + }, + { + "epoch": 0.34571793779292714, + "grad_norm": 16.125, + "learning_rate": 2.3272851141837343e-06, + "loss": 0.8646, + "step": 4057 + }, + { + "epoch": 0.3458031529612271, + "grad_norm": 16.0, + "learning_rate": 2.3271998767359407e-06, + "loss": 1.0957, + "step": 4058 + }, + { + "epoch": 0.34588836812952706, + "grad_norm": 15.0625, + "learning_rate": 2.327114619822062e-06, + "loss": 0.5483, + "step": 4059 + }, + { + "epoch": 0.345973583297827, + "grad_norm": 16.125, + "learning_rate": 2.3270293434436387e-06, + "loss": 1.0238, + "step": 4060 + }, + { + "epoch": 0.34605879846612697, + "grad_norm": 11.0625, + "learning_rate": 2.3269440476022124e-06, + "loss": 0.4951, + "step": 4061 + }, + { + "epoch": 0.34614401363442693, + "grad_norm": 11.875, + "learning_rate": 2.326858732299325e-06, + "loss": 0.5912, + "step": 4062 + }, + { + "epoch": 0.3462292288027269, + "grad_norm": 9.5625, + "learning_rate": 2.3267733975365164e-06, + "loss": 0.3455, + "step": 4063 + }, + { + "epoch": 0.34631444397102684, + "grad_norm": 17.375, + "learning_rate": 2.32668804331533e-06, + "loss": 0.8845, + "step": 4064 + }, + { + "epoch": 0.3463996591393268, + "grad_norm": 11.625, + "learning_rate": 2.3266026696373086e-06, + "loss": 0.9656, + "step": 4065 + }, + { + "epoch": 0.34648487430762676, + "grad_norm": 15.1875, + "learning_rate": 2.3265172765039936e-06, + "loss": 0.8902, + "step": 4066 + }, + { + "epoch": 0.3465700894759267, + "grad_norm": 12.3125, + "learning_rate": 2.326431863916929e-06, + "loss": 0.5413, + "step": 4067 + }, + { + "epoch": 0.34665530464422667, + "grad_norm": 9.9375, + "learning_rate": 2.3263464318776583e-06, + "loss": 0.3425, + "step": 4068 + }, + { + "epoch": 0.3467405198125266, + "grad_norm": 15.375, + "learning_rate": 2.326260980387725e-06, + "loss": 1.1196, + "step": 4069 + }, + { + "epoch": 0.3468257349808266, + "grad_norm": 17.125, + "learning_rate": 2.326175509448674e-06, + "loss": 1.0365, + "step": 4070 + }, + { + "epoch": 0.34691095014912654, + "grad_norm": 18.25, + "learning_rate": 2.326090019062049e-06, + "loss": 0.4944, + "step": 4071 + }, + { + "epoch": 0.3469961653174265, + "grad_norm": 16.0, + "learning_rate": 2.326004509229395e-06, + "loss": 0.9812, + "step": 4072 + }, + { + "epoch": 0.34708138048572645, + "grad_norm": 14.8125, + "learning_rate": 2.325918979952258e-06, + "loss": 0.7273, + "step": 4073 + }, + { + "epoch": 0.3471665956540264, + "grad_norm": 12.75, + "learning_rate": 2.3258334312321832e-06, + "loss": 0.5361, + "step": 4074 + }, + { + "epoch": 0.34725181082232637, + "grad_norm": 13.4375, + "learning_rate": 2.3257478630707164e-06, + "loss": 0.6556, + "step": 4075 + }, + { + "epoch": 0.3473370259906263, + "grad_norm": 19.25, + "learning_rate": 2.3256622754694037e-06, + "loss": 0.6668, + "step": 4076 + }, + { + "epoch": 0.3474222411589263, + "grad_norm": 20.75, + "learning_rate": 2.3255766684297924e-06, + "loss": 0.7155, + "step": 4077 + }, + { + "epoch": 0.34750745632722624, + "grad_norm": 11.1875, + "learning_rate": 2.3254910419534288e-06, + "loss": 0.4792, + "step": 4078 + }, + { + "epoch": 0.3475926714955262, + "grad_norm": 12.9375, + "learning_rate": 2.3254053960418603e-06, + "loss": 0.577, + "step": 4079 + }, + { + "epoch": 0.34767788666382615, + "grad_norm": 15.5625, + "learning_rate": 2.3253197306966357e-06, + "loss": 0.7516, + "step": 4080 + }, + { + "epoch": 0.3477631018321261, + "grad_norm": 21.875, + "learning_rate": 2.3252340459193017e-06, + "loss": 0.8775, + "step": 4081 + }, + { + "epoch": 0.34784831700042607, + "grad_norm": 17.875, + "learning_rate": 2.325148341711407e-06, + "loss": 0.8316, + "step": 4082 + }, + { + "epoch": 0.347933532168726, + "grad_norm": 10.4375, + "learning_rate": 2.3250626180745013e-06, + "loss": 0.3671, + "step": 4083 + }, + { + "epoch": 0.348018747337026, + "grad_norm": 11.1875, + "learning_rate": 2.3249768750101325e-06, + "loss": 0.585, + "step": 4084 + }, + { + "epoch": 0.34810396250532594, + "grad_norm": 20.0, + "learning_rate": 2.3248911125198507e-06, + "loss": 1.0553, + "step": 4085 + }, + { + "epoch": 0.3481891776736259, + "grad_norm": 12.625, + "learning_rate": 2.324805330605206e-06, + "loss": 0.6835, + "step": 4086 + }, + { + "epoch": 0.34827439284192585, + "grad_norm": 13.5625, + "learning_rate": 2.3247195292677475e-06, + "loss": 0.5109, + "step": 4087 + }, + { + "epoch": 0.3483596080102258, + "grad_norm": 16.0, + "learning_rate": 2.324633708509027e-06, + "loss": 0.9071, + "step": 4088 + }, + { + "epoch": 0.34844482317852576, + "grad_norm": 19.25, + "learning_rate": 2.3245478683305948e-06, + "loss": 0.853, + "step": 4089 + }, + { + "epoch": 0.3485300383468257, + "grad_norm": 13.625, + "learning_rate": 2.3244620087340017e-06, + "loss": 0.844, + "step": 4090 + }, + { + "epoch": 0.3486152535151257, + "grad_norm": 17.5, + "learning_rate": 2.3243761297207995e-06, + "loss": 0.8373, + "step": 4091 + }, + { + "epoch": 0.34870046868342564, + "grad_norm": 16.0, + "learning_rate": 2.3242902312925408e-06, + "loss": 0.7101, + "step": 4092 + }, + { + "epoch": 0.3487856838517256, + "grad_norm": 20.625, + "learning_rate": 2.324204313450777e-06, + "loss": 0.6203, + "step": 4093 + }, + { + "epoch": 0.34887089902002555, + "grad_norm": 11.5625, + "learning_rate": 2.324118376197061e-06, + "loss": 0.6374, + "step": 4094 + }, + { + "epoch": 0.3489561141883255, + "grad_norm": 14.1875, + "learning_rate": 2.324032419532946e-06, + "loss": 0.5917, + "step": 4095 + }, + { + "epoch": 0.34904132935662546, + "grad_norm": 13.375, + "learning_rate": 2.323946443459985e-06, + "loss": 0.7285, + "step": 4096 + }, + { + "epoch": 0.3491265445249254, + "grad_norm": 14.3125, + "learning_rate": 2.323860447979732e-06, + "loss": 0.7153, + "step": 4097 + }, + { + "epoch": 0.3492117596932254, + "grad_norm": 14.0625, + "learning_rate": 2.3237744330937407e-06, + "loss": 0.5982, + "step": 4098 + }, + { + "epoch": 0.34929697486152533, + "grad_norm": 9.5, + "learning_rate": 2.323688398803566e-06, + "loss": 0.5317, + "step": 4099 + }, + { + "epoch": 0.3493821900298253, + "grad_norm": 14.125, + "learning_rate": 2.323602345110762e-06, + "loss": 0.6596, + "step": 4100 + }, + { + "epoch": 0.34946740519812525, + "grad_norm": 14.0, + "learning_rate": 2.323516272016884e-06, + "loss": 0.6538, + "step": 4101 + }, + { + "epoch": 0.3495526203664252, + "grad_norm": 21.125, + "learning_rate": 2.3234301795234875e-06, + "loss": 0.5747, + "step": 4102 + }, + { + "epoch": 0.34963783553472516, + "grad_norm": 13.375, + "learning_rate": 2.3233440676321285e-06, + "loss": 0.7308, + "step": 4103 + }, + { + "epoch": 0.3497230507030251, + "grad_norm": 13.125, + "learning_rate": 2.3232579363443627e-06, + "loss": 0.6837, + "step": 4104 + }, + { + "epoch": 0.3498082658713251, + "grad_norm": 15.875, + "learning_rate": 2.3231717856617463e-06, + "loss": 0.4639, + "step": 4105 + }, + { + "epoch": 0.34989348103962503, + "grad_norm": 17.625, + "learning_rate": 2.3230856155858366e-06, + "loss": 0.5849, + "step": 4106 + }, + { + "epoch": 0.349978696207925, + "grad_norm": 11.9375, + "learning_rate": 2.322999426118191e-06, + "loss": 0.5821, + "step": 4107 + }, + { + "epoch": 0.35006391137622495, + "grad_norm": 10.1875, + "learning_rate": 2.3229132172603667e-06, + "loss": 0.3634, + "step": 4108 + }, + { + "epoch": 0.3501491265445249, + "grad_norm": 14.875, + "learning_rate": 2.3228269890139222e-06, + "loss": 0.8087, + "step": 4109 + }, + { + "epoch": 0.35023434171282486, + "grad_norm": 16.75, + "learning_rate": 2.322740741380415e-06, + "loss": 0.8326, + "step": 4110 + }, + { + "epoch": 0.3503195568811248, + "grad_norm": 18.25, + "learning_rate": 2.3226544743614036e-06, + "loss": 0.8339, + "step": 4111 + }, + { + "epoch": 0.3504047720494248, + "grad_norm": 11.4375, + "learning_rate": 2.3225681879584473e-06, + "loss": 0.6817, + "step": 4112 + }, + { + "epoch": 0.35048998721772473, + "grad_norm": 18.125, + "learning_rate": 2.3224818821731055e-06, + "loss": 1.0308, + "step": 4113 + }, + { + "epoch": 0.3505752023860247, + "grad_norm": 14.375, + "learning_rate": 2.3223955570069375e-06, + "loss": 0.8212, + "step": 4114 + }, + { + "epoch": 0.35066041755432464, + "grad_norm": 14.0625, + "learning_rate": 2.322309212461503e-06, + "loss": 0.8081, + "step": 4115 + }, + { + "epoch": 0.3507456327226246, + "grad_norm": 15.375, + "learning_rate": 2.322222848538363e-06, + "loss": 0.7543, + "step": 4116 + }, + { + "epoch": 0.35083084789092456, + "grad_norm": 14.4375, + "learning_rate": 2.3221364652390782e-06, + "loss": 0.7831, + "step": 4117 + }, + { + "epoch": 0.3509160630592245, + "grad_norm": 11.6875, + "learning_rate": 2.322050062565209e-06, + "loss": 0.5824, + "step": 4118 + }, + { + "epoch": 0.3510012782275245, + "grad_norm": 16.75, + "learning_rate": 2.321963640518317e-06, + "loss": 1.0646, + "step": 4119 + }, + { + "epoch": 0.3510864933958245, + "grad_norm": 16.25, + "learning_rate": 2.3218771990999644e-06, + "loss": 0.977, + "step": 4120 + }, + { + "epoch": 0.35117170856412444, + "grad_norm": 13.25, + "learning_rate": 2.3217907383117132e-06, + "loss": 0.4908, + "step": 4121 + }, + { + "epoch": 0.3512569237324244, + "grad_norm": 11.5625, + "learning_rate": 2.3217042581551253e-06, + "loss": 0.4513, + "step": 4122 + }, + { + "epoch": 0.35134213890072435, + "grad_norm": 15.9375, + "learning_rate": 2.3216177586317635e-06, + "loss": 0.7099, + "step": 4123 + }, + { + "epoch": 0.3514273540690243, + "grad_norm": 15.25, + "learning_rate": 2.3215312397431914e-06, + "loss": 0.7217, + "step": 4124 + }, + { + "epoch": 0.35151256923732427, + "grad_norm": 20.25, + "learning_rate": 2.3214447014909727e-06, + "loss": 0.7117, + "step": 4125 + }, + { + "epoch": 0.3515977844056242, + "grad_norm": 10.1875, + "learning_rate": 2.3213581438766704e-06, + "loss": 0.5743, + "step": 4126 + }, + { + "epoch": 0.3516829995739242, + "grad_norm": 15.4375, + "learning_rate": 2.321271566901849e-06, + "loss": 0.8363, + "step": 4127 + }, + { + "epoch": 0.35176821474222414, + "grad_norm": 15.4375, + "learning_rate": 2.3211849705680735e-06, + "loss": 0.7295, + "step": 4128 + }, + { + "epoch": 0.3518534299105241, + "grad_norm": 14.125, + "learning_rate": 2.3210983548769085e-06, + "loss": 0.539, + "step": 4129 + }, + { + "epoch": 0.35193864507882405, + "grad_norm": 16.125, + "learning_rate": 2.321011719829919e-06, + "loss": 0.8455, + "step": 4130 + }, + { + "epoch": 0.352023860247124, + "grad_norm": 13.0625, + "learning_rate": 2.3209250654286705e-06, + "loss": 0.8935, + "step": 4131 + }, + { + "epoch": 0.35210907541542397, + "grad_norm": 17.375, + "learning_rate": 2.3208383916747294e-06, + "loss": 0.7844, + "step": 4132 + }, + { + "epoch": 0.3521942905837239, + "grad_norm": 14.4375, + "learning_rate": 2.3207516985696616e-06, + "loss": 0.6487, + "step": 4133 + }, + { + "epoch": 0.3522795057520239, + "grad_norm": 11.375, + "learning_rate": 2.320664986115034e-06, + "loss": 0.8521, + "step": 4134 + }, + { + "epoch": 0.35236472092032384, + "grad_norm": 13.5, + "learning_rate": 2.3205782543124132e-06, + "loss": 0.641, + "step": 4135 + }, + { + "epoch": 0.3524499360886238, + "grad_norm": 13.125, + "learning_rate": 2.3204915031633675e-06, + "loss": 0.4281, + "step": 4136 + }, + { + "epoch": 0.35253515125692375, + "grad_norm": 16.0, + "learning_rate": 2.320404732669463e-06, + "loss": 1.0232, + "step": 4137 + }, + { + "epoch": 0.3526203664252237, + "grad_norm": 14.5, + "learning_rate": 2.3203179428322693e-06, + "loss": 0.5465, + "step": 4138 + }, + { + "epoch": 0.35270558159352366, + "grad_norm": 13.9375, + "learning_rate": 2.3202311336533538e-06, + "loss": 0.6673, + "step": 4139 + }, + { + "epoch": 0.3527907967618236, + "grad_norm": 15.4375, + "learning_rate": 2.3201443051342853e-06, + "loss": 0.8589, + "step": 4140 + }, + { + "epoch": 0.3528760119301236, + "grad_norm": 13.5625, + "learning_rate": 2.320057457276633e-06, + "loss": 0.7259, + "step": 4141 + }, + { + "epoch": 0.35296122709842354, + "grad_norm": 18.25, + "learning_rate": 2.319970590081967e-06, + "loss": 0.9552, + "step": 4142 + }, + { + "epoch": 0.3530464422667235, + "grad_norm": 12.1875, + "learning_rate": 2.3198837035518562e-06, + "loss": 0.4087, + "step": 4143 + }, + { + "epoch": 0.35313165743502345, + "grad_norm": 12.8125, + "learning_rate": 2.319796797687871e-06, + "loss": 0.6732, + "step": 4144 + }, + { + "epoch": 0.3532168726033234, + "grad_norm": 20.875, + "learning_rate": 2.319709872491582e-06, + "loss": 0.8255, + "step": 4145 + }, + { + "epoch": 0.35330208777162336, + "grad_norm": 11.1875, + "learning_rate": 2.31962292796456e-06, + "loss": 0.5269, + "step": 4146 + }, + { + "epoch": 0.3533873029399233, + "grad_norm": 14.6875, + "learning_rate": 2.3195359641083757e-06, + "loss": 0.5091, + "step": 4147 + }, + { + "epoch": 0.3534725181082233, + "grad_norm": 16.25, + "learning_rate": 2.3194489809246017e-06, + "loss": 0.7482, + "step": 4148 + }, + { + "epoch": 0.35355773327652323, + "grad_norm": 17.25, + "learning_rate": 2.3193619784148087e-06, + "loss": 0.908, + "step": 4149 + }, + { + "epoch": 0.3536429484448232, + "grad_norm": 12.5625, + "learning_rate": 2.3192749565805693e-06, + "loss": 0.4892, + "step": 4150 + }, + { + "epoch": 0.35372816361312315, + "grad_norm": 11.875, + "learning_rate": 2.3191879154234564e-06, + "loss": 0.6026, + "step": 4151 + }, + { + "epoch": 0.3538133787814231, + "grad_norm": 16.125, + "learning_rate": 2.3191008549450426e-06, + "loss": 0.6892, + "step": 4152 + }, + { + "epoch": 0.35389859394972306, + "grad_norm": 10.75, + "learning_rate": 2.3190137751469016e-06, + "loss": 0.4586, + "step": 4153 + }, + { + "epoch": 0.353983809118023, + "grad_norm": 14.25, + "learning_rate": 2.318926676030607e-06, + "loss": 0.8373, + "step": 4154 + }, + { + "epoch": 0.354069024286323, + "grad_norm": 23.75, + "learning_rate": 2.3188395575977316e-06, + "loss": 0.9728, + "step": 4155 + }, + { + "epoch": 0.35415423945462293, + "grad_norm": 16.25, + "learning_rate": 2.318752419849851e-06, + "loss": 0.582, + "step": 4156 + }, + { + "epoch": 0.3542394546229229, + "grad_norm": 10.875, + "learning_rate": 2.3186652627885393e-06, + "loss": 0.5042, + "step": 4157 + }, + { + "epoch": 0.35432466979122285, + "grad_norm": 17.625, + "learning_rate": 2.318578086415372e-06, + "loss": 1.0334, + "step": 4158 + }, + { + "epoch": 0.3544098849595228, + "grad_norm": 15.5, + "learning_rate": 2.318490890731924e-06, + "loss": 0.5908, + "step": 4159 + }, + { + "epoch": 0.35449510012782276, + "grad_norm": 18.5, + "learning_rate": 2.318403675739771e-06, + "loss": 0.9349, + "step": 4160 + }, + { + "epoch": 0.3545803152961227, + "grad_norm": 15.5625, + "learning_rate": 2.3183164414404895e-06, + "loss": 0.6485, + "step": 4161 + }, + { + "epoch": 0.3546655304644227, + "grad_norm": 16.0, + "learning_rate": 2.3182291878356554e-06, + "loss": 0.7995, + "step": 4162 + }, + { + "epoch": 0.35475074563272263, + "grad_norm": 14.5625, + "learning_rate": 2.318141914926846e-06, + "loss": 0.8688, + "step": 4163 + }, + { + "epoch": 0.3548359608010226, + "grad_norm": 13.0625, + "learning_rate": 2.3180546227156376e-06, + "loss": 0.7732, + "step": 4164 + }, + { + "epoch": 0.35492117596932254, + "grad_norm": 12.4375, + "learning_rate": 2.3179673112036083e-06, + "loss": 0.669, + "step": 4165 + }, + { + "epoch": 0.3550063911376225, + "grad_norm": 16.125, + "learning_rate": 2.317879980392336e-06, + "loss": 0.5297, + "step": 4166 + }, + { + "epoch": 0.35509160630592246, + "grad_norm": 16.25, + "learning_rate": 2.317792630283399e-06, + "loss": 0.7031, + "step": 4167 + }, + { + "epoch": 0.3551768214742224, + "grad_norm": 20.625, + "learning_rate": 2.3177052608783747e-06, + "loss": 0.9453, + "step": 4168 + }, + { + "epoch": 0.35526203664252237, + "grad_norm": 27.375, + "learning_rate": 2.3176178721788426e-06, + "loss": 1.2809, + "step": 4169 + }, + { + "epoch": 0.35534725181082233, + "grad_norm": 10.625, + "learning_rate": 2.317530464186382e-06, + "loss": 0.357, + "step": 4170 + }, + { + "epoch": 0.3554324669791223, + "grad_norm": 13.25, + "learning_rate": 2.317443036902573e-06, + "loss": 0.7564, + "step": 4171 + }, + { + "epoch": 0.35551768214742224, + "grad_norm": 15.375, + "learning_rate": 2.3173555903289945e-06, + "loss": 0.6833, + "step": 4172 + }, + { + "epoch": 0.3556028973157222, + "grad_norm": 14.4375, + "learning_rate": 2.3172681244672274e-06, + "loss": 0.667, + "step": 4173 + }, + { + "epoch": 0.35568811248402216, + "grad_norm": 14.625, + "learning_rate": 2.317180639318852e-06, + "loss": 0.8292, + "step": 4174 + }, + { + "epoch": 0.3557733276523221, + "grad_norm": 18.0, + "learning_rate": 2.3170931348854494e-06, + "loss": 0.966, + "step": 4175 + }, + { + "epoch": 0.35585854282062207, + "grad_norm": 21.125, + "learning_rate": 2.3170056111686008e-06, + "loss": 1.1164, + "step": 4176 + }, + { + "epoch": 0.355943757988922, + "grad_norm": 12.0, + "learning_rate": 2.3169180681698874e-06, + "loss": 0.5023, + "step": 4177 + }, + { + "epoch": 0.356028973157222, + "grad_norm": 14.875, + "learning_rate": 2.316830505890892e-06, + "loss": 0.6285, + "step": 4178 + }, + { + "epoch": 0.35611418832552194, + "grad_norm": 14.1875, + "learning_rate": 2.3167429243331966e-06, + "loss": 0.8044, + "step": 4179 + }, + { + "epoch": 0.3561994034938219, + "grad_norm": 12.375, + "learning_rate": 2.3166553234983836e-06, + "loss": 0.6846, + "step": 4180 + }, + { + "epoch": 0.35628461866212185, + "grad_norm": 12.1875, + "learning_rate": 2.3165677033880366e-06, + "loss": 0.5404, + "step": 4181 + }, + { + "epoch": 0.3563698338304218, + "grad_norm": 17.5, + "learning_rate": 2.3164800640037387e-06, + "loss": 0.776, + "step": 4182 + }, + { + "epoch": 0.35645504899872177, + "grad_norm": 11.5625, + "learning_rate": 2.3163924053470737e-06, + "loss": 0.5105, + "step": 4183 + }, + { + "epoch": 0.3565402641670217, + "grad_norm": 13.6875, + "learning_rate": 2.3163047274196253e-06, + "loss": 0.7045, + "step": 4184 + }, + { + "epoch": 0.3566254793353217, + "grad_norm": 17.25, + "learning_rate": 2.316217030222978e-06, + "loss": 0.4571, + "step": 4185 + }, + { + "epoch": 0.35671069450362164, + "grad_norm": 11.75, + "learning_rate": 2.3161293137587176e-06, + "loss": 0.5128, + "step": 4186 + }, + { + "epoch": 0.3567959096719216, + "grad_norm": 14.0625, + "learning_rate": 2.316041578028428e-06, + "loss": 0.8485, + "step": 4187 + }, + { + "epoch": 0.35688112484022155, + "grad_norm": 12.5625, + "learning_rate": 2.315953823033695e-06, + "loss": 0.5175, + "step": 4188 + }, + { + "epoch": 0.3569663400085215, + "grad_norm": 17.375, + "learning_rate": 2.3158660487761043e-06, + "loss": 0.9817, + "step": 4189 + }, + { + "epoch": 0.35705155517682147, + "grad_norm": 23.875, + "learning_rate": 2.3157782552572426e-06, + "loss": 1.1117, + "step": 4190 + }, + { + "epoch": 0.3571367703451214, + "grad_norm": 16.75, + "learning_rate": 2.315690442478696e-06, + "loss": 0.8224, + "step": 4191 + }, + { + "epoch": 0.3572219855134214, + "grad_norm": 14.4375, + "learning_rate": 2.3156026104420515e-06, + "loss": 1.0135, + "step": 4192 + }, + { + "epoch": 0.35730720068172134, + "grad_norm": 20.875, + "learning_rate": 2.3155147591488965e-06, + "loss": 1.2968, + "step": 4193 + }, + { + "epoch": 0.3573924158500213, + "grad_norm": 16.875, + "learning_rate": 2.315426888600818e-06, + "loss": 0.8112, + "step": 4194 + }, + { + "epoch": 0.35747763101832125, + "grad_norm": 14.6875, + "learning_rate": 2.3153389987994046e-06, + "loss": 0.6947, + "step": 4195 + }, + { + "epoch": 0.3575628461866212, + "grad_norm": 11.0, + "learning_rate": 2.3152510897462437e-06, + "loss": 0.5024, + "step": 4196 + }, + { + "epoch": 0.35764806135492117, + "grad_norm": 12.875, + "learning_rate": 2.315163161442925e-06, + "loss": 0.599, + "step": 4197 + }, + { + "epoch": 0.3577332765232211, + "grad_norm": 14.3125, + "learning_rate": 2.3150752138910366e-06, + "loss": 0.8405, + "step": 4198 + }, + { + "epoch": 0.3578184916915211, + "grad_norm": 14.25, + "learning_rate": 2.314987247092168e-06, + "loss": 0.8189, + "step": 4199 + }, + { + "epoch": 0.35790370685982104, + "grad_norm": 14.0625, + "learning_rate": 2.314899261047909e-06, + "loss": 0.6091, + "step": 4200 + }, + { + "epoch": 0.357988922028121, + "grad_norm": 11.25, + "learning_rate": 2.3148112557598497e-06, + "loss": 0.3602, + "step": 4201 + }, + { + "epoch": 0.35807413719642095, + "grad_norm": 12.625, + "learning_rate": 2.31472323122958e-06, + "loss": 0.9743, + "step": 4202 + }, + { + "epoch": 0.3581593523647209, + "grad_norm": 11.6875, + "learning_rate": 2.314635187458691e-06, + "loss": 0.4343, + "step": 4203 + }, + { + "epoch": 0.35824456753302086, + "grad_norm": 11.0625, + "learning_rate": 2.314547124448773e-06, + "loss": 0.4612, + "step": 4204 + }, + { + "epoch": 0.3583297827013208, + "grad_norm": 14.0, + "learning_rate": 2.314459042201419e-06, + "loss": 0.7736, + "step": 4205 + }, + { + "epoch": 0.3584149978696208, + "grad_norm": 12.0, + "learning_rate": 2.314370940718219e-06, + "loss": 0.6694, + "step": 4206 + }, + { + "epoch": 0.35850021303792073, + "grad_norm": 17.375, + "learning_rate": 2.314282820000766e-06, + "loss": 0.6235, + "step": 4207 + }, + { + "epoch": 0.3585854282062207, + "grad_norm": 14.1875, + "learning_rate": 2.314194680050652e-06, + "loss": 0.8902, + "step": 4208 + }, + { + "epoch": 0.35867064337452065, + "grad_norm": 13.6875, + "learning_rate": 2.3141065208694704e-06, + "loss": 0.5892, + "step": 4209 + }, + { + "epoch": 0.3587558585428206, + "grad_norm": 24.0, + "learning_rate": 2.3140183424588136e-06, + "loss": 0.6459, + "step": 4210 + }, + { + "epoch": 0.35884107371112056, + "grad_norm": 15.3125, + "learning_rate": 2.3139301448202757e-06, + "loss": 0.6411, + "step": 4211 + }, + { + "epoch": 0.3589262888794205, + "grad_norm": 12.3125, + "learning_rate": 2.31384192795545e-06, + "loss": 0.5189, + "step": 4212 + }, + { + "epoch": 0.3590115040477205, + "grad_norm": 28.125, + "learning_rate": 2.3137536918659307e-06, + "loss": 1.1267, + "step": 4213 + }, + { + "epoch": 0.35909671921602043, + "grad_norm": 13.75, + "learning_rate": 2.313665436553313e-06, + "loss": 0.6345, + "step": 4214 + }, + { + "epoch": 0.3591819343843204, + "grad_norm": 13.5625, + "learning_rate": 2.313577162019191e-06, + "loss": 0.5901, + "step": 4215 + }, + { + "epoch": 0.35926714955262035, + "grad_norm": 15.875, + "learning_rate": 2.31348886826516e-06, + "loss": 0.8198, + "step": 4216 + }, + { + "epoch": 0.3593523647209203, + "grad_norm": 16.875, + "learning_rate": 2.313400555292816e-06, + "loss": 0.8849, + "step": 4217 + }, + { + "epoch": 0.35943757988922026, + "grad_norm": 12.1875, + "learning_rate": 2.3133122231037547e-06, + "loss": 0.4802, + "step": 4218 + }, + { + "epoch": 0.3595227950575202, + "grad_norm": 13.1875, + "learning_rate": 2.313223871699572e-06, + "loss": 0.592, + "step": 4219 + }, + { + "epoch": 0.3596080102258202, + "grad_norm": 14.6875, + "learning_rate": 2.313135501081865e-06, + "loss": 0.8414, + "step": 4220 + }, + { + "epoch": 0.35969322539412013, + "grad_norm": 14.125, + "learning_rate": 2.3130471112522303e-06, + "loss": 0.7617, + "step": 4221 + }, + { + "epoch": 0.3597784405624201, + "grad_norm": 13.75, + "learning_rate": 2.3129587022122652e-06, + "loss": 0.7365, + "step": 4222 + }, + { + "epoch": 0.35986365573072004, + "grad_norm": 10.875, + "learning_rate": 2.3128702739635677e-06, + "loss": 0.4697, + "step": 4223 + }, + { + "epoch": 0.35994887089902, + "grad_norm": 17.625, + "learning_rate": 2.3127818265077357e-06, + "loss": 0.6869, + "step": 4224 + }, + { + "epoch": 0.36003408606731996, + "grad_norm": 24.125, + "learning_rate": 2.312693359846367e-06, + "loss": 0.863, + "step": 4225 + }, + { + "epoch": 0.3601193012356199, + "grad_norm": 17.5, + "learning_rate": 2.312604873981061e-06, + "loss": 0.8449, + "step": 4226 + }, + { + "epoch": 0.36020451640391987, + "grad_norm": 20.5, + "learning_rate": 2.312516368913416e-06, + "loss": 0.9352, + "step": 4227 + }, + { + "epoch": 0.36028973157221983, + "grad_norm": 14.4375, + "learning_rate": 2.312427844645032e-06, + "loss": 1.0289, + "step": 4228 + }, + { + "epoch": 0.3603749467405198, + "grad_norm": 13.875, + "learning_rate": 2.3123393011775085e-06, + "loss": 0.7591, + "step": 4229 + }, + { + "epoch": 0.36046016190881974, + "grad_norm": 11.4375, + "learning_rate": 2.3122507385124453e-06, + "loss": 0.5287, + "step": 4230 + }, + { + "epoch": 0.3605453770771197, + "grad_norm": 11.5, + "learning_rate": 2.3121621566514433e-06, + "loss": 0.5529, + "step": 4231 + }, + { + "epoch": 0.3606305922454197, + "grad_norm": 15.25, + "learning_rate": 2.3120735555961034e-06, + "loss": 0.7931, + "step": 4232 + }, + { + "epoch": 0.36071580741371967, + "grad_norm": 15.8125, + "learning_rate": 2.311984935348026e-06, + "loss": 1.0952, + "step": 4233 + }, + { + "epoch": 0.3608010225820196, + "grad_norm": 15.4375, + "learning_rate": 2.311896295908813e-06, + "loss": 0.6672, + "step": 4234 + }, + { + "epoch": 0.3608862377503196, + "grad_norm": 11.625, + "learning_rate": 2.311807637280066e-06, + "loss": 0.4892, + "step": 4235 + }, + { + "epoch": 0.36097145291861954, + "grad_norm": 14.75, + "learning_rate": 2.3117189594633873e-06, + "loss": 0.9461, + "step": 4236 + }, + { + "epoch": 0.3610566680869195, + "grad_norm": 57.0, + "learning_rate": 2.3116302624603797e-06, + "loss": 1.0312, + "step": 4237 + }, + { + "epoch": 0.36114188325521945, + "grad_norm": 13.3125, + "learning_rate": 2.311541546272645e-06, + "loss": 0.7095, + "step": 4238 + }, + { + "epoch": 0.3612270984235194, + "grad_norm": 12.5, + "learning_rate": 2.3114528109017876e-06, + "loss": 0.733, + "step": 4239 + }, + { + "epoch": 0.36131231359181937, + "grad_norm": 15.3125, + "learning_rate": 2.3113640563494105e-06, + "loss": 0.7853, + "step": 4240 + }, + { + "epoch": 0.3613975287601193, + "grad_norm": 11.9375, + "learning_rate": 2.3112752826171174e-06, + "loss": 0.3749, + "step": 4241 + }, + { + "epoch": 0.3614827439284193, + "grad_norm": 21.375, + "learning_rate": 2.3111864897065127e-06, + "loss": 0.9213, + "step": 4242 + }, + { + "epoch": 0.36156795909671924, + "grad_norm": 21.625, + "learning_rate": 2.3110976776192016e-06, + "loss": 0.694, + "step": 4243 + }, + { + "epoch": 0.3616531742650192, + "grad_norm": 11.25, + "learning_rate": 2.311008846356788e-06, + "loss": 0.5355, + "step": 4244 + }, + { + "epoch": 0.36173838943331915, + "grad_norm": 19.625, + "learning_rate": 2.3109199959208777e-06, + "loss": 1.1726, + "step": 4245 + }, + { + "epoch": 0.3618236046016191, + "grad_norm": 13.0625, + "learning_rate": 2.310831126313076e-06, + "loss": 0.6114, + "step": 4246 + }, + { + "epoch": 0.36190881976991907, + "grad_norm": 14.375, + "learning_rate": 2.3107422375349894e-06, + "loss": 0.9449, + "step": 4247 + }, + { + "epoch": 0.361994034938219, + "grad_norm": 17.25, + "learning_rate": 2.310653329588224e-06, + "loss": 0.9274, + "step": 4248 + }, + { + "epoch": 0.362079250106519, + "grad_norm": 16.625, + "learning_rate": 2.310564402474386e-06, + "loss": 0.9639, + "step": 4249 + }, + { + "epoch": 0.36216446527481894, + "grad_norm": 15.0625, + "learning_rate": 2.3104754561950828e-06, + "loss": 0.8019, + "step": 4250 + }, + { + "epoch": 0.3622496804431189, + "grad_norm": 15.375, + "learning_rate": 2.310386490751922e-06, + "loss": 0.8768, + "step": 4251 + }, + { + "epoch": 0.36233489561141885, + "grad_norm": 13.125, + "learning_rate": 2.310297506146511e-06, + "loss": 0.7227, + "step": 4252 + }, + { + "epoch": 0.3624201107797188, + "grad_norm": 12.375, + "learning_rate": 2.3102085023804576e-06, + "loss": 0.4608, + "step": 4253 + }, + { + "epoch": 0.36250532594801876, + "grad_norm": 15.4375, + "learning_rate": 2.3101194794553706e-06, + "loss": 0.9351, + "step": 4254 + }, + { + "epoch": 0.3625905411163187, + "grad_norm": 15.125, + "learning_rate": 2.3100304373728585e-06, + "loss": 0.6416, + "step": 4255 + }, + { + "epoch": 0.3626757562846187, + "grad_norm": 15.0, + "learning_rate": 2.3099413761345303e-06, + "loss": 0.781, + "step": 4256 + }, + { + "epoch": 0.36276097145291863, + "grad_norm": 16.625, + "learning_rate": 2.309852295741996e-06, + "loss": 0.5596, + "step": 4257 + }, + { + "epoch": 0.3628461866212186, + "grad_norm": 16.875, + "learning_rate": 2.3097631961968646e-06, + "loss": 0.8913, + "step": 4258 + }, + { + "epoch": 0.36293140178951855, + "grad_norm": 20.375, + "learning_rate": 2.3096740775007472e-06, + "loss": 0.9099, + "step": 4259 + }, + { + "epoch": 0.3630166169578185, + "grad_norm": 15.8125, + "learning_rate": 2.309584939655253e-06, + "loss": 0.9584, + "step": 4260 + }, + { + "epoch": 0.36310183212611846, + "grad_norm": 14.8125, + "learning_rate": 2.3094957826619933e-06, + "loss": 0.5139, + "step": 4261 + }, + { + "epoch": 0.3631870472944184, + "grad_norm": 15.3125, + "learning_rate": 2.3094066065225794e-06, + "loss": 1.0313, + "step": 4262 + }, + { + "epoch": 0.3632722624627184, + "grad_norm": 14.5, + "learning_rate": 2.3093174112386234e-06, + "loss": 0.6139, + "step": 4263 + }, + { + "epoch": 0.36335747763101833, + "grad_norm": 13.4375, + "learning_rate": 2.3092281968117363e-06, + "loss": 0.6167, + "step": 4264 + }, + { + "epoch": 0.3634426927993183, + "grad_norm": 13.0, + "learning_rate": 2.3091389632435305e-06, + "loss": 0.7595, + "step": 4265 + }, + { + "epoch": 0.36352790796761825, + "grad_norm": 15.75, + "learning_rate": 2.3090497105356187e-06, + "loss": 0.8667, + "step": 4266 + }, + { + "epoch": 0.3636131231359182, + "grad_norm": 13.1875, + "learning_rate": 2.3089604386896132e-06, + "loss": 0.5804, + "step": 4267 + }, + { + "epoch": 0.36369833830421816, + "grad_norm": 16.25, + "learning_rate": 2.308871147707128e-06, + "loss": 0.7245, + "step": 4268 + }, + { + "epoch": 0.3637835534725181, + "grad_norm": 17.125, + "learning_rate": 2.3087818375897767e-06, + "loss": 0.4998, + "step": 4269 + }, + { + "epoch": 0.3638687686408181, + "grad_norm": 14.4375, + "learning_rate": 2.3086925083391727e-06, + "loss": 0.5352, + "step": 4270 + }, + { + "epoch": 0.36395398380911803, + "grad_norm": 12.875, + "learning_rate": 2.3086031599569303e-06, + "loss": 0.6056, + "step": 4271 + }, + { + "epoch": 0.364039198977418, + "grad_norm": 16.125, + "learning_rate": 2.3085137924446644e-06, + "loss": 0.7291, + "step": 4272 + }, + { + "epoch": 0.36412441414571795, + "grad_norm": 13.4375, + "learning_rate": 2.3084244058039902e-06, + "loss": 0.7364, + "step": 4273 + }, + { + "epoch": 0.3642096293140179, + "grad_norm": 13.0, + "learning_rate": 2.3083350000365223e-06, + "loss": 0.6466, + "step": 4274 + }, + { + "epoch": 0.36429484448231786, + "grad_norm": 12.75, + "learning_rate": 2.3082455751438766e-06, + "loss": 0.3638, + "step": 4275 + }, + { + "epoch": 0.3643800596506178, + "grad_norm": 10.75, + "learning_rate": 2.308156131127669e-06, + "loss": 0.5419, + "step": 4276 + }, + { + "epoch": 0.3644652748189178, + "grad_norm": 13.125, + "learning_rate": 2.3080666679895165e-06, + "loss": 0.622, + "step": 4277 + }, + { + "epoch": 0.36455048998721773, + "grad_norm": 12.6875, + "learning_rate": 2.307977185731035e-06, + "loss": 0.6674, + "step": 4278 + }, + { + "epoch": 0.3646357051555177, + "grad_norm": 13.875, + "learning_rate": 2.3078876843538423e-06, + "loss": 0.4298, + "step": 4279 + }, + { + "epoch": 0.36472092032381764, + "grad_norm": 11.9375, + "learning_rate": 2.3077981638595554e-06, + "loss": 0.5843, + "step": 4280 + }, + { + "epoch": 0.3648061354921176, + "grad_norm": 17.25, + "learning_rate": 2.3077086242497916e-06, + "loss": 0.7285, + "step": 4281 + }, + { + "epoch": 0.36489135066041756, + "grad_norm": 17.0, + "learning_rate": 2.3076190655261694e-06, + "loss": 0.5327, + "step": 4282 + }, + { + "epoch": 0.3649765658287175, + "grad_norm": 15.625, + "learning_rate": 2.3075294876903076e-06, + "loss": 1.0993, + "step": 4283 + }, + { + "epoch": 0.36506178099701747, + "grad_norm": 12.8125, + "learning_rate": 2.307439890743824e-06, + "loss": 0.6302, + "step": 4284 + }, + { + "epoch": 0.36514699616531743, + "grad_norm": 13.6875, + "learning_rate": 2.3073502746883387e-06, + "loss": 0.6014, + "step": 4285 + }, + { + "epoch": 0.3652322113336174, + "grad_norm": 12.25, + "learning_rate": 2.30726063952547e-06, + "loss": 0.5546, + "step": 4286 + }, + { + "epoch": 0.36531742650191734, + "grad_norm": 15.1875, + "learning_rate": 2.3071709852568392e-06, + "loss": 0.7087, + "step": 4287 + }, + { + "epoch": 0.3654026416702173, + "grad_norm": 14.3125, + "learning_rate": 2.307081311884065e-06, + "loss": 0.9438, + "step": 4288 + }, + { + "epoch": 0.36548785683851726, + "grad_norm": 10.125, + "learning_rate": 2.3069916194087687e-06, + "loss": 0.2662, + "step": 4289 + }, + { + "epoch": 0.3655730720068172, + "grad_norm": 18.0, + "learning_rate": 2.3069019078325713e-06, + "loss": 0.9825, + "step": 4290 + }, + { + "epoch": 0.36565828717511717, + "grad_norm": 11.0, + "learning_rate": 2.306812177157093e-06, + "loss": 0.4794, + "step": 4291 + }, + { + "epoch": 0.3657435023434171, + "grad_norm": 23.625, + "learning_rate": 2.3067224273839562e-06, + "loss": 0.8132, + "step": 4292 + }, + { + "epoch": 0.3658287175117171, + "grad_norm": 13.0625, + "learning_rate": 2.3066326585147827e-06, + "loss": 0.6164, + "step": 4293 + }, + { + "epoch": 0.36591393268001704, + "grad_norm": 13.875, + "learning_rate": 2.3065428705511946e-06, + "loss": 0.8203, + "step": 4294 + }, + { + "epoch": 0.365999147848317, + "grad_norm": 15.5625, + "learning_rate": 2.306453063494814e-06, + "loss": 0.5603, + "step": 4295 + }, + { + "epoch": 0.36608436301661695, + "grad_norm": 25.875, + "learning_rate": 2.3063632373472645e-06, + "loss": 0.9298, + "step": 4296 + }, + { + "epoch": 0.3661695781849169, + "grad_norm": 11.5625, + "learning_rate": 2.306273392110169e-06, + "loss": 0.4272, + "step": 4297 + }, + { + "epoch": 0.36625479335321687, + "grad_norm": 9.5625, + "learning_rate": 2.3061835277851514e-06, + "loss": 0.4046, + "step": 4298 + }, + { + "epoch": 0.3663400085215168, + "grad_norm": 27.0, + "learning_rate": 2.3060936443738347e-06, + "loss": 0.7893, + "step": 4299 + }, + { + "epoch": 0.3664252236898168, + "grad_norm": 9.875, + "learning_rate": 2.3060037418778446e-06, + "loss": 0.9641, + "step": 4300 + }, + { + "epoch": 0.36651043885811674, + "grad_norm": 10.75, + "learning_rate": 2.3059138202988044e-06, + "loss": 0.5021, + "step": 4301 + }, + { + "epoch": 0.3665956540264167, + "grad_norm": 30.375, + "learning_rate": 2.3058238796383396e-06, + "loss": 0.8703, + "step": 4302 + }, + { + "epoch": 0.36668086919471665, + "grad_norm": 18.0, + "learning_rate": 2.305733919898076e-06, + "loss": 1.1952, + "step": 4303 + }, + { + "epoch": 0.3667660843630166, + "grad_norm": 18.375, + "learning_rate": 2.305643941079639e-06, + "loss": 0.6789, + "step": 4304 + }, + { + "epoch": 0.36685129953131657, + "grad_norm": 14.4375, + "learning_rate": 2.305553943184654e-06, + "loss": 0.8883, + "step": 4305 + }, + { + "epoch": 0.3669365146996165, + "grad_norm": 12.9375, + "learning_rate": 2.3054639262147483e-06, + "loss": 0.4696, + "step": 4306 + }, + { + "epoch": 0.3670217298679165, + "grad_norm": 14.6875, + "learning_rate": 2.305373890171548e-06, + "loss": 0.9156, + "step": 4307 + }, + { + "epoch": 0.36710694503621644, + "grad_norm": 21.5, + "learning_rate": 2.3052838350566797e-06, + "loss": 0.8899, + "step": 4308 + }, + { + "epoch": 0.3671921602045164, + "grad_norm": 14.625, + "learning_rate": 2.3051937608717716e-06, + "loss": 0.5692, + "step": 4309 + }, + { + "epoch": 0.36727737537281635, + "grad_norm": 15.4375, + "learning_rate": 2.3051036676184514e-06, + "loss": 0.7317, + "step": 4310 + }, + { + "epoch": 0.3673625905411163, + "grad_norm": 11.6875, + "learning_rate": 2.3050135552983467e-06, + "loss": 0.4207, + "step": 4311 + }, + { + "epoch": 0.36744780570941626, + "grad_norm": 16.25, + "learning_rate": 2.3049234239130864e-06, + "loss": 0.6036, + "step": 4312 + }, + { + "epoch": 0.3675330208777162, + "grad_norm": 16.5, + "learning_rate": 2.304833273464299e-06, + "loss": 0.8124, + "step": 4313 + }, + { + "epoch": 0.3676182360460162, + "grad_norm": 19.0, + "learning_rate": 2.304743103953613e-06, + "loss": 0.7817, + "step": 4314 + }, + { + "epoch": 0.36770345121431613, + "grad_norm": 17.75, + "learning_rate": 2.304652915382659e-06, + "loss": 0.9005, + "step": 4315 + }, + { + "epoch": 0.3677886663826161, + "grad_norm": 13.625, + "learning_rate": 2.3045627077530664e-06, + "loss": 0.859, + "step": 4316 + }, + { + "epoch": 0.36787388155091605, + "grad_norm": 15.0625, + "learning_rate": 2.3044724810664647e-06, + "loss": 0.6488, + "step": 4317 + }, + { + "epoch": 0.367959096719216, + "grad_norm": 19.0, + "learning_rate": 2.3043822353244856e-06, + "loss": 0.6512, + "step": 4318 + }, + { + "epoch": 0.36804431188751596, + "grad_norm": 14.125, + "learning_rate": 2.304291970528759e-06, + "loss": 0.5366, + "step": 4319 + }, + { + "epoch": 0.3681295270558159, + "grad_norm": 11.4375, + "learning_rate": 2.3042016866809165e-06, + "loss": 0.5852, + "step": 4320 + }, + { + "epoch": 0.3682147422241159, + "grad_norm": 14.5625, + "learning_rate": 2.304111383782589e-06, + "loss": 0.6552, + "step": 4321 + }, + { + "epoch": 0.36829995739241583, + "grad_norm": 13.5, + "learning_rate": 2.304021061835409e-06, + "loss": 0.7635, + "step": 4322 + }, + { + "epoch": 0.3683851725607158, + "grad_norm": 14.0, + "learning_rate": 2.303930720841008e-06, + "loss": 0.5679, + "step": 4323 + }, + { + "epoch": 0.36847038772901575, + "grad_norm": 12.625, + "learning_rate": 2.30384036080102e-06, + "loss": 0.7854, + "step": 4324 + }, + { + "epoch": 0.3685556028973157, + "grad_norm": 13.3125, + "learning_rate": 2.3037499817170765e-06, + "loss": 0.7057, + "step": 4325 + }, + { + "epoch": 0.36864081806561566, + "grad_norm": 15.9375, + "learning_rate": 2.303659583590811e-06, + "loss": 0.7431, + "step": 4326 + }, + { + "epoch": 0.3687260332339156, + "grad_norm": 12.4375, + "learning_rate": 2.303569166423858e-06, + "loss": 0.6521, + "step": 4327 + }, + { + "epoch": 0.3688112484022156, + "grad_norm": 12.25, + "learning_rate": 2.3034787302178503e-06, + "loss": 0.559, + "step": 4328 + }, + { + "epoch": 0.36889646357051553, + "grad_norm": 14.25, + "learning_rate": 2.3033882749744223e-06, + "loss": 0.7696, + "step": 4329 + }, + { + "epoch": 0.3689816787388155, + "grad_norm": 13.8125, + "learning_rate": 2.3032978006952087e-06, + "loss": 0.6558, + "step": 4330 + }, + { + "epoch": 0.36906689390711545, + "grad_norm": 14.375, + "learning_rate": 2.3032073073818456e-06, + "loss": 0.8422, + "step": 4331 + }, + { + "epoch": 0.3691521090754154, + "grad_norm": 19.625, + "learning_rate": 2.3031167950359668e-06, + "loss": 1.0157, + "step": 4332 + }, + { + "epoch": 0.36923732424371536, + "grad_norm": 15.8125, + "learning_rate": 2.3030262636592084e-06, + "loss": 0.9339, + "step": 4333 + }, + { + "epoch": 0.3693225394120153, + "grad_norm": 20.25, + "learning_rate": 2.3029357132532072e-06, + "loss": 0.6715, + "step": 4334 + }, + { + "epoch": 0.3694077545803153, + "grad_norm": 16.875, + "learning_rate": 2.302845143819598e-06, + "loss": 0.7813, + "step": 4335 + }, + { + "epoch": 0.36949296974861523, + "grad_norm": 20.75, + "learning_rate": 2.302754555360019e-06, + "loss": 1.2755, + "step": 4336 + }, + { + "epoch": 0.3695781849169152, + "grad_norm": 13.125, + "learning_rate": 2.302663947876106e-06, + "loss": 0.504, + "step": 4337 + }, + { + "epoch": 0.36966340008521514, + "grad_norm": 9.25, + "learning_rate": 2.3025733213694972e-06, + "loss": 0.5204, + "step": 4338 + }, + { + "epoch": 0.3697486152535151, + "grad_norm": 12.75, + "learning_rate": 2.30248267584183e-06, + "loss": 0.5122, + "step": 4339 + }, + { + "epoch": 0.36983383042181506, + "grad_norm": 24.625, + "learning_rate": 2.3023920112947426e-06, + "loss": 0.917, + "step": 4340 + }, + { + "epoch": 0.369919045590115, + "grad_norm": 11.75, + "learning_rate": 2.3023013277298733e-06, + "loss": 0.7611, + "step": 4341 + }, + { + "epoch": 0.37000426075841497, + "grad_norm": 15.625, + "learning_rate": 2.3022106251488606e-06, + "loss": 0.9692, + "step": 4342 + }, + { + "epoch": 0.37008947592671493, + "grad_norm": 12.6875, + "learning_rate": 2.3021199035533444e-06, + "loss": 0.5794, + "step": 4343 + }, + { + "epoch": 0.3701746910950149, + "grad_norm": 12.125, + "learning_rate": 2.302029162944963e-06, + "loss": 0.6374, + "step": 4344 + }, + { + "epoch": 0.3702599062633149, + "grad_norm": 13.0625, + "learning_rate": 2.301938403325357e-06, + "loss": 0.6702, + "step": 4345 + }, + { + "epoch": 0.37034512143161485, + "grad_norm": 12.4375, + "learning_rate": 2.301847624696166e-06, + "loss": 0.5704, + "step": 4346 + }, + { + "epoch": 0.3704303365999148, + "grad_norm": 26.25, + "learning_rate": 2.301756827059031e-06, + "loss": 1.0872, + "step": 4347 + }, + { + "epoch": 0.37051555176821477, + "grad_norm": 15.125, + "learning_rate": 2.3016660104155917e-06, + "loss": 0.7761, + "step": 4348 + }, + { + "epoch": 0.3706007669365147, + "grad_norm": 11.6875, + "learning_rate": 2.3015751747674907e-06, + "loss": 0.5179, + "step": 4349 + }, + { + "epoch": 0.3706859821048147, + "grad_norm": 14.625, + "learning_rate": 2.301484320116369e-06, + "loss": 0.9032, + "step": 4350 + }, + { + "epoch": 0.37077119727311464, + "grad_norm": 15.0625, + "learning_rate": 2.301393446463868e-06, + "loss": 1.1451, + "step": 4351 + }, + { + "epoch": 0.3708564124414146, + "grad_norm": 27.125, + "learning_rate": 2.3013025538116305e-06, + "loss": 1.2107, + "step": 4352 + }, + { + "epoch": 0.37094162760971455, + "grad_norm": 12.6875, + "learning_rate": 2.3012116421612985e-06, + "loss": 0.9511, + "step": 4353 + }, + { + "epoch": 0.3710268427780145, + "grad_norm": 14.25, + "learning_rate": 2.301120711514515e-06, + "loss": 0.7931, + "step": 4354 + }, + { + "epoch": 0.37111205794631447, + "grad_norm": 13.125, + "learning_rate": 2.3010297618729233e-06, + "loss": 0.5906, + "step": 4355 + }, + { + "epoch": 0.3711972731146144, + "grad_norm": 14.0, + "learning_rate": 2.3009387932381665e-06, + "loss": 0.6769, + "step": 4356 + }, + { + "epoch": 0.3712824882829144, + "grad_norm": 15.875, + "learning_rate": 2.3008478056118896e-06, + "loss": 0.8732, + "step": 4357 + }, + { + "epoch": 0.37136770345121434, + "grad_norm": 11.3125, + "learning_rate": 2.3007567989957358e-06, + "loss": 0.3986, + "step": 4358 + }, + { + "epoch": 0.3714529186195143, + "grad_norm": 12.8125, + "learning_rate": 2.30066577339135e-06, + "loss": 0.7206, + "step": 4359 + }, + { + "epoch": 0.37153813378781425, + "grad_norm": 16.375, + "learning_rate": 2.3005747288003774e-06, + "loss": 0.8717, + "step": 4360 + }, + { + "epoch": 0.3716233489561142, + "grad_norm": 12.3125, + "learning_rate": 2.3004836652244626e-06, + "loss": 0.6998, + "step": 4361 + }, + { + "epoch": 0.37170856412441416, + "grad_norm": 13.1875, + "learning_rate": 2.300392582665252e-06, + "loss": 0.579, + "step": 4362 + }, + { + "epoch": 0.3717937792927141, + "grad_norm": 14.5625, + "learning_rate": 2.300301481124391e-06, + "loss": 0.8073, + "step": 4363 + }, + { + "epoch": 0.3718789944610141, + "grad_norm": 13.75, + "learning_rate": 2.3002103606035265e-06, + "loss": 0.6729, + "step": 4364 + }, + { + "epoch": 0.37196420962931404, + "grad_norm": 11.6875, + "learning_rate": 2.300119221104304e-06, + "loss": 0.3388, + "step": 4365 + }, + { + "epoch": 0.372049424797614, + "grad_norm": 16.125, + "learning_rate": 2.3000280626283718e-06, + "loss": 0.6479, + "step": 4366 + }, + { + "epoch": 0.37213463996591395, + "grad_norm": 13.3125, + "learning_rate": 2.2999368851773764e-06, + "loss": 0.8203, + "step": 4367 + }, + { + "epoch": 0.3722198551342139, + "grad_norm": 16.75, + "learning_rate": 2.2998456887529654e-06, + "loss": 0.7895, + "step": 4368 + }, + { + "epoch": 0.37230507030251386, + "grad_norm": 14.5625, + "learning_rate": 2.299754473356787e-06, + "loss": 0.8478, + "step": 4369 + }, + { + "epoch": 0.3723902854708138, + "grad_norm": 20.625, + "learning_rate": 2.2996632389904903e-06, + "loss": 1.0137, + "step": 4370 + }, + { + "epoch": 0.3724755006391138, + "grad_norm": 13.75, + "learning_rate": 2.2995719856557232e-06, + "loss": 1.0345, + "step": 4371 + }, + { + "epoch": 0.37256071580741373, + "grad_norm": 12.6875, + "learning_rate": 2.2994807133541347e-06, + "loss": 0.7889, + "step": 4372 + }, + { + "epoch": 0.3726459309757137, + "grad_norm": 19.25, + "learning_rate": 2.2993894220873743e-06, + "loss": 0.9149, + "step": 4373 + }, + { + "epoch": 0.37273114614401365, + "grad_norm": 11.4375, + "learning_rate": 2.299298111857092e-06, + "loss": 0.4575, + "step": 4374 + }, + { + "epoch": 0.3728163613123136, + "grad_norm": 28.0, + "learning_rate": 2.299206782664938e-06, + "loss": 0.4605, + "step": 4375 + }, + { + "epoch": 0.37290157648061356, + "grad_norm": 16.125, + "learning_rate": 2.2991154345125617e-06, + "loss": 0.9573, + "step": 4376 + }, + { + "epoch": 0.3729867916489135, + "grad_norm": 10.1875, + "learning_rate": 2.299024067401615e-06, + "loss": 0.4191, + "step": 4377 + }, + { + "epoch": 0.3730720068172135, + "grad_norm": 13.5, + "learning_rate": 2.2989326813337483e-06, + "loss": 0.6309, + "step": 4378 + }, + { + "epoch": 0.37315722198551343, + "grad_norm": 12.5625, + "learning_rate": 2.2988412763106135e-06, + "loss": 0.458, + "step": 4379 + }, + { + "epoch": 0.3732424371538134, + "grad_norm": 13.6875, + "learning_rate": 2.2987498523338615e-06, + "loss": 0.7933, + "step": 4380 + }, + { + "epoch": 0.37332765232211335, + "grad_norm": 12.875, + "learning_rate": 2.2986584094051455e-06, + "loss": 0.6122, + "step": 4381 + }, + { + "epoch": 0.3734128674904133, + "grad_norm": 13.75, + "learning_rate": 2.2985669475261175e-06, + "loss": 0.6858, + "step": 4382 + }, + { + "epoch": 0.37349808265871326, + "grad_norm": 12.8125, + "learning_rate": 2.2984754666984306e-06, + "loss": 0.7758, + "step": 4383 + }, + { + "epoch": 0.3735832978270132, + "grad_norm": 16.75, + "learning_rate": 2.2983839669237372e-06, + "loss": 0.813, + "step": 4384 + }, + { + "epoch": 0.3736685129953132, + "grad_norm": 16.25, + "learning_rate": 2.2982924482036917e-06, + "loss": 0.5505, + "step": 4385 + }, + { + "epoch": 0.37375372816361313, + "grad_norm": 12.5625, + "learning_rate": 2.298200910539947e-06, + "loss": 0.7563, + "step": 4386 + }, + { + "epoch": 0.3738389433319131, + "grad_norm": 13.9375, + "learning_rate": 2.2981093539341588e-06, + "loss": 0.7459, + "step": 4387 + }, + { + "epoch": 0.37392415850021304, + "grad_norm": 18.25, + "learning_rate": 2.29801777838798e-06, + "loss": 0.9212, + "step": 4388 + }, + { + "epoch": 0.374009373668513, + "grad_norm": 19.125, + "learning_rate": 2.2979261839030657e-06, + "loss": 0.8884, + "step": 4389 + }, + { + "epoch": 0.37409458883681296, + "grad_norm": 16.375, + "learning_rate": 2.297834570481072e-06, + "loss": 0.8419, + "step": 4390 + }, + { + "epoch": 0.3741798040051129, + "grad_norm": 18.75, + "learning_rate": 2.297742938123654e-06, + "loss": 0.7831, + "step": 4391 + }, + { + "epoch": 0.37426501917341287, + "grad_norm": 12.9375, + "learning_rate": 2.2976512868324675e-06, + "loss": 0.4998, + "step": 4392 + }, + { + "epoch": 0.37435023434171283, + "grad_norm": 10.5, + "learning_rate": 2.2975596166091686e-06, + "loss": 0.3399, + "step": 4393 + }, + { + "epoch": 0.3744354495100128, + "grad_norm": 10.625, + "learning_rate": 2.297467927455414e-06, + "loss": 0.3401, + "step": 4394 + }, + { + "epoch": 0.37452066467831274, + "grad_norm": 16.625, + "learning_rate": 2.2973762193728606e-06, + "loss": 1.1352, + "step": 4395 + }, + { + "epoch": 0.3746058798466127, + "grad_norm": 14.9375, + "learning_rate": 2.2972844923631655e-06, + "loss": 0.8937, + "step": 4396 + }, + { + "epoch": 0.37469109501491266, + "grad_norm": 12.875, + "learning_rate": 2.297192746427987e-06, + "loss": 0.8557, + "step": 4397 + }, + { + "epoch": 0.3747763101832126, + "grad_norm": 28.25, + "learning_rate": 2.297100981568982e-06, + "loss": 0.9301, + "step": 4398 + }, + { + "epoch": 0.37486152535151257, + "grad_norm": 15.5, + "learning_rate": 2.2970091977878097e-06, + "loss": 0.7017, + "step": 4399 + }, + { + "epoch": 0.3749467405198125, + "grad_norm": 13.375, + "learning_rate": 2.296917395086128e-06, + "loss": 0.8333, + "step": 4400 + }, + { + "epoch": 0.3750319556881125, + "grad_norm": 17.625, + "learning_rate": 2.2968255734655963e-06, + "loss": 0.895, + "step": 4401 + }, + { + "epoch": 0.37511717085641244, + "grad_norm": 13.875, + "learning_rate": 2.2967337329278746e-06, + "loss": 0.9576, + "step": 4402 + }, + { + "epoch": 0.3752023860247124, + "grad_norm": 18.5, + "learning_rate": 2.296641873474621e-06, + "loss": 0.9124, + "step": 4403 + }, + { + "epoch": 0.37528760119301235, + "grad_norm": 14.3125, + "learning_rate": 2.2965499951074963e-06, + "loss": 0.7816, + "step": 4404 + }, + { + "epoch": 0.3753728163613123, + "grad_norm": 19.625, + "learning_rate": 2.296458097828161e-06, + "loss": 0.8399, + "step": 4405 + }, + { + "epoch": 0.37545803152961227, + "grad_norm": 16.625, + "learning_rate": 2.2963661816382755e-06, + "loss": 1.1387, + "step": 4406 + }, + { + "epoch": 0.3755432466979122, + "grad_norm": 16.75, + "learning_rate": 2.296274246539501e-06, + "loss": 1.0137, + "step": 4407 + }, + { + "epoch": 0.3756284618662122, + "grad_norm": 14.25, + "learning_rate": 2.2961822925334985e-06, + "loss": 0.9218, + "step": 4408 + }, + { + "epoch": 0.37571367703451214, + "grad_norm": 9.4375, + "learning_rate": 2.29609031962193e-06, + "loss": 0.3351, + "step": 4409 + }, + { + "epoch": 0.3757988922028121, + "grad_norm": 13.1875, + "learning_rate": 2.2959983278064576e-06, + "loss": 1.0254, + "step": 4410 + }, + { + "epoch": 0.37588410737111205, + "grad_norm": 14.3125, + "learning_rate": 2.2959063170887435e-06, + "loss": 0.7664, + "step": 4411 + }, + { + "epoch": 0.375969322539412, + "grad_norm": 14.25, + "learning_rate": 2.295814287470451e-06, + "loss": 0.4341, + "step": 4412 + }, + { + "epoch": 0.37605453770771197, + "grad_norm": 13.8125, + "learning_rate": 2.2957222389532423e-06, + "loss": 0.8277, + "step": 4413 + }, + { + "epoch": 0.3761397528760119, + "grad_norm": 19.375, + "learning_rate": 2.295630171538781e-06, + "loss": 0.5334, + "step": 4414 + }, + { + "epoch": 0.3762249680443119, + "grad_norm": 11.0625, + "learning_rate": 2.2955380852287314e-06, + "loss": 0.572, + "step": 4415 + }, + { + "epoch": 0.37631018321261184, + "grad_norm": 15.6875, + "learning_rate": 2.295445980024757e-06, + "loss": 0.5813, + "step": 4416 + }, + { + "epoch": 0.3763953983809118, + "grad_norm": 18.125, + "learning_rate": 2.2953538559285223e-06, + "loss": 1.1529, + "step": 4417 + }, + { + "epoch": 0.37648061354921175, + "grad_norm": 14.875, + "learning_rate": 2.2952617129416926e-06, + "loss": 0.7758, + "step": 4418 + }, + { + "epoch": 0.3765658287175117, + "grad_norm": 12.1875, + "learning_rate": 2.295169551065932e-06, + "loss": 0.3814, + "step": 4419 + }, + { + "epoch": 0.37665104388581166, + "grad_norm": 20.625, + "learning_rate": 2.295077370302907e-06, + "loss": 0.6561, + "step": 4420 + }, + { + "epoch": 0.3767362590541116, + "grad_norm": 16.0, + "learning_rate": 2.2949851706542837e-06, + "loss": 1.204, + "step": 4421 + }, + { + "epoch": 0.3768214742224116, + "grad_norm": 13.0, + "learning_rate": 2.2948929521217268e-06, + "loss": 0.7617, + "step": 4422 + }, + { + "epoch": 0.37690668939071154, + "grad_norm": 11.5625, + "learning_rate": 2.294800714706904e-06, + "loss": 0.4446, + "step": 4423 + }, + { + "epoch": 0.3769919045590115, + "grad_norm": 16.25, + "learning_rate": 2.2947084584114808e-06, + "loss": 1.0536, + "step": 4424 + }, + { + "epoch": 0.37707711972731145, + "grad_norm": 11.375, + "learning_rate": 2.294616183237126e-06, + "loss": 0.3647, + "step": 4425 + }, + { + "epoch": 0.3771623348956114, + "grad_norm": 16.625, + "learning_rate": 2.294523889185506e-06, + "loss": 0.9839, + "step": 4426 + }, + { + "epoch": 0.37724755006391136, + "grad_norm": 13.375, + "learning_rate": 2.2944315762582893e-06, + "loss": 0.5052, + "step": 4427 + }, + { + "epoch": 0.3773327652322113, + "grad_norm": 14.9375, + "learning_rate": 2.2943392444571435e-06, + "loss": 0.8716, + "step": 4428 + }, + { + "epoch": 0.3774179804005113, + "grad_norm": 18.125, + "learning_rate": 2.2942468937837377e-06, + "loss": 0.9572, + "step": 4429 + }, + { + "epoch": 0.37750319556881123, + "grad_norm": 14.0, + "learning_rate": 2.29415452423974e-06, + "loss": 0.4185, + "step": 4430 + }, + { + "epoch": 0.3775884107371112, + "grad_norm": 17.75, + "learning_rate": 2.29406213582682e-06, + "loss": 1.0528, + "step": 4431 + }, + { + "epoch": 0.37767362590541115, + "grad_norm": 11.625, + "learning_rate": 2.293969728546648e-06, + "loss": 0.4112, + "step": 4432 + }, + { + "epoch": 0.3777588410737111, + "grad_norm": 18.25, + "learning_rate": 2.2938773024008924e-06, + "loss": 1.0375, + "step": 4433 + }, + { + "epoch": 0.37784405624201106, + "grad_norm": 18.125, + "learning_rate": 2.2937848573912246e-06, + "loss": 0.776, + "step": 4434 + }, + { + "epoch": 0.377929271410311, + "grad_norm": 14.4375, + "learning_rate": 2.293692393519315e-06, + "loss": 0.6619, + "step": 4435 + }, + { + "epoch": 0.378014486578611, + "grad_norm": 24.5, + "learning_rate": 2.293599910786834e-06, + "loss": 1.2235, + "step": 4436 + }, + { + "epoch": 0.37809970174691093, + "grad_norm": 14.25, + "learning_rate": 2.2935074091954535e-06, + "loss": 0.6463, + "step": 4437 + }, + { + "epoch": 0.3781849169152109, + "grad_norm": 14.5625, + "learning_rate": 2.2934148887468448e-06, + "loss": 0.5819, + "step": 4438 + }, + { + "epoch": 0.37827013208351085, + "grad_norm": 10.0, + "learning_rate": 2.29332234944268e-06, + "loss": 0.3957, + "step": 4439 + }, + { + "epoch": 0.3783553472518108, + "grad_norm": 12.0625, + "learning_rate": 2.2932297912846307e-06, + "loss": 0.5868, + "step": 4440 + }, + { + "epoch": 0.37844056242011076, + "grad_norm": 18.875, + "learning_rate": 2.2931372142743703e-06, + "loss": 0.9452, + "step": 4441 + }, + { + "epoch": 0.3785257775884107, + "grad_norm": 17.25, + "learning_rate": 2.2930446184135712e-06, + "loss": 1.2007, + "step": 4442 + }, + { + "epoch": 0.3786109927567107, + "grad_norm": 15.8125, + "learning_rate": 2.292952003703907e-06, + "loss": 0.769, + "step": 4443 + }, + { + "epoch": 0.37869620792501063, + "grad_norm": 12.1875, + "learning_rate": 2.2928593701470516e-06, + "loss": 0.5402, + "step": 4444 + }, + { + "epoch": 0.3787814230933106, + "grad_norm": 17.5, + "learning_rate": 2.292766717744679e-06, + "loss": 0.5832, + "step": 4445 + }, + { + "epoch": 0.37886663826161054, + "grad_norm": 13.6875, + "learning_rate": 2.2926740464984625e-06, + "loss": 0.7219, + "step": 4446 + }, + { + "epoch": 0.3789518534299105, + "grad_norm": 15.6875, + "learning_rate": 2.292581356410078e-06, + "loss": 0.5683, + "step": 4447 + }, + { + "epoch": 0.37903706859821046, + "grad_norm": 13.5, + "learning_rate": 2.2924886474811995e-06, + "loss": 0.4263, + "step": 4448 + }, + { + "epoch": 0.3791222837665104, + "grad_norm": 11.3125, + "learning_rate": 2.2923959197135033e-06, + "loss": 0.4058, + "step": 4449 + }, + { + "epoch": 0.37920749893481037, + "grad_norm": 15.875, + "learning_rate": 2.2923031731086646e-06, + "loss": 0.7828, + "step": 4450 + }, + { + "epoch": 0.37929271410311033, + "grad_norm": 11.75, + "learning_rate": 2.2922104076683594e-06, + "loss": 0.5502, + "step": 4451 + }, + { + "epoch": 0.3793779292714103, + "grad_norm": 14.25, + "learning_rate": 2.292117623394264e-06, + "loss": 0.7631, + "step": 4452 + }, + { + "epoch": 0.37946314443971024, + "grad_norm": 18.125, + "learning_rate": 2.2920248202880552e-06, + "loss": 0.8136, + "step": 4453 + }, + { + "epoch": 0.3795483596080102, + "grad_norm": 20.0, + "learning_rate": 2.29193199835141e-06, + "loss": 1.064, + "step": 4454 + }, + { + "epoch": 0.37963357477631016, + "grad_norm": 12.9375, + "learning_rate": 2.291839157586006e-06, + "loss": 0.7133, + "step": 4455 + }, + { + "epoch": 0.3797187899446101, + "grad_norm": 12.625, + "learning_rate": 2.2917462979935207e-06, + "loss": 0.6152, + "step": 4456 + }, + { + "epoch": 0.37980400511291007, + "grad_norm": 18.75, + "learning_rate": 2.2916534195756323e-06, + "loss": 0.8759, + "step": 4457 + }, + { + "epoch": 0.3798892202812101, + "grad_norm": 19.25, + "learning_rate": 2.2915605223340193e-06, + "loss": 0.964, + "step": 4458 + }, + { + "epoch": 0.37997443544951004, + "grad_norm": 20.5, + "learning_rate": 2.2914676062703602e-06, + "loss": 0.7727, + "step": 4459 + }, + { + "epoch": 0.38005965061781, + "grad_norm": 12.0625, + "learning_rate": 2.2913746713863338e-06, + "loss": 0.6933, + "step": 4460 + }, + { + "epoch": 0.38014486578610995, + "grad_norm": 19.625, + "learning_rate": 2.2912817176836204e-06, + "loss": 1.1998, + "step": 4461 + }, + { + "epoch": 0.3802300809544099, + "grad_norm": 12.75, + "learning_rate": 2.291188745163899e-06, + "loss": 0.6076, + "step": 4462 + }, + { + "epoch": 0.38031529612270987, + "grad_norm": 14.4375, + "learning_rate": 2.29109575382885e-06, + "loss": 0.5314, + "step": 4463 + }, + { + "epoch": 0.3804005112910098, + "grad_norm": 14.375, + "learning_rate": 2.291002743680154e-06, + "loss": 0.4575, + "step": 4464 + }, + { + "epoch": 0.3804857264593098, + "grad_norm": 14.0, + "learning_rate": 2.2909097147194912e-06, + "loss": 1.0536, + "step": 4465 + }, + { + "epoch": 0.38057094162760974, + "grad_norm": 13.0625, + "learning_rate": 2.2908166669485433e-06, + "loss": 0.7613, + "step": 4466 + }, + { + "epoch": 0.3806561567959097, + "grad_norm": 15.9375, + "learning_rate": 2.290723600368992e-06, + "loss": 0.9637, + "step": 4467 + }, + { + "epoch": 0.38074137196420965, + "grad_norm": 13.625, + "learning_rate": 2.2906305149825185e-06, + "loss": 0.8168, + "step": 4468 + }, + { + "epoch": 0.3808265871325096, + "grad_norm": 12.75, + "learning_rate": 2.2905374107908047e-06, + "loss": 0.7389, + "step": 4469 + }, + { + "epoch": 0.38091180230080957, + "grad_norm": 20.25, + "learning_rate": 2.290444287795534e-06, + "loss": 1.0844, + "step": 4470 + }, + { + "epoch": 0.3809970174691095, + "grad_norm": 11.9375, + "learning_rate": 2.290351145998389e-06, + "loss": 0.4004, + "step": 4471 + }, + { + "epoch": 0.3810822326374095, + "grad_norm": 15.3125, + "learning_rate": 2.290257985401052e-06, + "loss": 0.5302, + "step": 4472 + }, + { + "epoch": 0.38116744780570944, + "grad_norm": 13.625, + "learning_rate": 2.2901648060052077e-06, + "loss": 0.711, + "step": 4473 + }, + { + "epoch": 0.3812526629740094, + "grad_norm": 23.625, + "learning_rate": 2.290071607812539e-06, + "loss": 0.9568, + "step": 4474 + }, + { + "epoch": 0.38133787814230935, + "grad_norm": 15.3125, + "learning_rate": 2.289978390824731e-06, + "loss": 0.8809, + "step": 4475 + }, + { + "epoch": 0.3814230933106093, + "grad_norm": 16.625, + "learning_rate": 2.289885155043467e-06, + "loss": 0.6751, + "step": 4476 + }, + { + "epoch": 0.38150830847890926, + "grad_norm": 17.625, + "learning_rate": 2.289791900470433e-06, + "loss": 1.0595, + "step": 4477 + }, + { + "epoch": 0.3815935236472092, + "grad_norm": 15.75, + "learning_rate": 2.289698627107314e-06, + "loss": 0.8042, + "step": 4478 + }, + { + "epoch": 0.3816787388155092, + "grad_norm": 17.875, + "learning_rate": 2.289605334955795e-06, + "loss": 1.0193, + "step": 4479 + }, + { + "epoch": 0.38176395398380913, + "grad_norm": 14.75, + "learning_rate": 2.2895120240175622e-06, + "loss": 0.6235, + "step": 4480 + }, + { + "epoch": 0.3818491691521091, + "grad_norm": 12.5, + "learning_rate": 2.2894186942943015e-06, + "loss": 0.6292, + "step": 4481 + }, + { + "epoch": 0.38193438432040905, + "grad_norm": 12.0625, + "learning_rate": 2.2893253457877e-06, + "loss": 0.4333, + "step": 4482 + }, + { + "epoch": 0.382019599488709, + "grad_norm": 18.625, + "learning_rate": 2.2892319784994447e-06, + "loss": 0.7367, + "step": 4483 + }, + { + "epoch": 0.38210481465700896, + "grad_norm": 15.5625, + "learning_rate": 2.2891385924312224e-06, + "loss": 0.7588, + "step": 4484 + }, + { + "epoch": 0.3821900298253089, + "grad_norm": 15.3125, + "learning_rate": 2.2890451875847205e-06, + "loss": 0.7638, + "step": 4485 + }, + { + "epoch": 0.3822752449936089, + "grad_norm": 16.375, + "learning_rate": 2.2889517639616277e-06, + "loss": 0.916, + "step": 4486 + }, + { + "epoch": 0.38236046016190883, + "grad_norm": 14.125, + "learning_rate": 2.288858321563631e-06, + "loss": 0.5469, + "step": 4487 + }, + { + "epoch": 0.3824456753302088, + "grad_norm": 13.0625, + "learning_rate": 2.2887648603924206e-06, + "loss": 0.4791, + "step": 4488 + }, + { + "epoch": 0.38253089049850875, + "grad_norm": 19.375, + "learning_rate": 2.288671380449684e-06, + "loss": 1.0754, + "step": 4489 + }, + { + "epoch": 0.3826161056668087, + "grad_norm": 19.625, + "learning_rate": 2.2885778817371115e-06, + "loss": 0.995, + "step": 4490 + }, + { + "epoch": 0.38270132083510866, + "grad_norm": 12.6875, + "learning_rate": 2.2884843642563924e-06, + "loss": 0.4355, + "step": 4491 + }, + { + "epoch": 0.3827865360034086, + "grad_norm": 17.0, + "learning_rate": 2.288390828009216e-06, + "loss": 1.1606, + "step": 4492 + }, + { + "epoch": 0.3828717511717086, + "grad_norm": 15.4375, + "learning_rate": 2.2882972729972737e-06, + "loss": 0.7709, + "step": 4493 + }, + { + "epoch": 0.38295696634000853, + "grad_norm": 12.8125, + "learning_rate": 2.2882036992222554e-06, + "loss": 0.7014, + "step": 4494 + }, + { + "epoch": 0.3830421815083085, + "grad_norm": 13.75, + "learning_rate": 2.288110106685852e-06, + "loss": 0.687, + "step": 4495 + }, + { + "epoch": 0.38312739667660844, + "grad_norm": 19.375, + "learning_rate": 2.2880164953897556e-06, + "loss": 1.0258, + "step": 4496 + }, + { + "epoch": 0.3832126118449084, + "grad_norm": 19.0, + "learning_rate": 2.287922865335657e-06, + "loss": 1.1486, + "step": 4497 + }, + { + "epoch": 0.38329782701320836, + "grad_norm": 24.5, + "learning_rate": 2.2878292165252486e-06, + "loss": 1.1778, + "step": 4498 + }, + { + "epoch": 0.3833830421815083, + "grad_norm": 11.0, + "learning_rate": 2.2877355489602223e-06, + "loss": 0.6552, + "step": 4499 + }, + { + "epoch": 0.3834682573498083, + "grad_norm": 14.1875, + "learning_rate": 2.2876418626422716e-06, + "loss": 0.6636, + "step": 4500 + }, + { + "epoch": 0.38355347251810823, + "grad_norm": 14.625, + "learning_rate": 2.2875481575730883e-06, + "loss": 0.6893, + "step": 4501 + }, + { + "epoch": 0.3836386876864082, + "grad_norm": 14.4375, + "learning_rate": 2.2874544337543674e-06, + "loss": 0.8543, + "step": 4502 + }, + { + "epoch": 0.38372390285470814, + "grad_norm": 10.9375, + "learning_rate": 2.287360691187801e-06, + "loss": 0.5006, + "step": 4503 + }, + { + "epoch": 0.3838091180230081, + "grad_norm": 14.1875, + "learning_rate": 2.287266929875084e-06, + "loss": 0.5273, + "step": 4504 + }, + { + "epoch": 0.38389433319130806, + "grad_norm": 18.875, + "learning_rate": 2.2871731498179105e-06, + "loss": 0.9154, + "step": 4505 + }, + { + "epoch": 0.383979548359608, + "grad_norm": 16.875, + "learning_rate": 2.287079351017975e-06, + "loss": 0.9958, + "step": 4506 + }, + { + "epoch": 0.38406476352790797, + "grad_norm": 18.25, + "learning_rate": 2.286985533476973e-06, + "loss": 0.9313, + "step": 4507 + }, + { + "epoch": 0.3841499786962079, + "grad_norm": 30.875, + "learning_rate": 2.2868916971966e-06, + "loss": 0.5992, + "step": 4508 + }, + { + "epoch": 0.3842351938645079, + "grad_norm": 12.1875, + "learning_rate": 2.286797842178551e-06, + "loss": 0.495, + "step": 4509 + }, + { + "epoch": 0.38432040903280784, + "grad_norm": 15.375, + "learning_rate": 2.2867039684245224e-06, + "loss": 0.8634, + "step": 4510 + }, + { + "epoch": 0.3844056242011078, + "grad_norm": 13.5, + "learning_rate": 2.2866100759362103e-06, + "loss": 0.61, + "step": 4511 + }, + { + "epoch": 0.38449083936940776, + "grad_norm": 13.4375, + "learning_rate": 2.2865161647153124e-06, + "loss": 1.007, + "step": 4512 + }, + { + "epoch": 0.3845760545377077, + "grad_norm": 13.1875, + "learning_rate": 2.2864222347635246e-06, + "loss": 0.7041, + "step": 4513 + }, + { + "epoch": 0.38466126970600767, + "grad_norm": 13.3125, + "learning_rate": 2.2863282860825446e-06, + "loss": 0.6604, + "step": 4514 + }, + { + "epoch": 0.3847464848743076, + "grad_norm": 14.5, + "learning_rate": 2.286234318674071e-06, + "loss": 0.8003, + "step": 4515 + }, + { + "epoch": 0.3848317000426076, + "grad_norm": 15.75, + "learning_rate": 2.2861403325398008e-06, + "loss": 0.9918, + "step": 4516 + }, + { + "epoch": 0.38491691521090754, + "grad_norm": 17.0, + "learning_rate": 2.286046327681433e-06, + "loss": 0.8309, + "step": 4517 + }, + { + "epoch": 0.3850021303792075, + "grad_norm": 14.125, + "learning_rate": 2.2859523041006665e-06, + "loss": 0.8778, + "step": 4518 + }, + { + "epoch": 0.38508734554750745, + "grad_norm": 24.625, + "learning_rate": 2.2858582617992e-06, + "loss": 1.1862, + "step": 4519 + }, + { + "epoch": 0.3851725607158074, + "grad_norm": 16.25, + "learning_rate": 2.2857642007787324e-06, + "loss": 0.7936, + "step": 4520 + }, + { + "epoch": 0.38525777588410737, + "grad_norm": 15.5625, + "learning_rate": 2.285670121040965e-06, + "loss": 1.0927, + "step": 4521 + }, + { + "epoch": 0.3853429910524073, + "grad_norm": 13.125, + "learning_rate": 2.2855760225875967e-06, + "loss": 0.523, + "step": 4522 + }, + { + "epoch": 0.3854282062207073, + "grad_norm": 17.75, + "learning_rate": 2.2854819054203285e-06, + "loss": 1.069, + "step": 4523 + }, + { + "epoch": 0.38551342138900724, + "grad_norm": 14.1875, + "learning_rate": 2.2853877695408608e-06, + "loss": 0.79, + "step": 4524 + }, + { + "epoch": 0.3855986365573072, + "grad_norm": 13.0, + "learning_rate": 2.285293614950895e-06, + "loss": 0.7708, + "step": 4525 + }, + { + "epoch": 0.38568385172560715, + "grad_norm": 17.625, + "learning_rate": 2.285199441652133e-06, + "loss": 1.1711, + "step": 4526 + }, + { + "epoch": 0.3857690668939071, + "grad_norm": 14.5625, + "learning_rate": 2.2851052496462754e-06, + "loss": 0.8402, + "step": 4527 + }, + { + "epoch": 0.38585428206220707, + "grad_norm": 12.3125, + "learning_rate": 2.2850110389350254e-06, + "loss": 0.5961, + "step": 4528 + }, + { + "epoch": 0.385939497230507, + "grad_norm": 16.875, + "learning_rate": 2.284916809520085e-06, + "loss": 0.9529, + "step": 4529 + }, + { + "epoch": 0.386024712398807, + "grad_norm": 12.6875, + "learning_rate": 2.2848225614031573e-06, + "loss": 0.6457, + "step": 4530 + }, + { + "epoch": 0.38610992756710694, + "grad_norm": 16.75, + "learning_rate": 2.284728294585945e-06, + "loss": 0.9751, + "step": 4531 + }, + { + "epoch": 0.3861951427354069, + "grad_norm": 11.8125, + "learning_rate": 2.2846340090701522e-06, + "loss": 0.5709, + "step": 4532 + }, + { + "epoch": 0.38628035790370685, + "grad_norm": 15.5, + "learning_rate": 2.2845397048574823e-06, + "loss": 0.8488, + "step": 4533 + }, + { + "epoch": 0.3863655730720068, + "grad_norm": 12.6875, + "learning_rate": 2.28444538194964e-06, + "loss": 0.5948, + "step": 4534 + }, + { + "epoch": 0.38645078824030676, + "grad_norm": 14.625, + "learning_rate": 2.284351040348329e-06, + "loss": 0.7408, + "step": 4535 + }, + { + "epoch": 0.3865360034086067, + "grad_norm": 13.1875, + "learning_rate": 2.284256680055255e-06, + "loss": 0.8908, + "step": 4536 + }, + { + "epoch": 0.3866212185769067, + "grad_norm": 19.375, + "learning_rate": 2.2841623010721228e-06, + "loss": 1.0765, + "step": 4537 + }, + { + "epoch": 0.38670643374520663, + "grad_norm": 13.125, + "learning_rate": 2.2840679034006373e-06, + "loss": 0.9415, + "step": 4538 + }, + { + "epoch": 0.3867916489135066, + "grad_norm": 17.75, + "learning_rate": 2.283973487042505e-06, + "loss": 0.945, + "step": 4539 + }, + { + "epoch": 0.38687686408180655, + "grad_norm": 20.25, + "learning_rate": 2.2838790519994326e-06, + "loss": 1.1383, + "step": 4540 + }, + { + "epoch": 0.3869620792501065, + "grad_norm": 23.125, + "learning_rate": 2.2837845982731257e-06, + "loss": 0.7768, + "step": 4541 + }, + { + "epoch": 0.38704729441840646, + "grad_norm": 12.875, + "learning_rate": 2.2836901258652915e-06, + "loss": 0.7491, + "step": 4542 + }, + { + "epoch": 0.3871325095867064, + "grad_norm": 17.25, + "learning_rate": 2.2835956347776377e-06, + "loss": 0.8438, + "step": 4543 + }, + { + "epoch": 0.3872177247550064, + "grad_norm": 16.625, + "learning_rate": 2.283501125011871e-06, + "loss": 0.7922, + "step": 4544 + }, + { + "epoch": 0.38730293992330633, + "grad_norm": 16.125, + "learning_rate": 2.2834065965697e-06, + "loss": 0.4668, + "step": 4545 + }, + { + "epoch": 0.3873881550916063, + "grad_norm": 14.375, + "learning_rate": 2.2833120494528325e-06, + "loss": 0.7096, + "step": 4546 + }, + { + "epoch": 0.38747337025990625, + "grad_norm": 18.625, + "learning_rate": 2.2832174836629774e-06, + "loss": 0.7342, + "step": 4547 + }, + { + "epoch": 0.3875585854282062, + "grad_norm": 14.1875, + "learning_rate": 2.283122899201843e-06, + "loss": 0.7097, + "step": 4548 + }, + { + "epoch": 0.38764380059650616, + "grad_norm": 16.125, + "learning_rate": 2.283028296071139e-06, + "loss": 1.0786, + "step": 4549 + }, + { + "epoch": 0.3877290157648061, + "grad_norm": 13.75, + "learning_rate": 2.282933674272575e-06, + "loss": 0.791, + "step": 4550 + }, + { + "epoch": 0.3878142309331061, + "grad_norm": 11.8125, + "learning_rate": 2.282839033807861e-06, + "loss": 0.5015, + "step": 4551 + }, + { + "epoch": 0.38789944610140603, + "grad_norm": 18.625, + "learning_rate": 2.2827443746787068e-06, + "loss": 1.1809, + "step": 4552 + }, + { + "epoch": 0.387984661269706, + "grad_norm": 12.0, + "learning_rate": 2.2826496968868234e-06, + "loss": 0.524, + "step": 4553 + }, + { + "epoch": 0.38806987643800595, + "grad_norm": 16.375, + "learning_rate": 2.282555000433921e-06, + "loss": 0.8558, + "step": 4554 + }, + { + "epoch": 0.3881550916063059, + "grad_norm": 12.625, + "learning_rate": 2.282460285321712e-06, + "loss": 0.5823, + "step": 4555 + }, + { + "epoch": 0.38824030677460586, + "grad_norm": 10.5, + "learning_rate": 2.2823655515519076e-06, + "loss": 0.5241, + "step": 4556 + }, + { + "epoch": 0.3883255219429058, + "grad_norm": 13.3125, + "learning_rate": 2.2822707991262196e-06, + "loss": 0.6566, + "step": 4557 + }, + { + "epoch": 0.3884107371112058, + "grad_norm": 11.5625, + "learning_rate": 2.2821760280463596e-06, + "loss": 0.5527, + "step": 4558 + }, + { + "epoch": 0.38849595227950573, + "grad_norm": 15.0625, + "learning_rate": 2.2820812383140413e-06, + "loss": 0.8569, + "step": 4559 + }, + { + "epoch": 0.3885811674478057, + "grad_norm": 14.1875, + "learning_rate": 2.281986429930977e-06, + "loss": 0.8562, + "step": 4560 + }, + { + "epoch": 0.38866638261610564, + "grad_norm": 12.9375, + "learning_rate": 2.2818916028988798e-06, + "loss": 0.8177, + "step": 4561 + }, + { + "epoch": 0.3887515977844056, + "grad_norm": 15.3125, + "learning_rate": 2.2817967572194644e-06, + "loss": 1.1979, + "step": 4562 + }, + { + "epoch": 0.38883681295270556, + "grad_norm": 16.0, + "learning_rate": 2.2817018928944433e-06, + "loss": 0.8594, + "step": 4563 + }, + { + "epoch": 0.3889220281210055, + "grad_norm": 14.0, + "learning_rate": 2.2816070099255324e-06, + "loss": 0.9151, + "step": 4564 + }, + { + "epoch": 0.38900724328930547, + "grad_norm": 10.75, + "learning_rate": 2.281512108314445e-06, + "loss": 0.3492, + "step": 4565 + }, + { + "epoch": 0.38909245845760543, + "grad_norm": 15.6875, + "learning_rate": 2.2814171880628964e-06, + "loss": 0.8376, + "step": 4566 + }, + { + "epoch": 0.3891776736259054, + "grad_norm": 14.1875, + "learning_rate": 2.2813222491726022e-06, + "loss": 1.0273, + "step": 4567 + }, + { + "epoch": 0.38926288879420534, + "grad_norm": 14.5, + "learning_rate": 2.281227291645278e-06, + "loss": 0.8885, + "step": 4568 + }, + { + "epoch": 0.3893481039625053, + "grad_norm": 10.3125, + "learning_rate": 2.28113231548264e-06, + "loss": 0.3477, + "step": 4569 + }, + { + "epoch": 0.3894333191308053, + "grad_norm": 16.0, + "learning_rate": 2.2810373206864034e-06, + "loss": 0.7715, + "step": 4570 + }, + { + "epoch": 0.38951853429910527, + "grad_norm": 16.75, + "learning_rate": 2.280942307258286e-06, + "loss": 1.1484, + "step": 4571 + }, + { + "epoch": 0.3896037494674052, + "grad_norm": 14.25, + "learning_rate": 2.2808472752000043e-06, + "loss": 0.78, + "step": 4572 + }, + { + "epoch": 0.3896889646357052, + "grad_norm": 15.375, + "learning_rate": 2.280752224513276e-06, + "loss": 0.635, + "step": 4573 + }, + { + "epoch": 0.38977417980400514, + "grad_norm": 35.25, + "learning_rate": 2.280657155199818e-06, + "loss": 0.9081, + "step": 4574 + }, + { + "epoch": 0.3898593949723051, + "grad_norm": 11.375, + "learning_rate": 2.2805620672613494e-06, + "loss": 0.5269, + "step": 4575 + }, + { + "epoch": 0.38994461014060505, + "grad_norm": 15.5, + "learning_rate": 2.280466960699588e-06, + "loss": 0.7133, + "step": 4576 + }, + { + "epoch": 0.390029825308905, + "grad_norm": 13.8125, + "learning_rate": 2.280371835516252e-06, + "loss": 0.7546, + "step": 4577 + }, + { + "epoch": 0.39011504047720497, + "grad_norm": 14.0625, + "learning_rate": 2.280276691713061e-06, + "loss": 0.6967, + "step": 4578 + }, + { + "epoch": 0.3902002556455049, + "grad_norm": 17.375, + "learning_rate": 2.280181529291734e-06, + "loss": 1.1751, + "step": 4579 + }, + { + "epoch": 0.3902854708138049, + "grad_norm": 18.75, + "learning_rate": 2.2800863482539905e-06, + "loss": 0.8412, + "step": 4580 + }, + { + "epoch": 0.39037068598210484, + "grad_norm": 11.8125, + "learning_rate": 2.279991148601551e-06, + "loss": 0.516, + "step": 4581 + }, + { + "epoch": 0.3904559011504048, + "grad_norm": 10.5, + "learning_rate": 2.2798959303361364e-06, + "loss": 0.4011, + "step": 4582 + }, + { + "epoch": 0.39054111631870475, + "grad_norm": 13.75, + "learning_rate": 2.2798006934594663e-06, + "loss": 0.6922, + "step": 4583 + }, + { + "epoch": 0.3906263314870047, + "grad_norm": 14.5, + "learning_rate": 2.279705437973262e-06, + "loss": 0.7145, + "step": 4584 + }, + { + "epoch": 0.39071154665530466, + "grad_norm": 15.25, + "learning_rate": 2.279610163879245e-06, + "loss": 0.7449, + "step": 4585 + }, + { + "epoch": 0.3907967618236046, + "grad_norm": 16.5, + "learning_rate": 2.279514871179137e-06, + "loss": 0.9176, + "step": 4586 + }, + { + "epoch": 0.3908819769919046, + "grad_norm": 11.625, + "learning_rate": 2.27941955987466e-06, + "loss": 0.5752, + "step": 4587 + }, + { + "epoch": 0.39096719216020454, + "grad_norm": 14.9375, + "learning_rate": 2.2793242299675366e-06, + "loss": 0.8102, + "step": 4588 + }, + { + "epoch": 0.3910524073285045, + "grad_norm": 15.375, + "learning_rate": 2.279228881459489e-06, + "loss": 0.8104, + "step": 4589 + }, + { + "epoch": 0.39113762249680445, + "grad_norm": 12.875, + "learning_rate": 2.279133514352241e-06, + "loss": 0.6299, + "step": 4590 + }, + { + "epoch": 0.3912228376651044, + "grad_norm": 12.3125, + "learning_rate": 2.279038128647515e-06, + "loss": 0.6081, + "step": 4591 + }, + { + "epoch": 0.39130805283340436, + "grad_norm": 21.25, + "learning_rate": 2.2789427243470356e-06, + "loss": 0.8856, + "step": 4592 + }, + { + "epoch": 0.3913932680017043, + "grad_norm": 12.8125, + "learning_rate": 2.278847301452526e-06, + "loss": 0.7382, + "step": 4593 + }, + { + "epoch": 0.3914784831700043, + "grad_norm": 14.25, + "learning_rate": 2.2787518599657116e-06, + "loss": 0.9935, + "step": 4594 + }, + { + "epoch": 0.39156369833830423, + "grad_norm": 15.0, + "learning_rate": 2.2786563998883167e-06, + "loss": 0.3733, + "step": 4595 + }, + { + "epoch": 0.3916489135066042, + "grad_norm": 9.25, + "learning_rate": 2.278560921222066e-06, + "loss": 0.3683, + "step": 4596 + }, + { + "epoch": 0.39173412867490415, + "grad_norm": 10.8125, + "learning_rate": 2.278465423968685e-06, + "loss": 0.5061, + "step": 4597 + }, + { + "epoch": 0.3918193438432041, + "grad_norm": 14.5625, + "learning_rate": 2.2783699081298995e-06, + "loss": 0.7671, + "step": 4598 + }, + { + "epoch": 0.39190455901150406, + "grad_norm": 15.1875, + "learning_rate": 2.278274373707436e-06, + "loss": 1.0247, + "step": 4599 + }, + { + "epoch": 0.391989774179804, + "grad_norm": 16.5, + "learning_rate": 2.2781788207030203e-06, + "loss": 0.7135, + "step": 4600 + }, + { + "epoch": 0.392074989348104, + "grad_norm": 17.25, + "learning_rate": 2.2780832491183796e-06, + "loss": 0.9954, + "step": 4601 + }, + { + "epoch": 0.39216020451640393, + "grad_norm": 13.25, + "learning_rate": 2.277987658955241e-06, + "loss": 0.7754, + "step": 4602 + }, + { + "epoch": 0.3922454196847039, + "grad_norm": 15.875, + "learning_rate": 2.277892050215331e-06, + "loss": 0.8307, + "step": 4603 + }, + { + "epoch": 0.39233063485300385, + "grad_norm": 15.75, + "learning_rate": 2.2777964229003786e-06, + "loss": 0.7016, + "step": 4604 + }, + { + "epoch": 0.3924158500213038, + "grad_norm": 16.0, + "learning_rate": 2.277700777012111e-06, + "loss": 0.91, + "step": 4605 + }, + { + "epoch": 0.39250106518960376, + "grad_norm": 15.1875, + "learning_rate": 2.277605112552257e-06, + "loss": 0.6765, + "step": 4606 + }, + { + "epoch": 0.3925862803579037, + "grad_norm": 11.25, + "learning_rate": 2.2775094295225452e-06, + "loss": 0.4097, + "step": 4607 + }, + { + "epoch": 0.3926714955262037, + "grad_norm": 16.875, + "learning_rate": 2.277413727924705e-06, + "loss": 0.9516, + "step": 4608 + }, + { + "epoch": 0.39275671069450363, + "grad_norm": 15.6875, + "learning_rate": 2.2773180077604657e-06, + "loss": 0.8189, + "step": 4609 + }, + { + "epoch": 0.3928419258628036, + "grad_norm": 14.1875, + "learning_rate": 2.2772222690315564e-06, + "loss": 0.5581, + "step": 4610 + }, + { + "epoch": 0.39292714103110354, + "grad_norm": 13.0625, + "learning_rate": 2.2771265117397083e-06, + "loss": 0.5962, + "step": 4611 + }, + { + "epoch": 0.3930123561994035, + "grad_norm": 21.125, + "learning_rate": 2.277030735886651e-06, + "loss": 0.6761, + "step": 4612 + }, + { + "epoch": 0.39309757136770346, + "grad_norm": 12.375, + "learning_rate": 2.276934941474116e-06, + "loss": 0.6512, + "step": 4613 + }, + { + "epoch": 0.3931827865360034, + "grad_norm": 18.375, + "learning_rate": 2.2768391285038334e-06, + "loss": 0.5349, + "step": 4614 + }, + { + "epoch": 0.39326800170430337, + "grad_norm": 17.5, + "learning_rate": 2.2767432969775354e-06, + "loss": 0.9304, + "step": 4615 + }, + { + "epoch": 0.39335321687260333, + "grad_norm": 17.375, + "learning_rate": 2.2766474468969536e-06, + "loss": 0.5558, + "step": 4616 + }, + { + "epoch": 0.3934384320409033, + "grad_norm": 30.875, + "learning_rate": 2.2765515782638197e-06, + "loss": 0.9864, + "step": 4617 + }, + { + "epoch": 0.39352364720920324, + "grad_norm": 13.0, + "learning_rate": 2.2764556910798668e-06, + "loss": 0.6642, + "step": 4618 + }, + { + "epoch": 0.3936088623775032, + "grad_norm": 16.5, + "learning_rate": 2.2763597853468274e-06, + "loss": 0.7602, + "step": 4619 + }, + { + "epoch": 0.39369407754580316, + "grad_norm": 21.875, + "learning_rate": 2.2762638610664343e-06, + "loss": 0.8022, + "step": 4620 + }, + { + "epoch": 0.3937792927141031, + "grad_norm": 14.5, + "learning_rate": 2.2761679182404216e-06, + "loss": 0.8596, + "step": 4621 + }, + { + "epoch": 0.39386450788240307, + "grad_norm": 15.625, + "learning_rate": 2.2760719568705225e-06, + "loss": 0.8398, + "step": 4622 + }, + { + "epoch": 0.393949723050703, + "grad_norm": 12.6875, + "learning_rate": 2.275975976958471e-06, + "loss": 0.7829, + "step": 4623 + }, + { + "epoch": 0.394034938219003, + "grad_norm": 11.5625, + "learning_rate": 2.2758799785060028e-06, + "loss": 0.4472, + "step": 4624 + }, + { + "epoch": 0.39412015338730294, + "grad_norm": 16.625, + "learning_rate": 2.2757839615148512e-06, + "loss": 0.8803, + "step": 4625 + }, + { + "epoch": 0.3942053685556029, + "grad_norm": 15.625, + "learning_rate": 2.275687925986752e-06, + "loss": 0.7726, + "step": 4626 + }, + { + "epoch": 0.39429058372390285, + "grad_norm": 10.25, + "learning_rate": 2.2755918719234405e-06, + "loss": 0.3376, + "step": 4627 + }, + { + "epoch": 0.3943757988922028, + "grad_norm": 14.8125, + "learning_rate": 2.275495799326653e-06, + "loss": 0.8363, + "step": 4628 + }, + { + "epoch": 0.39446101406050277, + "grad_norm": 19.625, + "learning_rate": 2.275399708198125e-06, + "loss": 1.1049, + "step": 4629 + }, + { + "epoch": 0.3945462292288027, + "grad_norm": 10.9375, + "learning_rate": 2.2753035985395928e-06, + "loss": 0.5373, + "step": 4630 + }, + { + "epoch": 0.3946314443971027, + "grad_norm": 16.75, + "learning_rate": 2.275207470352794e-06, + "loss": 0.8961, + "step": 4631 + }, + { + "epoch": 0.39471665956540264, + "grad_norm": 16.125, + "learning_rate": 2.2751113236394647e-06, + "loss": 0.7463, + "step": 4632 + }, + { + "epoch": 0.3948018747337026, + "grad_norm": 11.3125, + "learning_rate": 2.2750151584013432e-06, + "loss": 0.5158, + "step": 4633 + }, + { + "epoch": 0.39488708990200255, + "grad_norm": 16.0, + "learning_rate": 2.274918974640167e-06, + "loss": 0.7595, + "step": 4634 + }, + { + "epoch": 0.3949723050703025, + "grad_norm": 17.0, + "learning_rate": 2.2748227723576747e-06, + "loss": 1.0543, + "step": 4635 + }, + { + "epoch": 0.39505752023860247, + "grad_norm": 27.0, + "learning_rate": 2.274726551555604e-06, + "loss": 1.278, + "step": 4636 + }, + { + "epoch": 0.3951427354069024, + "grad_norm": 12.625, + "learning_rate": 2.2746303122356946e-06, + "loss": 0.6901, + "step": 4637 + }, + { + "epoch": 0.3952279505752024, + "grad_norm": 16.5, + "learning_rate": 2.2745340543996844e-06, + "loss": 0.9716, + "step": 4638 + }, + { + "epoch": 0.39531316574350234, + "grad_norm": 13.0, + "learning_rate": 2.2744377780493136e-06, + "loss": 0.6546, + "step": 4639 + }, + { + "epoch": 0.3953983809118023, + "grad_norm": 19.5, + "learning_rate": 2.2743414831863225e-06, + "loss": 1.1621, + "step": 4640 + }, + { + "epoch": 0.39548359608010225, + "grad_norm": 15.3125, + "learning_rate": 2.2742451698124507e-06, + "loss": 0.7983, + "step": 4641 + }, + { + "epoch": 0.3955688112484022, + "grad_norm": 13.75, + "learning_rate": 2.2741488379294386e-06, + "loss": 0.6898, + "step": 4642 + }, + { + "epoch": 0.39565402641670216, + "grad_norm": 14.3125, + "learning_rate": 2.274052487539027e-06, + "loss": 0.715, + "step": 4643 + }, + { + "epoch": 0.3957392415850021, + "grad_norm": 12.875, + "learning_rate": 2.273956118642957e-06, + "loss": 0.4977, + "step": 4644 + }, + { + "epoch": 0.3958244567533021, + "grad_norm": 19.25, + "learning_rate": 2.2738597312429706e-06, + "loss": 0.599, + "step": 4645 + }, + { + "epoch": 0.39590967192160204, + "grad_norm": 12.8125, + "learning_rate": 2.273763325340809e-06, + "loss": 0.5835, + "step": 4646 + }, + { + "epoch": 0.395994887089902, + "grad_norm": 10.9375, + "learning_rate": 2.273666900938215e-06, + "loss": 0.4101, + "step": 4647 + }, + { + "epoch": 0.39608010225820195, + "grad_norm": 11.8125, + "learning_rate": 2.2735704580369305e-06, + "loss": 0.5664, + "step": 4648 + }, + { + "epoch": 0.3961653174265019, + "grad_norm": 13.3125, + "learning_rate": 2.273473996638699e-06, + "loss": 0.3924, + "step": 4649 + }, + { + "epoch": 0.39625053259480186, + "grad_norm": 10.0625, + "learning_rate": 2.2733775167452627e-06, + "loss": 0.4271, + "step": 4650 + }, + { + "epoch": 0.3963357477631018, + "grad_norm": 20.625, + "learning_rate": 2.273281018358366e-06, + "loss": 1.0851, + "step": 4651 + }, + { + "epoch": 0.3964209629314018, + "grad_norm": 12.125, + "learning_rate": 2.2731845014797515e-06, + "loss": 0.5392, + "step": 4652 + }, + { + "epoch": 0.39650617809970173, + "grad_norm": 13.5625, + "learning_rate": 2.273087966111165e-06, + "loss": 0.6814, + "step": 4653 + }, + { + "epoch": 0.3965913932680017, + "grad_norm": 14.75, + "learning_rate": 2.27299141225435e-06, + "loss": 0.7036, + "step": 4654 + }, + { + "epoch": 0.39667660843630165, + "grad_norm": 21.375, + "learning_rate": 2.272894839911051e-06, + "loss": 0.9787, + "step": 4655 + }, + { + "epoch": 0.3967618236046016, + "grad_norm": 15.6875, + "learning_rate": 2.2727982490830143e-06, + "loss": 0.8725, + "step": 4656 + }, + { + "epoch": 0.39684703877290156, + "grad_norm": 16.125, + "learning_rate": 2.272701639771984e-06, + "loss": 1.0321, + "step": 4657 + }, + { + "epoch": 0.3969322539412015, + "grad_norm": 12.3125, + "learning_rate": 2.2726050119797072e-06, + "loss": 0.6422, + "step": 4658 + }, + { + "epoch": 0.3970174691095015, + "grad_norm": 12.6875, + "learning_rate": 2.2725083657079294e-06, + "loss": 0.5762, + "step": 4659 + }, + { + "epoch": 0.39710268427780143, + "grad_norm": 14.75, + "learning_rate": 2.272411700958397e-06, + "loss": 0.7513, + "step": 4660 + }, + { + "epoch": 0.3971878994461014, + "grad_norm": 18.125, + "learning_rate": 2.2723150177328575e-06, + "loss": 0.9375, + "step": 4661 + }, + { + "epoch": 0.39727311461440135, + "grad_norm": 16.25, + "learning_rate": 2.2722183160330576e-06, + "loss": 0.8393, + "step": 4662 + }, + { + "epoch": 0.3973583297827013, + "grad_norm": 17.25, + "learning_rate": 2.2721215958607446e-06, + "loss": 1.2914, + "step": 4663 + }, + { + "epoch": 0.39744354495100126, + "grad_norm": 11.75, + "learning_rate": 2.272024857217666e-06, + "loss": 0.6641, + "step": 4664 + }, + { + "epoch": 0.3975287601193012, + "grad_norm": 12.875, + "learning_rate": 2.2719281001055717e-06, + "loss": 0.6941, + "step": 4665 + }, + { + "epoch": 0.3976139752876012, + "grad_norm": 13.3125, + "learning_rate": 2.271831324526208e-06, + "loss": 0.7783, + "step": 4666 + }, + { + "epoch": 0.39769919045590113, + "grad_norm": 12.5625, + "learning_rate": 2.271734530481325e-06, + "loss": 0.7481, + "step": 4667 + }, + { + "epoch": 0.3977844056242011, + "grad_norm": 13.3125, + "learning_rate": 2.2716377179726716e-06, + "loss": 0.8422, + "step": 4668 + }, + { + "epoch": 0.39786962079250104, + "grad_norm": 18.375, + "learning_rate": 2.2715408870019977e-06, + "loss": 1.0709, + "step": 4669 + }, + { + "epoch": 0.397954835960801, + "grad_norm": 14.6875, + "learning_rate": 2.271444037571052e-06, + "loss": 0.7559, + "step": 4670 + }, + { + "epoch": 0.39804005112910096, + "grad_norm": 13.4375, + "learning_rate": 2.2713471696815863e-06, + "loss": 0.7677, + "step": 4671 + }, + { + "epoch": 0.3981252662974009, + "grad_norm": 12.0, + "learning_rate": 2.2712502833353496e-06, + "loss": 0.7987, + "step": 4672 + }, + { + "epoch": 0.39821048146570087, + "grad_norm": 9.5, + "learning_rate": 2.2711533785340933e-06, + "loss": 0.231, + "step": 4673 + }, + { + "epoch": 0.39829569663400083, + "grad_norm": 16.625, + "learning_rate": 2.2710564552795686e-06, + "loss": 0.8606, + "step": 4674 + }, + { + "epoch": 0.3983809118023008, + "grad_norm": 15.1875, + "learning_rate": 2.270959513573527e-06, + "loss": 0.8815, + "step": 4675 + }, + { + "epoch": 0.39846612697060074, + "grad_norm": 12.75, + "learning_rate": 2.270862553417721e-06, + "loss": 0.4789, + "step": 4676 + }, + { + "epoch": 0.3985513421389007, + "grad_norm": 14.1875, + "learning_rate": 2.2707655748139014e-06, + "loss": 0.7227, + "step": 4677 + }, + { + "epoch": 0.39863655730720066, + "grad_norm": 15.6875, + "learning_rate": 2.2706685777638217e-06, + "loss": 0.7824, + "step": 4678 + }, + { + "epoch": 0.3987217724755006, + "grad_norm": 19.75, + "learning_rate": 2.2705715622692344e-06, + "loss": 0.8465, + "step": 4679 + }, + { + "epoch": 0.39880698764380057, + "grad_norm": 15.4375, + "learning_rate": 2.2704745283318925e-06, + "loss": 0.7303, + "step": 4680 + }, + { + "epoch": 0.3988922028121005, + "grad_norm": 15.125, + "learning_rate": 2.2703774759535495e-06, + "loss": 0.8414, + "step": 4681 + }, + { + "epoch": 0.3989774179804005, + "grad_norm": 17.25, + "learning_rate": 2.27028040513596e-06, + "loss": 0.7061, + "step": 4682 + }, + { + "epoch": 0.3990626331487005, + "grad_norm": 14.75, + "learning_rate": 2.2701833158808774e-06, + "loss": 1.028, + "step": 4683 + }, + { + "epoch": 0.39914784831700045, + "grad_norm": 12.5625, + "learning_rate": 2.2700862081900564e-06, + "loss": 0.3536, + "step": 4684 + }, + { + "epoch": 0.3992330634853004, + "grad_norm": 12.25, + "learning_rate": 2.269989082065252e-06, + "loss": 0.5192, + "step": 4685 + }, + { + "epoch": 0.39931827865360037, + "grad_norm": 16.5, + "learning_rate": 2.269891937508219e-06, + "loss": 0.7274, + "step": 4686 + }, + { + "epoch": 0.3994034938219003, + "grad_norm": 12.1875, + "learning_rate": 2.269794774520713e-06, + "loss": 0.6608, + "step": 4687 + }, + { + "epoch": 0.3994887089902003, + "grad_norm": 19.75, + "learning_rate": 2.26969759310449e-06, + "loss": 0.9892, + "step": 4688 + }, + { + "epoch": 0.39957392415850024, + "grad_norm": 18.25, + "learning_rate": 2.269600393261306e-06, + "loss": 0.7239, + "step": 4689 + }, + { + "epoch": 0.3996591393268002, + "grad_norm": 15.5625, + "learning_rate": 2.269503174992918e-06, + "loss": 0.7861, + "step": 4690 + }, + { + "epoch": 0.39974435449510015, + "grad_norm": 8.625, + "learning_rate": 2.269405938301082e-06, + "loss": 0.3024, + "step": 4691 + }, + { + "epoch": 0.3998295696634001, + "grad_norm": 12.875, + "learning_rate": 2.269308683187556e-06, + "loss": 0.8666, + "step": 4692 + }, + { + "epoch": 0.39991478483170007, + "grad_norm": 35.5, + "learning_rate": 2.2692114096540966e-06, + "loss": 1.2893, + "step": 4693 + }, + { + "epoch": 0.4, + "grad_norm": 14.75, + "learning_rate": 2.269114117702463e-06, + "loss": 0.9085, + "step": 4694 + }, + { + "epoch": 0.4000852151683, + "grad_norm": 14.9375, + "learning_rate": 2.269016807334412e-06, + "loss": 0.967, + "step": 4695 + }, + { + "epoch": 0.40017043033659994, + "grad_norm": 12.4375, + "learning_rate": 2.2689194785517023e-06, + "loss": 0.7117, + "step": 4696 + }, + { + "epoch": 0.4002556455048999, + "grad_norm": 15.4375, + "learning_rate": 2.268822131356093e-06, + "loss": 0.9099, + "step": 4697 + }, + { + "epoch": 0.40034086067319985, + "grad_norm": 14.0625, + "learning_rate": 2.2687247657493437e-06, + "loss": 0.5604, + "step": 4698 + }, + { + "epoch": 0.4004260758414998, + "grad_norm": 12.625, + "learning_rate": 2.2686273817332136e-06, + "loss": 0.3514, + "step": 4699 + }, + { + "epoch": 0.40051129100979976, + "grad_norm": 18.5, + "learning_rate": 2.268529979309462e-06, + "loss": 0.7446, + "step": 4700 + }, + { + "epoch": 0.4005965061780997, + "grad_norm": 16.0, + "learning_rate": 2.2684325584798496e-06, + "loss": 0.9357, + "step": 4701 + }, + { + "epoch": 0.4006817213463997, + "grad_norm": 13.375, + "learning_rate": 2.268335119246137e-06, + "loss": 0.587, + "step": 4702 + }, + { + "epoch": 0.40076693651469963, + "grad_norm": 13.3125, + "learning_rate": 2.2682376616100842e-06, + "loss": 0.8557, + "step": 4703 + }, + { + "epoch": 0.4008521516829996, + "grad_norm": 13.1875, + "learning_rate": 2.2681401855734535e-06, + "loss": 0.6055, + "step": 4704 + }, + { + "epoch": 0.40093736685129955, + "grad_norm": 11.125, + "learning_rate": 2.2680426911380055e-06, + "loss": 0.4985, + "step": 4705 + }, + { + "epoch": 0.4010225820195995, + "grad_norm": 10.8125, + "learning_rate": 2.2679451783055025e-06, + "loss": 0.4013, + "step": 4706 + }, + { + "epoch": 0.40110779718789946, + "grad_norm": 16.5, + "learning_rate": 2.2678476470777067e-06, + "loss": 0.9624, + "step": 4707 + }, + { + "epoch": 0.4011930123561994, + "grad_norm": 13.9375, + "learning_rate": 2.26775009745638e-06, + "loss": 0.7059, + "step": 4708 + }, + { + "epoch": 0.4012782275244994, + "grad_norm": 14.3125, + "learning_rate": 2.2676525294432853e-06, + "loss": 0.8862, + "step": 4709 + }, + { + "epoch": 0.40136344269279933, + "grad_norm": 21.875, + "learning_rate": 2.267554943040187e-06, + "loss": 1.2246, + "step": 4710 + }, + { + "epoch": 0.4014486578610993, + "grad_norm": 12.25, + "learning_rate": 2.267457338248847e-06, + "loss": 0.5507, + "step": 4711 + }, + { + "epoch": 0.40153387302939925, + "grad_norm": 18.625, + "learning_rate": 2.2673597150710302e-06, + "loss": 0.7336, + "step": 4712 + }, + { + "epoch": 0.4016190881976992, + "grad_norm": 18.125, + "learning_rate": 2.2672620735085e-06, + "loss": 0.8719, + "step": 4713 + }, + { + "epoch": 0.40170430336599916, + "grad_norm": 19.125, + "learning_rate": 2.267164413563021e-06, + "loss": 0.7144, + "step": 4714 + }, + { + "epoch": 0.4017895185342991, + "grad_norm": 12.8125, + "learning_rate": 2.2670667352363586e-06, + "loss": 0.6919, + "step": 4715 + }, + { + "epoch": 0.4018747337025991, + "grad_norm": 17.375, + "learning_rate": 2.2669690385302775e-06, + "loss": 0.9576, + "step": 4716 + }, + { + "epoch": 0.40195994887089903, + "grad_norm": 16.875, + "learning_rate": 2.266871323446543e-06, + "loss": 0.5785, + "step": 4717 + }, + { + "epoch": 0.402045164039199, + "grad_norm": 11.6875, + "learning_rate": 2.2667735899869216e-06, + "loss": 0.6427, + "step": 4718 + }, + { + "epoch": 0.40213037920749894, + "grad_norm": 10.5, + "learning_rate": 2.2666758381531783e-06, + "loss": 0.3087, + "step": 4719 + }, + { + "epoch": 0.4022155943757989, + "grad_norm": 10.625, + "learning_rate": 2.266578067947081e-06, + "loss": 0.4566, + "step": 4720 + }, + { + "epoch": 0.40230080954409886, + "grad_norm": 11.875, + "learning_rate": 2.266480279370395e-06, + "loss": 0.5908, + "step": 4721 + }, + { + "epoch": 0.4023860247123988, + "grad_norm": 12.1875, + "learning_rate": 2.266382472424889e-06, + "loss": 0.4141, + "step": 4722 + }, + { + "epoch": 0.4024712398806988, + "grad_norm": 23.0, + "learning_rate": 2.266284647112329e-06, + "loss": 0.9698, + "step": 4723 + }, + { + "epoch": 0.40255645504899873, + "grad_norm": 15.625, + "learning_rate": 2.266186803434484e-06, + "loss": 0.8946, + "step": 4724 + }, + { + "epoch": 0.4026416702172987, + "grad_norm": 14.75, + "learning_rate": 2.2660889413931213e-06, + "loss": 0.8829, + "step": 4725 + }, + { + "epoch": 0.40272688538559864, + "grad_norm": 16.625, + "learning_rate": 2.2659910609900097e-06, + "loss": 0.8776, + "step": 4726 + }, + { + "epoch": 0.4028121005538986, + "grad_norm": 13.625, + "learning_rate": 2.2658931622269177e-06, + "loss": 0.8368, + "step": 4727 + }, + { + "epoch": 0.40289731572219856, + "grad_norm": 10.625, + "learning_rate": 2.2657952451056153e-06, + "loss": 0.3565, + "step": 4728 + }, + { + "epoch": 0.4029825308904985, + "grad_norm": 14.5625, + "learning_rate": 2.265697309627871e-06, + "loss": 0.7575, + "step": 4729 + }, + { + "epoch": 0.40306774605879847, + "grad_norm": 15.3125, + "learning_rate": 2.265599355795455e-06, + "loss": 0.8153, + "step": 4730 + }, + { + "epoch": 0.4031529612270984, + "grad_norm": 11.5, + "learning_rate": 2.2655013836101375e-06, + "loss": 0.4993, + "step": 4731 + }, + { + "epoch": 0.4032381763953984, + "grad_norm": 15.4375, + "learning_rate": 2.2654033930736883e-06, + "loss": 1.1303, + "step": 4732 + }, + { + "epoch": 0.40332339156369834, + "grad_norm": 13.625, + "learning_rate": 2.2653053841878793e-06, + "loss": 0.8431, + "step": 4733 + }, + { + "epoch": 0.4034086067319983, + "grad_norm": 16.0, + "learning_rate": 2.265207356954481e-06, + "loss": 0.5506, + "step": 4734 + }, + { + "epoch": 0.40349382190029826, + "grad_norm": 12.6875, + "learning_rate": 2.265109311375265e-06, + "loss": 0.6505, + "step": 4735 + }, + { + "epoch": 0.4035790370685982, + "grad_norm": 15.625, + "learning_rate": 2.2650112474520024e-06, + "loss": 0.924, + "step": 4736 + }, + { + "epoch": 0.40366425223689817, + "grad_norm": 17.25, + "learning_rate": 2.2649131651864666e-06, + "loss": 0.6731, + "step": 4737 + }, + { + "epoch": 0.4037494674051981, + "grad_norm": 15.6875, + "learning_rate": 2.264815064580429e-06, + "loss": 0.6855, + "step": 4738 + }, + { + "epoch": 0.4038346825734981, + "grad_norm": 12.125, + "learning_rate": 2.2647169456356623e-06, + "loss": 0.4179, + "step": 4739 + }, + { + "epoch": 0.40391989774179804, + "grad_norm": 13.5625, + "learning_rate": 2.2646188083539407e-06, + "loss": 0.8236, + "step": 4740 + }, + { + "epoch": 0.404005112910098, + "grad_norm": 16.75, + "learning_rate": 2.264520652737036e-06, + "loss": 0.8586, + "step": 4741 + }, + { + "epoch": 0.40409032807839795, + "grad_norm": 19.5, + "learning_rate": 2.264422478786724e-06, + "loss": 0.7602, + "step": 4742 + }, + { + "epoch": 0.4041755432466979, + "grad_norm": 13.25, + "learning_rate": 2.2643242865047773e-06, + "loss": 0.628, + "step": 4743 + }, + { + "epoch": 0.40426075841499787, + "grad_norm": 17.625, + "learning_rate": 2.2642260758929706e-06, + "loss": 0.9205, + "step": 4744 + }, + { + "epoch": 0.4043459735832978, + "grad_norm": 10.0, + "learning_rate": 2.264127846953079e-06, + "loss": 0.2715, + "step": 4745 + }, + { + "epoch": 0.4044311887515978, + "grad_norm": 12.0625, + "learning_rate": 2.2640295996868774e-06, + "loss": 0.4329, + "step": 4746 + }, + { + "epoch": 0.40451640391989774, + "grad_norm": 9.9375, + "learning_rate": 2.263931334096141e-06, + "loss": 0.2412, + "step": 4747 + }, + { + "epoch": 0.4046016190881977, + "grad_norm": 19.5, + "learning_rate": 2.263833050182646e-06, + "loss": 1.1018, + "step": 4748 + }, + { + "epoch": 0.40468683425649765, + "grad_norm": 13.6875, + "learning_rate": 2.2637347479481687e-06, + "loss": 0.7363, + "step": 4749 + }, + { + "epoch": 0.4047720494247976, + "grad_norm": 14.0, + "learning_rate": 2.2636364273944848e-06, + "loss": 0.9824, + "step": 4750 + }, + { + "epoch": 0.40485726459309757, + "grad_norm": 21.625, + "learning_rate": 2.263538088523371e-06, + "loss": 1.0033, + "step": 4751 + }, + { + "epoch": 0.4049424797613975, + "grad_norm": 12.8125, + "learning_rate": 2.2634397313366054e-06, + "loss": 0.6889, + "step": 4752 + }, + { + "epoch": 0.4050276949296975, + "grad_norm": 16.625, + "learning_rate": 2.263341355835964e-06, + "loss": 0.677, + "step": 4753 + }, + { + "epoch": 0.40511291009799744, + "grad_norm": 22.375, + "learning_rate": 2.263242962023226e-06, + "loss": 1.0805, + "step": 4754 + }, + { + "epoch": 0.4051981252662974, + "grad_norm": 20.875, + "learning_rate": 2.2631445499001687e-06, + "loss": 1.0173, + "step": 4755 + }, + { + "epoch": 0.40528334043459735, + "grad_norm": 18.125, + "learning_rate": 2.2630461194685705e-06, + "loss": 0.6526, + "step": 4756 + }, + { + "epoch": 0.4053685556028973, + "grad_norm": 26.375, + "learning_rate": 2.26294767073021e-06, + "loss": 0.8009, + "step": 4757 + }, + { + "epoch": 0.40545377077119726, + "grad_norm": 16.125, + "learning_rate": 2.2628492036868667e-06, + "loss": 0.876, + "step": 4758 + }, + { + "epoch": 0.4055389859394972, + "grad_norm": 18.5, + "learning_rate": 2.2627507183403195e-06, + "loss": 0.5533, + "step": 4759 + }, + { + "epoch": 0.4056242011077972, + "grad_norm": 10.0, + "learning_rate": 2.2626522146923487e-06, + "loss": 0.3191, + "step": 4760 + }, + { + "epoch": 0.40570941627609713, + "grad_norm": 17.75, + "learning_rate": 2.2625536927447337e-06, + "loss": 1.0784, + "step": 4761 + }, + { + "epoch": 0.4057946314443971, + "grad_norm": 13.25, + "learning_rate": 2.262455152499256e-06, + "loss": 0.793, + "step": 4762 + }, + { + "epoch": 0.40587984661269705, + "grad_norm": 18.5, + "learning_rate": 2.262356593957695e-06, + "loss": 0.9842, + "step": 4763 + }, + { + "epoch": 0.405965061780997, + "grad_norm": 12.4375, + "learning_rate": 2.2622580171218325e-06, + "loss": 0.7935, + "step": 4764 + }, + { + "epoch": 0.40605027694929696, + "grad_norm": 12.8125, + "learning_rate": 2.26215942199345e-06, + "loss": 0.4932, + "step": 4765 + }, + { + "epoch": 0.4061354921175969, + "grad_norm": 11.8125, + "learning_rate": 2.2620608085743286e-06, + "loss": 0.5764, + "step": 4766 + }, + { + "epoch": 0.4062207072858969, + "grad_norm": 11.1875, + "learning_rate": 2.261962176866251e-06, + "loss": 0.5877, + "step": 4767 + }, + { + "epoch": 0.40630592245419683, + "grad_norm": 17.875, + "learning_rate": 2.261863526870999e-06, + "loss": 0.8367, + "step": 4768 + }, + { + "epoch": 0.4063911376224968, + "grad_norm": 13.25, + "learning_rate": 2.2617648585903556e-06, + "loss": 0.853, + "step": 4769 + }, + { + "epoch": 0.40647635279079675, + "grad_norm": 13.625, + "learning_rate": 2.2616661720261034e-06, + "loss": 0.7978, + "step": 4770 + }, + { + "epoch": 0.4065615679590967, + "grad_norm": 11.75, + "learning_rate": 2.2615674671800266e-06, + "loss": 0.4212, + "step": 4771 + }, + { + "epoch": 0.40664678312739666, + "grad_norm": 15.0, + "learning_rate": 2.2614687440539085e-06, + "loss": 0.6508, + "step": 4772 + }, + { + "epoch": 0.4067319982956966, + "grad_norm": 23.0, + "learning_rate": 2.261370002649533e-06, + "loss": 0.9489, + "step": 4773 + }, + { + "epoch": 0.4068172134639966, + "grad_norm": 11.6875, + "learning_rate": 2.2612712429686846e-06, + "loss": 0.4592, + "step": 4774 + }, + { + "epoch": 0.40690242863229653, + "grad_norm": 13.75, + "learning_rate": 2.2611724650131477e-06, + "loss": 0.7842, + "step": 4775 + }, + { + "epoch": 0.4069876438005965, + "grad_norm": 11.4375, + "learning_rate": 2.261073668784708e-06, + "loss": 0.4253, + "step": 4776 + }, + { + "epoch": 0.40707285896889644, + "grad_norm": 15.3125, + "learning_rate": 2.26097485428515e-06, + "loss": 0.7862, + "step": 4777 + }, + { + "epoch": 0.4071580741371964, + "grad_norm": 13.875, + "learning_rate": 2.2608760215162594e-06, + "loss": 0.8884, + "step": 4778 + }, + { + "epoch": 0.40724328930549636, + "grad_norm": 10.9375, + "learning_rate": 2.2607771704798234e-06, + "loss": 0.4183, + "step": 4779 + }, + { + "epoch": 0.4073285044737963, + "grad_norm": 15.5, + "learning_rate": 2.2606783011776272e-06, + "loss": 0.6357, + "step": 4780 + }, + { + "epoch": 0.4074137196420963, + "grad_norm": 15.375, + "learning_rate": 2.260579413611458e-06, + "loss": 0.8026, + "step": 4781 + }, + { + "epoch": 0.40749893481039623, + "grad_norm": 18.0, + "learning_rate": 2.260480507783102e-06, + "loss": 0.8144, + "step": 4782 + }, + { + "epoch": 0.4075841499786962, + "grad_norm": 14.125, + "learning_rate": 2.2603815836943474e-06, + "loss": 0.4435, + "step": 4783 + }, + { + "epoch": 0.40766936514699614, + "grad_norm": 24.25, + "learning_rate": 2.2602826413469813e-06, + "loss": 0.9548, + "step": 4784 + }, + { + "epoch": 0.4077545803152961, + "grad_norm": 24.0, + "learning_rate": 2.2601836807427925e-06, + "loss": 0.9395, + "step": 4785 + }, + { + "epoch": 0.40783979548359606, + "grad_norm": 14.1875, + "learning_rate": 2.2600847018835685e-06, + "loss": 0.9131, + "step": 4786 + }, + { + "epoch": 0.407925010651896, + "grad_norm": 19.125, + "learning_rate": 2.2599857047710983e-06, + "loss": 1.0284, + "step": 4787 + }, + { + "epoch": 0.40801022582019597, + "grad_norm": 13.125, + "learning_rate": 2.259886689407171e-06, + "loss": 0.718, + "step": 4788 + }, + { + "epoch": 0.4080954409884959, + "grad_norm": 18.75, + "learning_rate": 2.259787655793575e-06, + "loss": 1.0749, + "step": 4789 + }, + { + "epoch": 0.4081806561567959, + "grad_norm": 18.125, + "learning_rate": 2.2596886039321013e-06, + "loss": 0.8422, + "step": 4790 + }, + { + "epoch": 0.40826587132509584, + "grad_norm": 12.125, + "learning_rate": 2.2595895338245387e-06, + "loss": 0.5409, + "step": 4791 + }, + { + "epoch": 0.4083510864933958, + "grad_norm": 17.125, + "learning_rate": 2.2594904454726784e-06, + "loss": 1.057, + "step": 4792 + }, + { + "epoch": 0.40843630166169576, + "grad_norm": 14.75, + "learning_rate": 2.2593913388783105e-06, + "loss": 0.5485, + "step": 4793 + }, + { + "epoch": 0.4085215168299957, + "grad_norm": 15.5625, + "learning_rate": 2.2592922140432262e-06, + "loss": 0.9442, + "step": 4794 + }, + { + "epoch": 0.40860673199829567, + "grad_norm": 19.5, + "learning_rate": 2.2591930709692166e-06, + "loss": 0.7642, + "step": 4795 + }, + { + "epoch": 0.4086919471665957, + "grad_norm": 16.75, + "learning_rate": 2.2590939096580734e-06, + "loss": 0.6897, + "step": 4796 + }, + { + "epoch": 0.40877716233489564, + "grad_norm": 15.1875, + "learning_rate": 2.258994730111588e-06, + "loss": 0.5593, + "step": 4797 + }, + { + "epoch": 0.4088623775031956, + "grad_norm": 12.25, + "learning_rate": 2.258895532331554e-06, + "loss": 0.506, + "step": 4798 + }, + { + "epoch": 0.40894759267149555, + "grad_norm": 12.6875, + "learning_rate": 2.258796316319763e-06, + "loss": 0.6796, + "step": 4799 + }, + { + "epoch": 0.4090328078397955, + "grad_norm": 14.5625, + "learning_rate": 2.258697082078008e-06, + "loss": 1.0086, + "step": 4800 + }, + { + "epoch": 0.40911802300809547, + "grad_norm": 18.25, + "learning_rate": 2.2585978296080823e-06, + "loss": 1.0874, + "step": 4801 + }, + { + "epoch": 0.4092032381763954, + "grad_norm": 14.8125, + "learning_rate": 2.25849855891178e-06, + "loss": 0.649, + "step": 4802 + }, + { + "epoch": 0.4092884533446954, + "grad_norm": 12.5625, + "learning_rate": 2.2583992699908943e-06, + "loss": 0.7, + "step": 4803 + }, + { + "epoch": 0.40937366851299534, + "grad_norm": 13.9375, + "learning_rate": 2.25829996284722e-06, + "loss": 0.6285, + "step": 4804 + }, + { + "epoch": 0.4094588836812953, + "grad_norm": 18.75, + "learning_rate": 2.258200637482551e-06, + "loss": 0.9545, + "step": 4805 + }, + { + "epoch": 0.40954409884959525, + "grad_norm": 14.5625, + "learning_rate": 2.258101293898683e-06, + "loss": 0.8332, + "step": 4806 + }, + { + "epoch": 0.4096293140178952, + "grad_norm": 14.3125, + "learning_rate": 2.258001932097411e-06, + "loss": 0.9381, + "step": 4807 + }, + { + "epoch": 0.40971452918619516, + "grad_norm": 18.5, + "learning_rate": 2.2579025520805303e-06, + "loss": 0.8526, + "step": 4808 + }, + { + "epoch": 0.4097997443544951, + "grad_norm": 15.9375, + "learning_rate": 2.2578031538498367e-06, + "loss": 0.7659, + "step": 4809 + }, + { + "epoch": 0.4098849595227951, + "grad_norm": 14.125, + "learning_rate": 2.257703737407127e-06, + "loss": 0.7903, + "step": 4810 + }, + { + "epoch": 0.40997017469109504, + "grad_norm": 18.125, + "learning_rate": 2.2576043027541972e-06, + "loss": 0.879, + "step": 4811 + }, + { + "epoch": 0.410055389859395, + "grad_norm": 13.0625, + "learning_rate": 2.257504849892845e-06, + "loss": 0.3422, + "step": 4812 + }, + { + "epoch": 0.41014060502769495, + "grad_norm": 13.4375, + "learning_rate": 2.257405378824866e-06, + "loss": 0.7472, + "step": 4813 + }, + { + "epoch": 0.4102258201959949, + "grad_norm": 14.4375, + "learning_rate": 2.25730588955206e-06, + "loss": 0.6635, + "step": 4814 + }, + { + "epoch": 0.41031103536429486, + "grad_norm": 15.5, + "learning_rate": 2.257206382076223e-06, + "loss": 0.7449, + "step": 4815 + }, + { + "epoch": 0.4103962505325948, + "grad_norm": 12.5625, + "learning_rate": 2.257106856399154e-06, + "loss": 0.6581, + "step": 4816 + }, + { + "epoch": 0.4104814657008948, + "grad_norm": 11.0625, + "learning_rate": 2.257007312522651e-06, + "loss": 0.5158, + "step": 4817 + }, + { + "epoch": 0.41056668086919473, + "grad_norm": 14.1875, + "learning_rate": 2.2569077504485136e-06, + "loss": 0.7873, + "step": 4818 + }, + { + "epoch": 0.4106518960374947, + "grad_norm": 10.4375, + "learning_rate": 2.2568081701785405e-06, + "loss": 0.4938, + "step": 4819 + }, + { + "epoch": 0.41073711120579465, + "grad_norm": 17.875, + "learning_rate": 2.2567085717145314e-06, + "loss": 0.675, + "step": 4820 + }, + { + "epoch": 0.4108223263740946, + "grad_norm": 11.0, + "learning_rate": 2.256608955058286e-06, + "loss": 0.4564, + "step": 4821 + }, + { + "epoch": 0.41090754154239456, + "grad_norm": 20.125, + "learning_rate": 2.2565093202116048e-06, + "loss": 0.6711, + "step": 4822 + }, + { + "epoch": 0.4109927567106945, + "grad_norm": 18.625, + "learning_rate": 2.256409667176288e-06, + "loss": 0.9432, + "step": 4823 + }, + { + "epoch": 0.4110779718789945, + "grad_norm": 16.0, + "learning_rate": 2.2563099959541363e-06, + "loss": 0.6405, + "step": 4824 + }, + { + "epoch": 0.41116318704729443, + "grad_norm": 14.4375, + "learning_rate": 2.256210306546951e-06, + "loss": 0.5645, + "step": 4825 + }, + { + "epoch": 0.4112484022155944, + "grad_norm": 18.625, + "learning_rate": 2.256110598956534e-06, + "loss": 1.231, + "step": 4826 + }, + { + "epoch": 0.41133361738389435, + "grad_norm": 25.125, + "learning_rate": 2.2560108731846862e-06, + "loss": 0.8709, + "step": 4827 + }, + { + "epoch": 0.4114188325521943, + "grad_norm": 20.0, + "learning_rate": 2.255911129233211e-06, + "loss": 0.9058, + "step": 4828 + }, + { + "epoch": 0.41150404772049426, + "grad_norm": 16.625, + "learning_rate": 2.2558113671039097e-06, + "loss": 0.9308, + "step": 4829 + }, + { + "epoch": 0.4115892628887942, + "grad_norm": 11.25, + "learning_rate": 2.255711586798586e-06, + "loss": 0.5423, + "step": 4830 + }, + { + "epoch": 0.4116744780570942, + "grad_norm": 16.25, + "learning_rate": 2.255611788319042e-06, + "loss": 1.1954, + "step": 4831 + }, + { + "epoch": 0.41175969322539413, + "grad_norm": 20.625, + "learning_rate": 2.255511971667082e-06, + "loss": 1.227, + "step": 4832 + }, + { + "epoch": 0.4118449083936941, + "grad_norm": 13.125, + "learning_rate": 2.2554121368445097e-06, + "loss": 0.8236, + "step": 4833 + }, + { + "epoch": 0.41193012356199404, + "grad_norm": 17.375, + "learning_rate": 2.2553122838531286e-06, + "loss": 1.2031, + "step": 4834 + }, + { + "epoch": 0.412015338730294, + "grad_norm": 18.0, + "learning_rate": 2.2552124126947446e-06, + "loss": 1.0453, + "step": 4835 + }, + { + "epoch": 0.41210055389859396, + "grad_norm": 15.0, + "learning_rate": 2.2551125233711605e-06, + "loss": 0.7285, + "step": 4836 + }, + { + "epoch": 0.4121857690668939, + "grad_norm": 14.0625, + "learning_rate": 2.2550126158841827e-06, + "loss": 0.8544, + "step": 4837 + }, + { + "epoch": 0.41227098423519387, + "grad_norm": 20.25, + "learning_rate": 2.2549126902356165e-06, + "loss": 0.985, + "step": 4838 + }, + { + "epoch": 0.41235619940349383, + "grad_norm": 15.0625, + "learning_rate": 2.2548127464272674e-06, + "loss": 0.6943, + "step": 4839 + }, + { + "epoch": 0.4124414145717938, + "grad_norm": 15.25, + "learning_rate": 2.2547127844609413e-06, + "loss": 0.8087, + "step": 4840 + }, + { + "epoch": 0.41252662974009374, + "grad_norm": 70.5, + "learning_rate": 2.2546128043384454e-06, + "loss": 0.5554, + "step": 4841 + }, + { + "epoch": 0.4126118449083937, + "grad_norm": 11.875, + "learning_rate": 2.2545128060615855e-06, + "loss": 0.6621, + "step": 4842 + }, + { + "epoch": 0.41269706007669366, + "grad_norm": 13.4375, + "learning_rate": 2.2544127896321687e-06, + "loss": 0.691, + "step": 4843 + }, + { + "epoch": 0.4127822752449936, + "grad_norm": 17.375, + "learning_rate": 2.254312755052003e-06, + "loss": 0.6695, + "step": 4844 + }, + { + "epoch": 0.41286749041329357, + "grad_norm": 11.75, + "learning_rate": 2.2542127023228967e-06, + "loss": 0.4374, + "step": 4845 + }, + { + "epoch": 0.4129527055815935, + "grad_norm": 24.125, + "learning_rate": 2.2541126314466565e-06, + "loss": 0.8464, + "step": 4846 + }, + { + "epoch": 0.4130379207498935, + "grad_norm": 23.125, + "learning_rate": 2.2540125424250914e-06, + "loss": 0.702, + "step": 4847 + }, + { + "epoch": 0.41312313591819344, + "grad_norm": 13.4375, + "learning_rate": 2.25391243526001e-06, + "loss": 0.7888, + "step": 4848 + }, + { + "epoch": 0.4132083510864934, + "grad_norm": 14.875, + "learning_rate": 2.2538123099532212e-06, + "loss": 0.9714, + "step": 4849 + }, + { + "epoch": 0.41329356625479335, + "grad_norm": 17.0, + "learning_rate": 2.2537121665065346e-06, + "loss": 0.2549, + "step": 4850 + }, + { + "epoch": 0.4133787814230933, + "grad_norm": 10.0, + "learning_rate": 2.25361200492176e-06, + "loss": 0.4591, + "step": 4851 + }, + { + "epoch": 0.41346399659139327, + "grad_norm": 15.5, + "learning_rate": 2.253511825200707e-06, + "loss": 0.7933, + "step": 4852 + }, + { + "epoch": 0.4135492117596932, + "grad_norm": 13.9375, + "learning_rate": 2.2534116273451865e-06, + "loss": 0.8653, + "step": 4853 + }, + { + "epoch": 0.4136344269279932, + "grad_norm": 15.3125, + "learning_rate": 2.253311411357009e-06, + "loss": 0.7351, + "step": 4854 + }, + { + "epoch": 0.41371964209629314, + "grad_norm": 16.75, + "learning_rate": 2.253211177237985e-06, + "loss": 0.8661, + "step": 4855 + }, + { + "epoch": 0.4138048572645931, + "grad_norm": 21.125, + "learning_rate": 2.2531109249899265e-06, + "loss": 0.9118, + "step": 4856 + }, + { + "epoch": 0.41389007243289305, + "grad_norm": 15.5, + "learning_rate": 2.2530106546146446e-06, + "loss": 0.6502, + "step": 4857 + }, + { + "epoch": 0.413975287601193, + "grad_norm": 13.75, + "learning_rate": 2.252910366113952e-06, + "loss": 0.8105, + "step": 4858 + }, + { + "epoch": 0.41406050276949297, + "grad_norm": 16.0, + "learning_rate": 2.25281005948966e-06, + "loss": 1.1773, + "step": 4859 + }, + { + "epoch": 0.4141457179377929, + "grad_norm": 33.5, + "learning_rate": 2.2527097347435823e-06, + "loss": 1.5125, + "step": 4860 + }, + { + "epoch": 0.4142309331060929, + "grad_norm": 16.75, + "learning_rate": 2.2526093918775313e-06, + "loss": 0.7659, + "step": 4861 + }, + { + "epoch": 0.41431614827439284, + "grad_norm": 12.5, + "learning_rate": 2.25250903089332e-06, + "loss": 0.5829, + "step": 4862 + }, + { + "epoch": 0.4144013634426928, + "grad_norm": 12.125, + "learning_rate": 2.252408651792763e-06, + "loss": 0.533, + "step": 4863 + }, + { + "epoch": 0.41448657861099275, + "grad_norm": 23.25, + "learning_rate": 2.252308254577673e-06, + "loss": 0.9993, + "step": 4864 + }, + { + "epoch": 0.4145717937792927, + "grad_norm": 12.625, + "learning_rate": 2.252207839249866e-06, + "loss": 0.3742, + "step": 4865 + }, + { + "epoch": 0.41465700894759266, + "grad_norm": 11.25, + "learning_rate": 2.252107405811155e-06, + "loss": 0.4757, + "step": 4866 + }, + { + "epoch": 0.4147422241158926, + "grad_norm": 14.9375, + "learning_rate": 2.252006954263355e-06, + "loss": 0.7955, + "step": 4867 + }, + { + "epoch": 0.4148274392841926, + "grad_norm": 11.9375, + "learning_rate": 2.251906484608282e-06, + "loss": 0.5482, + "step": 4868 + }, + { + "epoch": 0.41491265445249254, + "grad_norm": 15.9375, + "learning_rate": 2.2518059968477516e-06, + "loss": 0.8476, + "step": 4869 + }, + { + "epoch": 0.4149978696207925, + "grad_norm": 10.9375, + "learning_rate": 2.251705490983579e-06, + "loss": 0.3362, + "step": 4870 + }, + { + "epoch": 0.41508308478909245, + "grad_norm": 18.375, + "learning_rate": 2.251604967017581e-06, + "loss": 0.7088, + "step": 4871 + }, + { + "epoch": 0.4151682999573924, + "grad_norm": 13.125, + "learning_rate": 2.2515044249515745e-06, + "loss": 0.6508, + "step": 4872 + }, + { + "epoch": 0.41525351512569236, + "grad_norm": 15.375, + "learning_rate": 2.2514038647873758e-06, + "loss": 1.0892, + "step": 4873 + }, + { + "epoch": 0.4153387302939923, + "grad_norm": 13.5625, + "learning_rate": 2.2513032865268018e-06, + "loss": 0.8347, + "step": 4874 + }, + { + "epoch": 0.4154239454622923, + "grad_norm": 15.625, + "learning_rate": 2.251202690171671e-06, + "loss": 0.667, + "step": 4875 + }, + { + "epoch": 0.41550916063059223, + "grad_norm": 14.0625, + "learning_rate": 2.251102075723801e-06, + "loss": 0.5272, + "step": 4876 + }, + { + "epoch": 0.4155943757988922, + "grad_norm": 12.1875, + "learning_rate": 2.2510014431850093e-06, + "loss": 0.351, + "step": 4877 + }, + { + "epoch": 0.41567959096719215, + "grad_norm": 19.0, + "learning_rate": 2.2509007925571154e-06, + "loss": 1.1781, + "step": 4878 + }, + { + "epoch": 0.4157648061354921, + "grad_norm": 15.375, + "learning_rate": 2.250800123841937e-06, + "loss": 0.5345, + "step": 4879 + }, + { + "epoch": 0.41585002130379206, + "grad_norm": 14.0625, + "learning_rate": 2.2506994370412944e-06, + "loss": 0.5204, + "step": 4880 + }, + { + "epoch": 0.415935236472092, + "grad_norm": 17.0, + "learning_rate": 2.2505987321570066e-06, + "loss": 0.6551, + "step": 4881 + }, + { + "epoch": 0.416020451640392, + "grad_norm": 11.5, + "learning_rate": 2.250498009190894e-06, + "loss": 0.6634, + "step": 4882 + }, + { + "epoch": 0.41610566680869193, + "grad_norm": 14.0, + "learning_rate": 2.2503972681447756e-06, + "loss": 0.4377, + "step": 4883 + }, + { + "epoch": 0.4161908819769919, + "grad_norm": 15.125, + "learning_rate": 2.250296509020473e-06, + "loss": 0.8151, + "step": 4884 + }, + { + "epoch": 0.41627609714529185, + "grad_norm": 14.5625, + "learning_rate": 2.2501957318198063e-06, + "loss": 0.8733, + "step": 4885 + }, + { + "epoch": 0.4163613123135918, + "grad_norm": 15.375, + "learning_rate": 2.2500949365445974e-06, + "loss": 0.6452, + "step": 4886 + }, + { + "epoch": 0.41644652748189176, + "grad_norm": 16.375, + "learning_rate": 2.249994123196667e-06, + "loss": 0.9311, + "step": 4887 + }, + { + "epoch": 0.4165317426501917, + "grad_norm": 19.25, + "learning_rate": 2.2498932917778373e-06, + "loss": 0.7952, + "step": 4888 + }, + { + "epoch": 0.4166169578184917, + "grad_norm": 13.5, + "learning_rate": 2.2497924422899304e-06, + "loss": 0.735, + "step": 4889 + }, + { + "epoch": 0.41670217298679163, + "grad_norm": 15.0, + "learning_rate": 2.2496915747347684e-06, + "loss": 0.6473, + "step": 4890 + }, + { + "epoch": 0.4167873881550916, + "grad_norm": 13.625, + "learning_rate": 2.2495906891141746e-06, + "loss": 0.3715, + "step": 4891 + }, + { + "epoch": 0.41687260332339154, + "grad_norm": 16.5, + "learning_rate": 2.249489785429972e-06, + "loss": 0.5998, + "step": 4892 + }, + { + "epoch": 0.4169578184916915, + "grad_norm": 16.5, + "learning_rate": 2.2493888636839837e-06, + "loss": 0.9581, + "step": 4893 + }, + { + "epoch": 0.41704303365999146, + "grad_norm": 14.625, + "learning_rate": 2.249287923878034e-06, + "loss": 0.504, + "step": 4894 + }, + { + "epoch": 0.4171282488282914, + "grad_norm": 11.9375, + "learning_rate": 2.2491869660139463e-06, + "loss": 0.6031, + "step": 4895 + }, + { + "epoch": 0.41721346399659137, + "grad_norm": 11.5625, + "learning_rate": 2.2490859900935455e-06, + "loss": 0.5202, + "step": 4896 + }, + { + "epoch": 0.41729867916489133, + "grad_norm": 13.8125, + "learning_rate": 2.248984996118656e-06, + "loss": 0.6677, + "step": 4897 + }, + { + "epoch": 0.4173838943331913, + "grad_norm": 14.3125, + "learning_rate": 2.248883984091103e-06, + "loss": 0.5755, + "step": 4898 + }, + { + "epoch": 0.41746910950149124, + "grad_norm": 14.4375, + "learning_rate": 2.2487829540127124e-06, + "loss": 0.5265, + "step": 4899 + }, + { + "epoch": 0.4175543246697912, + "grad_norm": 13.0, + "learning_rate": 2.248681905885309e-06, + "loss": 0.5974, + "step": 4900 + }, + { + "epoch": 0.41763953983809116, + "grad_norm": 14.0625, + "learning_rate": 2.24858083971072e-06, + "loss": 0.9805, + "step": 4901 + }, + { + "epoch": 0.4177247550063911, + "grad_norm": 14.0625, + "learning_rate": 2.2484797554907706e-06, + "loss": 0.5058, + "step": 4902 + }, + { + "epoch": 0.41780997017469107, + "grad_norm": 17.0, + "learning_rate": 2.2483786532272885e-06, + "loss": 0.6768, + "step": 4903 + }, + { + "epoch": 0.417895185342991, + "grad_norm": 14.1875, + "learning_rate": 2.2482775329220995e-06, + "loss": 0.6575, + "step": 4904 + }, + { + "epoch": 0.417980400511291, + "grad_norm": 20.25, + "learning_rate": 2.248176394577032e-06, + "loss": 1.009, + "step": 4905 + }, + { + "epoch": 0.41806561567959094, + "grad_norm": 12.25, + "learning_rate": 2.2480752381939132e-06, + "loss": 0.3105, + "step": 4906 + }, + { + "epoch": 0.4181508308478909, + "grad_norm": 15.1875, + "learning_rate": 2.247974063774571e-06, + "loss": 0.6516, + "step": 4907 + }, + { + "epoch": 0.41823604601619085, + "grad_norm": 19.875, + "learning_rate": 2.247872871320834e-06, + "loss": 0.9633, + "step": 4908 + }, + { + "epoch": 0.41832126118449087, + "grad_norm": 13.4375, + "learning_rate": 2.2477716608345307e-06, + "loss": 0.6915, + "step": 4909 + }, + { + "epoch": 0.4184064763527908, + "grad_norm": 13.1875, + "learning_rate": 2.2476704323174904e-06, + "loss": 0.7064, + "step": 4910 + }, + { + "epoch": 0.4184916915210908, + "grad_norm": 18.875, + "learning_rate": 2.247569185771542e-06, + "loss": 0.8686, + "step": 4911 + }, + { + "epoch": 0.41857690668939074, + "grad_norm": 13.125, + "learning_rate": 2.247467921198516e-06, + "loss": 0.7016, + "step": 4912 + }, + { + "epoch": 0.4186621218576907, + "grad_norm": 19.25, + "learning_rate": 2.2473666386002403e-06, + "loss": 1.0099, + "step": 4913 + }, + { + "epoch": 0.41874733702599065, + "grad_norm": 14.5625, + "learning_rate": 2.2472653379785475e-06, + "loss": 0.6496, + "step": 4914 + }, + { + "epoch": 0.4188325521942906, + "grad_norm": 11.0625, + "learning_rate": 2.2471640193352665e-06, + "loss": 0.3736, + "step": 4915 + }, + { + "epoch": 0.41891776736259057, + "grad_norm": 14.8125, + "learning_rate": 2.24706268267223e-06, + "loss": 1.0165, + "step": 4916 + }, + { + "epoch": 0.4190029825308905, + "grad_norm": 15.625, + "learning_rate": 2.2469613279912672e-06, + "loss": 0.7243, + "step": 4917 + }, + { + "epoch": 0.4190881976991905, + "grad_norm": 14.75, + "learning_rate": 2.246859955294211e-06, + "loss": 0.5785, + "step": 4918 + }, + { + "epoch": 0.41917341286749044, + "grad_norm": 18.875, + "learning_rate": 2.246758564582893e-06, + "loss": 1.0706, + "step": 4919 + }, + { + "epoch": 0.4192586280357904, + "grad_norm": 16.625, + "learning_rate": 2.2466571558591457e-06, + "loss": 0.7378, + "step": 4920 + }, + { + "epoch": 0.41934384320409035, + "grad_norm": 12.4375, + "learning_rate": 2.2465557291248006e-06, + "loss": 0.6348, + "step": 4921 + }, + { + "epoch": 0.4194290583723903, + "grad_norm": 12.875, + "learning_rate": 2.2464542843816923e-06, + "loss": 0.5927, + "step": 4922 + }, + { + "epoch": 0.41951427354069026, + "grad_norm": 13.875, + "learning_rate": 2.246352821631653e-06, + "loss": 0.325, + "step": 4923 + }, + { + "epoch": 0.4195994887089902, + "grad_norm": 17.75, + "learning_rate": 2.2462513408765153e-06, + "loss": 0.8192, + "step": 4924 + }, + { + "epoch": 0.4196847038772902, + "grad_norm": 12.5, + "learning_rate": 2.246149842118115e-06, + "loss": 0.9298, + "step": 4925 + }, + { + "epoch": 0.41976991904559013, + "grad_norm": 13.75, + "learning_rate": 2.246048325358285e-06, + "loss": 0.8035, + "step": 4926 + }, + { + "epoch": 0.4198551342138901, + "grad_norm": 16.25, + "learning_rate": 2.2459467905988604e-06, + "loss": 0.7296, + "step": 4927 + }, + { + "epoch": 0.41994034938219005, + "grad_norm": 24.875, + "learning_rate": 2.2458452378416753e-06, + "loss": 0.9956, + "step": 4928 + }, + { + "epoch": 0.42002556455049, + "grad_norm": 20.375, + "learning_rate": 2.245743667088566e-06, + "loss": 0.8621, + "step": 4929 + }, + { + "epoch": 0.42011077971878996, + "grad_norm": 15.0, + "learning_rate": 2.2456420783413667e-06, + "loss": 0.7533, + "step": 4930 + }, + { + "epoch": 0.4201959948870899, + "grad_norm": 15.1875, + "learning_rate": 2.245540471601914e-06, + "loss": 1.0693, + "step": 4931 + }, + { + "epoch": 0.4202812100553899, + "grad_norm": 13.4375, + "learning_rate": 2.2454388468720444e-06, + "loss": 0.9583, + "step": 4932 + }, + { + "epoch": 0.42036642522368983, + "grad_norm": 11.375, + "learning_rate": 2.245337204153593e-06, + "loss": 0.4919, + "step": 4933 + }, + { + "epoch": 0.4204516403919898, + "grad_norm": 12.875, + "learning_rate": 2.245235543448398e-06, + "loss": 0.5027, + "step": 4934 + }, + { + "epoch": 0.42053685556028975, + "grad_norm": 12.1875, + "learning_rate": 2.245133864758296e-06, + "loss": 0.6087, + "step": 4935 + }, + { + "epoch": 0.4206220707285897, + "grad_norm": 15.6875, + "learning_rate": 2.245032168085124e-06, + "loss": 1.1637, + "step": 4936 + }, + { + "epoch": 0.42070728589688966, + "grad_norm": 11.9375, + "learning_rate": 2.2449304534307206e-06, + "loss": 0.4725, + "step": 4937 + }, + { + "epoch": 0.4207925010651896, + "grad_norm": 16.375, + "learning_rate": 2.2448287207969234e-06, + "loss": 1.0574, + "step": 4938 + }, + { + "epoch": 0.4208777162334896, + "grad_norm": 16.875, + "learning_rate": 2.2447269701855706e-06, + "loss": 1.0514, + "step": 4939 + }, + { + "epoch": 0.42096293140178953, + "grad_norm": 14.0, + "learning_rate": 2.244625201598501e-06, + "loss": 0.764, + "step": 4940 + }, + { + "epoch": 0.4210481465700895, + "grad_norm": 16.875, + "learning_rate": 2.2445234150375542e-06, + "loss": 0.9954, + "step": 4941 + }, + { + "epoch": 0.42113336173838944, + "grad_norm": 13.6875, + "learning_rate": 2.244421610504569e-06, + "loss": 0.5302, + "step": 4942 + }, + { + "epoch": 0.4212185769066894, + "grad_norm": 13.4375, + "learning_rate": 2.244319788001386e-06, + "loss": 0.4155, + "step": 4943 + }, + { + "epoch": 0.42130379207498936, + "grad_norm": 13.8125, + "learning_rate": 2.2442179475298433e-06, + "loss": 0.6276, + "step": 4944 + }, + { + "epoch": 0.4213890072432893, + "grad_norm": 17.75, + "learning_rate": 2.2441160890917833e-06, + "loss": 0.7554, + "step": 4945 + }, + { + "epoch": 0.42147422241158927, + "grad_norm": 15.5625, + "learning_rate": 2.2440142126890455e-06, + "loss": 0.7116, + "step": 4946 + }, + { + "epoch": 0.42155943757988923, + "grad_norm": 11.8125, + "learning_rate": 2.243912318323472e-06, + "loss": 0.4652, + "step": 4947 + }, + { + "epoch": 0.4216446527481892, + "grad_norm": 15.5, + "learning_rate": 2.243810405996903e-06, + "loss": 1.0827, + "step": 4948 + }, + { + "epoch": 0.42172986791648914, + "grad_norm": 17.75, + "learning_rate": 2.2437084757111804e-06, + "loss": 0.9546, + "step": 4949 + }, + { + "epoch": 0.4218150830847891, + "grad_norm": 11.0, + "learning_rate": 2.2436065274681467e-06, + "loss": 0.6164, + "step": 4950 + }, + { + "epoch": 0.42190029825308906, + "grad_norm": 12.875, + "learning_rate": 2.243504561269644e-06, + "loss": 0.7249, + "step": 4951 + }, + { + "epoch": 0.421985513421389, + "grad_norm": 12.9375, + "learning_rate": 2.2434025771175147e-06, + "loss": 0.6985, + "step": 4952 + }, + { + "epoch": 0.42207072858968897, + "grad_norm": 12.5, + "learning_rate": 2.243300575013602e-06, + "loss": 0.6818, + "step": 4953 + }, + { + "epoch": 0.4221559437579889, + "grad_norm": 29.25, + "learning_rate": 2.243198554959749e-06, + "loss": 1.2069, + "step": 4954 + }, + { + "epoch": 0.4222411589262889, + "grad_norm": 16.875, + "learning_rate": 2.2430965169577993e-06, + "loss": 1.3152, + "step": 4955 + }, + { + "epoch": 0.42232637409458884, + "grad_norm": 14.1875, + "learning_rate": 2.242994461009597e-06, + "loss": 0.7069, + "step": 4956 + }, + { + "epoch": 0.4224115892628888, + "grad_norm": 16.625, + "learning_rate": 2.242892387116986e-06, + "loss": 0.6766, + "step": 4957 + }, + { + "epoch": 0.42249680443118875, + "grad_norm": 17.125, + "learning_rate": 2.2427902952818117e-06, + "loss": 0.7159, + "step": 4958 + }, + { + "epoch": 0.4225820195994887, + "grad_norm": 14.0, + "learning_rate": 2.242688185505918e-06, + "loss": 0.8027, + "step": 4959 + }, + { + "epoch": 0.42266723476778867, + "grad_norm": 15.25, + "learning_rate": 2.2425860577911504e-06, + "loss": 1.0403, + "step": 4960 + }, + { + "epoch": 0.4227524499360886, + "grad_norm": 14.3125, + "learning_rate": 2.242483912139355e-06, + "loss": 0.8111, + "step": 4961 + }, + { + "epoch": 0.4228376651043886, + "grad_norm": 15.5, + "learning_rate": 2.242381748552377e-06, + "loss": 0.7659, + "step": 4962 + }, + { + "epoch": 0.42292288027268854, + "grad_norm": 13.75, + "learning_rate": 2.242279567032063e-06, + "loss": 0.9226, + "step": 4963 + }, + { + "epoch": 0.4230080954409885, + "grad_norm": 13.5625, + "learning_rate": 2.2421773675802596e-06, + "loss": 0.5856, + "step": 4964 + }, + { + "epoch": 0.42309331060928845, + "grad_norm": 29.25, + "learning_rate": 2.242075150198813e-06, + "loss": 1.1184, + "step": 4965 + }, + { + "epoch": 0.4231785257775884, + "grad_norm": 15.3125, + "learning_rate": 2.241972914889571e-06, + "loss": 0.9654, + "step": 4966 + }, + { + "epoch": 0.42326374094588837, + "grad_norm": 12.1875, + "learning_rate": 2.241870661654381e-06, + "loss": 0.5952, + "step": 4967 + }, + { + "epoch": 0.4233489561141883, + "grad_norm": 16.25, + "learning_rate": 2.241768390495091e-06, + "loss": 0.8882, + "step": 4968 + }, + { + "epoch": 0.4234341712824883, + "grad_norm": 18.625, + "learning_rate": 2.2416661014135484e-06, + "loss": 1.246, + "step": 4969 + }, + { + "epoch": 0.42351938645078824, + "grad_norm": 14.4375, + "learning_rate": 2.2415637944116025e-06, + "loss": 0.6923, + "step": 4970 + }, + { + "epoch": 0.4236046016190882, + "grad_norm": 13.9375, + "learning_rate": 2.2414614694911016e-06, + "loss": 0.5906, + "step": 4971 + }, + { + "epoch": 0.42368981678738815, + "grad_norm": 16.25, + "learning_rate": 2.241359126653895e-06, + "loss": 0.5943, + "step": 4972 + }, + { + "epoch": 0.4237750319556881, + "grad_norm": 14.3125, + "learning_rate": 2.241256765901832e-06, + "loss": 0.5606, + "step": 4973 + }, + { + "epoch": 0.42386024712398807, + "grad_norm": 23.25, + "learning_rate": 2.2411543872367623e-06, + "loss": 1.257, + "step": 4974 + }, + { + "epoch": 0.423945462292288, + "grad_norm": 20.375, + "learning_rate": 2.2410519906605368e-06, + "loss": 1.3547, + "step": 4975 + }, + { + "epoch": 0.424030677460588, + "grad_norm": 15.4375, + "learning_rate": 2.240949576175005e-06, + "loss": 0.858, + "step": 4976 + }, + { + "epoch": 0.42411589262888794, + "grad_norm": 17.375, + "learning_rate": 2.2408471437820175e-06, + "loss": 0.689, + "step": 4977 + }, + { + "epoch": 0.4242011077971879, + "grad_norm": 35.75, + "learning_rate": 2.240744693483426e-06, + "loss": 0.5849, + "step": 4978 + }, + { + "epoch": 0.42428632296548785, + "grad_norm": 16.875, + "learning_rate": 2.2406422252810815e-06, + "loss": 0.5204, + "step": 4979 + }, + { + "epoch": 0.4243715381337878, + "grad_norm": 13.875, + "learning_rate": 2.2405397391768357e-06, + "loss": 0.6337, + "step": 4980 + }, + { + "epoch": 0.42445675330208776, + "grad_norm": 13.0625, + "learning_rate": 2.2404372351725413e-06, + "loss": 0.7091, + "step": 4981 + }, + { + "epoch": 0.4245419684703877, + "grad_norm": 11.0, + "learning_rate": 2.2403347132700493e-06, + "loss": 0.5954, + "step": 4982 + }, + { + "epoch": 0.4246271836386877, + "grad_norm": 14.8125, + "learning_rate": 2.240232173471214e-06, + "loss": 0.7097, + "step": 4983 + }, + { + "epoch": 0.42471239880698763, + "grad_norm": 14.9375, + "learning_rate": 2.240129615777887e-06, + "loss": 0.7659, + "step": 4984 + }, + { + "epoch": 0.4247976139752876, + "grad_norm": 25.375, + "learning_rate": 2.240027040191922e-06, + "loss": 0.9313, + "step": 4985 + }, + { + "epoch": 0.42488282914358755, + "grad_norm": 13.9375, + "learning_rate": 2.2399244467151735e-06, + "loss": 0.8854, + "step": 4986 + }, + { + "epoch": 0.4249680443118875, + "grad_norm": 12.5, + "learning_rate": 2.2398218353494945e-06, + "loss": 0.4812, + "step": 4987 + }, + { + "epoch": 0.42505325948018746, + "grad_norm": 14.3125, + "learning_rate": 2.23971920609674e-06, + "loss": 0.629, + "step": 4988 + }, + { + "epoch": 0.4251384746484874, + "grad_norm": 15.25, + "learning_rate": 2.2396165589587636e-06, + "loss": 0.8386, + "step": 4989 + }, + { + "epoch": 0.4252236898167874, + "grad_norm": 11.0625, + "learning_rate": 2.2395138939374213e-06, + "loss": 0.5445, + "step": 4990 + }, + { + "epoch": 0.42530890498508733, + "grad_norm": 12.9375, + "learning_rate": 2.2394112110345672e-06, + "loss": 0.5404, + "step": 4991 + }, + { + "epoch": 0.4253941201533873, + "grad_norm": 14.25, + "learning_rate": 2.239308510252058e-06, + "loss": 0.4705, + "step": 4992 + }, + { + "epoch": 0.42547933532168725, + "grad_norm": 16.875, + "learning_rate": 2.239205791591749e-06, + "loss": 1.0347, + "step": 4993 + }, + { + "epoch": 0.4255645504899872, + "grad_norm": 12.8125, + "learning_rate": 2.2391030550554972e-06, + "loss": 0.6768, + "step": 4994 + }, + { + "epoch": 0.42564976565828716, + "grad_norm": 14.5625, + "learning_rate": 2.239000300645158e-06, + "loss": 0.6945, + "step": 4995 + }, + { + "epoch": 0.4257349808265871, + "grad_norm": 12.4375, + "learning_rate": 2.238897528362589e-06, + "loss": 0.5118, + "step": 4996 + }, + { + "epoch": 0.4258201959948871, + "grad_norm": 15.75, + "learning_rate": 2.2387947382096477e-06, + "loss": 1.0141, + "step": 4997 + }, + { + "epoch": 0.42590541116318703, + "grad_norm": 13.8125, + "learning_rate": 2.238691930188191e-06, + "loss": 0.6464, + "step": 4998 + }, + { + "epoch": 0.425990626331487, + "grad_norm": 15.0, + "learning_rate": 2.2385891043000766e-06, + "loss": 0.9191, + "step": 4999 + }, + { + "epoch": 0.42607584149978694, + "grad_norm": 16.5, + "learning_rate": 2.238486260547163e-06, + "loss": 0.4816, + "step": 5000 + }, + { + "epoch": 0.4261610566680869, + "grad_norm": 11.75, + "learning_rate": 2.2383833989313087e-06, + "loss": 0.5444, + "step": 5001 + }, + { + "epoch": 0.42624627183638686, + "grad_norm": 23.25, + "learning_rate": 2.2382805194543726e-06, + "loss": 0.8971, + "step": 5002 + }, + { + "epoch": 0.4263314870046868, + "grad_norm": 15.4375, + "learning_rate": 2.2381776221182136e-06, + "loss": 0.9627, + "step": 5003 + }, + { + "epoch": 0.4264167021729868, + "grad_norm": 15.8125, + "learning_rate": 2.2380747069246916e-06, + "loss": 0.6279, + "step": 5004 + }, + { + "epoch": 0.42650191734128673, + "grad_norm": 13.5625, + "learning_rate": 2.2379717738756656e-06, + "loss": 0.8512, + "step": 5005 + }, + { + "epoch": 0.4265871325095867, + "grad_norm": 14.375, + "learning_rate": 2.2378688229729965e-06, + "loss": 0.7839, + "step": 5006 + }, + { + "epoch": 0.42667234767788664, + "grad_norm": 16.75, + "learning_rate": 2.2377658542185442e-06, + "loss": 0.8248, + "step": 5007 + }, + { + "epoch": 0.4267575628461866, + "grad_norm": 12.6875, + "learning_rate": 2.23766286761417e-06, + "loss": 0.7751, + "step": 5008 + }, + { + "epoch": 0.42684277801448656, + "grad_norm": 17.125, + "learning_rate": 2.237559863161734e-06, + "loss": 1.0821, + "step": 5009 + }, + { + "epoch": 0.4269279931827865, + "grad_norm": 14.25, + "learning_rate": 2.2374568408630985e-06, + "loss": 0.7242, + "step": 5010 + }, + { + "epoch": 0.42701320835108647, + "grad_norm": 16.25, + "learning_rate": 2.237353800720125e-06, + "loss": 0.8592, + "step": 5011 + }, + { + "epoch": 0.4270984235193864, + "grad_norm": 12.5, + "learning_rate": 2.2372507427346753e-06, + "loss": 0.5808, + "step": 5012 + }, + { + "epoch": 0.4271836386876864, + "grad_norm": 15.125, + "learning_rate": 2.2371476669086123e-06, + "loss": 1.0118, + "step": 5013 + }, + { + "epoch": 0.42726885385598634, + "grad_norm": 14.3125, + "learning_rate": 2.237044573243798e-06, + "loss": 0.525, + "step": 5014 + }, + { + "epoch": 0.4273540690242863, + "grad_norm": 23.5, + "learning_rate": 2.236941461742096e-06, + "loss": 0.629, + "step": 5015 + }, + { + "epoch": 0.42743928419258626, + "grad_norm": 13.75, + "learning_rate": 2.2368383324053684e-06, + "loss": 0.6978, + "step": 5016 + }, + { + "epoch": 0.4275244993608862, + "grad_norm": 13.0625, + "learning_rate": 2.236735185235481e-06, + "loss": 0.5348, + "step": 5017 + }, + { + "epoch": 0.42760971452918617, + "grad_norm": 16.0, + "learning_rate": 2.2366320202342957e-06, + "loss": 0.9351, + "step": 5018 + }, + { + "epoch": 0.4276949296974861, + "grad_norm": 21.0, + "learning_rate": 2.236528837403678e-06, + "loss": 1.0668, + "step": 5019 + }, + { + "epoch": 0.4277801448657861, + "grad_norm": 23.75, + "learning_rate": 2.2364256367454925e-06, + "loss": 1.078, + "step": 5020 + }, + { + "epoch": 0.42786536003408604, + "grad_norm": 25.375, + "learning_rate": 2.2363224182616035e-06, + "loss": 1.1077, + "step": 5021 + }, + { + "epoch": 0.42795057520238605, + "grad_norm": 12.125, + "learning_rate": 2.2362191819538763e-06, + "loss": 0.5359, + "step": 5022 + }, + { + "epoch": 0.428035790370686, + "grad_norm": 11.5, + "learning_rate": 2.236115927824177e-06, + "loss": 0.4355, + "step": 5023 + }, + { + "epoch": 0.42812100553898597, + "grad_norm": 13.75, + "learning_rate": 2.236012655874371e-06, + "loss": 0.555, + "step": 5024 + }, + { + "epoch": 0.4282062207072859, + "grad_norm": 13.9375, + "learning_rate": 2.2359093661063254e-06, + "loss": 0.772, + "step": 5025 + }, + { + "epoch": 0.4282914358755859, + "grad_norm": 16.5, + "learning_rate": 2.235806058521906e-06, + "loss": 0.9323, + "step": 5026 + }, + { + "epoch": 0.42837665104388584, + "grad_norm": 17.125, + "learning_rate": 2.2357027331229796e-06, + "loss": 0.9413, + "step": 5027 + }, + { + "epoch": 0.4284618662121858, + "grad_norm": 13.3125, + "learning_rate": 2.235599389911414e-06, + "loss": 0.5359, + "step": 5028 + }, + { + "epoch": 0.42854708138048575, + "grad_norm": 10.375, + "learning_rate": 2.2354960288890754e-06, + "loss": 0.4342, + "step": 5029 + }, + { + "epoch": 0.4286322965487857, + "grad_norm": 13.8125, + "learning_rate": 2.2353926500578333e-06, + "loss": 0.766, + "step": 5030 + }, + { + "epoch": 0.42871751171708566, + "grad_norm": 14.0625, + "learning_rate": 2.2352892534195548e-06, + "loss": 0.5773, + "step": 5031 + }, + { + "epoch": 0.4288027268853856, + "grad_norm": 20.75, + "learning_rate": 2.2351858389761088e-06, + "loss": 0.7962, + "step": 5032 + }, + { + "epoch": 0.4288879420536856, + "grad_norm": 13.1875, + "learning_rate": 2.2350824067293643e-06, + "loss": 0.6885, + "step": 5033 + }, + { + "epoch": 0.42897315722198553, + "grad_norm": 16.375, + "learning_rate": 2.23497895668119e-06, + "loss": 0.7861, + "step": 5034 + }, + { + "epoch": 0.4290583723902855, + "grad_norm": 17.75, + "learning_rate": 2.2348754888334554e-06, + "loss": 0.8054, + "step": 5035 + }, + { + "epoch": 0.42914358755858545, + "grad_norm": 12.5625, + "learning_rate": 2.2347720031880303e-06, + "loss": 0.4438, + "step": 5036 + }, + { + "epoch": 0.4292288027268854, + "grad_norm": 17.375, + "learning_rate": 2.234668499746785e-06, + "loss": 0.8679, + "step": 5037 + }, + { + "epoch": 0.42931401789518536, + "grad_norm": 14.625, + "learning_rate": 2.2345649785115897e-06, + "loss": 0.6812, + "step": 5038 + }, + { + "epoch": 0.4293992330634853, + "grad_norm": 24.375, + "learning_rate": 2.234461439484315e-06, + "loss": 0.8815, + "step": 5039 + }, + { + "epoch": 0.4294844482317853, + "grad_norm": 11.5625, + "learning_rate": 2.2343578826668322e-06, + "loss": 0.4515, + "step": 5040 + }, + { + "epoch": 0.42956966340008523, + "grad_norm": 12.0625, + "learning_rate": 2.234254308061013e-06, + "loss": 0.5014, + "step": 5041 + }, + { + "epoch": 0.4296548785683852, + "grad_norm": 18.875, + "learning_rate": 2.234150715668728e-06, + "loss": 0.7075, + "step": 5042 + }, + { + "epoch": 0.42974009373668515, + "grad_norm": 20.625, + "learning_rate": 2.2340471054918506e-06, + "loss": 0.7983, + "step": 5043 + }, + { + "epoch": 0.4298253089049851, + "grad_norm": 14.0625, + "learning_rate": 2.233943477532252e-06, + "loss": 0.6592, + "step": 5044 + }, + { + "epoch": 0.42991052407328506, + "grad_norm": 16.5, + "learning_rate": 2.2338398317918054e-06, + "loss": 0.7493, + "step": 5045 + }, + { + "epoch": 0.429995739241585, + "grad_norm": 13.9375, + "learning_rate": 2.233736168272384e-06, + "loss": 0.4503, + "step": 5046 + }, + { + "epoch": 0.430080954409885, + "grad_norm": 12.8125, + "learning_rate": 2.2336324869758607e-06, + "loss": 0.4597, + "step": 5047 + }, + { + "epoch": 0.43016616957818493, + "grad_norm": 18.25, + "learning_rate": 2.233528787904109e-06, + "loss": 0.822, + "step": 5048 + }, + { + "epoch": 0.4302513847464849, + "grad_norm": 13.9375, + "learning_rate": 2.233425071059003e-06, + "loss": 0.6965, + "step": 5049 + }, + { + "epoch": 0.43033659991478485, + "grad_norm": 16.75, + "learning_rate": 2.2333213364424178e-06, + "loss": 0.9644, + "step": 5050 + }, + { + "epoch": 0.4304218150830848, + "grad_norm": 21.25, + "learning_rate": 2.2332175840562264e-06, + "loss": 1.0969, + "step": 5051 + }, + { + "epoch": 0.43050703025138476, + "grad_norm": 18.75, + "learning_rate": 2.233113813902305e-06, + "loss": 1.0114, + "step": 5052 + }, + { + "epoch": 0.4305922454196847, + "grad_norm": 12.625, + "learning_rate": 2.2330100259825285e-06, + "loss": 0.7766, + "step": 5053 + }, + { + "epoch": 0.4306774605879847, + "grad_norm": 22.125, + "learning_rate": 2.2329062202987718e-06, + "loss": 1.1876, + "step": 5054 + }, + { + "epoch": 0.43076267575628463, + "grad_norm": 15.9375, + "learning_rate": 2.232802396852912e-06, + "loss": 0.9081, + "step": 5055 + }, + { + "epoch": 0.4308478909245846, + "grad_norm": 17.5, + "learning_rate": 2.2326985556468238e-06, + "loss": 0.7736, + "step": 5056 + }, + { + "epoch": 0.43093310609288454, + "grad_norm": 18.0, + "learning_rate": 2.2325946966823853e-06, + "loss": 0.9413, + "step": 5057 + }, + { + "epoch": 0.4310183212611845, + "grad_norm": 19.125, + "learning_rate": 2.2324908199614722e-06, + "loss": 1.1306, + "step": 5058 + }, + { + "epoch": 0.43110353642948446, + "grad_norm": 11.75, + "learning_rate": 2.232386925485962e-06, + "loss": 0.6384, + "step": 5059 + }, + { + "epoch": 0.4311887515977844, + "grad_norm": 14.4375, + "learning_rate": 2.232283013257732e-06, + "loss": 0.7943, + "step": 5060 + }, + { + "epoch": 0.43127396676608437, + "grad_norm": 15.5, + "learning_rate": 2.2321790832786607e-06, + "loss": 0.9235, + "step": 5061 + }, + { + "epoch": 0.43135918193438433, + "grad_norm": 11.125, + "learning_rate": 2.2320751355506255e-06, + "loss": 0.5174, + "step": 5062 + }, + { + "epoch": 0.4314443971026843, + "grad_norm": 10.25, + "learning_rate": 2.231971170075505e-06, + "loss": 0.4515, + "step": 5063 + }, + { + "epoch": 0.43152961227098424, + "grad_norm": 12.75, + "learning_rate": 2.231867186855178e-06, + "loss": 0.6666, + "step": 5064 + }, + { + "epoch": 0.4316148274392842, + "grad_norm": 11.8125, + "learning_rate": 2.2317631858915235e-06, + "loss": 0.7227, + "step": 5065 + }, + { + "epoch": 0.43170004260758416, + "grad_norm": 14.375, + "learning_rate": 2.2316591671864213e-06, + "loss": 0.7342, + "step": 5066 + }, + { + "epoch": 0.4317852577758841, + "grad_norm": 15.3125, + "learning_rate": 2.2315551307417506e-06, + "loss": 0.7317, + "step": 5067 + }, + { + "epoch": 0.43187047294418407, + "grad_norm": 12.625, + "learning_rate": 2.2314510765593917e-06, + "loss": 0.5484, + "step": 5068 + }, + { + "epoch": 0.431955688112484, + "grad_norm": 17.875, + "learning_rate": 2.2313470046412247e-06, + "loss": 1.0629, + "step": 5069 + }, + { + "epoch": 0.432040903280784, + "grad_norm": 14.375, + "learning_rate": 2.2312429149891307e-06, + "loss": 0.74, + "step": 5070 + }, + { + "epoch": 0.43212611844908394, + "grad_norm": 28.875, + "learning_rate": 2.2311388076049903e-06, + "loss": 0.8439, + "step": 5071 + }, + { + "epoch": 0.4322113336173839, + "grad_norm": 11.1875, + "learning_rate": 2.2310346824906855e-06, + "loss": 0.519, + "step": 5072 + }, + { + "epoch": 0.43229654878568385, + "grad_norm": 11.4375, + "learning_rate": 2.2309305396480967e-06, + "loss": 0.4261, + "step": 5073 + }, + { + "epoch": 0.4323817639539838, + "grad_norm": 12.375, + "learning_rate": 2.2308263790791073e-06, + "loss": 0.6671, + "step": 5074 + }, + { + "epoch": 0.43246697912228377, + "grad_norm": 14.4375, + "learning_rate": 2.230722200785599e-06, + "loss": 0.8671, + "step": 5075 + }, + { + "epoch": 0.4325521942905837, + "grad_norm": 25.625, + "learning_rate": 2.230618004769454e-06, + "loss": 1.1036, + "step": 5076 + }, + { + "epoch": 0.4326374094588837, + "grad_norm": 11.3125, + "learning_rate": 2.230513791032556e-06, + "loss": 0.4433, + "step": 5077 + }, + { + "epoch": 0.43272262462718364, + "grad_norm": 15.875, + "learning_rate": 2.230409559576787e-06, + "loss": 0.6872, + "step": 5078 + }, + { + "epoch": 0.4328078397954836, + "grad_norm": 13.4375, + "learning_rate": 2.2303053104040322e-06, + "loss": 0.8941, + "step": 5079 + }, + { + "epoch": 0.43289305496378355, + "grad_norm": 13.5, + "learning_rate": 2.230201043516174e-06, + "loss": 0.6345, + "step": 5080 + }, + { + "epoch": 0.4329782701320835, + "grad_norm": 12.125, + "learning_rate": 2.230096758915098e-06, + "loss": 0.3736, + "step": 5081 + }, + { + "epoch": 0.43306348530038347, + "grad_norm": 15.3125, + "learning_rate": 2.2299924566026876e-06, + "loss": 0.869, + "step": 5082 + }, + { + "epoch": 0.4331487004686834, + "grad_norm": 14.8125, + "learning_rate": 2.2298881365808276e-06, + "loss": 0.9045, + "step": 5083 + }, + { + "epoch": 0.4332339156369834, + "grad_norm": 13.125, + "learning_rate": 2.229783798851404e-06, + "loss": 0.8778, + "step": 5084 + }, + { + "epoch": 0.43331913080528334, + "grad_norm": 14.25, + "learning_rate": 2.229679443416302e-06, + "loss": 0.7869, + "step": 5085 + }, + { + "epoch": 0.4334043459735833, + "grad_norm": 13.6875, + "learning_rate": 2.2295750702774077e-06, + "loss": 0.6352, + "step": 5086 + }, + { + "epoch": 0.43348956114188325, + "grad_norm": 16.875, + "learning_rate": 2.2294706794366066e-06, + "loss": 0.8027, + "step": 5087 + }, + { + "epoch": 0.4335747763101832, + "grad_norm": 15.9375, + "learning_rate": 2.229366270895785e-06, + "loss": 1.0057, + "step": 5088 + }, + { + "epoch": 0.43365999147848316, + "grad_norm": 11.8125, + "learning_rate": 2.2292618446568305e-06, + "loss": 0.3933, + "step": 5089 + }, + { + "epoch": 0.4337452066467831, + "grad_norm": 14.4375, + "learning_rate": 2.2291574007216295e-06, + "loss": 0.8552, + "step": 5090 + }, + { + "epoch": 0.4338304218150831, + "grad_norm": 16.375, + "learning_rate": 2.2290529390920696e-06, + "loss": 0.8523, + "step": 5091 + }, + { + "epoch": 0.43391563698338304, + "grad_norm": 14.8125, + "learning_rate": 2.2289484597700386e-06, + "loss": 0.7322, + "step": 5092 + }, + { + "epoch": 0.434000852151683, + "grad_norm": 11.0, + "learning_rate": 2.2288439627574247e-06, + "loss": 0.4172, + "step": 5093 + }, + { + "epoch": 0.43408606731998295, + "grad_norm": 13.375, + "learning_rate": 2.228739448056116e-06, + "loss": 0.5626, + "step": 5094 + }, + { + "epoch": 0.4341712824882829, + "grad_norm": 10.8125, + "learning_rate": 2.2286349156680014e-06, + "loss": 0.6058, + "step": 5095 + }, + { + "epoch": 0.43425649765658286, + "grad_norm": 15.25, + "learning_rate": 2.2285303655949693e-06, + "loss": 0.9247, + "step": 5096 + }, + { + "epoch": 0.4343417128248828, + "grad_norm": 13.5625, + "learning_rate": 2.22842579783891e-06, + "loss": 0.6181, + "step": 5097 + }, + { + "epoch": 0.4344269279931828, + "grad_norm": 14.375, + "learning_rate": 2.2283212124017124e-06, + "loss": 0.6834, + "step": 5098 + }, + { + "epoch": 0.43451214316148273, + "grad_norm": 17.5, + "learning_rate": 2.228216609285267e-06, + "loss": 0.6878, + "step": 5099 + }, + { + "epoch": 0.4345973583297827, + "grad_norm": 12.75, + "learning_rate": 2.2281119884914635e-06, + "loss": 0.6651, + "step": 5100 + }, + { + "epoch": 0.43468257349808265, + "grad_norm": 17.125, + "learning_rate": 2.2280073500221926e-06, + "loss": 1.0826, + "step": 5101 + }, + { + "epoch": 0.4347677886663826, + "grad_norm": 15.0625, + "learning_rate": 2.2279026938793458e-06, + "loss": 0.863, + "step": 5102 + }, + { + "epoch": 0.43485300383468256, + "grad_norm": 15.625, + "learning_rate": 2.2277980200648142e-06, + "loss": 0.8633, + "step": 5103 + }, + { + "epoch": 0.4349382190029825, + "grad_norm": 14.875, + "learning_rate": 2.2276933285804884e-06, + "loss": 0.6028, + "step": 5104 + }, + { + "epoch": 0.4350234341712825, + "grad_norm": 19.5, + "learning_rate": 2.2275886194282615e-06, + "loss": 0.7762, + "step": 5105 + }, + { + "epoch": 0.43510864933958243, + "grad_norm": 10.4375, + "learning_rate": 2.227483892610025e-06, + "loss": 0.5226, + "step": 5106 + }, + { + "epoch": 0.4351938645078824, + "grad_norm": 13.0625, + "learning_rate": 2.227379148127672e-06, + "loss": 0.8069, + "step": 5107 + }, + { + "epoch": 0.43527907967618235, + "grad_norm": 16.5, + "learning_rate": 2.227274385983095e-06, + "loss": 0.7514, + "step": 5108 + }, + { + "epoch": 0.4353642948444823, + "grad_norm": 36.25, + "learning_rate": 2.227169606178186e-06, + "loss": 0.8766, + "step": 5109 + }, + { + "epoch": 0.43544951001278226, + "grad_norm": 10.75, + "learning_rate": 2.2270648087148412e-06, + "loss": 0.6408, + "step": 5110 + }, + { + "epoch": 0.4355347251810822, + "grad_norm": 17.625, + "learning_rate": 2.2269599935949513e-06, + "loss": 1.1137, + "step": 5111 + }, + { + "epoch": 0.4356199403493822, + "grad_norm": 15.375, + "learning_rate": 2.226855160820413e-06, + "loss": 0.893, + "step": 5112 + }, + { + "epoch": 0.43570515551768213, + "grad_norm": 11.3125, + "learning_rate": 2.2267503103931194e-06, + "loss": 0.4471, + "step": 5113 + }, + { + "epoch": 0.4357903706859821, + "grad_norm": 18.5, + "learning_rate": 2.2266454423149653e-06, + "loss": 0.8746, + "step": 5114 + }, + { + "epoch": 0.43587558585428204, + "grad_norm": 13.6875, + "learning_rate": 2.2265405565878463e-06, + "loss": 0.6705, + "step": 5115 + }, + { + "epoch": 0.435960801022582, + "grad_norm": 15.1875, + "learning_rate": 2.226435653213657e-06, + "loss": 0.7691, + "step": 5116 + }, + { + "epoch": 0.43604601619088196, + "grad_norm": 17.5, + "learning_rate": 2.2263307321942935e-06, + "loss": 0.9185, + "step": 5117 + }, + { + "epoch": 0.4361312313591819, + "grad_norm": 13.4375, + "learning_rate": 2.2262257935316523e-06, + "loss": 0.6799, + "step": 5118 + }, + { + "epoch": 0.43621644652748187, + "grad_norm": 12.625, + "learning_rate": 2.226120837227629e-06, + "loss": 0.6413, + "step": 5119 + }, + { + "epoch": 0.43630166169578183, + "grad_norm": 13.125, + "learning_rate": 2.2260158632841207e-06, + "loss": 0.5855, + "step": 5120 + }, + { + "epoch": 0.4363868768640818, + "grad_norm": 17.125, + "learning_rate": 2.2259108717030245e-06, + "loss": 0.6167, + "step": 5121 + }, + { + "epoch": 0.43647209203238174, + "grad_norm": 15.125, + "learning_rate": 2.2258058624862374e-06, + "loss": 0.6559, + "step": 5122 + }, + { + "epoch": 0.4365573072006817, + "grad_norm": 18.75, + "learning_rate": 2.225700835635657e-06, + "loss": 0.9811, + "step": 5123 + }, + { + "epoch": 0.43664252236898166, + "grad_norm": 16.125, + "learning_rate": 2.2255957911531817e-06, + "loss": 0.7894, + "step": 5124 + }, + { + "epoch": 0.4367277375372816, + "grad_norm": 17.125, + "learning_rate": 2.225490729040709e-06, + "loss": 1.1962, + "step": 5125 + }, + { + "epoch": 0.43681295270558157, + "grad_norm": 12.0625, + "learning_rate": 2.225385649300138e-06, + "loss": 0.5451, + "step": 5126 + }, + { + "epoch": 0.4368981678738815, + "grad_norm": 12.75, + "learning_rate": 2.2252805519333676e-06, + "loss": 0.6483, + "step": 5127 + }, + { + "epoch": 0.4369833830421815, + "grad_norm": 15.125, + "learning_rate": 2.2251754369422968e-06, + "loss": 0.8555, + "step": 5128 + }, + { + "epoch": 0.43706859821048144, + "grad_norm": 14.75, + "learning_rate": 2.2250703043288253e-06, + "loss": 0.7367, + "step": 5129 + }, + { + "epoch": 0.4371538133787814, + "grad_norm": 12.75, + "learning_rate": 2.224965154094853e-06, + "loss": 0.524, + "step": 5130 + }, + { + "epoch": 0.43723902854708135, + "grad_norm": 15.875, + "learning_rate": 2.2248599862422798e-06, + "loss": 0.7958, + "step": 5131 + }, + { + "epoch": 0.4373242437153813, + "grad_norm": 13.625, + "learning_rate": 2.2247548007730063e-06, + "loss": 0.7912, + "step": 5132 + }, + { + "epoch": 0.43740945888368127, + "grad_norm": 18.375, + "learning_rate": 2.2246495976889335e-06, + "loss": 0.8527, + "step": 5133 + }, + { + "epoch": 0.4374946740519812, + "grad_norm": 18.875, + "learning_rate": 2.224544376991962e-06, + "loss": 0.9801, + "step": 5134 + }, + { + "epoch": 0.43757988922028124, + "grad_norm": 11.375, + "learning_rate": 2.224439138683994e-06, + "loss": 0.651, + "step": 5135 + }, + { + "epoch": 0.4376651043885812, + "grad_norm": 14.3125, + "learning_rate": 2.224333882766931e-06, + "loss": 0.7407, + "step": 5136 + }, + { + "epoch": 0.43775031955688115, + "grad_norm": 15.5625, + "learning_rate": 2.2242286092426747e-06, + "loss": 1.0061, + "step": 5137 + }, + { + "epoch": 0.4378355347251811, + "grad_norm": 14.25, + "learning_rate": 2.224123318113128e-06, + "loss": 0.9623, + "step": 5138 + }, + { + "epoch": 0.43792074989348106, + "grad_norm": 16.75, + "learning_rate": 2.2240180093801935e-06, + "loss": 0.4397, + "step": 5139 + }, + { + "epoch": 0.438005965061781, + "grad_norm": 19.25, + "learning_rate": 2.2239126830457736e-06, + "loss": 1.0909, + "step": 5140 + }, + { + "epoch": 0.438091180230081, + "grad_norm": 16.875, + "learning_rate": 2.2238073391117726e-06, + "loss": 1.1586, + "step": 5141 + }, + { + "epoch": 0.43817639539838094, + "grad_norm": 14.75, + "learning_rate": 2.2237019775800938e-06, + "loss": 0.9064, + "step": 5142 + }, + { + "epoch": 0.4382616105666809, + "grad_norm": 15.375, + "learning_rate": 2.2235965984526408e-06, + "loss": 0.7616, + "step": 5143 + }, + { + "epoch": 0.43834682573498085, + "grad_norm": 25.75, + "learning_rate": 2.2234912017313182e-06, + "loss": 1.1972, + "step": 5144 + }, + { + "epoch": 0.4384320409032808, + "grad_norm": 12.75, + "learning_rate": 2.223385787418031e-06, + "loss": 0.7455, + "step": 5145 + }, + { + "epoch": 0.43851725607158076, + "grad_norm": 13.5, + "learning_rate": 2.223280355514684e-06, + "loss": 0.7262, + "step": 5146 + }, + { + "epoch": 0.4386024712398807, + "grad_norm": 10.6875, + "learning_rate": 2.2231749060231816e-06, + "loss": 0.4949, + "step": 5147 + }, + { + "epoch": 0.4386876864081807, + "grad_norm": 13.875, + "learning_rate": 2.2230694389454305e-06, + "loss": 0.9748, + "step": 5148 + }, + { + "epoch": 0.43877290157648063, + "grad_norm": 15.0625, + "learning_rate": 2.222963954283336e-06, + "loss": 0.9741, + "step": 5149 + }, + { + "epoch": 0.4388581167447806, + "grad_norm": 11.5, + "learning_rate": 2.2228584520388043e-06, + "loss": 0.5132, + "step": 5150 + }, + { + "epoch": 0.43894333191308055, + "grad_norm": 11.9375, + "learning_rate": 2.2227529322137425e-06, + "loss": 0.6055, + "step": 5151 + }, + { + "epoch": 0.4390285470813805, + "grad_norm": 14.3125, + "learning_rate": 2.2226473948100564e-06, + "loss": 0.9458, + "step": 5152 + }, + { + "epoch": 0.43911376224968046, + "grad_norm": 11.4375, + "learning_rate": 2.2225418398296545e-06, + "loss": 0.5338, + "step": 5153 + }, + { + "epoch": 0.4391989774179804, + "grad_norm": 19.375, + "learning_rate": 2.222436267274443e-06, + "loss": 0.3762, + "step": 5154 + }, + { + "epoch": 0.4392841925862804, + "grad_norm": 11.75, + "learning_rate": 2.2223306771463305e-06, + "loss": 0.6674, + "step": 5155 + }, + { + "epoch": 0.43936940775458033, + "grad_norm": 12.0625, + "learning_rate": 2.222225069447225e-06, + "loss": 0.6346, + "step": 5156 + }, + { + "epoch": 0.4394546229228803, + "grad_norm": 18.0, + "learning_rate": 2.222119444179035e-06, + "loss": 1.0271, + "step": 5157 + }, + { + "epoch": 0.43953983809118025, + "grad_norm": 23.75, + "learning_rate": 2.222013801343669e-06, + "loss": 1.0399, + "step": 5158 + }, + { + "epoch": 0.4396250532594802, + "grad_norm": 15.625, + "learning_rate": 2.221908140943036e-06, + "loss": 0.9165, + "step": 5159 + }, + { + "epoch": 0.43971026842778016, + "grad_norm": 11.0, + "learning_rate": 2.221802462979046e-06, + "loss": 0.401, + "step": 5160 + }, + { + "epoch": 0.4397954835960801, + "grad_norm": 20.25, + "learning_rate": 2.2216967674536074e-06, + "loss": 0.7724, + "step": 5161 + }, + { + "epoch": 0.4398806987643801, + "grad_norm": 10.8125, + "learning_rate": 2.2215910543686317e-06, + "loss": 0.3656, + "step": 5162 + }, + { + "epoch": 0.43996591393268003, + "grad_norm": 11.375, + "learning_rate": 2.221485323726029e-06, + "loss": 0.3966, + "step": 5163 + }, + { + "epoch": 0.44005112910098, + "grad_norm": 11.1875, + "learning_rate": 2.221379575527709e-06, + "loss": 0.3691, + "step": 5164 + }, + { + "epoch": 0.44013634426927994, + "grad_norm": 12.0, + "learning_rate": 2.221273809775583e-06, + "loss": 0.4409, + "step": 5165 + }, + { + "epoch": 0.4402215594375799, + "grad_norm": 15.375, + "learning_rate": 2.221168026471563e-06, + "loss": 1.0218, + "step": 5166 + }, + { + "epoch": 0.44030677460587986, + "grad_norm": 16.625, + "learning_rate": 2.2210622256175606e-06, + "loss": 0.7113, + "step": 5167 + }, + { + "epoch": 0.4403919897741798, + "grad_norm": 13.4375, + "learning_rate": 2.2209564072154864e-06, + "loss": 0.7158, + "step": 5168 + }, + { + "epoch": 0.44047720494247977, + "grad_norm": 14.375, + "learning_rate": 2.2208505712672543e-06, + "loss": 0.739, + "step": 5169 + }, + { + "epoch": 0.44056242011077973, + "grad_norm": 22.5, + "learning_rate": 2.220744717774776e-06, + "loss": 1.0637, + "step": 5170 + }, + { + "epoch": 0.4406476352790797, + "grad_norm": 16.125, + "learning_rate": 2.2206388467399642e-06, + "loss": 0.7977, + "step": 5171 + }, + { + "epoch": 0.44073285044737964, + "grad_norm": 12.5625, + "learning_rate": 2.2205329581647323e-06, + "loss": 0.649, + "step": 5172 + }, + { + "epoch": 0.4408180656156796, + "grad_norm": 10.5625, + "learning_rate": 2.220427052050994e-06, + "loss": 0.4986, + "step": 5173 + }, + { + "epoch": 0.44090328078397956, + "grad_norm": 14.75, + "learning_rate": 2.2203211284006632e-06, + "loss": 0.9302, + "step": 5174 + }, + { + "epoch": 0.4409884959522795, + "grad_norm": 11.3125, + "learning_rate": 2.2202151872156542e-06, + "loss": 0.4803, + "step": 5175 + }, + { + "epoch": 0.44107371112057947, + "grad_norm": 13.1875, + "learning_rate": 2.22010922849788e-06, + "loss": 0.5799, + "step": 5176 + }, + { + "epoch": 0.4411589262888794, + "grad_norm": 10.1875, + "learning_rate": 2.220003252249258e-06, + "loss": 0.5021, + "step": 5177 + }, + { + "epoch": 0.4412441414571794, + "grad_norm": 16.25, + "learning_rate": 2.219897258471701e-06, + "loss": 0.8595, + "step": 5178 + }, + { + "epoch": 0.44132935662547934, + "grad_norm": 17.25, + "learning_rate": 2.219791247167125e-06, + "loss": 0.7109, + "step": 5179 + }, + { + "epoch": 0.4414145717937793, + "grad_norm": 12.125, + "learning_rate": 2.2196852183374466e-06, + "loss": 0.6619, + "step": 5180 + }, + { + "epoch": 0.44149978696207925, + "grad_norm": 13.0625, + "learning_rate": 2.219579171984581e-06, + "loss": 0.5405, + "step": 5181 + }, + { + "epoch": 0.4415850021303792, + "grad_norm": 11.6875, + "learning_rate": 2.2194731081104446e-06, + "loss": 0.5325, + "step": 5182 + }, + { + "epoch": 0.44167021729867917, + "grad_norm": 12.25, + "learning_rate": 2.2193670267169544e-06, + "loss": 0.7233, + "step": 5183 + }, + { + "epoch": 0.4417554324669791, + "grad_norm": 18.75, + "learning_rate": 2.219260927806027e-06, + "loss": 0.7228, + "step": 5184 + }, + { + "epoch": 0.4418406476352791, + "grad_norm": 12.4375, + "learning_rate": 2.2191548113795804e-06, + "loss": 0.4754, + "step": 5185 + }, + { + "epoch": 0.44192586280357904, + "grad_norm": 12.0, + "learning_rate": 2.2190486774395315e-06, + "loss": 0.677, + "step": 5186 + }, + { + "epoch": 0.442011077971879, + "grad_norm": 26.375, + "learning_rate": 2.2189425259877984e-06, + "loss": 0.9605, + "step": 5187 + }, + { + "epoch": 0.44209629314017895, + "grad_norm": 13.9375, + "learning_rate": 2.2188363570263e-06, + "loss": 0.7215, + "step": 5188 + }, + { + "epoch": 0.4421815083084789, + "grad_norm": 12.5, + "learning_rate": 2.218730170556954e-06, + "loss": 0.5369, + "step": 5189 + }, + { + "epoch": 0.44226672347677887, + "grad_norm": 15.4375, + "learning_rate": 2.2186239665816793e-06, + "loss": 0.8766, + "step": 5190 + }, + { + "epoch": 0.4423519386450788, + "grad_norm": 24.0, + "learning_rate": 2.2185177451023965e-06, + "loss": 0.8766, + "step": 5191 + }, + { + "epoch": 0.4424371538133788, + "grad_norm": 8.75, + "learning_rate": 2.2184115061210235e-06, + "loss": 0.2994, + "step": 5192 + }, + { + "epoch": 0.44252236898167874, + "grad_norm": 12.875, + "learning_rate": 2.218305249639481e-06, + "loss": 0.5088, + "step": 5193 + }, + { + "epoch": 0.4426075841499787, + "grad_norm": 13.3125, + "learning_rate": 2.2181989756596887e-06, + "loss": 0.5516, + "step": 5194 + }, + { + "epoch": 0.44269279931827865, + "grad_norm": 13.875, + "learning_rate": 2.2180926841835674e-06, + "loss": 0.6515, + "step": 5195 + }, + { + "epoch": 0.4427780144865786, + "grad_norm": 12.8125, + "learning_rate": 2.2179863752130377e-06, + "loss": 0.6983, + "step": 5196 + }, + { + "epoch": 0.44286322965487857, + "grad_norm": 17.25, + "learning_rate": 2.217880048750021e-06, + "loss": 0.9351, + "step": 5197 + }, + { + "epoch": 0.4429484448231785, + "grad_norm": 15.625, + "learning_rate": 2.2177737047964385e-06, + "loss": 1.06, + "step": 5198 + }, + { + "epoch": 0.4430336599914785, + "grad_norm": 13.1875, + "learning_rate": 2.217667343354212e-06, + "loss": 0.6594, + "step": 5199 + }, + { + "epoch": 0.44311887515977844, + "grad_norm": 14.875, + "learning_rate": 2.2175609644252635e-06, + "loss": 0.5199, + "step": 5200 + }, + { + "epoch": 0.4432040903280784, + "grad_norm": 12.9375, + "learning_rate": 2.2174545680115157e-06, + "loss": 0.4343, + "step": 5201 + }, + { + "epoch": 0.44328930549637835, + "grad_norm": 11.25, + "learning_rate": 2.2173481541148906e-06, + "loss": 0.5872, + "step": 5202 + }, + { + "epoch": 0.4433745206646783, + "grad_norm": 12.1875, + "learning_rate": 2.2172417227373118e-06, + "loss": 0.6242, + "step": 5203 + }, + { + "epoch": 0.44345973583297826, + "grad_norm": 15.4375, + "learning_rate": 2.2171352738807024e-06, + "loss": 0.5877, + "step": 5204 + }, + { + "epoch": 0.4435449510012782, + "grad_norm": 14.3125, + "learning_rate": 2.217028807546986e-06, + "loss": 0.8408, + "step": 5205 + }, + { + "epoch": 0.4436301661695782, + "grad_norm": 15.5625, + "learning_rate": 2.2169223237380864e-06, + "loss": 0.829, + "step": 5206 + }, + { + "epoch": 0.44371538133787813, + "grad_norm": 11.5, + "learning_rate": 2.2168158224559283e-06, + "loss": 0.4527, + "step": 5207 + }, + { + "epoch": 0.4438005965061781, + "grad_norm": 10.3125, + "learning_rate": 2.2167093037024362e-06, + "loss": 0.3851, + "step": 5208 + }, + { + "epoch": 0.44388581167447805, + "grad_norm": 22.5, + "learning_rate": 2.2166027674795346e-06, + "loss": 1.1393, + "step": 5209 + }, + { + "epoch": 0.443971026842778, + "grad_norm": 13.375, + "learning_rate": 2.2164962137891493e-06, + "loss": 0.5187, + "step": 5210 + }, + { + "epoch": 0.44405624201107796, + "grad_norm": 17.75, + "learning_rate": 2.2163896426332054e-06, + "loss": 1.0089, + "step": 5211 + }, + { + "epoch": 0.4441414571793779, + "grad_norm": 19.75, + "learning_rate": 2.216283054013629e-06, + "loss": 0.9306, + "step": 5212 + }, + { + "epoch": 0.4442266723476779, + "grad_norm": 11.5625, + "learning_rate": 2.216176447932346e-06, + "loss": 0.394, + "step": 5213 + }, + { + "epoch": 0.44431188751597783, + "grad_norm": 27.0, + "learning_rate": 2.216069824391283e-06, + "loss": 1.0792, + "step": 5214 + }, + { + "epoch": 0.4443971026842778, + "grad_norm": 16.5, + "learning_rate": 2.2159631833923673e-06, + "loss": 0.7715, + "step": 5215 + }, + { + "epoch": 0.44448231785257775, + "grad_norm": 13.125, + "learning_rate": 2.215856524937525e-06, + "loss": 0.6128, + "step": 5216 + }, + { + "epoch": 0.4445675330208777, + "grad_norm": 12.125, + "learning_rate": 2.2157498490286845e-06, + "loss": 0.6043, + "step": 5217 + }, + { + "epoch": 0.44465274818917766, + "grad_norm": 13.0625, + "learning_rate": 2.2156431556677724e-06, + "loss": 0.6091, + "step": 5218 + }, + { + "epoch": 0.4447379633574776, + "grad_norm": 18.875, + "learning_rate": 2.2155364448567178e-06, + "loss": 0.6945, + "step": 5219 + }, + { + "epoch": 0.4448231785257776, + "grad_norm": 20.0, + "learning_rate": 2.2154297165974486e-06, + "loss": 0.7325, + "step": 5220 + }, + { + "epoch": 0.44490839369407753, + "grad_norm": 12.375, + "learning_rate": 2.215322970891894e-06, + "loss": 0.7625, + "step": 5221 + }, + { + "epoch": 0.4449936088623775, + "grad_norm": 15.5, + "learning_rate": 2.2152162077419822e-06, + "loss": 1.0698, + "step": 5222 + }, + { + "epoch": 0.44507882403067744, + "grad_norm": 18.375, + "learning_rate": 2.215109427149643e-06, + "loss": 0.8082, + "step": 5223 + }, + { + "epoch": 0.4451640391989774, + "grad_norm": 12.5625, + "learning_rate": 2.215002629116806e-06, + "loss": 0.5166, + "step": 5224 + }, + { + "epoch": 0.44524925436727736, + "grad_norm": 15.0625, + "learning_rate": 2.214895813645401e-06, + "loss": 0.8749, + "step": 5225 + }, + { + "epoch": 0.4453344695355773, + "grad_norm": 15.1875, + "learning_rate": 2.214788980737359e-06, + "loss": 0.966, + "step": 5226 + }, + { + "epoch": 0.44541968470387727, + "grad_norm": 13.0625, + "learning_rate": 2.2146821303946093e-06, + "loss": 0.7744, + "step": 5227 + }, + { + "epoch": 0.44550489987217723, + "grad_norm": 13.5625, + "learning_rate": 2.2145752626190835e-06, + "loss": 0.807, + "step": 5228 + }, + { + "epoch": 0.4455901150404772, + "grad_norm": 15.25, + "learning_rate": 2.214468377412713e-06, + "loss": 0.8426, + "step": 5229 + }, + { + "epoch": 0.44567533020877714, + "grad_norm": 19.125, + "learning_rate": 2.214361474777429e-06, + "loss": 0.8497, + "step": 5230 + }, + { + "epoch": 0.4457605453770771, + "grad_norm": 33.0, + "learning_rate": 2.2142545547151633e-06, + "loss": 1.0836, + "step": 5231 + }, + { + "epoch": 0.44584576054537706, + "grad_norm": 19.625, + "learning_rate": 2.214147617227848e-06, + "loss": 0.6466, + "step": 5232 + }, + { + "epoch": 0.445930975713677, + "grad_norm": 9.625, + "learning_rate": 2.214040662317416e-06, + "loss": 0.5071, + "step": 5233 + }, + { + "epoch": 0.44601619088197697, + "grad_norm": 15.0625, + "learning_rate": 2.2139336899858e-06, + "loss": 0.7577, + "step": 5234 + }, + { + "epoch": 0.4461014060502769, + "grad_norm": 15.75, + "learning_rate": 2.2138267002349324e-06, + "loss": 0.992, + "step": 5235 + }, + { + "epoch": 0.4461866212185769, + "grad_norm": 14.6875, + "learning_rate": 2.2137196930667477e-06, + "loss": 0.5531, + "step": 5236 + }, + { + "epoch": 0.44627183638687684, + "grad_norm": 13.5625, + "learning_rate": 2.213612668483178e-06, + "loss": 0.4828, + "step": 5237 + }, + { + "epoch": 0.4463570515551768, + "grad_norm": 13.5625, + "learning_rate": 2.2135056264861596e-06, + "loss": 0.72, + "step": 5238 + }, + { + "epoch": 0.44644226672347675, + "grad_norm": 17.125, + "learning_rate": 2.213398567077625e-06, + "loss": 0.6722, + "step": 5239 + }, + { + "epoch": 0.4465274818917767, + "grad_norm": 14.6875, + "learning_rate": 2.21329149025951e-06, + "loss": 1.135, + "step": 5240 + }, + { + "epoch": 0.44661269706007667, + "grad_norm": 18.375, + "learning_rate": 2.2131843960337483e-06, + "loss": 0.6865, + "step": 5241 + }, + { + "epoch": 0.4466979122283766, + "grad_norm": 11.125, + "learning_rate": 2.213077284402277e-06, + "loss": 0.511, + "step": 5242 + }, + { + "epoch": 0.4467831273966766, + "grad_norm": 25.0, + "learning_rate": 2.21297015536703e-06, + "loss": 0.7436, + "step": 5243 + }, + { + "epoch": 0.44686834256497654, + "grad_norm": 12.9375, + "learning_rate": 2.212863008929944e-06, + "loss": 0.5247, + "step": 5244 + }, + { + "epoch": 0.4469535577332765, + "grad_norm": 24.0, + "learning_rate": 2.2127558450929556e-06, + "loss": 1.2036, + "step": 5245 + }, + { + "epoch": 0.44703877290157645, + "grad_norm": 13.875, + "learning_rate": 2.212648663858001e-06, + "loss": 0.7338, + "step": 5246 + }, + { + "epoch": 0.44712398806987647, + "grad_norm": 20.375, + "learning_rate": 2.2125414652270162e-06, + "loss": 0.6491, + "step": 5247 + }, + { + "epoch": 0.4472092032381764, + "grad_norm": 16.25, + "learning_rate": 2.21243424920194e-06, + "loss": 0.9395, + "step": 5248 + }, + { + "epoch": 0.4472944184064764, + "grad_norm": 18.75, + "learning_rate": 2.212327015784709e-06, + "loss": 0.6624, + "step": 5249 + }, + { + "epoch": 0.44737963357477634, + "grad_norm": 19.5, + "learning_rate": 2.2122197649772608e-06, + "loss": 0.9869, + "step": 5250 + }, + { + "epoch": 0.4474648487430763, + "grad_norm": 19.0, + "learning_rate": 2.212112496781534e-06, + "loss": 0.717, + "step": 5251 + }, + { + "epoch": 0.44755006391137625, + "grad_norm": 13.8125, + "learning_rate": 2.212005211199467e-06, + "loss": 0.7496, + "step": 5252 + }, + { + "epoch": 0.4476352790796762, + "grad_norm": 16.5, + "learning_rate": 2.2118979082329984e-06, + "loss": 0.7765, + "step": 5253 + }, + { + "epoch": 0.44772049424797616, + "grad_norm": 14.75, + "learning_rate": 2.2117905878840678e-06, + "loss": 0.9994, + "step": 5254 + }, + { + "epoch": 0.4478057094162761, + "grad_norm": 18.0, + "learning_rate": 2.2116832501546135e-06, + "loss": 0.9808, + "step": 5255 + }, + { + "epoch": 0.4478909245845761, + "grad_norm": 20.875, + "learning_rate": 2.211575895046576e-06, + "loss": 0.8259, + "step": 5256 + }, + { + "epoch": 0.44797613975287603, + "grad_norm": 22.0, + "learning_rate": 2.2114685225618953e-06, + "loss": 1.1549, + "step": 5257 + }, + { + "epoch": 0.448061354921176, + "grad_norm": 12.875, + "learning_rate": 2.211361132702511e-06, + "loss": 0.6495, + "step": 5258 + }, + { + "epoch": 0.44814657008947595, + "grad_norm": 16.125, + "learning_rate": 2.2112537254703647e-06, + "loss": 0.8431, + "step": 5259 + }, + { + "epoch": 0.4482317852577759, + "grad_norm": 11.8125, + "learning_rate": 2.2111463008673974e-06, + "loss": 0.6806, + "step": 5260 + }, + { + "epoch": 0.44831700042607586, + "grad_norm": 13.1875, + "learning_rate": 2.2110388588955495e-06, + "loss": 0.5808, + "step": 5261 + }, + { + "epoch": 0.4484022155943758, + "grad_norm": 10.75, + "learning_rate": 2.2109313995567628e-06, + "loss": 0.4957, + "step": 5262 + }, + { + "epoch": 0.4484874307626758, + "grad_norm": 11.25, + "learning_rate": 2.2108239228529798e-06, + "loss": 0.3631, + "step": 5263 + }, + { + "epoch": 0.44857264593097573, + "grad_norm": 14.5625, + "learning_rate": 2.210716428786142e-06, + "loss": 0.7375, + "step": 5264 + }, + { + "epoch": 0.4486578610992757, + "grad_norm": 12.875, + "learning_rate": 2.2106089173581926e-06, + "loss": 0.4065, + "step": 5265 + }, + { + "epoch": 0.44874307626757565, + "grad_norm": 13.6875, + "learning_rate": 2.210501388571074e-06, + "loss": 0.7736, + "step": 5266 + }, + { + "epoch": 0.4488282914358756, + "grad_norm": 11.625, + "learning_rate": 2.2103938424267292e-06, + "loss": 0.5118, + "step": 5267 + }, + { + "epoch": 0.44891350660417556, + "grad_norm": 25.125, + "learning_rate": 2.2102862789271024e-06, + "loss": 0.8588, + "step": 5268 + }, + { + "epoch": 0.4489987217724755, + "grad_norm": 37.0, + "learning_rate": 2.210178698074136e-06, + "loss": 1.5926, + "step": 5269 + }, + { + "epoch": 0.4490839369407755, + "grad_norm": 19.125, + "learning_rate": 2.210071099869776e-06, + "loss": 1.0543, + "step": 5270 + }, + { + "epoch": 0.44916915210907543, + "grad_norm": 17.125, + "learning_rate": 2.2099634843159654e-06, + "loss": 0.9676, + "step": 5271 + }, + { + "epoch": 0.4492543672773754, + "grad_norm": 17.75, + "learning_rate": 2.2098558514146496e-06, + "loss": 0.5011, + "step": 5272 + }, + { + "epoch": 0.44933958244567535, + "grad_norm": 11.625, + "learning_rate": 2.2097482011677727e-06, + "loss": 0.6006, + "step": 5273 + }, + { + "epoch": 0.4494247976139753, + "grad_norm": 14.6875, + "learning_rate": 2.2096405335772813e-06, + "loss": 0.8925, + "step": 5274 + }, + { + "epoch": 0.44951001278227526, + "grad_norm": 14.75, + "learning_rate": 2.20953284864512e-06, + "loss": 0.6267, + "step": 5275 + }, + { + "epoch": 0.4495952279505752, + "grad_norm": 14.0625, + "learning_rate": 2.2094251463732357e-06, + "loss": 0.8305, + "step": 5276 + }, + { + "epoch": 0.4496804431188752, + "grad_norm": 13.875, + "learning_rate": 2.209317426763574e-06, + "loss": 0.9639, + "step": 5277 + }, + { + "epoch": 0.44976565828717513, + "grad_norm": 14.75, + "learning_rate": 2.2092096898180816e-06, + "loss": 0.7304, + "step": 5278 + }, + { + "epoch": 0.4498508734554751, + "grad_norm": 16.625, + "learning_rate": 2.209101935538706e-06, + "loss": 0.9641, + "step": 5279 + }, + { + "epoch": 0.44993608862377504, + "grad_norm": 33.75, + "learning_rate": 2.2089941639273936e-06, + "loss": 1.5391, + "step": 5280 + }, + { + "epoch": 0.450021303792075, + "grad_norm": 17.875, + "learning_rate": 2.208886374986092e-06, + "loss": 0.9008, + "step": 5281 + }, + { + "epoch": 0.45010651896037496, + "grad_norm": 14.8125, + "learning_rate": 2.20877856871675e-06, + "loss": 0.8533, + "step": 5282 + }, + { + "epoch": 0.4501917341286749, + "grad_norm": 13.0625, + "learning_rate": 2.208670745121315e-06, + "loss": 0.7125, + "step": 5283 + }, + { + "epoch": 0.45027694929697487, + "grad_norm": 15.0, + "learning_rate": 2.2085629042017353e-06, + "loss": 0.8686, + "step": 5284 + }, + { + "epoch": 0.45036216446527483, + "grad_norm": 15.5, + "learning_rate": 2.2084550459599606e-06, + "loss": 0.8119, + "step": 5285 + }, + { + "epoch": 0.4504473796335748, + "grad_norm": 21.875, + "learning_rate": 2.208347170397939e-06, + "loss": 1.0876, + "step": 5286 + }, + { + "epoch": 0.45053259480187474, + "grad_norm": 16.25, + "learning_rate": 2.2082392775176205e-06, + "loss": 1.1659, + "step": 5287 + }, + { + "epoch": 0.4506178099701747, + "grad_norm": 11.75, + "learning_rate": 2.2081313673209544e-06, + "loss": 0.6577, + "step": 5288 + }, + { + "epoch": 0.45070302513847466, + "grad_norm": 11.0, + "learning_rate": 2.208023439809891e-06, + "loss": 0.3374, + "step": 5289 + }, + { + "epoch": 0.4507882403067746, + "grad_norm": 15.1875, + "learning_rate": 2.2079154949863806e-06, + "loss": 0.8558, + "step": 5290 + }, + { + "epoch": 0.45087345547507457, + "grad_norm": 18.5, + "learning_rate": 2.2078075328523744e-06, + "loss": 0.9127, + "step": 5291 + }, + { + "epoch": 0.4509586706433745, + "grad_norm": 25.75, + "learning_rate": 2.207699553409823e-06, + "loss": 1.2323, + "step": 5292 + }, + { + "epoch": 0.4510438858116745, + "grad_norm": 10.125, + "learning_rate": 2.207591556660677e-06, + "loss": 0.4612, + "step": 5293 + }, + { + "epoch": 0.45112910097997444, + "grad_norm": 15.125, + "learning_rate": 2.207483542606889e-06, + "loss": 0.605, + "step": 5294 + }, + { + "epoch": 0.4512143161482744, + "grad_norm": 13.8125, + "learning_rate": 2.2073755112504105e-06, + "loss": 0.9879, + "step": 5295 + }, + { + "epoch": 0.45129953131657435, + "grad_norm": 15.3125, + "learning_rate": 2.207267462593194e-06, + "loss": 0.5696, + "step": 5296 + }, + { + "epoch": 0.4513847464848743, + "grad_norm": 13.8125, + "learning_rate": 2.207159396637192e-06, + "loss": 0.7613, + "step": 5297 + }, + { + "epoch": 0.45146996165317427, + "grad_norm": 13.8125, + "learning_rate": 2.2070513133843568e-06, + "loss": 0.7534, + "step": 5298 + }, + { + "epoch": 0.4515551768214742, + "grad_norm": 21.75, + "learning_rate": 2.206943212836642e-06, + "loss": 1.1458, + "step": 5299 + }, + { + "epoch": 0.4516403919897742, + "grad_norm": 13.375, + "learning_rate": 2.206835094996001e-06, + "loss": 0.5133, + "step": 5300 + }, + { + "epoch": 0.45172560715807414, + "grad_norm": 15.625, + "learning_rate": 2.2067269598643874e-06, + "loss": 0.8986, + "step": 5301 + }, + { + "epoch": 0.4518108223263741, + "grad_norm": 12.5625, + "learning_rate": 2.2066188074437562e-06, + "loss": 0.6559, + "step": 5302 + }, + { + "epoch": 0.45189603749467405, + "grad_norm": 13.75, + "learning_rate": 2.2065106377360614e-06, + "loss": 0.6126, + "step": 5303 + }, + { + "epoch": 0.451981252662974, + "grad_norm": 16.125, + "learning_rate": 2.206402450743257e-06, + "loss": 0.7747, + "step": 5304 + }, + { + "epoch": 0.45206646783127397, + "grad_norm": 16.25, + "learning_rate": 2.2062942464672984e-06, + "loss": 0.9428, + "step": 5305 + }, + { + "epoch": 0.4521516829995739, + "grad_norm": 18.375, + "learning_rate": 2.2061860249101414e-06, + "loss": 0.9978, + "step": 5306 + }, + { + "epoch": 0.4522368981678739, + "grad_norm": 14.1875, + "learning_rate": 2.2060777860737415e-06, + "loss": 0.5861, + "step": 5307 + }, + { + "epoch": 0.45232211333617384, + "grad_norm": 15.0, + "learning_rate": 2.2059695299600543e-06, + "loss": 0.836, + "step": 5308 + }, + { + "epoch": 0.4524073285044738, + "grad_norm": 16.625, + "learning_rate": 2.205861256571037e-06, + "loss": 0.8986, + "step": 5309 + }, + { + "epoch": 0.45249254367277375, + "grad_norm": 19.125, + "learning_rate": 2.205752965908645e-06, + "loss": 0.7699, + "step": 5310 + }, + { + "epoch": 0.4525777588410737, + "grad_norm": 18.625, + "learning_rate": 2.205644657974836e-06, + "loss": 0.5925, + "step": 5311 + }, + { + "epoch": 0.45266297400937366, + "grad_norm": 9.75, + "learning_rate": 2.205536332771567e-06, + "loss": 0.3757, + "step": 5312 + }, + { + "epoch": 0.4527481891776736, + "grad_norm": 11.3125, + "learning_rate": 2.205427990300796e-06, + "loss": 0.4661, + "step": 5313 + }, + { + "epoch": 0.4528334043459736, + "grad_norm": 11.3125, + "learning_rate": 2.20531963056448e-06, + "loss": 0.5236, + "step": 5314 + }, + { + "epoch": 0.45291861951427353, + "grad_norm": 17.5, + "learning_rate": 2.2052112535645777e-06, + "loss": 1.0352, + "step": 5315 + }, + { + "epoch": 0.4530038346825735, + "grad_norm": 14.375, + "learning_rate": 2.2051028593030474e-06, + "loss": 0.8625, + "step": 5316 + }, + { + "epoch": 0.45308904985087345, + "grad_norm": 15.375, + "learning_rate": 2.204994447781848e-06, + "loss": 0.6187, + "step": 5317 + }, + { + "epoch": 0.4531742650191734, + "grad_norm": 11.5625, + "learning_rate": 2.2048860190029393e-06, + "loss": 0.6228, + "step": 5318 + }, + { + "epoch": 0.45325948018747336, + "grad_norm": 17.625, + "learning_rate": 2.204777572968279e-06, + "loss": 0.6301, + "step": 5319 + }, + { + "epoch": 0.4533446953557733, + "grad_norm": 14.5625, + "learning_rate": 2.2046691096798285e-06, + "loss": 1.1456, + "step": 5320 + }, + { + "epoch": 0.4534299105240733, + "grad_norm": 19.625, + "learning_rate": 2.2045606291395472e-06, + "loss": 0.897, + "step": 5321 + }, + { + "epoch": 0.45351512569237323, + "grad_norm": 13.6875, + "learning_rate": 2.204452131349395e-06, + "loss": 0.5846, + "step": 5322 + }, + { + "epoch": 0.4536003408606732, + "grad_norm": 19.875, + "learning_rate": 2.2043436163113334e-06, + "loss": 0.7718, + "step": 5323 + }, + { + "epoch": 0.45368555602897315, + "grad_norm": 18.625, + "learning_rate": 2.204235084027323e-06, + "loss": 0.6552, + "step": 5324 + }, + { + "epoch": 0.4537707711972731, + "grad_norm": 10.5625, + "learning_rate": 2.204126534499325e-06, + "loss": 0.3773, + "step": 5325 + }, + { + "epoch": 0.45385598636557306, + "grad_norm": 16.75, + "learning_rate": 2.204017967729301e-06, + "loss": 0.8311, + "step": 5326 + }, + { + "epoch": 0.453941201533873, + "grad_norm": 14.5625, + "learning_rate": 2.203909383719213e-06, + "loss": 1.0692, + "step": 5327 + }, + { + "epoch": 0.454026416702173, + "grad_norm": 16.875, + "learning_rate": 2.2038007824710232e-06, + "loss": 0.342, + "step": 5328 + }, + { + "epoch": 0.45411163187047293, + "grad_norm": 11.5625, + "learning_rate": 2.203692163986694e-06, + "loss": 0.3911, + "step": 5329 + }, + { + "epoch": 0.4541968470387729, + "grad_norm": 11.25, + "learning_rate": 2.2035835282681888e-06, + "loss": 0.4331, + "step": 5330 + }, + { + "epoch": 0.45428206220707285, + "grad_norm": 12.1875, + "learning_rate": 2.2034748753174703e-06, + "loss": 0.2653, + "step": 5331 + }, + { + "epoch": 0.4543672773753728, + "grad_norm": 16.5, + "learning_rate": 2.203366205136502e-06, + "loss": 0.8055, + "step": 5332 + }, + { + "epoch": 0.45445249254367276, + "grad_norm": 17.875, + "learning_rate": 2.2032575177272473e-06, + "loss": 1.0661, + "step": 5333 + }, + { + "epoch": 0.4545377077119727, + "grad_norm": 14.625, + "learning_rate": 2.2031488130916706e-06, + "loss": 0.9893, + "step": 5334 + }, + { + "epoch": 0.4546229228802727, + "grad_norm": 13.0, + "learning_rate": 2.203040091231737e-06, + "loss": 0.6344, + "step": 5335 + }, + { + "epoch": 0.45470813804857263, + "grad_norm": 10.4375, + "learning_rate": 2.20293135214941e-06, + "loss": 0.4345, + "step": 5336 + }, + { + "epoch": 0.4547933532168726, + "grad_norm": 15.0625, + "learning_rate": 2.202822595846656e-06, + "loss": 0.5901, + "step": 5337 + }, + { + "epoch": 0.45487856838517254, + "grad_norm": 28.0, + "learning_rate": 2.2027138223254386e-06, + "loss": 1.2241, + "step": 5338 + }, + { + "epoch": 0.4549637835534725, + "grad_norm": 19.375, + "learning_rate": 2.2026050315877253e-06, + "loss": 0.8944, + "step": 5339 + }, + { + "epoch": 0.45504899872177246, + "grad_norm": 23.0, + "learning_rate": 2.20249622363548e-06, + "loss": 0.8549, + "step": 5340 + }, + { + "epoch": 0.4551342138900724, + "grad_norm": 21.75, + "learning_rate": 2.2023873984706713e-06, + "loss": 1.0132, + "step": 5341 + }, + { + "epoch": 0.45521942905837237, + "grad_norm": 16.375, + "learning_rate": 2.2022785560952643e-06, + "loss": 0.8309, + "step": 5342 + }, + { + "epoch": 0.45530464422667233, + "grad_norm": 15.5625, + "learning_rate": 2.202169696511226e-06, + "loss": 0.9, + "step": 5343 + }, + { + "epoch": 0.4553898593949723, + "grad_norm": 17.375, + "learning_rate": 2.2020608197205234e-06, + "loss": 0.6411, + "step": 5344 + }, + { + "epoch": 0.45547507456327224, + "grad_norm": 14.8125, + "learning_rate": 2.201951925725125e-06, + "loss": 0.6685, + "step": 5345 + }, + { + "epoch": 0.4555602897315722, + "grad_norm": 13.6875, + "learning_rate": 2.201843014526998e-06, + "loss": 0.7465, + "step": 5346 + }, + { + "epoch": 0.45564550489987216, + "grad_norm": 19.5, + "learning_rate": 2.2017340861281106e-06, + "loss": 0.7149, + "step": 5347 + }, + { + "epoch": 0.4557307200681721, + "grad_norm": 12.375, + "learning_rate": 2.201625140530431e-06, + "loss": 0.5987, + "step": 5348 + }, + { + "epoch": 0.45581593523647207, + "grad_norm": 18.0, + "learning_rate": 2.201516177735928e-06, + "loss": 0.9223, + "step": 5349 + }, + { + "epoch": 0.455901150404772, + "grad_norm": 11.25, + "learning_rate": 2.201407197746571e-06, + "loss": 0.5194, + "step": 5350 + }, + { + "epoch": 0.455986365573072, + "grad_norm": 20.0, + "learning_rate": 2.2012982005643296e-06, + "loss": 1.2225, + "step": 5351 + }, + { + "epoch": 0.45607158074137194, + "grad_norm": 13.9375, + "learning_rate": 2.2011891861911725e-06, + "loss": 0.7037, + "step": 5352 + }, + { + "epoch": 0.4561567959096719, + "grad_norm": 15.875, + "learning_rate": 2.2010801546290712e-06, + "loss": 0.6425, + "step": 5353 + }, + { + "epoch": 0.45624201107797185, + "grad_norm": 11.75, + "learning_rate": 2.2009711058799943e-06, + "loss": 0.5349, + "step": 5354 + }, + { + "epoch": 0.4563272262462718, + "grad_norm": 24.0, + "learning_rate": 2.200862039945914e-06, + "loss": 1.1827, + "step": 5355 + }, + { + "epoch": 0.45641244141457177, + "grad_norm": 12.75, + "learning_rate": 2.2007529568287995e-06, + "loss": 0.7241, + "step": 5356 + }, + { + "epoch": 0.4564976565828717, + "grad_norm": 18.0, + "learning_rate": 2.200643856530624e-06, + "loss": 0.511, + "step": 5357 + }, + { + "epoch": 0.4565828717511717, + "grad_norm": 15.0, + "learning_rate": 2.200534739053357e-06, + "loss": 0.7864, + "step": 5358 + }, + { + "epoch": 0.45666808691947164, + "grad_norm": 14.25, + "learning_rate": 2.2004256043989723e-06, + "loss": 0.8832, + "step": 5359 + }, + { + "epoch": 0.45675330208777165, + "grad_norm": 14.8125, + "learning_rate": 2.200316452569441e-06, + "loss": 0.6291, + "step": 5360 + }, + { + "epoch": 0.4568385172560716, + "grad_norm": 17.125, + "learning_rate": 2.2002072835667353e-06, + "loss": 0.8866, + "step": 5361 + }, + { + "epoch": 0.45692373242437156, + "grad_norm": 14.4375, + "learning_rate": 2.2000980973928294e-06, + "loss": 0.7363, + "step": 5362 + }, + { + "epoch": 0.4570089475926715, + "grad_norm": 14.1875, + "learning_rate": 2.199988894049695e-06, + "loss": 0.8368, + "step": 5363 + }, + { + "epoch": 0.4570941627609715, + "grad_norm": 14.125, + "learning_rate": 2.1998796735393056e-06, + "loss": 0.929, + "step": 5364 + }, + { + "epoch": 0.45717937792927144, + "grad_norm": 13.9375, + "learning_rate": 2.1997704358636357e-06, + "loss": 0.5921, + "step": 5365 + }, + { + "epoch": 0.4572645930975714, + "grad_norm": 14.75, + "learning_rate": 2.1996611810246592e-06, + "loss": 0.7622, + "step": 5366 + }, + { + "epoch": 0.45734980826587135, + "grad_norm": 15.0, + "learning_rate": 2.1995519090243497e-06, + "loss": 0.4529, + "step": 5367 + }, + { + "epoch": 0.4574350234341713, + "grad_norm": 14.1875, + "learning_rate": 2.1994426198646828e-06, + "loss": 0.7882, + "step": 5368 + }, + { + "epoch": 0.45752023860247126, + "grad_norm": 11.1875, + "learning_rate": 2.1993333135476325e-06, + "loss": 0.4896, + "step": 5369 + }, + { + "epoch": 0.4576054537707712, + "grad_norm": 11.3125, + "learning_rate": 2.199223990075175e-06, + "loss": 0.469, + "step": 5370 + }, + { + "epoch": 0.4576906689390712, + "grad_norm": 16.0, + "learning_rate": 2.1991146494492853e-06, + "loss": 0.7332, + "step": 5371 + }, + { + "epoch": 0.45777588410737113, + "grad_norm": 12.75, + "learning_rate": 2.1990052916719397e-06, + "loss": 0.6996, + "step": 5372 + }, + { + "epoch": 0.4578610992756711, + "grad_norm": 14.6875, + "learning_rate": 2.198895916745114e-06, + "loss": 0.7664, + "step": 5373 + }, + { + "epoch": 0.45794631444397105, + "grad_norm": 32.5, + "learning_rate": 2.198786524670785e-06, + "loss": 0.9411, + "step": 5374 + }, + { + "epoch": 0.458031529612271, + "grad_norm": 15.0625, + "learning_rate": 2.1986771154509293e-06, + "loss": 0.9394, + "step": 5375 + }, + { + "epoch": 0.45811674478057096, + "grad_norm": 13.625, + "learning_rate": 2.1985676890875243e-06, + "loss": 0.6836, + "step": 5376 + }, + { + "epoch": 0.4582019599488709, + "grad_norm": 15.625, + "learning_rate": 2.198458245582547e-06, + "loss": 0.5596, + "step": 5377 + }, + { + "epoch": 0.4582871751171709, + "grad_norm": 12.125, + "learning_rate": 2.198348784937976e-06, + "loss": 0.6925, + "step": 5378 + }, + { + "epoch": 0.45837239028547083, + "grad_norm": 13.375, + "learning_rate": 2.1982393071557884e-06, + "loss": 0.7237, + "step": 5379 + }, + { + "epoch": 0.4584576054537708, + "grad_norm": 15.125, + "learning_rate": 2.198129812237963e-06, + "loss": 0.9014, + "step": 5380 + }, + { + "epoch": 0.45854282062207075, + "grad_norm": 15.5625, + "learning_rate": 2.1980203001864784e-06, + "loss": 0.7805, + "step": 5381 + }, + { + "epoch": 0.4586280357903707, + "grad_norm": 18.875, + "learning_rate": 2.1979107710033135e-06, + "loss": 0.6593, + "step": 5382 + }, + { + "epoch": 0.45871325095867066, + "grad_norm": 17.625, + "learning_rate": 2.1978012246904485e-06, + "loss": 0.7923, + "step": 5383 + }, + { + "epoch": 0.4587984661269706, + "grad_norm": 12.9375, + "learning_rate": 2.197691661249862e-06, + "loss": 0.5387, + "step": 5384 + }, + { + "epoch": 0.4588836812952706, + "grad_norm": 13.125, + "learning_rate": 2.197582080683534e-06, + "loss": 0.6742, + "step": 5385 + }, + { + "epoch": 0.45896889646357053, + "grad_norm": 13.125, + "learning_rate": 2.197472482993445e-06, + "loss": 0.6558, + "step": 5386 + }, + { + "epoch": 0.4590541116318705, + "grad_norm": 14.3125, + "learning_rate": 2.197362868181576e-06, + "loss": 0.735, + "step": 5387 + }, + { + "epoch": 0.45913932680017044, + "grad_norm": 13.0, + "learning_rate": 2.1972532362499064e-06, + "loss": 0.712, + "step": 5388 + }, + { + "epoch": 0.4592245419684704, + "grad_norm": 12.1875, + "learning_rate": 2.197143587200419e-06, + "loss": 0.5596, + "step": 5389 + }, + { + "epoch": 0.45930975713677036, + "grad_norm": 18.875, + "learning_rate": 2.197033921035094e-06, + "loss": 1.0656, + "step": 5390 + }, + { + "epoch": 0.4593949723050703, + "grad_norm": 9.8125, + "learning_rate": 2.196924237755914e-06, + "loss": 0.9935, + "step": 5391 + }, + { + "epoch": 0.45948018747337027, + "grad_norm": 14.0625, + "learning_rate": 2.1968145373648607e-06, + "loss": 0.6395, + "step": 5392 + }, + { + "epoch": 0.45956540264167023, + "grad_norm": 14.3125, + "learning_rate": 2.1967048198639167e-06, + "loss": 0.8226, + "step": 5393 + }, + { + "epoch": 0.4596506178099702, + "grad_norm": 10.3125, + "learning_rate": 2.196595085255064e-06, + "loss": 0.4611, + "step": 5394 + }, + { + "epoch": 0.45973583297827014, + "grad_norm": 14.625, + "learning_rate": 2.1964853335402873e-06, + "loss": 0.6716, + "step": 5395 + }, + { + "epoch": 0.4598210481465701, + "grad_norm": 11.875, + "learning_rate": 2.1963755647215684e-06, + "loss": 0.3015, + "step": 5396 + }, + { + "epoch": 0.45990626331487006, + "grad_norm": 9.625, + "learning_rate": 2.1962657788008915e-06, + "loss": 0.3835, + "step": 5397 + }, + { + "epoch": 0.45999147848317, + "grad_norm": 11.6875, + "learning_rate": 2.1961559757802405e-06, + "loss": 0.4582, + "step": 5398 + }, + { + "epoch": 0.46007669365146997, + "grad_norm": 13.5625, + "learning_rate": 2.1960461556615993e-06, + "loss": 0.8695, + "step": 5399 + }, + { + "epoch": 0.4601619088197699, + "grad_norm": 10.75, + "learning_rate": 2.195936318446953e-06, + "loss": 0.5271, + "step": 5400 + }, + { + "epoch": 0.4602471239880699, + "grad_norm": 13.1875, + "learning_rate": 2.195826464138286e-06, + "loss": 0.6792, + "step": 5401 + }, + { + "epoch": 0.46033233915636984, + "grad_norm": 10.375, + "learning_rate": 2.1957165927375836e-06, + "loss": 0.458, + "step": 5402 + }, + { + "epoch": 0.4604175543246698, + "grad_norm": 13.0, + "learning_rate": 2.1956067042468317e-06, + "loss": 0.6812, + "step": 5403 + }, + { + "epoch": 0.46050276949296975, + "grad_norm": 17.25, + "learning_rate": 2.1954967986680155e-06, + "loss": 0.7756, + "step": 5404 + }, + { + "epoch": 0.4605879846612697, + "grad_norm": 13.875, + "learning_rate": 2.1953868760031218e-06, + "loss": 0.8689, + "step": 5405 + }, + { + "epoch": 0.46067319982956967, + "grad_norm": 16.375, + "learning_rate": 2.1952769362541363e-06, + "loss": 0.6738, + "step": 5406 + }, + { + "epoch": 0.4607584149978696, + "grad_norm": 10.375, + "learning_rate": 2.195166979423046e-06, + "loss": 0.442, + "step": 5407 + }, + { + "epoch": 0.4608436301661696, + "grad_norm": 13.8125, + "learning_rate": 2.1950570055118376e-06, + "loss": 0.7507, + "step": 5408 + }, + { + "epoch": 0.46092884533446954, + "grad_norm": 16.125, + "learning_rate": 2.1949470145224995e-06, + "loss": 0.6132, + "step": 5409 + }, + { + "epoch": 0.4610140605027695, + "grad_norm": 11.75, + "learning_rate": 2.194837006457018e-06, + "loss": 0.8065, + "step": 5410 + }, + { + "epoch": 0.46109927567106945, + "grad_norm": 15.125, + "learning_rate": 2.1947269813173823e-06, + "loss": 0.6637, + "step": 5411 + }, + { + "epoch": 0.4611844908393694, + "grad_norm": 11.5625, + "learning_rate": 2.1946169391055795e-06, + "loss": 0.4672, + "step": 5412 + }, + { + "epoch": 0.46126970600766937, + "grad_norm": 12.1875, + "learning_rate": 2.194506879823599e-06, + "loss": 0.4659, + "step": 5413 + }, + { + "epoch": 0.4613549211759693, + "grad_norm": 14.8125, + "learning_rate": 2.1943968034734292e-06, + "loss": 0.7119, + "step": 5414 + }, + { + "epoch": 0.4614401363442693, + "grad_norm": 15.0625, + "learning_rate": 2.1942867100570597e-06, + "loss": 0.8455, + "step": 5415 + }, + { + "epoch": 0.46152535151256924, + "grad_norm": 13.125, + "learning_rate": 2.1941765995764796e-06, + "loss": 0.8413, + "step": 5416 + }, + { + "epoch": 0.4616105666808692, + "grad_norm": 13.25, + "learning_rate": 2.194066472033679e-06, + "loss": 0.7852, + "step": 5417 + }, + { + "epoch": 0.46169578184916915, + "grad_norm": 9.875, + "learning_rate": 2.193956327430648e-06, + "loss": 0.3056, + "step": 5418 + }, + { + "epoch": 0.4617809970174691, + "grad_norm": 13.3125, + "learning_rate": 2.1938461657693767e-06, + "loss": 0.7517, + "step": 5419 + }, + { + "epoch": 0.46186621218576907, + "grad_norm": 33.0, + "learning_rate": 2.1937359870518567e-06, + "loss": 0.5952, + "step": 5420 + }, + { + "epoch": 0.461951427354069, + "grad_norm": 15.5, + "learning_rate": 2.193625791280078e-06, + "loss": 0.5737, + "step": 5421 + }, + { + "epoch": 0.462036642522369, + "grad_norm": 15.1875, + "learning_rate": 2.193515578456032e-06, + "loss": 0.8829, + "step": 5422 + }, + { + "epoch": 0.46212185769066894, + "grad_norm": 16.125, + "learning_rate": 2.1934053485817113e-06, + "loss": 0.715, + "step": 5423 + }, + { + "epoch": 0.4622070728589689, + "grad_norm": 14.1875, + "learning_rate": 2.193295101659107e-06, + "loss": 0.985, + "step": 5424 + }, + { + "epoch": 0.46229228802726885, + "grad_norm": 12.0625, + "learning_rate": 2.1931848376902115e-06, + "loss": 0.4106, + "step": 5425 + }, + { + "epoch": 0.4623775031955688, + "grad_norm": 19.0, + "learning_rate": 2.193074556677018e-06, + "loss": 0.8552, + "step": 5426 + }, + { + "epoch": 0.46246271836386876, + "grad_norm": 14.375, + "learning_rate": 2.1929642586215186e-06, + "loss": 0.7395, + "step": 5427 + }, + { + "epoch": 0.4625479335321687, + "grad_norm": 14.625, + "learning_rate": 2.192853943525707e-06, + "loss": 0.6328, + "step": 5428 + }, + { + "epoch": 0.4626331487004687, + "grad_norm": 10.6875, + "learning_rate": 2.1927436113915762e-06, + "loss": 0.3064, + "step": 5429 + }, + { + "epoch": 0.46271836386876863, + "grad_norm": 23.375, + "learning_rate": 2.1926332622211206e-06, + "loss": 0.9315, + "step": 5430 + }, + { + "epoch": 0.4628035790370686, + "grad_norm": 13.375, + "learning_rate": 2.192522896016334e-06, + "loss": 0.6818, + "step": 5431 + }, + { + "epoch": 0.46288879420536855, + "grad_norm": 14.875, + "learning_rate": 2.192412512779211e-06, + "loss": 0.6938, + "step": 5432 + }, + { + "epoch": 0.4629740093736685, + "grad_norm": 13.875, + "learning_rate": 2.192302112511746e-06, + "loss": 0.6912, + "step": 5433 + }, + { + "epoch": 0.46305922454196846, + "grad_norm": 14.8125, + "learning_rate": 2.1921916952159345e-06, + "loss": 0.651, + "step": 5434 + }, + { + "epoch": 0.4631444397102684, + "grad_norm": 12.875, + "learning_rate": 2.1920812608937713e-06, + "loss": 0.59, + "step": 5435 + }, + { + "epoch": 0.4632296548785684, + "grad_norm": 11.875, + "learning_rate": 2.1919708095472526e-06, + "loss": 0.6347, + "step": 5436 + }, + { + "epoch": 0.46331487004686833, + "grad_norm": 13.8125, + "learning_rate": 2.1918603411783743e-06, + "loss": 0.5928, + "step": 5437 + }, + { + "epoch": 0.4634000852151683, + "grad_norm": 14.3125, + "learning_rate": 2.191749855789132e-06, + "loss": 0.7762, + "step": 5438 + }, + { + "epoch": 0.46348530038346825, + "grad_norm": 16.75, + "learning_rate": 2.1916393533815234e-06, + "loss": 0.8205, + "step": 5439 + }, + { + "epoch": 0.4635705155517682, + "grad_norm": 12.25, + "learning_rate": 2.1915288339575445e-06, + "loss": 0.2623, + "step": 5440 + }, + { + "epoch": 0.46365573072006816, + "grad_norm": 13.3125, + "learning_rate": 2.1914182975191927e-06, + "loss": 0.3928, + "step": 5441 + }, + { + "epoch": 0.4637409458883681, + "grad_norm": 12.5, + "learning_rate": 2.191307744068466e-06, + "loss": 0.6305, + "step": 5442 + }, + { + "epoch": 0.4638261610566681, + "grad_norm": 15.625, + "learning_rate": 2.1911971736073617e-06, + "loss": 0.6561, + "step": 5443 + }, + { + "epoch": 0.46391137622496803, + "grad_norm": 19.125, + "learning_rate": 2.191086586137878e-06, + "loss": 0.9618, + "step": 5444 + }, + { + "epoch": 0.463996591393268, + "grad_norm": 12.3125, + "learning_rate": 2.190975981662013e-06, + "loss": 0.3177, + "step": 5445 + }, + { + "epoch": 0.46408180656156794, + "grad_norm": 15.3125, + "learning_rate": 2.190865360181766e-06, + "loss": 0.8093, + "step": 5446 + }, + { + "epoch": 0.4641670217298679, + "grad_norm": 13.6875, + "learning_rate": 2.190754721699136e-06, + "loss": 0.7158, + "step": 5447 + }, + { + "epoch": 0.46425223689816786, + "grad_norm": 16.125, + "learning_rate": 2.1906440662161215e-06, + "loss": 0.5968, + "step": 5448 + }, + { + "epoch": 0.4643374520664678, + "grad_norm": 19.75, + "learning_rate": 2.1905333937347235e-06, + "loss": 1.0111, + "step": 5449 + }, + { + "epoch": 0.46442266723476777, + "grad_norm": 14.875, + "learning_rate": 2.190422704256941e-06, + "loss": 0.7625, + "step": 5450 + }, + { + "epoch": 0.46450788240306773, + "grad_norm": 14.0625, + "learning_rate": 2.190311997784775e-06, + "loss": 0.7636, + "step": 5451 + }, + { + "epoch": 0.4645930975713677, + "grad_norm": 16.375, + "learning_rate": 2.1902012743202254e-06, + "loss": 1.1334, + "step": 5452 + }, + { + "epoch": 0.46467831273966764, + "grad_norm": 13.125, + "learning_rate": 2.190090533865293e-06, + "loss": 0.7281, + "step": 5453 + }, + { + "epoch": 0.4647635279079676, + "grad_norm": 21.25, + "learning_rate": 2.18997977642198e-06, + "loss": 0.9653, + "step": 5454 + }, + { + "epoch": 0.46484874307626756, + "grad_norm": 11.75, + "learning_rate": 2.189869001992287e-06, + "loss": 0.4201, + "step": 5455 + }, + { + "epoch": 0.4649339582445675, + "grad_norm": 15.5625, + "learning_rate": 2.1897582105782156e-06, + "loss": 0.6026, + "step": 5456 + }, + { + "epoch": 0.46501917341286747, + "grad_norm": 11.875, + "learning_rate": 2.1896474021817686e-06, + "loss": 0.5422, + "step": 5457 + }, + { + "epoch": 0.4651043885811674, + "grad_norm": 15.5, + "learning_rate": 2.189536576804948e-06, + "loss": 0.4348, + "step": 5458 + }, + { + "epoch": 0.4651896037494674, + "grad_norm": 15.1875, + "learning_rate": 2.189425734449757e-06, + "loss": 0.5475, + "step": 5459 + }, + { + "epoch": 0.46527481891776734, + "grad_norm": 19.5, + "learning_rate": 2.189314875118198e-06, + "loss": 0.68, + "step": 5460 + }, + { + "epoch": 0.4653600340860673, + "grad_norm": 15.6875, + "learning_rate": 2.189203998812275e-06, + "loss": 0.7498, + "step": 5461 + }, + { + "epoch": 0.46544524925436725, + "grad_norm": 15.6875, + "learning_rate": 2.189093105533991e-06, + "loss": 1.0508, + "step": 5462 + }, + { + "epoch": 0.4655304644226672, + "grad_norm": 13.6875, + "learning_rate": 2.18898219528535e-06, + "loss": 0.6773, + "step": 5463 + }, + { + "epoch": 0.46561567959096717, + "grad_norm": 16.0, + "learning_rate": 2.188871268068357e-06, + "loss": 0.7554, + "step": 5464 + }, + { + "epoch": 0.4657008947592671, + "grad_norm": 12.1875, + "learning_rate": 2.188760323885016e-06, + "loss": 0.6219, + "step": 5465 + }, + { + "epoch": 0.4657861099275671, + "grad_norm": 11.875, + "learning_rate": 2.188649362737332e-06, + "loss": 0.5974, + "step": 5466 + }, + { + "epoch": 0.46587132509586704, + "grad_norm": 11.6875, + "learning_rate": 2.18853838462731e-06, + "loss": 0.5459, + "step": 5467 + }, + { + "epoch": 0.465956540264167, + "grad_norm": 28.375, + "learning_rate": 2.1884273895569553e-06, + "loss": 0.9978, + "step": 5468 + }, + { + "epoch": 0.46604175543246695, + "grad_norm": 10.5625, + "learning_rate": 2.1883163775282744e-06, + "loss": 0.4482, + "step": 5469 + }, + { + "epoch": 0.4661269706007669, + "grad_norm": 16.875, + "learning_rate": 2.188205348543273e-06, + "loss": 0.6727, + "step": 5470 + }, + { + "epoch": 0.46621218576906687, + "grad_norm": 14.3125, + "learning_rate": 2.1880943026039575e-06, + "loss": 0.5859, + "step": 5471 + }, + { + "epoch": 0.4662974009373668, + "grad_norm": 12.875, + "learning_rate": 2.1879832397123347e-06, + "loss": 0.4882, + "step": 5472 + }, + { + "epoch": 0.46638261610566684, + "grad_norm": 14.125, + "learning_rate": 2.1878721598704112e-06, + "loss": 0.9089, + "step": 5473 + }, + { + "epoch": 0.4664678312739668, + "grad_norm": 13.3125, + "learning_rate": 2.187761063080195e-06, + "loss": 0.659, + "step": 5474 + }, + { + "epoch": 0.46655304644226675, + "grad_norm": 18.125, + "learning_rate": 2.187649949343693e-06, + "loss": 1.1222, + "step": 5475 + }, + { + "epoch": 0.4666382616105667, + "grad_norm": 24.0, + "learning_rate": 2.187538818662914e-06, + "loss": 1.0194, + "step": 5476 + }, + { + "epoch": 0.46672347677886666, + "grad_norm": 14.6875, + "learning_rate": 2.1874276710398657e-06, + "loss": 0.8416, + "step": 5477 + }, + { + "epoch": 0.4668086919471666, + "grad_norm": 15.6875, + "learning_rate": 2.1873165064765566e-06, + "loss": 0.5463, + "step": 5478 + }, + { + "epoch": 0.4668939071154666, + "grad_norm": 21.875, + "learning_rate": 2.1872053249749957e-06, + "loss": 1.1472, + "step": 5479 + }, + { + "epoch": 0.46697912228376653, + "grad_norm": 13.125, + "learning_rate": 2.187094126537192e-06, + "loss": 0.6365, + "step": 5480 + }, + { + "epoch": 0.4670643374520665, + "grad_norm": 18.125, + "learning_rate": 2.1869829111651553e-06, + "loss": 1.2176, + "step": 5481 + }, + { + "epoch": 0.46714955262036645, + "grad_norm": 11.1875, + "learning_rate": 2.1868716788608952e-06, + "loss": 0.5142, + "step": 5482 + }, + { + "epoch": 0.4672347677886664, + "grad_norm": 17.625, + "learning_rate": 2.186760429626422e-06, + "loss": 1.1302, + "step": 5483 + }, + { + "epoch": 0.46731998295696636, + "grad_norm": 12.375, + "learning_rate": 2.1866491634637456e-06, + "loss": 0.5261, + "step": 5484 + }, + { + "epoch": 0.4674051981252663, + "grad_norm": 15.0, + "learning_rate": 2.1865378803748764e-06, + "loss": 0.9297, + "step": 5485 + }, + { + "epoch": 0.4674904132935663, + "grad_norm": 13.25, + "learning_rate": 2.186426580361827e-06, + "loss": 0.7438, + "step": 5486 + }, + { + "epoch": 0.46757562846186623, + "grad_norm": 12.8125, + "learning_rate": 2.1863152634266066e-06, + "loss": 0.8134, + "step": 5487 + }, + { + "epoch": 0.4676608436301662, + "grad_norm": 16.875, + "learning_rate": 2.1862039295712285e-06, + "loss": 0.5644, + "step": 5488 + }, + { + "epoch": 0.46774605879846615, + "grad_norm": 12.0625, + "learning_rate": 2.186092578797704e-06, + "loss": 0.5421, + "step": 5489 + }, + { + "epoch": 0.4678312739667661, + "grad_norm": 14.0625, + "learning_rate": 2.1859812111080446e-06, + "loss": 0.45, + "step": 5490 + }, + { + "epoch": 0.46791648913506606, + "grad_norm": 15.75, + "learning_rate": 2.1858698265042643e-06, + "loss": 0.6468, + "step": 5491 + }, + { + "epoch": 0.468001704303366, + "grad_norm": 21.125, + "learning_rate": 2.1857584249883746e-06, + "loss": 0.8439, + "step": 5492 + }, + { + "epoch": 0.468086919471666, + "grad_norm": 14.9375, + "learning_rate": 2.1856470065623896e-06, + "loss": 0.8426, + "step": 5493 + }, + { + "epoch": 0.46817213463996593, + "grad_norm": 17.375, + "learning_rate": 2.185535571228322e-06, + "loss": 0.9854, + "step": 5494 + }, + { + "epoch": 0.4682573498082659, + "grad_norm": 14.5625, + "learning_rate": 2.1854241189881862e-06, + "loss": 0.6358, + "step": 5495 + }, + { + "epoch": 0.46834256497656584, + "grad_norm": 19.625, + "learning_rate": 2.185312649843996e-06, + "loss": 0.9284, + "step": 5496 + }, + { + "epoch": 0.4684277801448658, + "grad_norm": 14.5, + "learning_rate": 2.1852011637977656e-06, + "loss": 0.7636, + "step": 5497 + }, + { + "epoch": 0.46851299531316576, + "grad_norm": 27.375, + "learning_rate": 2.1850896608515093e-06, + "loss": 0.5462, + "step": 5498 + }, + { + "epoch": 0.4685982104814657, + "grad_norm": 10.875, + "learning_rate": 2.184978141007243e-06, + "loss": 0.4294, + "step": 5499 + }, + { + "epoch": 0.4686834256497657, + "grad_norm": 10.375, + "learning_rate": 2.1848666042669814e-06, + "loss": 0.4634, + "step": 5500 + }, + { + "epoch": 0.46876864081806563, + "grad_norm": 12.625, + "learning_rate": 2.18475505063274e-06, + "loss": 0.5324, + "step": 5501 + }, + { + "epoch": 0.4688538559863656, + "grad_norm": 14.1875, + "learning_rate": 2.1846434801065347e-06, + "loss": 0.3975, + "step": 5502 + }, + { + "epoch": 0.46893907115466554, + "grad_norm": 15.1875, + "learning_rate": 2.1845318926903823e-06, + "loss": 0.7668, + "step": 5503 + }, + { + "epoch": 0.4690242863229655, + "grad_norm": 14.125, + "learning_rate": 2.184420288386299e-06, + "loss": 0.7345, + "step": 5504 + }, + { + "epoch": 0.46910950149126546, + "grad_norm": 11.125, + "learning_rate": 2.184308667196301e-06, + "loss": 0.4143, + "step": 5505 + }, + { + "epoch": 0.4691947166595654, + "grad_norm": 13.0, + "learning_rate": 2.184197029122406e-06, + "loss": 0.7244, + "step": 5506 + }, + { + "epoch": 0.46927993182786537, + "grad_norm": 17.25, + "learning_rate": 2.1840853741666314e-06, + "loss": 1.0593, + "step": 5507 + }, + { + "epoch": 0.4693651469961653, + "grad_norm": 13.5625, + "learning_rate": 2.183973702330994e-06, + "loss": 0.8069, + "step": 5508 + }, + { + "epoch": 0.4694503621644653, + "grad_norm": 12.6875, + "learning_rate": 2.183862013617514e-06, + "loss": 0.7345, + "step": 5509 + }, + { + "epoch": 0.46953557733276524, + "grad_norm": 13.6875, + "learning_rate": 2.1837503080282074e-06, + "loss": 0.9162, + "step": 5510 + }, + { + "epoch": 0.4696207925010652, + "grad_norm": 12.4375, + "learning_rate": 2.1836385855650936e-06, + "loss": 0.5146, + "step": 5511 + }, + { + "epoch": 0.46970600766936516, + "grad_norm": 11.625, + "learning_rate": 2.1835268462301928e-06, + "loss": 0.556, + "step": 5512 + }, + { + "epoch": 0.4697912228376651, + "grad_norm": 13.4375, + "learning_rate": 2.183415090025522e-06, + "loss": 0.5341, + "step": 5513 + }, + { + "epoch": 0.46987643800596507, + "grad_norm": 11.8125, + "learning_rate": 2.1833033169531027e-06, + "loss": 0.6581, + "step": 5514 + }, + { + "epoch": 0.469961653174265, + "grad_norm": 16.875, + "learning_rate": 2.1831915270149535e-06, + "loss": 0.6958, + "step": 5515 + }, + { + "epoch": 0.470046868342565, + "grad_norm": 11.75, + "learning_rate": 2.1830797202130953e-06, + "loss": 0.54, + "step": 5516 + }, + { + "epoch": 0.47013208351086494, + "grad_norm": 12.6875, + "learning_rate": 2.182967896549548e-06, + "loss": 0.529, + "step": 5517 + }, + { + "epoch": 0.4702172986791649, + "grad_norm": 11.25, + "learning_rate": 2.1828560560263327e-06, + "loss": 0.3441, + "step": 5518 + }, + { + "epoch": 0.47030251384746485, + "grad_norm": 13.3125, + "learning_rate": 2.1827441986454708e-06, + "loss": 0.6511, + "step": 5519 + }, + { + "epoch": 0.4703877290157648, + "grad_norm": 14.75, + "learning_rate": 2.1826323244089827e-06, + "loss": 0.8134, + "step": 5520 + }, + { + "epoch": 0.47047294418406477, + "grad_norm": 17.875, + "learning_rate": 2.182520433318891e-06, + "loss": 0.8867, + "step": 5521 + }, + { + "epoch": 0.4705581593523647, + "grad_norm": 13.75, + "learning_rate": 2.182408525377217e-06, + "loss": 0.8701, + "step": 5522 + }, + { + "epoch": 0.4706433745206647, + "grad_norm": 10.8125, + "learning_rate": 2.1822966005859835e-06, + "loss": 0.4297, + "step": 5523 + }, + { + "epoch": 0.47072858968896464, + "grad_norm": 13.75, + "learning_rate": 2.1821846589472133e-06, + "loss": 0.537, + "step": 5524 + }, + { + "epoch": 0.4708138048572646, + "grad_norm": 13.9375, + "learning_rate": 2.1820727004629284e-06, + "loss": 0.7734, + "step": 5525 + }, + { + "epoch": 0.47089902002556455, + "grad_norm": 14.0, + "learning_rate": 2.181960725135153e-06, + "loss": 0.7162, + "step": 5526 + }, + { + "epoch": 0.4709842351938645, + "grad_norm": 11.5, + "learning_rate": 2.18184873296591e-06, + "loss": 0.5148, + "step": 5527 + }, + { + "epoch": 0.47106945036216447, + "grad_norm": 15.4375, + "learning_rate": 2.181736723957223e-06, + "loss": 0.6375, + "step": 5528 + }, + { + "epoch": 0.4711546655304644, + "grad_norm": 16.25, + "learning_rate": 2.1816246981111168e-06, + "loss": 0.7674, + "step": 5529 + }, + { + "epoch": 0.4712398806987644, + "grad_norm": 11.625, + "learning_rate": 2.1815126554296156e-06, + "loss": 0.5394, + "step": 5530 + }, + { + "epoch": 0.47132509586706434, + "grad_norm": 15.4375, + "learning_rate": 2.181400595914744e-06, + "loss": 0.9407, + "step": 5531 + }, + { + "epoch": 0.4714103110353643, + "grad_norm": 16.625, + "learning_rate": 2.1812885195685265e-06, + "loss": 0.8713, + "step": 5532 + }, + { + "epoch": 0.47149552620366425, + "grad_norm": 19.0, + "learning_rate": 2.181176426392989e-06, + "loss": 0.7733, + "step": 5533 + }, + { + "epoch": 0.4715807413719642, + "grad_norm": 12.6875, + "learning_rate": 2.181064316390158e-06, + "loss": 0.5568, + "step": 5534 + }, + { + "epoch": 0.47166595654026416, + "grad_norm": 16.75, + "learning_rate": 2.1809521895620577e-06, + "loss": 0.853, + "step": 5535 + }, + { + "epoch": 0.4717511717085641, + "grad_norm": 13.75, + "learning_rate": 2.1808400459107157e-06, + "loss": 0.5345, + "step": 5536 + }, + { + "epoch": 0.4718363868768641, + "grad_norm": 26.5, + "learning_rate": 2.180727885438158e-06, + "loss": 1.1742, + "step": 5537 + }, + { + "epoch": 0.47192160204516403, + "grad_norm": 9.6875, + "learning_rate": 2.180615708146411e-06, + "loss": 0.2815, + "step": 5538 + }, + { + "epoch": 0.472006817213464, + "grad_norm": 17.125, + "learning_rate": 2.1805035140375028e-06, + "loss": 0.9562, + "step": 5539 + }, + { + "epoch": 0.47209203238176395, + "grad_norm": 18.75, + "learning_rate": 2.1803913031134603e-06, + "loss": 1.208, + "step": 5540 + }, + { + "epoch": 0.4721772475500639, + "grad_norm": 11.875, + "learning_rate": 2.180279075376311e-06, + "loss": 0.6441, + "step": 5541 + }, + { + "epoch": 0.47226246271836386, + "grad_norm": 15.125, + "learning_rate": 2.1801668308280836e-06, + "loss": 0.7921, + "step": 5542 + }, + { + "epoch": 0.4723476778866638, + "grad_norm": 12.1875, + "learning_rate": 2.1800545694708063e-06, + "loss": 0.2958, + "step": 5543 + }, + { + "epoch": 0.4724328930549638, + "grad_norm": 14.125, + "learning_rate": 2.179942291306508e-06, + "loss": 0.8886, + "step": 5544 + }, + { + "epoch": 0.47251810822326373, + "grad_norm": 16.125, + "learning_rate": 2.179829996337217e-06, + "loss": 0.7413, + "step": 5545 + }, + { + "epoch": 0.4726033233915637, + "grad_norm": 14.4375, + "learning_rate": 2.179717684564963e-06, + "loss": 0.5756, + "step": 5546 + }, + { + "epoch": 0.47268853855986365, + "grad_norm": 12.8125, + "learning_rate": 2.1796053559917755e-06, + "loss": 0.5664, + "step": 5547 + }, + { + "epoch": 0.4727737537281636, + "grad_norm": 15.75, + "learning_rate": 2.179493010619684e-06, + "loss": 0.6699, + "step": 5548 + }, + { + "epoch": 0.47285896889646356, + "grad_norm": 10.6875, + "learning_rate": 2.1793806484507197e-06, + "loss": 0.4686, + "step": 5549 + }, + { + "epoch": 0.4729441840647635, + "grad_norm": 16.25, + "learning_rate": 2.1792682694869127e-06, + "loss": 0.7772, + "step": 5550 + }, + { + "epoch": 0.4730293992330635, + "grad_norm": 13.9375, + "learning_rate": 2.1791558737302933e-06, + "loss": 0.7444, + "step": 5551 + }, + { + "epoch": 0.47311461440136343, + "grad_norm": 15.375, + "learning_rate": 2.1790434611828933e-06, + "loss": 0.9241, + "step": 5552 + }, + { + "epoch": 0.4731998295696634, + "grad_norm": 15.0625, + "learning_rate": 2.178931031846743e-06, + "loss": 0.7668, + "step": 5553 + }, + { + "epoch": 0.47328504473796335, + "grad_norm": 14.625, + "learning_rate": 2.1788185857238754e-06, + "loss": 0.8363, + "step": 5554 + }, + { + "epoch": 0.4733702599062633, + "grad_norm": 12.375, + "learning_rate": 2.1787061228163216e-06, + "loss": 0.3811, + "step": 5555 + }, + { + "epoch": 0.47345547507456326, + "grad_norm": 14.75, + "learning_rate": 2.1785936431261144e-06, + "loss": 0.8861, + "step": 5556 + }, + { + "epoch": 0.4735406902428632, + "grad_norm": 13.625, + "learning_rate": 2.1784811466552863e-06, + "loss": 0.6705, + "step": 5557 + }, + { + "epoch": 0.4736259054111632, + "grad_norm": 12.25, + "learning_rate": 2.17836863340587e-06, + "loss": 0.3512, + "step": 5558 + }, + { + "epoch": 0.47371112057946313, + "grad_norm": 12.375, + "learning_rate": 2.1782561033798994e-06, + "loss": 0.7501, + "step": 5559 + }, + { + "epoch": 0.4737963357477631, + "grad_norm": 18.5, + "learning_rate": 2.1781435565794074e-06, + "loss": 1.0052, + "step": 5560 + }, + { + "epoch": 0.47388155091606304, + "grad_norm": 15.4375, + "learning_rate": 2.178030993006428e-06, + "loss": 0.7937, + "step": 5561 + }, + { + "epoch": 0.473966766084363, + "grad_norm": 12.25, + "learning_rate": 2.1779184126629954e-06, + "loss": 0.6188, + "step": 5562 + }, + { + "epoch": 0.47405198125266296, + "grad_norm": 12.625, + "learning_rate": 2.1778058155511437e-06, + "loss": 0.2805, + "step": 5563 + }, + { + "epoch": 0.4741371964209629, + "grad_norm": 11.5625, + "learning_rate": 2.1776932016729085e-06, + "loss": 0.6485, + "step": 5564 + }, + { + "epoch": 0.47422241158926287, + "grad_norm": 17.625, + "learning_rate": 2.177580571030324e-06, + "loss": 0.9021, + "step": 5565 + }, + { + "epoch": 0.47430762675756283, + "grad_norm": 11.0, + "learning_rate": 2.1774679236254255e-06, + "loss": 0.4493, + "step": 5566 + }, + { + "epoch": 0.4743928419258628, + "grad_norm": 15.5625, + "learning_rate": 2.1773552594602495e-06, + "loss": 0.7791, + "step": 5567 + }, + { + "epoch": 0.47447805709416274, + "grad_norm": 11.625, + "learning_rate": 2.177242578536831e-06, + "loss": 0.6009, + "step": 5568 + }, + { + "epoch": 0.4745632722624627, + "grad_norm": 18.0, + "learning_rate": 2.1771298808572065e-06, + "loss": 0.8438, + "step": 5569 + }, + { + "epoch": 0.47464848743076266, + "grad_norm": 13.0625, + "learning_rate": 2.1770171664234134e-06, + "loss": 0.5735, + "step": 5570 + }, + { + "epoch": 0.4747337025990626, + "grad_norm": 14.625, + "learning_rate": 2.1769044352374877e-06, + "loss": 0.607, + "step": 5571 + }, + { + "epoch": 0.47481891776736257, + "grad_norm": 11.75, + "learning_rate": 2.1767916873014667e-06, + "loss": 0.5811, + "step": 5572 + }, + { + "epoch": 0.4749041329356625, + "grad_norm": 14.625, + "learning_rate": 2.1766789226173877e-06, + "loss": 0.7509, + "step": 5573 + }, + { + "epoch": 0.4749893481039625, + "grad_norm": 9.5, + "learning_rate": 2.176566141187289e-06, + "loss": 0.5692, + "step": 5574 + }, + { + "epoch": 0.47507456327226244, + "grad_norm": 13.625, + "learning_rate": 2.1764533430132087e-06, + "loss": 0.7993, + "step": 5575 + }, + { + "epoch": 0.4751597784405624, + "grad_norm": 21.125, + "learning_rate": 2.1763405280971846e-06, + "loss": 1.1098, + "step": 5576 + }, + { + "epoch": 0.47524499360886235, + "grad_norm": 14.5, + "learning_rate": 2.1762276964412556e-06, + "loss": 0.5298, + "step": 5577 + }, + { + "epoch": 0.4753302087771623, + "grad_norm": 15.5, + "learning_rate": 2.1761148480474604e-06, + "loss": 0.9666, + "step": 5578 + }, + { + "epoch": 0.47541542394546227, + "grad_norm": 12.5, + "learning_rate": 2.176001982917839e-06, + "loss": 0.4802, + "step": 5579 + }, + { + "epoch": 0.4755006391137622, + "grad_norm": 17.0, + "learning_rate": 2.1758891010544307e-06, + "loss": 1.1795, + "step": 5580 + }, + { + "epoch": 0.4755858542820622, + "grad_norm": 25.25, + "learning_rate": 2.175776202459275e-06, + "loss": 0.6754, + "step": 5581 + }, + { + "epoch": 0.47567106945036214, + "grad_norm": 15.6875, + "learning_rate": 2.175663287134413e-06, + "loss": 0.8893, + "step": 5582 + }, + { + "epoch": 0.4757562846186621, + "grad_norm": 20.0, + "learning_rate": 2.175550355081884e-06, + "loss": 0.8174, + "step": 5583 + }, + { + "epoch": 0.47584149978696205, + "grad_norm": 10.75, + "learning_rate": 2.1754374063037294e-06, + "loss": 0.3816, + "step": 5584 + }, + { + "epoch": 0.475926714955262, + "grad_norm": 12.375, + "learning_rate": 2.17532444080199e-06, + "loss": 0.6221, + "step": 5585 + }, + { + "epoch": 0.476011930123562, + "grad_norm": 15.0625, + "learning_rate": 2.175211458578708e-06, + "loss": 0.7942, + "step": 5586 + }, + { + "epoch": 0.476097145291862, + "grad_norm": 16.75, + "learning_rate": 2.175098459635924e-06, + "loss": 1.0592, + "step": 5587 + }, + { + "epoch": 0.47618236046016194, + "grad_norm": 11.9375, + "learning_rate": 2.1749854439756813e-06, + "loss": 0.4611, + "step": 5588 + }, + { + "epoch": 0.4762675756284619, + "grad_norm": 16.25, + "learning_rate": 2.1748724116000213e-06, + "loss": 0.7988, + "step": 5589 + }, + { + "epoch": 0.47635279079676185, + "grad_norm": 13.0625, + "learning_rate": 2.1747593625109863e-06, + "loss": 0.5668, + "step": 5590 + }, + { + "epoch": 0.4764380059650618, + "grad_norm": 11.6875, + "learning_rate": 2.17464629671062e-06, + "loss": 0.406, + "step": 5591 + }, + { + "epoch": 0.47652322113336176, + "grad_norm": 13.5625, + "learning_rate": 2.1745332142009654e-06, + "loss": 0.7513, + "step": 5592 + }, + { + "epoch": 0.4766084363016617, + "grad_norm": 14.5625, + "learning_rate": 2.174420114984066e-06, + "loss": 0.8797, + "step": 5593 + }, + { + "epoch": 0.4766936514699617, + "grad_norm": 12.75, + "learning_rate": 2.1743069990619653e-06, + "loss": 0.5748, + "step": 5594 + }, + { + "epoch": 0.47677886663826163, + "grad_norm": 17.0, + "learning_rate": 2.174193866436708e-06, + "loss": 1.0065, + "step": 5595 + }, + { + "epoch": 0.4768640818065616, + "grad_norm": 16.25, + "learning_rate": 2.174080717110338e-06, + "loss": 1.0059, + "step": 5596 + }, + { + "epoch": 0.47694929697486155, + "grad_norm": 15.5625, + "learning_rate": 2.1739675510849002e-06, + "loss": 0.7129, + "step": 5597 + }, + { + "epoch": 0.4770345121431615, + "grad_norm": 15.9375, + "learning_rate": 2.1738543683624396e-06, + "loss": 0.8302, + "step": 5598 + }, + { + "epoch": 0.47711972731146146, + "grad_norm": 13.75, + "learning_rate": 2.1737411689450015e-06, + "loss": 0.6182, + "step": 5599 + }, + { + "epoch": 0.4772049424797614, + "grad_norm": 19.5, + "learning_rate": 2.1736279528346315e-06, + "loss": 0.9148, + "step": 5600 + }, + { + "epoch": 0.4772901576480614, + "grad_norm": 28.0, + "learning_rate": 2.1735147200333757e-06, + "loss": 1.5502, + "step": 5601 + }, + { + "epoch": 0.47737537281636133, + "grad_norm": 13.8125, + "learning_rate": 2.1734014705432804e-06, + "loss": 0.7342, + "step": 5602 + }, + { + "epoch": 0.4774605879846613, + "grad_norm": 12.0, + "learning_rate": 2.1732882043663917e-06, + "loss": 0.7289, + "step": 5603 + }, + { + "epoch": 0.47754580315296125, + "grad_norm": 15.625, + "learning_rate": 2.173174921504757e-06, + "loss": 0.7083, + "step": 5604 + }, + { + "epoch": 0.4776310183212612, + "grad_norm": 15.5625, + "learning_rate": 2.173061621960423e-06, + "loss": 0.6174, + "step": 5605 + }, + { + "epoch": 0.47771623348956116, + "grad_norm": 14.3125, + "learning_rate": 2.1729483057354374e-06, + "loss": 1.0593, + "step": 5606 + }, + { + "epoch": 0.4778014486578611, + "grad_norm": 14.125, + "learning_rate": 2.1728349728318478e-06, + "loss": 0.6123, + "step": 5607 + }, + { + "epoch": 0.4778866638261611, + "grad_norm": 15.3125, + "learning_rate": 2.172721623251702e-06, + "loss": 1.1667, + "step": 5608 + }, + { + "epoch": 0.47797187899446103, + "grad_norm": 15.9375, + "learning_rate": 2.1726082569970493e-06, + "loss": 0.6684, + "step": 5609 + }, + { + "epoch": 0.478057094162761, + "grad_norm": 10.8125, + "learning_rate": 2.1724948740699367e-06, + "loss": 0.4906, + "step": 5610 + }, + { + "epoch": 0.47814230933106094, + "grad_norm": 20.5, + "learning_rate": 2.172381474472415e-06, + "loss": 0.9473, + "step": 5611 + }, + { + "epoch": 0.4782275244993609, + "grad_norm": 22.875, + "learning_rate": 2.1722680582065315e-06, + "loss": 1.3324, + "step": 5612 + }, + { + "epoch": 0.47831273966766086, + "grad_norm": 16.125, + "learning_rate": 2.1721546252743376e-06, + "loss": 1.0274, + "step": 5613 + }, + { + "epoch": 0.4783979548359608, + "grad_norm": 16.25, + "learning_rate": 2.172041175677882e-06, + "loss": 0.8132, + "step": 5614 + }, + { + "epoch": 0.47848317000426077, + "grad_norm": 20.125, + "learning_rate": 2.171927709419215e-06, + "loss": 1.0301, + "step": 5615 + }, + { + "epoch": 0.47856838517256073, + "grad_norm": 10.25, + "learning_rate": 2.1718142265003875e-06, + "loss": 0.7921, + "step": 5616 + }, + { + "epoch": 0.4786536003408607, + "grad_norm": 20.375, + "learning_rate": 2.17170072692345e-06, + "loss": 0.7704, + "step": 5617 + }, + { + "epoch": 0.47873881550916064, + "grad_norm": 13.1875, + "learning_rate": 2.1715872106904535e-06, + "loss": 0.7501, + "step": 5618 + }, + { + "epoch": 0.4788240306774606, + "grad_norm": 10.0625, + "learning_rate": 2.171473677803449e-06, + "loss": 0.3499, + "step": 5619 + }, + { + "epoch": 0.47890924584576056, + "grad_norm": 15.1875, + "learning_rate": 2.171360128264489e-06, + "loss": 1.0791, + "step": 5620 + }, + { + "epoch": 0.4789944610140605, + "grad_norm": 12.625, + "learning_rate": 2.171246562075625e-06, + "loss": 0.7571, + "step": 5621 + }, + { + "epoch": 0.47907967618236047, + "grad_norm": 15.75, + "learning_rate": 2.1711329792389088e-06, + "loss": 0.7993, + "step": 5622 + }, + { + "epoch": 0.4791648913506604, + "grad_norm": 15.3125, + "learning_rate": 2.171019379756394e-06, + "loss": 0.7689, + "step": 5623 + }, + { + "epoch": 0.4792501065189604, + "grad_norm": 19.25, + "learning_rate": 2.1709057636301325e-06, + "loss": 0.9547, + "step": 5624 + }, + { + "epoch": 0.47933532168726034, + "grad_norm": 15.9375, + "learning_rate": 2.1707921308621775e-06, + "loss": 0.9161, + "step": 5625 + }, + { + "epoch": 0.4794205368555603, + "grad_norm": 17.75, + "learning_rate": 2.170678481454583e-06, + "loss": 1.0343, + "step": 5626 + }, + { + "epoch": 0.47950575202386025, + "grad_norm": 13.625, + "learning_rate": 2.1705648154094028e-06, + "loss": 0.5476, + "step": 5627 + }, + { + "epoch": 0.4795909671921602, + "grad_norm": 11.3125, + "learning_rate": 2.17045113272869e-06, + "loss": 0.3372, + "step": 5628 + }, + { + "epoch": 0.47967618236046017, + "grad_norm": 15.3125, + "learning_rate": 2.1703374334145006e-06, + "loss": 1.1081, + "step": 5629 + }, + { + "epoch": 0.4797613975287601, + "grad_norm": 14.6875, + "learning_rate": 2.1702237174688877e-06, + "loss": 0.5139, + "step": 5630 + }, + { + "epoch": 0.4798466126970601, + "grad_norm": 12.6875, + "learning_rate": 2.1701099848939073e-06, + "loss": 0.4015, + "step": 5631 + }, + { + "epoch": 0.47993182786536004, + "grad_norm": 13.625, + "learning_rate": 2.169996235691614e-06, + "loss": 0.4946, + "step": 5632 + }, + { + "epoch": 0.48001704303366, + "grad_norm": 16.25, + "learning_rate": 2.169882469864063e-06, + "loss": 0.6849, + "step": 5633 + }, + { + "epoch": 0.48010225820195995, + "grad_norm": 19.0, + "learning_rate": 2.1697686874133114e-06, + "loss": 0.9169, + "step": 5634 + }, + { + "epoch": 0.4801874733702599, + "grad_norm": 12.1875, + "learning_rate": 2.169654888341415e-06, + "loss": 0.6028, + "step": 5635 + }, + { + "epoch": 0.48027268853855987, + "grad_norm": 16.125, + "learning_rate": 2.1695410726504295e-06, + "loss": 0.7547, + "step": 5636 + }, + { + "epoch": 0.4803579037068598, + "grad_norm": 17.125, + "learning_rate": 2.1694272403424123e-06, + "loss": 1.2083, + "step": 5637 + }, + { + "epoch": 0.4804431188751598, + "grad_norm": 16.5, + "learning_rate": 2.16931339141942e-06, + "loss": 0.7565, + "step": 5638 + }, + { + "epoch": 0.48052833404345974, + "grad_norm": 19.5, + "learning_rate": 2.1691995258835106e-06, + "loss": 0.8016, + "step": 5639 + }, + { + "epoch": 0.4806135492117597, + "grad_norm": 14.75, + "learning_rate": 2.1690856437367414e-06, + "loss": 0.6593, + "step": 5640 + }, + { + "epoch": 0.48069876438005965, + "grad_norm": 14.9375, + "learning_rate": 2.1689717449811702e-06, + "loss": 0.9464, + "step": 5641 + }, + { + "epoch": 0.4807839795483596, + "grad_norm": 15.5, + "learning_rate": 2.1688578296188557e-06, + "loss": 1.011, + "step": 5642 + }, + { + "epoch": 0.48086919471665956, + "grad_norm": 16.5, + "learning_rate": 2.168743897651856e-06, + "loss": 0.8007, + "step": 5643 + }, + { + "epoch": 0.4809544098849595, + "grad_norm": 11.25, + "learning_rate": 2.1686299490822304e-06, + "loss": 0.5918, + "step": 5644 + }, + { + "epoch": 0.4810396250532595, + "grad_norm": 16.5, + "learning_rate": 2.1685159839120378e-06, + "loss": 0.9377, + "step": 5645 + }, + { + "epoch": 0.48112484022155944, + "grad_norm": 13.0625, + "learning_rate": 2.168402002143338e-06, + "loss": 0.624, + "step": 5646 + }, + { + "epoch": 0.4812100553898594, + "grad_norm": 14.6875, + "learning_rate": 2.16828800377819e-06, + "loss": 0.685, + "step": 5647 + }, + { + "epoch": 0.48129527055815935, + "grad_norm": 12.75, + "learning_rate": 2.1681739888186547e-06, + "loss": 0.4311, + "step": 5648 + }, + { + "epoch": 0.4813804857264593, + "grad_norm": 17.5, + "learning_rate": 2.1680599572667925e-06, + "loss": 0.7615, + "step": 5649 + }, + { + "epoch": 0.48146570089475926, + "grad_norm": 14.625, + "learning_rate": 2.167945909124663e-06, + "loss": 0.9038, + "step": 5650 + }, + { + "epoch": 0.4815509160630592, + "grad_norm": 14.25, + "learning_rate": 2.167831844394328e-06, + "loss": 0.5393, + "step": 5651 + }, + { + "epoch": 0.4816361312313592, + "grad_norm": 12.0625, + "learning_rate": 2.167717763077849e-06, + "loss": 0.3941, + "step": 5652 + }, + { + "epoch": 0.48172134639965913, + "grad_norm": 16.5, + "learning_rate": 2.167603665177287e-06, + "loss": 1.1596, + "step": 5653 + }, + { + "epoch": 0.4818065615679591, + "grad_norm": 14.3125, + "learning_rate": 2.1674895506947034e-06, + "loss": 0.6373, + "step": 5654 + }, + { + "epoch": 0.48189177673625905, + "grad_norm": 13.75, + "learning_rate": 2.167375419632162e-06, + "loss": 0.4344, + "step": 5655 + }, + { + "epoch": 0.481976991904559, + "grad_norm": 20.25, + "learning_rate": 2.1672612719917237e-06, + "loss": 1.0426, + "step": 5656 + }, + { + "epoch": 0.48206220707285896, + "grad_norm": 20.5, + "learning_rate": 2.1671471077754517e-06, + "loss": 0.9335, + "step": 5657 + }, + { + "epoch": 0.4821474222411589, + "grad_norm": 14.0, + "learning_rate": 2.1670329269854097e-06, + "loss": 0.8758, + "step": 5658 + }, + { + "epoch": 0.4822326374094589, + "grad_norm": 15.6875, + "learning_rate": 2.1669187296236606e-06, + "loss": 0.8131, + "step": 5659 + }, + { + "epoch": 0.48231785257775883, + "grad_norm": 11.625, + "learning_rate": 2.166804515692268e-06, + "loss": 0.5703, + "step": 5660 + }, + { + "epoch": 0.4824030677460588, + "grad_norm": 17.5, + "learning_rate": 2.166690285193295e-06, + "loss": 0.6691, + "step": 5661 + }, + { + "epoch": 0.48248828291435875, + "grad_norm": 13.75, + "learning_rate": 2.1665760381288077e-06, + "loss": 0.785, + "step": 5662 + }, + { + "epoch": 0.4825734980826587, + "grad_norm": 17.875, + "learning_rate": 2.1664617745008696e-06, + "loss": 0.3756, + "step": 5663 + }, + { + "epoch": 0.48265871325095866, + "grad_norm": 37.75, + "learning_rate": 2.1663474943115455e-06, + "loss": 1.4977, + "step": 5664 + }, + { + "epoch": 0.4827439284192586, + "grad_norm": 13.0, + "learning_rate": 2.1662331975629006e-06, + "loss": 0.8138, + "step": 5665 + }, + { + "epoch": 0.4828291435875586, + "grad_norm": 13.6875, + "learning_rate": 2.1661188842570006e-06, + "loss": 0.8917, + "step": 5666 + }, + { + "epoch": 0.48291435875585853, + "grad_norm": 13.0625, + "learning_rate": 2.166004554395911e-06, + "loss": 0.6281, + "step": 5667 + }, + { + "epoch": 0.4829995739241585, + "grad_norm": 12.9375, + "learning_rate": 2.165890207981698e-06, + "loss": 0.4999, + "step": 5668 + }, + { + "epoch": 0.48308478909245844, + "grad_norm": 15.875, + "learning_rate": 2.165775845016428e-06, + "loss": 0.7781, + "step": 5669 + }, + { + "epoch": 0.4831700042607584, + "grad_norm": 16.0, + "learning_rate": 2.1656614655021675e-06, + "loss": 0.9867, + "step": 5670 + }, + { + "epoch": 0.48325521942905836, + "grad_norm": 17.0, + "learning_rate": 2.165547069440984e-06, + "loss": 0.6159, + "step": 5671 + }, + { + "epoch": 0.4833404345973583, + "grad_norm": 11.4375, + "learning_rate": 2.1654326568349436e-06, + "loss": 0.5578, + "step": 5672 + }, + { + "epoch": 0.48342564976565827, + "grad_norm": 11.5, + "learning_rate": 2.1653182276861153e-06, + "loss": 0.555, + "step": 5673 + }, + { + "epoch": 0.48351086493395823, + "grad_norm": 23.5, + "learning_rate": 2.1652037819965657e-06, + "loss": 0.9754, + "step": 5674 + }, + { + "epoch": 0.4835960801022582, + "grad_norm": 32.5, + "learning_rate": 2.1650893197683636e-06, + "loss": 1.3365, + "step": 5675 + }, + { + "epoch": 0.48368129527055814, + "grad_norm": 15.8125, + "learning_rate": 2.164974841003577e-06, + "loss": 0.6482, + "step": 5676 + }, + { + "epoch": 0.4837665104388581, + "grad_norm": 19.125, + "learning_rate": 2.1648603457042756e-06, + "loss": 1.0686, + "step": 5677 + }, + { + "epoch": 0.48385172560715806, + "grad_norm": 11.4375, + "learning_rate": 2.1647458338725273e-06, + "loss": 0.3889, + "step": 5678 + }, + { + "epoch": 0.483936940775458, + "grad_norm": 17.125, + "learning_rate": 2.164631305510402e-06, + "loss": 0.5939, + "step": 5679 + }, + { + "epoch": 0.48402215594375797, + "grad_norm": 21.375, + "learning_rate": 2.1645167606199693e-06, + "loss": 0.9593, + "step": 5680 + }, + { + "epoch": 0.4841073711120579, + "grad_norm": 36.0, + "learning_rate": 2.1644021992032994e-06, + "loss": 1.0025, + "step": 5681 + }, + { + "epoch": 0.4841925862803579, + "grad_norm": 20.75, + "learning_rate": 2.164287621262462e-06, + "loss": 0.9502, + "step": 5682 + }, + { + "epoch": 0.48427780144865784, + "grad_norm": 13.1875, + "learning_rate": 2.1641730267995277e-06, + "loss": 0.6198, + "step": 5683 + }, + { + "epoch": 0.4843630166169578, + "grad_norm": 10.625, + "learning_rate": 2.164058415816568e-06, + "loss": 0.4153, + "step": 5684 + }, + { + "epoch": 0.48444823178525775, + "grad_norm": 14.0, + "learning_rate": 2.163943788315653e-06, + "loss": 0.3983, + "step": 5685 + }, + { + "epoch": 0.4845334469535577, + "grad_norm": 21.375, + "learning_rate": 2.163829144298855e-06, + "loss": 1.0344, + "step": 5686 + }, + { + "epoch": 0.48461866212185767, + "grad_norm": 11.5, + "learning_rate": 2.1637144837682456e-06, + "loss": 0.5683, + "step": 5687 + }, + { + "epoch": 0.4847038772901576, + "grad_norm": 17.125, + "learning_rate": 2.163599806725896e-06, + "loss": 0.879, + "step": 5688 + }, + { + "epoch": 0.4847890924584576, + "grad_norm": 13.4375, + "learning_rate": 2.16348511317388e-06, + "loss": 0.3956, + "step": 5689 + }, + { + "epoch": 0.48487430762675754, + "grad_norm": 17.375, + "learning_rate": 2.163370403114269e-06, + "loss": 0.7775, + "step": 5690 + }, + { + "epoch": 0.4849595227950575, + "grad_norm": 15.5, + "learning_rate": 2.163255676549137e-06, + "loss": 0.6783, + "step": 5691 + }, + { + "epoch": 0.48504473796335745, + "grad_norm": 15.75, + "learning_rate": 2.163140933480556e-06, + "loss": 0.6896, + "step": 5692 + }, + { + "epoch": 0.4851299531316574, + "grad_norm": 25.125, + "learning_rate": 2.1630261739105996e-06, + "loss": 1.416, + "step": 5693 + }, + { + "epoch": 0.48521516829995737, + "grad_norm": 14.0625, + "learning_rate": 2.162911397841343e-06, + "loss": 0.4953, + "step": 5694 + }, + { + "epoch": 0.4853003834682573, + "grad_norm": 19.375, + "learning_rate": 2.162796605274859e-06, + "loss": 1.0947, + "step": 5695 + }, + { + "epoch": 0.4853855986365573, + "grad_norm": 15.8125, + "learning_rate": 2.162681796213223e-06, + "loss": 0.8249, + "step": 5696 + }, + { + "epoch": 0.48547081380485724, + "grad_norm": 15.125, + "learning_rate": 2.1625669706585085e-06, + "loss": 0.8246, + "step": 5697 + }, + { + "epoch": 0.4855560289731572, + "grad_norm": 21.375, + "learning_rate": 2.1624521286127915e-06, + "loss": 1.2245, + "step": 5698 + }, + { + "epoch": 0.4856412441414572, + "grad_norm": 12.5625, + "learning_rate": 2.1623372700781463e-06, + "loss": 0.5612, + "step": 5699 + }, + { + "epoch": 0.48572645930975716, + "grad_norm": 20.5, + "learning_rate": 2.1622223950566503e-06, + "loss": 1.0498, + "step": 5700 + }, + { + "epoch": 0.4858116744780571, + "grad_norm": 13.5, + "learning_rate": 2.1621075035503777e-06, + "loss": 0.6437, + "step": 5701 + }, + { + "epoch": 0.4858968896463571, + "grad_norm": 18.0, + "learning_rate": 2.161992595561405e-06, + "loss": 0.6918, + "step": 5702 + }, + { + "epoch": 0.48598210481465703, + "grad_norm": 16.375, + "learning_rate": 2.16187767109181e-06, + "loss": 0.8423, + "step": 5703 + }, + { + "epoch": 0.486067319982957, + "grad_norm": 16.75, + "learning_rate": 2.1617627301436677e-06, + "loss": 0.8333, + "step": 5704 + }, + { + "epoch": 0.48615253515125695, + "grad_norm": 11.8125, + "learning_rate": 2.1616477727190563e-06, + "loss": 0.69, + "step": 5705 + }, + { + "epoch": 0.4862377503195569, + "grad_norm": 12.1875, + "learning_rate": 2.161532798820053e-06, + "loss": 0.5914, + "step": 5706 + }, + { + "epoch": 0.48632296548785686, + "grad_norm": 13.5625, + "learning_rate": 2.161417808448735e-06, + "loss": 0.7777, + "step": 5707 + }, + { + "epoch": 0.4864081806561568, + "grad_norm": 15.125, + "learning_rate": 2.1613028016071814e-06, + "loss": 0.7, + "step": 5708 + }, + { + "epoch": 0.4864933958244568, + "grad_norm": 13.8125, + "learning_rate": 2.1611877782974696e-06, + "loss": 0.7747, + "step": 5709 + }, + { + "epoch": 0.48657861099275673, + "grad_norm": 14.6875, + "learning_rate": 2.161072738521678e-06, + "loss": 0.9512, + "step": 5710 + }, + { + "epoch": 0.4866638261610567, + "grad_norm": 13.6875, + "learning_rate": 2.160957682281886e-06, + "loss": 0.7135, + "step": 5711 + }, + { + "epoch": 0.48674904132935665, + "grad_norm": 9.4375, + "learning_rate": 2.1608426095801727e-06, + "loss": 0.376, + "step": 5712 + }, + { + "epoch": 0.4868342564976566, + "grad_norm": 14.1875, + "learning_rate": 2.1607275204186174e-06, + "loss": 0.7216, + "step": 5713 + }, + { + "epoch": 0.48691947166595656, + "grad_norm": 12.75, + "learning_rate": 2.1606124147993e-06, + "loss": 0.4498, + "step": 5714 + }, + { + "epoch": 0.4870046868342565, + "grad_norm": 11.875, + "learning_rate": 2.160497292724301e-06, + "loss": 0.609, + "step": 5715 + }, + { + "epoch": 0.4870899020025565, + "grad_norm": 15.6875, + "learning_rate": 2.1603821541956997e-06, + "loss": 0.5074, + "step": 5716 + }, + { + "epoch": 0.48717511717085643, + "grad_norm": 14.5625, + "learning_rate": 2.1602669992155776e-06, + "loss": 0.6741, + "step": 5717 + }, + { + "epoch": 0.4872603323391564, + "grad_norm": 16.0, + "learning_rate": 2.160151827786016e-06, + "loss": 0.9134, + "step": 5718 + }, + { + "epoch": 0.48734554750745634, + "grad_norm": 17.75, + "learning_rate": 2.160036639909095e-06, + "loss": 1.1331, + "step": 5719 + }, + { + "epoch": 0.4874307626757563, + "grad_norm": 21.25, + "learning_rate": 2.159921435586897e-06, + "loss": 0.7145, + "step": 5720 + }, + { + "epoch": 0.48751597784405626, + "grad_norm": 14.875, + "learning_rate": 2.1598062148215037e-06, + "loss": 0.9386, + "step": 5721 + }, + { + "epoch": 0.4876011930123562, + "grad_norm": 25.375, + "learning_rate": 2.159690977614997e-06, + "loss": 1.2735, + "step": 5722 + }, + { + "epoch": 0.4876864081806562, + "grad_norm": 12.125, + "learning_rate": 2.1595757239694603e-06, + "loss": 0.659, + "step": 5723 + }, + { + "epoch": 0.48777162334895613, + "grad_norm": 15.75, + "learning_rate": 2.159460453886975e-06, + "loss": 0.748, + "step": 5724 + }, + { + "epoch": 0.4878568385172561, + "grad_norm": 16.625, + "learning_rate": 2.159345167369625e-06, + "loss": 0.9094, + "step": 5725 + }, + { + "epoch": 0.48794205368555604, + "grad_norm": 14.75, + "learning_rate": 2.159229864419493e-06, + "loss": 0.8414, + "step": 5726 + }, + { + "epoch": 0.488027268853856, + "grad_norm": 14.375, + "learning_rate": 2.1591145450386636e-06, + "loss": 0.2189, + "step": 5727 + }, + { + "epoch": 0.48811248402215596, + "grad_norm": 10.6875, + "learning_rate": 2.15899920922922e-06, + "loss": 0.4003, + "step": 5728 + }, + { + "epoch": 0.4881976991904559, + "grad_norm": 12.3125, + "learning_rate": 2.1588838569932466e-06, + "loss": 0.5217, + "step": 5729 + }, + { + "epoch": 0.48828291435875587, + "grad_norm": 16.5, + "learning_rate": 2.158768488332828e-06, + "loss": 0.9054, + "step": 5730 + }, + { + "epoch": 0.4883681295270558, + "grad_norm": 12.375, + "learning_rate": 2.1586531032500488e-06, + "loss": 0.536, + "step": 5731 + }, + { + "epoch": 0.4884533446953558, + "grad_norm": 13.6875, + "learning_rate": 2.1585377017469946e-06, + "loss": 0.7898, + "step": 5732 + }, + { + "epoch": 0.48853855986365574, + "grad_norm": 13.625, + "learning_rate": 2.15842228382575e-06, + "loss": 0.4113, + "step": 5733 + }, + { + "epoch": 0.4886237750319557, + "grad_norm": 19.625, + "learning_rate": 2.1583068494884016e-06, + "loss": 0.8966, + "step": 5734 + }, + { + "epoch": 0.48870899020025566, + "grad_norm": 17.5, + "learning_rate": 2.158191398737035e-06, + "loss": 1.003, + "step": 5735 + }, + { + "epoch": 0.4887942053685556, + "grad_norm": 14.3125, + "learning_rate": 2.1580759315737367e-06, + "loss": 0.9048, + "step": 5736 + }, + { + "epoch": 0.48887942053685557, + "grad_norm": 12.875, + "learning_rate": 2.157960448000593e-06, + "loss": 0.5135, + "step": 5737 + }, + { + "epoch": 0.4889646357051555, + "grad_norm": 14.9375, + "learning_rate": 2.1578449480196913e-06, + "loss": 0.7703, + "step": 5738 + }, + { + "epoch": 0.4890498508734555, + "grad_norm": 21.375, + "learning_rate": 2.157729431633118e-06, + "loss": 0.7969, + "step": 5739 + }, + { + "epoch": 0.48913506604175544, + "grad_norm": 17.875, + "learning_rate": 2.1576138988429614e-06, + "loss": 1.0738, + "step": 5740 + }, + { + "epoch": 0.4892202812100554, + "grad_norm": 16.0, + "learning_rate": 2.1574983496513086e-06, + "loss": 0.8701, + "step": 5741 + }, + { + "epoch": 0.48930549637835535, + "grad_norm": 13.6875, + "learning_rate": 2.1573827840602483e-06, + "loss": 0.8098, + "step": 5742 + }, + { + "epoch": 0.4893907115466553, + "grad_norm": 17.625, + "learning_rate": 2.157267202071869e-06, + "loss": 1.2905, + "step": 5743 + }, + { + "epoch": 0.48947592671495527, + "grad_norm": 15.875, + "learning_rate": 2.1571516036882585e-06, + "loss": 0.9116, + "step": 5744 + }, + { + "epoch": 0.4895611418832552, + "grad_norm": 13.1875, + "learning_rate": 2.1570359889115065e-06, + "loss": 0.7612, + "step": 5745 + }, + { + "epoch": 0.4896463570515552, + "grad_norm": 14.375, + "learning_rate": 2.156920357743702e-06, + "loss": 0.5782, + "step": 5746 + }, + { + "epoch": 0.48973157221985514, + "grad_norm": 14.0, + "learning_rate": 2.156804710186935e-06, + "loss": 0.5309, + "step": 5747 + }, + { + "epoch": 0.4898167873881551, + "grad_norm": 15.5625, + "learning_rate": 2.1566890462432944e-06, + "loss": 0.6784, + "step": 5748 + }, + { + "epoch": 0.48990200255645505, + "grad_norm": 13.0, + "learning_rate": 2.156573365914872e-06, + "loss": 0.6522, + "step": 5749 + }, + { + "epoch": 0.489987217724755, + "grad_norm": 14.625, + "learning_rate": 2.1564576692037564e-06, + "loss": 0.9226, + "step": 5750 + }, + { + "epoch": 0.49007243289305497, + "grad_norm": 16.25, + "learning_rate": 2.1563419561120393e-06, + "loss": 0.6082, + "step": 5751 + }, + { + "epoch": 0.4901576480613549, + "grad_norm": 11.8125, + "learning_rate": 2.1562262266418115e-06, + "loss": 0.492, + "step": 5752 + }, + { + "epoch": 0.4902428632296549, + "grad_norm": 12.875, + "learning_rate": 2.1561104807951648e-06, + "loss": 0.6654, + "step": 5753 + }, + { + "epoch": 0.49032807839795484, + "grad_norm": 13.875, + "learning_rate": 2.155994718574191e-06, + "loss": 0.6855, + "step": 5754 + }, + { + "epoch": 0.4904132935662548, + "grad_norm": 24.125, + "learning_rate": 2.155878939980981e-06, + "loss": 0.7849, + "step": 5755 + }, + { + "epoch": 0.49049850873455475, + "grad_norm": 15.0, + "learning_rate": 2.1557631450176274e-06, + "loss": 0.831, + "step": 5756 + }, + { + "epoch": 0.4905837239028547, + "grad_norm": 13.75, + "learning_rate": 2.1556473336862236e-06, + "loss": 0.4793, + "step": 5757 + }, + { + "epoch": 0.49066893907115466, + "grad_norm": 17.25, + "learning_rate": 2.1555315059888616e-06, + "loss": 0.9881, + "step": 5758 + }, + { + "epoch": 0.4907541542394546, + "grad_norm": 14.125, + "learning_rate": 2.1554156619276346e-06, + "loss": 0.8504, + "step": 5759 + }, + { + "epoch": 0.4908393694077546, + "grad_norm": 15.4375, + "learning_rate": 2.155299801504636e-06, + "loss": 0.8327, + "step": 5760 + }, + { + "epoch": 0.49092458457605453, + "grad_norm": 14.0625, + "learning_rate": 2.15518392472196e-06, + "loss": 0.8131, + "step": 5761 + }, + { + "epoch": 0.4910097997443545, + "grad_norm": 11.25, + "learning_rate": 2.1550680315816996e-06, + "loss": 0.6647, + "step": 5762 + }, + { + "epoch": 0.49109501491265445, + "grad_norm": 26.125, + "learning_rate": 2.15495212208595e-06, + "loss": 0.7175, + "step": 5763 + }, + { + "epoch": 0.4911802300809544, + "grad_norm": 15.9375, + "learning_rate": 2.154836196236806e-06, + "loss": 0.6945, + "step": 5764 + }, + { + "epoch": 0.49126544524925436, + "grad_norm": 15.375, + "learning_rate": 2.1547202540363617e-06, + "loss": 0.7726, + "step": 5765 + }, + { + "epoch": 0.4913506604175543, + "grad_norm": 16.875, + "learning_rate": 2.1546042954867124e-06, + "loss": 0.9963, + "step": 5766 + }, + { + "epoch": 0.4914358755858543, + "grad_norm": 15.5625, + "learning_rate": 2.154488320589954e-06, + "loss": 0.9346, + "step": 5767 + }, + { + "epoch": 0.49152109075415423, + "grad_norm": 19.125, + "learning_rate": 2.1543723293481826e-06, + "loss": 1.0091, + "step": 5768 + }, + { + "epoch": 0.4916063059224542, + "grad_norm": 17.875, + "learning_rate": 2.1542563217634936e-06, + "loss": 0.541, + "step": 5769 + }, + { + "epoch": 0.49169152109075415, + "grad_norm": 14.5625, + "learning_rate": 2.1541402978379835e-06, + "loss": 0.8654, + "step": 5770 + }, + { + "epoch": 0.4917767362590541, + "grad_norm": 24.625, + "learning_rate": 2.1540242575737486e-06, + "loss": 1.3225, + "step": 5771 + }, + { + "epoch": 0.49186195142735406, + "grad_norm": 10.4375, + "learning_rate": 2.1539082009728863e-06, + "loss": 0.3247, + "step": 5772 + }, + { + "epoch": 0.491947166595654, + "grad_norm": 18.75, + "learning_rate": 2.1537921280374946e-06, + "loss": 0.6543, + "step": 5773 + }, + { + "epoch": 0.492032381763954, + "grad_norm": 20.875, + "learning_rate": 2.15367603876967e-06, + "loss": 1.0828, + "step": 5774 + }, + { + "epoch": 0.49211759693225393, + "grad_norm": 14.5, + "learning_rate": 2.15355993317151e-06, + "loss": 0.9117, + "step": 5775 + }, + { + "epoch": 0.4922028121005539, + "grad_norm": 13.3125, + "learning_rate": 2.153443811245114e-06, + "loss": 0.7582, + "step": 5776 + }, + { + "epoch": 0.49228802726885384, + "grad_norm": 14.875, + "learning_rate": 2.15332767299258e-06, + "loss": 0.9515, + "step": 5777 + }, + { + "epoch": 0.4923732424371538, + "grad_norm": 18.25, + "learning_rate": 2.1532115184160063e-06, + "loss": 0.9592, + "step": 5778 + }, + { + "epoch": 0.49245845760545376, + "grad_norm": 24.25, + "learning_rate": 2.1530953475174926e-06, + "loss": 0.9465, + "step": 5779 + }, + { + "epoch": 0.4925436727737537, + "grad_norm": 12.5, + "learning_rate": 2.1529791602991375e-06, + "loss": 0.5654, + "step": 5780 + }, + { + "epoch": 0.4926288879420537, + "grad_norm": 12.0625, + "learning_rate": 2.152862956763041e-06, + "loss": 0.5259, + "step": 5781 + }, + { + "epoch": 0.49271410311035363, + "grad_norm": 10.0625, + "learning_rate": 2.152746736911303e-06, + "loss": 0.5468, + "step": 5782 + }, + { + "epoch": 0.4927993182786536, + "grad_norm": 21.375, + "learning_rate": 2.1526305007460235e-06, + "loss": 0.8994, + "step": 5783 + }, + { + "epoch": 0.49288453344695354, + "grad_norm": 16.875, + "learning_rate": 2.1525142482693033e-06, + "loss": 0.8481, + "step": 5784 + }, + { + "epoch": 0.4929697486152535, + "grad_norm": 14.3125, + "learning_rate": 2.152397979483243e-06, + "loss": 0.839, + "step": 5785 + }, + { + "epoch": 0.49305496378355346, + "grad_norm": 14.8125, + "learning_rate": 2.1522816943899443e-06, + "loss": 0.7765, + "step": 5786 + }, + { + "epoch": 0.4931401789518534, + "grad_norm": 14.125, + "learning_rate": 2.1521653929915077e-06, + "loss": 0.757, + "step": 5787 + }, + { + "epoch": 0.49322539412015337, + "grad_norm": 11.6875, + "learning_rate": 2.152049075290035e-06, + "loss": 0.4022, + "step": 5788 + }, + { + "epoch": 0.49331060928845333, + "grad_norm": 11.75, + "learning_rate": 2.151932741287629e-06, + "loss": 0.473, + "step": 5789 + }, + { + "epoch": 0.4933958244567533, + "grad_norm": 25.0, + "learning_rate": 2.151816390986391e-06, + "loss": 0.8495, + "step": 5790 + }, + { + "epoch": 0.49348103962505324, + "grad_norm": 10.125, + "learning_rate": 2.151700024388424e-06, + "loss": 0.4126, + "step": 5791 + }, + { + "epoch": 0.4935662547933532, + "grad_norm": 17.75, + "learning_rate": 2.1515836414958308e-06, + "loss": 0.6038, + "step": 5792 + }, + { + "epoch": 0.49365146996165316, + "grad_norm": 25.125, + "learning_rate": 2.1514672423107146e-06, + "loss": 1.2276, + "step": 5793 + }, + { + "epoch": 0.4937366851299531, + "grad_norm": 13.1875, + "learning_rate": 2.1513508268351787e-06, + "loss": 0.656, + "step": 5794 + }, + { + "epoch": 0.49382190029825307, + "grad_norm": 13.25, + "learning_rate": 2.151234395071327e-06, + "loss": 0.426, + "step": 5795 + }, + { + "epoch": 0.493907115466553, + "grad_norm": 10.5625, + "learning_rate": 2.151117947021264e-06, + "loss": 0.3618, + "step": 5796 + }, + { + "epoch": 0.493992330634853, + "grad_norm": 15.25, + "learning_rate": 2.151001482687093e-06, + "loss": 0.7949, + "step": 5797 + }, + { + "epoch": 0.49407754580315294, + "grad_norm": 33.5, + "learning_rate": 2.1508850020709196e-06, + "loss": 0.962, + "step": 5798 + }, + { + "epoch": 0.4941627609714529, + "grad_norm": 14.125, + "learning_rate": 2.150768505174848e-06, + "loss": 1.0273, + "step": 5799 + }, + { + "epoch": 0.49424797613975285, + "grad_norm": 16.0, + "learning_rate": 2.1506519920009838e-06, + "loss": 0.6742, + "step": 5800 + }, + { + "epoch": 0.4943331913080528, + "grad_norm": 17.75, + "learning_rate": 2.1505354625514325e-06, + "loss": 0.7906, + "step": 5801 + }, + { + "epoch": 0.49441840647635277, + "grad_norm": 13.125, + "learning_rate": 2.1504189168283e-06, + "loss": 0.749, + "step": 5802 + }, + { + "epoch": 0.4945036216446527, + "grad_norm": 12.4375, + "learning_rate": 2.150302354833692e-06, + "loss": 0.6195, + "step": 5803 + }, + { + "epoch": 0.4945888368129527, + "grad_norm": 17.0, + "learning_rate": 2.150185776569715e-06, + "loss": 0.7822, + "step": 5804 + }, + { + "epoch": 0.49467405198125264, + "grad_norm": 14.125, + "learning_rate": 2.1500691820384762e-06, + "loss": 0.7311, + "step": 5805 + }, + { + "epoch": 0.4947592671495526, + "grad_norm": 14.0, + "learning_rate": 2.1499525712420818e-06, + "loss": 0.6468, + "step": 5806 + }, + { + "epoch": 0.49484448231785255, + "grad_norm": 11.625, + "learning_rate": 2.1498359441826395e-06, + "loss": 0.473, + "step": 5807 + }, + { + "epoch": 0.4949296974861525, + "grad_norm": 19.625, + "learning_rate": 2.149719300862257e-06, + "loss": 0.763, + "step": 5808 + }, + { + "epoch": 0.49501491265445247, + "grad_norm": 20.25, + "learning_rate": 2.149602641283042e-06, + "loss": 0.6283, + "step": 5809 + }, + { + "epoch": 0.4951001278227524, + "grad_norm": 15.6875, + "learning_rate": 2.1494859654471027e-06, + "loss": 0.8069, + "step": 5810 + }, + { + "epoch": 0.4951853429910524, + "grad_norm": 13.625, + "learning_rate": 2.1493692733565474e-06, + "loss": 0.7092, + "step": 5811 + }, + { + "epoch": 0.4952705581593524, + "grad_norm": 13.25, + "learning_rate": 2.149252565013485e-06, + "loss": 0.8032, + "step": 5812 + }, + { + "epoch": 0.49535577332765235, + "grad_norm": 23.25, + "learning_rate": 2.1491358404200245e-06, + "loss": 1.2107, + "step": 5813 + }, + { + "epoch": 0.4954409884959523, + "grad_norm": 15.875, + "learning_rate": 2.149019099578275e-06, + "loss": 0.7964, + "step": 5814 + }, + { + "epoch": 0.49552620366425226, + "grad_norm": 13.25, + "learning_rate": 2.1489023424903466e-06, + "loss": 0.7263, + "step": 5815 + }, + { + "epoch": 0.4956114188325522, + "grad_norm": 17.375, + "learning_rate": 2.1487855691583487e-06, + "loss": 0.859, + "step": 5816 + }, + { + "epoch": 0.4956966340008522, + "grad_norm": 13.8125, + "learning_rate": 2.148668779584392e-06, + "loss": 0.871, + "step": 5817 + }, + { + "epoch": 0.49578184916915213, + "grad_norm": 25.0, + "learning_rate": 2.1485519737705868e-06, + "loss": 0.8851, + "step": 5818 + }, + { + "epoch": 0.4958670643374521, + "grad_norm": 15.6875, + "learning_rate": 2.1484351517190435e-06, + "loss": 0.7336, + "step": 5819 + }, + { + "epoch": 0.49595227950575205, + "grad_norm": 14.9375, + "learning_rate": 2.148318313431874e-06, + "loss": 0.6069, + "step": 5820 + }, + { + "epoch": 0.496037494674052, + "grad_norm": 21.25, + "learning_rate": 2.148201458911189e-06, + "loss": 0.9736, + "step": 5821 + }, + { + "epoch": 0.49612270984235196, + "grad_norm": 20.125, + "learning_rate": 2.1480845881591e-06, + "loss": 1.0662, + "step": 5822 + }, + { + "epoch": 0.4962079250106519, + "grad_norm": 13.875, + "learning_rate": 2.14796770117772e-06, + "loss": 0.8106, + "step": 5823 + }, + { + "epoch": 0.4962931401789519, + "grad_norm": 14.3125, + "learning_rate": 2.147850797969161e-06, + "loss": 0.4575, + "step": 5824 + }, + { + "epoch": 0.49637835534725183, + "grad_norm": 15.1875, + "learning_rate": 2.1477338785355345e-06, + "loss": 0.6992, + "step": 5825 + }, + { + "epoch": 0.4964635705155518, + "grad_norm": 12.8125, + "learning_rate": 2.1476169428789545e-06, + "loss": 0.645, + "step": 5826 + }, + { + "epoch": 0.49654878568385175, + "grad_norm": 12.6875, + "learning_rate": 2.1474999910015334e-06, + "loss": 0.823, + "step": 5827 + }, + { + "epoch": 0.4966340008521517, + "grad_norm": 12.5, + "learning_rate": 2.147383022905385e-06, + "loss": 0.5045, + "step": 5828 + }, + { + "epoch": 0.49671921602045166, + "grad_norm": 21.875, + "learning_rate": 2.1472660385926232e-06, + "loss": 1.1154, + "step": 5829 + }, + { + "epoch": 0.4968044311887516, + "grad_norm": 17.25, + "learning_rate": 2.1471490380653617e-06, + "loss": 0.6726, + "step": 5830 + }, + { + "epoch": 0.4968896463570516, + "grad_norm": 13.875, + "learning_rate": 2.147032021325715e-06, + "loss": 0.6526, + "step": 5831 + }, + { + "epoch": 0.49697486152535153, + "grad_norm": 14.625, + "learning_rate": 2.146914988375798e-06, + "loss": 0.7688, + "step": 5832 + }, + { + "epoch": 0.4970600766936515, + "grad_norm": 15.9375, + "learning_rate": 2.1467979392177245e-06, + "loss": 0.9518, + "step": 5833 + }, + { + "epoch": 0.49714529186195144, + "grad_norm": 12.1875, + "learning_rate": 2.1466808738536107e-06, + "loss": 0.9798, + "step": 5834 + }, + { + "epoch": 0.4972305070302514, + "grad_norm": 11.6875, + "learning_rate": 2.146563792285572e-06, + "loss": 0.4721, + "step": 5835 + }, + { + "epoch": 0.49731572219855136, + "grad_norm": 12.25, + "learning_rate": 2.1464466945157237e-06, + "loss": 0.6144, + "step": 5836 + }, + { + "epoch": 0.4974009373668513, + "grad_norm": 12.75, + "learning_rate": 2.1463295805461828e-06, + "loss": 0.7683, + "step": 5837 + }, + { + "epoch": 0.49748615253515127, + "grad_norm": 14.75, + "learning_rate": 2.1462124503790645e-06, + "loss": 0.8325, + "step": 5838 + }, + { + "epoch": 0.49757136770345123, + "grad_norm": 15.4375, + "learning_rate": 2.146095304016486e-06, + "loss": 0.8617, + "step": 5839 + }, + { + "epoch": 0.4976565828717512, + "grad_norm": 14.8125, + "learning_rate": 2.1459781414605645e-06, + "loss": 0.9438, + "step": 5840 + }, + { + "epoch": 0.49774179804005114, + "grad_norm": 13.8125, + "learning_rate": 2.1458609627134166e-06, + "loss": 0.8293, + "step": 5841 + }, + { + "epoch": 0.4978270132083511, + "grad_norm": 13.5, + "learning_rate": 2.1457437677771605e-06, + "loss": 0.6026, + "step": 5842 + }, + { + "epoch": 0.49791222837665106, + "grad_norm": 9.875, + "learning_rate": 2.1456265566539137e-06, + "loss": 0.3548, + "step": 5843 + }, + { + "epoch": 0.497997443544951, + "grad_norm": 13.3125, + "learning_rate": 2.1455093293457947e-06, + "loss": 0.707, + "step": 5844 + }, + { + "epoch": 0.49808265871325097, + "grad_norm": 15.5625, + "learning_rate": 2.1453920858549216e-06, + "loss": 0.953, + "step": 5845 + }, + { + "epoch": 0.4981678738815509, + "grad_norm": 16.125, + "learning_rate": 2.1452748261834127e-06, + "loss": 0.8664, + "step": 5846 + }, + { + "epoch": 0.4982530890498509, + "grad_norm": 14.3125, + "learning_rate": 2.145157550333388e-06, + "loss": 0.6303, + "step": 5847 + }, + { + "epoch": 0.49833830421815084, + "grad_norm": 23.75, + "learning_rate": 2.1450402583069655e-06, + "loss": 0.5487, + "step": 5848 + }, + { + "epoch": 0.4984235193864508, + "grad_norm": 17.5, + "learning_rate": 2.1449229501062662e-06, + "loss": 0.6514, + "step": 5849 + }, + { + "epoch": 0.49850873455475075, + "grad_norm": 10.4375, + "learning_rate": 2.1448056257334087e-06, + "loss": 0.3962, + "step": 5850 + }, + { + "epoch": 0.4985939497230507, + "grad_norm": 14.9375, + "learning_rate": 2.1446882851905144e-06, + "loss": 0.5081, + "step": 5851 + }, + { + "epoch": 0.49867916489135067, + "grad_norm": 14.5, + "learning_rate": 2.144570928479703e-06, + "loss": 0.696, + "step": 5852 + }, + { + "epoch": 0.4987643800596506, + "grad_norm": 15.6875, + "learning_rate": 2.144453555603095e-06, + "loss": 0.6514, + "step": 5853 + }, + { + "epoch": 0.4988495952279506, + "grad_norm": 15.3125, + "learning_rate": 2.1443361665628115e-06, + "loss": 0.7258, + "step": 5854 + }, + { + "epoch": 0.49893481039625054, + "grad_norm": 14.5625, + "learning_rate": 2.144218761360975e-06, + "loss": 0.9006, + "step": 5855 + }, + { + "epoch": 0.4990200255645505, + "grad_norm": 12.9375, + "learning_rate": 2.144101339999706e-06, + "loss": 0.669, + "step": 5856 + }, + { + "epoch": 0.49910524073285045, + "grad_norm": 15.25, + "learning_rate": 2.143983902481126e-06, + "loss": 0.7959, + "step": 5857 + }, + { + "epoch": 0.4991904559011504, + "grad_norm": 12.625, + "learning_rate": 2.1438664488073585e-06, + "loss": 0.8162, + "step": 5858 + }, + { + "epoch": 0.49927567106945037, + "grad_norm": 14.75, + "learning_rate": 2.1437489789805257e-06, + "loss": 0.6857, + "step": 5859 + }, + { + "epoch": 0.4993608862377503, + "grad_norm": 11.0, + "learning_rate": 2.1436314930027494e-06, + "loss": 0.4244, + "step": 5860 + }, + { + "epoch": 0.4994461014060503, + "grad_norm": 14.0, + "learning_rate": 2.1435139908761535e-06, + "loss": 0.5832, + "step": 5861 + }, + { + "epoch": 0.49953131657435024, + "grad_norm": 13.125, + "learning_rate": 2.1433964726028617e-06, + "loss": 0.5866, + "step": 5862 + }, + { + "epoch": 0.4996165317426502, + "grad_norm": 18.625, + "learning_rate": 2.1432789381849974e-06, + "loss": 0.9645, + "step": 5863 + }, + { + "epoch": 0.49970174691095015, + "grad_norm": 25.875, + "learning_rate": 2.143161387624684e-06, + "loss": 0.6659, + "step": 5864 + }, + { + "epoch": 0.4997869620792501, + "grad_norm": 13.125, + "learning_rate": 2.1430438209240467e-06, + "loss": 0.7143, + "step": 5865 + }, + { + "epoch": 0.49987217724755006, + "grad_norm": 14.3125, + "learning_rate": 2.142926238085209e-06, + "loss": 0.5149, + "step": 5866 + }, + { + "epoch": 0.49995739241585, + "grad_norm": 15.1875, + "learning_rate": 2.1428086391102965e-06, + "loss": 0.8181, + "step": 5867 + }, + { + "epoch": 0.50004260758415, + "grad_norm": 11.25, + "learning_rate": 2.1426910240014335e-06, + "loss": 0.2639, + "step": 5868 + }, + { + "epoch": 0.5001278227524499, + "grad_norm": 10.625, + "learning_rate": 2.1425733927607466e-06, + "loss": 0.3995, + "step": 5869 + }, + { + "epoch": 0.5002130379207499, + "grad_norm": 15.5, + "learning_rate": 2.142455745390361e-06, + "loss": 0.8227, + "step": 5870 + }, + { + "epoch": 0.5002982530890498, + "grad_norm": 12.0625, + "learning_rate": 2.1423380818924028e-06, + "loss": 0.7189, + "step": 5871 + }, + { + "epoch": 0.5003834682573498, + "grad_norm": 29.5, + "learning_rate": 2.142220402268998e-06, + "loss": 0.8472, + "step": 5872 + }, + { + "epoch": 0.5004686834256498, + "grad_norm": 11.5625, + "learning_rate": 2.142102706522273e-06, + "loss": 0.5844, + "step": 5873 + }, + { + "epoch": 0.5005538985939497, + "grad_norm": 18.375, + "learning_rate": 2.1419849946543554e-06, + "loss": 1.0016, + "step": 5874 + }, + { + "epoch": 0.5006391137622497, + "grad_norm": 18.25, + "learning_rate": 2.1418672666673716e-06, + "loss": 1.0531, + "step": 5875 + }, + { + "epoch": 0.5007243289305496, + "grad_norm": 14.25, + "learning_rate": 2.1417495225634497e-06, + "loss": 0.7746, + "step": 5876 + }, + { + "epoch": 0.5008095440988496, + "grad_norm": 16.5, + "learning_rate": 2.1416317623447176e-06, + "loss": 0.9421, + "step": 5877 + }, + { + "epoch": 0.5008947592671495, + "grad_norm": 21.375, + "learning_rate": 2.141513986013302e-06, + "loss": 0.5886, + "step": 5878 + }, + { + "epoch": 0.5009799744354495, + "grad_norm": 10.0625, + "learning_rate": 2.141396193571333e-06, + "loss": 0.3807, + "step": 5879 + }, + { + "epoch": 0.5010651896037495, + "grad_norm": 12.9375, + "learning_rate": 2.141278385020938e-06, + "loss": 0.5249, + "step": 5880 + }, + { + "epoch": 0.5011504047720494, + "grad_norm": 18.625, + "learning_rate": 2.1411605603642466e-06, + "loss": 0.9046, + "step": 5881 + }, + { + "epoch": 0.5012356199403494, + "grad_norm": 13.0625, + "learning_rate": 2.1410427196033877e-06, + "loss": 0.5197, + "step": 5882 + }, + { + "epoch": 0.5013208351086493, + "grad_norm": 12.3125, + "learning_rate": 2.1409248627404905e-06, + "loss": 0.4229, + "step": 5883 + }, + { + "epoch": 0.5014060502769493, + "grad_norm": 16.125, + "learning_rate": 2.1408069897776857e-06, + "loss": 0.9829, + "step": 5884 + }, + { + "epoch": 0.5014912654452492, + "grad_norm": 16.0, + "learning_rate": 2.1406891007171023e-06, + "loss": 0.9808, + "step": 5885 + }, + { + "epoch": 0.5015764806135492, + "grad_norm": 12.4375, + "learning_rate": 2.1405711955608714e-06, + "loss": 0.5924, + "step": 5886 + }, + { + "epoch": 0.5016616957818492, + "grad_norm": 11.9375, + "learning_rate": 2.1404532743111235e-06, + "loss": 0.532, + "step": 5887 + }, + { + "epoch": 0.5017469109501491, + "grad_norm": 10.4375, + "learning_rate": 2.1403353369699894e-06, + "loss": 0.4992, + "step": 5888 + }, + { + "epoch": 0.5018321261184491, + "grad_norm": 23.5, + "learning_rate": 2.1402173835396006e-06, + "loss": 1.0601, + "step": 5889 + }, + { + "epoch": 0.501917341286749, + "grad_norm": 10.8125, + "learning_rate": 2.1400994140220883e-06, + "loss": 0.5125, + "step": 5890 + }, + { + "epoch": 0.502002556455049, + "grad_norm": 20.25, + "learning_rate": 2.1399814284195848e-06, + "loss": 0.8205, + "step": 5891 + }, + { + "epoch": 0.5020877716233489, + "grad_norm": 12.75, + "learning_rate": 2.139863426734222e-06, + "loss": 0.535, + "step": 5892 + }, + { + "epoch": 0.5021729867916489, + "grad_norm": 12.8125, + "learning_rate": 2.139745408968132e-06, + "loss": 0.3571, + "step": 5893 + }, + { + "epoch": 0.5022582019599489, + "grad_norm": 22.125, + "learning_rate": 2.1396273751234475e-06, + "loss": 1.0411, + "step": 5894 + }, + { + "epoch": 0.5023434171282488, + "grad_norm": 15.4375, + "learning_rate": 2.1395093252023022e-06, + "loss": 0.9942, + "step": 5895 + }, + { + "epoch": 0.5024286322965488, + "grad_norm": 14.1875, + "learning_rate": 2.1393912592068285e-06, + "loss": 0.8437, + "step": 5896 + }, + { + "epoch": 0.5025138474648487, + "grad_norm": 13.8125, + "learning_rate": 2.1392731771391607e-06, + "loss": 0.4644, + "step": 5897 + }, + { + "epoch": 0.5025990626331487, + "grad_norm": 17.0, + "learning_rate": 2.1391550790014327e-06, + "loss": 0.7442, + "step": 5898 + }, + { + "epoch": 0.5026842778014486, + "grad_norm": 14.1875, + "learning_rate": 2.1390369647957777e-06, + "loss": 1.0194, + "step": 5899 + }, + { + "epoch": 0.5027694929697486, + "grad_norm": 16.0, + "learning_rate": 2.138918834524331e-06, + "loss": 0.9234, + "step": 5900 + }, + { + "epoch": 0.5028547081380486, + "grad_norm": 13.5625, + "learning_rate": 2.138800688189227e-06, + "loss": 0.8514, + "step": 5901 + }, + { + "epoch": 0.5029399233063485, + "grad_norm": 13.8125, + "learning_rate": 2.138682525792601e-06, + "loss": 0.7345, + "step": 5902 + }, + { + "epoch": 0.5030251384746485, + "grad_norm": 14.125, + "learning_rate": 2.138564347336588e-06, + "loss": 0.6518, + "step": 5903 + }, + { + "epoch": 0.5031103536429484, + "grad_norm": 12.1875, + "learning_rate": 2.1384461528233235e-06, + "loss": 0.7844, + "step": 5904 + }, + { + "epoch": 0.5031955688112484, + "grad_norm": 14.375, + "learning_rate": 2.138327942254944e-06, + "loss": 0.7696, + "step": 5905 + }, + { + "epoch": 0.5032807839795483, + "grad_norm": 10.0625, + "learning_rate": 2.138209715633585e-06, + "loss": 0.3982, + "step": 5906 + }, + { + "epoch": 0.5033659991478483, + "grad_norm": 13.25, + "learning_rate": 2.138091472961383e-06, + "loss": 0.8193, + "step": 5907 + }, + { + "epoch": 0.5034512143161483, + "grad_norm": 17.5, + "learning_rate": 2.1379732142404758e-06, + "loss": 0.7489, + "step": 5908 + }, + { + "epoch": 0.5035364294844482, + "grad_norm": 13.8125, + "learning_rate": 2.1378549394729993e-06, + "loss": 0.7697, + "step": 5909 + }, + { + "epoch": 0.5036216446527482, + "grad_norm": 17.125, + "learning_rate": 2.1377366486610913e-06, + "loss": 1.1737, + "step": 5910 + }, + { + "epoch": 0.5037068598210481, + "grad_norm": 14.8125, + "learning_rate": 2.137618341806889e-06, + "loss": 0.7671, + "step": 5911 + }, + { + "epoch": 0.5037920749893481, + "grad_norm": 28.125, + "learning_rate": 2.1375000189125316e-06, + "loss": 0.918, + "step": 5912 + }, + { + "epoch": 0.503877290157648, + "grad_norm": 14.4375, + "learning_rate": 2.1373816799801554e-06, + "loss": 0.8216, + "step": 5913 + }, + { + "epoch": 0.503962505325948, + "grad_norm": 12.8125, + "learning_rate": 2.1372633250119005e-06, + "loss": 0.4578, + "step": 5914 + }, + { + "epoch": 0.504047720494248, + "grad_norm": 21.25, + "learning_rate": 2.137144954009905e-06, + "loss": 0.7276, + "step": 5915 + }, + { + "epoch": 0.5041329356625479, + "grad_norm": 15.0625, + "learning_rate": 2.137026566976308e-06, + "loss": 0.6661, + "step": 5916 + }, + { + "epoch": 0.5042181508308479, + "grad_norm": 32.25, + "learning_rate": 2.136908163913249e-06, + "loss": 1.0548, + "step": 5917 + }, + { + "epoch": 0.5043033659991478, + "grad_norm": 21.25, + "learning_rate": 2.1367897448228676e-06, + "loss": 0.9312, + "step": 5918 + }, + { + "epoch": 0.5043885811674478, + "grad_norm": 23.625, + "learning_rate": 2.1366713097073037e-06, + "loss": 1.1637, + "step": 5919 + }, + { + "epoch": 0.5044737963357477, + "grad_norm": 24.0, + "learning_rate": 2.136552858568698e-06, + "loss": 1.2734, + "step": 5920 + }, + { + "epoch": 0.5045590115040477, + "grad_norm": 12.4375, + "learning_rate": 2.13643439140919e-06, + "loss": 0.6037, + "step": 5921 + }, + { + "epoch": 0.5046442266723477, + "grad_norm": 12.75, + "learning_rate": 2.136315908230922e-06, + "loss": 0.4343, + "step": 5922 + }, + { + "epoch": 0.5047294418406476, + "grad_norm": 16.25, + "learning_rate": 2.136197409036034e-06, + "loss": 0.3127, + "step": 5923 + }, + { + "epoch": 0.5048146570089476, + "grad_norm": 11.8125, + "learning_rate": 2.136078893826667e-06, + "loss": 0.5275, + "step": 5924 + }, + { + "epoch": 0.5048998721772475, + "grad_norm": 17.0, + "learning_rate": 2.135960362604964e-06, + "loss": 0.7268, + "step": 5925 + }, + { + "epoch": 0.5049850873455475, + "grad_norm": 11.9375, + "learning_rate": 2.135841815373066e-06, + "loss": 0.6535, + "step": 5926 + }, + { + "epoch": 0.5050703025138474, + "grad_norm": 18.375, + "learning_rate": 2.1357232521331157e-06, + "loss": 1.2585, + "step": 5927 + }, + { + "epoch": 0.5051555176821474, + "grad_norm": 14.125, + "learning_rate": 2.135604672887256e-06, + "loss": 0.7749, + "step": 5928 + }, + { + "epoch": 0.5052407328504473, + "grad_norm": 12.25, + "learning_rate": 2.135486077637629e-06, + "loss": 0.5934, + "step": 5929 + }, + { + "epoch": 0.5053259480187473, + "grad_norm": 9.875, + "learning_rate": 2.135367466386378e-06, + "loss": 0.3538, + "step": 5930 + }, + { + "epoch": 0.5054111631870473, + "grad_norm": 13.4375, + "learning_rate": 2.135248839135647e-06, + "loss": 0.6921, + "step": 5931 + }, + { + "epoch": 0.5054963783553472, + "grad_norm": 14.6875, + "learning_rate": 2.1351301958875788e-06, + "loss": 0.7804, + "step": 5932 + }, + { + "epoch": 0.5055815935236472, + "grad_norm": 11.875, + "learning_rate": 2.135011536644318e-06, + "loss": 0.6082, + "step": 5933 + }, + { + "epoch": 0.5056668086919471, + "grad_norm": 11.625, + "learning_rate": 2.1348928614080093e-06, + "loss": 0.3377, + "step": 5934 + }, + { + "epoch": 0.5057520238602471, + "grad_norm": 11.6875, + "learning_rate": 2.134774170180796e-06, + "loss": 0.9423, + "step": 5935 + }, + { + "epoch": 0.505837239028547, + "grad_norm": 23.125, + "learning_rate": 2.134655462964824e-06, + "loss": 0.7924, + "step": 5936 + }, + { + "epoch": 0.505922454196847, + "grad_norm": 13.75, + "learning_rate": 2.1345367397622386e-06, + "loss": 0.7311, + "step": 5937 + }, + { + "epoch": 0.506007669365147, + "grad_norm": 16.625, + "learning_rate": 2.134418000575185e-06, + "loss": 1.3445, + "step": 5938 + }, + { + "epoch": 0.5060928845334469, + "grad_norm": 11.3125, + "learning_rate": 2.1342992454058077e-06, + "loss": 0.4964, + "step": 5939 + }, + { + "epoch": 0.5061780997017469, + "grad_norm": 15.9375, + "learning_rate": 2.1341804742562545e-06, + "loss": 0.9353, + "step": 5940 + }, + { + "epoch": 0.5062633148700468, + "grad_norm": 12.4375, + "learning_rate": 2.134061687128671e-06, + "loss": 0.8831, + "step": 5941 + }, + { + "epoch": 0.5063485300383468, + "grad_norm": 11.0, + "learning_rate": 2.1339428840252036e-06, + "loss": 0.595, + "step": 5942 + }, + { + "epoch": 0.5064337452066467, + "grad_norm": 17.125, + "learning_rate": 2.1338240649479995e-06, + "loss": 0.6384, + "step": 5943 + }, + { + "epoch": 0.5065189603749467, + "grad_norm": 18.125, + "learning_rate": 2.133705229899206e-06, + "loss": 1.064, + "step": 5944 + }, + { + "epoch": 0.5066041755432467, + "grad_norm": 16.125, + "learning_rate": 2.1335863788809703e-06, + "loss": 0.9516, + "step": 5945 + }, + { + "epoch": 0.5066893907115466, + "grad_norm": 16.375, + "learning_rate": 2.13346751189544e-06, + "loss": 0.5388, + "step": 5946 + }, + { + "epoch": 0.5067746058798466, + "grad_norm": 13.1875, + "learning_rate": 2.1333486289447634e-06, + "loss": 0.3838, + "step": 5947 + }, + { + "epoch": 0.5068598210481465, + "grad_norm": 15.0, + "learning_rate": 2.133229730031089e-06, + "loss": 0.6363, + "step": 5948 + }, + { + "epoch": 0.5069450362164465, + "grad_norm": 34.25, + "learning_rate": 2.1331108151565655e-06, + "loss": 1.8226, + "step": 5949 + }, + { + "epoch": 0.5070302513847464, + "grad_norm": 14.5625, + "learning_rate": 2.1329918843233412e-06, + "loss": 0.7174, + "step": 5950 + }, + { + "epoch": 0.5071154665530464, + "grad_norm": 23.875, + "learning_rate": 2.1328729375335655e-06, + "loss": 0.849, + "step": 5951 + }, + { + "epoch": 0.5072006817213464, + "grad_norm": 33.5, + "learning_rate": 2.1327539747893875e-06, + "loss": 0.9369, + "step": 5952 + }, + { + "epoch": 0.5072858968896463, + "grad_norm": 12.1875, + "learning_rate": 2.132634996092959e-06, + "loss": 0.5689, + "step": 5953 + }, + { + "epoch": 0.5073711120579463, + "grad_norm": 11.6875, + "learning_rate": 2.1325160014464273e-06, + "loss": 0.3275, + "step": 5954 + }, + { + "epoch": 0.5074563272262462, + "grad_norm": 14.8125, + "learning_rate": 2.1323969908519445e-06, + "loss": 0.7506, + "step": 5955 + }, + { + "epoch": 0.5075415423945462, + "grad_norm": 11.125, + "learning_rate": 2.1322779643116603e-06, + "loss": 0.5071, + "step": 5956 + }, + { + "epoch": 0.5076267575628461, + "grad_norm": 20.375, + "learning_rate": 2.1321589218277266e-06, + "loss": 0.6119, + "step": 5957 + }, + { + "epoch": 0.5077119727311461, + "grad_norm": 13.75, + "learning_rate": 2.132039863402294e-06, + "loss": 0.912, + "step": 5958 + }, + { + "epoch": 0.5077971878994461, + "grad_norm": 12.25, + "learning_rate": 2.131920789037514e-06, + "loss": 0.5427, + "step": 5959 + }, + { + "epoch": 0.507882403067746, + "grad_norm": 13.6875, + "learning_rate": 2.1318016987355385e-06, + "loss": 0.5217, + "step": 5960 + }, + { + "epoch": 0.507967618236046, + "grad_norm": 19.75, + "learning_rate": 2.1316825924985198e-06, + "loss": 0.5625, + "step": 5961 + }, + { + "epoch": 0.5080528334043459, + "grad_norm": 11.0, + "learning_rate": 2.1315634703286097e-06, + "loss": 0.565, + "step": 5962 + }, + { + "epoch": 0.5081380485726459, + "grad_norm": 13.375, + "learning_rate": 2.1314443322279615e-06, + "loss": 0.7379, + "step": 5963 + }, + { + "epoch": 0.5082232637409458, + "grad_norm": 16.125, + "learning_rate": 2.131325178198728e-06, + "loss": 0.6521, + "step": 5964 + }, + { + "epoch": 0.5083084789092458, + "grad_norm": 19.0, + "learning_rate": 2.131206008243062e-06, + "loss": 1.0253, + "step": 5965 + }, + { + "epoch": 0.5083936940775458, + "grad_norm": 14.875, + "learning_rate": 2.131086822363117e-06, + "loss": 0.4935, + "step": 5966 + }, + { + "epoch": 0.5084789092458457, + "grad_norm": 29.625, + "learning_rate": 2.1309676205610474e-06, + "loss": 1.0021, + "step": 5967 + }, + { + "epoch": 0.5085641244141457, + "grad_norm": 12.5625, + "learning_rate": 2.1308484028390073e-06, + "loss": 0.7737, + "step": 5968 + }, + { + "epoch": 0.5086493395824456, + "grad_norm": 15.375, + "learning_rate": 2.1307291691991506e-06, + "loss": 0.7197, + "step": 5969 + }, + { + "epoch": 0.5087345547507456, + "grad_norm": 14.9375, + "learning_rate": 2.130609919643632e-06, + "loss": 0.6636, + "step": 5970 + }, + { + "epoch": 0.5088197699190455, + "grad_norm": 10.8125, + "learning_rate": 2.130490654174607e-06, + "loss": 0.5925, + "step": 5971 + }, + { + "epoch": 0.5089049850873455, + "grad_norm": 13.75, + "learning_rate": 2.13037137279423e-06, + "loss": 0.9755, + "step": 5972 + }, + { + "epoch": 0.5089902002556455, + "grad_norm": 14.8125, + "learning_rate": 2.1302520755046572e-06, + "loss": 0.7805, + "step": 5973 + }, + { + "epoch": 0.5090754154239454, + "grad_norm": 15.5, + "learning_rate": 2.130132762308045e-06, + "loss": 0.9451, + "step": 5974 + }, + { + "epoch": 0.5091606305922454, + "grad_norm": 16.125, + "learning_rate": 2.1300134332065476e-06, + "loss": 0.5044, + "step": 5975 + }, + { + "epoch": 0.5092458457605453, + "grad_norm": 21.125, + "learning_rate": 2.129894088202323e-06, + "loss": 1.1951, + "step": 5976 + }, + { + "epoch": 0.5093310609288453, + "grad_norm": 11.1875, + "learning_rate": 2.1297747272975274e-06, + "loss": 0.4616, + "step": 5977 + }, + { + "epoch": 0.5094162760971452, + "grad_norm": 16.5, + "learning_rate": 2.129655350494318e-06, + "loss": 1.068, + "step": 5978 + }, + { + "epoch": 0.5095014912654452, + "grad_norm": 12.75, + "learning_rate": 2.1295359577948515e-06, + "loss": 0.4224, + "step": 5979 + }, + { + "epoch": 0.5095867064337452, + "grad_norm": 14.375, + "learning_rate": 2.1294165492012856e-06, + "loss": 0.9224, + "step": 5980 + }, + { + "epoch": 0.5096719216020452, + "grad_norm": 10.875, + "learning_rate": 2.129297124715779e-06, + "loss": 0.4246, + "step": 5981 + }, + { + "epoch": 0.5097571367703452, + "grad_norm": 13.0625, + "learning_rate": 2.129177684340489e-06, + "loss": 0.7747, + "step": 5982 + }, + { + "epoch": 0.5098423519386451, + "grad_norm": 16.0, + "learning_rate": 2.129058228077574e-06, + "loss": 0.7793, + "step": 5983 + }, + { + "epoch": 0.5099275671069451, + "grad_norm": 18.375, + "learning_rate": 2.1289387559291927e-06, + "loss": 0.492, + "step": 5984 + }, + { + "epoch": 0.510012782275245, + "grad_norm": 21.875, + "learning_rate": 2.1288192678975046e-06, + "loss": 0.8946, + "step": 5985 + }, + { + "epoch": 0.510097997443545, + "grad_norm": 13.0625, + "learning_rate": 2.1286997639846683e-06, + "loss": 0.7928, + "step": 5986 + }, + { + "epoch": 0.510183212611845, + "grad_norm": 10.125, + "learning_rate": 2.1285802441928437e-06, + "loss": 0.4138, + "step": 5987 + }, + { + "epoch": 0.5102684277801449, + "grad_norm": 13.9375, + "learning_rate": 2.128460708524191e-06, + "loss": 0.8671, + "step": 5988 + }, + { + "epoch": 0.5103536429484449, + "grad_norm": 16.375, + "learning_rate": 2.1283411569808695e-06, + "loss": 0.9474, + "step": 5989 + }, + { + "epoch": 0.5104388581167448, + "grad_norm": 22.125, + "learning_rate": 2.1282215895650405e-06, + "loss": 0.9463, + "step": 5990 + }, + { + "epoch": 0.5105240732850448, + "grad_norm": 15.4375, + "learning_rate": 2.128102006278864e-06, + "loss": 1.1134, + "step": 5991 + }, + { + "epoch": 0.5106092884533447, + "grad_norm": 19.5, + "learning_rate": 2.1279824071245007e-06, + "loss": 0.8526, + "step": 5992 + }, + { + "epoch": 0.5106945036216447, + "grad_norm": 20.125, + "learning_rate": 2.1278627921041128e-06, + "loss": 0.8309, + "step": 5993 + }, + { + "epoch": 0.5107797187899447, + "grad_norm": 9.9375, + "learning_rate": 2.127743161219862e-06, + "loss": 0.3082, + "step": 5994 + }, + { + "epoch": 0.5108649339582446, + "grad_norm": 11.125, + "learning_rate": 2.1276235144739087e-06, + "loss": 0.3851, + "step": 5995 + }, + { + "epoch": 0.5109501491265446, + "grad_norm": 13.4375, + "learning_rate": 2.1275038518684168e-06, + "loss": 0.6475, + "step": 5996 + }, + { + "epoch": 0.5110353642948445, + "grad_norm": 14.0625, + "learning_rate": 2.1273841734055478e-06, + "loss": 0.6217, + "step": 5997 + }, + { + "epoch": 0.5111205794631445, + "grad_norm": 13.5625, + "learning_rate": 2.127264479087464e-06, + "loss": 0.7222, + "step": 5998 + }, + { + "epoch": 0.5112057946314444, + "grad_norm": 14.5, + "learning_rate": 2.1271447689163293e-06, + "loss": 0.5706, + "step": 5999 + }, + { + "epoch": 0.5112910097997444, + "grad_norm": 11.8125, + "learning_rate": 2.1270250428943064e-06, + "loss": 0.5834, + "step": 6000 + }, + { + "epoch": 0.5113762249680444, + "grad_norm": 18.0, + "learning_rate": 2.126905301023559e-06, + "loss": 0.9105, + "step": 6001 + }, + { + "epoch": 0.5114614401363443, + "grad_norm": 13.1875, + "learning_rate": 2.126785543306252e-06, + "loss": 0.8711, + "step": 6002 + }, + { + "epoch": 0.5115466553046443, + "grad_norm": 15.875, + "learning_rate": 2.126665769744547e-06, + "loss": 0.7354, + "step": 6003 + }, + { + "epoch": 0.5116318704729442, + "grad_norm": 15.75, + "learning_rate": 2.126545980340611e-06, + "loss": 0.615, + "step": 6004 + }, + { + "epoch": 0.5117170856412442, + "grad_norm": 11.8125, + "learning_rate": 2.1264261750966073e-06, + "loss": 0.5414, + "step": 6005 + }, + { + "epoch": 0.5118023008095441, + "grad_norm": 19.625, + "learning_rate": 2.1263063540147018e-06, + "loss": 0.7888, + "step": 6006 + }, + { + "epoch": 0.5118875159778441, + "grad_norm": 14.0625, + "learning_rate": 2.1261865170970587e-06, + "loss": 0.5355, + "step": 6007 + }, + { + "epoch": 0.5119727311461441, + "grad_norm": 13.25, + "learning_rate": 2.1260666643458445e-06, + "loss": 0.552, + "step": 6008 + }, + { + "epoch": 0.512057946314444, + "grad_norm": 14.3125, + "learning_rate": 2.1259467957632244e-06, + "loss": 0.9799, + "step": 6009 + }, + { + "epoch": 0.512143161482744, + "grad_norm": 13.9375, + "learning_rate": 2.1258269113513653e-06, + "loss": 0.6855, + "step": 6010 + }, + { + "epoch": 0.5122283766510439, + "grad_norm": 12.1875, + "learning_rate": 2.125707011112433e-06, + "loss": 0.5838, + "step": 6011 + }, + { + "epoch": 0.5123135918193439, + "grad_norm": 13.6875, + "learning_rate": 2.125587095048594e-06, + "loss": 0.6412, + "step": 6012 + }, + { + "epoch": 0.5123988069876438, + "grad_norm": 12.1875, + "learning_rate": 2.125467163162016e-06, + "loss": 0.5421, + "step": 6013 + }, + { + "epoch": 0.5124840221559438, + "grad_norm": 19.875, + "learning_rate": 2.1253472154548665e-06, + "loss": 0.6201, + "step": 6014 + }, + { + "epoch": 0.5125692373242438, + "grad_norm": 15.0, + "learning_rate": 2.1252272519293123e-06, + "loss": 0.6306, + "step": 6015 + }, + { + "epoch": 0.5126544524925437, + "grad_norm": 15.5, + "learning_rate": 2.125107272587521e-06, + "loss": 0.4924, + "step": 6016 + }, + { + "epoch": 0.5127396676608437, + "grad_norm": 21.25, + "learning_rate": 2.1249872774316623e-06, + "loss": 0.8329, + "step": 6017 + }, + { + "epoch": 0.5128248828291436, + "grad_norm": 12.1875, + "learning_rate": 2.124867266463903e-06, + "loss": 0.7385, + "step": 6018 + }, + { + "epoch": 0.5129100979974436, + "grad_norm": 14.0, + "learning_rate": 2.1247472396864123e-06, + "loss": 0.7668, + "step": 6019 + }, + { + "epoch": 0.5129953131657435, + "grad_norm": 14.875, + "learning_rate": 2.1246271971013596e-06, + "loss": 0.9232, + "step": 6020 + }, + { + "epoch": 0.5130805283340435, + "grad_norm": 14.6875, + "learning_rate": 2.124507138710914e-06, + "loss": 1.0421, + "step": 6021 + }, + { + "epoch": 0.5131657435023435, + "grad_norm": 19.125, + "learning_rate": 2.124387064517245e-06, + "loss": 0.8987, + "step": 6022 + }, + { + "epoch": 0.5132509586706434, + "grad_norm": 17.0, + "learning_rate": 2.1242669745225223e-06, + "loss": 0.9868, + "step": 6023 + }, + { + "epoch": 0.5133361738389434, + "grad_norm": 14.8125, + "learning_rate": 2.124146868728916e-06, + "loss": 0.7766, + "step": 6024 + }, + { + "epoch": 0.5134213890072433, + "grad_norm": 16.0, + "learning_rate": 2.124026747138597e-06, + "loss": 0.9565, + "step": 6025 + }, + { + "epoch": 0.5135066041755433, + "grad_norm": 16.375, + "learning_rate": 2.1239066097537363e-06, + "loss": 0.9806, + "step": 6026 + }, + { + "epoch": 0.5135918193438432, + "grad_norm": 10.4375, + "learning_rate": 2.123786456576504e-06, + "loss": 0.3605, + "step": 6027 + }, + { + "epoch": 0.5136770345121432, + "grad_norm": 19.25, + "learning_rate": 2.123666287609072e-06, + "loss": 0.8089, + "step": 6028 + }, + { + "epoch": 0.5137622496804432, + "grad_norm": 17.875, + "learning_rate": 2.1235461028536112e-06, + "loss": 0.8724, + "step": 6029 + }, + { + "epoch": 0.5138474648487431, + "grad_norm": 15.625, + "learning_rate": 2.123425902312294e-06, + "loss": 0.736, + "step": 6030 + }, + { + "epoch": 0.5139326800170431, + "grad_norm": 14.5625, + "learning_rate": 2.1233056859872925e-06, + "loss": 0.8992, + "step": 6031 + }, + { + "epoch": 0.514017895185343, + "grad_norm": 15.6875, + "learning_rate": 2.123185453880779e-06, + "loss": 1.0508, + "step": 6032 + }, + { + "epoch": 0.514103110353643, + "grad_norm": 11.625, + "learning_rate": 2.1230652059949265e-06, + "loss": 0.5298, + "step": 6033 + }, + { + "epoch": 0.5141883255219429, + "grad_norm": 11.625, + "learning_rate": 2.1229449423319075e-06, + "loss": 0.7364, + "step": 6034 + }, + { + "epoch": 0.5142735406902429, + "grad_norm": 17.5, + "learning_rate": 2.122824662893896e-06, + "loss": 0.864, + "step": 6035 + }, + { + "epoch": 0.5143587558585428, + "grad_norm": 11.1875, + "learning_rate": 2.122704367683065e-06, + "loss": 0.9077, + "step": 6036 + }, + { + "epoch": 0.5144439710268428, + "grad_norm": 15.6875, + "learning_rate": 2.122584056701588e-06, + "loss": 0.8837, + "step": 6037 + }, + { + "epoch": 0.5145291861951428, + "grad_norm": 23.0, + "learning_rate": 2.1224637299516403e-06, + "loss": 1.0585, + "step": 6038 + }, + { + "epoch": 0.5146144013634427, + "grad_norm": 17.0, + "learning_rate": 2.1223433874353956e-06, + "loss": 1.0097, + "step": 6039 + }, + { + "epoch": 0.5146996165317427, + "grad_norm": 14.125, + "learning_rate": 2.1222230291550284e-06, + "loss": 0.9164, + "step": 6040 + }, + { + "epoch": 0.5147848317000426, + "grad_norm": 16.125, + "learning_rate": 2.122102655112714e-06, + "loss": 0.6209, + "step": 6041 + }, + { + "epoch": 0.5148700468683426, + "grad_norm": 15.0625, + "learning_rate": 2.121982265310628e-06, + "loss": 0.6848, + "step": 6042 + }, + { + "epoch": 0.5149552620366425, + "grad_norm": 11.0625, + "learning_rate": 2.1218618597509453e-06, + "loss": 0.503, + "step": 6043 + }, + { + "epoch": 0.5150404772049425, + "grad_norm": 15.125, + "learning_rate": 2.121741438435842e-06, + "loss": 0.8655, + "step": 6044 + }, + { + "epoch": 0.5151256923732425, + "grad_norm": 14.3125, + "learning_rate": 2.121621001367494e-06, + "loss": 0.9471, + "step": 6045 + }, + { + "epoch": 0.5152109075415424, + "grad_norm": 11.25, + "learning_rate": 2.121500548548079e-06, + "loss": 0.4627, + "step": 6046 + }, + { + "epoch": 0.5152961227098424, + "grad_norm": 21.125, + "learning_rate": 2.121380079979772e-06, + "loss": 0.9475, + "step": 6047 + }, + { + "epoch": 0.5153813378781423, + "grad_norm": 16.125, + "learning_rate": 2.121259595664751e-06, + "loss": 0.8405, + "step": 6048 + }, + { + "epoch": 0.5154665530464423, + "grad_norm": 19.625, + "learning_rate": 2.1211390956051927e-06, + "loss": 1.0412, + "step": 6049 + }, + { + "epoch": 0.5155517682147422, + "grad_norm": 16.625, + "learning_rate": 2.121018579803275e-06, + "loss": 0.9891, + "step": 6050 + }, + { + "epoch": 0.5156369833830422, + "grad_norm": 14.25, + "learning_rate": 2.1208980482611763e-06, + "loss": 0.6184, + "step": 6051 + }, + { + "epoch": 0.5157221985513422, + "grad_norm": 15.125, + "learning_rate": 2.1207775009810736e-06, + "loss": 1.0893, + "step": 6052 + }, + { + "epoch": 0.5158074137196421, + "grad_norm": 14.0, + "learning_rate": 2.1206569379651456e-06, + "loss": 0.6063, + "step": 6053 + }, + { + "epoch": 0.5158926288879421, + "grad_norm": 16.75, + "learning_rate": 2.1205363592155713e-06, + "loss": 1.0436, + "step": 6054 + }, + { + "epoch": 0.515977844056242, + "grad_norm": 16.625, + "learning_rate": 2.12041576473453e-06, + "loss": 0.7825, + "step": 6055 + }, + { + "epoch": 0.516063059224542, + "grad_norm": 18.125, + "learning_rate": 2.1202951545242006e-06, + "loss": 0.8175, + "step": 6056 + }, + { + "epoch": 0.5161482743928419, + "grad_norm": 12.75, + "learning_rate": 2.120174528586762e-06, + "loss": 0.3533, + "step": 6057 + }, + { + "epoch": 0.5162334895611419, + "grad_norm": 13.1875, + "learning_rate": 2.120053886924395e-06, + "loss": 0.498, + "step": 6058 + }, + { + "epoch": 0.5163187047294419, + "grad_norm": 15.125, + "learning_rate": 2.1199332295392796e-06, + "loss": 1.0237, + "step": 6059 + }, + { + "epoch": 0.5164039198977418, + "grad_norm": 25.125, + "learning_rate": 2.1198125564335956e-06, + "loss": 1.0128, + "step": 6060 + }, + { + "epoch": 0.5164891350660418, + "grad_norm": 13.8125, + "learning_rate": 2.119691867609524e-06, + "loss": 0.6675, + "step": 6061 + }, + { + "epoch": 0.5165743502343417, + "grad_norm": 14.8125, + "learning_rate": 2.1195711630692463e-06, + "loss": 0.7392, + "step": 6062 + }, + { + "epoch": 0.5166595654026417, + "grad_norm": 17.625, + "learning_rate": 2.1194504428149425e-06, + "loss": 0.7134, + "step": 6063 + }, + { + "epoch": 0.5167447805709416, + "grad_norm": 22.25, + "learning_rate": 2.1193297068487958e-06, + "loss": 0.8278, + "step": 6064 + }, + { + "epoch": 0.5168299957392416, + "grad_norm": 14.9375, + "learning_rate": 2.119208955172986e-06, + "loss": 0.7356, + "step": 6065 + }, + { + "epoch": 0.5169152109075416, + "grad_norm": 25.375, + "learning_rate": 2.119088187789697e-06, + "loss": 1.1296, + "step": 6066 + }, + { + "epoch": 0.5170004260758415, + "grad_norm": 9.3125, + "learning_rate": 2.1189674047011106e-06, + "loss": 0.3042, + "step": 6067 + }, + { + "epoch": 0.5170856412441415, + "grad_norm": 10.4375, + "learning_rate": 2.1188466059094092e-06, + "loss": 0.4142, + "step": 6068 + }, + { + "epoch": 0.5171708564124414, + "grad_norm": 18.375, + "learning_rate": 2.118725791416776e-06, + "loss": 0.8698, + "step": 6069 + }, + { + "epoch": 0.5172560715807414, + "grad_norm": 17.125, + "learning_rate": 2.118604961225394e-06, + "loss": 0.7608, + "step": 6070 + }, + { + "epoch": 0.5173412867490413, + "grad_norm": 17.0, + "learning_rate": 2.1184841153374475e-06, + "loss": 0.8325, + "step": 6071 + }, + { + "epoch": 0.5174265019173413, + "grad_norm": 13.625, + "learning_rate": 2.118363253755119e-06, + "loss": 0.4737, + "step": 6072 + }, + { + "epoch": 0.5175117170856413, + "grad_norm": 15.3125, + "learning_rate": 2.1182423764805937e-06, + "loss": 0.8842, + "step": 6073 + }, + { + "epoch": 0.5175969322539412, + "grad_norm": 14.3125, + "learning_rate": 2.118121483516055e-06, + "loss": 0.6919, + "step": 6074 + }, + { + "epoch": 0.5176821474222412, + "grad_norm": 14.0, + "learning_rate": 2.118000574863689e-06, + "loss": 0.7782, + "step": 6075 + }, + { + "epoch": 0.5177673625905411, + "grad_norm": 20.375, + "learning_rate": 2.1178796505256794e-06, + "loss": 1.2393, + "step": 6076 + }, + { + "epoch": 0.5178525777588411, + "grad_norm": 13.0, + "learning_rate": 2.1177587105042114e-06, + "loss": 0.6245, + "step": 6077 + }, + { + "epoch": 0.517937792927141, + "grad_norm": 13.3125, + "learning_rate": 2.1176377548014717e-06, + "loss": 0.6985, + "step": 6078 + }, + { + "epoch": 0.518023008095441, + "grad_norm": 10.875, + "learning_rate": 2.1175167834196443e-06, + "loss": 0.2675, + "step": 6079 + }, + { + "epoch": 0.518108223263741, + "grad_norm": 10.3125, + "learning_rate": 2.117395796360917e-06, + "loss": 0.4387, + "step": 6080 + }, + { + "epoch": 0.5181934384320409, + "grad_norm": 16.125, + "learning_rate": 2.117274793627475e-06, + "loss": 0.7577, + "step": 6081 + }, + { + "epoch": 0.5182786536003409, + "grad_norm": 23.375, + "learning_rate": 2.117153775221506e-06, + "loss": 1.2516, + "step": 6082 + }, + { + "epoch": 0.5183638687686408, + "grad_norm": 16.625, + "learning_rate": 2.1170327411451957e-06, + "loss": 0.7569, + "step": 6083 + }, + { + "epoch": 0.5184490839369408, + "grad_norm": 15.625, + "learning_rate": 2.116911691400732e-06, + "loss": 0.8209, + "step": 6084 + }, + { + "epoch": 0.5185342991052407, + "grad_norm": 29.75, + "learning_rate": 2.116790625990302e-06, + "loss": 0.9579, + "step": 6085 + }, + { + "epoch": 0.5186195142735407, + "grad_norm": 13.0625, + "learning_rate": 2.1166695449160946e-06, + "loss": 0.7305, + "step": 6086 + }, + { + "epoch": 0.5187047294418407, + "grad_norm": 11.6875, + "learning_rate": 2.116548448180297e-06, + "loss": 0.4928, + "step": 6087 + }, + { + "epoch": 0.5187899446101406, + "grad_norm": 14.75, + "learning_rate": 2.1164273357850965e-06, + "loss": 0.6753, + "step": 6088 + }, + { + "epoch": 0.5188751597784406, + "grad_norm": 15.5625, + "learning_rate": 2.1163062077326836e-06, + "loss": 0.6788, + "step": 6089 + }, + { + "epoch": 0.5189603749467405, + "grad_norm": 20.875, + "learning_rate": 2.116185064025246e-06, + "loss": 1.0482, + "step": 6090 + }, + { + "epoch": 0.5190455901150405, + "grad_norm": 18.75, + "learning_rate": 2.1160639046649733e-06, + "loss": 0.6192, + "step": 6091 + }, + { + "epoch": 0.5191308052833404, + "grad_norm": 17.75, + "learning_rate": 2.1159427296540554e-06, + "loss": 1.3317, + "step": 6092 + }, + { + "epoch": 0.5192160204516404, + "grad_norm": 10.8125, + "learning_rate": 2.115821538994681e-06, + "loss": 0.4594, + "step": 6093 + }, + { + "epoch": 0.5193012356199403, + "grad_norm": 17.625, + "learning_rate": 2.1157003326890413e-06, + "loss": 0.6357, + "step": 6094 + }, + { + "epoch": 0.5193864507882403, + "grad_norm": 10.875, + "learning_rate": 2.115579110739326e-06, + "loss": 0.439, + "step": 6095 + }, + { + "epoch": 0.5194716659565403, + "grad_norm": 16.25, + "learning_rate": 2.115457873147725e-06, + "loss": 0.8373, + "step": 6096 + }, + { + "epoch": 0.5195568811248402, + "grad_norm": 16.875, + "learning_rate": 2.1153366199164306e-06, + "loss": 0.8145, + "step": 6097 + }, + { + "epoch": 0.5196420962931402, + "grad_norm": 15.0625, + "learning_rate": 2.1152153510476333e-06, + "loss": 1.0841, + "step": 6098 + }, + { + "epoch": 0.5197273114614401, + "grad_norm": 16.125, + "learning_rate": 2.1150940665435247e-06, + "loss": 0.5388, + "step": 6099 + }, + { + "epoch": 0.5198125266297401, + "grad_norm": 16.125, + "learning_rate": 2.114972766406296e-06, + "loss": 0.7404, + "step": 6100 + }, + { + "epoch": 0.51989774179804, + "grad_norm": 17.125, + "learning_rate": 2.11485145063814e-06, + "loss": 0.9856, + "step": 6101 + }, + { + "epoch": 0.51998295696634, + "grad_norm": 17.5, + "learning_rate": 2.1147301192412485e-06, + "loss": 0.845, + "step": 6102 + }, + { + "epoch": 0.52006817213464, + "grad_norm": 18.5, + "learning_rate": 2.1146087722178138e-06, + "loss": 0.9277, + "step": 6103 + }, + { + "epoch": 0.5201533873029399, + "grad_norm": 18.0, + "learning_rate": 2.1144874095700297e-06, + "loss": 0.7724, + "step": 6104 + }, + { + "epoch": 0.5202386024712399, + "grad_norm": 14.1875, + "learning_rate": 2.1143660313000887e-06, + "loss": 0.8382, + "step": 6105 + }, + { + "epoch": 0.5203238176395398, + "grad_norm": 15.75, + "learning_rate": 2.114244637410184e-06, + "loss": 0.7275, + "step": 6106 + }, + { + "epoch": 0.5204090328078398, + "grad_norm": 12.5625, + "learning_rate": 2.1141232279025097e-06, + "loss": 0.5651, + "step": 6107 + }, + { + "epoch": 0.5204942479761397, + "grad_norm": 17.625, + "learning_rate": 2.1140018027792597e-06, + "loss": 0.6045, + "step": 6108 + }, + { + "epoch": 0.5205794631444397, + "grad_norm": 11.6875, + "learning_rate": 2.1138803620426285e-06, + "loss": 0.5992, + "step": 6109 + }, + { + "epoch": 0.5206646783127397, + "grad_norm": 15.875, + "learning_rate": 2.1137589056948104e-06, + "loss": 0.5883, + "step": 6110 + }, + { + "epoch": 0.5207498934810396, + "grad_norm": 15.4375, + "learning_rate": 2.1136374337380005e-06, + "loss": 1.1205, + "step": 6111 + }, + { + "epoch": 0.5208351086493396, + "grad_norm": 14.0625, + "learning_rate": 2.1135159461743937e-06, + "loss": 0.863, + "step": 6112 + }, + { + "epoch": 0.5209203238176395, + "grad_norm": 14.875, + "learning_rate": 2.1133944430061853e-06, + "loss": 0.6164, + "step": 6113 + }, + { + "epoch": 0.5210055389859395, + "grad_norm": 17.25, + "learning_rate": 2.1132729242355707e-06, + "loss": 0.672, + "step": 6114 + }, + { + "epoch": 0.5210907541542394, + "grad_norm": 16.875, + "learning_rate": 2.1131513898647466e-06, + "loss": 0.7632, + "step": 6115 + }, + { + "epoch": 0.5211759693225394, + "grad_norm": 15.125, + "learning_rate": 2.113029839895909e-06, + "loss": 0.8079, + "step": 6116 + }, + { + "epoch": 0.5212611844908394, + "grad_norm": 15.125, + "learning_rate": 2.112908274331254e-06, + "loss": 0.6959, + "step": 6117 + }, + { + "epoch": 0.5213463996591393, + "grad_norm": 19.0, + "learning_rate": 2.112786693172979e-06, + "loss": 0.6458, + "step": 6118 + }, + { + "epoch": 0.5214316148274393, + "grad_norm": 13.75, + "learning_rate": 2.112665096423281e-06, + "loss": 0.4932, + "step": 6119 + }, + { + "epoch": 0.5215168299957392, + "grad_norm": 22.625, + "learning_rate": 2.112543484084357e-06, + "loss": 1.2836, + "step": 6120 + }, + { + "epoch": 0.5216020451640392, + "grad_norm": 19.0, + "learning_rate": 2.1124218561584053e-06, + "loss": 1.1012, + "step": 6121 + }, + { + "epoch": 0.5216872603323391, + "grad_norm": 17.25, + "learning_rate": 2.112300212647623e-06, + "loss": 1.1612, + "step": 6122 + }, + { + "epoch": 0.5217724755006391, + "grad_norm": 21.375, + "learning_rate": 2.1121785535542093e-06, + "loss": 0.5436, + "step": 6123 + }, + { + "epoch": 0.5218576906689391, + "grad_norm": 14.125, + "learning_rate": 2.1120568788803617e-06, + "loss": 0.5055, + "step": 6124 + }, + { + "epoch": 0.521942905837239, + "grad_norm": 15.0625, + "learning_rate": 2.1119351886282797e-06, + "loss": 0.6316, + "step": 6125 + }, + { + "epoch": 0.522028121005539, + "grad_norm": 13.75, + "learning_rate": 2.1118134828001618e-06, + "loss": 0.7559, + "step": 6126 + }, + { + "epoch": 0.5221133361738389, + "grad_norm": 16.25, + "learning_rate": 2.1116917613982072e-06, + "loss": 0.568, + "step": 6127 + }, + { + "epoch": 0.5221985513421389, + "grad_norm": 12.3125, + "learning_rate": 2.1115700244246167e-06, + "loss": 0.6198, + "step": 6128 + }, + { + "epoch": 0.5222837665104388, + "grad_norm": 14.8125, + "learning_rate": 2.1114482718815895e-06, + "loss": 0.5334, + "step": 6129 + }, + { + "epoch": 0.5223689816787388, + "grad_norm": 23.75, + "learning_rate": 2.1113265037713253e-06, + "loss": 0.7541, + "step": 6130 + }, + { + "epoch": 0.5224541968470388, + "grad_norm": 16.625, + "learning_rate": 2.1112047200960256e-06, + "loss": 0.7181, + "step": 6131 + }, + { + "epoch": 0.5225394120153387, + "grad_norm": 16.625, + "learning_rate": 2.1110829208578905e-06, + "loss": 0.7677, + "step": 6132 + }, + { + "epoch": 0.5226246271836387, + "grad_norm": 14.5625, + "learning_rate": 2.1109611060591207e-06, + "loss": 0.7057, + "step": 6133 + }, + { + "epoch": 0.5227098423519386, + "grad_norm": 16.625, + "learning_rate": 2.110839275701918e-06, + "loss": 0.726, + "step": 6134 + }, + { + "epoch": 0.5227950575202386, + "grad_norm": 12.625, + "learning_rate": 2.1107174297884843e-06, + "loss": 0.4921, + "step": 6135 + }, + { + "epoch": 0.5228802726885385, + "grad_norm": 14.0, + "learning_rate": 2.110595568321021e-06, + "loss": 0.7788, + "step": 6136 + }, + { + "epoch": 0.5229654878568385, + "grad_norm": 17.75, + "learning_rate": 2.1104736913017304e-06, + "loss": 1.323, + "step": 6137 + }, + { + "epoch": 0.5230507030251385, + "grad_norm": 20.875, + "learning_rate": 2.110351798732815e-06, + "loss": 0.7678, + "step": 6138 + }, + { + "epoch": 0.5231359181934384, + "grad_norm": 19.75, + "learning_rate": 2.110229890616477e-06, + "loss": 1.2404, + "step": 6139 + }, + { + "epoch": 0.5232211333617384, + "grad_norm": 20.0, + "learning_rate": 2.11010796695492e-06, + "loss": 0.6425, + "step": 6140 + }, + { + "epoch": 0.5233063485300383, + "grad_norm": 14.5, + "learning_rate": 2.1099860277503475e-06, + "loss": 0.7397, + "step": 6141 + }, + { + "epoch": 0.5233915636983383, + "grad_norm": 16.25, + "learning_rate": 2.1098640730049623e-06, + "loss": 0.7646, + "step": 6142 + }, + { + "epoch": 0.5234767788666382, + "grad_norm": 11.6875, + "learning_rate": 2.109742102720969e-06, + "loss": 0.4993, + "step": 6143 + }, + { + "epoch": 0.5235619940349382, + "grad_norm": 14.3125, + "learning_rate": 2.1096201169005716e-06, + "loss": 0.614, + "step": 6144 + }, + { + "epoch": 0.5236472092032382, + "grad_norm": 13.4375, + "learning_rate": 2.1094981155459736e-06, + "loss": 0.5377, + "step": 6145 + }, + { + "epoch": 0.5237324243715381, + "grad_norm": 13.0625, + "learning_rate": 2.1093760986593804e-06, + "loss": 0.8189, + "step": 6146 + }, + { + "epoch": 0.5238176395398381, + "grad_norm": 14.75, + "learning_rate": 2.109254066242997e-06, + "loss": 0.7903, + "step": 6147 + }, + { + "epoch": 0.523902854708138, + "grad_norm": 15.625, + "learning_rate": 2.109132018299029e-06, + "loss": 0.8813, + "step": 6148 + }, + { + "epoch": 0.523988069876438, + "grad_norm": 9.1875, + "learning_rate": 2.1090099548296814e-06, + "loss": 0.3579, + "step": 6149 + }, + { + "epoch": 0.5240732850447379, + "grad_norm": 16.5, + "learning_rate": 2.10888787583716e-06, + "loss": 0.6973, + "step": 6150 + }, + { + "epoch": 0.5241585002130379, + "grad_norm": 16.0, + "learning_rate": 2.108765781323671e-06, + "loss": 0.6114, + "step": 6151 + }, + { + "epoch": 0.5242437153813378, + "grad_norm": 16.375, + "learning_rate": 2.1086436712914206e-06, + "loss": 0.9928, + "step": 6152 + }, + { + "epoch": 0.5243289305496378, + "grad_norm": 12.5, + "learning_rate": 2.1085215457426157e-06, + "loss": 0.3729, + "step": 6153 + }, + { + "epoch": 0.5244141457179378, + "grad_norm": 13.9375, + "learning_rate": 2.108399404679463e-06, + "loss": 0.9815, + "step": 6154 + }, + { + "epoch": 0.5244993608862377, + "grad_norm": 16.125, + "learning_rate": 2.10827724810417e-06, + "loss": 0.9905, + "step": 6155 + }, + { + "epoch": 0.5245845760545377, + "grad_norm": 16.0, + "learning_rate": 2.1081550760189444e-06, + "loss": 0.7105, + "step": 6156 + }, + { + "epoch": 0.5246697912228376, + "grad_norm": 14.375, + "learning_rate": 2.108032888425994e-06, + "loss": 0.552, + "step": 6157 + }, + { + "epoch": 0.5247550063911376, + "grad_norm": 13.25, + "learning_rate": 2.1079106853275255e-06, + "loss": 0.5625, + "step": 6158 + }, + { + "epoch": 0.5248402215594375, + "grad_norm": 25.25, + "learning_rate": 2.1077884667257483e-06, + "loss": 1.145, + "step": 6159 + }, + { + "epoch": 0.5249254367277375, + "grad_norm": 11.5625, + "learning_rate": 2.1076662326228713e-06, + "loss": 0.3689, + "step": 6160 + }, + { + "epoch": 0.5250106518960375, + "grad_norm": 14.25, + "learning_rate": 2.107543983021103e-06, + "loss": 0.5504, + "step": 6161 + }, + { + "epoch": 0.5250958670643374, + "grad_norm": 14.875, + "learning_rate": 2.1074217179226525e-06, + "loss": 0.6711, + "step": 6162 + }, + { + "epoch": 0.5251810822326374, + "grad_norm": 13.3125, + "learning_rate": 2.1072994373297297e-06, + "loss": 0.5921, + "step": 6163 + }, + { + "epoch": 0.5252662974009373, + "grad_norm": 13.375, + "learning_rate": 2.1071771412445434e-06, + "loss": 0.6126, + "step": 6164 + }, + { + "epoch": 0.5253515125692373, + "grad_norm": 9.25, + "learning_rate": 2.1070548296693043e-06, + "loss": 0.2824, + "step": 6165 + }, + { + "epoch": 0.5254367277375372, + "grad_norm": 11.375, + "learning_rate": 2.1069325026062232e-06, + "loss": 0.4675, + "step": 6166 + }, + { + "epoch": 0.5255219429058372, + "grad_norm": 16.125, + "learning_rate": 2.1068101600575096e-06, + "loss": 0.8072, + "step": 6167 + }, + { + "epoch": 0.5256071580741372, + "grad_norm": 18.125, + "learning_rate": 2.106687802025375e-06, + "loss": 1.2718, + "step": 6168 + }, + { + "epoch": 0.5256923732424371, + "grad_norm": 12.9375, + "learning_rate": 2.10656542851203e-06, + "loss": 0.5753, + "step": 6169 + }, + { + "epoch": 0.5257775884107371, + "grad_norm": 15.6875, + "learning_rate": 2.1064430395196867e-06, + "loss": 1.0124, + "step": 6170 + }, + { + "epoch": 0.525862803579037, + "grad_norm": 11.25, + "learning_rate": 2.106320635050556e-06, + "loss": 0.5928, + "step": 6171 + }, + { + "epoch": 0.525948018747337, + "grad_norm": 13.25, + "learning_rate": 2.106198215106851e-06, + "loss": 0.728, + "step": 6172 + }, + { + "epoch": 0.5260332339156369, + "grad_norm": 11.4375, + "learning_rate": 2.106075779690783e-06, + "loss": 0.3916, + "step": 6173 + }, + { + "epoch": 0.5261184490839369, + "grad_norm": 15.125, + "learning_rate": 2.1059533288045648e-06, + "loss": 0.9074, + "step": 6174 + }, + { + "epoch": 0.5262036642522369, + "grad_norm": 18.5, + "learning_rate": 2.105830862450409e-06, + "loss": 1.1213, + "step": 6175 + }, + { + "epoch": 0.5262888794205368, + "grad_norm": 16.25, + "learning_rate": 2.105708380630529e-06, + "loss": 0.7609, + "step": 6176 + }, + { + "epoch": 0.5263740945888368, + "grad_norm": 15.25, + "learning_rate": 2.105585883347138e-06, + "loss": 0.7624, + "step": 6177 + }, + { + "epoch": 0.5264593097571367, + "grad_norm": 25.625, + "learning_rate": 2.1054633706024504e-06, + "loss": 0.9504, + "step": 6178 + }, + { + "epoch": 0.5265445249254367, + "grad_norm": 14.9375, + "learning_rate": 2.1053408423986786e-06, + "loss": 0.9271, + "step": 6179 + }, + { + "epoch": 0.5266297400937366, + "grad_norm": 13.5, + "learning_rate": 2.1052182987380383e-06, + "loss": 0.8987, + "step": 6180 + }, + { + "epoch": 0.5267149552620366, + "grad_norm": 13.3125, + "learning_rate": 2.1050957396227427e-06, + "loss": 0.5143, + "step": 6181 + }, + { + "epoch": 0.5268001704303366, + "grad_norm": 12.5, + "learning_rate": 2.104973165055008e-06, + "loss": 0.4736, + "step": 6182 + }, + { + "epoch": 0.5268853855986365, + "grad_norm": 15.0, + "learning_rate": 2.104850575037048e-06, + "loss": 0.9557, + "step": 6183 + }, + { + "epoch": 0.5269706007669365, + "grad_norm": 11.5, + "learning_rate": 2.1047279695710786e-06, + "loss": 0.7713, + "step": 6184 + }, + { + "epoch": 0.5270558159352364, + "grad_norm": 16.25, + "learning_rate": 2.1046053486593153e-06, + "loss": 1.0148, + "step": 6185 + }, + { + "epoch": 0.5271410311035364, + "grad_norm": 13.1875, + "learning_rate": 2.1044827123039738e-06, + "loss": 0.6012, + "step": 6186 + }, + { + "epoch": 0.5272262462718363, + "grad_norm": 12.4375, + "learning_rate": 2.1043600605072708e-06, + "loss": 0.7012, + "step": 6187 + }, + { + "epoch": 0.5273114614401363, + "grad_norm": 18.0, + "learning_rate": 2.1042373932714217e-06, + "loss": 1.245, + "step": 6188 + }, + { + "epoch": 0.5273966766084363, + "grad_norm": 9.5625, + "learning_rate": 2.104114710598645e-06, + "loss": 0.4105, + "step": 6189 + }, + { + "epoch": 0.5274818917767362, + "grad_norm": 18.375, + "learning_rate": 2.103992012491156e-06, + "loss": 0.7026, + "step": 6190 + }, + { + "epoch": 0.5275671069450362, + "grad_norm": 17.5, + "learning_rate": 2.1038692989511724e-06, + "loss": 0.7081, + "step": 6191 + }, + { + "epoch": 0.5276523221133361, + "grad_norm": 16.875, + "learning_rate": 2.1037465699809117e-06, + "loss": 0.5986, + "step": 6192 + }, + { + "epoch": 0.5277375372816361, + "grad_norm": 18.625, + "learning_rate": 2.1036238255825926e-06, + "loss": 1.2049, + "step": 6193 + }, + { + "epoch": 0.527822752449936, + "grad_norm": 20.125, + "learning_rate": 2.1035010657584322e-06, + "loss": 0.7627, + "step": 6194 + }, + { + "epoch": 0.527907967618236, + "grad_norm": 15.375, + "learning_rate": 2.1033782905106493e-06, + "loss": 0.8761, + "step": 6195 + }, + { + "epoch": 0.527993182786536, + "grad_norm": 19.0, + "learning_rate": 2.1032554998414627e-06, + "loss": 0.9158, + "step": 6196 + }, + { + "epoch": 0.5280783979548359, + "grad_norm": 12.375, + "learning_rate": 2.103132693753091e-06, + "loss": 0.6206, + "step": 6197 + }, + { + "epoch": 0.5281636131231359, + "grad_norm": 9.6875, + "learning_rate": 2.103009872247754e-06, + "loss": 0.2813, + "step": 6198 + }, + { + "epoch": 0.5282488282914358, + "grad_norm": 13.625, + "learning_rate": 2.1028870353276708e-06, + "loss": 0.7611, + "step": 6199 + }, + { + "epoch": 0.5283340434597358, + "grad_norm": 13.4375, + "learning_rate": 2.1027641829950607e-06, + "loss": 0.6344, + "step": 6200 + }, + { + "epoch": 0.5284192586280357, + "grad_norm": 13.8125, + "learning_rate": 2.102641315252145e-06, + "loss": 0.5089, + "step": 6201 + }, + { + "epoch": 0.5285044737963357, + "grad_norm": 11.1875, + "learning_rate": 2.1025184321011427e-06, + "loss": 0.7062, + "step": 6202 + }, + { + "epoch": 0.5285896889646357, + "grad_norm": 12.4375, + "learning_rate": 2.102395533544275e-06, + "loss": 0.548, + "step": 6203 + }, + { + "epoch": 0.5286749041329356, + "grad_norm": 10.125, + "learning_rate": 2.1022726195837633e-06, + "loss": 0.5138, + "step": 6204 + }, + { + "epoch": 0.5287601193012356, + "grad_norm": 12.5, + "learning_rate": 2.1021496902218284e-06, + "loss": 0.557, + "step": 6205 + }, + { + "epoch": 0.5288453344695355, + "grad_norm": 15.6875, + "learning_rate": 2.1020267454606918e-06, + "loss": 0.7341, + "step": 6206 + }, + { + "epoch": 0.5289305496378356, + "grad_norm": 19.5, + "learning_rate": 2.101903785302575e-06, + "loss": 1.0213, + "step": 6207 + }, + { + "epoch": 0.5290157648061355, + "grad_norm": 15.375, + "learning_rate": 2.1017808097496996e-06, + "loss": 0.6237, + "step": 6208 + }, + { + "epoch": 0.5291009799744355, + "grad_norm": 13.25, + "learning_rate": 2.101657818804289e-06, + "loss": 0.6032, + "step": 6209 + }, + { + "epoch": 0.5291861951427355, + "grad_norm": 23.5, + "learning_rate": 2.1015348124685653e-06, + "loss": 1.1062, + "step": 6210 + }, + { + "epoch": 0.5292714103110354, + "grad_norm": 10.25, + "learning_rate": 2.101411790744751e-06, + "loss": 0.3556, + "step": 6211 + }, + { + "epoch": 0.5293566254793354, + "grad_norm": 43.5, + "learning_rate": 2.101288753635069e-06, + "loss": 0.6076, + "step": 6212 + }, + { + "epoch": 0.5294418406476353, + "grad_norm": 13.0625, + "learning_rate": 2.101165701141744e-06, + "loss": 0.9004, + "step": 6213 + }, + { + "epoch": 0.5295270558159353, + "grad_norm": 13.375, + "learning_rate": 2.1010426332669986e-06, + "loss": 0.7381, + "step": 6214 + }, + { + "epoch": 0.5296122709842352, + "grad_norm": 12.75, + "learning_rate": 2.100919550013057e-06, + "loss": 0.6725, + "step": 6215 + }, + { + "epoch": 0.5296974861525352, + "grad_norm": 15.0, + "learning_rate": 2.1007964513821434e-06, + "loss": 0.624, + "step": 6216 + }, + { + "epoch": 0.5297827013208352, + "grad_norm": 11.25, + "learning_rate": 2.1006733373764822e-06, + "loss": 0.4557, + "step": 6217 + }, + { + "epoch": 0.5298679164891351, + "grad_norm": 14.4375, + "learning_rate": 2.100550207998299e-06, + "loss": 1.0441, + "step": 6218 + }, + { + "epoch": 0.5299531316574351, + "grad_norm": 15.875, + "learning_rate": 2.1004270632498174e-06, + "loss": 0.9227, + "step": 6219 + }, + { + "epoch": 0.530038346825735, + "grad_norm": 13.75, + "learning_rate": 2.100303903133264e-06, + "loss": 0.6257, + "step": 6220 + }, + { + "epoch": 0.530123561994035, + "grad_norm": 18.875, + "learning_rate": 2.1001807276508642e-06, + "loss": 0.9663, + "step": 6221 + }, + { + "epoch": 0.5302087771623349, + "grad_norm": 13.875, + "learning_rate": 2.1000575368048436e-06, + "loss": 0.6762, + "step": 6222 + }, + { + "epoch": 0.5302939923306349, + "grad_norm": 13.9375, + "learning_rate": 2.0999343305974287e-06, + "loss": 1.0094, + "step": 6223 + }, + { + "epoch": 0.5303792074989349, + "grad_norm": 14.875, + "learning_rate": 2.0998111090308453e-06, + "loss": 0.675, + "step": 6224 + }, + { + "epoch": 0.5304644226672348, + "grad_norm": 19.0, + "learning_rate": 2.0996878721073212e-06, + "loss": 0.8701, + "step": 6225 + }, + { + "epoch": 0.5305496378355348, + "grad_norm": 9.875, + "learning_rate": 2.0995646198290825e-06, + "loss": 0.4679, + "step": 6226 + }, + { + "epoch": 0.5306348530038347, + "grad_norm": 17.25, + "learning_rate": 2.0994413521983567e-06, + "loss": 0.8929, + "step": 6227 + }, + { + "epoch": 0.5307200681721347, + "grad_norm": 13.6875, + "learning_rate": 2.0993180692173713e-06, + "loss": 0.7287, + "step": 6228 + }, + { + "epoch": 0.5308052833404346, + "grad_norm": 19.0, + "learning_rate": 2.099194770888355e-06, + "loss": 0.894, + "step": 6229 + }, + { + "epoch": 0.5308904985087346, + "grad_norm": 16.125, + "learning_rate": 2.099071457213535e-06, + "loss": 1.2041, + "step": 6230 + }, + { + "epoch": 0.5309757136770346, + "grad_norm": 18.125, + "learning_rate": 2.0989481281951395e-06, + "loss": 0.7065, + "step": 6231 + }, + { + "epoch": 0.5310609288453345, + "grad_norm": 18.25, + "learning_rate": 2.098824783835398e-06, + "loss": 0.7604, + "step": 6232 + }, + { + "epoch": 0.5311461440136345, + "grad_norm": 12.875, + "learning_rate": 2.0987014241365396e-06, + "loss": 0.727, + "step": 6233 + }, + { + "epoch": 0.5312313591819344, + "grad_norm": 11.5, + "learning_rate": 2.0985780491007923e-06, + "loss": 0.5016, + "step": 6234 + }, + { + "epoch": 0.5313165743502344, + "grad_norm": 17.0, + "learning_rate": 2.0984546587303863e-06, + "loss": 0.9232, + "step": 6235 + }, + { + "epoch": 0.5314017895185343, + "grad_norm": 12.4375, + "learning_rate": 2.098331253027552e-06, + "loss": 0.5898, + "step": 6236 + }, + { + "epoch": 0.5314870046868343, + "grad_norm": 14.75, + "learning_rate": 2.0982078319945185e-06, + "loss": 0.5249, + "step": 6237 + }, + { + "epoch": 0.5315722198551343, + "grad_norm": 15.6875, + "learning_rate": 2.098084395633517e-06, + "loss": 0.6885, + "step": 6238 + }, + { + "epoch": 0.5316574350234342, + "grad_norm": 10.6875, + "learning_rate": 2.097960943946777e-06, + "loss": 0.728, + "step": 6239 + }, + { + "epoch": 0.5317426501917342, + "grad_norm": 17.375, + "learning_rate": 2.09783747693653e-06, + "loss": 0.9854, + "step": 6240 + }, + { + "epoch": 0.5318278653600341, + "grad_norm": 15.5625, + "learning_rate": 2.097713994605008e-06, + "loss": 0.7116, + "step": 6241 + }, + { + "epoch": 0.5319130805283341, + "grad_norm": 21.125, + "learning_rate": 2.0975904969544413e-06, + "loss": 0.9684, + "step": 6242 + }, + { + "epoch": 0.531998295696634, + "grad_norm": 12.625, + "learning_rate": 2.0974669839870615e-06, + "loss": 0.666, + "step": 6243 + }, + { + "epoch": 0.532083510864934, + "grad_norm": 16.25, + "learning_rate": 2.0973434557051016e-06, + "loss": 0.6634, + "step": 6244 + }, + { + "epoch": 0.532168726033234, + "grad_norm": 13.75, + "learning_rate": 2.0972199121107934e-06, + "loss": 0.6413, + "step": 6245 + }, + { + "epoch": 0.5322539412015339, + "grad_norm": 14.4375, + "learning_rate": 2.0970963532063694e-06, + "loss": 0.729, + "step": 6246 + }, + { + "epoch": 0.5323391563698339, + "grad_norm": 9.625, + "learning_rate": 2.0969727789940623e-06, + "loss": 0.3331, + "step": 6247 + }, + { + "epoch": 0.5324243715381338, + "grad_norm": 14.125, + "learning_rate": 2.096849189476106e-06, + "loss": 0.757, + "step": 6248 + }, + { + "epoch": 0.5325095867064338, + "grad_norm": 12.5625, + "learning_rate": 2.0967255846547326e-06, + "loss": 0.686, + "step": 6249 + }, + { + "epoch": 0.5325948018747337, + "grad_norm": 16.25, + "learning_rate": 2.0966019645321767e-06, + "loss": 0.9446, + "step": 6250 + }, + { + "epoch": 0.5326800170430337, + "grad_norm": 14.3125, + "learning_rate": 2.0964783291106714e-06, + "loss": 0.616, + "step": 6251 + }, + { + "epoch": 0.5327652322113337, + "grad_norm": 18.5, + "learning_rate": 2.0963546783924525e-06, + "loss": 0.9818, + "step": 6252 + }, + { + "epoch": 0.5328504473796336, + "grad_norm": 12.9375, + "learning_rate": 2.096231012379753e-06, + "loss": 0.5134, + "step": 6253 + }, + { + "epoch": 0.5329356625479336, + "grad_norm": 11.1875, + "learning_rate": 2.096107331074808e-06, + "loss": 0.4026, + "step": 6254 + }, + { + "epoch": 0.5330208777162335, + "grad_norm": 17.25, + "learning_rate": 2.0959836344798527e-06, + "loss": 0.8716, + "step": 6255 + }, + { + "epoch": 0.5331060928845335, + "grad_norm": 14.875, + "learning_rate": 2.095859922597123e-06, + "loss": 0.4164, + "step": 6256 + }, + { + "epoch": 0.5331913080528334, + "grad_norm": 16.125, + "learning_rate": 2.095736195428853e-06, + "loss": 0.8771, + "step": 6257 + }, + { + "epoch": 0.5332765232211334, + "grad_norm": 13.625, + "learning_rate": 2.0956124529772805e-06, + "loss": 0.8138, + "step": 6258 + }, + { + "epoch": 0.5333617383894333, + "grad_norm": 16.625, + "learning_rate": 2.09548869524464e-06, + "loss": 1.078, + "step": 6259 + }, + { + "epoch": 0.5334469535577333, + "grad_norm": 18.125, + "learning_rate": 2.0953649222331686e-06, + "loss": 1.2338, + "step": 6260 + }, + { + "epoch": 0.5335321687260333, + "grad_norm": 12.9375, + "learning_rate": 2.0952411339451027e-06, + "loss": 0.4029, + "step": 6261 + }, + { + "epoch": 0.5336173838943332, + "grad_norm": 13.6875, + "learning_rate": 2.0951173303826806e-06, + "loss": 0.5247, + "step": 6262 + }, + { + "epoch": 0.5337025990626332, + "grad_norm": 11.5, + "learning_rate": 2.0949935115481376e-06, + "loss": 0.5547, + "step": 6263 + }, + { + "epoch": 0.5337878142309331, + "grad_norm": 24.875, + "learning_rate": 2.0948696774437123e-06, + "loss": 0.9971, + "step": 6264 + }, + { + "epoch": 0.5338730293992331, + "grad_norm": 13.125, + "learning_rate": 2.0947458280716427e-06, + "loss": 0.5835, + "step": 6265 + }, + { + "epoch": 0.533958244567533, + "grad_norm": 13.75, + "learning_rate": 2.094621963434166e-06, + "loss": 0.8904, + "step": 6266 + }, + { + "epoch": 0.534043459735833, + "grad_norm": 17.0, + "learning_rate": 2.0944980835335214e-06, + "loss": 0.9664, + "step": 6267 + }, + { + "epoch": 0.534128674904133, + "grad_norm": 21.75, + "learning_rate": 2.094374188371947e-06, + "loss": 1.1452, + "step": 6268 + }, + { + "epoch": 0.5342138900724329, + "grad_norm": 14.1875, + "learning_rate": 2.0942502779516825e-06, + "loss": 0.9308, + "step": 6269 + }, + { + "epoch": 0.5342991052407329, + "grad_norm": 17.75, + "learning_rate": 2.0941263522749662e-06, + "loss": 0.4857, + "step": 6270 + }, + { + "epoch": 0.5343843204090328, + "grad_norm": 13.8125, + "learning_rate": 2.094002411344038e-06, + "loss": 0.4573, + "step": 6271 + }, + { + "epoch": 0.5344695355773328, + "grad_norm": 11.9375, + "learning_rate": 2.093878455161137e-06, + "loss": 0.644, + "step": 6272 + }, + { + "epoch": 0.5345547507456327, + "grad_norm": 14.625, + "learning_rate": 2.093754483728504e-06, + "loss": 0.8602, + "step": 6273 + }, + { + "epoch": 0.5346399659139327, + "grad_norm": 18.625, + "learning_rate": 2.093630497048379e-06, + "loss": 0.9489, + "step": 6274 + }, + { + "epoch": 0.5347251810822327, + "grad_norm": 10.0, + "learning_rate": 2.0935064951230026e-06, + "loss": 0.3423, + "step": 6275 + }, + { + "epoch": 0.5348103962505326, + "grad_norm": 12.375, + "learning_rate": 2.093382477954616e-06, + "loss": 0.6262, + "step": 6276 + }, + { + "epoch": 0.5348956114188326, + "grad_norm": 16.0, + "learning_rate": 2.0932584455454594e-06, + "loss": 0.2831, + "step": 6277 + }, + { + "epoch": 0.5349808265871325, + "grad_norm": 13.25, + "learning_rate": 2.093134397897775e-06, + "loss": 0.5382, + "step": 6278 + }, + { + "epoch": 0.5350660417554325, + "grad_norm": 14.3125, + "learning_rate": 2.093010335013804e-06, + "loss": 0.6745, + "step": 6279 + }, + { + "epoch": 0.5351512569237324, + "grad_norm": 30.25, + "learning_rate": 2.0928862568957886e-06, + "loss": 0.8815, + "step": 6280 + }, + { + "epoch": 0.5352364720920324, + "grad_norm": 14.875, + "learning_rate": 2.092762163545971e-06, + "loss": 1.1011, + "step": 6281 + }, + { + "epoch": 0.5353216872603324, + "grad_norm": 14.625, + "learning_rate": 2.092638054966594e-06, + "loss": 0.5702, + "step": 6282 + }, + { + "epoch": 0.5354069024286323, + "grad_norm": 26.75, + "learning_rate": 2.0925139311598996e-06, + "loss": 0.805, + "step": 6283 + }, + { + "epoch": 0.5354921175969323, + "grad_norm": 15.375, + "learning_rate": 2.092389792128131e-06, + "loss": 0.7906, + "step": 6284 + }, + { + "epoch": 0.5355773327652322, + "grad_norm": 20.75, + "learning_rate": 2.0922656378735324e-06, + "loss": 0.9157, + "step": 6285 + }, + { + "epoch": 0.5356625479335322, + "grad_norm": 19.375, + "learning_rate": 2.0921414683983464e-06, + "loss": 0.4856, + "step": 6286 + }, + { + "epoch": 0.5357477631018321, + "grad_norm": 10.9375, + "learning_rate": 2.092017283704817e-06, + "loss": 0.4744, + "step": 6287 + }, + { + "epoch": 0.5358329782701321, + "grad_norm": 13.375, + "learning_rate": 2.091893083795189e-06, + "loss": 0.4649, + "step": 6288 + }, + { + "epoch": 0.5359181934384321, + "grad_norm": 13.4375, + "learning_rate": 2.091768868671706e-06, + "loss": 0.3992, + "step": 6289 + }, + { + "epoch": 0.536003408606732, + "grad_norm": 24.625, + "learning_rate": 2.091644638336614e-06, + "loss": 1.0316, + "step": 6290 + }, + { + "epoch": 0.536088623775032, + "grad_norm": 15.75, + "learning_rate": 2.0915203927921563e-06, + "loss": 0.856, + "step": 6291 + }, + { + "epoch": 0.5361738389433319, + "grad_norm": 17.25, + "learning_rate": 2.0913961320405783e-06, + "loss": 1.061, + "step": 6292 + }, + { + "epoch": 0.5362590541116319, + "grad_norm": 14.0625, + "learning_rate": 2.091271856084127e-06, + "loss": 0.5562, + "step": 6293 + }, + { + "epoch": 0.5363442692799318, + "grad_norm": 22.125, + "learning_rate": 2.091147564925047e-06, + "loss": 1.3397, + "step": 6294 + }, + { + "epoch": 0.5364294844482318, + "grad_norm": 14.4375, + "learning_rate": 2.091023258565585e-06, + "loss": 0.7106, + "step": 6295 + }, + { + "epoch": 0.5365146996165318, + "grad_norm": 22.0, + "learning_rate": 2.0908989370079867e-06, + "loss": 0.6138, + "step": 6296 + }, + { + "epoch": 0.5365999147848317, + "grad_norm": 29.625, + "learning_rate": 2.090774600254499e-06, + "loss": 1.3506, + "step": 6297 + }, + { + "epoch": 0.5366851299531317, + "grad_norm": 12.625, + "learning_rate": 2.090650248307369e-06, + "loss": 0.6136, + "step": 6298 + }, + { + "epoch": 0.5367703451214316, + "grad_norm": 12.25, + "learning_rate": 2.0905258811688432e-06, + "loss": 0.3978, + "step": 6299 + }, + { + "epoch": 0.5368555602897316, + "grad_norm": 15.1875, + "learning_rate": 2.09040149884117e-06, + "loss": 0.7418, + "step": 6300 + }, + { + "epoch": 0.5369407754580315, + "grad_norm": 12.5625, + "learning_rate": 2.0902771013265964e-06, + "loss": 0.5526, + "step": 6301 + }, + { + "epoch": 0.5370259906263315, + "grad_norm": 12.1875, + "learning_rate": 2.090152688627371e-06, + "loss": 0.6491, + "step": 6302 + }, + { + "epoch": 0.5371112057946315, + "grad_norm": 14.0, + "learning_rate": 2.0900282607457413e-06, + "loss": 0.7835, + "step": 6303 + }, + { + "epoch": 0.5371964209629314, + "grad_norm": 13.0, + "learning_rate": 2.089903817683956e-06, + "loss": 0.3832, + "step": 6304 + }, + { + "epoch": 0.5372816361312314, + "grad_norm": 20.125, + "learning_rate": 2.089779359444264e-06, + "loss": 0.9013, + "step": 6305 + }, + { + "epoch": 0.5373668512995313, + "grad_norm": 20.5, + "learning_rate": 2.0896548860289153e-06, + "loss": 0.9356, + "step": 6306 + }, + { + "epoch": 0.5374520664678313, + "grad_norm": 14.4375, + "learning_rate": 2.0895303974401582e-06, + "loss": 0.5911, + "step": 6307 + }, + { + "epoch": 0.5375372816361312, + "grad_norm": 22.625, + "learning_rate": 2.0894058936802424e-06, + "loss": 1.011, + "step": 6308 + }, + { + "epoch": 0.5376224968044312, + "grad_norm": 15.125, + "learning_rate": 2.089281374751418e-06, + "loss": 0.855, + "step": 6309 + }, + { + "epoch": 0.5377077119727312, + "grad_norm": 14.6875, + "learning_rate": 2.0891568406559355e-06, + "loss": 0.8929, + "step": 6310 + }, + { + "epoch": 0.5377929271410311, + "grad_norm": 11.5, + "learning_rate": 2.089032291396045e-06, + "loss": 0.3536, + "step": 6311 + }, + { + "epoch": 0.5378781423093311, + "grad_norm": 14.6875, + "learning_rate": 2.0889077269739967e-06, + "loss": 0.6752, + "step": 6312 + }, + { + "epoch": 0.537963357477631, + "grad_norm": 20.375, + "learning_rate": 2.0887831473920426e-06, + "loss": 1.3051, + "step": 6313 + }, + { + "epoch": 0.538048572645931, + "grad_norm": 14.375, + "learning_rate": 2.0886585526524335e-06, + "loss": 0.6454, + "step": 6314 + }, + { + "epoch": 0.5381337878142309, + "grad_norm": 13.0625, + "learning_rate": 2.088533942757421e-06, + "loss": 0.4837, + "step": 6315 + }, + { + "epoch": 0.5382190029825309, + "grad_norm": 22.75, + "learning_rate": 2.0884093177092575e-06, + "loss": 0.9437, + "step": 6316 + }, + { + "epoch": 0.5383042181508308, + "grad_norm": 29.75, + "learning_rate": 2.088284677510194e-06, + "loss": 1.1829, + "step": 6317 + }, + { + "epoch": 0.5383894333191308, + "grad_norm": 21.625, + "learning_rate": 2.088160022162483e-06, + "loss": 0.9178, + "step": 6318 + }, + { + "epoch": 0.5384746484874308, + "grad_norm": 11.9375, + "learning_rate": 2.0880353516683782e-06, + "loss": 0.5022, + "step": 6319 + }, + { + "epoch": 0.5385598636557307, + "grad_norm": 15.4375, + "learning_rate": 2.087910666030132e-06, + "loss": 0.7753, + "step": 6320 + }, + { + "epoch": 0.5386450788240307, + "grad_norm": 14.5, + "learning_rate": 2.087785965249997e-06, + "loss": 0.7404, + "step": 6321 + }, + { + "epoch": 0.5387302939923306, + "grad_norm": 15.625, + "learning_rate": 2.087661249330227e-06, + "loss": 0.6213, + "step": 6322 + }, + { + "epoch": 0.5388155091606306, + "grad_norm": 14.5, + "learning_rate": 2.0875365182730767e-06, + "loss": 0.596, + "step": 6323 + }, + { + "epoch": 0.5389007243289305, + "grad_norm": 18.25, + "learning_rate": 2.087411772080799e-06, + "loss": 1.1188, + "step": 6324 + }, + { + "epoch": 0.5389859394972305, + "grad_norm": 17.0, + "learning_rate": 2.0872870107556483e-06, + "loss": 0.776, + "step": 6325 + }, + { + "epoch": 0.5390711546655305, + "grad_norm": 22.75, + "learning_rate": 2.0871622342998794e-06, + "loss": 1.0287, + "step": 6326 + }, + { + "epoch": 0.5391563698338304, + "grad_norm": 10.625, + "learning_rate": 2.087037442715747e-06, + "loss": 0.5038, + "step": 6327 + }, + { + "epoch": 0.5392415850021304, + "grad_norm": 15.375, + "learning_rate": 2.0869126360055065e-06, + "loss": 0.9662, + "step": 6328 + }, + { + "epoch": 0.5393268001704303, + "grad_norm": 10.75, + "learning_rate": 2.086787814171413e-06, + "loss": 0.4304, + "step": 6329 + }, + { + "epoch": 0.5394120153387303, + "grad_norm": 13.3125, + "learning_rate": 2.086662977215722e-06, + "loss": 0.6223, + "step": 6330 + }, + { + "epoch": 0.5394972305070302, + "grad_norm": 9.1875, + "learning_rate": 2.08653812514069e-06, + "loss": 0.3308, + "step": 6331 + }, + { + "epoch": 0.5395824456753302, + "grad_norm": 15.625, + "learning_rate": 2.086413257948573e-06, + "loss": 0.7595, + "step": 6332 + }, + { + "epoch": 0.5396676608436302, + "grad_norm": 14.5, + "learning_rate": 2.086288375641627e-06, + "loss": 0.8933, + "step": 6333 + }, + { + "epoch": 0.5397528760119301, + "grad_norm": 12.5625, + "learning_rate": 2.0861634782221093e-06, + "loss": 0.7803, + "step": 6334 + }, + { + "epoch": 0.5398380911802301, + "grad_norm": 11.8125, + "learning_rate": 2.0860385656922767e-06, + "loss": 0.4623, + "step": 6335 + }, + { + "epoch": 0.53992330634853, + "grad_norm": 12.6875, + "learning_rate": 2.0859136380543866e-06, + "loss": 0.4133, + "step": 6336 + }, + { + "epoch": 0.54000852151683, + "grad_norm": 13.875, + "learning_rate": 2.0857886953106965e-06, + "loss": 0.7697, + "step": 6337 + }, + { + "epoch": 0.5400937366851299, + "grad_norm": 13.875, + "learning_rate": 2.085663737463464e-06, + "loss": 0.7458, + "step": 6338 + }, + { + "epoch": 0.5401789518534299, + "grad_norm": 13.4375, + "learning_rate": 2.0855387645149475e-06, + "loss": 0.5537, + "step": 6339 + }, + { + "epoch": 0.5402641670217299, + "grad_norm": 13.5625, + "learning_rate": 2.0854137764674055e-06, + "loss": 0.4072, + "step": 6340 + }, + { + "epoch": 0.5403493821900298, + "grad_norm": 11.625, + "learning_rate": 2.085288773323096e-06, + "loss": 0.5515, + "step": 6341 + }, + { + "epoch": 0.5404345973583298, + "grad_norm": 16.375, + "learning_rate": 2.085163755084279e-06, + "loss": 0.6574, + "step": 6342 + }, + { + "epoch": 0.5405198125266297, + "grad_norm": 17.25, + "learning_rate": 2.0850387217532127e-06, + "loss": 0.8315, + "step": 6343 + }, + { + "epoch": 0.5406050276949297, + "grad_norm": 21.125, + "learning_rate": 2.0849136733321573e-06, + "loss": 1.0187, + "step": 6344 + }, + { + "epoch": 0.5406902428632296, + "grad_norm": 15.6875, + "learning_rate": 2.0847886098233724e-06, + "loss": 0.7853, + "step": 6345 + }, + { + "epoch": 0.5407754580315296, + "grad_norm": 12.0625, + "learning_rate": 2.084663531229117e-06, + "loss": 0.6363, + "step": 6346 + }, + { + "epoch": 0.5408606731998296, + "grad_norm": 20.75, + "learning_rate": 2.084538437551653e-06, + "loss": 1.0188, + "step": 6347 + }, + { + "epoch": 0.5409458883681295, + "grad_norm": 13.1875, + "learning_rate": 2.0844133287932404e-06, + "loss": 0.7096, + "step": 6348 + }, + { + "epoch": 0.5410311035364295, + "grad_norm": 11.6875, + "learning_rate": 2.0842882049561398e-06, + "loss": 0.6152, + "step": 6349 + }, + { + "epoch": 0.5411163187047294, + "grad_norm": 14.375, + "learning_rate": 2.0841630660426116e-06, + "loss": 0.9194, + "step": 6350 + }, + { + "epoch": 0.5412015338730294, + "grad_norm": 13.4375, + "learning_rate": 2.084037912054919e-06, + "loss": 0.887, + "step": 6351 + }, + { + "epoch": 0.5412867490413293, + "grad_norm": 11.8125, + "learning_rate": 2.083912742995322e-06, + "loss": 0.6242, + "step": 6352 + }, + { + "epoch": 0.5413719642096293, + "grad_norm": 12.0625, + "learning_rate": 2.0837875588660834e-06, + "loss": 0.7515, + "step": 6353 + }, + { + "epoch": 0.5414571793779293, + "grad_norm": 14.125, + "learning_rate": 2.083662359669465e-06, + "loss": 0.6306, + "step": 6354 + }, + { + "epoch": 0.5415423945462292, + "grad_norm": 11.875, + "learning_rate": 2.08353714540773e-06, + "loss": 0.5489, + "step": 6355 + }, + { + "epoch": 0.5416276097145292, + "grad_norm": 13.0625, + "learning_rate": 2.08341191608314e-06, + "loss": 0.6301, + "step": 6356 + }, + { + "epoch": 0.5417128248828291, + "grad_norm": 14.625, + "learning_rate": 2.083286671697959e-06, + "loss": 1.1312, + "step": 6357 + }, + { + "epoch": 0.5417980400511291, + "grad_norm": 15.125, + "learning_rate": 2.0831614122544496e-06, + "loss": 0.8191, + "step": 6358 + }, + { + "epoch": 0.541883255219429, + "grad_norm": 17.125, + "learning_rate": 2.0830361377548757e-06, + "loss": 0.7818, + "step": 6359 + }, + { + "epoch": 0.541968470387729, + "grad_norm": 10.4375, + "learning_rate": 2.0829108482015015e-06, + "loss": 0.2524, + "step": 6360 + }, + { + "epoch": 0.542053685556029, + "grad_norm": 25.875, + "learning_rate": 2.0827855435965907e-06, + "loss": 1.318, + "step": 6361 + }, + { + "epoch": 0.5421389007243289, + "grad_norm": 17.75, + "learning_rate": 2.082660223942407e-06, + "loss": 0.9532, + "step": 6362 + }, + { + "epoch": 0.5422241158926289, + "grad_norm": 14.1875, + "learning_rate": 2.082534889241216e-06, + "loss": 0.8932, + "step": 6363 + }, + { + "epoch": 0.5423093310609288, + "grad_norm": 15.5625, + "learning_rate": 2.0824095394952828e-06, + "loss": 0.8502, + "step": 6364 + }, + { + "epoch": 0.5423945462292288, + "grad_norm": 12.1875, + "learning_rate": 2.082284174706872e-06, + "loss": 0.6244, + "step": 6365 + }, + { + "epoch": 0.5424797613975287, + "grad_norm": 18.0, + "learning_rate": 2.082158794878249e-06, + "loss": 0.9709, + "step": 6366 + }, + { + "epoch": 0.5425649765658287, + "grad_norm": 19.25, + "learning_rate": 2.0820334000116803e-06, + "loss": 1.0619, + "step": 6367 + }, + { + "epoch": 0.5426501917341287, + "grad_norm": 13.75, + "learning_rate": 2.0819079901094306e-06, + "loss": 0.7625, + "step": 6368 + }, + { + "epoch": 0.5427354069024286, + "grad_norm": 13.6875, + "learning_rate": 2.0817825651737674e-06, + "loss": 0.9346, + "step": 6369 + }, + { + "epoch": 0.5428206220707286, + "grad_norm": 15.0625, + "learning_rate": 2.0816571252069565e-06, + "loss": 0.6954, + "step": 6370 + }, + { + "epoch": 0.5429058372390285, + "grad_norm": 14.75, + "learning_rate": 2.081531670211266e-06, + "loss": 0.8664, + "step": 6371 + }, + { + "epoch": 0.5429910524073285, + "grad_norm": 15.875, + "learning_rate": 2.0814062001889613e-06, + "loss": 0.775, + "step": 6372 + }, + { + "epoch": 0.5430762675756284, + "grad_norm": 14.3125, + "learning_rate": 2.0812807151423103e-06, + "loss": 0.7374, + "step": 6373 + }, + { + "epoch": 0.5431614827439284, + "grad_norm": 17.5, + "learning_rate": 2.0811552150735813e-06, + "loss": 0.9329, + "step": 6374 + }, + { + "epoch": 0.5432466979122283, + "grad_norm": 14.3125, + "learning_rate": 2.081029699985042e-06, + "loss": 0.8339, + "step": 6375 + }, + { + "epoch": 0.5433319130805283, + "grad_norm": 18.125, + "learning_rate": 2.0809041698789598e-06, + "loss": 0.7687, + "step": 6376 + }, + { + "epoch": 0.5434171282488283, + "grad_norm": 14.1875, + "learning_rate": 2.080778624757604e-06, + "loss": 0.7427, + "step": 6377 + }, + { + "epoch": 0.5435023434171282, + "grad_norm": 15.5625, + "learning_rate": 2.080653064623243e-06, + "loss": 0.908, + "step": 6378 + }, + { + "epoch": 0.5435875585854282, + "grad_norm": 17.875, + "learning_rate": 2.0805274894781458e-06, + "loss": 0.9505, + "step": 6379 + }, + { + "epoch": 0.5436727737537281, + "grad_norm": 10.1875, + "learning_rate": 2.0804018993245816e-06, + "loss": 0.307, + "step": 6380 + }, + { + "epoch": 0.5437579889220281, + "grad_norm": 15.5625, + "learning_rate": 2.0802762941648203e-06, + "loss": 1.0888, + "step": 6381 + }, + { + "epoch": 0.543843204090328, + "grad_norm": 14.6875, + "learning_rate": 2.0801506740011312e-06, + "loss": 0.9752, + "step": 6382 + }, + { + "epoch": 0.543928419258628, + "grad_norm": 14.625, + "learning_rate": 2.0800250388357847e-06, + "loss": 0.7256, + "step": 6383 + }, + { + "epoch": 0.544013634426928, + "grad_norm": 17.375, + "learning_rate": 2.079899388671051e-06, + "loss": 1.1044, + "step": 6384 + }, + { + "epoch": 0.5440988495952279, + "grad_norm": 12.6875, + "learning_rate": 2.0797737235092015e-06, + "loss": 0.6349, + "step": 6385 + }, + { + "epoch": 0.5441840647635279, + "grad_norm": 16.0, + "learning_rate": 2.0796480433525057e-06, + "loss": 0.7979, + "step": 6386 + }, + { + "epoch": 0.5442692799318278, + "grad_norm": 66.5, + "learning_rate": 2.0795223482032357e-06, + "loss": 1.5932, + "step": 6387 + }, + { + "epoch": 0.5443544951001278, + "grad_norm": 12.625, + "learning_rate": 2.079396638063663e-06, + "loss": 0.7549, + "step": 6388 + }, + { + "epoch": 0.5444397102684277, + "grad_norm": 23.875, + "learning_rate": 2.0792709129360583e-06, + "loss": 1.1931, + "step": 6389 + }, + { + "epoch": 0.5445249254367277, + "grad_norm": 13.0625, + "learning_rate": 2.079145172822695e-06, + "loss": 0.7229, + "step": 6390 + }, + { + "epoch": 0.5446101406050277, + "grad_norm": 16.375, + "learning_rate": 2.0790194177258436e-06, + "loss": 0.7531, + "step": 6391 + }, + { + "epoch": 0.5446953557733276, + "grad_norm": 14.0625, + "learning_rate": 2.078893647647779e-06, + "loss": 0.759, + "step": 6392 + }, + { + "epoch": 0.5447805709416276, + "grad_norm": 13.8125, + "learning_rate": 2.078767862590772e-06, + "loss": 0.6109, + "step": 6393 + }, + { + "epoch": 0.5448657861099275, + "grad_norm": 14.1875, + "learning_rate": 2.0786420625570963e-06, + "loss": 0.9057, + "step": 6394 + }, + { + "epoch": 0.5449510012782275, + "grad_norm": 17.375, + "learning_rate": 2.0785162475490254e-06, + "loss": 0.8169, + "step": 6395 + }, + { + "epoch": 0.5450362164465274, + "grad_norm": 19.875, + "learning_rate": 2.078390417568833e-06, + "loss": 0.3474, + "step": 6396 + }, + { + "epoch": 0.5451214316148274, + "grad_norm": 25.375, + "learning_rate": 2.0782645726187926e-06, + "loss": 0.8701, + "step": 6397 + }, + { + "epoch": 0.5452066467831274, + "grad_norm": 13.9375, + "learning_rate": 2.078138712701178e-06, + "loss": 0.5955, + "step": 6398 + }, + { + "epoch": 0.5452918619514273, + "grad_norm": 13.75, + "learning_rate": 2.0780128378182645e-06, + "loss": 0.6416, + "step": 6399 + }, + { + "epoch": 0.5453770771197273, + "grad_norm": 14.375, + "learning_rate": 2.0778869479723267e-06, + "loss": 0.9216, + "step": 6400 + }, + { + "epoch": 0.5454622922880272, + "grad_norm": 16.25, + "learning_rate": 2.077761043165639e-06, + "loss": 0.8919, + "step": 6401 + }, + { + "epoch": 0.5455475074563272, + "grad_norm": 13.75, + "learning_rate": 2.077635123400477e-06, + "loss": 0.7583, + "step": 6402 + }, + { + "epoch": 0.5456327226246271, + "grad_norm": 14.6875, + "learning_rate": 2.0775091886791158e-06, + "loss": 0.7405, + "step": 6403 + }, + { + "epoch": 0.5457179377929271, + "grad_norm": 19.5, + "learning_rate": 2.0773832390038316e-06, + "loss": 0.8922, + "step": 6404 + }, + { + "epoch": 0.5458031529612271, + "grad_norm": 21.375, + "learning_rate": 2.0772572743769002e-06, + "loss": 1.1329, + "step": 6405 + }, + { + "epoch": 0.545888368129527, + "grad_norm": 15.25, + "learning_rate": 2.077131294800598e-06, + "loss": 0.904, + "step": 6406 + }, + { + "epoch": 0.545973583297827, + "grad_norm": 18.0, + "learning_rate": 2.0770053002772016e-06, + "loss": 0.8525, + "step": 6407 + }, + { + "epoch": 0.5460587984661269, + "grad_norm": 14.8125, + "learning_rate": 2.0768792908089876e-06, + "loss": 0.5698, + "step": 6408 + }, + { + "epoch": 0.5461440136344269, + "grad_norm": 18.25, + "learning_rate": 2.0767532663982332e-06, + "loss": 0.9145, + "step": 6409 + }, + { + "epoch": 0.5462292288027268, + "grad_norm": 14.9375, + "learning_rate": 2.0766272270472165e-06, + "loss": 0.9785, + "step": 6410 + }, + { + "epoch": 0.5463144439710268, + "grad_norm": 13.875, + "learning_rate": 2.076501172758214e-06, + "loss": 0.7995, + "step": 6411 + }, + { + "epoch": 0.5463996591393268, + "grad_norm": 18.0, + "learning_rate": 2.0763751035335047e-06, + "loss": 0.9431, + "step": 6412 + }, + { + "epoch": 0.5464848743076267, + "grad_norm": 17.375, + "learning_rate": 2.076249019375366e-06, + "loss": 0.8848, + "step": 6413 + }, + { + "epoch": 0.5465700894759267, + "grad_norm": 12.8125, + "learning_rate": 2.0761229202860765e-06, + "loss": 0.7104, + "step": 6414 + }, + { + "epoch": 0.5466553046442266, + "grad_norm": 20.375, + "learning_rate": 2.0759968062679154e-06, + "loss": 0.9082, + "step": 6415 + }, + { + "epoch": 0.5467405198125266, + "grad_norm": 19.25, + "learning_rate": 2.0758706773231614e-06, + "loss": 0.8975, + "step": 6416 + }, + { + "epoch": 0.5468257349808265, + "grad_norm": 19.0, + "learning_rate": 2.075744533454094e-06, + "loss": 0.9932, + "step": 6417 + }, + { + "epoch": 0.5469109501491265, + "grad_norm": 15.0625, + "learning_rate": 2.075618374662992e-06, + "loss": 0.8165, + "step": 6418 + }, + { + "epoch": 0.5469961653174265, + "grad_norm": 10.5625, + "learning_rate": 2.0754922009521357e-06, + "loss": 0.4637, + "step": 6419 + }, + { + "epoch": 0.5470813804857264, + "grad_norm": 31.5, + "learning_rate": 2.075366012323806e-06, + "loss": 0.9428, + "step": 6420 + }, + { + "epoch": 0.5471665956540264, + "grad_norm": 29.5, + "learning_rate": 2.0752398087802817e-06, + "loss": 0.8586, + "step": 6421 + }, + { + "epoch": 0.5472518108223263, + "grad_norm": 17.5, + "learning_rate": 2.0751135903238446e-06, + "loss": 0.2941, + "step": 6422 + }, + { + "epoch": 0.5473370259906263, + "grad_norm": 14.5625, + "learning_rate": 2.0749873569567753e-06, + "loss": 0.6178, + "step": 6423 + }, + { + "epoch": 0.5474222411589262, + "grad_norm": 12.1875, + "learning_rate": 2.0748611086813548e-06, + "loss": 0.5331, + "step": 6424 + }, + { + "epoch": 0.5475074563272262, + "grad_norm": 15.125, + "learning_rate": 2.074734845499864e-06, + "loss": 0.7394, + "step": 6425 + }, + { + "epoch": 0.5475926714955262, + "grad_norm": 15.25, + "learning_rate": 2.0746085674145856e-06, + "loss": 0.8424, + "step": 6426 + }, + { + "epoch": 0.5476778866638261, + "grad_norm": 20.0, + "learning_rate": 2.0744822744278017e-06, + "loss": 1.0727, + "step": 6427 + }, + { + "epoch": 0.5477631018321261, + "grad_norm": 19.125, + "learning_rate": 2.0743559665417933e-06, + "loss": 0.4943, + "step": 6428 + }, + { + "epoch": 0.547848317000426, + "grad_norm": 16.5, + "learning_rate": 2.074229643758844e-06, + "loss": 0.9638, + "step": 6429 + }, + { + "epoch": 0.547933532168726, + "grad_norm": 14.875, + "learning_rate": 2.074103306081236e-06, + "loss": 0.9248, + "step": 6430 + }, + { + "epoch": 0.5480187473370259, + "grad_norm": 13.375, + "learning_rate": 2.0739769535112527e-06, + "loss": 0.6848, + "step": 6431 + }, + { + "epoch": 0.548103962505326, + "grad_norm": 16.125, + "learning_rate": 2.0738505860511776e-06, + "loss": 0.7043, + "step": 6432 + }, + { + "epoch": 0.548189177673626, + "grad_norm": 16.625, + "learning_rate": 2.0737242037032935e-06, + "loss": 0.7008, + "step": 6433 + }, + { + "epoch": 0.5482743928419259, + "grad_norm": 12.125, + "learning_rate": 2.0735978064698846e-06, + "loss": 0.6398, + "step": 6434 + }, + { + "epoch": 0.5483596080102259, + "grad_norm": 14.75, + "learning_rate": 2.0734713943532355e-06, + "loss": 0.8238, + "step": 6435 + }, + { + "epoch": 0.5484448231785258, + "grad_norm": 15.0625, + "learning_rate": 2.0733449673556305e-06, + "loss": 0.7854, + "step": 6436 + }, + { + "epoch": 0.5485300383468258, + "grad_norm": 13.9375, + "learning_rate": 2.0732185254793534e-06, + "loss": 1.0776, + "step": 6437 + }, + { + "epoch": 0.5486152535151257, + "grad_norm": 15.3125, + "learning_rate": 2.07309206872669e-06, + "loss": 0.8542, + "step": 6438 + }, + { + "epoch": 0.5487004686834257, + "grad_norm": 10.875, + "learning_rate": 2.0729655970999252e-06, + "loss": 0.4787, + "step": 6439 + }, + { + "epoch": 0.5487856838517257, + "grad_norm": 16.25, + "learning_rate": 2.0728391106013448e-06, + "loss": 0.6915, + "step": 6440 + }, + { + "epoch": 0.5488708990200256, + "grad_norm": 17.125, + "learning_rate": 2.072712609233234e-06, + "loss": 0.864, + "step": 6441 + }, + { + "epoch": 0.5489561141883256, + "grad_norm": 19.125, + "learning_rate": 2.0725860929978787e-06, + "loss": 0.6895, + "step": 6442 + }, + { + "epoch": 0.5490413293566255, + "grad_norm": 12.625, + "learning_rate": 2.072459561897566e-06, + "loss": 0.5052, + "step": 6443 + }, + { + "epoch": 0.5491265445249255, + "grad_norm": 15.25, + "learning_rate": 2.0723330159345815e-06, + "loss": 0.9757, + "step": 6444 + }, + { + "epoch": 0.5492117596932254, + "grad_norm": 16.25, + "learning_rate": 2.072206455111213e-06, + "loss": 0.8129, + "step": 6445 + }, + { + "epoch": 0.5492969748615254, + "grad_norm": 17.5, + "learning_rate": 2.072079879429747e-06, + "loss": 0.8724, + "step": 6446 + }, + { + "epoch": 0.5493821900298254, + "grad_norm": 12.6875, + "learning_rate": 2.0719532888924705e-06, + "loss": 0.5526, + "step": 6447 + }, + { + "epoch": 0.5494674051981253, + "grad_norm": 16.5, + "learning_rate": 2.071826683501672e-06, + "loss": 0.9774, + "step": 6448 + }, + { + "epoch": 0.5495526203664253, + "grad_norm": 10.3125, + "learning_rate": 2.071700063259639e-06, + "loss": 0.4172, + "step": 6449 + }, + { + "epoch": 0.5496378355347252, + "grad_norm": 15.1875, + "learning_rate": 2.071573428168659e-06, + "loss": 0.3816, + "step": 6450 + }, + { + "epoch": 0.5497230507030252, + "grad_norm": 16.0, + "learning_rate": 2.071446778231021e-06, + "loss": 0.7898, + "step": 6451 + }, + { + "epoch": 0.5498082658713251, + "grad_norm": 11.25, + "learning_rate": 2.0713201134490143e-06, + "loss": 0.5958, + "step": 6452 + }, + { + "epoch": 0.5498934810396251, + "grad_norm": 24.375, + "learning_rate": 2.0711934338249266e-06, + "loss": 1.0397, + "step": 6453 + }, + { + "epoch": 0.549978696207925, + "grad_norm": 14.4375, + "learning_rate": 2.0710667393610483e-06, + "loss": 0.8518, + "step": 6454 + }, + { + "epoch": 0.550063911376225, + "grad_norm": 14.9375, + "learning_rate": 2.070940030059668e-06, + "loss": 0.7177, + "step": 6455 + }, + { + "epoch": 0.550149126544525, + "grad_norm": 14.0625, + "learning_rate": 2.070813305923076e-06, + "loss": 1.0817, + "step": 6456 + }, + { + "epoch": 0.5502343417128249, + "grad_norm": 16.75, + "learning_rate": 2.0706865669535627e-06, + "loss": 0.9834, + "step": 6457 + }, + { + "epoch": 0.5503195568811249, + "grad_norm": 15.3125, + "learning_rate": 2.070559813153417e-06, + "loss": 0.5247, + "step": 6458 + }, + { + "epoch": 0.5504047720494248, + "grad_norm": 17.875, + "learning_rate": 2.0704330445249306e-06, + "loss": 0.5587, + "step": 6459 + }, + { + "epoch": 0.5504899872177248, + "grad_norm": 14.5625, + "learning_rate": 2.0703062610703944e-06, + "loss": 0.6146, + "step": 6460 + }, + { + "epoch": 0.5505752023860248, + "grad_norm": 16.75, + "learning_rate": 2.070179462792099e-06, + "loss": 0.3817, + "step": 6461 + }, + { + "epoch": 0.5506604175543247, + "grad_norm": 18.25, + "learning_rate": 2.0700526496923355e-06, + "loss": 1.0213, + "step": 6462 + }, + { + "epoch": 0.5507456327226247, + "grad_norm": 19.375, + "learning_rate": 2.069925821773397e-06, + "loss": 1.2363, + "step": 6463 + }, + { + "epoch": 0.5508308478909246, + "grad_norm": 16.5, + "learning_rate": 2.0697989790375734e-06, + "loss": 0.8194, + "step": 6464 + }, + { + "epoch": 0.5509160630592246, + "grad_norm": 17.125, + "learning_rate": 2.0696721214871586e-06, + "loss": 0.3935, + "step": 6465 + }, + { + "epoch": 0.5510012782275245, + "grad_norm": 14.1875, + "learning_rate": 2.069545249124444e-06, + "loss": 0.5125, + "step": 6466 + }, + { + "epoch": 0.5510864933958245, + "grad_norm": 10.6875, + "learning_rate": 2.069418361951722e-06, + "loss": 0.4013, + "step": 6467 + }, + { + "epoch": 0.5511717085641245, + "grad_norm": 13.6875, + "learning_rate": 2.069291459971287e-06, + "loss": 0.5735, + "step": 6468 + }, + { + "epoch": 0.5512569237324244, + "grad_norm": 19.5, + "learning_rate": 2.069164543185432e-06, + "loss": 0.5107, + "step": 6469 + }, + { + "epoch": 0.5513421389007244, + "grad_norm": 24.125, + "learning_rate": 2.069037611596449e-06, + "loss": 0.5876, + "step": 6470 + }, + { + "epoch": 0.5514273540690243, + "grad_norm": 16.25, + "learning_rate": 2.0689106652066327e-06, + "loss": 0.5836, + "step": 6471 + }, + { + "epoch": 0.5515125692373243, + "grad_norm": 35.0, + "learning_rate": 2.0687837040182775e-06, + "loss": 1.0261, + "step": 6472 + }, + { + "epoch": 0.5515977844056242, + "grad_norm": 17.375, + "learning_rate": 2.068656728033678e-06, + "loss": 0.525, + "step": 6473 + }, + { + "epoch": 0.5516829995739242, + "grad_norm": 13.1875, + "learning_rate": 2.068529737255127e-06, + "loss": 0.2721, + "step": 6474 + }, + { + "epoch": 0.5517682147422242, + "grad_norm": 11.0, + "learning_rate": 2.0684027316849212e-06, + "loss": 0.371, + "step": 6475 + }, + { + "epoch": 0.5518534299105241, + "grad_norm": 11.0625, + "learning_rate": 2.0682757113253554e-06, + "loss": 0.374, + "step": 6476 + }, + { + "epoch": 0.5519386450788241, + "grad_norm": 17.75, + "learning_rate": 2.068148676178724e-06, + "loss": 0.8038, + "step": 6477 + }, + { + "epoch": 0.552023860247124, + "grad_norm": 17.0, + "learning_rate": 2.0680216262473237e-06, + "loss": 0.8735, + "step": 6478 + }, + { + "epoch": 0.552109075415424, + "grad_norm": 13.6875, + "learning_rate": 2.06789456153345e-06, + "loss": 0.7367, + "step": 6479 + }, + { + "epoch": 0.5521942905837239, + "grad_norm": 10.0, + "learning_rate": 2.067767482039399e-06, + "loss": 0.302, + "step": 6480 + }, + { + "epoch": 0.5522795057520239, + "grad_norm": 14.1875, + "learning_rate": 2.0676403877674677e-06, + "loss": 0.9204, + "step": 6481 + }, + { + "epoch": 0.5523647209203238, + "grad_norm": 17.5, + "learning_rate": 2.067513278719952e-06, + "loss": 0.6165, + "step": 6482 + }, + { + "epoch": 0.5524499360886238, + "grad_norm": 25.125, + "learning_rate": 2.067386154899149e-06, + "loss": 0.969, + "step": 6483 + }, + { + "epoch": 0.5525351512569238, + "grad_norm": 17.875, + "learning_rate": 2.0672590163073566e-06, + "loss": 0.919, + "step": 6484 + }, + { + "epoch": 0.5526203664252237, + "grad_norm": 14.0, + "learning_rate": 2.067131862946872e-06, + "loss": 0.6028, + "step": 6485 + }, + { + "epoch": 0.5527055815935237, + "grad_norm": 18.625, + "learning_rate": 2.067004694819993e-06, + "loss": 0.8, + "step": 6486 + }, + { + "epoch": 0.5527907967618236, + "grad_norm": 13.5, + "learning_rate": 2.0668775119290176e-06, + "loss": 0.6528, + "step": 6487 + }, + { + "epoch": 0.5528760119301236, + "grad_norm": 18.125, + "learning_rate": 2.0667503142762435e-06, + "loss": 0.7239, + "step": 6488 + }, + { + "epoch": 0.5529612270984235, + "grad_norm": 12.125, + "learning_rate": 2.0666231018639707e-06, + "loss": 0.7263, + "step": 6489 + }, + { + "epoch": 0.5530464422667235, + "grad_norm": 14.9375, + "learning_rate": 2.066495874694497e-06, + "loss": 0.6555, + "step": 6490 + }, + { + "epoch": 0.5531316574350235, + "grad_norm": 14.6875, + "learning_rate": 2.0663686327701215e-06, + "loss": 0.5581, + "step": 6491 + }, + { + "epoch": 0.5532168726033234, + "grad_norm": 13.0625, + "learning_rate": 2.0662413760931444e-06, + "loss": 0.7057, + "step": 6492 + }, + { + "epoch": 0.5533020877716234, + "grad_norm": 12.8125, + "learning_rate": 2.0661141046658644e-06, + "loss": 0.528, + "step": 6493 + }, + { + "epoch": 0.5533873029399233, + "grad_norm": 10.5625, + "learning_rate": 2.0659868184905825e-06, + "loss": 0.554, + "step": 6494 + }, + { + "epoch": 0.5534725181082233, + "grad_norm": 15.0625, + "learning_rate": 2.0658595175695975e-06, + "loss": 0.9603, + "step": 6495 + }, + { + "epoch": 0.5535577332765232, + "grad_norm": 11.5, + "learning_rate": 2.0657322019052108e-06, + "loss": 0.4121, + "step": 6496 + }, + { + "epoch": 0.5536429484448232, + "grad_norm": 10.5, + "learning_rate": 2.0656048714997233e-06, + "loss": 0.4158, + "step": 6497 + }, + { + "epoch": 0.5537281636131232, + "grad_norm": 18.75, + "learning_rate": 2.0654775263554353e-06, + "loss": 0.738, + "step": 6498 + }, + { + "epoch": 0.5538133787814231, + "grad_norm": 18.5, + "learning_rate": 2.0653501664746485e-06, + "loss": 0.8041, + "step": 6499 + }, + { + "epoch": 0.5538985939497231, + "grad_norm": 13.3125, + "learning_rate": 2.065222791859664e-06, + "loss": 0.5907, + "step": 6500 + }, + { + "epoch": 0.553983809118023, + "grad_norm": 13.125, + "learning_rate": 2.0650954025127844e-06, + "loss": 0.5922, + "step": 6501 + }, + { + "epoch": 0.554069024286323, + "grad_norm": 16.0, + "learning_rate": 2.0649679984363108e-06, + "loss": 0.8694, + "step": 6502 + }, + { + "epoch": 0.5541542394546229, + "grad_norm": 14.0, + "learning_rate": 2.064840579632546e-06, + "loss": 0.5095, + "step": 6503 + }, + { + "epoch": 0.5542394546229229, + "grad_norm": 21.0, + "learning_rate": 2.064713146103793e-06, + "loss": 0.5398, + "step": 6504 + }, + { + "epoch": 0.5543246697912229, + "grad_norm": 17.75, + "learning_rate": 2.064585697852354e-06, + "loss": 0.7407, + "step": 6505 + }, + { + "epoch": 0.5544098849595228, + "grad_norm": 14.3125, + "learning_rate": 2.0644582348805316e-06, + "loss": 0.6482, + "step": 6506 + }, + { + "epoch": 0.5544951001278228, + "grad_norm": 22.0, + "learning_rate": 2.0643307571906306e-06, + "loss": 1.1116, + "step": 6507 + }, + { + "epoch": 0.5545803152961227, + "grad_norm": 10.75, + "learning_rate": 2.0642032647849535e-06, + "loss": 0.3904, + "step": 6508 + }, + { + "epoch": 0.5546655304644227, + "grad_norm": 16.5, + "learning_rate": 2.064075757665805e-06, + "loss": 0.6855, + "step": 6509 + }, + { + "epoch": 0.5547507456327226, + "grad_norm": 13.125, + "learning_rate": 2.063948235835489e-06, + "loss": 0.56, + "step": 6510 + }, + { + "epoch": 0.5548359608010226, + "grad_norm": 17.75, + "learning_rate": 2.0638206992963098e-06, + "loss": 1.0035, + "step": 6511 + }, + { + "epoch": 0.5549211759693226, + "grad_norm": 16.75, + "learning_rate": 2.063693148050572e-06, + "loss": 0.7331, + "step": 6512 + }, + { + "epoch": 0.5550063911376225, + "grad_norm": 15.3125, + "learning_rate": 2.0635655821005813e-06, + "loss": 0.8202, + "step": 6513 + }, + { + "epoch": 0.5550916063059225, + "grad_norm": 12.375, + "learning_rate": 2.063438001448642e-06, + "loss": 0.4944, + "step": 6514 + }, + { + "epoch": 0.5551768214742224, + "grad_norm": 12.25, + "learning_rate": 2.06331040609706e-06, + "loss": 0.6298, + "step": 6515 + }, + { + "epoch": 0.5552620366425224, + "grad_norm": 13.6875, + "learning_rate": 2.0631827960481414e-06, + "loss": 0.8438, + "step": 6516 + }, + { + "epoch": 0.5553472518108223, + "grad_norm": 20.25, + "learning_rate": 2.0630551713041915e-06, + "loss": 0.7935, + "step": 6517 + }, + { + "epoch": 0.5554324669791223, + "grad_norm": 22.375, + "learning_rate": 2.062927531867518e-06, + "loss": 1.115, + "step": 6518 + }, + { + "epoch": 0.5555176821474223, + "grad_norm": 28.25, + "learning_rate": 2.0627998777404252e-06, + "loss": 1.2354, + "step": 6519 + }, + { + "epoch": 0.5556028973157222, + "grad_norm": 15.8125, + "learning_rate": 2.0626722089252223e-06, + "loss": 1.0337, + "step": 6520 + }, + { + "epoch": 0.5556881124840222, + "grad_norm": 12.875, + "learning_rate": 2.062544525424215e-06, + "loss": 0.6304, + "step": 6521 + }, + { + "epoch": 0.5557733276523221, + "grad_norm": 14.5, + "learning_rate": 2.062416827239711e-06, + "loss": 0.7317, + "step": 6522 + }, + { + "epoch": 0.5558585428206221, + "grad_norm": 15.3125, + "learning_rate": 2.0622891143740182e-06, + "loss": 0.5654, + "step": 6523 + }, + { + "epoch": 0.555943757988922, + "grad_norm": 17.75, + "learning_rate": 2.062161386829444e-06, + "loss": 0.8162, + "step": 6524 + }, + { + "epoch": 0.556028973157222, + "grad_norm": 14.0, + "learning_rate": 2.062033644608297e-06, + "loss": 0.9634, + "step": 6525 + }, + { + "epoch": 0.556114188325522, + "grad_norm": 12.375, + "learning_rate": 2.0619058877128858e-06, + "loss": 0.4097, + "step": 6526 + }, + { + "epoch": 0.5561994034938219, + "grad_norm": 13.5625, + "learning_rate": 2.0617781161455183e-06, + "loss": 0.8011, + "step": 6527 + }, + { + "epoch": 0.5562846186621219, + "grad_norm": 12.5, + "learning_rate": 2.061650329908504e-06, + "loss": 0.6, + "step": 6528 + }, + { + "epoch": 0.5563698338304218, + "grad_norm": 16.5, + "learning_rate": 2.0615225290041525e-06, + "loss": 0.3198, + "step": 6529 + }, + { + "epoch": 0.5564550489987218, + "grad_norm": 41.5, + "learning_rate": 2.0613947134347723e-06, + "loss": 1.2413, + "step": 6530 + }, + { + "epoch": 0.5565402641670217, + "grad_norm": 22.625, + "learning_rate": 2.0612668832026737e-06, + "loss": 1.0818, + "step": 6531 + }, + { + "epoch": 0.5566254793353217, + "grad_norm": 17.125, + "learning_rate": 2.061139038310167e-06, + "loss": 0.8843, + "step": 6532 + }, + { + "epoch": 0.5567106945036217, + "grad_norm": 17.25, + "learning_rate": 2.0610111787595623e-06, + "loss": 0.8752, + "step": 6533 + }, + { + "epoch": 0.5567959096719216, + "grad_norm": 18.625, + "learning_rate": 2.06088330455317e-06, + "loss": 0.8696, + "step": 6534 + }, + { + "epoch": 0.5568811248402216, + "grad_norm": 18.75, + "learning_rate": 2.060755415693301e-06, + "loss": 1.0898, + "step": 6535 + }, + { + "epoch": 0.5569663400085215, + "grad_norm": 29.0, + "learning_rate": 2.060627512182266e-06, + "loss": 0.8938, + "step": 6536 + }, + { + "epoch": 0.5570515551768215, + "grad_norm": 12.75, + "learning_rate": 2.060499594022377e-06, + "loss": 0.6198, + "step": 6537 + }, + { + "epoch": 0.5571367703451214, + "grad_norm": 24.0, + "learning_rate": 2.0603716612159458e-06, + "loss": 0.743, + "step": 6538 + }, + { + "epoch": 0.5572219855134214, + "grad_norm": 11.1875, + "learning_rate": 2.060243713765283e-06, + "loss": 0.3479, + "step": 6539 + }, + { + "epoch": 0.5573072006817213, + "grad_norm": 33.25, + "learning_rate": 2.0601157516727018e-06, + "loss": 0.789, + "step": 6540 + }, + { + "epoch": 0.5573924158500213, + "grad_norm": 27.875, + "learning_rate": 2.0599877749405146e-06, + "loss": 0.8828, + "step": 6541 + }, + { + "epoch": 0.5574776310183213, + "grad_norm": 13.0, + "learning_rate": 2.059859783571033e-06, + "loss": 0.7172, + "step": 6542 + }, + { + "epoch": 0.5575628461866212, + "grad_norm": 16.125, + "learning_rate": 2.0597317775665714e-06, + "loss": 0.6149, + "step": 6543 + }, + { + "epoch": 0.5576480613549212, + "grad_norm": 15.9375, + "learning_rate": 2.0596037569294423e-06, + "loss": 1.1744, + "step": 6544 + }, + { + "epoch": 0.5577332765232211, + "grad_norm": 14.25, + "learning_rate": 2.0594757216619595e-06, + "loss": 0.7805, + "step": 6545 + }, + { + "epoch": 0.5578184916915211, + "grad_norm": 13.4375, + "learning_rate": 2.059347671766436e-06, + "loss": 0.7751, + "step": 6546 + }, + { + "epoch": 0.557903706859821, + "grad_norm": 14.25, + "learning_rate": 2.0592196072451857e-06, + "loss": 0.6645, + "step": 6547 + }, + { + "epoch": 0.557988922028121, + "grad_norm": 14.1875, + "learning_rate": 2.0590915281005243e-06, + "loss": 0.8242, + "step": 6548 + }, + { + "epoch": 0.558074137196421, + "grad_norm": 15.5625, + "learning_rate": 2.0589634343347646e-06, + "loss": 0.9557, + "step": 6549 + }, + { + "epoch": 0.5581593523647209, + "grad_norm": 13.9375, + "learning_rate": 2.058835325950223e-06, + "loss": 0.6361, + "step": 6550 + }, + { + "epoch": 0.5582445675330209, + "grad_norm": 16.0, + "learning_rate": 2.0587072029492134e-06, + "loss": 1.0102, + "step": 6551 + }, + { + "epoch": 0.5583297827013208, + "grad_norm": 11.4375, + "learning_rate": 2.058579065334051e-06, + "loss": 0.5505, + "step": 6552 + }, + { + "epoch": 0.5584149978696208, + "grad_norm": 14.0625, + "learning_rate": 2.0584509131070522e-06, + "loss": 0.3969, + "step": 6553 + }, + { + "epoch": 0.5585002130379207, + "grad_norm": 13.9375, + "learning_rate": 2.0583227462705326e-06, + "loss": 0.689, + "step": 6554 + }, + { + "epoch": 0.5585854282062207, + "grad_norm": 18.75, + "learning_rate": 2.0581945648268077e-06, + "loss": 0.892, + "step": 6555 + }, + { + "epoch": 0.5586706433745207, + "grad_norm": 13.5, + "learning_rate": 2.0580663687781944e-06, + "loss": 0.6407, + "step": 6556 + }, + { + "epoch": 0.5587558585428206, + "grad_norm": 13.875, + "learning_rate": 2.057938158127009e-06, + "loss": 0.5784, + "step": 6557 + }, + { + "epoch": 0.5588410737111206, + "grad_norm": 11.75, + "learning_rate": 2.057809932875569e-06, + "loss": 0.5397, + "step": 6558 + }, + { + "epoch": 0.5589262888794205, + "grad_norm": 19.875, + "learning_rate": 2.0576816930261917e-06, + "loss": 0.9131, + "step": 6559 + }, + { + "epoch": 0.5590115040477205, + "grad_norm": 13.1875, + "learning_rate": 2.0575534385811932e-06, + "loss": 0.5312, + "step": 6560 + }, + { + "epoch": 0.5590967192160204, + "grad_norm": 15.5, + "learning_rate": 2.057425169542892e-06, + "loss": 0.9385, + "step": 6561 + }, + { + "epoch": 0.5591819343843204, + "grad_norm": 12.8125, + "learning_rate": 2.0572968859136066e-06, + "loss": 0.7566, + "step": 6562 + }, + { + "epoch": 0.5592671495526204, + "grad_norm": 16.875, + "learning_rate": 2.0571685876956542e-06, + "loss": 0.6952, + "step": 6563 + }, + { + "epoch": 0.5593523647209203, + "grad_norm": 20.0, + "learning_rate": 2.057040274891354e-06, + "loss": 1.1375, + "step": 6564 + }, + { + "epoch": 0.5594375798892203, + "grad_norm": 17.625, + "learning_rate": 2.056911947503025e-06, + "loss": 1.2299, + "step": 6565 + }, + { + "epoch": 0.5595227950575202, + "grad_norm": 11.4375, + "learning_rate": 2.056783605532985e-06, + "loss": 0.3876, + "step": 6566 + }, + { + "epoch": 0.5596080102258202, + "grad_norm": 13.5, + "learning_rate": 2.056655248983554e-06, + "loss": 0.4853, + "step": 6567 + }, + { + "epoch": 0.5596932253941201, + "grad_norm": 13.75, + "learning_rate": 2.0565268778570518e-06, + "loss": 0.9482, + "step": 6568 + }, + { + "epoch": 0.5597784405624201, + "grad_norm": 12.125, + "learning_rate": 2.0563984921557973e-06, + "loss": 0.5212, + "step": 6569 + }, + { + "epoch": 0.5598636557307201, + "grad_norm": 13.4375, + "learning_rate": 2.0562700918821117e-06, + "loss": 0.7779, + "step": 6570 + }, + { + "epoch": 0.55994887089902, + "grad_norm": 14.0, + "learning_rate": 2.0561416770383147e-06, + "loss": 0.9288, + "step": 6571 + }, + { + "epoch": 0.56003408606732, + "grad_norm": 14.9375, + "learning_rate": 2.056013247626727e-06, + "loss": 1.0304, + "step": 6572 + }, + { + "epoch": 0.5601193012356199, + "grad_norm": 14.5625, + "learning_rate": 2.0558848036496693e-06, + "loss": 0.7677, + "step": 6573 + }, + { + "epoch": 0.5602045164039199, + "grad_norm": 22.0, + "learning_rate": 2.0557563451094627e-06, + "loss": 0.4116, + "step": 6574 + }, + { + "epoch": 0.5602897315722198, + "grad_norm": 22.375, + "learning_rate": 2.055627872008429e-06, + "loss": 1.3887, + "step": 6575 + }, + { + "epoch": 0.5603749467405198, + "grad_norm": 20.875, + "learning_rate": 2.055499384348889e-06, + "loss": 1.071, + "step": 6576 + }, + { + "epoch": 0.5604601619088198, + "grad_norm": 12.9375, + "learning_rate": 2.0553708821331657e-06, + "loss": 0.7193, + "step": 6577 + }, + { + "epoch": 0.5605453770771197, + "grad_norm": 17.75, + "learning_rate": 2.0552423653635805e-06, + "loss": 0.8723, + "step": 6578 + }, + { + "epoch": 0.5606305922454197, + "grad_norm": 16.5, + "learning_rate": 2.055113834042456e-06, + "loss": 0.5388, + "step": 6579 + }, + { + "epoch": 0.5607158074137196, + "grad_norm": 11.25, + "learning_rate": 2.0549852881721148e-06, + "loss": 0.4838, + "step": 6580 + }, + { + "epoch": 0.5608010225820196, + "grad_norm": 15.125, + "learning_rate": 2.05485672775488e-06, + "loss": 0.8816, + "step": 6581 + }, + { + "epoch": 0.5608862377503195, + "grad_norm": 11.3125, + "learning_rate": 2.0547281527930753e-06, + "loss": 0.5521, + "step": 6582 + }, + { + "epoch": 0.5609714529186195, + "grad_norm": 19.375, + "learning_rate": 2.054599563289023e-06, + "loss": 1.0904, + "step": 6583 + }, + { + "epoch": 0.5610566680869195, + "grad_norm": 13.375, + "learning_rate": 2.0544709592450473e-06, + "loss": 0.6103, + "step": 6584 + }, + { + "epoch": 0.5611418832552194, + "grad_norm": 16.0, + "learning_rate": 2.054342340663473e-06, + "loss": 0.9819, + "step": 6585 + }, + { + "epoch": 0.5612270984235194, + "grad_norm": 13.5625, + "learning_rate": 2.0542137075466234e-06, + "loss": 0.8006, + "step": 6586 + }, + { + "epoch": 0.5613123135918193, + "grad_norm": 14.875, + "learning_rate": 2.054085059896824e-06, + "loss": 0.698, + "step": 6587 + }, + { + "epoch": 0.5613975287601193, + "grad_norm": 19.5, + "learning_rate": 2.0539563977163983e-06, + "loss": 0.7512, + "step": 6588 + }, + { + "epoch": 0.5614827439284192, + "grad_norm": 19.375, + "learning_rate": 2.0538277210076722e-06, + "loss": 1.2143, + "step": 6589 + }, + { + "epoch": 0.5615679590967192, + "grad_norm": 31.5, + "learning_rate": 2.0536990297729715e-06, + "loss": 1.0487, + "step": 6590 + }, + { + "epoch": 0.5616531742650192, + "grad_norm": 17.125, + "learning_rate": 2.0535703240146207e-06, + "loss": 1.0768, + "step": 6591 + }, + { + "epoch": 0.5617383894333191, + "grad_norm": 13.75, + "learning_rate": 2.053441603734946e-06, + "loss": 0.8349, + "step": 6592 + }, + { + "epoch": 0.5618236046016191, + "grad_norm": 13.25, + "learning_rate": 2.0533128689362738e-06, + "loss": 0.6898, + "step": 6593 + }, + { + "epoch": 0.561908819769919, + "grad_norm": 12.9375, + "learning_rate": 2.05318411962093e-06, + "loss": 0.6294, + "step": 6594 + }, + { + "epoch": 0.561994034938219, + "grad_norm": 22.25, + "learning_rate": 2.0530553557912417e-06, + "loss": 1.0585, + "step": 6595 + }, + { + "epoch": 0.5620792501065189, + "grad_norm": 13.3125, + "learning_rate": 2.0529265774495357e-06, + "loss": 0.6274, + "step": 6596 + }, + { + "epoch": 0.5621644652748189, + "grad_norm": 15.6875, + "learning_rate": 2.0527977845981383e-06, + "loss": 0.5907, + "step": 6597 + }, + { + "epoch": 0.5622496804431188, + "grad_norm": 13.125, + "learning_rate": 2.052668977239378e-06, + "loss": 0.4528, + "step": 6598 + }, + { + "epoch": 0.5623348956114188, + "grad_norm": 14.0625, + "learning_rate": 2.052540155375583e-06, + "loss": 0.7443, + "step": 6599 + }, + { + "epoch": 0.5624201107797188, + "grad_norm": 14.0, + "learning_rate": 2.0524113190090795e-06, + "loss": 0.7788, + "step": 6600 + }, + { + "epoch": 0.5625053259480187, + "grad_norm": 12.375, + "learning_rate": 2.0522824681421967e-06, + "loss": 0.5758, + "step": 6601 + }, + { + "epoch": 0.5625905411163187, + "grad_norm": 18.5, + "learning_rate": 2.0521536027772633e-06, + "loss": 0.8529, + "step": 6602 + }, + { + "epoch": 0.5626757562846186, + "grad_norm": 12.9375, + "learning_rate": 2.0520247229166075e-06, + "loss": 0.489, + "step": 6603 + }, + { + "epoch": 0.5627609714529186, + "grad_norm": 17.0, + "learning_rate": 2.0518958285625585e-06, + "loss": 0.9931, + "step": 6604 + }, + { + "epoch": 0.5628461866212185, + "grad_norm": 16.125, + "learning_rate": 2.051766919717445e-06, + "loss": 0.8584, + "step": 6605 + }, + { + "epoch": 0.5629314017895185, + "grad_norm": 21.625, + "learning_rate": 2.0516379963835977e-06, + "loss": 0.7719, + "step": 6606 + }, + { + "epoch": 0.5630166169578185, + "grad_norm": 13.75, + "learning_rate": 2.0515090585633453e-06, + "loss": 1.024, + "step": 6607 + }, + { + "epoch": 0.5631018321261184, + "grad_norm": 12.375, + "learning_rate": 2.0513801062590184e-06, + "loss": 0.4838, + "step": 6608 + }, + { + "epoch": 0.5631870472944184, + "grad_norm": 12.25, + "learning_rate": 2.051251139472947e-06, + "loss": 0.6111, + "step": 6609 + }, + { + "epoch": 0.5632722624627183, + "grad_norm": 13.5, + "learning_rate": 2.051122158207462e-06, + "loss": 0.5268, + "step": 6610 + }, + { + "epoch": 0.5633574776310183, + "grad_norm": 11.25, + "learning_rate": 2.050993162464894e-06, + "loss": 0.4595, + "step": 6611 + }, + { + "epoch": 0.5634426927993182, + "grad_norm": 12.625, + "learning_rate": 2.0508641522475734e-06, + "loss": 0.3808, + "step": 6612 + }, + { + "epoch": 0.5635279079676182, + "grad_norm": 11.875, + "learning_rate": 2.050735127557833e-06, + "loss": 0.525, + "step": 6613 + }, + { + "epoch": 0.5636131231359182, + "grad_norm": 14.8125, + "learning_rate": 2.0506060883980034e-06, + "loss": 0.6565, + "step": 6614 + }, + { + "epoch": 0.5636983383042181, + "grad_norm": 17.125, + "learning_rate": 2.0504770347704167e-06, + "loss": 1.0895, + "step": 6615 + }, + { + "epoch": 0.5637835534725181, + "grad_norm": 10.8125, + "learning_rate": 2.050347966677405e-06, + "loss": 0.4161, + "step": 6616 + }, + { + "epoch": 0.563868768640818, + "grad_norm": 15.3125, + "learning_rate": 2.0502188841213012e-06, + "loss": 0.9376, + "step": 6617 + }, + { + "epoch": 0.563953983809118, + "grad_norm": 15.1875, + "learning_rate": 2.0500897871044366e-06, + "loss": 0.7401, + "step": 6618 + }, + { + "epoch": 0.5640391989774179, + "grad_norm": 14.8125, + "learning_rate": 2.049960675629146e-06, + "loss": 0.6603, + "step": 6619 + }, + { + "epoch": 0.5641244141457179, + "grad_norm": 16.5, + "learning_rate": 2.049831549697761e-06, + "loss": 0.6978, + "step": 6620 + }, + { + "epoch": 0.5642096293140179, + "grad_norm": 11.75, + "learning_rate": 2.0497024093126155e-06, + "loss": 0.4691, + "step": 6621 + }, + { + "epoch": 0.5642948444823178, + "grad_norm": 15.3125, + "learning_rate": 2.0495732544760435e-06, + "loss": 0.8976, + "step": 6622 + }, + { + "epoch": 0.5643800596506178, + "grad_norm": 18.75, + "learning_rate": 2.049444085190379e-06, + "loss": 0.2405, + "step": 6623 + }, + { + "epoch": 0.5644652748189177, + "grad_norm": 12.1875, + "learning_rate": 2.0493149014579557e-06, + "loss": 0.2641, + "step": 6624 + }, + { + "epoch": 0.5645504899872177, + "grad_norm": 11.1875, + "learning_rate": 2.049185703281108e-06, + "loss": 0.4319, + "step": 6625 + }, + { + "epoch": 0.5646357051555176, + "grad_norm": 11.0625, + "learning_rate": 2.0490564906621717e-06, + "loss": 0.4829, + "step": 6626 + }, + { + "epoch": 0.5647209203238176, + "grad_norm": 14.5, + "learning_rate": 2.0489272636034808e-06, + "loss": 0.7447, + "step": 6627 + }, + { + "epoch": 0.5648061354921176, + "grad_norm": 16.375, + "learning_rate": 2.0487980221073707e-06, + "loss": 0.9176, + "step": 6628 + }, + { + "epoch": 0.5648913506604175, + "grad_norm": 17.25, + "learning_rate": 2.0486687661761773e-06, + "loss": 1.0748, + "step": 6629 + }, + { + "epoch": 0.5649765658287175, + "grad_norm": 17.375, + "learning_rate": 2.0485394958122357e-06, + "loss": 0.6985, + "step": 6630 + }, + { + "epoch": 0.5650617809970174, + "grad_norm": 15.4375, + "learning_rate": 2.048410211017883e-06, + "loss": 0.9069, + "step": 6631 + }, + { + "epoch": 0.5651469961653174, + "grad_norm": 15.5, + "learning_rate": 2.048280911795455e-06, + "loss": 0.538, + "step": 6632 + }, + { + "epoch": 0.5652322113336173, + "grad_norm": 11.6875, + "learning_rate": 2.0481515981472877e-06, + "loss": 0.3747, + "step": 6633 + }, + { + "epoch": 0.5653174265019173, + "grad_norm": 10.375, + "learning_rate": 2.0480222700757186e-06, + "loss": 0.434, + "step": 6634 + }, + { + "epoch": 0.5654026416702173, + "grad_norm": 20.5, + "learning_rate": 2.0478929275830845e-06, + "loss": 0.8986, + "step": 6635 + }, + { + "epoch": 0.5654878568385172, + "grad_norm": 11.625, + "learning_rate": 2.047763570671723e-06, + "loss": 0.4706, + "step": 6636 + }, + { + "epoch": 0.5655730720068172, + "grad_norm": 16.125, + "learning_rate": 2.0476341993439718e-06, + "loss": 1.1335, + "step": 6637 + }, + { + "epoch": 0.5656582871751171, + "grad_norm": 15.3125, + "learning_rate": 2.047504813602168e-06, + "loss": 1.0077, + "step": 6638 + }, + { + "epoch": 0.5657435023434171, + "grad_norm": 15.5, + "learning_rate": 2.047375413448651e-06, + "loss": 0.6693, + "step": 6639 + }, + { + "epoch": 0.565828717511717, + "grad_norm": 14.5, + "learning_rate": 2.047245998885758e-06, + "loss": 0.9674, + "step": 6640 + }, + { + "epoch": 0.565913932680017, + "grad_norm": 13.6875, + "learning_rate": 2.047116569915828e-06, + "loss": 0.5483, + "step": 6641 + }, + { + "epoch": 0.565999147848317, + "grad_norm": 9.5625, + "learning_rate": 2.0469871265412e-06, + "loss": 0.8026, + "step": 6642 + }, + { + "epoch": 0.5660843630166169, + "grad_norm": 19.75, + "learning_rate": 2.0468576687642135e-06, + "loss": 0.9924, + "step": 6643 + }, + { + "epoch": 0.5661695781849169, + "grad_norm": 13.0625, + "learning_rate": 2.0467281965872073e-06, + "loss": 0.6448, + "step": 6644 + }, + { + "epoch": 0.5662547933532168, + "grad_norm": 16.625, + "learning_rate": 2.046598710012522e-06, + "loss": 0.8509, + "step": 6645 + }, + { + "epoch": 0.5663400085215168, + "grad_norm": 14.5, + "learning_rate": 2.0464692090424963e-06, + "loss": 0.5845, + "step": 6646 + }, + { + "epoch": 0.5664252236898167, + "grad_norm": 15.4375, + "learning_rate": 2.046339693679471e-06, + "loss": 0.8635, + "step": 6647 + }, + { + "epoch": 0.5665104388581167, + "grad_norm": 15.25, + "learning_rate": 2.046210163925787e-06, + "loss": 0.8912, + "step": 6648 + }, + { + "epoch": 0.5665956540264167, + "grad_norm": 14.75, + "learning_rate": 2.0460806197837845e-06, + "loss": 0.6407, + "step": 6649 + }, + { + "epoch": 0.5666808691947166, + "grad_norm": 16.25, + "learning_rate": 2.045951061255805e-06, + "loss": 0.7689, + "step": 6650 + }, + { + "epoch": 0.5667660843630166, + "grad_norm": 14.5, + "learning_rate": 2.0458214883441897e-06, + "loss": 0.8556, + "step": 6651 + }, + { + "epoch": 0.5668512995313165, + "grad_norm": 13.6875, + "learning_rate": 2.0456919010512793e-06, + "loss": 0.4139, + "step": 6652 + }, + { + "epoch": 0.5669365146996165, + "grad_norm": 16.25, + "learning_rate": 2.0455622993794162e-06, + "loss": 0.83, + "step": 6653 + }, + { + "epoch": 0.5670217298679164, + "grad_norm": 14.25, + "learning_rate": 2.0454326833309426e-06, + "loss": 1.0854, + "step": 6654 + }, + { + "epoch": 0.5671069450362164, + "grad_norm": 19.375, + "learning_rate": 2.0453030529082e-06, + "loss": 0.8774, + "step": 6655 + }, + { + "epoch": 0.5671921602045163, + "grad_norm": 19.25, + "learning_rate": 2.045173408113532e-06, + "loss": 0.7203, + "step": 6656 + }, + { + "epoch": 0.5672773753728163, + "grad_norm": 16.375, + "learning_rate": 2.0450437489492806e-06, + "loss": 0.932, + "step": 6657 + }, + { + "epoch": 0.5673625905411164, + "grad_norm": 10.0625, + "learning_rate": 2.0449140754177893e-06, + "loss": 0.4206, + "step": 6658 + }, + { + "epoch": 0.5674478057094163, + "grad_norm": 12.6875, + "learning_rate": 2.0447843875214014e-06, + "loss": 0.7666, + "step": 6659 + }, + { + "epoch": 0.5675330208777163, + "grad_norm": 15.6875, + "learning_rate": 2.0446546852624603e-06, + "loss": 0.71, + "step": 6660 + }, + { + "epoch": 0.5676182360460162, + "grad_norm": 22.625, + "learning_rate": 2.04452496864331e-06, + "loss": 0.8318, + "step": 6661 + }, + { + "epoch": 0.5677034512143162, + "grad_norm": 13.5, + "learning_rate": 2.0443952376662946e-06, + "loss": 0.7203, + "step": 6662 + }, + { + "epoch": 0.5677886663826162, + "grad_norm": 12.0, + "learning_rate": 2.044265492333758e-06, + "loss": 0.5695, + "step": 6663 + }, + { + "epoch": 0.5678738815509161, + "grad_norm": 12.0625, + "learning_rate": 2.044135732648046e-06, + "loss": 0.4586, + "step": 6664 + }, + { + "epoch": 0.5679590967192161, + "grad_norm": 15.5, + "learning_rate": 2.044005958611502e-06, + "loss": 0.9225, + "step": 6665 + }, + { + "epoch": 0.568044311887516, + "grad_norm": 10.5, + "learning_rate": 2.0438761702264722e-06, + "loss": 0.3583, + "step": 6666 + }, + { + "epoch": 0.568129527055816, + "grad_norm": 12.5625, + "learning_rate": 2.043746367495302e-06, + "loss": 0.6217, + "step": 6667 + }, + { + "epoch": 0.5682147422241159, + "grad_norm": 14.1875, + "learning_rate": 2.0436165504203366e-06, + "loss": 0.9287, + "step": 6668 + }, + { + "epoch": 0.5682999573924159, + "grad_norm": 11.4375, + "learning_rate": 2.043486719003922e-06, + "loss": 0.4519, + "step": 6669 + }, + { + "epoch": 0.5683851725607159, + "grad_norm": 13.875, + "learning_rate": 2.043356873248404e-06, + "loss": 0.7129, + "step": 6670 + }, + { + "epoch": 0.5684703877290158, + "grad_norm": 14.0, + "learning_rate": 2.0432270131561304e-06, + "loss": 0.3966, + "step": 6671 + }, + { + "epoch": 0.5685556028973158, + "grad_norm": 19.0, + "learning_rate": 2.043097138729447e-06, + "loss": 1.1598, + "step": 6672 + }, + { + "epoch": 0.5686408180656157, + "grad_norm": 16.125, + "learning_rate": 2.0429672499707e-06, + "loss": 0.8472, + "step": 6673 + }, + { + "epoch": 0.5687260332339157, + "grad_norm": 11.9375, + "learning_rate": 2.0428373468822376e-06, + "loss": 0.3286, + "step": 6674 + }, + { + "epoch": 0.5688112484022156, + "grad_norm": 13.75, + "learning_rate": 2.0427074294664074e-06, + "loss": 0.8259, + "step": 6675 + }, + { + "epoch": 0.5688964635705156, + "grad_norm": 13.375, + "learning_rate": 2.0425774977255567e-06, + "loss": 0.7435, + "step": 6676 + }, + { + "epoch": 0.5689816787388156, + "grad_norm": 14.3125, + "learning_rate": 2.042447551662034e-06, + "loss": 0.6939, + "step": 6677 + }, + { + "epoch": 0.5690668939071155, + "grad_norm": 12.3125, + "learning_rate": 2.0423175912781867e-06, + "loss": 0.6179, + "step": 6678 + }, + { + "epoch": 0.5691521090754155, + "grad_norm": 13.3125, + "learning_rate": 2.0421876165763637e-06, + "loss": 0.7054, + "step": 6679 + }, + { + "epoch": 0.5692373242437154, + "grad_norm": 18.625, + "learning_rate": 2.042057627558914e-06, + "loss": 0.9556, + "step": 6680 + }, + { + "epoch": 0.5693225394120154, + "grad_norm": 18.75, + "learning_rate": 2.0419276242281866e-06, + "loss": 1.0551, + "step": 6681 + }, + { + "epoch": 0.5694077545803153, + "grad_norm": 20.25, + "learning_rate": 2.041797606586531e-06, + "loss": 0.8718, + "step": 6682 + }, + { + "epoch": 0.5694929697486153, + "grad_norm": 19.625, + "learning_rate": 2.0416675746362956e-06, + "loss": 1.1494, + "step": 6683 + }, + { + "epoch": 0.5695781849169153, + "grad_norm": 16.5, + "learning_rate": 2.0415375283798316e-06, + "loss": 0.7423, + "step": 6684 + }, + { + "epoch": 0.5696634000852152, + "grad_norm": 11.5625, + "learning_rate": 2.0414074678194884e-06, + "loss": 0.6575, + "step": 6685 + }, + { + "epoch": 0.5697486152535152, + "grad_norm": 16.5, + "learning_rate": 2.0412773929576165e-06, + "loss": 0.8467, + "step": 6686 + }, + { + "epoch": 0.5698338304218151, + "grad_norm": 13.3125, + "learning_rate": 2.0411473037965663e-06, + "loss": 0.6279, + "step": 6687 + }, + { + "epoch": 0.5699190455901151, + "grad_norm": 12.25, + "learning_rate": 2.0410172003386892e-06, + "loss": 0.4611, + "step": 6688 + }, + { + "epoch": 0.570004260758415, + "grad_norm": 11.0, + "learning_rate": 2.0408870825863354e-06, + "loss": 0.5633, + "step": 6689 + }, + { + "epoch": 0.570089475926715, + "grad_norm": 17.375, + "learning_rate": 2.0407569505418566e-06, + "loss": 0.6571, + "step": 6690 + }, + { + "epoch": 0.570174691095015, + "grad_norm": 21.625, + "learning_rate": 2.040626804207605e-06, + "loss": 0.6264, + "step": 6691 + }, + { + "epoch": 0.5702599062633149, + "grad_norm": 13.4375, + "learning_rate": 2.040496643585932e-06, + "loss": 0.5343, + "step": 6692 + }, + { + "epoch": 0.5703451214316149, + "grad_norm": 11.5625, + "learning_rate": 2.040366468679189e-06, + "loss": 0.5013, + "step": 6693 + }, + { + "epoch": 0.5704303365999148, + "grad_norm": 12.5625, + "learning_rate": 2.0402362794897294e-06, + "loss": 0.494, + "step": 6694 + }, + { + "epoch": 0.5705155517682148, + "grad_norm": 14.375, + "learning_rate": 2.0401060760199063e-06, + "loss": 0.6601, + "step": 6695 + }, + { + "epoch": 0.5706007669365147, + "grad_norm": 12.3125, + "learning_rate": 2.0399758582720715e-06, + "loss": 0.5622, + "step": 6696 + }, + { + "epoch": 0.5706859821048147, + "grad_norm": 14.3125, + "learning_rate": 2.0398456262485784e-06, + "loss": 0.5121, + "step": 6697 + }, + { + "epoch": 0.5707711972731146, + "grad_norm": 26.25, + "learning_rate": 2.0397153799517806e-06, + "loss": 0.86, + "step": 6698 + }, + { + "epoch": 0.5708564124414146, + "grad_norm": 14.875, + "learning_rate": 2.039585119384032e-06, + "loss": 0.8094, + "step": 6699 + }, + { + "epoch": 0.5709416276097146, + "grad_norm": 13.5, + "learning_rate": 2.039454844547686e-06, + "loss": 0.4824, + "step": 6700 + }, + { + "epoch": 0.5710268427780145, + "grad_norm": 13.9375, + "learning_rate": 2.039324555445097e-06, + "loss": 0.5616, + "step": 6701 + }, + { + "epoch": 0.5711120579463145, + "grad_norm": 16.375, + "learning_rate": 2.03919425207862e-06, + "loss": 1.2018, + "step": 6702 + }, + { + "epoch": 0.5711972731146144, + "grad_norm": 16.125, + "learning_rate": 2.039063934450609e-06, + "loss": 0.7871, + "step": 6703 + }, + { + "epoch": 0.5712824882829144, + "grad_norm": 13.125, + "learning_rate": 2.0389336025634194e-06, + "loss": 0.7451, + "step": 6704 + }, + { + "epoch": 0.5713677034512143, + "grad_norm": 20.25, + "learning_rate": 2.038803256419406e-06, + "loss": 0.7874, + "step": 6705 + }, + { + "epoch": 0.5714529186195143, + "grad_norm": 9.9375, + "learning_rate": 2.0386728960209247e-06, + "loss": 0.3616, + "step": 6706 + }, + { + "epoch": 0.5715381337878143, + "grad_norm": 12.5, + "learning_rate": 2.0385425213703313e-06, + "loss": 0.5933, + "step": 6707 + }, + { + "epoch": 0.5716233489561142, + "grad_norm": 17.375, + "learning_rate": 2.038412132469981e-06, + "loss": 0.989, + "step": 6708 + }, + { + "epoch": 0.5717085641244142, + "grad_norm": 11.5, + "learning_rate": 2.0382817293222314e-06, + "loss": 0.5472, + "step": 6709 + }, + { + "epoch": 0.5717937792927141, + "grad_norm": 16.25, + "learning_rate": 2.038151311929438e-06, + "loss": 0.7496, + "step": 6710 + }, + { + "epoch": 0.5718789944610141, + "grad_norm": 15.8125, + "learning_rate": 2.038020880293958e-06, + "loss": 0.9538, + "step": 6711 + }, + { + "epoch": 0.571964209629314, + "grad_norm": 13.875, + "learning_rate": 2.037890434418148e-06, + "loss": 0.8813, + "step": 6712 + }, + { + "epoch": 0.572049424797614, + "grad_norm": 10.5625, + "learning_rate": 2.0377599743043653e-06, + "loss": 0.3264, + "step": 6713 + }, + { + "epoch": 0.572134639965914, + "grad_norm": 14.4375, + "learning_rate": 2.0376294999549684e-06, + "loss": 1.0381, + "step": 6714 + }, + { + "epoch": 0.5722198551342139, + "grad_norm": 14.9375, + "learning_rate": 2.0374990113723136e-06, + "loss": 0.9154, + "step": 6715 + }, + { + "epoch": 0.5723050703025139, + "grad_norm": 16.75, + "learning_rate": 2.0373685085587604e-06, + "loss": 0.7372, + "step": 6716 + }, + { + "epoch": 0.5723902854708138, + "grad_norm": 13.4375, + "learning_rate": 2.0372379915166664e-06, + "loss": 0.9768, + "step": 6717 + }, + { + "epoch": 0.5724755006391138, + "grad_norm": 12.9375, + "learning_rate": 2.03710746024839e-06, + "loss": 0.5258, + "step": 6718 + }, + { + "epoch": 0.5725607158074137, + "grad_norm": 13.0, + "learning_rate": 2.0369769147562908e-06, + "loss": 0.7403, + "step": 6719 + }, + { + "epoch": 0.5726459309757137, + "grad_norm": 20.0, + "learning_rate": 2.0368463550427267e-06, + "loss": 0.8875, + "step": 6720 + }, + { + "epoch": 0.5727311461440137, + "grad_norm": 11.8125, + "learning_rate": 2.0367157811100587e-06, + "loss": 0.5844, + "step": 6721 + }, + { + "epoch": 0.5728163613123136, + "grad_norm": 12.5, + "learning_rate": 2.0365851929606444e-06, + "loss": 0.351, + "step": 6722 + }, + { + "epoch": 0.5729015764806136, + "grad_norm": 13.1875, + "learning_rate": 2.0364545905968454e-06, + "loss": 0.895, + "step": 6723 + }, + { + "epoch": 0.5729867916489135, + "grad_norm": 15.6875, + "learning_rate": 2.0363239740210206e-06, + "loss": 0.7221, + "step": 6724 + }, + { + "epoch": 0.5730720068172135, + "grad_norm": 16.5, + "learning_rate": 2.0361933432355312e-06, + "loss": 0.857, + "step": 6725 + }, + { + "epoch": 0.5731572219855134, + "grad_norm": 14.75, + "learning_rate": 2.0360626982427375e-06, + "loss": 0.8438, + "step": 6726 + }, + { + "epoch": 0.5732424371538134, + "grad_norm": 20.875, + "learning_rate": 2.035932039045e-06, + "loss": 1.0177, + "step": 6727 + }, + { + "epoch": 0.5733276523221134, + "grad_norm": 16.5, + "learning_rate": 2.035801365644681e-06, + "loss": 0.9042, + "step": 6728 + }, + { + "epoch": 0.5734128674904133, + "grad_norm": 15.1875, + "learning_rate": 2.0356706780441406e-06, + "loss": 0.8774, + "step": 6729 + }, + { + "epoch": 0.5734980826587133, + "grad_norm": 14.375, + "learning_rate": 2.0355399762457408e-06, + "loss": 0.7066, + "step": 6730 + }, + { + "epoch": 0.5735832978270132, + "grad_norm": 14.375, + "learning_rate": 2.035409260251844e-06, + "loss": 0.617, + "step": 6731 + }, + { + "epoch": 0.5736685129953132, + "grad_norm": 20.125, + "learning_rate": 2.035278530064812e-06, + "loss": 0.7918, + "step": 6732 + }, + { + "epoch": 0.5737537281636131, + "grad_norm": 12.8125, + "learning_rate": 2.035147785687007e-06, + "loss": 0.4654, + "step": 6733 + }, + { + "epoch": 0.5738389433319131, + "grad_norm": 17.0, + "learning_rate": 2.0350170271207924e-06, + "loss": 0.5677, + "step": 6734 + }, + { + "epoch": 0.573924158500213, + "grad_norm": 19.25, + "learning_rate": 2.034886254368531e-06, + "loss": 0.9539, + "step": 6735 + }, + { + "epoch": 0.574009373668513, + "grad_norm": 13.125, + "learning_rate": 2.034755467432585e-06, + "loss": 0.4915, + "step": 6736 + }, + { + "epoch": 0.574094588836813, + "grad_norm": 14.9375, + "learning_rate": 2.0346246663153193e-06, + "loss": 0.5336, + "step": 6737 + }, + { + "epoch": 0.5741798040051129, + "grad_norm": 22.25, + "learning_rate": 2.0344938510190964e-06, + "loss": 1.0443, + "step": 6738 + }, + { + "epoch": 0.5742650191734129, + "grad_norm": 13.9375, + "learning_rate": 2.034363021546281e-06, + "loss": 0.7953, + "step": 6739 + }, + { + "epoch": 0.5743502343417128, + "grad_norm": 19.5, + "learning_rate": 2.034232177899237e-06, + "loss": 0.9649, + "step": 6740 + }, + { + "epoch": 0.5744354495100128, + "grad_norm": 18.625, + "learning_rate": 2.034101320080329e-06, + "loss": 0.8548, + "step": 6741 + }, + { + "epoch": 0.5745206646783128, + "grad_norm": 15.5625, + "learning_rate": 2.0339704480919216e-06, + "loss": 0.7896, + "step": 6742 + }, + { + "epoch": 0.5746058798466127, + "grad_norm": 14.0, + "learning_rate": 2.0338395619363794e-06, + "loss": 0.7382, + "step": 6743 + }, + { + "epoch": 0.5746910950149127, + "grad_norm": 30.25, + "learning_rate": 2.0337086616160687e-06, + "loss": 0.9601, + "step": 6744 + }, + { + "epoch": 0.5747763101832126, + "grad_norm": 16.375, + "learning_rate": 2.0335777471333545e-06, + "loss": 1.0277, + "step": 6745 + }, + { + "epoch": 0.5748615253515126, + "grad_norm": 15.25, + "learning_rate": 2.033446818490602e-06, + "loss": 0.6509, + "step": 6746 + }, + { + "epoch": 0.5749467405198125, + "grad_norm": 11.6875, + "learning_rate": 2.033315875690178e-06, + "loss": 0.4963, + "step": 6747 + }, + { + "epoch": 0.5750319556881125, + "grad_norm": 18.375, + "learning_rate": 2.0331849187344483e-06, + "loss": 1.1608, + "step": 6748 + }, + { + "epoch": 0.5751171708564125, + "grad_norm": 19.5, + "learning_rate": 2.0330539476257795e-06, + "loss": 0.8022, + "step": 6749 + }, + { + "epoch": 0.5752023860247124, + "grad_norm": 36.5, + "learning_rate": 2.0329229623665384e-06, + "loss": 1.1027, + "step": 6750 + }, + { + "epoch": 0.5752876011930124, + "grad_norm": 14.25, + "learning_rate": 2.032791962959092e-06, + "loss": 0.844, + "step": 6751 + }, + { + "epoch": 0.5753728163613123, + "grad_norm": 21.75, + "learning_rate": 2.032660949405808e-06, + "loss": 0.732, + "step": 6752 + }, + { + "epoch": 0.5754580315296123, + "grad_norm": 30.625, + "learning_rate": 2.0325299217090534e-06, + "loss": 1.2567, + "step": 6753 + }, + { + "epoch": 0.5755432466979122, + "grad_norm": 23.75, + "learning_rate": 2.0323988798711963e-06, + "loss": 0.8542, + "step": 6754 + }, + { + "epoch": 0.5756284618662122, + "grad_norm": 17.0, + "learning_rate": 2.0322678238946047e-06, + "loss": 0.6581, + "step": 6755 + }, + { + "epoch": 0.5757136770345122, + "grad_norm": 13.75, + "learning_rate": 2.032136753781647e-06, + "loss": 0.7792, + "step": 6756 + }, + { + "epoch": 0.5757988922028121, + "grad_norm": 11.4375, + "learning_rate": 2.032005669534692e-06, + "loss": 0.32, + "step": 6757 + }, + { + "epoch": 0.5758841073711121, + "grad_norm": 14.5625, + "learning_rate": 2.031874571156107e-06, + "loss": 0.8013, + "step": 6758 + }, + { + "epoch": 0.575969322539412, + "grad_norm": 12.0, + "learning_rate": 2.031743458648263e-06, + "loss": 0.4304, + "step": 6759 + }, + { + "epoch": 0.576054537707712, + "grad_norm": 14.0, + "learning_rate": 2.031612332013529e-06, + "loss": 0.9376, + "step": 6760 + }, + { + "epoch": 0.5761397528760119, + "grad_norm": 22.25, + "learning_rate": 2.031481191254274e-06, + "loss": 1.0684, + "step": 6761 + }, + { + "epoch": 0.5762249680443119, + "grad_norm": 20.625, + "learning_rate": 2.0313500363728675e-06, + "loss": 0.6924, + "step": 6762 + }, + { + "epoch": 0.5763101832126118, + "grad_norm": 22.875, + "learning_rate": 2.0312188673716802e-06, + "loss": 0.9757, + "step": 6763 + }, + { + "epoch": 0.5763953983809118, + "grad_norm": 13.8125, + "learning_rate": 2.0310876842530826e-06, + "loss": 1.0137, + "step": 6764 + }, + { + "epoch": 0.5764806135492118, + "grad_norm": 13.5, + "learning_rate": 2.0309564870194453e-06, + "loss": 0.6846, + "step": 6765 + }, + { + "epoch": 0.5765658287175117, + "grad_norm": 14.9375, + "learning_rate": 2.0308252756731388e-06, + "loss": 0.8261, + "step": 6766 + }, + { + "epoch": 0.5766510438858117, + "grad_norm": 16.375, + "learning_rate": 2.0306940502165343e-06, + "loss": 0.4896, + "step": 6767 + }, + { + "epoch": 0.5767362590541116, + "grad_norm": 13.75, + "learning_rate": 2.0305628106520033e-06, + "loss": 0.7851, + "step": 6768 + }, + { + "epoch": 0.5768214742224116, + "grad_norm": 14.1875, + "learning_rate": 2.0304315569819176e-06, + "loss": 0.6587, + "step": 6769 + }, + { + "epoch": 0.5769066893907115, + "grad_norm": 17.375, + "learning_rate": 2.0303002892086487e-06, + "loss": 0.4979, + "step": 6770 + }, + { + "epoch": 0.5769919045590115, + "grad_norm": 11.75, + "learning_rate": 2.0301690073345687e-06, + "loss": 0.3979, + "step": 6771 + }, + { + "epoch": 0.5770771197273115, + "grad_norm": 14.9375, + "learning_rate": 2.030037711362051e-06, + "loss": 0.558, + "step": 6772 + }, + { + "epoch": 0.5771623348956114, + "grad_norm": 22.875, + "learning_rate": 2.0299064012934665e-06, + "loss": 0.9738, + "step": 6773 + }, + { + "epoch": 0.5772475500639114, + "grad_norm": 16.375, + "learning_rate": 2.02977507713119e-06, + "loss": 0.8899, + "step": 6774 + }, + { + "epoch": 0.5773327652322113, + "grad_norm": 9.75, + "learning_rate": 2.0296437388775927e-06, + "loss": 0.3427, + "step": 6775 + }, + { + "epoch": 0.5774179804005113, + "grad_norm": 10.625, + "learning_rate": 2.0295123865350496e-06, + "loss": 0.2844, + "step": 6776 + }, + { + "epoch": 0.5775031955688112, + "grad_norm": 10.1875, + "learning_rate": 2.029381020105934e-06, + "loss": 0.4853, + "step": 6777 + }, + { + "epoch": 0.5775884107371112, + "grad_norm": 20.125, + "learning_rate": 2.029249639592619e-06, + "loss": 0.8696, + "step": 6778 + }, + { + "epoch": 0.5776736259054112, + "grad_norm": 13.25, + "learning_rate": 2.02911824499748e-06, + "loss": 0.5983, + "step": 6779 + }, + { + "epoch": 0.5777588410737111, + "grad_norm": 13.125, + "learning_rate": 2.028986836322891e-06, + "loss": 0.8721, + "step": 6780 + }, + { + "epoch": 0.5778440562420111, + "grad_norm": 14.625, + "learning_rate": 2.028855413571226e-06, + "loss": 0.6158, + "step": 6781 + }, + { + "epoch": 0.577929271410311, + "grad_norm": 22.25, + "learning_rate": 2.0287239767448608e-06, + "loss": 0.9306, + "step": 6782 + }, + { + "epoch": 0.578014486578611, + "grad_norm": 13.5, + "learning_rate": 2.02859252584617e-06, + "loss": 0.7195, + "step": 6783 + }, + { + "epoch": 0.5780997017469109, + "grad_norm": 14.375, + "learning_rate": 2.0284610608775294e-06, + "loss": 0.6131, + "step": 6784 + }, + { + "epoch": 0.5781849169152109, + "grad_norm": 22.5, + "learning_rate": 2.0283295818413152e-06, + "loss": 1.3303, + "step": 6785 + }, + { + "epoch": 0.5782701320835109, + "grad_norm": 14.6875, + "learning_rate": 2.028198088739902e-06, + "loss": 0.6883, + "step": 6786 + }, + { + "epoch": 0.5783553472518108, + "grad_norm": 18.375, + "learning_rate": 2.0280665815756672e-06, + "loss": 0.8534, + "step": 6787 + }, + { + "epoch": 0.5784405624201108, + "grad_norm": 13.0625, + "learning_rate": 2.027935060350987e-06, + "loss": 0.5742, + "step": 6788 + }, + { + "epoch": 0.5785257775884107, + "grad_norm": 20.75, + "learning_rate": 2.0278035250682373e-06, + "loss": 0.8672, + "step": 6789 + }, + { + "epoch": 0.5786109927567107, + "grad_norm": 10.4375, + "learning_rate": 2.027671975729796e-06, + "loss": 0.5317, + "step": 6790 + }, + { + "epoch": 0.5786962079250106, + "grad_norm": 14.125, + "learning_rate": 2.027540412338041e-06, + "loss": 0.7013, + "step": 6791 + }, + { + "epoch": 0.5787814230933106, + "grad_norm": 16.875, + "learning_rate": 2.027408834895348e-06, + "loss": 0.9326, + "step": 6792 + }, + { + "epoch": 0.5788666382616106, + "grad_norm": 9.875, + "learning_rate": 2.0272772434040955e-06, + "loss": 0.3284, + "step": 6793 + }, + { + "epoch": 0.5789518534299105, + "grad_norm": 17.0, + "learning_rate": 2.027145637866662e-06, + "loss": 0.9317, + "step": 6794 + }, + { + "epoch": 0.5790370685982105, + "grad_norm": 16.5, + "learning_rate": 2.027014018285425e-06, + "loss": 0.8467, + "step": 6795 + }, + { + "epoch": 0.5791222837665104, + "grad_norm": 22.875, + "learning_rate": 2.026882384662764e-06, + "loss": 1.1981, + "step": 6796 + }, + { + "epoch": 0.5792074989348104, + "grad_norm": 14.5, + "learning_rate": 2.0267507370010564e-06, + "loss": 0.7329, + "step": 6797 + }, + { + "epoch": 0.5792927141031103, + "grad_norm": 13.0625, + "learning_rate": 2.026619075302682e-06, + "loss": 0.7239, + "step": 6798 + }, + { + "epoch": 0.5793779292714103, + "grad_norm": 20.375, + "learning_rate": 2.0264873995700203e-06, + "loss": 1.0197, + "step": 6799 + }, + { + "epoch": 0.5794631444397103, + "grad_norm": 16.75, + "learning_rate": 2.0263557098054503e-06, + "loss": 0.992, + "step": 6800 + }, + { + "epoch": 0.5795483596080102, + "grad_norm": 14.3125, + "learning_rate": 2.0262240060113517e-06, + "loss": 0.5332, + "step": 6801 + }, + { + "epoch": 0.5796335747763102, + "grad_norm": 16.125, + "learning_rate": 2.0260922881901055e-06, + "loss": 0.488, + "step": 6802 + }, + { + "epoch": 0.5797187899446101, + "grad_norm": 13.875, + "learning_rate": 2.0259605563440904e-06, + "loss": 0.8797, + "step": 6803 + }, + { + "epoch": 0.5798040051129101, + "grad_norm": 13.1875, + "learning_rate": 2.025828810475688e-06, + "loss": 0.5762, + "step": 6804 + }, + { + "epoch": 0.57988922028121, + "grad_norm": 15.625, + "learning_rate": 2.0256970505872793e-06, + "loss": 0.8089, + "step": 6805 + }, + { + "epoch": 0.57997443544951, + "grad_norm": 12.0, + "learning_rate": 2.0255652766812444e-06, + "loss": 0.3175, + "step": 6806 + }, + { + "epoch": 0.58005965061781, + "grad_norm": 16.25, + "learning_rate": 2.0254334887599652e-06, + "loss": 0.9716, + "step": 6807 + }, + { + "epoch": 0.5801448657861099, + "grad_norm": 14.0625, + "learning_rate": 2.025301686825823e-06, + "loss": 0.7905, + "step": 6808 + }, + { + "epoch": 0.5802300809544099, + "grad_norm": 15.4375, + "learning_rate": 2.0251698708811996e-06, + "loss": 0.7279, + "step": 6809 + }, + { + "epoch": 0.5803152961227098, + "grad_norm": 24.25, + "learning_rate": 2.0250380409284777e-06, + "loss": 1.1662, + "step": 6810 + }, + { + "epoch": 0.5804005112910098, + "grad_norm": 14.125, + "learning_rate": 2.0249061969700386e-06, + "loss": 0.603, + "step": 6811 + }, + { + "epoch": 0.5804857264593097, + "grad_norm": 11.4375, + "learning_rate": 2.0247743390082648e-06, + "loss": 0.6623, + "step": 6812 + }, + { + "epoch": 0.5805709416276097, + "grad_norm": 13.0625, + "learning_rate": 2.0246424670455404e-06, + "loss": 0.8346, + "step": 6813 + }, + { + "epoch": 0.5806561567959097, + "grad_norm": 18.75, + "learning_rate": 2.0245105810842476e-06, + "loss": 0.6573, + "step": 6814 + }, + { + "epoch": 0.5807413719642096, + "grad_norm": 11.9375, + "learning_rate": 2.024378681126769e-06, + "loss": 0.5768, + "step": 6815 + }, + { + "epoch": 0.5808265871325096, + "grad_norm": 14.375, + "learning_rate": 2.02424676717549e-06, + "loss": 0.7226, + "step": 6816 + }, + { + "epoch": 0.5809118023008095, + "grad_norm": 13.0625, + "learning_rate": 2.0241148392327926e-06, + "loss": 0.8274, + "step": 6817 + }, + { + "epoch": 0.5809970174691095, + "grad_norm": 14.875, + "learning_rate": 2.0239828973010615e-06, + "loss": 0.8791, + "step": 6818 + }, + { + "epoch": 0.5810822326374094, + "grad_norm": 12.3125, + "learning_rate": 2.023850941382682e-06, + "loss": 0.3899, + "step": 6819 + }, + { + "epoch": 0.5811674478057094, + "grad_norm": 12.25, + "learning_rate": 2.023718971480037e-06, + "loss": 0.8178, + "step": 6820 + }, + { + "epoch": 0.5812526629740093, + "grad_norm": 11.5, + "learning_rate": 2.0235869875955123e-06, + "loss": 0.3628, + "step": 6821 + }, + { + "epoch": 0.5813378781423093, + "grad_norm": 15.75, + "learning_rate": 2.0234549897314933e-06, + "loss": 0.6951, + "step": 6822 + }, + { + "epoch": 0.5814230933106093, + "grad_norm": 17.375, + "learning_rate": 2.0233229778903646e-06, + "loss": 0.7554, + "step": 6823 + }, + { + "epoch": 0.5815083084789092, + "grad_norm": 12.4375, + "learning_rate": 2.023190952074512e-06, + "loss": 0.5659, + "step": 6824 + }, + { + "epoch": 0.5815935236472092, + "grad_norm": 16.125, + "learning_rate": 2.0230589122863215e-06, + "loss": 1.1069, + "step": 6825 + }, + { + "epoch": 0.5816787388155091, + "grad_norm": 14.0625, + "learning_rate": 2.022926858528179e-06, + "loss": 0.8917, + "step": 6826 + }, + { + "epoch": 0.5817639539838091, + "grad_norm": 13.1875, + "learning_rate": 2.022794790802471e-06, + "loss": 0.656, + "step": 6827 + }, + { + "epoch": 0.581849169152109, + "grad_norm": 26.125, + "learning_rate": 2.0226627091115837e-06, + "loss": 1.0499, + "step": 6828 + }, + { + "epoch": 0.581934384320409, + "grad_norm": 15.6875, + "learning_rate": 2.0225306134579044e-06, + "loss": 0.6488, + "step": 6829 + }, + { + "epoch": 0.582019599488709, + "grad_norm": 42.5, + "learning_rate": 2.0223985038438202e-06, + "loss": 1.4108, + "step": 6830 + }, + { + "epoch": 0.5821048146570089, + "grad_norm": 20.75, + "learning_rate": 2.0222663802717185e-06, + "loss": 0.899, + "step": 6831 + }, + { + "epoch": 0.5821900298253089, + "grad_norm": 13.875, + "learning_rate": 2.0221342427439867e-06, + "loss": 0.6804, + "step": 6832 + }, + { + "epoch": 0.5822752449936088, + "grad_norm": 12.4375, + "learning_rate": 2.022002091263012e-06, + "loss": 0.547, + "step": 6833 + }, + { + "epoch": 0.5823604601619088, + "grad_norm": 13.1875, + "learning_rate": 2.0218699258311843e-06, + "loss": 0.5511, + "step": 6834 + }, + { + "epoch": 0.5824456753302087, + "grad_norm": 16.625, + "learning_rate": 2.0217377464508904e-06, + "loss": 0.9553, + "step": 6835 + }, + { + "epoch": 0.5825308904985087, + "grad_norm": 16.75, + "learning_rate": 2.021605553124519e-06, + "loss": 0.6424, + "step": 6836 + }, + { + "epoch": 0.5826161056668087, + "grad_norm": 13.9375, + "learning_rate": 2.0214733458544595e-06, + "loss": 0.7372, + "step": 6837 + }, + { + "epoch": 0.5827013208351086, + "grad_norm": 19.375, + "learning_rate": 2.0213411246431014e-06, + "loss": 0.7079, + "step": 6838 + }, + { + "epoch": 0.5827865360034086, + "grad_norm": 12.375, + "learning_rate": 2.0212088894928328e-06, + "loss": 0.6697, + "step": 6839 + }, + { + "epoch": 0.5828717511717085, + "grad_norm": 19.125, + "learning_rate": 2.0210766404060443e-06, + "loss": 1.0168, + "step": 6840 + }, + { + "epoch": 0.5829569663400085, + "grad_norm": 16.75, + "learning_rate": 2.020944377385126e-06, + "loss": 0.689, + "step": 6841 + }, + { + "epoch": 0.5830421815083084, + "grad_norm": 12.75, + "learning_rate": 2.020812100432467e-06, + "loss": 0.3973, + "step": 6842 + }, + { + "epoch": 0.5831273966766084, + "grad_norm": 10.6875, + "learning_rate": 2.0206798095504587e-06, + "loss": 0.2996, + "step": 6843 + }, + { + "epoch": 0.5832126118449084, + "grad_norm": 14.375, + "learning_rate": 2.020547504741491e-06, + "loss": 0.5681, + "step": 6844 + }, + { + "epoch": 0.5832978270132083, + "grad_norm": 19.5, + "learning_rate": 2.0204151860079545e-06, + "loss": 0.9279, + "step": 6845 + }, + { + "epoch": 0.5833830421815083, + "grad_norm": 17.875, + "learning_rate": 2.0202828533522415e-06, + "loss": 0.6386, + "step": 6846 + }, + { + "epoch": 0.5834682573498082, + "grad_norm": 13.4375, + "learning_rate": 2.020150506776743e-06, + "loss": 0.7953, + "step": 6847 + }, + { + "epoch": 0.5835534725181082, + "grad_norm": 10.6875, + "learning_rate": 2.02001814628385e-06, + "loss": 0.5392, + "step": 6848 + }, + { + "epoch": 0.5836386876864081, + "grad_norm": 18.0, + "learning_rate": 2.0198857718759545e-06, + "loss": 1.1605, + "step": 6849 + }, + { + "epoch": 0.5837239028547081, + "grad_norm": 13.3125, + "learning_rate": 2.0197533835554496e-06, + "loss": 0.5234, + "step": 6850 + }, + { + "epoch": 0.5838091180230081, + "grad_norm": 15.75, + "learning_rate": 2.0196209813247265e-06, + "loss": 0.8265, + "step": 6851 + }, + { + "epoch": 0.583894333191308, + "grad_norm": 12.8125, + "learning_rate": 2.0194885651861785e-06, + "loss": 0.7643, + "step": 6852 + }, + { + "epoch": 0.583979548359608, + "grad_norm": 12.6875, + "learning_rate": 2.0193561351421985e-06, + "loss": 0.3946, + "step": 6853 + }, + { + "epoch": 0.5840647635279079, + "grad_norm": 18.875, + "learning_rate": 2.0192236911951794e-06, + "loss": 1.0265, + "step": 6854 + }, + { + "epoch": 0.5841499786962079, + "grad_norm": 18.375, + "learning_rate": 2.019091233347515e-06, + "loss": 1.0825, + "step": 6855 + }, + { + "epoch": 0.5842351938645078, + "grad_norm": 12.0, + "learning_rate": 2.018958761601598e-06, + "loss": 0.6365, + "step": 6856 + }, + { + "epoch": 0.5843204090328078, + "grad_norm": 13.75, + "learning_rate": 2.018826275959823e-06, + "loss": 0.4975, + "step": 6857 + }, + { + "epoch": 0.5844056242011078, + "grad_norm": 16.75, + "learning_rate": 2.0186937764245847e-06, + "loss": 0.8235, + "step": 6858 + }, + { + "epoch": 0.5844908393694077, + "grad_norm": 21.5, + "learning_rate": 2.0185612629982765e-06, + "loss": 0.6771, + "step": 6859 + }, + { + "epoch": 0.5845760545377077, + "grad_norm": 15.25, + "learning_rate": 2.018428735683293e-06, + "loss": 0.788, + "step": 6860 + }, + { + "epoch": 0.5846612697060076, + "grad_norm": 14.625, + "learning_rate": 2.01829619448203e-06, + "loss": 0.921, + "step": 6861 + }, + { + "epoch": 0.5847464848743076, + "grad_norm": 17.25, + "learning_rate": 2.0181636393968823e-06, + "loss": 0.9471, + "step": 6862 + }, + { + "epoch": 0.5848317000426075, + "grad_norm": 14.125, + "learning_rate": 2.018031070430245e-06, + "loss": 0.5874, + "step": 6863 + }, + { + "epoch": 0.5849169152109075, + "grad_norm": 11.5625, + "learning_rate": 2.017898487584514e-06, + "loss": 0.6185, + "step": 6864 + }, + { + "epoch": 0.5850021303792075, + "grad_norm": 17.0, + "learning_rate": 2.0177658908620853e-06, + "loss": 1.0259, + "step": 6865 + }, + { + "epoch": 0.5850873455475074, + "grad_norm": 14.5, + "learning_rate": 2.0176332802653542e-06, + "loss": 0.8108, + "step": 6866 + }, + { + "epoch": 0.5851725607158074, + "grad_norm": 17.125, + "learning_rate": 2.0175006557967184e-06, + "loss": 1.0411, + "step": 6867 + }, + { + "epoch": 0.5852577758841073, + "grad_norm": 14.625, + "learning_rate": 2.017368017458574e-06, + "loss": 0.8712, + "step": 6868 + }, + { + "epoch": 0.5853429910524073, + "grad_norm": 12.0, + "learning_rate": 2.0172353652533177e-06, + "loss": 0.6059, + "step": 6869 + }, + { + "epoch": 0.5854282062207072, + "grad_norm": 17.25, + "learning_rate": 2.017102699183347e-06, + "loss": 0.8399, + "step": 6870 + }, + { + "epoch": 0.5855134213890072, + "grad_norm": 16.75, + "learning_rate": 2.016970019251059e-06, + "loss": 0.8368, + "step": 6871 + }, + { + "epoch": 0.5855986365573072, + "grad_norm": 11.1875, + "learning_rate": 2.0168373254588514e-06, + "loss": 0.4935, + "step": 6872 + }, + { + "epoch": 0.5856838517256071, + "grad_norm": 10.5, + "learning_rate": 2.0167046178091225e-06, + "loss": 0.5049, + "step": 6873 + }, + { + "epoch": 0.5857690668939071, + "grad_norm": 14.8125, + "learning_rate": 2.01657189630427e-06, + "loss": 0.7636, + "step": 6874 + }, + { + "epoch": 0.585854282062207, + "grad_norm": 15.9375, + "learning_rate": 2.0164391609466928e-06, + "loss": 0.6882, + "step": 6875 + }, + { + "epoch": 0.585939497230507, + "grad_norm": 21.125, + "learning_rate": 2.0163064117387886e-06, + "loss": 1.1915, + "step": 6876 + }, + { + "epoch": 0.5860247123988069, + "grad_norm": 12.375, + "learning_rate": 2.0161736486829576e-06, + "loss": 0.5292, + "step": 6877 + }, + { + "epoch": 0.5861099275671069, + "grad_norm": 28.125, + "learning_rate": 2.0160408717815975e-06, + "loss": 0.9053, + "step": 6878 + }, + { + "epoch": 0.5861951427354068, + "grad_norm": 31.875, + "learning_rate": 2.0159080810371094e-06, + "loss": 0.8698, + "step": 6879 + }, + { + "epoch": 0.5862803579037068, + "grad_norm": 23.0, + "learning_rate": 2.0157752764518916e-06, + "loss": 0.7061, + "step": 6880 + }, + { + "epoch": 0.5863655730720068, + "grad_norm": 17.5, + "learning_rate": 2.015642458028345e-06, + "loss": 0.9361, + "step": 6881 + }, + { + "epoch": 0.5864507882403067, + "grad_norm": 15.6875, + "learning_rate": 2.015509625768869e-06, + "loss": 0.8305, + "step": 6882 + }, + { + "epoch": 0.5865360034086067, + "grad_norm": 15.9375, + "learning_rate": 2.015376779675864e-06, + "loss": 0.7644, + "step": 6883 + }, + { + "epoch": 0.5866212185769067, + "grad_norm": 12.1875, + "learning_rate": 2.015243919751731e-06, + "loss": 0.5548, + "step": 6884 + }, + { + "epoch": 0.5867064337452067, + "grad_norm": 14.25, + "learning_rate": 2.0151110459988714e-06, + "loss": 0.96, + "step": 6885 + }, + { + "epoch": 0.5867916489135067, + "grad_norm": 12.125, + "learning_rate": 2.0149781584196853e-06, + "loss": 0.4868, + "step": 6886 + }, + { + "epoch": 0.5868768640818066, + "grad_norm": 23.375, + "learning_rate": 2.014845257016575e-06, + "loss": 0.7223, + "step": 6887 + }, + { + "epoch": 0.5869620792501066, + "grad_norm": 12.25, + "learning_rate": 2.0147123417919412e-06, + "loss": 0.5124, + "step": 6888 + }, + { + "epoch": 0.5870472944184065, + "grad_norm": 12.5625, + "learning_rate": 2.0145794127481868e-06, + "loss": 0.4794, + "step": 6889 + }, + { + "epoch": 0.5871325095867065, + "grad_norm": 18.25, + "learning_rate": 2.0144464698877134e-06, + "loss": 0.9951, + "step": 6890 + }, + { + "epoch": 0.5872177247550064, + "grad_norm": 19.875, + "learning_rate": 2.0143135132129243e-06, + "loss": 0.7204, + "step": 6891 + }, + { + "epoch": 0.5873029399233064, + "grad_norm": 19.0, + "learning_rate": 2.0141805427262207e-06, + "loss": 0.8634, + "step": 6892 + }, + { + "epoch": 0.5873881550916064, + "grad_norm": 15.625, + "learning_rate": 2.014047558430006e-06, + "loss": 0.6356, + "step": 6893 + }, + { + "epoch": 0.5874733702599063, + "grad_norm": 17.75, + "learning_rate": 2.0139145603266843e-06, + "loss": 0.6298, + "step": 6894 + }, + { + "epoch": 0.5875585854282063, + "grad_norm": 16.5, + "learning_rate": 2.0137815484186584e-06, + "loss": 0.8022, + "step": 6895 + }, + { + "epoch": 0.5876438005965062, + "grad_norm": 22.25, + "learning_rate": 2.0136485227083314e-06, + "loss": 1.3751, + "step": 6896 + }, + { + "epoch": 0.5877290157648062, + "grad_norm": 19.875, + "learning_rate": 2.0135154831981077e-06, + "loss": 0.6795, + "step": 6897 + }, + { + "epoch": 0.5878142309331061, + "grad_norm": 11.375, + "learning_rate": 2.0133824298903915e-06, + "loss": 0.5237, + "step": 6898 + }, + { + "epoch": 0.5878994461014061, + "grad_norm": 13.4375, + "learning_rate": 2.0132493627875875e-06, + "loss": 0.5565, + "step": 6899 + }, + { + "epoch": 0.587984661269706, + "grad_norm": 12.5, + "learning_rate": 2.0131162818920997e-06, + "loss": 0.8173, + "step": 6900 + }, + { + "epoch": 0.588069876438006, + "grad_norm": 12.375, + "learning_rate": 2.012983187206333e-06, + "loss": 0.5747, + "step": 6901 + }, + { + "epoch": 0.588155091606306, + "grad_norm": 13.5625, + "learning_rate": 2.012850078732693e-06, + "loss": 0.7632, + "step": 6902 + }, + { + "epoch": 0.5882403067746059, + "grad_norm": 17.875, + "learning_rate": 2.0127169564735853e-06, + "loss": 0.8048, + "step": 6903 + }, + { + "epoch": 0.5883255219429059, + "grad_norm": 42.75, + "learning_rate": 2.0125838204314152e-06, + "loss": 1.3908, + "step": 6904 + }, + { + "epoch": 0.5884107371112058, + "grad_norm": 12.0, + "learning_rate": 2.0124506706085887e-06, + "loss": 0.4324, + "step": 6905 + }, + { + "epoch": 0.5884959522795058, + "grad_norm": 14.9375, + "learning_rate": 2.0123175070075113e-06, + "loss": 0.6727, + "step": 6906 + }, + { + "epoch": 0.5885811674478058, + "grad_norm": 16.625, + "learning_rate": 2.012184329630591e-06, + "loss": 0.9828, + "step": 6907 + }, + { + "epoch": 0.5886663826161057, + "grad_norm": 14.875, + "learning_rate": 2.0120511384802327e-06, + "loss": 0.8603, + "step": 6908 + }, + { + "epoch": 0.5887515977844057, + "grad_norm": 15.0625, + "learning_rate": 2.011917933558844e-06, + "loss": 0.5823, + "step": 6909 + }, + { + "epoch": 0.5888368129527056, + "grad_norm": 15.1875, + "learning_rate": 2.011784714868832e-06, + "loss": 0.5983, + "step": 6910 + }, + { + "epoch": 0.5889220281210056, + "grad_norm": 15.75, + "learning_rate": 2.0116514824126047e-06, + "loss": 0.5843, + "step": 6911 + }, + { + "epoch": 0.5890072432893055, + "grad_norm": 16.125, + "learning_rate": 2.011518236192569e-06, + "loss": 0.7228, + "step": 6912 + }, + { + "epoch": 0.5890924584576055, + "grad_norm": 10.6875, + "learning_rate": 2.0113849762111328e-06, + "loss": 0.5681, + "step": 6913 + }, + { + "epoch": 0.5891776736259055, + "grad_norm": 16.25, + "learning_rate": 2.0112517024707047e-06, + "loss": 0.6344, + "step": 6914 + }, + { + "epoch": 0.5892628887942054, + "grad_norm": 13.375, + "learning_rate": 2.0111184149736925e-06, + "loss": 0.6513, + "step": 6915 + }, + { + "epoch": 0.5893481039625054, + "grad_norm": 15.4375, + "learning_rate": 2.0109851137225055e-06, + "loss": 0.7996, + "step": 6916 + }, + { + "epoch": 0.5894333191308053, + "grad_norm": 15.9375, + "learning_rate": 2.010851798719552e-06, + "loss": 0.5699, + "step": 6917 + }, + { + "epoch": 0.5895185342991053, + "grad_norm": 13.0625, + "learning_rate": 2.010718469967242e-06, + "loss": 0.534, + "step": 6918 + }, + { + "epoch": 0.5896037494674052, + "grad_norm": 12.5625, + "learning_rate": 2.0105851274679834e-06, + "loss": 0.63, + "step": 6919 + }, + { + "epoch": 0.5896889646357052, + "grad_norm": 10.25, + "learning_rate": 2.0104517712241873e-06, + "loss": 0.4163, + "step": 6920 + }, + { + "epoch": 0.5897741798040051, + "grad_norm": 13.75, + "learning_rate": 2.0103184012382625e-06, + "loss": 0.6829, + "step": 6921 + }, + { + "epoch": 0.5898593949723051, + "grad_norm": 13.125, + "learning_rate": 2.01018501751262e-06, + "loss": 0.6144, + "step": 6922 + }, + { + "epoch": 0.5899446101406051, + "grad_norm": 13.8125, + "learning_rate": 2.01005162004967e-06, + "loss": 0.7802, + "step": 6923 + }, + { + "epoch": 0.590029825308905, + "grad_norm": 23.0, + "learning_rate": 2.0099182088518225e-06, + "loss": 1.0607, + "step": 6924 + }, + { + "epoch": 0.590115040477205, + "grad_norm": 16.125, + "learning_rate": 2.009784783921489e-06, + "loss": 0.6496, + "step": 6925 + }, + { + "epoch": 0.5902002556455049, + "grad_norm": 14.375, + "learning_rate": 2.0096513452610805e-06, + "loss": 0.7623, + "step": 6926 + }, + { + "epoch": 0.5902854708138049, + "grad_norm": 12.0, + "learning_rate": 2.009517892873008e-06, + "loss": 0.2801, + "step": 6927 + }, + { + "epoch": 0.5903706859821048, + "grad_norm": 19.875, + "learning_rate": 2.009384426759684e-06, + "loss": 1.1026, + "step": 6928 + }, + { + "epoch": 0.5904559011504048, + "grad_norm": 11.8125, + "learning_rate": 2.0092509469235193e-06, + "loss": 0.4951, + "step": 6929 + }, + { + "epoch": 0.5905411163187048, + "grad_norm": 18.625, + "learning_rate": 2.0091174533669267e-06, + "loss": 0.9784, + "step": 6930 + }, + { + "epoch": 0.5906263314870047, + "grad_norm": 15.25, + "learning_rate": 2.0089839460923183e-06, + "loss": 1.0034, + "step": 6931 + }, + { + "epoch": 0.5907115466553047, + "grad_norm": 15.8125, + "learning_rate": 2.0088504251021073e-06, + "loss": 0.7972, + "step": 6932 + }, + { + "epoch": 0.5907967618236046, + "grad_norm": 14.9375, + "learning_rate": 2.008716890398705e-06, + "loss": 0.7977, + "step": 6933 + }, + { + "epoch": 0.5908819769919046, + "grad_norm": 12.0, + "learning_rate": 2.0085833419845265e-06, + "loss": 0.3837, + "step": 6934 + }, + { + "epoch": 0.5909671921602045, + "grad_norm": 15.6875, + "learning_rate": 2.0084497798619835e-06, + "loss": 0.991, + "step": 6935 + }, + { + "epoch": 0.5910524073285045, + "grad_norm": 11.9375, + "learning_rate": 2.0083162040334906e-06, + "loss": 0.5306, + "step": 6936 + }, + { + "epoch": 0.5911376224968045, + "grad_norm": 13.625, + "learning_rate": 2.0081826145014616e-06, + "loss": 0.8169, + "step": 6937 + }, + { + "epoch": 0.5912228376651044, + "grad_norm": 13.1875, + "learning_rate": 2.00804901126831e-06, + "loss": 0.5817, + "step": 6938 + }, + { + "epoch": 0.5913080528334044, + "grad_norm": 12.5625, + "learning_rate": 2.0079153943364504e-06, + "loss": 0.6393, + "step": 6939 + }, + { + "epoch": 0.5913932680017043, + "grad_norm": 18.75, + "learning_rate": 2.007781763708298e-06, + "loss": 0.882, + "step": 6940 + }, + { + "epoch": 0.5914784831700043, + "grad_norm": 13.25, + "learning_rate": 2.0076481193862667e-06, + "loss": 0.852, + "step": 6941 + }, + { + "epoch": 0.5915636983383042, + "grad_norm": 15.4375, + "learning_rate": 2.0075144613727717e-06, + "loss": 0.7561, + "step": 6942 + }, + { + "epoch": 0.5916489135066042, + "grad_norm": 12.4375, + "learning_rate": 2.007380789670229e-06, + "loss": 0.5374, + "step": 6943 + }, + { + "epoch": 0.5917341286749042, + "grad_norm": 14.8125, + "learning_rate": 2.0072471042810542e-06, + "loss": 0.7475, + "step": 6944 + }, + { + "epoch": 0.5918193438432041, + "grad_norm": 15.625, + "learning_rate": 2.0071134052076625e-06, + "loss": 0.7733, + "step": 6945 + }, + { + "epoch": 0.5919045590115041, + "grad_norm": 14.1875, + "learning_rate": 2.00697969245247e-06, + "loss": 1.0881, + "step": 6946 + }, + { + "epoch": 0.591989774179804, + "grad_norm": 17.125, + "learning_rate": 2.0068459660178932e-06, + "loss": 1.1784, + "step": 6947 + }, + { + "epoch": 0.592074989348104, + "grad_norm": 13.4375, + "learning_rate": 2.006712225906349e-06, + "loss": 0.4161, + "step": 6948 + }, + { + "epoch": 0.5921602045164039, + "grad_norm": 32.5, + "learning_rate": 2.006578472120254e-06, + "loss": 0.9841, + "step": 6949 + }, + { + "epoch": 0.5922454196847039, + "grad_norm": 13.0625, + "learning_rate": 2.006444704662025e-06, + "loss": 0.6029, + "step": 6950 + }, + { + "epoch": 0.5923306348530039, + "grad_norm": 14.6875, + "learning_rate": 2.0063109235340794e-06, + "loss": 0.6538, + "step": 6951 + }, + { + "epoch": 0.5924158500213038, + "grad_norm": 15.5625, + "learning_rate": 2.0061771287388355e-06, + "loss": 0.7806, + "step": 6952 + }, + { + "epoch": 0.5925010651896038, + "grad_norm": 17.5, + "learning_rate": 2.00604332027871e-06, + "loss": 0.7478, + "step": 6953 + }, + { + "epoch": 0.5925862803579037, + "grad_norm": 16.875, + "learning_rate": 2.005909498156121e-06, + "loss": 0.9095, + "step": 6954 + }, + { + "epoch": 0.5926714955262037, + "grad_norm": 11.8125, + "learning_rate": 2.005775662373488e-06, + "loss": 0.5839, + "step": 6955 + }, + { + "epoch": 0.5927567106945036, + "grad_norm": 14.5625, + "learning_rate": 2.0056418129332283e-06, + "loss": 0.4359, + "step": 6956 + }, + { + "epoch": 0.5928419258628036, + "grad_norm": 19.0, + "learning_rate": 2.005507949837762e-06, + "loss": 0.7483, + "step": 6957 + }, + { + "epoch": 0.5929271410311036, + "grad_norm": 15.9375, + "learning_rate": 2.005374073089506e-06, + "loss": 0.9208, + "step": 6958 + }, + { + "epoch": 0.5930123561994035, + "grad_norm": 16.5, + "learning_rate": 2.0052401826908817e-06, + "loss": 0.5269, + "step": 6959 + }, + { + "epoch": 0.5930975713677035, + "grad_norm": 16.125, + "learning_rate": 2.005106278644308e-06, + "loss": 0.4589, + "step": 6960 + }, + { + "epoch": 0.5931827865360034, + "grad_norm": 15.6875, + "learning_rate": 2.0049723609522043e-06, + "loss": 0.535, + "step": 6961 + }, + { + "epoch": 0.5932680017043034, + "grad_norm": 13.25, + "learning_rate": 2.004838429616991e-06, + "loss": 0.7606, + "step": 6962 + }, + { + "epoch": 0.5933532168726033, + "grad_norm": 12.6875, + "learning_rate": 2.004704484641088e-06, + "loss": 0.5362, + "step": 6963 + }, + { + "epoch": 0.5934384320409033, + "grad_norm": 18.0, + "learning_rate": 2.0045705260269164e-06, + "loss": 1.0528, + "step": 6964 + }, + { + "epoch": 0.5935236472092033, + "grad_norm": 20.25, + "learning_rate": 2.0044365537768962e-06, + "loss": 1.1418, + "step": 6965 + }, + { + "epoch": 0.5936088623775032, + "grad_norm": 14.5, + "learning_rate": 2.004302567893449e-06, + "loss": 0.7125, + "step": 6966 + }, + { + "epoch": 0.5936940775458032, + "grad_norm": 34.75, + "learning_rate": 2.0041685683789957e-06, + "loss": 1.3935, + "step": 6967 + }, + { + "epoch": 0.5937792927141031, + "grad_norm": 12.125, + "learning_rate": 2.0040345552359587e-06, + "loss": 0.3361, + "step": 6968 + }, + { + "epoch": 0.5938645078824031, + "grad_norm": 16.375, + "learning_rate": 2.0039005284667586e-06, + "loss": 1.0431, + "step": 6969 + }, + { + "epoch": 0.593949723050703, + "grad_norm": 12.375, + "learning_rate": 2.0037664880738174e-06, + "loss": 0.5184, + "step": 6970 + }, + { + "epoch": 0.594034938219003, + "grad_norm": 12.0, + "learning_rate": 2.003632434059558e-06, + "loss": 0.3794, + "step": 6971 + }, + { + "epoch": 0.594120153387303, + "grad_norm": 12.625, + "learning_rate": 2.003498366426403e-06, + "loss": 0.6815, + "step": 6972 + }, + { + "epoch": 0.5942053685556029, + "grad_norm": 11.875, + "learning_rate": 2.003364285176775e-06, + "loss": 0.8421, + "step": 6973 + }, + { + "epoch": 0.5942905837239029, + "grad_norm": 12.1875, + "learning_rate": 2.003230190313096e-06, + "loss": 0.5783, + "step": 6974 + }, + { + "epoch": 0.5943757988922028, + "grad_norm": 12.75, + "learning_rate": 2.0030960818377906e-06, + "loss": 0.8218, + "step": 6975 + }, + { + "epoch": 0.5944610140605028, + "grad_norm": 16.75, + "learning_rate": 2.002961959753282e-06, + "loss": 1.0087, + "step": 6976 + }, + { + "epoch": 0.5945462292288027, + "grad_norm": 16.125, + "learning_rate": 2.002827824061993e-06, + "loss": 0.653, + "step": 6977 + }, + { + "epoch": 0.5946314443971027, + "grad_norm": 12.5625, + "learning_rate": 2.0026936747663484e-06, + "loss": 0.5757, + "step": 6978 + }, + { + "epoch": 0.5947166595654026, + "grad_norm": 20.75, + "learning_rate": 2.002559511868772e-06, + "loss": 1.0626, + "step": 6979 + }, + { + "epoch": 0.5948018747337026, + "grad_norm": 15.0625, + "learning_rate": 2.002425335371689e-06, + "loss": 0.7168, + "step": 6980 + }, + { + "epoch": 0.5948870899020026, + "grad_norm": 16.0, + "learning_rate": 2.002291145277523e-06, + "loss": 0.769, + "step": 6981 + }, + { + "epoch": 0.5949723050703025, + "grad_norm": 19.25, + "learning_rate": 2.0021569415887e-06, + "loss": 0.717, + "step": 6982 + }, + { + "epoch": 0.5950575202386025, + "grad_norm": 13.9375, + "learning_rate": 2.0020227243076444e-06, + "loss": 0.3966, + "step": 6983 + }, + { + "epoch": 0.5951427354069024, + "grad_norm": 11.0625, + "learning_rate": 2.001888493436782e-06, + "loss": 0.6422, + "step": 6984 + }, + { + "epoch": 0.5952279505752024, + "grad_norm": 16.875, + "learning_rate": 2.0017542489785387e-06, + "loss": 0.6107, + "step": 6985 + }, + { + "epoch": 0.5953131657435023, + "grad_norm": 13.0, + "learning_rate": 2.00161999093534e-06, + "loss": 0.482, + "step": 6986 + }, + { + "epoch": 0.5953983809118023, + "grad_norm": 18.375, + "learning_rate": 2.001485719309612e-06, + "loss": 0.6762, + "step": 6987 + }, + { + "epoch": 0.5954835960801023, + "grad_norm": 15.1875, + "learning_rate": 2.0013514341037817e-06, + "loss": 0.8348, + "step": 6988 + }, + { + "epoch": 0.5955688112484022, + "grad_norm": 19.5, + "learning_rate": 2.0012171353202754e-06, + "loss": 1.065, + "step": 6989 + }, + { + "epoch": 0.5956540264167022, + "grad_norm": 15.875, + "learning_rate": 2.00108282296152e-06, + "loss": 0.7257, + "step": 6990 + }, + { + "epoch": 0.5957392415850021, + "grad_norm": 10.9375, + "learning_rate": 2.0009484970299424e-06, + "loss": 0.3832, + "step": 6991 + }, + { + "epoch": 0.5958244567533021, + "grad_norm": 13.0, + "learning_rate": 2.0008141575279704e-06, + "loss": 0.6952, + "step": 6992 + }, + { + "epoch": 0.595909671921602, + "grad_norm": 16.0, + "learning_rate": 2.0006798044580324e-06, + "loss": 1.0096, + "step": 6993 + }, + { + "epoch": 0.595994887089902, + "grad_norm": 15.9375, + "learning_rate": 2.0005454378225547e-06, + "loss": 0.7708, + "step": 6994 + }, + { + "epoch": 0.596080102258202, + "grad_norm": 15.6875, + "learning_rate": 2.000411057623966e-06, + "loss": 0.7953, + "step": 6995 + }, + { + "epoch": 0.5961653174265019, + "grad_norm": 13.75, + "learning_rate": 2.000276663864695e-06, + "loss": 0.6034, + "step": 6996 + }, + { + "epoch": 0.5962505325948019, + "grad_norm": 15.9375, + "learning_rate": 2.000142256547171e-06, + "loss": 0.6687, + "step": 6997 + }, + { + "epoch": 0.5963357477631018, + "grad_norm": 19.5, + "learning_rate": 2.0000078356738215e-06, + "loss": 1.3052, + "step": 6998 + }, + { + "epoch": 0.5964209629314018, + "grad_norm": 12.6875, + "learning_rate": 1.999873401247076e-06, + "loss": 0.6121, + "step": 6999 + }, + { + "epoch": 0.5965061780997017, + "grad_norm": 13.3125, + "learning_rate": 1.9997389532693644e-06, + "loss": 0.6854, + "step": 7000 + }, + { + "epoch": 0.5965913932680017, + "grad_norm": 18.75, + "learning_rate": 1.999604491743116e-06, + "loss": 0.7471, + "step": 7001 + }, + { + "epoch": 0.5966766084363017, + "grad_norm": 16.875, + "learning_rate": 1.999470016670761e-06, + "loss": 0.6186, + "step": 7002 + }, + { + "epoch": 0.5967618236046016, + "grad_norm": 14.5625, + "learning_rate": 1.9993355280547287e-06, + "loss": 0.6835, + "step": 7003 + }, + { + "epoch": 0.5968470387729016, + "grad_norm": 16.125, + "learning_rate": 1.9992010258974497e-06, + "loss": 0.9148, + "step": 7004 + }, + { + "epoch": 0.5969322539412015, + "grad_norm": 16.125, + "learning_rate": 1.9990665102013553e-06, + "loss": 0.9025, + "step": 7005 + }, + { + "epoch": 0.5970174691095015, + "grad_norm": 10.875, + "learning_rate": 1.9989319809688757e-06, + "loss": 0.4374, + "step": 7006 + }, + { + "epoch": 0.5971026842778014, + "grad_norm": 15.75, + "learning_rate": 1.9987974382024418e-06, + "loss": 1.1291, + "step": 7007 + }, + { + "epoch": 0.5971878994461014, + "grad_norm": 14.625, + "learning_rate": 1.9986628819044854e-06, + "loss": 0.6285, + "step": 7008 + }, + { + "epoch": 0.5972731146144014, + "grad_norm": 17.5, + "learning_rate": 1.998528312077438e-06, + "loss": 0.8571, + "step": 7009 + }, + { + "epoch": 0.5973583297827013, + "grad_norm": 15.9375, + "learning_rate": 1.998393728723731e-06, + "loss": 0.7511, + "step": 7010 + }, + { + "epoch": 0.5974435449510013, + "grad_norm": 21.125, + "learning_rate": 1.9982591318457968e-06, + "loss": 1.0017, + "step": 7011 + }, + { + "epoch": 0.5975287601193012, + "grad_norm": 14.25, + "learning_rate": 1.9981245214460678e-06, + "loss": 1.0681, + "step": 7012 + }, + { + "epoch": 0.5976139752876012, + "grad_norm": 11.3125, + "learning_rate": 1.9979898975269767e-06, + "loss": 0.3837, + "step": 7013 + }, + { + "epoch": 0.5976991904559011, + "grad_norm": 11.25, + "learning_rate": 1.9978552600909555e-06, + "loss": 0.4004, + "step": 7014 + }, + { + "epoch": 0.5977844056242011, + "grad_norm": 14.5625, + "learning_rate": 1.9977206091404377e-06, + "loss": 0.6966, + "step": 7015 + }, + { + "epoch": 0.597869620792501, + "grad_norm": 25.75, + "learning_rate": 1.9975859446778567e-06, + "loss": 0.8567, + "step": 7016 + }, + { + "epoch": 0.597954835960801, + "grad_norm": 15.375, + "learning_rate": 1.997451266705646e-06, + "loss": 0.8897, + "step": 7017 + }, + { + "epoch": 0.598040051129101, + "grad_norm": 14.75, + "learning_rate": 1.9973165752262395e-06, + "loss": 0.887, + "step": 7018 + }, + { + "epoch": 0.5981252662974009, + "grad_norm": 14.75, + "learning_rate": 1.997181870242071e-06, + "loss": 0.6838, + "step": 7019 + }, + { + "epoch": 0.5982104814657009, + "grad_norm": 11.5625, + "learning_rate": 1.9970471517555743e-06, + "loss": 0.6087, + "step": 7020 + }, + { + "epoch": 0.5982956966340008, + "grad_norm": 16.625, + "learning_rate": 1.996912419769185e-06, + "loss": 0.9808, + "step": 7021 + }, + { + "epoch": 0.5983809118023008, + "grad_norm": 15.875, + "learning_rate": 1.9967776742853364e-06, + "loss": 0.4564, + "step": 7022 + }, + { + "epoch": 0.5984661269706008, + "grad_norm": 21.75, + "learning_rate": 1.996642915306465e-06, + "loss": 1.0245, + "step": 7023 + }, + { + "epoch": 0.5985513421389007, + "grad_norm": 12.5625, + "learning_rate": 1.9965081428350046e-06, + "loss": 0.761, + "step": 7024 + }, + { + "epoch": 0.5986365573072007, + "grad_norm": 16.25, + "learning_rate": 1.9963733568733917e-06, + "loss": 1.0195, + "step": 7025 + }, + { + "epoch": 0.5987217724755006, + "grad_norm": 13.125, + "learning_rate": 1.9962385574240616e-06, + "loss": 0.5404, + "step": 7026 + }, + { + "epoch": 0.5988069876438006, + "grad_norm": 14.6875, + "learning_rate": 1.996103744489451e-06, + "loss": 0.6135, + "step": 7027 + }, + { + "epoch": 0.5988922028121005, + "grad_norm": 19.0, + "learning_rate": 1.995968918071995e-06, + "loss": 0.9434, + "step": 7028 + }, + { + "epoch": 0.5989774179804005, + "grad_norm": 15.5625, + "learning_rate": 1.9958340781741302e-06, + "loss": 0.9301, + "step": 7029 + }, + { + "epoch": 0.5990626331487005, + "grad_norm": 18.25, + "learning_rate": 1.995699224798294e-06, + "loss": 0.7505, + "step": 7030 + }, + { + "epoch": 0.5991478483170004, + "grad_norm": 10.25, + "learning_rate": 1.995564357946923e-06, + "loss": 0.5414, + "step": 7031 + }, + { + "epoch": 0.5992330634853004, + "grad_norm": 13.1875, + "learning_rate": 1.995429477622454e-06, + "loss": 0.9013, + "step": 7032 + }, + { + "epoch": 0.5993182786536003, + "grad_norm": 16.625, + "learning_rate": 1.9952945838273256e-06, + "loss": 1.0507, + "step": 7033 + }, + { + "epoch": 0.5994034938219003, + "grad_norm": 11.875, + "learning_rate": 1.9951596765639734e-06, + "loss": 0.5178, + "step": 7034 + }, + { + "epoch": 0.5994887089902002, + "grad_norm": 13.5625, + "learning_rate": 1.9950247558348374e-06, + "loss": 0.6673, + "step": 7035 + }, + { + "epoch": 0.5995739241585002, + "grad_norm": 15.5625, + "learning_rate": 1.9948898216423546e-06, + "loss": 0.817, + "step": 7036 + }, + { + "epoch": 0.5996591393268002, + "grad_norm": 14.5625, + "learning_rate": 1.994754873988964e-06, + "loss": 0.548, + "step": 7037 + }, + { + "epoch": 0.5997443544951001, + "grad_norm": 12.6875, + "learning_rate": 1.9946199128771036e-06, + "loss": 0.7274, + "step": 7038 + }, + { + "epoch": 0.5998295696634001, + "grad_norm": 11.6875, + "learning_rate": 1.9944849383092123e-06, + "loss": 0.5219, + "step": 7039 + }, + { + "epoch": 0.5999147848317, + "grad_norm": 35.75, + "learning_rate": 1.99434995028773e-06, + "loss": 0.9749, + "step": 7040 + }, + { + "epoch": 0.6, + "grad_norm": 23.125, + "learning_rate": 1.994214948815095e-06, + "loss": 0.8435, + "step": 7041 + }, + { + "epoch": 0.6000852151682999, + "grad_norm": 13.375, + "learning_rate": 1.994079933893748e-06, + "loss": 0.8088, + "step": 7042 + }, + { + "epoch": 0.6001704303365999, + "grad_norm": 17.375, + "learning_rate": 1.9939449055261284e-06, + "loss": 0.513, + "step": 7043 + }, + { + "epoch": 0.6002556455048998, + "grad_norm": 18.625, + "learning_rate": 1.993809863714676e-06, + "loss": 1.2511, + "step": 7044 + }, + { + "epoch": 0.6003408606731998, + "grad_norm": 13.4375, + "learning_rate": 1.993674808461832e-06, + "loss": 0.5832, + "step": 7045 + }, + { + "epoch": 0.6004260758414998, + "grad_norm": 14.0, + "learning_rate": 1.993539739770036e-06, + "loss": 0.7236, + "step": 7046 + }, + { + "epoch": 0.6005112910097997, + "grad_norm": 10.75, + "learning_rate": 1.993404657641729e-06, + "loss": 0.3398, + "step": 7047 + }, + { + "epoch": 0.6005965061780997, + "grad_norm": 16.875, + "learning_rate": 1.9932695620793525e-06, + "loss": 0.9036, + "step": 7048 + }, + { + "epoch": 0.6006817213463996, + "grad_norm": 19.625, + "learning_rate": 1.993134453085348e-06, + "loss": 1.061, + "step": 7049 + }, + { + "epoch": 0.6007669365146996, + "grad_norm": 14.5625, + "learning_rate": 1.992999330662156e-06, + "loss": 0.568, + "step": 7050 + }, + { + "epoch": 0.6008521516829995, + "grad_norm": 10.75, + "learning_rate": 1.992864194812219e-06, + "loss": 0.411, + "step": 7051 + }, + { + "epoch": 0.6009373668512995, + "grad_norm": 10.3125, + "learning_rate": 1.9927290455379792e-06, + "loss": 0.3216, + "step": 7052 + }, + { + "epoch": 0.6010225820195995, + "grad_norm": 25.375, + "learning_rate": 1.992593882841879e-06, + "loss": 0.9012, + "step": 7053 + }, + { + "epoch": 0.6011077971878994, + "grad_norm": 26.0, + "learning_rate": 1.9924587067263603e-06, + "loss": 1.1129, + "step": 7054 + }, + { + "epoch": 0.6011930123561994, + "grad_norm": 13.125, + "learning_rate": 1.992323517193866e-06, + "loss": 0.7143, + "step": 7055 + }, + { + "epoch": 0.6012782275244993, + "grad_norm": 18.625, + "learning_rate": 1.992188314246839e-06, + "loss": 0.7174, + "step": 7056 + }, + { + "epoch": 0.6013634426927993, + "grad_norm": 11.375, + "learning_rate": 1.9920530978877235e-06, + "loss": 0.4578, + "step": 7057 + }, + { + "epoch": 0.6014486578610992, + "grad_norm": 15.8125, + "learning_rate": 1.991917868118962e-06, + "loss": 0.8299, + "step": 7058 + }, + { + "epoch": 0.6015338730293992, + "grad_norm": 15.4375, + "learning_rate": 1.991782624942998e-06, + "loss": 0.9186, + "step": 7059 + }, + { + "epoch": 0.6016190881976992, + "grad_norm": 13.875, + "learning_rate": 1.9916473683622773e-06, + "loss": 0.6989, + "step": 7060 + }, + { + "epoch": 0.6017043033659991, + "grad_norm": 13.375, + "learning_rate": 1.991512098379242e-06, + "loss": 0.4594, + "step": 7061 + }, + { + "epoch": 0.6017895185342991, + "grad_norm": 13.9375, + "learning_rate": 1.991376814996338e-06, + "loss": 0.5327, + "step": 7062 + }, + { + "epoch": 0.601874733702599, + "grad_norm": 16.125, + "learning_rate": 1.991241518216009e-06, + "loss": 1.1869, + "step": 7063 + }, + { + "epoch": 0.601959948870899, + "grad_norm": 14.125, + "learning_rate": 1.9911062080407003e-06, + "loss": 0.703, + "step": 7064 + }, + { + "epoch": 0.6020451640391989, + "grad_norm": 12.6875, + "learning_rate": 1.9909708844728578e-06, + "loss": 0.7683, + "step": 7065 + }, + { + "epoch": 0.6021303792074989, + "grad_norm": 12.6875, + "learning_rate": 1.9908355475149257e-06, + "loss": 0.6584, + "step": 7066 + }, + { + "epoch": 0.6022155943757989, + "grad_norm": 17.125, + "learning_rate": 1.9907001971693503e-06, + "loss": 0.9931, + "step": 7067 + }, + { + "epoch": 0.6023008095440988, + "grad_norm": 13.6875, + "learning_rate": 1.990564833438578e-06, + "loss": 0.7852, + "step": 7068 + }, + { + "epoch": 0.6023860247123988, + "grad_norm": 12.625, + "learning_rate": 1.9904294563250537e-06, + "loss": 0.533, + "step": 7069 + }, + { + "epoch": 0.6024712398806987, + "grad_norm": 17.0, + "learning_rate": 1.9902940658312254e-06, + "loss": 0.7965, + "step": 7070 + }, + { + "epoch": 0.6025564550489987, + "grad_norm": 17.875, + "learning_rate": 1.9901586619595383e-06, + "loss": 1.1383, + "step": 7071 + }, + { + "epoch": 0.6026416702172986, + "grad_norm": 14.8125, + "learning_rate": 1.9900232447124403e-06, + "loss": 0.6484, + "step": 7072 + }, + { + "epoch": 0.6027268853855986, + "grad_norm": 19.75, + "learning_rate": 1.9898878140923777e-06, + "loss": 1.0655, + "step": 7073 + }, + { + "epoch": 0.6028121005538986, + "grad_norm": 20.625, + "learning_rate": 1.9897523701017983e-06, + "loss": 0.9423, + "step": 7074 + }, + { + "epoch": 0.6028973157221985, + "grad_norm": 14.375, + "learning_rate": 1.98961691274315e-06, + "loss": 0.7268, + "step": 7075 + }, + { + "epoch": 0.6029825308904985, + "grad_norm": 14.125, + "learning_rate": 1.9894814420188808e-06, + "loss": 0.7933, + "step": 7076 + }, + { + "epoch": 0.6030677460587984, + "grad_norm": 20.75, + "learning_rate": 1.989345957931437e-06, + "loss": 0.9622, + "step": 7077 + }, + { + "epoch": 0.6031529612270984, + "grad_norm": 13.8125, + "learning_rate": 1.9892104604832698e-06, + "loss": 0.591, + "step": 7078 + }, + { + "epoch": 0.6032381763953983, + "grad_norm": 15.25, + "learning_rate": 1.9890749496768253e-06, + "loss": 0.698, + "step": 7079 + }, + { + "epoch": 0.6033233915636983, + "grad_norm": 14.1875, + "learning_rate": 1.9889394255145534e-06, + "loss": 1.0469, + "step": 7080 + }, + { + "epoch": 0.6034086067319983, + "grad_norm": 16.875, + "learning_rate": 1.988803887998903e-06, + "loss": 0.8548, + "step": 7081 + }, + { + "epoch": 0.6034938219002982, + "grad_norm": 12.5625, + "learning_rate": 1.988668337132324e-06, + "loss": 0.6733, + "step": 7082 + }, + { + "epoch": 0.6035790370685982, + "grad_norm": 17.625, + "learning_rate": 1.9885327729172642e-06, + "loss": 0.8537, + "step": 7083 + }, + { + "epoch": 0.6036642522368981, + "grad_norm": 16.75, + "learning_rate": 1.9883971953561752e-06, + "loss": 0.8663, + "step": 7084 + }, + { + "epoch": 0.6037494674051981, + "grad_norm": 13.1875, + "learning_rate": 1.9882616044515064e-06, + "loss": 0.844, + "step": 7085 + }, + { + "epoch": 0.603834682573498, + "grad_norm": 15.125, + "learning_rate": 1.9881260002057082e-06, + "loss": 0.8308, + "step": 7086 + }, + { + "epoch": 0.603919897741798, + "grad_norm": 15.75, + "learning_rate": 1.987990382621231e-06, + "loss": 0.8632, + "step": 7087 + }, + { + "epoch": 0.604005112910098, + "grad_norm": 17.125, + "learning_rate": 1.987854751700525e-06, + "loss": 0.6834, + "step": 7088 + }, + { + "epoch": 0.6040903280783979, + "grad_norm": 12.125, + "learning_rate": 1.9877191074460416e-06, + "loss": 0.7014, + "step": 7089 + }, + { + "epoch": 0.6041755432466979, + "grad_norm": 16.75, + "learning_rate": 1.9875834498602325e-06, + "loss": 0.7654, + "step": 7090 + }, + { + "epoch": 0.6042607584149978, + "grad_norm": 14.6875, + "learning_rate": 1.987447778945549e-06, + "loss": 0.7517, + "step": 7091 + }, + { + "epoch": 0.6043459735832978, + "grad_norm": 14.25, + "learning_rate": 1.987312094704442e-06, + "loss": 1.0279, + "step": 7092 + }, + { + "epoch": 0.6044311887515977, + "grad_norm": 18.375, + "learning_rate": 1.987176397139364e-06, + "loss": 0.6282, + "step": 7093 + }, + { + "epoch": 0.6045164039198977, + "grad_norm": 18.125, + "learning_rate": 1.987040686252768e-06, + "loss": 0.845, + "step": 7094 + }, + { + "epoch": 0.6046016190881977, + "grad_norm": 11.875, + "learning_rate": 1.9869049620471048e-06, + "loss": 0.4352, + "step": 7095 + }, + { + "epoch": 0.6046868342564976, + "grad_norm": 13.3125, + "learning_rate": 1.9867692245248283e-06, + "loss": 0.4568, + "step": 7096 + }, + { + "epoch": 0.6047720494247976, + "grad_norm": 12.9375, + "learning_rate": 1.986633473688391e-06, + "loss": 0.533, + "step": 7097 + }, + { + "epoch": 0.6048572645930975, + "grad_norm": 13.9375, + "learning_rate": 1.986497709540246e-06, + "loss": 0.7173, + "step": 7098 + }, + { + "epoch": 0.6049424797613975, + "grad_norm": 14.4375, + "learning_rate": 1.986361932082847e-06, + "loss": 0.7891, + "step": 7099 + }, + { + "epoch": 0.6050276949296974, + "grad_norm": 15.75, + "learning_rate": 1.986226141318647e-06, + "loss": 0.6882, + "step": 7100 + }, + { + "epoch": 0.6051129100979974, + "grad_norm": 19.0, + "learning_rate": 1.986090337250101e-06, + "loss": 0.9855, + "step": 7101 + }, + { + "epoch": 0.6051981252662973, + "grad_norm": 19.75, + "learning_rate": 1.9859545198796615e-06, + "loss": 0.995, + "step": 7102 + }, + { + "epoch": 0.6052833404345973, + "grad_norm": 27.5, + "learning_rate": 1.9858186892097843e-06, + "loss": 1.2267, + "step": 7103 + }, + { + "epoch": 0.6053685556028973, + "grad_norm": 17.5, + "learning_rate": 1.9856828452429234e-06, + "loss": 0.9777, + "step": 7104 + }, + { + "epoch": 0.6054537707711972, + "grad_norm": 14.6875, + "learning_rate": 1.9855469879815335e-06, + "loss": 0.7694, + "step": 7105 + }, + { + "epoch": 0.6055389859394972, + "grad_norm": 17.375, + "learning_rate": 1.9854111174280706e-06, + "loss": 0.7356, + "step": 7106 + }, + { + "epoch": 0.6056242011077971, + "grad_norm": 12.625, + "learning_rate": 1.9852752335849886e-06, + "loss": 0.6151, + "step": 7107 + }, + { + "epoch": 0.6057094162760971, + "grad_norm": 22.375, + "learning_rate": 1.9851393364547434e-06, + "loss": 1.1245, + "step": 7108 + }, + { + "epoch": 0.6057946314443972, + "grad_norm": 13.5625, + "learning_rate": 1.985003426039792e-06, + "loss": 0.6953, + "step": 7109 + }, + { + "epoch": 0.6058798466126971, + "grad_norm": 18.0, + "learning_rate": 1.984867502342589e-06, + "loss": 0.6803, + "step": 7110 + }, + { + "epoch": 0.6059650617809971, + "grad_norm": 10.0, + "learning_rate": 1.9847315653655916e-06, + "loss": 0.8856, + "step": 7111 + }, + { + "epoch": 0.606050276949297, + "grad_norm": 14.9375, + "learning_rate": 1.984595615111256e-06, + "loss": 0.8024, + "step": 7112 + }, + { + "epoch": 0.606135492117597, + "grad_norm": 17.75, + "learning_rate": 1.9844596515820388e-06, + "loss": 0.7841, + "step": 7113 + }, + { + "epoch": 0.6062207072858969, + "grad_norm": 13.4375, + "learning_rate": 1.984323674780397e-06, + "loss": 0.8036, + "step": 7114 + }, + { + "epoch": 0.6063059224541969, + "grad_norm": 11.4375, + "learning_rate": 1.984187684708788e-06, + "loss": 0.4548, + "step": 7115 + }, + { + "epoch": 0.6063911376224969, + "grad_norm": 15.6875, + "learning_rate": 1.98405168136967e-06, + "loss": 0.7197, + "step": 7116 + }, + { + "epoch": 0.6064763527907968, + "grad_norm": 19.0, + "learning_rate": 1.983915664765499e-06, + "loss": 0.9, + "step": 7117 + }, + { + "epoch": 0.6065615679590968, + "grad_norm": 17.125, + "learning_rate": 1.983779634898734e-06, + "loss": 0.7093, + "step": 7118 + }, + { + "epoch": 0.6066467831273967, + "grad_norm": 24.0, + "learning_rate": 1.983643591771834e-06, + "loss": 1.4911, + "step": 7119 + }, + { + "epoch": 0.6067319982956967, + "grad_norm": 15.4375, + "learning_rate": 1.983507535387256e-06, + "loss": 1.0796, + "step": 7120 + }, + { + "epoch": 0.6068172134639966, + "grad_norm": 13.4375, + "learning_rate": 1.983371465747459e-06, + "loss": 0.6264, + "step": 7121 + }, + { + "epoch": 0.6069024286322966, + "grad_norm": 11.75, + "learning_rate": 1.9832353828549025e-06, + "loss": 0.4159, + "step": 7122 + }, + { + "epoch": 0.6069876438005966, + "grad_norm": 16.0, + "learning_rate": 1.9830992867120454e-06, + "loss": 0.833, + "step": 7123 + }, + { + "epoch": 0.6070728589688965, + "grad_norm": 12.5, + "learning_rate": 1.9829631773213466e-06, + "loss": 0.5283, + "step": 7124 + }, + { + "epoch": 0.6071580741371965, + "grad_norm": 13.5625, + "learning_rate": 1.982827054685266e-06, + "loss": 0.7466, + "step": 7125 + }, + { + "epoch": 0.6072432893054964, + "grad_norm": 15.0625, + "learning_rate": 1.982690918806264e-06, + "loss": 0.9566, + "step": 7126 + }, + { + "epoch": 0.6073285044737964, + "grad_norm": 13.0625, + "learning_rate": 1.9825547696868e-06, + "loss": 0.5742, + "step": 7127 + }, + { + "epoch": 0.6074137196420963, + "grad_norm": 16.0, + "learning_rate": 1.9824186073293354e-06, + "loss": 0.9767, + "step": 7128 + }, + { + "epoch": 0.6074989348103963, + "grad_norm": 17.875, + "learning_rate": 1.982282431736329e-06, + "loss": 0.8745, + "step": 7129 + }, + { + "epoch": 0.6075841499786963, + "grad_norm": 17.625, + "learning_rate": 1.9821462429102435e-06, + "loss": 1.0922, + "step": 7130 + }, + { + "epoch": 0.6076693651469962, + "grad_norm": 10.5, + "learning_rate": 1.9820100408535393e-06, + "loss": 0.4651, + "step": 7131 + }, + { + "epoch": 0.6077545803152962, + "grad_norm": 22.875, + "learning_rate": 1.981873825568677e-06, + "loss": 0.9883, + "step": 7132 + }, + { + "epoch": 0.6078397954835961, + "grad_norm": 11.3125, + "learning_rate": 1.9817375970581187e-06, + "loss": 0.3495, + "step": 7133 + }, + { + "epoch": 0.6079250106518961, + "grad_norm": 18.25, + "learning_rate": 1.9816013553243265e-06, + "loss": 1.1247, + "step": 7134 + }, + { + "epoch": 0.608010225820196, + "grad_norm": 15.6875, + "learning_rate": 1.981465100369762e-06, + "loss": 0.785, + "step": 7135 + }, + { + "epoch": 0.608095440988496, + "grad_norm": 16.375, + "learning_rate": 1.981328832196888e-06, + "loss": 0.7935, + "step": 7136 + }, + { + "epoch": 0.608180656156796, + "grad_norm": 14.9375, + "learning_rate": 1.981192550808166e-06, + "loss": 0.9892, + "step": 7137 + }, + { + "epoch": 0.6082658713250959, + "grad_norm": 15.0, + "learning_rate": 1.98105625620606e-06, + "loss": 0.8352, + "step": 7138 + }, + { + "epoch": 0.6083510864933959, + "grad_norm": 13.5625, + "learning_rate": 1.9809199483930317e-06, + "loss": 0.4018, + "step": 7139 + }, + { + "epoch": 0.6084363016616958, + "grad_norm": 13.875, + "learning_rate": 1.980783627371545e-06, + "loss": 0.7114, + "step": 7140 + }, + { + "epoch": 0.6085215168299958, + "grad_norm": 13.1875, + "learning_rate": 1.9806472931440634e-06, + "loss": 0.3879, + "step": 7141 + }, + { + "epoch": 0.6086067319982957, + "grad_norm": 11.625, + "learning_rate": 1.9805109457130507e-06, + "loss": 0.5047, + "step": 7142 + }, + { + "epoch": 0.6086919471665957, + "grad_norm": 12.6875, + "learning_rate": 1.98037458508097e-06, + "loss": 0.6552, + "step": 7143 + }, + { + "epoch": 0.6087771623348956, + "grad_norm": 25.25, + "learning_rate": 1.9802382112502867e-06, + "loss": 0.9809, + "step": 7144 + }, + { + "epoch": 0.6088623775031956, + "grad_norm": 16.0, + "learning_rate": 1.9801018242234644e-06, + "loss": 0.8074, + "step": 7145 + }, + { + "epoch": 0.6089475926714956, + "grad_norm": 17.625, + "learning_rate": 1.979965424002968e-06, + "loss": 0.7705, + "step": 7146 + }, + { + "epoch": 0.6090328078397955, + "grad_norm": 13.875, + "learning_rate": 1.9798290105912625e-06, + "loss": 0.961, + "step": 7147 + }, + { + "epoch": 0.6091180230080955, + "grad_norm": 14.375, + "learning_rate": 1.9796925839908127e-06, + "loss": 0.6139, + "step": 7148 + }, + { + "epoch": 0.6092032381763954, + "grad_norm": 12.3125, + "learning_rate": 1.979556144204084e-06, + "loss": 0.5328, + "step": 7149 + }, + { + "epoch": 0.6092884533446954, + "grad_norm": 12.5625, + "learning_rate": 1.9794196912335422e-06, + "loss": 0.4705, + "step": 7150 + }, + { + "epoch": 0.6093736685129953, + "grad_norm": 9.9375, + "learning_rate": 1.9792832250816534e-06, + "loss": 0.3359, + "step": 7151 + }, + { + "epoch": 0.6094588836812953, + "grad_norm": 12.0, + "learning_rate": 1.9791467457508836e-06, + "loss": 0.5211, + "step": 7152 + }, + { + "epoch": 0.6095440988495953, + "grad_norm": 88.0, + "learning_rate": 1.9790102532436984e-06, + "loss": 1.0034, + "step": 7153 + }, + { + "epoch": 0.6096293140178952, + "grad_norm": 15.5625, + "learning_rate": 1.9788737475625645e-06, + "loss": 0.8336, + "step": 7154 + }, + { + "epoch": 0.6097145291861952, + "grad_norm": 19.875, + "learning_rate": 1.9787372287099497e-06, + "loss": 1.0461, + "step": 7155 + }, + { + "epoch": 0.6097997443544951, + "grad_norm": 13.25, + "learning_rate": 1.97860069668832e-06, + "loss": 0.5638, + "step": 7156 + }, + { + "epoch": 0.6098849595227951, + "grad_norm": 12.8125, + "learning_rate": 1.9784641515001436e-06, + "loss": 0.3514, + "step": 7157 + }, + { + "epoch": 0.609970174691095, + "grad_norm": 15.375, + "learning_rate": 1.9783275931478867e-06, + "loss": 0.6992, + "step": 7158 + }, + { + "epoch": 0.610055389859395, + "grad_norm": 19.125, + "learning_rate": 1.9781910216340185e-06, + "loss": 0.8709, + "step": 7159 + }, + { + "epoch": 0.610140605027695, + "grad_norm": 11.1875, + "learning_rate": 1.9780544369610055e-06, + "loss": 0.4385, + "step": 7160 + }, + { + "epoch": 0.6102258201959949, + "grad_norm": 22.625, + "learning_rate": 1.9779178391313174e-06, + "loss": 0.789, + "step": 7161 + }, + { + "epoch": 0.6103110353642949, + "grad_norm": 12.5, + "learning_rate": 1.977781228147422e-06, + "loss": 0.6892, + "step": 7162 + }, + { + "epoch": 0.6103962505325948, + "grad_norm": 13.125, + "learning_rate": 1.9776446040117877e-06, + "loss": 0.7682, + "step": 7163 + }, + { + "epoch": 0.6104814657008948, + "grad_norm": 14.875, + "learning_rate": 1.977507966726883e-06, + "loss": 0.7456, + "step": 7164 + }, + { + "epoch": 0.6105666808691947, + "grad_norm": 18.75, + "learning_rate": 1.9773713162951787e-06, + "loss": 0.9256, + "step": 7165 + }, + { + "epoch": 0.6106518960374947, + "grad_norm": 15.5625, + "learning_rate": 1.977234652719143e-06, + "loss": 0.913, + "step": 7166 + }, + { + "epoch": 0.6107371112057947, + "grad_norm": 17.25, + "learning_rate": 1.977097976001246e-06, + "loss": 0.9924, + "step": 7167 + }, + { + "epoch": 0.6108223263740946, + "grad_norm": 17.875, + "learning_rate": 1.9769612861439573e-06, + "loss": 0.9682, + "step": 7168 + }, + { + "epoch": 0.6109075415423946, + "grad_norm": 11.625, + "learning_rate": 1.976824583149747e-06, + "loss": 0.4987, + "step": 7169 + }, + { + "epoch": 0.6109927567106945, + "grad_norm": 13.5625, + "learning_rate": 1.976687867021086e-06, + "loss": 0.8199, + "step": 7170 + }, + { + "epoch": 0.6110779718789945, + "grad_norm": 10.625, + "learning_rate": 1.976551137760444e-06, + "loss": 0.5617, + "step": 7171 + }, + { + "epoch": 0.6111631870472944, + "grad_norm": 17.125, + "learning_rate": 1.976414395370293e-06, + "loss": 0.8479, + "step": 7172 + }, + { + "epoch": 0.6112484022155944, + "grad_norm": 14.5625, + "learning_rate": 1.9762776398531027e-06, + "loss": 0.9526, + "step": 7173 + }, + { + "epoch": 0.6113336173838944, + "grad_norm": 16.875, + "learning_rate": 1.9761408712113455e-06, + "loss": 0.574, + "step": 7174 + }, + { + "epoch": 0.6114188325521943, + "grad_norm": 11.625, + "learning_rate": 1.9760040894474923e-06, + "loss": 0.5072, + "step": 7175 + }, + { + "epoch": 0.6115040477204943, + "grad_norm": 14.3125, + "learning_rate": 1.9758672945640157e-06, + "loss": 0.7083, + "step": 7176 + }, + { + "epoch": 0.6115892628887942, + "grad_norm": 11.4375, + "learning_rate": 1.9757304865633864e-06, + "loss": 0.6307, + "step": 7177 + }, + { + "epoch": 0.6116744780570942, + "grad_norm": 11.4375, + "learning_rate": 1.975593665448078e-06, + "loss": 0.3968, + "step": 7178 + }, + { + "epoch": 0.6117596932253941, + "grad_norm": 13.8125, + "learning_rate": 1.9754568312205625e-06, + "loss": 0.5954, + "step": 7179 + }, + { + "epoch": 0.6118449083936941, + "grad_norm": 17.375, + "learning_rate": 1.975319983883312e-06, + "loss": 0.9735, + "step": 7180 + }, + { + "epoch": 0.611930123561994, + "grad_norm": 14.875, + "learning_rate": 1.9751831234388004e-06, + "loss": 0.7383, + "step": 7181 + }, + { + "epoch": 0.612015338730294, + "grad_norm": 12.9375, + "learning_rate": 1.9750462498895003e-06, + "loss": 0.708, + "step": 7182 + }, + { + "epoch": 0.612100553898594, + "grad_norm": 18.125, + "learning_rate": 1.974909363237886e-06, + "loss": 0.6898, + "step": 7183 + }, + { + "epoch": 0.6121857690668939, + "grad_norm": 12.3125, + "learning_rate": 1.9747724634864297e-06, + "loss": 0.5799, + "step": 7184 + }, + { + "epoch": 0.6122709842351939, + "grad_norm": 11.3125, + "learning_rate": 1.9746355506376063e-06, + "loss": 0.3727, + "step": 7185 + }, + { + "epoch": 0.6123561994034938, + "grad_norm": 14.9375, + "learning_rate": 1.9744986246938902e-06, + "loss": 0.8564, + "step": 7186 + }, + { + "epoch": 0.6124414145717938, + "grad_norm": 23.125, + "learning_rate": 1.9743616856577557e-06, + "loss": 0.9636, + "step": 7187 + }, + { + "epoch": 0.6125266297400938, + "grad_norm": 13.25, + "learning_rate": 1.9742247335316766e-06, + "loss": 0.7136, + "step": 7188 + }, + { + "epoch": 0.6126118449083937, + "grad_norm": 15.5625, + "learning_rate": 1.974087768318128e-06, + "loss": 0.9755, + "step": 7189 + }, + { + "epoch": 0.6126970600766937, + "grad_norm": 15.0, + "learning_rate": 1.973950790019586e-06, + "loss": 0.7343, + "step": 7190 + }, + { + "epoch": 0.6127822752449936, + "grad_norm": 20.75, + "learning_rate": 1.973813798638525e-06, + "loss": 0.9544, + "step": 7191 + }, + { + "epoch": 0.6128674904132936, + "grad_norm": 11.75, + "learning_rate": 1.973676794177421e-06, + "loss": 0.49, + "step": 7192 + }, + { + "epoch": 0.6129527055815935, + "grad_norm": 11.125, + "learning_rate": 1.9735397766387494e-06, + "loss": 0.3162, + "step": 7193 + }, + { + "epoch": 0.6130379207498935, + "grad_norm": 11.5, + "learning_rate": 1.9734027460249862e-06, + "loss": 0.4049, + "step": 7194 + }, + { + "epoch": 0.6131231359181935, + "grad_norm": 13.9375, + "learning_rate": 1.9732657023386083e-06, + "loss": 0.7312, + "step": 7195 + }, + { + "epoch": 0.6132083510864934, + "grad_norm": 13.0, + "learning_rate": 1.973128645582092e-06, + "loss": 0.8564, + "step": 7196 + }, + { + "epoch": 0.6132935662547934, + "grad_norm": 13.8125, + "learning_rate": 1.9729915757579133e-06, + "loss": 0.8346, + "step": 7197 + }, + { + "epoch": 0.6133787814230933, + "grad_norm": 17.5, + "learning_rate": 1.97285449286855e-06, + "loss": 0.8359, + "step": 7198 + }, + { + "epoch": 0.6134639965913933, + "grad_norm": 17.875, + "learning_rate": 1.9727173969164797e-06, + "loss": 0.7552, + "step": 7199 + }, + { + "epoch": 0.6135492117596932, + "grad_norm": 12.0625, + "learning_rate": 1.972580287904179e-06, + "loss": 0.445, + "step": 7200 + }, + { + "epoch": 0.6136344269279932, + "grad_norm": 14.5625, + "learning_rate": 1.9724431658341256e-06, + "loss": 0.9417, + "step": 7201 + }, + { + "epoch": 0.6137196420962931, + "grad_norm": 15.375, + "learning_rate": 1.972306030708798e-06, + "loss": 0.6247, + "step": 7202 + }, + { + "epoch": 0.6138048572645931, + "grad_norm": 15.75, + "learning_rate": 1.9721688825306737e-06, + "loss": 0.5556, + "step": 7203 + }, + { + "epoch": 0.6138900724328931, + "grad_norm": 19.75, + "learning_rate": 1.972031721302232e-06, + "loss": 0.7019, + "step": 7204 + }, + { + "epoch": 0.613975287601193, + "grad_norm": 13.4375, + "learning_rate": 1.971894547025951e-06, + "loss": 0.592, + "step": 7205 + }, + { + "epoch": 0.614060502769493, + "grad_norm": 12.625, + "learning_rate": 1.971757359704309e-06, + "loss": 0.4118, + "step": 7206 + }, + { + "epoch": 0.6141457179377929, + "grad_norm": 16.375, + "learning_rate": 1.9716201593397863e-06, + "loss": 0.9746, + "step": 7207 + }, + { + "epoch": 0.6142309331060929, + "grad_norm": 13.375, + "learning_rate": 1.9714829459348618e-06, + "loss": 0.5896, + "step": 7208 + }, + { + "epoch": 0.6143161482743928, + "grad_norm": 12.6875, + "learning_rate": 1.9713457194920143e-06, + "loss": 0.705, + "step": 7209 + }, + { + "epoch": 0.6144013634426928, + "grad_norm": 14.4375, + "learning_rate": 1.971208480013725e-06, + "loss": 0.8955, + "step": 7210 + }, + { + "epoch": 0.6144865786109928, + "grad_norm": 16.5, + "learning_rate": 1.9710712275024726e-06, + "loss": 0.9369, + "step": 7211 + }, + { + "epoch": 0.6145717937792927, + "grad_norm": 23.625, + "learning_rate": 1.9709339619607385e-06, + "loss": 1.138, + "step": 7212 + }, + { + "epoch": 0.6146570089475927, + "grad_norm": 18.0, + "learning_rate": 1.9707966833910025e-06, + "loss": 0.8803, + "step": 7213 + }, + { + "epoch": 0.6147422241158926, + "grad_norm": 12.875, + "learning_rate": 1.9706593917957457e-06, + "loss": 0.3852, + "step": 7214 + }, + { + "epoch": 0.6148274392841926, + "grad_norm": 11.75, + "learning_rate": 1.970522087177449e-06, + "loss": 0.3889, + "step": 7215 + }, + { + "epoch": 0.6149126544524925, + "grad_norm": 11.875, + "learning_rate": 1.970384769538594e-06, + "loss": 0.5226, + "step": 7216 + }, + { + "epoch": 0.6149978696207925, + "grad_norm": 14.75, + "learning_rate": 1.970247438881661e-06, + "loss": 0.6125, + "step": 7217 + }, + { + "epoch": 0.6150830847890925, + "grad_norm": 15.0625, + "learning_rate": 1.970110095209133e-06, + "loss": 0.7232, + "step": 7218 + }, + { + "epoch": 0.6151682999573924, + "grad_norm": 11.9375, + "learning_rate": 1.9699727385234915e-06, + "loss": 0.4787, + "step": 7219 + }, + { + "epoch": 0.6152535151256924, + "grad_norm": 12.75, + "learning_rate": 1.969835368827219e-06, + "loss": 0.3575, + "step": 7220 + }, + { + "epoch": 0.6153387302939923, + "grad_norm": 17.375, + "learning_rate": 1.969697986122797e-06, + "loss": 0.9611, + "step": 7221 + }, + { + "epoch": 0.6154239454622923, + "grad_norm": 17.375, + "learning_rate": 1.9695605904127085e-06, + "loss": 0.8128, + "step": 7222 + }, + { + "epoch": 0.6155091606305922, + "grad_norm": 24.5, + "learning_rate": 1.969423181699437e-06, + "loss": 1.3074, + "step": 7223 + }, + { + "epoch": 0.6155943757988922, + "grad_norm": 15.0625, + "learning_rate": 1.969285759985465e-06, + "loss": 0.7217, + "step": 7224 + }, + { + "epoch": 0.6156795909671922, + "grad_norm": 15.6875, + "learning_rate": 1.969148325273276e-06, + "loss": 0.8687, + "step": 7225 + }, + { + "epoch": 0.6157648061354921, + "grad_norm": 13.3125, + "learning_rate": 1.9690108775653535e-06, + "loss": 0.771, + "step": 7226 + }, + { + "epoch": 0.6158500213037921, + "grad_norm": 16.125, + "learning_rate": 1.9688734168641816e-06, + "loss": 0.8843, + "step": 7227 + }, + { + "epoch": 0.615935236472092, + "grad_norm": 17.625, + "learning_rate": 1.968735943172244e-06, + "loss": 0.6593, + "step": 7228 + }, + { + "epoch": 0.616020451640392, + "grad_norm": 17.375, + "learning_rate": 1.9685984564920253e-06, + "loss": 1.0138, + "step": 7229 + }, + { + "epoch": 0.6161056668086919, + "grad_norm": 17.875, + "learning_rate": 1.9684609568260097e-06, + "loss": 0.9872, + "step": 7230 + }, + { + "epoch": 0.6161908819769919, + "grad_norm": 9.25, + "learning_rate": 1.9683234441766824e-06, + "loss": 0.2184, + "step": 7231 + }, + { + "epoch": 0.6162760971452919, + "grad_norm": 12.375, + "learning_rate": 1.968185918546528e-06, + "loss": 0.4289, + "step": 7232 + }, + { + "epoch": 0.6163613123135918, + "grad_norm": 25.875, + "learning_rate": 1.968048379938032e-06, + "loss": 1.0873, + "step": 7233 + }, + { + "epoch": 0.6164465274818918, + "grad_norm": 17.25, + "learning_rate": 1.9679108283536796e-06, + "loss": 0.9169, + "step": 7234 + }, + { + "epoch": 0.6165317426501917, + "grad_norm": 10.125, + "learning_rate": 1.967773263795957e-06, + "loss": 0.4318, + "step": 7235 + }, + { + "epoch": 0.6166169578184917, + "grad_norm": 14.25, + "learning_rate": 1.967635686267349e-06, + "loss": 0.6577, + "step": 7236 + }, + { + "epoch": 0.6167021729867916, + "grad_norm": 12.125, + "learning_rate": 1.967498095770343e-06, + "loss": 0.5881, + "step": 7237 + }, + { + "epoch": 0.6167873881550916, + "grad_norm": 13.75, + "learning_rate": 1.967360492307425e-06, + "loss": 0.5848, + "step": 7238 + }, + { + "epoch": 0.6168726033233916, + "grad_norm": 18.375, + "learning_rate": 1.967222875881081e-06, + "loss": 1.015, + "step": 7239 + }, + { + "epoch": 0.6169578184916915, + "grad_norm": 18.875, + "learning_rate": 1.967085246493799e-06, + "loss": 0.959, + "step": 7240 + }, + { + "epoch": 0.6170430336599915, + "grad_norm": 14.8125, + "learning_rate": 1.9669476041480652e-06, + "loss": 0.4899, + "step": 7241 + }, + { + "epoch": 0.6171282488282914, + "grad_norm": 15.0, + "learning_rate": 1.9668099488463672e-06, + "loss": 0.833, + "step": 7242 + }, + { + "epoch": 0.6172134639965914, + "grad_norm": 18.75, + "learning_rate": 1.966672280591193e-06, + "loss": 0.8397, + "step": 7243 + }, + { + "epoch": 0.6172986791648913, + "grad_norm": 16.125, + "learning_rate": 1.96653459938503e-06, + "loss": 0.8139, + "step": 7244 + }, + { + "epoch": 0.6173838943331913, + "grad_norm": 16.75, + "learning_rate": 1.966396905230366e-06, + "loss": 0.9423, + "step": 7245 + }, + { + "epoch": 0.6174691095014913, + "grad_norm": 13.0, + "learning_rate": 1.966259198129689e-06, + "loss": 0.3414, + "step": 7246 + }, + { + "epoch": 0.6175543246697912, + "grad_norm": 13.0625, + "learning_rate": 1.966121478085489e-06, + "loss": 0.7184, + "step": 7247 + }, + { + "epoch": 0.6176395398380912, + "grad_norm": 17.25, + "learning_rate": 1.9659837451002532e-06, + "loss": 1.2353, + "step": 7248 + }, + { + "epoch": 0.6177247550063911, + "grad_norm": 22.5, + "learning_rate": 1.9658459991764713e-06, + "loss": 0.743, + "step": 7249 + }, + { + "epoch": 0.6178099701746911, + "grad_norm": 10.875, + "learning_rate": 1.9657082403166323e-06, + "loss": 0.3962, + "step": 7250 + }, + { + "epoch": 0.617895185342991, + "grad_norm": 19.5, + "learning_rate": 1.9655704685232257e-06, + "loss": 0.6572, + "step": 7251 + }, + { + "epoch": 0.617980400511291, + "grad_norm": 21.5, + "learning_rate": 1.9654326837987414e-06, + "loss": 1.2073, + "step": 7252 + }, + { + "epoch": 0.618065615679591, + "grad_norm": 20.875, + "learning_rate": 1.965294886145669e-06, + "loss": 1.2191, + "step": 7253 + }, + { + "epoch": 0.6181508308478909, + "grad_norm": 13.9375, + "learning_rate": 1.9651570755664983e-06, + "loss": 0.8051, + "step": 7254 + }, + { + "epoch": 0.6182360460161909, + "grad_norm": 12.375, + "learning_rate": 1.965019252063721e-06, + "loss": 0.5261, + "step": 7255 + }, + { + "epoch": 0.6183212611844908, + "grad_norm": 15.3125, + "learning_rate": 1.964881415639826e-06, + "loss": 0.9768, + "step": 7256 + }, + { + "epoch": 0.6184064763527908, + "grad_norm": 11.4375, + "learning_rate": 1.964743566297305e-06, + "loss": 0.4253, + "step": 7257 + }, + { + "epoch": 0.6184916915210907, + "grad_norm": 12.25, + "learning_rate": 1.964605704038649e-06, + "loss": 0.4746, + "step": 7258 + }, + { + "epoch": 0.6185769066893907, + "grad_norm": 11.6875, + "learning_rate": 1.9644678288663497e-06, + "loss": 0.6493, + "step": 7259 + }, + { + "epoch": 0.6186621218576907, + "grad_norm": 11.875, + "learning_rate": 1.964329940782898e-06, + "loss": 0.5322, + "step": 7260 + }, + { + "epoch": 0.6187473370259906, + "grad_norm": 15.1875, + "learning_rate": 1.9641920397907864e-06, + "loss": 0.6982, + "step": 7261 + }, + { + "epoch": 0.6188325521942906, + "grad_norm": 12.125, + "learning_rate": 1.9640541258925062e-06, + "loss": 0.4661, + "step": 7262 + }, + { + "epoch": 0.6189177673625905, + "grad_norm": 14.5, + "learning_rate": 1.96391619909055e-06, + "loss": 0.9326, + "step": 7263 + }, + { + "epoch": 0.6190029825308905, + "grad_norm": 10.875, + "learning_rate": 1.9637782593874097e-06, + "loss": 0.4321, + "step": 7264 + }, + { + "epoch": 0.6190881976991904, + "grad_norm": 12.8125, + "learning_rate": 1.963640306785579e-06, + "loss": 0.7314, + "step": 7265 + }, + { + "epoch": 0.6191734128674904, + "grad_norm": 17.875, + "learning_rate": 1.9635023412875497e-06, + "loss": 0.442, + "step": 7266 + }, + { + "epoch": 0.6192586280357903, + "grad_norm": 15.5, + "learning_rate": 1.9633643628958163e-06, + "loss": 0.8016, + "step": 7267 + }, + { + "epoch": 0.6193438432040903, + "grad_norm": 16.875, + "learning_rate": 1.9632263716128713e-06, + "loss": 0.8831, + "step": 7268 + }, + { + "epoch": 0.6194290583723903, + "grad_norm": 10.375, + "learning_rate": 1.9630883674412086e-06, + "loss": 0.2588, + "step": 7269 + }, + { + "epoch": 0.6195142735406902, + "grad_norm": 17.25, + "learning_rate": 1.962950350383322e-06, + "loss": 0.9381, + "step": 7270 + }, + { + "epoch": 0.6195994887089902, + "grad_norm": 13.1875, + "learning_rate": 1.9628123204417055e-06, + "loss": 0.928, + "step": 7271 + }, + { + "epoch": 0.6196847038772901, + "grad_norm": 16.875, + "learning_rate": 1.962674277618854e-06, + "loss": 0.8597, + "step": 7272 + }, + { + "epoch": 0.6197699190455901, + "grad_norm": 14.875, + "learning_rate": 1.9625362219172616e-06, + "loss": 0.9568, + "step": 7273 + }, + { + "epoch": 0.61985513421389, + "grad_norm": 15.4375, + "learning_rate": 1.9623981533394226e-06, + "loss": 0.798, + "step": 7274 + }, + { + "epoch": 0.61994034938219, + "grad_norm": 13.5, + "learning_rate": 1.9622600718878328e-06, + "loss": 0.6489, + "step": 7275 + }, + { + "epoch": 0.62002556455049, + "grad_norm": 15.25, + "learning_rate": 1.9621219775649873e-06, + "loss": 0.6409, + "step": 7276 + }, + { + "epoch": 0.6201107797187899, + "grad_norm": 15.25, + "learning_rate": 1.961983870373382e-06, + "loss": 0.775, + "step": 7277 + }, + { + "epoch": 0.6201959948870899, + "grad_norm": 12.6875, + "learning_rate": 1.961845750315512e-06, + "loss": 0.5108, + "step": 7278 + }, + { + "epoch": 0.6202812100553898, + "grad_norm": 11.0625, + "learning_rate": 1.961707617393873e-06, + "loss": 0.4268, + "step": 7279 + }, + { + "epoch": 0.6203664252236898, + "grad_norm": 23.25, + "learning_rate": 1.9615694716109622e-06, + "loss": 1.4864, + "step": 7280 + }, + { + "epoch": 0.6204516403919897, + "grad_norm": 13.125, + "learning_rate": 1.961431312969275e-06, + "loss": 0.7875, + "step": 7281 + }, + { + "epoch": 0.6205368555602897, + "grad_norm": 15.1875, + "learning_rate": 1.9612931414713092e-06, + "loss": 0.6852, + "step": 7282 + }, + { + "epoch": 0.6206220707285897, + "grad_norm": 16.5, + "learning_rate": 1.9611549571195608e-06, + "loss": 0.8346, + "step": 7283 + }, + { + "epoch": 0.6207072858968896, + "grad_norm": 14.375, + "learning_rate": 1.9610167599165268e-06, + "loss": 0.6892, + "step": 7284 + }, + { + "epoch": 0.6207925010651896, + "grad_norm": 13.625, + "learning_rate": 1.960878549864706e-06, + "loss": 0.6599, + "step": 7285 + }, + { + "epoch": 0.6208777162334895, + "grad_norm": 15.6875, + "learning_rate": 1.9607403269665943e-06, + "loss": 1.1817, + "step": 7286 + }, + { + "epoch": 0.6209629314017895, + "grad_norm": 15.6875, + "learning_rate": 1.96060209122469e-06, + "loss": 0.7801, + "step": 7287 + }, + { + "epoch": 0.6210481465700894, + "grad_norm": 12.5625, + "learning_rate": 1.960463842641491e-06, + "loss": 0.6956, + "step": 7288 + }, + { + "epoch": 0.6211333617383894, + "grad_norm": 14.5625, + "learning_rate": 1.960325581219497e-06, + "loss": 0.7612, + "step": 7289 + }, + { + "epoch": 0.6212185769066894, + "grad_norm": 13.5, + "learning_rate": 1.9601873069612047e-06, + "loss": 0.4868, + "step": 7290 + }, + { + "epoch": 0.6213037920749893, + "grad_norm": 17.375, + "learning_rate": 1.960049019869114e-06, + "loss": 0.7407, + "step": 7291 + }, + { + "epoch": 0.6213890072432893, + "grad_norm": 13.375, + "learning_rate": 1.9599107199457234e-06, + "loss": 0.8536, + "step": 7292 + }, + { + "epoch": 0.6214742224115892, + "grad_norm": 17.25, + "learning_rate": 1.9597724071935324e-06, + "loss": 0.927, + "step": 7293 + }, + { + "epoch": 0.6215594375798892, + "grad_norm": 17.25, + "learning_rate": 1.95963408161504e-06, + "loss": 0.7374, + "step": 7294 + }, + { + "epoch": 0.6216446527481891, + "grad_norm": 15.3125, + "learning_rate": 1.959495743212746e-06, + "loss": 1.0294, + "step": 7295 + }, + { + "epoch": 0.6217298679164891, + "grad_norm": 9.25, + "learning_rate": 1.9593573919891506e-06, + "loss": 0.2492, + "step": 7296 + }, + { + "epoch": 0.621815083084789, + "grad_norm": 14.4375, + "learning_rate": 1.959219027946754e-06, + "loss": 0.8218, + "step": 7297 + }, + { + "epoch": 0.621900298253089, + "grad_norm": 13.125, + "learning_rate": 1.959080651088056e-06, + "loss": 0.6199, + "step": 7298 + }, + { + "epoch": 0.621985513421389, + "grad_norm": 14.25, + "learning_rate": 1.958942261415558e-06, + "loss": 0.6136, + "step": 7299 + }, + { + "epoch": 0.6220707285896889, + "grad_norm": 14.1875, + "learning_rate": 1.9588038589317604e-06, + "loss": 0.7423, + "step": 7300 + }, + { + "epoch": 0.6221559437579889, + "grad_norm": 15.875, + "learning_rate": 1.9586654436391644e-06, + "loss": 0.7883, + "step": 7301 + }, + { + "epoch": 0.6222411589262888, + "grad_norm": 14.0, + "learning_rate": 1.958527015540271e-06, + "loss": 0.7382, + "step": 7302 + }, + { + "epoch": 0.6223263740945888, + "grad_norm": 12.3125, + "learning_rate": 1.958388574637582e-06, + "loss": 0.6697, + "step": 7303 + }, + { + "epoch": 0.6224115892628888, + "grad_norm": 23.375, + "learning_rate": 1.958250120933599e-06, + "loss": 1.1391, + "step": 7304 + }, + { + "epoch": 0.6224968044311887, + "grad_norm": 19.625, + "learning_rate": 1.9581116544308243e-06, + "loss": 0.8311, + "step": 7305 + }, + { + "epoch": 0.6225820195994887, + "grad_norm": 13.3125, + "learning_rate": 1.9579731751317597e-06, + "loss": 0.5279, + "step": 7306 + }, + { + "epoch": 0.6226672347677886, + "grad_norm": 12.8125, + "learning_rate": 1.9578346830389083e-06, + "loss": 0.6692, + "step": 7307 + }, + { + "epoch": 0.6227524499360886, + "grad_norm": 15.4375, + "learning_rate": 1.957696178154772e-06, + "loss": 0.7927, + "step": 7308 + }, + { + "epoch": 0.6228376651043885, + "grad_norm": 13.3125, + "learning_rate": 1.9575576604818544e-06, + "loss": 0.6831, + "step": 7309 + }, + { + "epoch": 0.6229228802726885, + "grad_norm": 40.25, + "learning_rate": 1.957419130022658e-06, + "loss": 0.8341, + "step": 7310 + }, + { + "epoch": 0.6230080954409885, + "grad_norm": 15.125, + "learning_rate": 1.9572805867796866e-06, + "loss": 1.093, + "step": 7311 + }, + { + "epoch": 0.6230933106092884, + "grad_norm": 23.125, + "learning_rate": 1.9571420307554446e-06, + "loss": 0.8997, + "step": 7312 + }, + { + "epoch": 0.6231785257775884, + "grad_norm": 18.625, + "learning_rate": 1.957003461952434e-06, + "loss": 1.2009, + "step": 7313 + }, + { + "epoch": 0.6232637409458883, + "grad_norm": 11.75, + "learning_rate": 1.9568648803731603e-06, + "loss": 0.4673, + "step": 7314 + }, + { + "epoch": 0.6233489561141883, + "grad_norm": 17.0, + "learning_rate": 1.9567262860201273e-06, + "loss": 0.7243, + "step": 7315 + }, + { + "epoch": 0.6234341712824882, + "grad_norm": 14.0625, + "learning_rate": 1.95658767889584e-06, + "loss": 1.0163, + "step": 7316 + }, + { + "epoch": 0.6235193864507882, + "grad_norm": 14.5, + "learning_rate": 1.9564490590028026e-06, + "loss": 0.8731, + "step": 7317 + }, + { + "epoch": 0.6236046016190882, + "grad_norm": 14.375, + "learning_rate": 1.9563104263435203e-06, + "loss": 0.8048, + "step": 7318 + }, + { + "epoch": 0.6236898167873881, + "grad_norm": 28.375, + "learning_rate": 1.9561717809204984e-06, + "loss": 0.8914, + "step": 7319 + }, + { + "epoch": 0.6237750319556881, + "grad_norm": 11.1875, + "learning_rate": 1.956033122736242e-06, + "loss": 0.3814, + "step": 7320 + }, + { + "epoch": 0.623860247123988, + "grad_norm": 17.375, + "learning_rate": 1.9558944517932574e-06, + "loss": 0.8612, + "step": 7321 + }, + { + "epoch": 0.623945462292288, + "grad_norm": 30.25, + "learning_rate": 1.9557557680940503e-06, + "loss": 1.2453, + "step": 7322 + }, + { + "epoch": 0.6240306774605879, + "grad_norm": 13.5, + "learning_rate": 1.9556170716411266e-06, + "loss": 0.5864, + "step": 7323 + }, + { + "epoch": 0.6241158926288879, + "grad_norm": 12.0, + "learning_rate": 1.9554783624369926e-06, + "loss": 0.3767, + "step": 7324 + }, + { + "epoch": 0.6242011077971878, + "grad_norm": 15.375, + "learning_rate": 1.9553396404841555e-06, + "loss": 0.6376, + "step": 7325 + }, + { + "epoch": 0.6242863229654878, + "grad_norm": 13.0, + "learning_rate": 1.9552009057851218e-06, + "loss": 0.6214, + "step": 7326 + }, + { + "epoch": 0.6243715381337878, + "grad_norm": 12.875, + "learning_rate": 1.9550621583423985e-06, + "loss": 0.7112, + "step": 7327 + }, + { + "epoch": 0.6244567533020877, + "grad_norm": 15.5625, + "learning_rate": 1.954923398158493e-06, + "loss": 0.8004, + "step": 7328 + }, + { + "epoch": 0.6245419684703877, + "grad_norm": 11.8125, + "learning_rate": 1.954784625235913e-06, + "loss": 0.4881, + "step": 7329 + }, + { + "epoch": 0.6246271836386876, + "grad_norm": 11.9375, + "learning_rate": 1.954645839577166e-06, + "loss": 0.55, + "step": 7330 + }, + { + "epoch": 0.6247123988069876, + "grad_norm": 12.4375, + "learning_rate": 1.9545070411847604e-06, + "loss": 0.4759, + "step": 7331 + }, + { + "epoch": 0.6247976139752875, + "grad_norm": 12.5, + "learning_rate": 1.9543682300612037e-06, + "loss": 0.4297, + "step": 7332 + }, + { + "epoch": 0.6248828291435875, + "grad_norm": 15.9375, + "learning_rate": 1.954229406209005e-06, + "loss": 0.8195, + "step": 7333 + }, + { + "epoch": 0.6249680443118875, + "grad_norm": 11.125, + "learning_rate": 1.954090569630673e-06, + "loss": 0.5054, + "step": 7334 + }, + { + "epoch": 0.6250532594801875, + "grad_norm": 16.375, + "learning_rate": 1.9539517203287157e-06, + "loss": 0.7437, + "step": 7335 + }, + { + "epoch": 0.6251384746484875, + "grad_norm": 11.4375, + "learning_rate": 1.9538128583056435e-06, + "loss": 0.5343, + "step": 7336 + }, + { + "epoch": 0.6252236898167874, + "grad_norm": 9.75, + "learning_rate": 1.953673983563965e-06, + "loss": 0.2756, + "step": 7337 + }, + { + "epoch": 0.6253089049850874, + "grad_norm": 14.8125, + "learning_rate": 1.9535350961061903e-06, + "loss": 0.9079, + "step": 7338 + }, + { + "epoch": 0.6253941201533874, + "grad_norm": 30.125, + "learning_rate": 1.9533961959348284e-06, + "loss": 1.1096, + "step": 7339 + }, + { + "epoch": 0.6254793353216873, + "grad_norm": 12.3125, + "learning_rate": 1.9532572830523904e-06, + "loss": 0.8761, + "step": 7340 + }, + { + "epoch": 0.6255645504899873, + "grad_norm": 11.8125, + "learning_rate": 1.953118357461386e-06, + "loss": 0.7214, + "step": 7341 + }, + { + "epoch": 0.6256497656582872, + "grad_norm": 11.625, + "learning_rate": 1.952979419164326e-06, + "loss": 0.4724, + "step": 7342 + }, + { + "epoch": 0.6257349808265872, + "grad_norm": 11.375, + "learning_rate": 1.9528404681637205e-06, + "loss": 0.4248, + "step": 7343 + }, + { + "epoch": 0.6258201959948871, + "grad_norm": 18.5, + "learning_rate": 1.9527015044620814e-06, + "loss": 0.9084, + "step": 7344 + }, + { + "epoch": 0.6259054111631871, + "grad_norm": 17.25, + "learning_rate": 1.952562528061919e-06, + "loss": 0.897, + "step": 7345 + }, + { + "epoch": 0.625990626331487, + "grad_norm": 25.125, + "learning_rate": 1.952423538965746e-06, + "loss": 0.9288, + "step": 7346 + }, + { + "epoch": 0.626075841499787, + "grad_norm": 74.5, + "learning_rate": 1.952284537176073e-06, + "loss": 0.8996, + "step": 7347 + }, + { + "epoch": 0.626161056668087, + "grad_norm": 14.5625, + "learning_rate": 1.9521455226954124e-06, + "loss": 0.6105, + "step": 7348 + }, + { + "epoch": 0.6262462718363869, + "grad_norm": 9.125, + "learning_rate": 1.9520064955262756e-06, + "loss": 0.6165, + "step": 7349 + }, + { + "epoch": 0.6263314870046869, + "grad_norm": 17.0, + "learning_rate": 1.951867455671176e-06, + "loss": 0.8216, + "step": 7350 + }, + { + "epoch": 0.6264167021729868, + "grad_norm": 12.1875, + "learning_rate": 1.951728403132625e-06, + "loss": 0.528, + "step": 7351 + }, + { + "epoch": 0.6265019173412868, + "grad_norm": 13.5, + "learning_rate": 1.951589337913137e-06, + "loss": 0.582, + "step": 7352 + }, + { + "epoch": 0.6265871325095868, + "grad_norm": 15.0, + "learning_rate": 1.9514502600152236e-06, + "loss": 0.7162, + "step": 7353 + }, + { + "epoch": 0.6266723476778867, + "grad_norm": 18.75, + "learning_rate": 1.951311169441399e-06, + "loss": 0.9798, + "step": 7354 + }, + { + "epoch": 0.6267575628461867, + "grad_norm": 15.8125, + "learning_rate": 1.9511720661941757e-06, + "loss": 0.8901, + "step": 7355 + }, + { + "epoch": 0.6268427780144866, + "grad_norm": 12.1875, + "learning_rate": 1.9510329502760687e-06, + "loss": 0.5513, + "step": 7356 + }, + { + "epoch": 0.6269279931827866, + "grad_norm": 12.8125, + "learning_rate": 1.9508938216895913e-06, + "loss": 0.643, + "step": 7357 + }, + { + "epoch": 0.6270132083510865, + "grad_norm": 11.25, + "learning_rate": 1.9507546804372574e-06, + "loss": 0.6086, + "step": 7358 + }, + { + "epoch": 0.6270984235193865, + "grad_norm": 16.25, + "learning_rate": 1.950615526521582e-06, + "loss": 1.0132, + "step": 7359 + }, + { + "epoch": 0.6271836386876865, + "grad_norm": 13.5, + "learning_rate": 1.9504763599450792e-06, + "loss": 0.6273, + "step": 7360 + }, + { + "epoch": 0.6272688538559864, + "grad_norm": 14.5625, + "learning_rate": 1.9503371807102646e-06, + "loss": 0.5556, + "step": 7361 + }, + { + "epoch": 0.6273540690242864, + "grad_norm": 12.4375, + "learning_rate": 1.9501979888196533e-06, + "loss": 0.6749, + "step": 7362 + }, + { + "epoch": 0.6274392841925863, + "grad_norm": 18.0, + "learning_rate": 1.9500587842757594e-06, + "loss": 0.6568, + "step": 7363 + }, + { + "epoch": 0.6275244993608863, + "grad_norm": 11.375, + "learning_rate": 1.9499195670810993e-06, + "loss": 0.608, + "step": 7364 + }, + { + "epoch": 0.6276097145291862, + "grad_norm": 18.25, + "learning_rate": 1.949780337238189e-06, + "loss": 0.8802, + "step": 7365 + }, + { + "epoch": 0.6276949296974862, + "grad_norm": 14.6875, + "learning_rate": 1.9496410947495447e-06, + "loss": 0.5076, + "step": 7366 + }, + { + "epoch": 0.6277801448657861, + "grad_norm": 13.125, + "learning_rate": 1.949501839617682e-06, + "loss": 0.6984, + "step": 7367 + }, + { + "epoch": 0.6278653600340861, + "grad_norm": 16.375, + "learning_rate": 1.9493625718451176e-06, + "loss": 0.9003, + "step": 7368 + }, + { + "epoch": 0.6279505752023861, + "grad_norm": 13.625, + "learning_rate": 1.9492232914343685e-06, + "loss": 0.6419, + "step": 7369 + }, + { + "epoch": 0.628035790370686, + "grad_norm": 13.375, + "learning_rate": 1.949083998387951e-06, + "loss": 0.5162, + "step": 7370 + }, + { + "epoch": 0.628121005538986, + "grad_norm": 11.875, + "learning_rate": 1.948944692708383e-06, + "loss": 0.4872, + "step": 7371 + }, + { + "epoch": 0.6282062207072859, + "grad_norm": 14.875, + "learning_rate": 1.9488053743981814e-06, + "loss": 0.8946, + "step": 7372 + }, + { + "epoch": 0.6282914358755859, + "grad_norm": 23.25, + "learning_rate": 1.9486660434598638e-06, + "loss": 1.0278, + "step": 7373 + }, + { + "epoch": 0.6283766510438858, + "grad_norm": 16.125, + "learning_rate": 1.948526699895949e-06, + "loss": 0.929, + "step": 7374 + }, + { + "epoch": 0.6284618662121858, + "grad_norm": 22.375, + "learning_rate": 1.9483873437089536e-06, + "loss": 0.8232, + "step": 7375 + }, + { + "epoch": 0.6285470813804858, + "grad_norm": 13.5625, + "learning_rate": 1.948247974901397e-06, + "loss": 0.815, + "step": 7376 + }, + { + "epoch": 0.6286322965487857, + "grad_norm": 12.0, + "learning_rate": 1.948108593475797e-06, + "loss": 0.3755, + "step": 7377 + }, + { + "epoch": 0.6287175117170857, + "grad_norm": 17.75, + "learning_rate": 1.9479691994346727e-06, + "loss": 0.6934, + "step": 7378 + }, + { + "epoch": 0.6288027268853856, + "grad_norm": 15.75, + "learning_rate": 1.9478297927805433e-06, + "loss": 0.7856, + "step": 7379 + }, + { + "epoch": 0.6288879420536856, + "grad_norm": 14.3125, + "learning_rate": 1.947690373515928e-06, + "loss": 0.7932, + "step": 7380 + }, + { + "epoch": 0.6289731572219855, + "grad_norm": 17.25, + "learning_rate": 1.9475509416433457e-06, + "loss": 0.7672, + "step": 7381 + }, + { + "epoch": 0.6290583723902855, + "grad_norm": 10.5, + "learning_rate": 1.947411497165316e-06, + "loss": 0.3226, + "step": 7382 + }, + { + "epoch": 0.6291435875585855, + "grad_norm": 11.4375, + "learning_rate": 1.9472720400843604e-06, + "loss": 0.5668, + "step": 7383 + }, + { + "epoch": 0.6292288027268854, + "grad_norm": 17.5, + "learning_rate": 1.9471325704029973e-06, + "loss": 0.6698, + "step": 7384 + }, + { + "epoch": 0.6293140178951854, + "grad_norm": 15.0, + "learning_rate": 1.9469930881237474e-06, + "loss": 0.929, + "step": 7385 + }, + { + "epoch": 0.6293992330634853, + "grad_norm": 10.4375, + "learning_rate": 1.946853593249132e-06, + "loss": 0.3636, + "step": 7386 + }, + { + "epoch": 0.6294844482317853, + "grad_norm": 16.5, + "learning_rate": 1.946714085781671e-06, + "loss": 0.7777, + "step": 7387 + }, + { + "epoch": 0.6295696634000852, + "grad_norm": 15.9375, + "learning_rate": 1.946574565723886e-06, + "loss": 1.119, + "step": 7388 + }, + { + "epoch": 0.6296548785683852, + "grad_norm": 14.25, + "learning_rate": 1.9464350330782985e-06, + "loss": 0.7391, + "step": 7389 + }, + { + "epoch": 0.6297400937366852, + "grad_norm": 19.25, + "learning_rate": 1.9462954878474295e-06, + "loss": 0.9352, + "step": 7390 + }, + { + "epoch": 0.6298253089049851, + "grad_norm": 11.8125, + "learning_rate": 1.9461559300338005e-06, + "loss": 0.6283, + "step": 7391 + }, + { + "epoch": 0.6299105240732851, + "grad_norm": 15.75, + "learning_rate": 1.946016359639934e-06, + "loss": 0.4493, + "step": 7392 + }, + { + "epoch": 0.629995739241585, + "grad_norm": 17.125, + "learning_rate": 1.9458767766683517e-06, + "loss": 0.5668, + "step": 7393 + }, + { + "epoch": 0.630080954409885, + "grad_norm": 14.25, + "learning_rate": 1.945737181121577e-06, + "loss": 0.4691, + "step": 7394 + }, + { + "epoch": 0.6301661695781849, + "grad_norm": 16.0, + "learning_rate": 1.9455975730021308e-06, + "loss": 0.7759, + "step": 7395 + }, + { + "epoch": 0.6302513847464849, + "grad_norm": 14.125, + "learning_rate": 1.9454579523125376e-06, + "loss": 0.7293, + "step": 7396 + }, + { + "epoch": 0.6303365999147849, + "grad_norm": 15.125, + "learning_rate": 1.94531831905532e-06, + "loss": 0.926, + "step": 7397 + }, + { + "epoch": 0.6304218150830848, + "grad_norm": 13.75, + "learning_rate": 1.9451786732330006e-06, + "loss": 0.8237, + "step": 7398 + }, + { + "epoch": 0.6305070302513848, + "grad_norm": 15.25, + "learning_rate": 1.9450390148481043e-06, + "loss": 1.0485, + "step": 7399 + }, + { + "epoch": 0.6305922454196847, + "grad_norm": 14.5625, + "learning_rate": 1.944899343903154e-06, + "loss": 0.6306, + "step": 7400 + }, + { + "epoch": 0.6306774605879847, + "grad_norm": 14.25, + "learning_rate": 1.9447596604006735e-06, + "loss": 0.8967, + "step": 7401 + }, + { + "epoch": 0.6307626757562846, + "grad_norm": 14.375, + "learning_rate": 1.944619964343187e-06, + "loss": 0.561, + "step": 7402 + }, + { + "epoch": 0.6308478909245846, + "grad_norm": 12.8125, + "learning_rate": 1.9444802557332195e-06, + "loss": 0.6972, + "step": 7403 + }, + { + "epoch": 0.6309331060928846, + "grad_norm": 12.6875, + "learning_rate": 1.9443405345732957e-06, + "loss": 0.4426, + "step": 7404 + }, + { + "epoch": 0.6310183212611845, + "grad_norm": 22.375, + "learning_rate": 1.9442008008659395e-06, + "loss": 0.7587, + "step": 7405 + }, + { + "epoch": 0.6311035364294845, + "grad_norm": 15.4375, + "learning_rate": 1.944061054613677e-06, + "loss": 0.7279, + "step": 7406 + }, + { + "epoch": 0.6311887515977844, + "grad_norm": 15.875, + "learning_rate": 1.943921295819034e-06, + "loss": 0.5185, + "step": 7407 + }, + { + "epoch": 0.6312739667660844, + "grad_norm": 15.0625, + "learning_rate": 1.943781524484535e-06, + "loss": 0.7107, + "step": 7408 + }, + { + "epoch": 0.6313591819343843, + "grad_norm": 12.6875, + "learning_rate": 1.943641740612706e-06, + "loss": 0.6492, + "step": 7409 + }, + { + "epoch": 0.6314443971026843, + "grad_norm": 12.0, + "learning_rate": 1.9435019442060733e-06, + "loss": 0.5004, + "step": 7410 + }, + { + "epoch": 0.6315296122709843, + "grad_norm": 22.125, + "learning_rate": 1.9433621352671634e-06, + "loss": 1.1887, + "step": 7411 + }, + { + "epoch": 0.6316148274392842, + "grad_norm": 11.3125, + "learning_rate": 1.9432223137985024e-06, + "loss": 0.4793, + "step": 7412 + }, + { + "epoch": 0.6317000426075842, + "grad_norm": 10.0, + "learning_rate": 1.943082479802617e-06, + "loss": 0.4498, + "step": 7413 + }, + { + "epoch": 0.6317852577758841, + "grad_norm": 15.3125, + "learning_rate": 1.942942633282034e-06, + "loss": 0.8601, + "step": 7414 + }, + { + "epoch": 0.6318704729441841, + "grad_norm": 12.75, + "learning_rate": 1.9428027742392813e-06, + "loss": 0.4812, + "step": 7415 + }, + { + "epoch": 0.631955688112484, + "grad_norm": 13.25, + "learning_rate": 1.942662902676885e-06, + "loss": 0.7221, + "step": 7416 + }, + { + "epoch": 0.632040903280784, + "grad_norm": 14.3125, + "learning_rate": 1.942523018597374e-06, + "loss": 0.5319, + "step": 7417 + }, + { + "epoch": 0.632126118449084, + "grad_norm": 12.1875, + "learning_rate": 1.942383122003276e-06, + "loss": 0.6062, + "step": 7418 + }, + { + "epoch": 0.6322113336173839, + "grad_norm": 12.5625, + "learning_rate": 1.9422432128971187e-06, + "loss": 0.6049, + "step": 7419 + }, + { + "epoch": 0.6322965487856839, + "grad_norm": 13.75, + "learning_rate": 1.94210329128143e-06, + "loss": 0.8464, + "step": 7420 + }, + { + "epoch": 0.6323817639539838, + "grad_norm": 12.0, + "learning_rate": 1.9419633571587392e-06, + "loss": 0.4576, + "step": 7421 + }, + { + "epoch": 0.6324669791222838, + "grad_norm": 14.0, + "learning_rate": 1.9418234105315744e-06, + "loss": 0.7391, + "step": 7422 + }, + { + "epoch": 0.6325521942905837, + "grad_norm": 13.5, + "learning_rate": 1.9416834514024654e-06, + "loss": 0.5661, + "step": 7423 + }, + { + "epoch": 0.6326374094588837, + "grad_norm": 13.3125, + "learning_rate": 1.94154347977394e-06, + "loss": 0.706, + "step": 7424 + }, + { + "epoch": 0.6327226246271836, + "grad_norm": 14.125, + "learning_rate": 1.9414034956485293e-06, + "loss": 0.8299, + "step": 7425 + }, + { + "epoch": 0.6328078397954836, + "grad_norm": 15.6875, + "learning_rate": 1.9412634990287616e-06, + "loss": 0.9554, + "step": 7426 + }, + { + "epoch": 0.6328930549637836, + "grad_norm": 12.6875, + "learning_rate": 1.9411234899171685e-06, + "loss": 0.7038, + "step": 7427 + }, + { + "epoch": 0.6329782701320835, + "grad_norm": 18.375, + "learning_rate": 1.9409834683162777e-06, + "loss": 0.8385, + "step": 7428 + }, + { + "epoch": 0.6330634853003835, + "grad_norm": 13.5, + "learning_rate": 1.9408434342286213e-06, + "loss": 0.9287, + "step": 7429 + }, + { + "epoch": 0.6331487004686834, + "grad_norm": 14.4375, + "learning_rate": 1.9407033876567292e-06, + "loss": 0.7782, + "step": 7430 + }, + { + "epoch": 0.6332339156369834, + "grad_norm": 15.75, + "learning_rate": 1.9405633286031324e-06, + "loss": 0.566, + "step": 7431 + }, + { + "epoch": 0.6333191308052833, + "grad_norm": 12.125, + "learning_rate": 1.940423257070362e-06, + "loss": 0.53, + "step": 7432 + }, + { + "epoch": 0.6334043459735833, + "grad_norm": 15.9375, + "learning_rate": 1.940283173060949e-06, + "loss": 0.8393, + "step": 7433 + }, + { + "epoch": 0.6334895611418833, + "grad_norm": 18.5, + "learning_rate": 1.9401430765774247e-06, + "loss": 0.957, + "step": 7434 + }, + { + "epoch": 0.6335747763101832, + "grad_norm": 13.5, + "learning_rate": 1.940002967622321e-06, + "loss": 0.6762, + "step": 7435 + }, + { + "epoch": 0.6336599914784832, + "grad_norm": 13.3125, + "learning_rate": 1.93986284619817e-06, + "loss": 0.965, + "step": 7436 + }, + { + "epoch": 0.6337452066467831, + "grad_norm": 14.1875, + "learning_rate": 1.9397227123075035e-06, + "loss": 0.8258, + "step": 7437 + }, + { + "epoch": 0.6338304218150831, + "grad_norm": 11.1875, + "learning_rate": 1.9395825659528543e-06, + "loss": 0.5571, + "step": 7438 + }, + { + "epoch": 0.633915636983383, + "grad_norm": 10.8125, + "learning_rate": 1.9394424071367545e-06, + "loss": 0.4198, + "step": 7439 + }, + { + "epoch": 0.634000852151683, + "grad_norm": 15.625, + "learning_rate": 1.9393022358617374e-06, + "loss": 0.6657, + "step": 7440 + }, + { + "epoch": 0.634086067319983, + "grad_norm": 18.25, + "learning_rate": 1.9391620521303355e-06, + "loss": 0.9386, + "step": 7441 + }, + { + "epoch": 0.6341712824882829, + "grad_norm": 13.875, + "learning_rate": 1.9390218559450823e-06, + "loss": 0.3725, + "step": 7442 + }, + { + "epoch": 0.6342564976565829, + "grad_norm": 31.5, + "learning_rate": 1.9388816473085116e-06, + "loss": 1.049, + "step": 7443 + }, + { + "epoch": 0.6343417128248828, + "grad_norm": 11.125, + "learning_rate": 1.9387414262231562e-06, + "loss": 0.4952, + "step": 7444 + }, + { + "epoch": 0.6344269279931828, + "grad_norm": 17.625, + "learning_rate": 1.938601192691551e-06, + "loss": 0.9978, + "step": 7445 + }, + { + "epoch": 0.6345121431614827, + "grad_norm": 16.0, + "learning_rate": 1.93846094671623e-06, + "loss": 0.7123, + "step": 7446 + }, + { + "epoch": 0.6345973583297827, + "grad_norm": 20.75, + "learning_rate": 1.9383206882997275e-06, + "loss": 1.2092, + "step": 7447 + }, + { + "epoch": 0.6346825734980827, + "grad_norm": 11.625, + "learning_rate": 1.9381804174445775e-06, + "loss": 0.5536, + "step": 7448 + }, + { + "epoch": 0.6347677886663826, + "grad_norm": 12.0625, + "learning_rate": 1.9380401341533155e-06, + "loss": 0.5701, + "step": 7449 + }, + { + "epoch": 0.6348530038346826, + "grad_norm": 21.5, + "learning_rate": 1.9378998384284764e-06, + "loss": 0.8794, + "step": 7450 + }, + { + "epoch": 0.6349382190029825, + "grad_norm": 15.375, + "learning_rate": 1.9377595302725954e-06, + "loss": 0.8057, + "step": 7451 + }, + { + "epoch": 0.6350234341712825, + "grad_norm": 14.1875, + "learning_rate": 1.9376192096882083e-06, + "loss": 0.8704, + "step": 7452 + }, + { + "epoch": 0.6351086493395824, + "grad_norm": 14.9375, + "learning_rate": 1.9374788766778505e-06, + "loss": 0.8706, + "step": 7453 + }, + { + "epoch": 0.6351938645078824, + "grad_norm": 15.75, + "learning_rate": 1.9373385312440583e-06, + "loss": 0.9047, + "step": 7454 + }, + { + "epoch": 0.6352790796761824, + "grad_norm": 15.4375, + "learning_rate": 1.9371981733893674e-06, + "loss": 0.9157, + "step": 7455 + }, + { + "epoch": 0.6353642948444823, + "grad_norm": 18.0, + "learning_rate": 1.9370578031163145e-06, + "loss": 0.6883, + "step": 7456 + }, + { + "epoch": 0.6354495100127823, + "grad_norm": 19.375, + "learning_rate": 1.9369174204274362e-06, + "loss": 1.1392, + "step": 7457 + }, + { + "epoch": 0.6355347251810822, + "grad_norm": 10.375, + "learning_rate": 1.9367770253252694e-06, + "loss": 0.4064, + "step": 7458 + }, + { + "epoch": 0.6356199403493822, + "grad_norm": 14.3125, + "learning_rate": 1.936636617812351e-06, + "loss": 0.5448, + "step": 7459 + }, + { + "epoch": 0.6357051555176821, + "grad_norm": 18.5, + "learning_rate": 1.936496197891219e-06, + "loss": 0.9048, + "step": 7460 + }, + { + "epoch": 0.6357903706859821, + "grad_norm": 14.625, + "learning_rate": 1.93635576556441e-06, + "loss": 0.9042, + "step": 7461 + }, + { + "epoch": 0.635875585854282, + "grad_norm": 15.1875, + "learning_rate": 1.936215320834462e-06, + "loss": 0.6657, + "step": 7462 + }, + { + "epoch": 0.635960801022582, + "grad_norm": 12.5, + "learning_rate": 1.9360748637039136e-06, + "loss": 0.5172, + "step": 7463 + }, + { + "epoch": 0.636046016190882, + "grad_norm": 13.25, + "learning_rate": 1.9359343941753024e-06, + "loss": 0.629, + "step": 7464 + }, + { + "epoch": 0.6361312313591819, + "grad_norm": 11.9375, + "learning_rate": 1.9357939122511666e-06, + "loss": 0.7695, + "step": 7465 + }, + { + "epoch": 0.6362164465274819, + "grad_norm": 19.875, + "learning_rate": 1.935653417934045e-06, + "loss": 0.6998, + "step": 7466 + }, + { + "epoch": 0.6363016616957818, + "grad_norm": 13.3125, + "learning_rate": 1.935512911226477e-06, + "loss": 0.5618, + "step": 7467 + }, + { + "epoch": 0.6363868768640818, + "grad_norm": 22.25, + "learning_rate": 1.935372392131001e-06, + "loss": 0.9994, + "step": 7468 + }, + { + "epoch": 0.6364720920323818, + "grad_norm": 13.0625, + "learning_rate": 1.9352318606501576e-06, + "loss": 0.5832, + "step": 7469 + }, + { + "epoch": 0.6365573072006817, + "grad_norm": 19.375, + "learning_rate": 1.9350913167864842e-06, + "loss": 1.0666, + "step": 7470 + }, + { + "epoch": 0.6366425223689817, + "grad_norm": 19.875, + "learning_rate": 1.9349507605425227e-06, + "loss": 1.0364, + "step": 7471 + }, + { + "epoch": 0.6367277375372816, + "grad_norm": 14.625, + "learning_rate": 1.934810191920812e-06, + "loss": 0.9384, + "step": 7472 + }, + { + "epoch": 0.6368129527055816, + "grad_norm": 19.5, + "learning_rate": 1.9346696109238923e-06, + "loss": 0.7831, + "step": 7473 + }, + { + "epoch": 0.6368981678738815, + "grad_norm": 15.375, + "learning_rate": 1.9345290175543043e-06, + "loss": 0.8197, + "step": 7474 + }, + { + "epoch": 0.6369833830421815, + "grad_norm": 14.125, + "learning_rate": 1.9343884118145883e-06, + "loss": 0.4731, + "step": 7475 + }, + { + "epoch": 0.6370685982104815, + "grad_norm": 13.5625, + "learning_rate": 1.934247793707286e-06, + "loss": 0.6042, + "step": 7476 + }, + { + "epoch": 0.6371538133787814, + "grad_norm": 22.375, + "learning_rate": 1.934107163234938e-06, + "loss": 1.2094, + "step": 7477 + }, + { + "epoch": 0.6372390285470814, + "grad_norm": 11.625, + "learning_rate": 1.9339665204000856e-06, + "loss": 0.5361, + "step": 7478 + }, + { + "epoch": 0.6373242437153813, + "grad_norm": 13.75, + "learning_rate": 1.93382586520527e-06, + "loss": 0.7193, + "step": 7479 + }, + { + "epoch": 0.6374094588836813, + "grad_norm": 15.5, + "learning_rate": 1.9336851976530342e-06, + "loss": 1.0151, + "step": 7480 + }, + { + "epoch": 0.6374946740519812, + "grad_norm": 13.4375, + "learning_rate": 1.933544517745918e-06, + "loss": 0.7259, + "step": 7481 + }, + { + "epoch": 0.6375798892202812, + "grad_norm": 14.6875, + "learning_rate": 1.933403825486466e-06, + "loss": 0.2779, + "step": 7482 + }, + { + "epoch": 0.6376651043885811, + "grad_norm": 19.0, + "learning_rate": 1.9332631208772197e-06, + "loss": 0.8455, + "step": 7483 + }, + { + "epoch": 0.6377503195568811, + "grad_norm": 13.5, + "learning_rate": 1.9331224039207217e-06, + "loss": 0.6364, + "step": 7484 + }, + { + "epoch": 0.6378355347251811, + "grad_norm": 15.8125, + "learning_rate": 1.9329816746195146e-06, + "loss": 0.7591, + "step": 7485 + }, + { + "epoch": 0.637920749893481, + "grad_norm": 12.25, + "learning_rate": 1.932840932976142e-06, + "loss": 0.6894, + "step": 7486 + }, + { + "epoch": 0.638005965061781, + "grad_norm": 14.0625, + "learning_rate": 1.932700178993147e-06, + "loss": 0.5843, + "step": 7487 + }, + { + "epoch": 0.6380911802300809, + "grad_norm": 13.3125, + "learning_rate": 1.9325594126730736e-06, + "loss": 0.5927, + "step": 7488 + }, + { + "epoch": 0.6381763953983809, + "grad_norm": 17.0, + "learning_rate": 1.932418634018465e-06, + "loss": 0.5767, + "step": 7489 + }, + { + "epoch": 0.6382616105666808, + "grad_norm": 11.75, + "learning_rate": 1.932277843031865e-06, + "loss": 0.4747, + "step": 7490 + }, + { + "epoch": 0.6383468257349808, + "grad_norm": 15.375, + "learning_rate": 1.932137039715819e-06, + "loss": 0.6273, + "step": 7491 + }, + { + "epoch": 0.6384320409032808, + "grad_norm": 22.25, + "learning_rate": 1.931996224072871e-06, + "loss": 0.851, + "step": 7492 + }, + { + "epoch": 0.6385172560715807, + "grad_norm": 11.25, + "learning_rate": 1.9318553961055645e-06, + "loss": 0.5278, + "step": 7493 + }, + { + "epoch": 0.6386024712398807, + "grad_norm": 16.375, + "learning_rate": 1.9317145558164458e-06, + "loss": 0.989, + "step": 7494 + }, + { + "epoch": 0.6386876864081806, + "grad_norm": 13.0625, + "learning_rate": 1.9315737032080595e-06, + "loss": 0.7262, + "step": 7495 + }, + { + "epoch": 0.6387729015764806, + "grad_norm": 18.5, + "learning_rate": 1.931432838282951e-06, + "loss": 1.0054, + "step": 7496 + }, + { + "epoch": 0.6388581167447805, + "grad_norm": 17.875, + "learning_rate": 1.931291961043666e-06, + "loss": 0.7548, + "step": 7497 + }, + { + "epoch": 0.6389433319130805, + "grad_norm": 20.0, + "learning_rate": 1.93115107149275e-06, + "loss": 0.9563, + "step": 7498 + }, + { + "epoch": 0.6390285470813805, + "grad_norm": 15.1875, + "learning_rate": 1.9310101696327494e-06, + "loss": 0.7491, + "step": 7499 + }, + { + "epoch": 0.6391137622496804, + "grad_norm": 22.25, + "learning_rate": 1.93086925546621e-06, + "loss": 0.6004, + "step": 7500 + }, + { + "epoch": 0.6391989774179804, + "grad_norm": 12.375, + "learning_rate": 1.9307283289956786e-06, + "loss": 0.5954, + "step": 7501 + }, + { + "epoch": 0.6392841925862803, + "grad_norm": 10.875, + "learning_rate": 1.9305873902237014e-06, + "loss": 0.3651, + "step": 7502 + }, + { + "epoch": 0.6393694077545803, + "grad_norm": 12.9375, + "learning_rate": 1.930446439152826e-06, + "loss": 0.6012, + "step": 7503 + }, + { + "epoch": 0.6394546229228802, + "grad_norm": 22.375, + "learning_rate": 1.9303054757855994e-06, + "loss": 0.9328, + "step": 7504 + }, + { + "epoch": 0.6395398380911802, + "grad_norm": 13.8125, + "learning_rate": 1.9301645001245686e-06, + "loss": 0.5892, + "step": 7505 + }, + { + "epoch": 0.6396250532594802, + "grad_norm": 15.0, + "learning_rate": 1.9300235121722813e-06, + "loss": 0.7726, + "step": 7506 + }, + { + "epoch": 0.6397102684277801, + "grad_norm": 13.25, + "learning_rate": 1.9298825119312853e-06, + "loss": 0.8592, + "step": 7507 + }, + { + "epoch": 0.6397954835960801, + "grad_norm": 17.625, + "learning_rate": 1.929741499404129e-06, + "loss": 0.5337, + "step": 7508 + }, + { + "epoch": 0.63988069876438, + "grad_norm": 12.75, + "learning_rate": 1.9296004745933596e-06, + "loss": 0.4694, + "step": 7509 + }, + { + "epoch": 0.63996591393268, + "grad_norm": 31.875, + "learning_rate": 1.9294594375015266e-06, + "loss": 1.1662, + "step": 7510 + }, + { + "epoch": 0.6400511291009799, + "grad_norm": 29.125, + "learning_rate": 1.9293183881311784e-06, + "loss": 1.1891, + "step": 7511 + }, + { + "epoch": 0.6401363442692799, + "grad_norm": 11.1875, + "learning_rate": 1.929177326484864e-06, + "loss": 0.5254, + "step": 7512 + }, + { + "epoch": 0.6402215594375799, + "grad_norm": 12.125, + "learning_rate": 1.9290362525651318e-06, + "loss": 0.9344, + "step": 7513 + }, + { + "epoch": 0.6403067746058798, + "grad_norm": 14.875, + "learning_rate": 1.9288951663745316e-06, + "loss": 0.7495, + "step": 7514 + }, + { + "epoch": 0.6403919897741798, + "grad_norm": 12.5, + "learning_rate": 1.9287540679156134e-06, + "loss": 0.5234, + "step": 7515 + }, + { + "epoch": 0.6404772049424797, + "grad_norm": 12.125, + "learning_rate": 1.9286129571909267e-06, + "loss": 0.7819, + "step": 7516 + }, + { + "epoch": 0.6405624201107797, + "grad_norm": 13.5, + "learning_rate": 1.928471834203022e-06, + "loss": 0.424, + "step": 7517 + }, + { + "epoch": 0.6406476352790796, + "grad_norm": 24.125, + "learning_rate": 1.928330698954448e-06, + "loss": 0.8901, + "step": 7518 + }, + { + "epoch": 0.6407328504473796, + "grad_norm": 15.6875, + "learning_rate": 1.9281895514477568e-06, + "loss": 0.6623, + "step": 7519 + }, + { + "epoch": 0.6408180656156796, + "grad_norm": 17.375, + "learning_rate": 1.928048391685498e-06, + "loss": 0.6539, + "step": 7520 + }, + { + "epoch": 0.6409032807839795, + "grad_norm": 19.0, + "learning_rate": 1.927907219670223e-06, + "loss": 0.9933, + "step": 7521 + }, + { + "epoch": 0.6409884959522795, + "grad_norm": 16.625, + "learning_rate": 1.927766035404483e-06, + "loss": 0.7251, + "step": 7522 + }, + { + "epoch": 0.6410737111205794, + "grad_norm": 12.5, + "learning_rate": 1.9276248388908293e-06, + "loss": 0.6019, + "step": 7523 + }, + { + "epoch": 0.6411589262888794, + "grad_norm": 10.75, + "learning_rate": 1.9274836301318135e-06, + "loss": 0.5628, + "step": 7524 + }, + { + "epoch": 0.6412441414571793, + "grad_norm": 11.25, + "learning_rate": 1.9273424091299867e-06, + "loss": 0.5809, + "step": 7525 + }, + { + "epoch": 0.6413293566254793, + "grad_norm": 21.25, + "learning_rate": 1.9272011758879015e-06, + "loss": 0.9849, + "step": 7526 + }, + { + "epoch": 0.6414145717937793, + "grad_norm": 16.0, + "learning_rate": 1.9270599304081105e-06, + "loss": 0.5735, + "step": 7527 + }, + { + "epoch": 0.6414997869620792, + "grad_norm": 14.125, + "learning_rate": 1.9269186726931653e-06, + "loss": 0.5916, + "step": 7528 + }, + { + "epoch": 0.6415850021303792, + "grad_norm": 14.8125, + "learning_rate": 1.9267774027456193e-06, + "loss": 0.8224, + "step": 7529 + }, + { + "epoch": 0.6416702172986791, + "grad_norm": 13.5625, + "learning_rate": 1.926636120568025e-06, + "loss": 0.6435, + "step": 7530 + }, + { + "epoch": 0.6417554324669791, + "grad_norm": 15.9375, + "learning_rate": 1.926494826162936e-06, + "loss": 0.5791, + "step": 7531 + }, + { + "epoch": 0.641840647635279, + "grad_norm": 13.25, + "learning_rate": 1.9263535195329044e-06, + "loss": 0.741, + "step": 7532 + }, + { + "epoch": 0.641925862803579, + "grad_norm": 16.75, + "learning_rate": 1.926212200680485e-06, + "loss": 1.0166, + "step": 7533 + }, + { + "epoch": 0.642011077971879, + "grad_norm": 13.9375, + "learning_rate": 1.9260708696082312e-06, + "loss": 0.9037, + "step": 7534 + }, + { + "epoch": 0.6420962931401789, + "grad_norm": 18.5, + "learning_rate": 1.925929526318697e-06, + "loss": 0.8912, + "step": 7535 + }, + { + "epoch": 0.6421815083084789, + "grad_norm": 17.75, + "learning_rate": 1.925788170814436e-06, + "loss": 1.0491, + "step": 7536 + }, + { + "epoch": 0.6422667234767788, + "grad_norm": 14.625, + "learning_rate": 1.9256468030980037e-06, + "loss": 0.7845, + "step": 7537 + }, + { + "epoch": 0.6423519386450788, + "grad_norm": 18.875, + "learning_rate": 1.9255054231719544e-06, + "loss": 0.9729, + "step": 7538 + }, + { + "epoch": 0.6424371538133787, + "grad_norm": 14.5625, + "learning_rate": 1.9253640310388427e-06, + "loss": 1.1976, + "step": 7539 + }, + { + "epoch": 0.6425223689816787, + "grad_norm": 21.5, + "learning_rate": 1.9252226267012238e-06, + "loss": 0.7589, + "step": 7540 + }, + { + "epoch": 0.6426075841499787, + "grad_norm": 22.25, + "learning_rate": 1.9250812101616527e-06, + "loss": 0.6298, + "step": 7541 + }, + { + "epoch": 0.6426927993182786, + "grad_norm": 30.625, + "learning_rate": 1.9249397814226857e-06, + "loss": 1.174, + "step": 7542 + }, + { + "epoch": 0.6427780144865786, + "grad_norm": 10.6875, + "learning_rate": 1.924798340486878e-06, + "loss": 0.4541, + "step": 7543 + }, + { + "epoch": 0.6428632296548785, + "grad_norm": 13.25, + "learning_rate": 1.9246568873567855e-06, + "loss": 0.4666, + "step": 7544 + }, + { + "epoch": 0.6429484448231785, + "grad_norm": 16.625, + "learning_rate": 1.924515422034965e-06, + "loss": 0.871, + "step": 7545 + }, + { + "epoch": 0.6430336599914784, + "grad_norm": 13.6875, + "learning_rate": 1.9243739445239725e-06, + "loss": 0.6814, + "step": 7546 + }, + { + "epoch": 0.6431188751597784, + "grad_norm": 13.5, + "learning_rate": 1.924232454826365e-06, + "loss": 0.7081, + "step": 7547 + }, + { + "epoch": 0.6432040903280783, + "grad_norm": 19.125, + "learning_rate": 1.924090952944698e-06, + "loss": 1.1679, + "step": 7548 + }, + { + "epoch": 0.6432893054963783, + "grad_norm": 20.375, + "learning_rate": 1.9239494388815304e-06, + "loss": 0.8666, + "step": 7549 + }, + { + "epoch": 0.6433745206646783, + "grad_norm": 13.9375, + "learning_rate": 1.923807912639419e-06, + "loss": 0.5498, + "step": 7550 + }, + { + "epoch": 0.6434597358329782, + "grad_norm": 16.125, + "learning_rate": 1.9236663742209203e-06, + "loss": 0.8145, + "step": 7551 + }, + { + "epoch": 0.6435449510012782, + "grad_norm": 12.9375, + "learning_rate": 1.923524823628593e-06, + "loss": 0.4805, + "step": 7552 + }, + { + "epoch": 0.6436301661695781, + "grad_norm": 14.875, + "learning_rate": 1.9233832608649955e-06, + "loss": 0.7224, + "step": 7553 + }, + { + "epoch": 0.6437153813378781, + "grad_norm": 15.8125, + "learning_rate": 1.9232416859326845e-06, + "loss": 0.7638, + "step": 7554 + }, + { + "epoch": 0.643800596506178, + "grad_norm": 15.125, + "learning_rate": 1.92310009883422e-06, + "loss": 0.7241, + "step": 7555 + }, + { + "epoch": 0.643885811674478, + "grad_norm": 8.3125, + "learning_rate": 1.922958499572159e-06, + "loss": 0.5145, + "step": 7556 + }, + { + "epoch": 0.643971026842778, + "grad_norm": 39.5, + "learning_rate": 1.9228168881490616e-06, + "loss": 0.9358, + "step": 7557 + }, + { + "epoch": 0.6440562420110779, + "grad_norm": 14.1875, + "learning_rate": 1.922675264567487e-06, + "loss": 0.7633, + "step": 7558 + }, + { + "epoch": 0.6441414571793779, + "grad_norm": 14.6875, + "learning_rate": 1.922533628829993e-06, + "loss": 0.9419, + "step": 7559 + }, + { + "epoch": 0.6442266723476778, + "grad_norm": 17.625, + "learning_rate": 1.9223919809391407e-06, + "loss": 0.7895, + "step": 7560 + }, + { + "epoch": 0.6443118875159779, + "grad_norm": 14.125, + "learning_rate": 1.922250320897489e-06, + "loss": 0.7973, + "step": 7561 + }, + { + "epoch": 0.6443971026842779, + "grad_norm": 12.75, + "learning_rate": 1.922108648707598e-06, + "loss": 0.7334, + "step": 7562 + }, + { + "epoch": 0.6444823178525778, + "grad_norm": 13.0, + "learning_rate": 1.9219669643720275e-06, + "loss": 0.5542, + "step": 7563 + }, + { + "epoch": 0.6445675330208778, + "grad_norm": 13.6875, + "learning_rate": 1.921825267893339e-06, + "loss": 0.6766, + "step": 7564 + }, + { + "epoch": 0.6446527481891777, + "grad_norm": 14.375, + "learning_rate": 1.9216835592740917e-06, + "loss": 0.7353, + "step": 7565 + }, + { + "epoch": 0.6447379633574777, + "grad_norm": 17.875, + "learning_rate": 1.9215418385168473e-06, + "loss": 1.0316, + "step": 7566 + }, + { + "epoch": 0.6448231785257776, + "grad_norm": 15.125, + "learning_rate": 1.9214001056241665e-06, + "loss": 0.5764, + "step": 7567 + }, + { + "epoch": 0.6449083936940776, + "grad_norm": 17.125, + "learning_rate": 1.9212583605986103e-06, + "loss": 0.8525, + "step": 7568 + }, + { + "epoch": 0.6449936088623776, + "grad_norm": 11.125, + "learning_rate": 1.9211166034427415e-06, + "loss": 0.3902, + "step": 7569 + }, + { + "epoch": 0.6450788240306775, + "grad_norm": 14.4375, + "learning_rate": 1.92097483415912e-06, + "loss": 0.6017, + "step": 7570 + }, + { + "epoch": 0.6451640391989775, + "grad_norm": 17.75, + "learning_rate": 1.920833052750309e-06, + "loss": 1.0817, + "step": 7571 + }, + { + "epoch": 0.6452492543672774, + "grad_norm": 18.25, + "learning_rate": 1.9206912592188693e-06, + "loss": 0.5975, + "step": 7572 + }, + { + "epoch": 0.6453344695355774, + "grad_norm": 15.375, + "learning_rate": 1.9205494535673653e-06, + "loss": 0.5062, + "step": 7573 + }, + { + "epoch": 0.6454196847038773, + "grad_norm": 14.4375, + "learning_rate": 1.9204076357983575e-06, + "loss": 0.51, + "step": 7574 + }, + { + "epoch": 0.6455048998721773, + "grad_norm": 10.6875, + "learning_rate": 1.9202658059144098e-06, + "loss": 0.3104, + "step": 7575 + }, + { + "epoch": 0.6455901150404773, + "grad_norm": 14.1875, + "learning_rate": 1.9201239639180853e-06, + "loss": 0.5224, + "step": 7576 + }, + { + "epoch": 0.6456753302087772, + "grad_norm": 21.75, + "learning_rate": 1.9199821098119468e-06, + "loss": 0.8429, + "step": 7577 + }, + { + "epoch": 0.6457605453770772, + "grad_norm": 20.875, + "learning_rate": 1.9198402435985573e-06, + "loss": 0.6916, + "step": 7578 + }, + { + "epoch": 0.6458457605453771, + "grad_norm": 14.3125, + "learning_rate": 1.9196983652804817e-06, + "loss": 0.5824, + "step": 7579 + }, + { + "epoch": 0.6459309757136771, + "grad_norm": 16.5, + "learning_rate": 1.9195564748602828e-06, + "loss": 0.6706, + "step": 7580 + }, + { + "epoch": 0.646016190881977, + "grad_norm": 17.125, + "learning_rate": 1.9194145723405256e-06, + "loss": 0.9729, + "step": 7581 + }, + { + "epoch": 0.646101406050277, + "grad_norm": 17.0, + "learning_rate": 1.9192726577237738e-06, + "loss": 1.0382, + "step": 7582 + }, + { + "epoch": 0.646186621218577, + "grad_norm": 16.375, + "learning_rate": 1.9191307310125913e-06, + "loss": 1.0176, + "step": 7583 + }, + { + "epoch": 0.6462718363868769, + "grad_norm": 14.25, + "learning_rate": 1.9189887922095443e-06, + "loss": 0.7483, + "step": 7584 + }, + { + "epoch": 0.6463570515551769, + "grad_norm": 12.3125, + "learning_rate": 1.918846841317197e-06, + "loss": 0.4302, + "step": 7585 + }, + { + "epoch": 0.6464422667234768, + "grad_norm": 11.3125, + "learning_rate": 1.9187048783381144e-06, + "loss": 0.3955, + "step": 7586 + }, + { + "epoch": 0.6465274818917768, + "grad_norm": 17.875, + "learning_rate": 1.9185629032748626e-06, + "loss": 0.904, + "step": 7587 + }, + { + "epoch": 0.6466126970600767, + "grad_norm": 15.6875, + "learning_rate": 1.918420916130007e-06, + "loss": 0.6171, + "step": 7588 + }, + { + "epoch": 0.6466979122283767, + "grad_norm": 17.75, + "learning_rate": 1.918278916906113e-06, + "loss": 1.0285, + "step": 7589 + }, + { + "epoch": 0.6467831273966766, + "grad_norm": 16.25, + "learning_rate": 1.9181369056057474e-06, + "loss": 0.7404, + "step": 7590 + }, + { + "epoch": 0.6468683425649766, + "grad_norm": 13.25, + "learning_rate": 1.9179948822314754e-06, + "loss": 0.4816, + "step": 7591 + }, + { + "epoch": 0.6469535577332766, + "grad_norm": 11.75, + "learning_rate": 1.917852846785864e-06, + "loss": 0.5135, + "step": 7592 + }, + { + "epoch": 0.6470387729015765, + "grad_norm": 12.4375, + "learning_rate": 1.917710799271481e-06, + "loss": 0.462, + "step": 7593 + }, + { + "epoch": 0.6471239880698765, + "grad_norm": 17.375, + "learning_rate": 1.917568739690892e-06, + "loss": 0.7491, + "step": 7594 + }, + { + "epoch": 0.6472092032381764, + "grad_norm": 13.5625, + "learning_rate": 1.917426668046664e-06, + "loss": 0.848, + "step": 7595 + }, + { + "epoch": 0.6472944184064764, + "grad_norm": 11.5, + "learning_rate": 1.9172845843413655e-06, + "loss": 0.4944, + "step": 7596 + }, + { + "epoch": 0.6473796335747763, + "grad_norm": 11.1875, + "learning_rate": 1.917142488577564e-06, + "loss": 0.4256, + "step": 7597 + }, + { + "epoch": 0.6474648487430763, + "grad_norm": 15.8125, + "learning_rate": 1.917000380757826e-06, + "loss": 1.0364, + "step": 7598 + }, + { + "epoch": 0.6475500639113763, + "grad_norm": 14.0, + "learning_rate": 1.916858260884721e-06, + "loss": 0.7619, + "step": 7599 + }, + { + "epoch": 0.6476352790796762, + "grad_norm": 17.0, + "learning_rate": 1.9167161289608164e-06, + "loss": 0.8506, + "step": 7600 + }, + { + "epoch": 0.6477204942479762, + "grad_norm": 24.375, + "learning_rate": 1.916573984988681e-06, + "loss": 0.9152, + "step": 7601 + }, + { + "epoch": 0.6478057094162761, + "grad_norm": 11.625, + "learning_rate": 1.9164318289708834e-06, + "loss": 0.5738, + "step": 7602 + }, + { + "epoch": 0.6478909245845761, + "grad_norm": 17.5, + "learning_rate": 1.9162896609099924e-06, + "loss": 1.0527, + "step": 7603 + }, + { + "epoch": 0.647976139752876, + "grad_norm": 17.0, + "learning_rate": 1.9161474808085773e-06, + "loss": 0.7811, + "step": 7604 + }, + { + "epoch": 0.648061354921176, + "grad_norm": 15.375, + "learning_rate": 1.916005288669207e-06, + "loss": 1.0938, + "step": 7605 + }, + { + "epoch": 0.648146570089476, + "grad_norm": 10.875, + "learning_rate": 1.915863084494452e-06, + "loss": 0.3323, + "step": 7606 + }, + { + "epoch": 0.6482317852577759, + "grad_norm": 14.5625, + "learning_rate": 1.915720868286881e-06, + "loss": 0.4926, + "step": 7607 + }, + { + "epoch": 0.6483170004260759, + "grad_norm": 13.75, + "learning_rate": 1.915578640049065e-06, + "loss": 0.8232, + "step": 7608 + }, + { + "epoch": 0.6484022155943758, + "grad_norm": 19.125, + "learning_rate": 1.9154363997835736e-06, + "loss": 0.8178, + "step": 7609 + }, + { + "epoch": 0.6484874307626758, + "grad_norm": 14.9375, + "learning_rate": 1.915294147492977e-06, + "loss": 1.0978, + "step": 7610 + }, + { + "epoch": 0.6485726459309757, + "grad_norm": 14.3125, + "learning_rate": 1.9151518831798462e-06, + "loss": 0.8506, + "step": 7611 + }, + { + "epoch": 0.6486578610992757, + "grad_norm": 15.8125, + "learning_rate": 1.915009606846752e-06, + "loss": 0.9594, + "step": 7612 + }, + { + "epoch": 0.6487430762675757, + "grad_norm": 16.0, + "learning_rate": 1.914867318496266e-06, + "loss": 0.8196, + "step": 7613 + }, + { + "epoch": 0.6488282914358756, + "grad_norm": 12.5625, + "learning_rate": 1.914725018130958e-06, + "loss": 0.6289, + "step": 7614 + }, + { + "epoch": 0.6489135066041756, + "grad_norm": 19.0, + "learning_rate": 1.914582705753401e-06, + "loss": 0.9905, + "step": 7615 + }, + { + "epoch": 0.6489987217724755, + "grad_norm": 11.625, + "learning_rate": 1.9144403813661665e-06, + "loss": 0.5533, + "step": 7616 + }, + { + "epoch": 0.6490839369407755, + "grad_norm": 19.625, + "learning_rate": 1.9142980449718256e-06, + "loss": 0.7281, + "step": 7617 + }, + { + "epoch": 0.6491691521090754, + "grad_norm": 13.25, + "learning_rate": 1.9141556965729518e-06, + "loss": 0.4316, + "step": 7618 + }, + { + "epoch": 0.6492543672773754, + "grad_norm": 15.3125, + "learning_rate": 1.9140133361721157e-06, + "loss": 0.8764, + "step": 7619 + }, + { + "epoch": 0.6493395824456754, + "grad_norm": 13.125, + "learning_rate": 1.9138709637718916e-06, + "loss": 1.0729, + "step": 7620 + }, + { + "epoch": 0.6494247976139753, + "grad_norm": 13.9375, + "learning_rate": 1.913728579374851e-06, + "loss": 0.4932, + "step": 7621 + }, + { + "epoch": 0.6495100127822753, + "grad_norm": 13.9375, + "learning_rate": 1.9135861829835686e-06, + "loss": 0.6855, + "step": 7622 + }, + { + "epoch": 0.6495952279505752, + "grad_norm": 18.625, + "learning_rate": 1.9134437746006158e-06, + "loss": 0.854, + "step": 7623 + }, + { + "epoch": 0.6496804431188752, + "grad_norm": 23.0, + "learning_rate": 1.9133013542285668e-06, + "loss": 1.3407, + "step": 7624 + }, + { + "epoch": 0.6497656582871751, + "grad_norm": 14.1875, + "learning_rate": 1.9131589218699957e-06, + "loss": 0.8345, + "step": 7625 + }, + { + "epoch": 0.6498508734554751, + "grad_norm": 17.25, + "learning_rate": 1.9130164775274763e-06, + "loss": 0.7151, + "step": 7626 + }, + { + "epoch": 0.649936088623775, + "grad_norm": 19.625, + "learning_rate": 1.9128740212035814e-06, + "loss": 1.013, + "step": 7627 + }, + { + "epoch": 0.650021303792075, + "grad_norm": 13.125, + "learning_rate": 1.9127315529008873e-06, + "loss": 0.8041, + "step": 7628 + }, + { + "epoch": 0.650106518960375, + "grad_norm": 12.8125, + "learning_rate": 1.912589072621967e-06, + "loss": 0.5704, + "step": 7629 + }, + { + "epoch": 0.6501917341286749, + "grad_norm": 12.375, + "learning_rate": 1.9124465803693962e-06, + "loss": 0.5162, + "step": 7630 + }, + { + "epoch": 0.6502769492969749, + "grad_norm": 16.375, + "learning_rate": 1.9123040761457497e-06, + "loss": 0.6625, + "step": 7631 + }, + { + "epoch": 0.6503621644652748, + "grad_norm": 10.6875, + "learning_rate": 1.912161559953602e-06, + "loss": 0.407, + "step": 7632 + }, + { + "epoch": 0.6504473796335748, + "grad_norm": 12.5625, + "learning_rate": 1.9120190317955293e-06, + "loss": 0.5644, + "step": 7633 + }, + { + "epoch": 0.6505325948018748, + "grad_norm": 15.3125, + "learning_rate": 1.911876491674107e-06, + "loss": 0.6873, + "step": 7634 + }, + { + "epoch": 0.6506178099701747, + "grad_norm": 14.5625, + "learning_rate": 1.9117339395919115e-06, + "loss": 0.7525, + "step": 7635 + }, + { + "epoch": 0.6507030251384747, + "grad_norm": 16.75, + "learning_rate": 1.9115913755515173e-06, + "loss": 0.9009, + "step": 7636 + }, + { + "epoch": 0.6507882403067746, + "grad_norm": 12.5, + "learning_rate": 1.9114487995555025e-06, + "loss": 0.749, + "step": 7637 + }, + { + "epoch": 0.6508734554750746, + "grad_norm": 16.125, + "learning_rate": 1.9113062116064426e-06, + "loss": 1.1835, + "step": 7638 + }, + { + "epoch": 0.6509586706433745, + "grad_norm": 15.3125, + "learning_rate": 1.9111636117069145e-06, + "loss": 0.9973, + "step": 7639 + }, + { + "epoch": 0.6510438858116745, + "grad_norm": 14.125, + "learning_rate": 1.911020999859494e-06, + "loss": 0.7477, + "step": 7640 + }, + { + "epoch": 0.6511291009799745, + "grad_norm": 27.25, + "learning_rate": 1.9108783760667607e-06, + "loss": 0.8946, + "step": 7641 + }, + { + "epoch": 0.6512143161482744, + "grad_norm": 14.3125, + "learning_rate": 1.91073574033129e-06, + "loss": 0.8122, + "step": 7642 + }, + { + "epoch": 0.6512995313165744, + "grad_norm": 11.0, + "learning_rate": 1.9105930926556606e-06, + "loss": 0.4939, + "step": 7643 + }, + { + "epoch": 0.6513847464848743, + "grad_norm": 12.1875, + "learning_rate": 1.910450433042449e-06, + "loss": 0.6006, + "step": 7644 + }, + { + "epoch": 0.6514699616531743, + "grad_norm": 15.6875, + "learning_rate": 1.910307761494234e-06, + "loss": 0.9568, + "step": 7645 + }, + { + "epoch": 0.6515551768214742, + "grad_norm": 69.5, + "learning_rate": 1.9101650780135937e-06, + "loss": 0.9464, + "step": 7646 + }, + { + "epoch": 0.6516403919897742, + "grad_norm": 34.0, + "learning_rate": 1.9100223826031074e-06, + "loss": 0.715, + "step": 7647 + }, + { + "epoch": 0.6517256071580741, + "grad_norm": 11.6875, + "learning_rate": 1.909879675265352e-06, + "loss": 0.4619, + "step": 7648 + }, + { + "epoch": 0.6518108223263741, + "grad_norm": 24.5, + "learning_rate": 1.9097369560029078e-06, + "loss": 1.0209, + "step": 7649 + }, + { + "epoch": 0.6518960374946741, + "grad_norm": 14.375, + "learning_rate": 1.9095942248183534e-06, + "loss": 0.5722, + "step": 7650 + }, + { + "epoch": 0.651981252662974, + "grad_norm": 18.25, + "learning_rate": 1.909451481714268e-06, + "loss": 0.7237, + "step": 7651 + }, + { + "epoch": 0.652066467831274, + "grad_norm": 12.5625, + "learning_rate": 1.909308726693231e-06, + "loss": 0.5568, + "step": 7652 + }, + { + "epoch": 0.6521516829995739, + "grad_norm": 12.9375, + "learning_rate": 1.9091659597578223e-06, + "loss": 0.4262, + "step": 7653 + }, + { + "epoch": 0.6522368981678739, + "grad_norm": 13.75, + "learning_rate": 1.9090231809106223e-06, + "loss": 0.6109, + "step": 7654 + }, + { + "epoch": 0.6523221133361738, + "grad_norm": 13.1875, + "learning_rate": 1.9088803901542103e-06, + "loss": 0.5379, + "step": 7655 + }, + { + "epoch": 0.6524073285044738, + "grad_norm": 13.3125, + "learning_rate": 1.9087375874911674e-06, + "loss": 0.6077, + "step": 7656 + }, + { + "epoch": 0.6524925436727738, + "grad_norm": 16.75, + "learning_rate": 1.9085947729240735e-06, + "loss": 0.8948, + "step": 7657 + }, + { + "epoch": 0.6525777588410737, + "grad_norm": 16.625, + "learning_rate": 1.9084519464555106e-06, + "loss": 0.8755, + "step": 7658 + }, + { + "epoch": 0.6526629740093737, + "grad_norm": 14.8125, + "learning_rate": 1.908309108088058e-06, + "loss": 0.7233, + "step": 7659 + }, + { + "epoch": 0.6527481891776736, + "grad_norm": 27.0, + "learning_rate": 1.9081662578242983e-06, + "loss": 0.3556, + "step": 7660 + }, + { + "epoch": 0.6528334043459736, + "grad_norm": 11.5, + "learning_rate": 1.908023395666812e-06, + "loss": 0.4448, + "step": 7661 + }, + { + "epoch": 0.6529186195142735, + "grad_norm": 16.625, + "learning_rate": 1.907880521618182e-06, + "loss": 0.8879, + "step": 7662 + }, + { + "epoch": 0.6530038346825735, + "grad_norm": 12.6875, + "learning_rate": 1.907737635680989e-06, + "loss": 0.5524, + "step": 7663 + }, + { + "epoch": 0.6530890498508735, + "grad_norm": 14.6875, + "learning_rate": 1.9075947378578157e-06, + "loss": 0.8419, + "step": 7664 + }, + { + "epoch": 0.6531742650191734, + "grad_norm": 14.125, + "learning_rate": 1.9074518281512438e-06, + "loss": 0.6376, + "step": 7665 + }, + { + "epoch": 0.6532594801874734, + "grad_norm": 20.25, + "learning_rate": 1.9073089065638563e-06, + "loss": 0.8453, + "step": 7666 + }, + { + "epoch": 0.6533446953557733, + "grad_norm": 13.375, + "learning_rate": 1.9071659730982362e-06, + "loss": 0.5867, + "step": 7667 + }, + { + "epoch": 0.6534299105240733, + "grad_norm": 19.125, + "learning_rate": 1.907023027756966e-06, + "loss": 0.5683, + "step": 7668 + }, + { + "epoch": 0.6535151256923732, + "grad_norm": 10.5, + "learning_rate": 1.906880070542629e-06, + "loss": 0.3983, + "step": 7669 + }, + { + "epoch": 0.6536003408606732, + "grad_norm": 19.375, + "learning_rate": 1.9067371014578087e-06, + "loss": 1.0957, + "step": 7670 + }, + { + "epoch": 0.6536855560289732, + "grad_norm": 19.375, + "learning_rate": 1.9065941205050886e-06, + "loss": 1.0778, + "step": 7671 + }, + { + "epoch": 0.6537707711972731, + "grad_norm": 13.5625, + "learning_rate": 1.9064511276870523e-06, + "loss": 0.6851, + "step": 7672 + }, + { + "epoch": 0.6538559863655731, + "grad_norm": 15.0625, + "learning_rate": 1.9063081230062841e-06, + "loss": 0.9018, + "step": 7673 + }, + { + "epoch": 0.653941201533873, + "grad_norm": 12.6875, + "learning_rate": 1.906165106465368e-06, + "loss": 0.6788, + "step": 7674 + }, + { + "epoch": 0.654026416702173, + "grad_norm": 12.625, + "learning_rate": 1.906022078066889e-06, + "loss": 0.6593, + "step": 7675 + }, + { + "epoch": 0.6541116318704729, + "grad_norm": 13.5625, + "learning_rate": 1.9058790378134309e-06, + "loss": 0.5763, + "step": 7676 + }, + { + "epoch": 0.6541968470387729, + "grad_norm": 14.875, + "learning_rate": 1.9057359857075794e-06, + "loss": 0.8212, + "step": 7677 + }, + { + "epoch": 0.6542820622070729, + "grad_norm": 17.0, + "learning_rate": 1.9055929217519186e-06, + "loss": 0.5392, + "step": 7678 + }, + { + "epoch": 0.6543672773753728, + "grad_norm": 8.1875, + "learning_rate": 1.9054498459490357e-06, + "loss": 0.1814, + "step": 7679 + }, + { + "epoch": 0.6544524925436728, + "grad_norm": 12.4375, + "learning_rate": 1.9053067583015138e-06, + "loss": 0.4883, + "step": 7680 + }, + { + "epoch": 0.6545377077119727, + "grad_norm": 16.75, + "learning_rate": 1.9051636588119405e-06, + "loss": 1.0578, + "step": 7681 + }, + { + "epoch": 0.6546229228802727, + "grad_norm": 11.875, + "learning_rate": 1.9050205474829005e-06, + "loss": 0.7196, + "step": 7682 + }, + { + "epoch": 0.6547081380485726, + "grad_norm": 11.9375, + "learning_rate": 1.904877424316981e-06, + "loss": 0.6645, + "step": 7683 + }, + { + "epoch": 0.6547933532168726, + "grad_norm": 15.0, + "learning_rate": 1.9047342893167678e-06, + "loss": 0.5131, + "step": 7684 + }, + { + "epoch": 0.6548785683851726, + "grad_norm": 16.125, + "learning_rate": 1.9045911424848476e-06, + "loss": 0.8631, + "step": 7685 + }, + { + "epoch": 0.6549637835534725, + "grad_norm": 15.6875, + "learning_rate": 1.904447983823807e-06, + "loss": 0.4195, + "step": 7686 + }, + { + "epoch": 0.6550489987217725, + "grad_norm": 21.0, + "learning_rate": 1.9043048133362334e-06, + "loss": 1.0715, + "step": 7687 + }, + { + "epoch": 0.6551342138900724, + "grad_norm": 17.75, + "learning_rate": 1.904161631024714e-06, + "loss": 1.0429, + "step": 7688 + }, + { + "epoch": 0.6552194290583724, + "grad_norm": 18.0, + "learning_rate": 1.9040184368918358e-06, + "loss": 0.9046, + "step": 7689 + }, + { + "epoch": 0.6553046442266723, + "grad_norm": 15.375, + "learning_rate": 1.903875230940187e-06, + "loss": 0.9655, + "step": 7690 + }, + { + "epoch": 0.6553898593949723, + "grad_norm": 16.625, + "learning_rate": 1.9037320131723552e-06, + "loss": 1.1764, + "step": 7691 + }, + { + "epoch": 0.6554750745632723, + "grad_norm": 18.125, + "learning_rate": 1.9035887835909289e-06, + "loss": 1.071, + "step": 7692 + }, + { + "epoch": 0.6555602897315722, + "grad_norm": 14.0, + "learning_rate": 1.9034455421984955e-06, + "loss": 0.63, + "step": 7693 + }, + { + "epoch": 0.6556455048998722, + "grad_norm": 12.5, + "learning_rate": 1.9033022889976444e-06, + "loss": 0.5354, + "step": 7694 + }, + { + "epoch": 0.6557307200681721, + "grad_norm": 14.1875, + "learning_rate": 1.903159023990964e-06, + "loss": 1.0363, + "step": 7695 + }, + { + "epoch": 0.6558159352364721, + "grad_norm": 9.6875, + "learning_rate": 1.9030157471810428e-06, + "loss": 0.6402, + "step": 7696 + }, + { + "epoch": 0.655901150404772, + "grad_norm": 17.125, + "learning_rate": 1.9028724585704708e-06, + "loss": 1.0597, + "step": 7697 + }, + { + "epoch": 0.655986365573072, + "grad_norm": 12.9375, + "learning_rate": 1.9027291581618368e-06, + "loss": 0.698, + "step": 7698 + }, + { + "epoch": 0.656071580741372, + "grad_norm": 12.25, + "learning_rate": 1.9025858459577306e-06, + "loss": 0.4694, + "step": 7699 + }, + { + "epoch": 0.6561567959096719, + "grad_norm": 13.5625, + "learning_rate": 1.9024425219607422e-06, + "loss": 0.6339, + "step": 7700 + }, + { + "epoch": 0.6562420110779719, + "grad_norm": 10.875, + "learning_rate": 1.9022991861734607e-06, + "loss": 0.2362, + "step": 7701 + }, + { + "epoch": 0.6563272262462718, + "grad_norm": 12.0625, + "learning_rate": 1.9021558385984776e-06, + "loss": 0.4518, + "step": 7702 + }, + { + "epoch": 0.6564124414145718, + "grad_norm": 16.25, + "learning_rate": 1.9020124792383826e-06, + "loss": 0.5414, + "step": 7703 + }, + { + "epoch": 0.6564976565828717, + "grad_norm": 16.5, + "learning_rate": 1.9018691080957663e-06, + "loss": 0.8466, + "step": 7704 + }, + { + "epoch": 0.6565828717511717, + "grad_norm": 12.3125, + "learning_rate": 1.9017257251732197e-06, + "loss": 0.4506, + "step": 7705 + }, + { + "epoch": 0.6566680869194716, + "grad_norm": 14.5625, + "learning_rate": 1.9015823304733337e-06, + "loss": 0.9776, + "step": 7706 + }, + { + "epoch": 0.6567533020877716, + "grad_norm": 11.3125, + "learning_rate": 1.9014389239987003e-06, + "loss": 0.5014, + "step": 7707 + }, + { + "epoch": 0.6568385172560716, + "grad_norm": 19.75, + "learning_rate": 1.90129550575191e-06, + "loss": 1.0714, + "step": 7708 + }, + { + "epoch": 0.6569237324243715, + "grad_norm": 15.1875, + "learning_rate": 1.901152075735555e-06, + "loss": 0.876, + "step": 7709 + }, + { + "epoch": 0.6570089475926715, + "grad_norm": 13.8125, + "learning_rate": 1.9010086339522272e-06, + "loss": 0.8053, + "step": 7710 + }, + { + "epoch": 0.6570941627609714, + "grad_norm": 13.625, + "learning_rate": 1.9008651804045192e-06, + "loss": 0.6776, + "step": 7711 + }, + { + "epoch": 0.6571793779292714, + "grad_norm": 17.375, + "learning_rate": 1.9007217150950224e-06, + "loss": 0.8731, + "step": 7712 + }, + { + "epoch": 0.6572645930975713, + "grad_norm": 15.875, + "learning_rate": 1.9005782380263301e-06, + "loss": 0.6487, + "step": 7713 + }, + { + "epoch": 0.6573498082658713, + "grad_norm": 11.5, + "learning_rate": 1.9004347492010344e-06, + "loss": 0.5321, + "step": 7714 + }, + { + "epoch": 0.6574350234341713, + "grad_norm": 13.875, + "learning_rate": 1.9002912486217294e-06, + "loss": 0.7253, + "step": 7715 + }, + { + "epoch": 0.6575202386024712, + "grad_norm": 19.5, + "learning_rate": 1.9001477362910072e-06, + "loss": 0.9307, + "step": 7716 + }, + { + "epoch": 0.6576054537707712, + "grad_norm": 24.125, + "learning_rate": 1.900004212211462e-06, + "loss": 0.8846, + "step": 7717 + }, + { + "epoch": 0.6576906689390711, + "grad_norm": 11.0, + "learning_rate": 1.8998606763856864e-06, + "loss": 0.375, + "step": 7718 + }, + { + "epoch": 0.6577758841073711, + "grad_norm": 13.1875, + "learning_rate": 1.8997171288162756e-06, + "loss": 0.5148, + "step": 7719 + }, + { + "epoch": 0.657861099275671, + "grad_norm": 10.5625, + "learning_rate": 1.8995735695058226e-06, + "loss": 0.4347, + "step": 7720 + }, + { + "epoch": 0.657946314443971, + "grad_norm": 10.625, + "learning_rate": 1.899429998456922e-06, + "loss": 0.3172, + "step": 7721 + }, + { + "epoch": 0.658031529612271, + "grad_norm": 16.125, + "learning_rate": 1.8992864156721682e-06, + "loss": 0.7518, + "step": 7722 + }, + { + "epoch": 0.6581167447805709, + "grad_norm": 18.25, + "learning_rate": 1.8991428211541564e-06, + "loss": 0.7375, + "step": 7723 + }, + { + "epoch": 0.6582019599488709, + "grad_norm": 12.8125, + "learning_rate": 1.8989992149054806e-06, + "loss": 0.7669, + "step": 7724 + }, + { + "epoch": 0.6582871751171708, + "grad_norm": 12.5625, + "learning_rate": 1.8988555969287365e-06, + "loss": 0.6209, + "step": 7725 + }, + { + "epoch": 0.6583723902854708, + "grad_norm": 16.75, + "learning_rate": 1.898711967226519e-06, + "loss": 0.8698, + "step": 7726 + }, + { + "epoch": 0.6584576054537707, + "grad_norm": 13.375, + "learning_rate": 1.8985683258014244e-06, + "loss": 0.7797, + "step": 7727 + }, + { + "epoch": 0.6585428206220707, + "grad_norm": 12.3125, + "learning_rate": 1.8984246726560478e-06, + "loss": 0.7394, + "step": 7728 + }, + { + "epoch": 0.6586280357903707, + "grad_norm": 17.875, + "learning_rate": 1.8982810077929854e-06, + "loss": 0.6197, + "step": 7729 + }, + { + "epoch": 0.6587132509586706, + "grad_norm": 15.75, + "learning_rate": 1.898137331214833e-06, + "loss": 0.4921, + "step": 7730 + }, + { + "epoch": 0.6587984661269706, + "grad_norm": 17.125, + "learning_rate": 1.8979936429241874e-06, + "loss": 0.9318, + "step": 7731 + }, + { + "epoch": 0.6588836812952705, + "grad_norm": 12.4375, + "learning_rate": 1.8978499429236451e-06, + "loss": 0.6064, + "step": 7732 + }, + { + "epoch": 0.6589688964635705, + "grad_norm": 15.75, + "learning_rate": 1.897706231215803e-06, + "loss": 0.3024, + "step": 7733 + }, + { + "epoch": 0.6590541116318704, + "grad_norm": 13.0625, + "learning_rate": 1.8975625078032577e-06, + "loss": 0.7487, + "step": 7734 + }, + { + "epoch": 0.6591393268001704, + "grad_norm": 14.125, + "learning_rate": 1.8974187726886068e-06, + "loss": 0.7034, + "step": 7735 + }, + { + "epoch": 0.6592245419684704, + "grad_norm": 15.8125, + "learning_rate": 1.8972750258744478e-06, + "loss": 0.7132, + "step": 7736 + }, + { + "epoch": 0.6593097571367703, + "grad_norm": 16.125, + "learning_rate": 1.8971312673633782e-06, + "loss": 0.8852, + "step": 7737 + }, + { + "epoch": 0.6593949723050703, + "grad_norm": 13.625, + "learning_rate": 1.8969874971579954e-06, + "loss": 0.8059, + "step": 7738 + }, + { + "epoch": 0.6594801874733702, + "grad_norm": 9.875, + "learning_rate": 1.8968437152608987e-06, + "loss": 0.3816, + "step": 7739 + }, + { + "epoch": 0.6595654026416702, + "grad_norm": 23.75, + "learning_rate": 1.896699921674685e-06, + "loss": 0.9941, + "step": 7740 + }, + { + "epoch": 0.6596506178099701, + "grad_norm": 18.25, + "learning_rate": 1.8965561164019538e-06, + "loss": 1.0777, + "step": 7741 + }, + { + "epoch": 0.6597358329782701, + "grad_norm": 15.5625, + "learning_rate": 1.8964122994453032e-06, + "loss": 0.8716, + "step": 7742 + }, + { + "epoch": 0.65982104814657, + "grad_norm": 10.9375, + "learning_rate": 1.8962684708073327e-06, + "loss": 0.4156, + "step": 7743 + }, + { + "epoch": 0.65990626331487, + "grad_norm": 18.375, + "learning_rate": 1.8961246304906406e-06, + "loss": 0.84, + "step": 7744 + }, + { + "epoch": 0.65999147848317, + "grad_norm": 22.25, + "learning_rate": 1.8959807784978268e-06, + "loss": 0.9273, + "step": 7745 + }, + { + "epoch": 0.6600766936514699, + "grad_norm": 14.1875, + "learning_rate": 1.8958369148314907e-06, + "loss": 0.7878, + "step": 7746 + }, + { + "epoch": 0.6601619088197699, + "grad_norm": 19.875, + "learning_rate": 1.8956930394942323e-06, + "loss": 1.1043, + "step": 7747 + }, + { + "epoch": 0.6602471239880698, + "grad_norm": 12.875, + "learning_rate": 1.8955491524886513e-06, + "loss": 0.6509, + "step": 7748 + }, + { + "epoch": 0.6603323391563698, + "grad_norm": 10.625, + "learning_rate": 1.8954052538173478e-06, + "loss": 0.3943, + "step": 7749 + }, + { + "epoch": 0.6604175543246698, + "grad_norm": 26.125, + "learning_rate": 1.8952613434829226e-06, + "loss": 1.2321, + "step": 7750 + }, + { + "epoch": 0.6605027694929697, + "grad_norm": 17.875, + "learning_rate": 1.8951174214879757e-06, + "loss": 1.2233, + "step": 7751 + }, + { + "epoch": 0.6605879846612697, + "grad_norm": 18.75, + "learning_rate": 1.894973487835109e-06, + "loss": 0.7206, + "step": 7752 + }, + { + "epoch": 0.6606731998295696, + "grad_norm": 11.9375, + "learning_rate": 1.894829542526922e-06, + "loss": 0.4861, + "step": 7753 + }, + { + "epoch": 0.6607584149978696, + "grad_norm": 14.6875, + "learning_rate": 1.8946855855660174e-06, + "loss": 0.6899, + "step": 7754 + }, + { + "epoch": 0.6608436301661695, + "grad_norm": 19.625, + "learning_rate": 1.8945416169549954e-06, + "loss": 1.0371, + "step": 7755 + }, + { + "epoch": 0.6609288453344695, + "grad_norm": 13.375, + "learning_rate": 1.8943976366964586e-06, + "loss": 0.8959, + "step": 7756 + }, + { + "epoch": 0.6610140605027695, + "grad_norm": 13.8125, + "learning_rate": 1.8942536447930081e-06, + "loss": 0.8161, + "step": 7757 + }, + { + "epoch": 0.6610992756710694, + "grad_norm": 13.1875, + "learning_rate": 1.8941096412472472e-06, + "loss": 0.4195, + "step": 7758 + }, + { + "epoch": 0.6611844908393694, + "grad_norm": 13.875, + "learning_rate": 1.8939656260617767e-06, + "loss": 1.0856, + "step": 7759 + }, + { + "epoch": 0.6612697060076693, + "grad_norm": 14.5, + "learning_rate": 1.8938215992392e-06, + "loss": 0.5116, + "step": 7760 + }, + { + "epoch": 0.6613549211759693, + "grad_norm": 14.75, + "learning_rate": 1.8936775607821194e-06, + "loss": 0.8593, + "step": 7761 + }, + { + "epoch": 0.6614401363442692, + "grad_norm": 17.5, + "learning_rate": 1.8935335106931385e-06, + "loss": 0.5395, + "step": 7762 + }, + { + "epoch": 0.6615253515125692, + "grad_norm": 12.1875, + "learning_rate": 1.8933894489748595e-06, + "loss": 0.649, + "step": 7763 + }, + { + "epoch": 0.6616105666808691, + "grad_norm": 15.625, + "learning_rate": 1.8932453756298864e-06, + "loss": 0.8281, + "step": 7764 + }, + { + "epoch": 0.6616957818491691, + "grad_norm": 16.0, + "learning_rate": 1.8931012906608222e-06, + "loss": 0.918, + "step": 7765 + }, + { + "epoch": 0.6617809970174691, + "grad_norm": 8.9375, + "learning_rate": 1.8929571940702712e-06, + "loss": 0.5456, + "step": 7766 + }, + { + "epoch": 0.661866212185769, + "grad_norm": 25.5, + "learning_rate": 1.8928130858608373e-06, + "loss": 1.1243, + "step": 7767 + }, + { + "epoch": 0.661951427354069, + "grad_norm": 9.3125, + "learning_rate": 1.8926689660351243e-06, + "loss": 0.2324, + "step": 7768 + }, + { + "epoch": 0.6620366425223689, + "grad_norm": 14.125, + "learning_rate": 1.892524834595737e-06, + "loss": 0.7314, + "step": 7769 + }, + { + "epoch": 0.6621218576906689, + "grad_norm": 15.5, + "learning_rate": 1.89238069154528e-06, + "loss": 0.8729, + "step": 7770 + }, + { + "epoch": 0.6622070728589688, + "grad_norm": 16.0, + "learning_rate": 1.8922365368863578e-06, + "loss": 0.6991, + "step": 7771 + }, + { + "epoch": 0.6622922880272688, + "grad_norm": 11.6875, + "learning_rate": 1.8920923706215751e-06, + "loss": 0.5239, + "step": 7772 + }, + { + "epoch": 0.6623775031955688, + "grad_norm": 13.0625, + "learning_rate": 1.8919481927535382e-06, + "loss": 0.6511, + "step": 7773 + }, + { + "epoch": 0.6624627183638687, + "grad_norm": 20.75, + "learning_rate": 1.8918040032848515e-06, + "loss": 0.684, + "step": 7774 + }, + { + "epoch": 0.6625479335321687, + "grad_norm": 21.25, + "learning_rate": 1.8916598022181213e-06, + "loss": 1.0745, + "step": 7775 + }, + { + "epoch": 0.6626331487004686, + "grad_norm": 20.375, + "learning_rate": 1.8915155895559534e-06, + "loss": 0.7396, + "step": 7776 + }, + { + "epoch": 0.6627183638687686, + "grad_norm": 17.25, + "learning_rate": 1.8913713653009536e-06, + "loss": 0.5338, + "step": 7777 + }, + { + "epoch": 0.6628035790370685, + "grad_norm": 12.0625, + "learning_rate": 1.8912271294557277e-06, + "loss": 0.5244, + "step": 7778 + }, + { + "epoch": 0.6628887942053685, + "grad_norm": 14.5, + "learning_rate": 1.8910828820228836e-06, + "loss": 0.4698, + "step": 7779 + }, + { + "epoch": 0.6629740093736685, + "grad_norm": 14.0625, + "learning_rate": 1.890938623005027e-06, + "loss": 0.7289, + "step": 7780 + }, + { + "epoch": 0.6630592245419684, + "grad_norm": 16.875, + "learning_rate": 1.8907943524047645e-06, + "loss": 0.9359, + "step": 7781 + }, + { + "epoch": 0.6631444397102684, + "grad_norm": 19.75, + "learning_rate": 1.8906500702247038e-06, + "loss": 1.0226, + "step": 7782 + }, + { + "epoch": 0.6632296548785683, + "grad_norm": 12.75, + "learning_rate": 1.8905057764674522e-06, + "loss": 0.416, + "step": 7783 + }, + { + "epoch": 0.6633148700468683, + "grad_norm": 14.8125, + "learning_rate": 1.8903614711356178e-06, + "loss": 0.4751, + "step": 7784 + }, + { + "epoch": 0.6634000852151682, + "grad_norm": 13.5, + "learning_rate": 1.890217154231807e-06, + "loss": 0.7512, + "step": 7785 + }, + { + "epoch": 0.6634853003834683, + "grad_norm": 13.9375, + "learning_rate": 1.8900728257586282e-06, + "loss": 0.7259, + "step": 7786 + }, + { + "epoch": 0.6635705155517683, + "grad_norm": 18.875, + "learning_rate": 1.8899284857186903e-06, + "loss": 0.9254, + "step": 7787 + }, + { + "epoch": 0.6636557307200682, + "grad_norm": 16.875, + "learning_rate": 1.8897841341146012e-06, + "loss": 0.849, + "step": 7788 + }, + { + "epoch": 0.6637409458883682, + "grad_norm": 13.0625, + "learning_rate": 1.8896397709489688e-06, + "loss": 1.0384, + "step": 7789 + }, + { + "epoch": 0.6638261610566681, + "grad_norm": 13.625, + "learning_rate": 1.889495396224403e-06, + "loss": 0.7918, + "step": 7790 + }, + { + "epoch": 0.6639113762249681, + "grad_norm": 25.0, + "learning_rate": 1.8893510099435122e-06, + "loss": 1.2188, + "step": 7791 + }, + { + "epoch": 0.663996591393268, + "grad_norm": 9.5625, + "learning_rate": 1.889206612108906e-06, + "loss": 0.2705, + "step": 7792 + }, + { + "epoch": 0.664081806561568, + "grad_norm": 13.125, + "learning_rate": 1.8890622027231937e-06, + "loss": 0.7039, + "step": 7793 + }, + { + "epoch": 0.664167021729868, + "grad_norm": 14.3125, + "learning_rate": 1.8889177817889841e-06, + "loss": 0.9733, + "step": 7794 + }, + { + "epoch": 0.6642522368981679, + "grad_norm": 14.3125, + "learning_rate": 1.8887733493088882e-06, + "loss": 0.621, + "step": 7795 + }, + { + "epoch": 0.6643374520664679, + "grad_norm": 11.0625, + "learning_rate": 1.8886289052855155e-06, + "loss": 0.4317, + "step": 7796 + }, + { + "epoch": 0.6644226672347678, + "grad_norm": 13.6875, + "learning_rate": 1.888484449721476e-06, + "loss": 0.7371, + "step": 7797 + }, + { + "epoch": 0.6645078824030678, + "grad_norm": 14.6875, + "learning_rate": 1.8883399826193807e-06, + "loss": 0.553, + "step": 7798 + }, + { + "epoch": 0.6645930975713678, + "grad_norm": 12.1875, + "learning_rate": 1.8881955039818398e-06, + "loss": 0.3941, + "step": 7799 + }, + { + "epoch": 0.6646783127396677, + "grad_norm": 13.375, + "learning_rate": 1.8880510138114646e-06, + "loss": 0.4722, + "step": 7800 + }, + { + "epoch": 0.6647635279079677, + "grad_norm": 10.3125, + "learning_rate": 1.8879065121108659e-06, + "loss": 0.4653, + "step": 7801 + }, + { + "epoch": 0.6648487430762676, + "grad_norm": 24.625, + "learning_rate": 1.887761998882655e-06, + "loss": 1.0067, + "step": 7802 + }, + { + "epoch": 0.6649339582445676, + "grad_norm": 15.5625, + "learning_rate": 1.8876174741294434e-06, + "loss": 0.991, + "step": 7803 + }, + { + "epoch": 0.6650191734128675, + "grad_norm": 13.6875, + "learning_rate": 1.8874729378538432e-06, + "loss": 0.396, + "step": 7804 + }, + { + "epoch": 0.6651043885811675, + "grad_norm": 19.75, + "learning_rate": 1.887328390058466e-06, + "loss": 1.0518, + "step": 7805 + }, + { + "epoch": 0.6651896037494675, + "grad_norm": 14.5625, + "learning_rate": 1.8871838307459234e-06, + "loss": 0.8238, + "step": 7806 + }, + { + "epoch": 0.6652748189177674, + "grad_norm": 10.1875, + "learning_rate": 1.8870392599188286e-06, + "loss": 0.3528, + "step": 7807 + }, + { + "epoch": 0.6653600340860674, + "grad_norm": 18.125, + "learning_rate": 1.8868946775797936e-06, + "loss": 0.4562, + "step": 7808 + }, + { + "epoch": 0.6654452492543673, + "grad_norm": 14.75, + "learning_rate": 1.8867500837314318e-06, + "loss": 0.941, + "step": 7809 + }, + { + "epoch": 0.6655304644226673, + "grad_norm": 17.75, + "learning_rate": 1.886605478376355e-06, + "loss": 0.6784, + "step": 7810 + }, + { + "epoch": 0.6656156795909672, + "grad_norm": 15.25, + "learning_rate": 1.8864608615171776e-06, + "loss": 0.7678, + "step": 7811 + }, + { + "epoch": 0.6657008947592672, + "grad_norm": 17.375, + "learning_rate": 1.886316233156512e-06, + "loss": 0.7316, + "step": 7812 + }, + { + "epoch": 0.6657861099275671, + "grad_norm": 29.625, + "learning_rate": 1.8861715932969727e-06, + "loss": 0.8083, + "step": 7813 + }, + { + "epoch": 0.6658713250958671, + "grad_norm": 13.1875, + "learning_rate": 1.8860269419411724e-06, + "loss": 0.4648, + "step": 7814 + }, + { + "epoch": 0.6659565402641671, + "grad_norm": 26.125, + "learning_rate": 1.885882279091726e-06, + "loss": 0.6458, + "step": 7815 + }, + { + "epoch": 0.666041755432467, + "grad_norm": 12.5625, + "learning_rate": 1.8857376047512471e-06, + "loss": 0.6986, + "step": 7816 + }, + { + "epoch": 0.666126970600767, + "grad_norm": 20.625, + "learning_rate": 1.885592918922351e-06, + "loss": 1.1474, + "step": 7817 + }, + { + "epoch": 0.6662121857690669, + "grad_norm": 23.875, + "learning_rate": 1.8854482216076514e-06, + "loss": 0.8777, + "step": 7818 + }, + { + "epoch": 0.6662974009373669, + "grad_norm": 16.0, + "learning_rate": 1.8853035128097635e-06, + "loss": 0.6345, + "step": 7819 + }, + { + "epoch": 0.6663826161056668, + "grad_norm": 19.0, + "learning_rate": 1.885158792531302e-06, + "loss": 0.8317, + "step": 7820 + }, + { + "epoch": 0.6664678312739668, + "grad_norm": 14.375, + "learning_rate": 1.8850140607748826e-06, + "loss": 0.9223, + "step": 7821 + }, + { + "epoch": 0.6665530464422668, + "grad_norm": 11.375, + "learning_rate": 1.8848693175431206e-06, + "loss": 0.5993, + "step": 7822 + }, + { + "epoch": 0.6666382616105667, + "grad_norm": 14.8125, + "learning_rate": 1.8847245628386313e-06, + "loss": 0.5059, + "step": 7823 + }, + { + "epoch": 0.6667234767788667, + "grad_norm": 11.3125, + "learning_rate": 1.8845797966640314e-06, + "loss": 0.4261, + "step": 7824 + }, + { + "epoch": 0.6668086919471666, + "grad_norm": 11.0, + "learning_rate": 1.8844350190219362e-06, + "loss": 0.5137, + "step": 7825 + }, + { + "epoch": 0.6668939071154666, + "grad_norm": 18.25, + "learning_rate": 1.8842902299149624e-06, + "loss": 1.0134, + "step": 7826 + }, + { + "epoch": 0.6669791222837665, + "grad_norm": 13.0, + "learning_rate": 1.8841454293457263e-06, + "loss": 0.5634, + "step": 7827 + }, + { + "epoch": 0.6670643374520665, + "grad_norm": 23.0, + "learning_rate": 1.8840006173168442e-06, + "loss": 0.4067, + "step": 7828 + }, + { + "epoch": 0.6671495526203665, + "grad_norm": 13.25, + "learning_rate": 1.8838557938309337e-06, + "loss": 0.4343, + "step": 7829 + }, + { + "epoch": 0.6672347677886664, + "grad_norm": 14.5625, + "learning_rate": 1.883710958890612e-06, + "loss": 0.817, + "step": 7830 + }, + { + "epoch": 0.6673199829569664, + "grad_norm": 19.5, + "learning_rate": 1.8835661124984956e-06, + "loss": 0.8981, + "step": 7831 + }, + { + "epoch": 0.6674051981252663, + "grad_norm": 14.375, + "learning_rate": 1.8834212546572023e-06, + "loss": 0.698, + "step": 7832 + }, + { + "epoch": 0.6674904132935663, + "grad_norm": 18.875, + "learning_rate": 1.8832763853693502e-06, + "loss": 0.7266, + "step": 7833 + }, + { + "epoch": 0.6675756284618662, + "grad_norm": 14.5625, + "learning_rate": 1.883131504637557e-06, + "loss": 0.715, + "step": 7834 + }, + { + "epoch": 0.6676608436301662, + "grad_norm": 12.4375, + "learning_rate": 1.8829866124644408e-06, + "loss": 0.6582, + "step": 7835 + }, + { + "epoch": 0.6677460587984662, + "grad_norm": 15.8125, + "learning_rate": 1.8828417088526203e-06, + "loss": 0.7126, + "step": 7836 + }, + { + "epoch": 0.6678312739667661, + "grad_norm": 19.0, + "learning_rate": 1.8826967938047134e-06, + "loss": 1.1873, + "step": 7837 + }, + { + "epoch": 0.6679164891350661, + "grad_norm": 18.0, + "learning_rate": 1.882551867323339e-06, + "loss": 1.1855, + "step": 7838 + }, + { + "epoch": 0.668001704303366, + "grad_norm": 15.625, + "learning_rate": 1.8824069294111166e-06, + "loss": 0.9735, + "step": 7839 + }, + { + "epoch": 0.668086919471666, + "grad_norm": 14.3125, + "learning_rate": 1.8822619800706649e-06, + "loss": 0.815, + "step": 7840 + }, + { + "epoch": 0.6681721346399659, + "grad_norm": 17.25, + "learning_rate": 1.8821170193046038e-06, + "loss": 0.6461, + "step": 7841 + }, + { + "epoch": 0.6682573498082659, + "grad_norm": 17.5, + "learning_rate": 1.8819720471155519e-06, + "loss": 0.8103, + "step": 7842 + }, + { + "epoch": 0.6683425649765659, + "grad_norm": 12.9375, + "learning_rate": 1.8818270635061303e-06, + "loss": 0.5115, + "step": 7843 + }, + { + "epoch": 0.6684277801448658, + "grad_norm": 21.5, + "learning_rate": 1.8816820684789578e-06, + "loss": 1.2185, + "step": 7844 + }, + { + "epoch": 0.6685129953131658, + "grad_norm": 17.5, + "learning_rate": 1.8815370620366554e-06, + "loss": 1.0734, + "step": 7845 + }, + { + "epoch": 0.6685982104814657, + "grad_norm": 13.25, + "learning_rate": 1.8813920441818428e-06, + "loss": 0.7405, + "step": 7846 + }, + { + "epoch": 0.6686834256497657, + "grad_norm": 10.5, + "learning_rate": 1.8812470149171416e-06, + "loss": 0.3884, + "step": 7847 + }, + { + "epoch": 0.6687686408180656, + "grad_norm": 13.0625, + "learning_rate": 1.881101974245172e-06, + "loss": 0.4499, + "step": 7848 + }, + { + "epoch": 0.6688538559863656, + "grad_norm": 26.625, + "learning_rate": 1.880956922168555e-06, + "loss": 0.769, + "step": 7849 + }, + { + "epoch": 0.6689390711546656, + "grad_norm": 13.0625, + "learning_rate": 1.8808118586899116e-06, + "loss": 0.709, + "step": 7850 + }, + { + "epoch": 0.6690242863229655, + "grad_norm": 12.25, + "learning_rate": 1.880666783811864e-06, + "loss": 0.5738, + "step": 7851 + }, + { + "epoch": 0.6691095014912655, + "grad_norm": 16.875, + "learning_rate": 1.8805216975370334e-06, + "loss": 0.7661, + "step": 7852 + }, + { + "epoch": 0.6691947166595654, + "grad_norm": 15.9375, + "learning_rate": 1.8803765998680414e-06, + "loss": 0.7259, + "step": 7853 + }, + { + "epoch": 0.6692799318278654, + "grad_norm": 13.0, + "learning_rate": 1.8802314908075107e-06, + "loss": 0.7106, + "step": 7854 + }, + { + "epoch": 0.6693651469961653, + "grad_norm": 18.25, + "learning_rate": 1.880086370358063e-06, + "loss": 0.5793, + "step": 7855 + }, + { + "epoch": 0.6694503621644653, + "grad_norm": 15.4375, + "learning_rate": 1.8799412385223208e-06, + "loss": 0.4874, + "step": 7856 + }, + { + "epoch": 0.6695355773327653, + "grad_norm": 10.75, + "learning_rate": 1.8797960953029076e-06, + "loss": 0.5157, + "step": 7857 + }, + { + "epoch": 0.6696207925010652, + "grad_norm": 13.0, + "learning_rate": 1.8796509407024452e-06, + "loss": 0.7985, + "step": 7858 + }, + { + "epoch": 0.6697060076693652, + "grad_norm": 18.75, + "learning_rate": 1.8795057747235569e-06, + "loss": 0.6469, + "step": 7859 + }, + { + "epoch": 0.6697912228376651, + "grad_norm": 19.375, + "learning_rate": 1.8793605973688666e-06, + "loss": 0.8221, + "step": 7860 + }, + { + "epoch": 0.6698764380059651, + "grad_norm": 14.375, + "learning_rate": 1.8792154086409974e-06, + "loss": 1.0967, + "step": 7861 + }, + { + "epoch": 0.669961653174265, + "grad_norm": 13.5, + "learning_rate": 1.8790702085425732e-06, + "loss": 0.5723, + "step": 7862 + }, + { + "epoch": 0.670046868342565, + "grad_norm": 14.3125, + "learning_rate": 1.8789249970762174e-06, + "loss": 0.5327, + "step": 7863 + }, + { + "epoch": 0.670132083510865, + "grad_norm": 15.25, + "learning_rate": 1.8787797742445545e-06, + "loss": 0.567, + "step": 7864 + }, + { + "epoch": 0.6702172986791649, + "grad_norm": 16.25, + "learning_rate": 1.8786345400502092e-06, + "loss": 0.5532, + "step": 7865 + }, + { + "epoch": 0.6703025138474649, + "grad_norm": 16.0, + "learning_rate": 1.8784892944958055e-06, + "loss": 1.0269, + "step": 7866 + }, + { + "epoch": 0.6703877290157648, + "grad_norm": 15.875, + "learning_rate": 1.8783440375839679e-06, + "loss": 1.0999, + "step": 7867 + }, + { + "epoch": 0.6704729441840648, + "grad_norm": 14.75, + "learning_rate": 1.8781987693173218e-06, + "loss": 0.7037, + "step": 7868 + }, + { + "epoch": 0.6705581593523647, + "grad_norm": 11.125, + "learning_rate": 1.878053489698492e-06, + "loss": 0.364, + "step": 7869 + }, + { + "epoch": 0.6706433745206647, + "grad_norm": 11.8125, + "learning_rate": 1.8779081987301049e-06, + "loss": 0.4882, + "step": 7870 + }, + { + "epoch": 0.6707285896889646, + "grad_norm": 17.0, + "learning_rate": 1.8777628964147845e-06, + "loss": 0.6133, + "step": 7871 + }, + { + "epoch": 0.6708138048572646, + "grad_norm": 31.75, + "learning_rate": 1.8776175827551577e-06, + "loss": 1.3091, + "step": 7872 + }, + { + "epoch": 0.6708990200255646, + "grad_norm": 10.875, + "learning_rate": 1.8774722577538497e-06, + "loss": 0.534, + "step": 7873 + }, + { + "epoch": 0.6709842351938645, + "grad_norm": 16.375, + "learning_rate": 1.8773269214134871e-06, + "loss": 0.8544, + "step": 7874 + }, + { + "epoch": 0.6710694503621645, + "grad_norm": 18.125, + "learning_rate": 1.8771815737366967e-06, + "loss": 0.8822, + "step": 7875 + }, + { + "epoch": 0.6711546655304644, + "grad_norm": 16.625, + "learning_rate": 1.877036214726104e-06, + "loss": 0.6763, + "step": 7876 + }, + { + "epoch": 0.6712398806987644, + "grad_norm": 16.25, + "learning_rate": 1.8768908443843363e-06, + "loss": 0.7085, + "step": 7877 + }, + { + "epoch": 0.6713250958670643, + "grad_norm": 9.6875, + "learning_rate": 1.8767454627140214e-06, + "loss": 0.2832, + "step": 7878 + }, + { + "epoch": 0.6714103110353643, + "grad_norm": 14.0625, + "learning_rate": 1.8766000697177853e-06, + "loss": 0.8317, + "step": 7879 + }, + { + "epoch": 0.6714955262036643, + "grad_norm": 14.25, + "learning_rate": 1.8764546653982558e-06, + "loss": 0.4419, + "step": 7880 + }, + { + "epoch": 0.6715807413719642, + "grad_norm": 12.375, + "learning_rate": 1.8763092497580603e-06, + "loss": 0.7417, + "step": 7881 + }, + { + "epoch": 0.6716659565402642, + "grad_norm": 17.375, + "learning_rate": 1.8761638227998274e-06, + "loss": 0.6472, + "step": 7882 + }, + { + "epoch": 0.6717511717085641, + "grad_norm": 26.25, + "learning_rate": 1.8760183845261842e-06, + "loss": 0.9428, + "step": 7883 + }, + { + "epoch": 0.6718363868768641, + "grad_norm": 13.875, + "learning_rate": 1.875872934939759e-06, + "loss": 0.6219, + "step": 7884 + }, + { + "epoch": 0.671921602045164, + "grad_norm": 18.5, + "learning_rate": 1.8757274740431813e-06, + "loss": 0.9919, + "step": 7885 + }, + { + "epoch": 0.672006817213464, + "grad_norm": 11.8125, + "learning_rate": 1.8755820018390786e-06, + "loss": 0.5577, + "step": 7886 + }, + { + "epoch": 0.672092032381764, + "grad_norm": 13.1875, + "learning_rate": 1.8754365183300798e-06, + "loss": 0.4373, + "step": 7887 + }, + { + "epoch": 0.6721772475500639, + "grad_norm": 13.3125, + "learning_rate": 1.8752910235188143e-06, + "loss": 0.652, + "step": 7888 + }, + { + "epoch": 0.6722624627183639, + "grad_norm": 10.25, + "learning_rate": 1.8751455174079115e-06, + "loss": 0.3525, + "step": 7889 + }, + { + "epoch": 0.6723476778866638, + "grad_norm": 19.5, + "learning_rate": 1.8750000000000003e-06, + "loss": 0.8254, + "step": 7890 + }, + { + "epoch": 0.6724328930549638, + "grad_norm": 13.5, + "learning_rate": 1.8748544712977104e-06, + "loss": 0.5942, + "step": 7891 + }, + { + "epoch": 0.6725181082232637, + "grad_norm": 24.625, + "learning_rate": 1.8747089313036722e-06, + "loss": 1.1972, + "step": 7892 + }, + { + "epoch": 0.6726033233915637, + "grad_norm": 14.8125, + "learning_rate": 1.874563380020515e-06, + "loss": 0.7444, + "step": 7893 + }, + { + "epoch": 0.6726885385598637, + "grad_norm": 14.9375, + "learning_rate": 1.8744178174508703e-06, + "loss": 0.4562, + "step": 7894 + }, + { + "epoch": 0.6727737537281636, + "grad_norm": 19.625, + "learning_rate": 1.874272243597367e-06, + "loss": 1.2864, + "step": 7895 + }, + { + "epoch": 0.6728589688964636, + "grad_norm": 26.0, + "learning_rate": 1.874126658462637e-06, + "loss": 0.9159, + "step": 7896 + }, + { + "epoch": 0.6729441840647635, + "grad_norm": 10.5, + "learning_rate": 1.8739810620493101e-06, + "loss": 0.3869, + "step": 7897 + }, + { + "epoch": 0.6730293992330635, + "grad_norm": 13.4375, + "learning_rate": 1.8738354543600187e-06, + "loss": 0.7865, + "step": 7898 + }, + { + "epoch": 0.6731146144013634, + "grad_norm": 14.375, + "learning_rate": 1.8736898353973925e-06, + "loss": 0.4406, + "step": 7899 + }, + { + "epoch": 0.6731998295696634, + "grad_norm": 10.4375, + "learning_rate": 1.8735442051640644e-06, + "loss": 0.4261, + "step": 7900 + }, + { + "epoch": 0.6732850447379634, + "grad_norm": 17.375, + "learning_rate": 1.8733985636626653e-06, + "loss": 0.6808, + "step": 7901 + }, + { + "epoch": 0.6733702599062633, + "grad_norm": 20.125, + "learning_rate": 1.8732529108958275e-06, + "loss": 0.9224, + "step": 7902 + }, + { + "epoch": 0.6734554750745633, + "grad_norm": 21.875, + "learning_rate": 1.8731072468661826e-06, + "loss": 1.3805, + "step": 7903 + }, + { + "epoch": 0.6735406902428632, + "grad_norm": 13.5, + "learning_rate": 1.8729615715763633e-06, + "loss": 0.4637, + "step": 7904 + }, + { + "epoch": 0.6736259054111632, + "grad_norm": 12.875, + "learning_rate": 1.8728158850290016e-06, + "loss": 0.4815, + "step": 7905 + }, + { + "epoch": 0.6737111205794631, + "grad_norm": 11.5625, + "learning_rate": 1.8726701872267309e-06, + "loss": 0.5942, + "step": 7906 + }, + { + "epoch": 0.6737963357477631, + "grad_norm": 12.9375, + "learning_rate": 1.8725244781721837e-06, + "loss": 0.5458, + "step": 7907 + }, + { + "epoch": 0.673881550916063, + "grad_norm": 14.875, + "learning_rate": 1.872378757867993e-06, + "loss": 0.7109, + "step": 7908 + }, + { + "epoch": 0.673966766084363, + "grad_norm": 18.125, + "learning_rate": 1.8722330263167926e-06, + "loss": 0.9752, + "step": 7909 + }, + { + "epoch": 0.674051981252663, + "grad_norm": 14.4375, + "learning_rate": 1.872087283521216e-06, + "loss": 0.6463, + "step": 7910 + }, + { + "epoch": 0.6741371964209629, + "grad_norm": 14.375, + "learning_rate": 1.8719415294838962e-06, + "loss": 0.5688, + "step": 7911 + }, + { + "epoch": 0.6742224115892629, + "grad_norm": 13.0625, + "learning_rate": 1.8717957642074675e-06, + "loss": 0.5499, + "step": 7912 + }, + { + "epoch": 0.6743076267575628, + "grad_norm": 13.6875, + "learning_rate": 1.871649987694564e-06, + "loss": 0.7316, + "step": 7913 + }, + { + "epoch": 0.6743928419258628, + "grad_norm": 18.125, + "learning_rate": 1.8715041999478204e-06, + "loss": 0.7385, + "step": 7914 + }, + { + "epoch": 0.6744780570941628, + "grad_norm": 14.75, + "learning_rate": 1.871358400969871e-06, + "loss": 0.6208, + "step": 7915 + }, + { + "epoch": 0.6745632722624627, + "grad_norm": 16.75, + "learning_rate": 1.87121259076335e-06, + "loss": 1.0545, + "step": 7916 + }, + { + "epoch": 0.6746484874307627, + "grad_norm": 13.9375, + "learning_rate": 1.8710667693308932e-06, + "loss": 0.6094, + "step": 7917 + }, + { + "epoch": 0.6747337025990626, + "grad_norm": 17.25, + "learning_rate": 1.8709209366751351e-06, + "loss": 0.787, + "step": 7918 + }, + { + "epoch": 0.6748189177673626, + "grad_norm": 16.5, + "learning_rate": 1.8707750927987116e-06, + "loss": 0.5271, + "step": 7919 + }, + { + "epoch": 0.6749041329356625, + "grad_norm": 12.8125, + "learning_rate": 1.8706292377042575e-06, + "loss": 0.4676, + "step": 7920 + }, + { + "epoch": 0.6749893481039625, + "grad_norm": 13.8125, + "learning_rate": 1.8704833713944095e-06, + "loss": 0.5728, + "step": 7921 + }, + { + "epoch": 0.6750745632722625, + "grad_norm": 13.8125, + "learning_rate": 1.8703374938718028e-06, + "loss": 0.779, + "step": 7922 + }, + { + "epoch": 0.6751597784405624, + "grad_norm": 15.0625, + "learning_rate": 1.8701916051390737e-06, + "loss": 0.7196, + "step": 7923 + }, + { + "epoch": 0.6752449936088624, + "grad_norm": 13.1875, + "learning_rate": 1.8700457051988585e-06, + "loss": 0.6841, + "step": 7924 + }, + { + "epoch": 0.6753302087771623, + "grad_norm": 16.375, + "learning_rate": 1.8698997940537944e-06, + "loss": 0.8734, + "step": 7925 + }, + { + "epoch": 0.6754154239454623, + "grad_norm": 13.6875, + "learning_rate": 1.8697538717065175e-06, + "loss": 0.8064, + "step": 7926 + }, + { + "epoch": 0.6755006391137622, + "grad_norm": 12.9375, + "learning_rate": 1.8696079381596647e-06, + "loss": 0.5771, + "step": 7927 + }, + { + "epoch": 0.6755858542820622, + "grad_norm": 10.5625, + "learning_rate": 1.8694619934158734e-06, + "loss": 0.3223, + "step": 7928 + }, + { + "epoch": 0.6756710694503621, + "grad_norm": 12.3125, + "learning_rate": 1.8693160374777809e-06, + "loss": 0.7083, + "step": 7929 + }, + { + "epoch": 0.6757562846186621, + "grad_norm": 12.875, + "learning_rate": 1.8691700703480254e-06, + "loss": 0.7299, + "step": 7930 + }, + { + "epoch": 0.6758414997869621, + "grad_norm": 13.5, + "learning_rate": 1.8690240920292436e-06, + "loss": 0.7834, + "step": 7931 + }, + { + "epoch": 0.675926714955262, + "grad_norm": 15.75, + "learning_rate": 1.868878102524074e-06, + "loss": 1.0039, + "step": 7932 + }, + { + "epoch": 0.676011930123562, + "grad_norm": 12.25, + "learning_rate": 1.8687321018351545e-06, + "loss": 0.8225, + "step": 7933 + }, + { + "epoch": 0.6760971452918619, + "grad_norm": 12.5, + "learning_rate": 1.8685860899651246e-06, + "loss": 0.423, + "step": 7934 + }, + { + "epoch": 0.6761823604601619, + "grad_norm": 18.375, + "learning_rate": 1.8684400669166214e-06, + "loss": 0.9114, + "step": 7935 + }, + { + "epoch": 0.6762675756284618, + "grad_norm": 10.3125, + "learning_rate": 1.8682940326922844e-06, + "loss": 0.8907, + "step": 7936 + }, + { + "epoch": 0.6763527907967618, + "grad_norm": 15.1875, + "learning_rate": 1.8681479872947523e-06, + "loss": 0.629, + "step": 7937 + }, + { + "epoch": 0.6764380059650618, + "grad_norm": 17.0, + "learning_rate": 1.8680019307266647e-06, + "loss": 0.7546, + "step": 7938 + }, + { + "epoch": 0.6765232211333617, + "grad_norm": 24.0, + "learning_rate": 1.8678558629906606e-06, + "loss": 1.1357, + "step": 7939 + }, + { + "epoch": 0.6766084363016617, + "grad_norm": 17.25, + "learning_rate": 1.8677097840893802e-06, + "loss": 1.2446, + "step": 7940 + }, + { + "epoch": 0.6766936514699616, + "grad_norm": 19.625, + "learning_rate": 1.8675636940254622e-06, + "loss": 1.1777, + "step": 7941 + }, + { + "epoch": 0.6767788666382616, + "grad_norm": 9.75, + "learning_rate": 1.8674175928015475e-06, + "loss": 0.4478, + "step": 7942 + }, + { + "epoch": 0.6768640818065615, + "grad_norm": 10.375, + "learning_rate": 1.8672714804202763e-06, + "loss": 0.2777, + "step": 7943 + }, + { + "epoch": 0.6769492969748615, + "grad_norm": 13.125, + "learning_rate": 1.8671253568842884e-06, + "loss": 0.498, + "step": 7944 + }, + { + "epoch": 0.6770345121431615, + "grad_norm": 13.5625, + "learning_rate": 1.8669792221962248e-06, + "loss": 0.4724, + "step": 7945 + }, + { + "epoch": 0.6771197273114614, + "grad_norm": 15.625, + "learning_rate": 1.866833076358726e-06, + "loss": 0.5586, + "step": 7946 + }, + { + "epoch": 0.6772049424797614, + "grad_norm": 12.4375, + "learning_rate": 1.8666869193744342e-06, + "loss": 0.6965, + "step": 7947 + }, + { + "epoch": 0.6772901576480613, + "grad_norm": 14.5625, + "learning_rate": 1.8665407512459887e-06, + "loss": 0.6138, + "step": 7948 + }, + { + "epoch": 0.6773753728163613, + "grad_norm": 13.125, + "learning_rate": 1.866394571976032e-06, + "loss": 0.6222, + "step": 7949 + }, + { + "epoch": 0.6774605879846612, + "grad_norm": 10.5, + "learning_rate": 1.8662483815672055e-06, + "loss": 0.4369, + "step": 7950 + }, + { + "epoch": 0.6775458031529612, + "grad_norm": 15.875, + "learning_rate": 1.866102180022152e-06, + "loss": 0.5167, + "step": 7951 + }, + { + "epoch": 0.6776310183212612, + "grad_norm": 16.375, + "learning_rate": 1.8659559673435113e-06, + "loss": 0.9183, + "step": 7952 + }, + { + "epoch": 0.6777162334895611, + "grad_norm": 15.6875, + "learning_rate": 1.8658097435339273e-06, + "loss": 0.808, + "step": 7953 + }, + { + "epoch": 0.6778014486578611, + "grad_norm": 12.1875, + "learning_rate": 1.865663508596042e-06, + "loss": 0.5715, + "step": 7954 + }, + { + "epoch": 0.677886663826161, + "grad_norm": 19.125, + "learning_rate": 1.8655172625324986e-06, + "loss": 1.043, + "step": 7955 + }, + { + "epoch": 0.677971878994461, + "grad_norm": 12.875, + "learning_rate": 1.8653710053459383e-06, + "loss": 0.6524, + "step": 7956 + }, + { + "epoch": 0.6780570941627609, + "grad_norm": 17.125, + "learning_rate": 1.8652247370390059e-06, + "loss": 0.853, + "step": 7957 + }, + { + "epoch": 0.6781423093310609, + "grad_norm": 14.875, + "learning_rate": 1.8650784576143433e-06, + "loss": 0.8039, + "step": 7958 + }, + { + "epoch": 0.6782275244993609, + "grad_norm": 13.0, + "learning_rate": 1.8649321670745948e-06, + "loss": 0.5593, + "step": 7959 + }, + { + "epoch": 0.6783127396676608, + "grad_norm": 14.0, + "learning_rate": 1.8647858654224039e-06, + "loss": 0.6878, + "step": 7960 + }, + { + "epoch": 0.6783979548359608, + "grad_norm": 15.3125, + "learning_rate": 1.8646395526604134e-06, + "loss": 0.69, + "step": 7961 + }, + { + "epoch": 0.6784831700042607, + "grad_norm": 29.25, + "learning_rate": 1.8644932287912686e-06, + "loss": 0.9136, + "step": 7962 + }, + { + "epoch": 0.6785683851725607, + "grad_norm": 13.625, + "learning_rate": 1.8643468938176134e-06, + "loss": 0.5083, + "step": 7963 + }, + { + "epoch": 0.6786536003408606, + "grad_norm": 18.5, + "learning_rate": 1.8642005477420922e-06, + "loss": 0.4323, + "step": 7964 + }, + { + "epoch": 0.6787388155091606, + "grad_norm": 17.25, + "learning_rate": 1.8640541905673487e-06, + "loss": 0.7278, + "step": 7965 + }, + { + "epoch": 0.6788240306774606, + "grad_norm": 16.0, + "learning_rate": 1.8639078222960292e-06, + "loss": 0.7939, + "step": 7966 + }, + { + "epoch": 0.6789092458457605, + "grad_norm": 13.625, + "learning_rate": 1.863761442930778e-06, + "loss": 0.6886, + "step": 7967 + }, + { + "epoch": 0.6789944610140605, + "grad_norm": 14.875, + "learning_rate": 1.86361505247424e-06, + "loss": 0.7742, + "step": 7968 + }, + { + "epoch": 0.6790796761823604, + "grad_norm": 18.75, + "learning_rate": 1.863468650929061e-06, + "loss": 0.9031, + "step": 7969 + }, + { + "epoch": 0.6791648913506604, + "grad_norm": 14.125, + "learning_rate": 1.8633222382978867e-06, + "loss": 0.8332, + "step": 7970 + }, + { + "epoch": 0.6792501065189603, + "grad_norm": 16.75, + "learning_rate": 1.8631758145833628e-06, + "loss": 0.5905, + "step": 7971 + }, + { + "epoch": 0.6793353216872603, + "grad_norm": 17.0, + "learning_rate": 1.8630293797881355e-06, + "loss": 1.1168, + "step": 7972 + }, + { + "epoch": 0.6794205368555603, + "grad_norm": 14.25, + "learning_rate": 1.8628829339148504e-06, + "loss": 0.6075, + "step": 7973 + }, + { + "epoch": 0.6795057520238602, + "grad_norm": 19.0, + "learning_rate": 1.862736476966155e-06, + "loss": 0.9521, + "step": 7974 + }, + { + "epoch": 0.6795909671921602, + "grad_norm": 19.0, + "learning_rate": 1.862590008944695e-06, + "loss": 0.9583, + "step": 7975 + }, + { + "epoch": 0.6796761823604601, + "grad_norm": 33.75, + "learning_rate": 1.8624435298531174e-06, + "loss": 1.3099, + "step": 7976 + }, + { + "epoch": 0.6797613975287601, + "grad_norm": 13.9375, + "learning_rate": 1.8622970396940693e-06, + "loss": 0.5251, + "step": 7977 + }, + { + "epoch": 0.67984661269706, + "grad_norm": 15.6875, + "learning_rate": 1.8621505384701982e-06, + "loss": 0.9311, + "step": 7978 + }, + { + "epoch": 0.67993182786536, + "grad_norm": 17.125, + "learning_rate": 1.8620040261841516e-06, + "loss": 0.9038, + "step": 7979 + }, + { + "epoch": 0.68001704303366, + "grad_norm": 12.5625, + "learning_rate": 1.861857502838576e-06, + "loss": 0.2994, + "step": 7980 + }, + { + "epoch": 0.6801022582019599, + "grad_norm": 10.5625, + "learning_rate": 1.8617109684361206e-06, + "loss": 0.3945, + "step": 7981 + }, + { + "epoch": 0.6801874733702599, + "grad_norm": 18.375, + "learning_rate": 1.8615644229794327e-06, + "loss": 1.1678, + "step": 7982 + }, + { + "epoch": 0.6802726885385598, + "grad_norm": 14.0625, + "learning_rate": 1.8614178664711613e-06, + "loss": 0.8588, + "step": 7983 + }, + { + "epoch": 0.6803579037068598, + "grad_norm": 20.5, + "learning_rate": 1.8612712989139534e-06, + "loss": 0.8733, + "step": 7984 + }, + { + "epoch": 0.6804431188751597, + "grad_norm": 13.125, + "learning_rate": 1.8611247203104587e-06, + "loss": 0.6981, + "step": 7985 + }, + { + "epoch": 0.6805283340434597, + "grad_norm": 11.4375, + "learning_rate": 1.860978130663326e-06, + "loss": 0.4357, + "step": 7986 + }, + { + "epoch": 0.6806135492117596, + "grad_norm": 12.75, + "learning_rate": 1.8608315299752042e-06, + "loss": 0.6975, + "step": 7987 + }, + { + "epoch": 0.6806987643800596, + "grad_norm": 16.25, + "learning_rate": 1.8606849182487418e-06, + "loss": 0.8316, + "step": 7988 + }, + { + "epoch": 0.6807839795483596, + "grad_norm": 17.0, + "learning_rate": 1.8605382954865894e-06, + "loss": 0.7927, + "step": 7989 + }, + { + "epoch": 0.6808691947166595, + "grad_norm": 18.875, + "learning_rate": 1.8603916616913959e-06, + "loss": 0.797, + "step": 7990 + }, + { + "epoch": 0.6809544098849595, + "grad_norm": 17.625, + "learning_rate": 1.8602450168658113e-06, + "loss": 0.904, + "step": 7991 + }, + { + "epoch": 0.6810396250532594, + "grad_norm": 17.0, + "learning_rate": 1.8600983610124853e-06, + "loss": 0.5518, + "step": 7992 + }, + { + "epoch": 0.6811248402215594, + "grad_norm": 17.875, + "learning_rate": 1.8599516941340689e-06, + "loss": 0.8319, + "step": 7993 + }, + { + "epoch": 0.6812100553898593, + "grad_norm": 18.75, + "learning_rate": 1.8598050162332116e-06, + "loss": 1.0023, + "step": 7994 + }, + { + "epoch": 0.6812952705581593, + "grad_norm": 11.4375, + "learning_rate": 1.8596583273125646e-06, + "loss": 0.4096, + "step": 7995 + }, + { + "epoch": 0.6813804857264593, + "grad_norm": 12.5625, + "learning_rate": 1.8595116273747787e-06, + "loss": 0.3719, + "step": 7996 + }, + { + "epoch": 0.6814657008947592, + "grad_norm": 15.5625, + "learning_rate": 1.8593649164225044e-06, + "loss": 0.8419, + "step": 7997 + }, + { + "epoch": 0.6815509160630592, + "grad_norm": 11.5625, + "learning_rate": 1.8592181944583938e-06, + "loss": 0.532, + "step": 7998 + }, + { + "epoch": 0.6816361312313591, + "grad_norm": 15.375, + "learning_rate": 1.8590714614850975e-06, + "loss": 0.5619, + "step": 7999 + }, + { + "epoch": 0.6817213463996591, + "grad_norm": 15.875, + "learning_rate": 1.8589247175052677e-06, + "loss": 0.9156, + "step": 8000 + }, + { + "epoch": 0.681806561567959, + "grad_norm": 13.375, + "learning_rate": 1.8587779625215554e-06, + "loss": 0.6071, + "step": 8001 + }, + { + "epoch": 0.681891776736259, + "grad_norm": 17.25, + "learning_rate": 1.8586311965366138e-06, + "loss": 0.9373, + "step": 8002 + }, + { + "epoch": 0.681976991904559, + "grad_norm": 21.5, + "learning_rate": 1.8584844195530939e-06, + "loss": 0.9696, + "step": 8003 + }, + { + "epoch": 0.6820622070728589, + "grad_norm": 16.625, + "learning_rate": 1.8583376315736493e-06, + "loss": 0.7568, + "step": 8004 + }, + { + "epoch": 0.6821474222411589, + "grad_norm": 12.875, + "learning_rate": 1.8581908326009312e-06, + "loss": 0.4698, + "step": 8005 + }, + { + "epoch": 0.6822326374094588, + "grad_norm": 11.6875, + "learning_rate": 1.8580440226375937e-06, + "loss": 0.5944, + "step": 8006 + }, + { + "epoch": 0.6823178525777588, + "grad_norm": 17.75, + "learning_rate": 1.857897201686289e-06, + "loss": 1.0195, + "step": 8007 + }, + { + "epoch": 0.6824030677460587, + "grad_norm": 17.125, + "learning_rate": 1.857750369749671e-06, + "loss": 0.8105, + "step": 8008 + }, + { + "epoch": 0.6824882829143587, + "grad_norm": 10.4375, + "learning_rate": 1.8576035268303922e-06, + "loss": 0.4069, + "step": 8009 + }, + { + "epoch": 0.6825734980826587, + "grad_norm": 11.5, + "learning_rate": 1.8574566729311066e-06, + "loss": 0.4938, + "step": 8010 + }, + { + "epoch": 0.6826587132509586, + "grad_norm": 16.125, + "learning_rate": 1.8573098080544682e-06, + "loss": 0.8314, + "step": 8011 + }, + { + "epoch": 0.6827439284192587, + "grad_norm": 13.625, + "learning_rate": 1.8571629322031312e-06, + "loss": 0.809, + "step": 8012 + }, + { + "epoch": 0.6828291435875586, + "grad_norm": 17.625, + "learning_rate": 1.8570160453797493e-06, + "loss": 0.9014, + "step": 8013 + }, + { + "epoch": 0.6829143587558586, + "grad_norm": 12.5625, + "learning_rate": 1.8568691475869768e-06, + "loss": 0.3936, + "step": 8014 + }, + { + "epoch": 0.6829995739241586, + "grad_norm": 11.3125, + "learning_rate": 1.856722238827469e-06, + "loss": 0.472, + "step": 8015 + }, + { + "epoch": 0.6830847890924585, + "grad_norm": 15.1875, + "learning_rate": 1.85657531910388e-06, + "loss": 0.6809, + "step": 8016 + }, + { + "epoch": 0.6831700042607585, + "grad_norm": 19.0, + "learning_rate": 1.856428388418865e-06, + "loss": 0.6864, + "step": 8017 + }, + { + "epoch": 0.6832552194290584, + "grad_norm": 25.625, + "learning_rate": 1.8562814467750791e-06, + "loss": 1.1295, + "step": 8018 + }, + { + "epoch": 0.6833404345973584, + "grad_norm": 13.9375, + "learning_rate": 1.856134494175178e-06, + "loss": 0.8235, + "step": 8019 + }, + { + "epoch": 0.6834256497656583, + "grad_norm": 13.25, + "learning_rate": 1.8559875306218172e-06, + "loss": 0.8916, + "step": 8020 + }, + { + "epoch": 0.6835108649339583, + "grad_norm": 13.9375, + "learning_rate": 1.855840556117652e-06, + "loss": 0.7067, + "step": 8021 + }, + { + "epoch": 0.6835960801022583, + "grad_norm": 11.0, + "learning_rate": 1.855693570665339e-06, + "loss": 0.5367, + "step": 8022 + }, + { + "epoch": 0.6836812952705582, + "grad_norm": 13.8125, + "learning_rate": 1.855546574267534e-06, + "loss": 0.8688, + "step": 8023 + }, + { + "epoch": 0.6837665104388582, + "grad_norm": 11.0625, + "learning_rate": 1.8553995669268935e-06, + "loss": 0.5116, + "step": 8024 + }, + { + "epoch": 0.6838517256071581, + "grad_norm": 12.5625, + "learning_rate": 1.855252548646074e-06, + "loss": 0.7576, + "step": 8025 + }, + { + "epoch": 0.6839369407754581, + "grad_norm": 14.5625, + "learning_rate": 1.8551055194277324e-06, + "loss": 0.7792, + "step": 8026 + }, + { + "epoch": 0.684022155943758, + "grad_norm": 19.0, + "learning_rate": 1.8549584792745254e-06, + "loss": 0.968, + "step": 8027 + }, + { + "epoch": 0.684107371112058, + "grad_norm": 12.1875, + "learning_rate": 1.8548114281891104e-06, + "loss": 0.5374, + "step": 8028 + }, + { + "epoch": 0.684192586280358, + "grad_norm": 16.25, + "learning_rate": 1.8546643661741448e-06, + "loss": 0.9379, + "step": 8029 + }, + { + "epoch": 0.6842778014486579, + "grad_norm": 16.0, + "learning_rate": 1.854517293232286e-06, + "loss": 0.7617, + "step": 8030 + }, + { + "epoch": 0.6843630166169579, + "grad_norm": 15.6875, + "learning_rate": 1.8543702093661918e-06, + "loss": 0.7383, + "step": 8031 + }, + { + "epoch": 0.6844482317852578, + "grad_norm": 14.0, + "learning_rate": 1.8542231145785205e-06, + "loss": 0.6043, + "step": 8032 + }, + { + "epoch": 0.6845334469535578, + "grad_norm": 14.0, + "learning_rate": 1.8540760088719295e-06, + "loss": 0.7912, + "step": 8033 + }, + { + "epoch": 0.6846186621218577, + "grad_norm": 14.875, + "learning_rate": 1.8539288922490778e-06, + "loss": 0.9933, + "step": 8034 + }, + { + "epoch": 0.6847038772901577, + "grad_norm": 12.25, + "learning_rate": 1.8537817647126235e-06, + "loss": 0.6488, + "step": 8035 + }, + { + "epoch": 0.6847890924584576, + "grad_norm": 12.0625, + "learning_rate": 1.853634626265226e-06, + "loss": 0.6229, + "step": 8036 + }, + { + "epoch": 0.6848743076267576, + "grad_norm": 10.25, + "learning_rate": 1.8534874769095434e-06, + "loss": 0.4627, + "step": 8037 + }, + { + "epoch": 0.6849595227950576, + "grad_norm": 12.5625, + "learning_rate": 1.8533403166482356e-06, + "loss": 1.0666, + "step": 8038 + }, + { + "epoch": 0.6850447379633575, + "grad_norm": 12.1875, + "learning_rate": 1.853193145483961e-06, + "loss": 0.4895, + "step": 8039 + }, + { + "epoch": 0.6851299531316575, + "grad_norm": 13.9375, + "learning_rate": 1.8530459634193805e-06, + "loss": 0.5945, + "step": 8040 + }, + { + "epoch": 0.6852151682999574, + "grad_norm": 30.75, + "learning_rate": 1.8528987704571522e-06, + "loss": 1.6107, + "step": 8041 + }, + { + "epoch": 0.6853003834682574, + "grad_norm": 11.0625, + "learning_rate": 1.8527515665999375e-06, + "loss": 0.6774, + "step": 8042 + }, + { + "epoch": 0.6853855986365573, + "grad_norm": 13.0, + "learning_rate": 1.8526043518503953e-06, + "loss": 0.6746, + "step": 8043 + }, + { + "epoch": 0.6854708138048573, + "grad_norm": 15.0, + "learning_rate": 1.8524571262111869e-06, + "loss": 0.5955, + "step": 8044 + }, + { + "epoch": 0.6855560289731573, + "grad_norm": 18.125, + "learning_rate": 1.8523098896849723e-06, + "loss": 1.0742, + "step": 8045 + }, + { + "epoch": 0.6856412441414572, + "grad_norm": 17.125, + "learning_rate": 1.852162642274412e-06, + "loss": 0.1923, + "step": 8046 + }, + { + "epoch": 0.6857264593097572, + "grad_norm": 13.375, + "learning_rate": 1.852015383982168e-06, + "loss": 0.41, + "step": 8047 + }, + { + "epoch": 0.6858116744780571, + "grad_norm": 20.375, + "learning_rate": 1.8518681148109002e-06, + "loss": 0.6981, + "step": 8048 + }, + { + "epoch": 0.6858968896463571, + "grad_norm": 23.0, + "learning_rate": 1.8517208347632705e-06, + "loss": 1.0914, + "step": 8049 + }, + { + "epoch": 0.685982104814657, + "grad_norm": 14.5, + "learning_rate": 1.8515735438419396e-06, + "loss": 0.7991, + "step": 8050 + }, + { + "epoch": 0.686067319982957, + "grad_norm": 15.125, + "learning_rate": 1.8514262420495704e-06, + "loss": 0.8384, + "step": 8051 + }, + { + "epoch": 0.686152535151257, + "grad_norm": 11.4375, + "learning_rate": 1.8512789293888244e-06, + "loss": 0.4543, + "step": 8052 + }, + { + "epoch": 0.6862377503195569, + "grad_norm": 13.9375, + "learning_rate": 1.8511316058623635e-06, + "loss": 0.8889, + "step": 8053 + }, + { + "epoch": 0.6863229654878569, + "grad_norm": 12.8125, + "learning_rate": 1.8509842714728493e-06, + "loss": 0.6106, + "step": 8054 + }, + { + "epoch": 0.6864081806561568, + "grad_norm": 13.625, + "learning_rate": 1.8508369262229457e-06, + "loss": 0.8317, + "step": 8055 + }, + { + "epoch": 0.6864933958244568, + "grad_norm": 17.75, + "learning_rate": 1.8506895701153144e-06, + "loss": 0.6371, + "step": 8056 + }, + { + "epoch": 0.6865786109927567, + "grad_norm": 10.875, + "learning_rate": 1.8505422031526186e-06, + "loss": 0.4233, + "step": 8057 + }, + { + "epoch": 0.6866638261610567, + "grad_norm": 11.5625, + "learning_rate": 1.8503948253375209e-06, + "loss": 0.4715, + "step": 8058 + }, + { + "epoch": 0.6867490413293567, + "grad_norm": 14.75, + "learning_rate": 1.8502474366726856e-06, + "loss": 0.7788, + "step": 8059 + }, + { + "epoch": 0.6868342564976566, + "grad_norm": 15.875, + "learning_rate": 1.8501000371607752e-06, + "loss": 0.7568, + "step": 8060 + }, + { + "epoch": 0.6869194716659566, + "grad_norm": 14.3125, + "learning_rate": 1.8499526268044534e-06, + "loss": 0.7253, + "step": 8061 + }, + { + "epoch": 0.6870046868342565, + "grad_norm": 16.125, + "learning_rate": 1.8498052056063847e-06, + "loss": 0.7275, + "step": 8062 + }, + { + "epoch": 0.6870899020025565, + "grad_norm": 28.625, + "learning_rate": 1.849657773569233e-06, + "loss": 1.1963, + "step": 8063 + }, + { + "epoch": 0.6871751171708564, + "grad_norm": 14.5, + "learning_rate": 1.8495103306956618e-06, + "loss": 0.7997, + "step": 8064 + }, + { + "epoch": 0.6872603323391564, + "grad_norm": 14.4375, + "learning_rate": 1.8493628769883363e-06, + "loss": 0.7399, + "step": 8065 + }, + { + "epoch": 0.6873455475074564, + "grad_norm": 21.125, + "learning_rate": 1.8492154124499208e-06, + "loss": 0.8936, + "step": 8066 + }, + { + "epoch": 0.6874307626757563, + "grad_norm": 14.25, + "learning_rate": 1.8490679370830804e-06, + "loss": 0.7956, + "step": 8067 + }, + { + "epoch": 0.6875159778440563, + "grad_norm": 12.0, + "learning_rate": 1.84892045089048e-06, + "loss": 0.4879, + "step": 8068 + }, + { + "epoch": 0.6876011930123562, + "grad_norm": 15.4375, + "learning_rate": 1.8487729538747847e-06, + "loss": 0.7588, + "step": 8069 + }, + { + "epoch": 0.6876864081806562, + "grad_norm": 11.6875, + "learning_rate": 1.8486254460386602e-06, + "loss": 0.4672, + "step": 8070 + }, + { + "epoch": 0.6877716233489561, + "grad_norm": 18.125, + "learning_rate": 1.8484779273847713e-06, + "loss": 1.003, + "step": 8071 + }, + { + "epoch": 0.6878568385172561, + "grad_norm": 18.625, + "learning_rate": 1.8483303979157851e-06, + "loss": 0.7874, + "step": 8072 + }, + { + "epoch": 0.687942053685556, + "grad_norm": 46.0, + "learning_rate": 1.8481828576343668e-06, + "loss": 1.5661, + "step": 8073 + }, + { + "epoch": 0.688027268853856, + "grad_norm": 15.75, + "learning_rate": 1.8480353065431828e-06, + "loss": 0.6623, + "step": 8074 + }, + { + "epoch": 0.688112484022156, + "grad_norm": 17.375, + "learning_rate": 1.847887744644899e-06, + "loss": 0.987, + "step": 8075 + }, + { + "epoch": 0.6881976991904559, + "grad_norm": 16.5, + "learning_rate": 1.8477401719421827e-06, + "loss": 0.8627, + "step": 8076 + }, + { + "epoch": 0.6882829143587559, + "grad_norm": 18.0, + "learning_rate": 1.8475925884377008e-06, + "loss": 0.9467, + "step": 8077 + }, + { + "epoch": 0.6883681295270558, + "grad_norm": 16.5, + "learning_rate": 1.8474449941341197e-06, + "loss": 0.7599, + "step": 8078 + }, + { + "epoch": 0.6884533446953558, + "grad_norm": 14.3125, + "learning_rate": 1.8472973890341062e-06, + "loss": 0.5263, + "step": 8079 + }, + { + "epoch": 0.6885385598636558, + "grad_norm": 18.125, + "learning_rate": 1.8471497731403287e-06, + "loss": 1.1068, + "step": 8080 + }, + { + "epoch": 0.6886237750319557, + "grad_norm": 11.75, + "learning_rate": 1.8470021464554545e-06, + "loss": 0.403, + "step": 8081 + }, + { + "epoch": 0.6887089902002557, + "grad_norm": 12.4375, + "learning_rate": 1.846854508982151e-06, + "loss": 0.5711, + "step": 8082 + }, + { + "epoch": 0.6887942053685556, + "grad_norm": 16.25, + "learning_rate": 1.8467068607230861e-06, + "loss": 0.3717, + "step": 8083 + }, + { + "epoch": 0.6888794205368556, + "grad_norm": 10.5625, + "learning_rate": 1.8465592016809283e-06, + "loss": 0.3918, + "step": 8084 + }, + { + "epoch": 0.6889646357051555, + "grad_norm": 23.0, + "learning_rate": 1.8464115318583464e-06, + "loss": 0.3721, + "step": 8085 + }, + { + "epoch": 0.6890498508734555, + "grad_norm": 11.9375, + "learning_rate": 1.8462638512580078e-06, + "loss": 0.575, + "step": 8086 + }, + { + "epoch": 0.6891350660417555, + "grad_norm": 16.125, + "learning_rate": 1.846116159882582e-06, + "loss": 0.8211, + "step": 8087 + }, + { + "epoch": 0.6892202812100554, + "grad_norm": 13.125, + "learning_rate": 1.8459684577347376e-06, + "loss": 0.7709, + "step": 8088 + }, + { + "epoch": 0.6893054963783554, + "grad_norm": 13.25, + "learning_rate": 1.8458207448171442e-06, + "loss": 0.5794, + "step": 8089 + }, + { + "epoch": 0.6893907115466553, + "grad_norm": 22.125, + "learning_rate": 1.8456730211324703e-06, + "loss": 1.0998, + "step": 8090 + }, + { + "epoch": 0.6894759267149553, + "grad_norm": 15.75, + "learning_rate": 1.8455252866833864e-06, + "loss": 0.7661, + "step": 8091 + }, + { + "epoch": 0.6895611418832552, + "grad_norm": 12.5625, + "learning_rate": 1.8453775414725613e-06, + "loss": 0.6825, + "step": 8092 + }, + { + "epoch": 0.6896463570515552, + "grad_norm": 30.5, + "learning_rate": 1.8452297855026656e-06, + "loss": 0.9084, + "step": 8093 + }, + { + "epoch": 0.6897315722198551, + "grad_norm": 14.0, + "learning_rate": 1.8450820187763691e-06, + "loss": 0.6193, + "step": 8094 + }, + { + "epoch": 0.6898167873881551, + "grad_norm": 13.375, + "learning_rate": 1.8449342412963422e-06, + "loss": 0.6635, + "step": 8095 + }, + { + "epoch": 0.6899020025564551, + "grad_norm": 13.125, + "learning_rate": 1.8447864530652552e-06, + "loss": 1.0097, + "step": 8096 + }, + { + "epoch": 0.689987217724755, + "grad_norm": 16.75, + "learning_rate": 1.844638654085779e-06, + "loss": 1.0979, + "step": 8097 + }, + { + "epoch": 0.690072432893055, + "grad_norm": 13.25, + "learning_rate": 1.8444908443605842e-06, + "loss": 0.6398, + "step": 8098 + }, + { + "epoch": 0.6901576480613549, + "grad_norm": 11.0625, + "learning_rate": 1.8443430238923418e-06, + "loss": 0.6056, + "step": 8099 + }, + { + "epoch": 0.6902428632296549, + "grad_norm": 30.5, + "learning_rate": 1.8441951926837237e-06, + "loss": 1.3652, + "step": 8100 + }, + { + "epoch": 0.6903280783979548, + "grad_norm": 16.75, + "learning_rate": 1.8440473507374007e-06, + "loss": 1.1108, + "step": 8101 + }, + { + "epoch": 0.6904132935662548, + "grad_norm": 15.0625, + "learning_rate": 1.8438994980560451e-06, + "loss": 0.9355, + "step": 8102 + }, + { + "epoch": 0.6904985087345548, + "grad_norm": 15.8125, + "learning_rate": 1.8437516346423279e-06, + "loss": 0.6826, + "step": 8103 + }, + { + "epoch": 0.6905837239028547, + "grad_norm": 26.875, + "learning_rate": 1.843603760498922e-06, + "loss": 0.7131, + "step": 8104 + }, + { + "epoch": 0.6906689390711547, + "grad_norm": 17.375, + "learning_rate": 1.843455875628499e-06, + "loss": 0.6287, + "step": 8105 + }, + { + "epoch": 0.6907541542394546, + "grad_norm": 15.8125, + "learning_rate": 1.8433079800337317e-06, + "loss": 0.8101, + "step": 8106 + }, + { + "epoch": 0.6908393694077546, + "grad_norm": 18.125, + "learning_rate": 1.8431600737172922e-06, + "loss": 1.0646, + "step": 8107 + }, + { + "epoch": 0.6909245845760545, + "grad_norm": 14.6875, + "learning_rate": 1.8430121566818543e-06, + "loss": 0.7475, + "step": 8108 + }, + { + "epoch": 0.6910097997443545, + "grad_norm": 17.625, + "learning_rate": 1.8428642289300903e-06, + "loss": 0.9519, + "step": 8109 + }, + { + "epoch": 0.6910950149126545, + "grad_norm": 14.6875, + "learning_rate": 1.8427162904646731e-06, + "loss": 0.7998, + "step": 8110 + }, + { + "epoch": 0.6911802300809544, + "grad_norm": 19.625, + "learning_rate": 1.8425683412882766e-06, + "loss": 1.1156, + "step": 8111 + }, + { + "epoch": 0.6912654452492544, + "grad_norm": 12.5, + "learning_rate": 1.8424203814035746e-06, + "loss": 0.569, + "step": 8112 + }, + { + "epoch": 0.6913506604175543, + "grad_norm": 11.9375, + "learning_rate": 1.8422724108132404e-06, + "loss": 0.7167, + "step": 8113 + }, + { + "epoch": 0.6914358755858543, + "grad_norm": 26.25, + "learning_rate": 1.8421244295199481e-06, + "loss": 0.7344, + "step": 8114 + }, + { + "epoch": 0.6915210907541542, + "grad_norm": 12.25, + "learning_rate": 1.8419764375263718e-06, + "loss": 0.6383, + "step": 8115 + }, + { + "epoch": 0.6916063059224542, + "grad_norm": 14.9375, + "learning_rate": 1.8418284348351862e-06, + "loss": 0.7088, + "step": 8116 + }, + { + "epoch": 0.6916915210907542, + "grad_norm": 21.25, + "learning_rate": 1.8416804214490658e-06, + "loss": 0.537, + "step": 8117 + }, + { + "epoch": 0.6917767362590541, + "grad_norm": 16.375, + "learning_rate": 1.841532397370685e-06, + "loss": 0.789, + "step": 8118 + }, + { + "epoch": 0.6918619514273541, + "grad_norm": 12.5625, + "learning_rate": 1.8413843626027189e-06, + "loss": 0.4099, + "step": 8119 + }, + { + "epoch": 0.691947166595654, + "grad_norm": 13.6875, + "learning_rate": 1.8412363171478426e-06, + "loss": 0.6251, + "step": 8120 + }, + { + "epoch": 0.692032381763954, + "grad_norm": 14.0625, + "learning_rate": 1.8410882610087322e-06, + "loss": 0.8444, + "step": 8121 + }, + { + "epoch": 0.6921175969322539, + "grad_norm": 10.6875, + "learning_rate": 1.8409401941880618e-06, + "loss": 0.5268, + "step": 8122 + }, + { + "epoch": 0.6922028121005539, + "grad_norm": 9.8125, + "learning_rate": 1.8407921166885082e-06, + "loss": 0.3121, + "step": 8123 + }, + { + "epoch": 0.6922880272688539, + "grad_norm": 16.125, + "learning_rate": 1.8406440285127466e-06, + "loss": 0.897, + "step": 8124 + }, + { + "epoch": 0.6923732424371538, + "grad_norm": 27.0, + "learning_rate": 1.840495929663454e-06, + "loss": 0.8406, + "step": 8125 + }, + { + "epoch": 0.6924584576054538, + "grad_norm": 14.625, + "learning_rate": 1.8403478201433055e-06, + "loss": 0.6323, + "step": 8126 + }, + { + "epoch": 0.6925436727737537, + "grad_norm": 18.375, + "learning_rate": 1.840199699954979e-06, + "loss": 0.4347, + "step": 8127 + }, + { + "epoch": 0.6926288879420537, + "grad_norm": 16.625, + "learning_rate": 1.8400515691011496e-06, + "loss": 0.8205, + "step": 8128 + }, + { + "epoch": 0.6927141031103536, + "grad_norm": 13.25, + "learning_rate": 1.8399034275844957e-06, + "loss": 0.6021, + "step": 8129 + }, + { + "epoch": 0.6927993182786536, + "grad_norm": 14.625, + "learning_rate": 1.8397552754076937e-06, + "loss": 0.8508, + "step": 8130 + }, + { + "epoch": 0.6928845334469536, + "grad_norm": 19.625, + "learning_rate": 1.8396071125734202e-06, + "loss": 0.7914, + "step": 8131 + }, + { + "epoch": 0.6929697486152535, + "grad_norm": 33.25, + "learning_rate": 1.8394589390843536e-06, + "loss": 1.1633, + "step": 8132 + }, + { + "epoch": 0.6930549637835535, + "grad_norm": 16.875, + "learning_rate": 1.8393107549431716e-06, + "loss": 0.692, + "step": 8133 + }, + { + "epoch": 0.6931401789518534, + "grad_norm": 12.375, + "learning_rate": 1.839162560152551e-06, + "loss": 0.4968, + "step": 8134 + }, + { + "epoch": 0.6932253941201534, + "grad_norm": 16.125, + "learning_rate": 1.839014354715171e-06, + "loss": 0.813, + "step": 8135 + }, + { + "epoch": 0.6933106092884533, + "grad_norm": 15.3125, + "learning_rate": 1.8388661386337093e-06, + "loss": 0.8119, + "step": 8136 + }, + { + "epoch": 0.6933958244567533, + "grad_norm": 13.875, + "learning_rate": 1.8387179119108439e-06, + "loss": 0.6631, + "step": 8137 + }, + { + "epoch": 0.6934810396250533, + "grad_norm": 10.6875, + "learning_rate": 1.8385696745492542e-06, + "loss": 0.4312, + "step": 8138 + }, + { + "epoch": 0.6935662547933532, + "grad_norm": 11.3125, + "learning_rate": 1.8384214265516184e-06, + "loss": 0.3797, + "step": 8139 + }, + { + "epoch": 0.6936514699616532, + "grad_norm": 9.25, + "learning_rate": 1.838273167920616e-06, + "loss": 0.3417, + "step": 8140 + }, + { + "epoch": 0.6937366851299531, + "grad_norm": 17.625, + "learning_rate": 1.8381248986589253e-06, + "loss": 1.0847, + "step": 8141 + }, + { + "epoch": 0.6938219002982531, + "grad_norm": 17.125, + "learning_rate": 1.8379766187692272e-06, + "loss": 0.7426, + "step": 8142 + }, + { + "epoch": 0.693907115466553, + "grad_norm": 13.75, + "learning_rate": 1.8378283282541997e-06, + "loss": 0.8764, + "step": 8143 + }, + { + "epoch": 0.693992330634853, + "grad_norm": 23.25, + "learning_rate": 1.8376800271165234e-06, + "loss": 0.6341, + "step": 8144 + }, + { + "epoch": 0.694077545803153, + "grad_norm": 9.625, + "learning_rate": 1.8375317153588778e-06, + "loss": 0.335, + "step": 8145 + }, + { + "epoch": 0.6941627609714529, + "grad_norm": 19.125, + "learning_rate": 1.8373833929839436e-06, + "loss": 0.8884, + "step": 8146 + }, + { + "epoch": 0.6942479761397529, + "grad_norm": 13.875, + "learning_rate": 1.8372350599944008e-06, + "loss": 0.4556, + "step": 8147 + }, + { + "epoch": 0.6943331913080528, + "grad_norm": 30.5, + "learning_rate": 1.83708671639293e-06, + "loss": 1.385, + "step": 8148 + }, + { + "epoch": 0.6944184064763528, + "grad_norm": 13.5625, + "learning_rate": 1.8369383621822116e-06, + "loss": 0.4308, + "step": 8149 + }, + { + "epoch": 0.6945036216446527, + "grad_norm": 13.9375, + "learning_rate": 1.8367899973649267e-06, + "loss": 0.8422, + "step": 8150 + }, + { + "epoch": 0.6945888368129527, + "grad_norm": 18.375, + "learning_rate": 1.836641621943757e-06, + "loss": 0.8391, + "step": 8151 + }, + { + "epoch": 0.6946740519812526, + "grad_norm": 12.8125, + "learning_rate": 1.8364932359213828e-06, + "loss": 0.5521, + "step": 8152 + }, + { + "epoch": 0.6947592671495526, + "grad_norm": 16.25, + "learning_rate": 1.8363448393004862e-06, + "loss": 0.9841, + "step": 8153 + }, + { + "epoch": 0.6948444823178526, + "grad_norm": 14.3125, + "learning_rate": 1.836196432083749e-06, + "loss": 0.5429, + "step": 8154 + }, + { + "epoch": 0.6949296974861525, + "grad_norm": 10.875, + "learning_rate": 1.8360480142738527e-06, + "loss": 0.4257, + "step": 8155 + }, + { + "epoch": 0.6950149126544525, + "grad_norm": 14.625, + "learning_rate": 1.835899585873479e-06, + "loss": 0.9718, + "step": 8156 + }, + { + "epoch": 0.6951001278227524, + "grad_norm": 15.1875, + "learning_rate": 1.835751146885311e-06, + "loss": 0.8682, + "step": 8157 + }, + { + "epoch": 0.6951853429910524, + "grad_norm": 14.6875, + "learning_rate": 1.835602697312031e-06, + "loss": 0.6956, + "step": 8158 + }, + { + "epoch": 0.6952705581593523, + "grad_norm": 12.625, + "learning_rate": 1.8354542371563212e-06, + "loss": 0.5683, + "step": 8159 + }, + { + "epoch": 0.6953557733276523, + "grad_norm": 13.25, + "learning_rate": 1.8353057664208646e-06, + "loss": 0.7104, + "step": 8160 + }, + { + "epoch": 0.6954409884959523, + "grad_norm": 15.5625, + "learning_rate": 1.8351572851083442e-06, + "loss": 0.5315, + "step": 8161 + }, + { + "epoch": 0.6955262036642522, + "grad_norm": 26.625, + "learning_rate": 1.835008793221443e-06, + "loss": 0.9861, + "step": 8162 + }, + { + "epoch": 0.6956114188325522, + "grad_norm": 18.125, + "learning_rate": 1.8348602907628448e-06, + "loss": 1.0019, + "step": 8163 + }, + { + "epoch": 0.6956966340008521, + "grad_norm": 17.375, + "learning_rate": 1.834711777735233e-06, + "loss": 1.0653, + "step": 8164 + }, + { + "epoch": 0.6957818491691521, + "grad_norm": 16.5, + "learning_rate": 1.8345632541412917e-06, + "loss": 0.6701, + "step": 8165 + }, + { + "epoch": 0.695867064337452, + "grad_norm": 9.375, + "learning_rate": 1.8344147199837041e-06, + "loss": 0.281, + "step": 8166 + }, + { + "epoch": 0.695952279505752, + "grad_norm": 15.9375, + "learning_rate": 1.8342661752651554e-06, + "loss": 0.9143, + "step": 8167 + }, + { + "epoch": 0.696037494674052, + "grad_norm": 13.25, + "learning_rate": 1.834117619988329e-06, + "loss": 0.7012, + "step": 8168 + }, + { + "epoch": 0.6961227098423519, + "grad_norm": 16.375, + "learning_rate": 1.8339690541559097e-06, + "loss": 0.7927, + "step": 8169 + }, + { + "epoch": 0.6962079250106519, + "grad_norm": 12.5, + "learning_rate": 1.833820477770583e-06, + "loss": 0.6952, + "step": 8170 + }, + { + "epoch": 0.6962931401789518, + "grad_norm": 12.1875, + "learning_rate": 1.8336718908350326e-06, + "loss": 0.5869, + "step": 8171 + }, + { + "epoch": 0.6963783553472518, + "grad_norm": 9.5, + "learning_rate": 1.8335232933519443e-06, + "loss": 0.2891, + "step": 8172 + }, + { + "epoch": 0.6964635705155517, + "grad_norm": 17.125, + "learning_rate": 1.833374685324003e-06, + "loss": 0.9662, + "step": 8173 + }, + { + "epoch": 0.6965487856838517, + "grad_norm": 13.9375, + "learning_rate": 1.8332260667538953e-06, + "loss": 0.6379, + "step": 8174 + }, + { + "epoch": 0.6966340008521517, + "grad_norm": 16.125, + "learning_rate": 1.8330774376443056e-06, + "loss": 0.9416, + "step": 8175 + }, + { + "epoch": 0.6967192160204516, + "grad_norm": 16.0, + "learning_rate": 1.8329287979979205e-06, + "loss": 0.9773, + "step": 8176 + }, + { + "epoch": 0.6968044311887516, + "grad_norm": 14.9375, + "learning_rate": 1.8327801478174253e-06, + "loss": 0.9329, + "step": 8177 + }, + { + "epoch": 0.6968896463570515, + "grad_norm": 14.4375, + "learning_rate": 1.8326314871055075e-06, + "loss": 0.9319, + "step": 8178 + }, + { + "epoch": 0.6969748615253515, + "grad_norm": 13.5625, + "learning_rate": 1.8324828158648525e-06, + "loss": 0.5919, + "step": 8179 + }, + { + "epoch": 0.6970600766936514, + "grad_norm": 13.4375, + "learning_rate": 1.8323341340981474e-06, + "loss": 0.6949, + "step": 8180 + }, + { + "epoch": 0.6971452918619514, + "grad_norm": 20.125, + "learning_rate": 1.8321854418080786e-06, + "loss": 1.0211, + "step": 8181 + }, + { + "epoch": 0.6972305070302514, + "grad_norm": 22.25, + "learning_rate": 1.8320367389973337e-06, + "loss": 1.0448, + "step": 8182 + }, + { + "epoch": 0.6973157221985513, + "grad_norm": 14.9375, + "learning_rate": 1.8318880256685998e-06, + "loss": 0.7677, + "step": 8183 + }, + { + "epoch": 0.6974009373668513, + "grad_norm": 15.5625, + "learning_rate": 1.8317393018245638e-06, + "loss": 0.7157, + "step": 8184 + }, + { + "epoch": 0.6974861525351512, + "grad_norm": 15.75, + "learning_rate": 1.8315905674679139e-06, + "loss": 0.8377, + "step": 8185 + }, + { + "epoch": 0.6975713677034512, + "grad_norm": 11.125, + "learning_rate": 1.8314418226013375e-06, + "loss": 0.3535, + "step": 8186 + }, + { + "epoch": 0.6976565828717511, + "grad_norm": 13.3125, + "learning_rate": 1.8312930672275229e-06, + "loss": 0.7206, + "step": 8187 + }, + { + "epoch": 0.6977417980400511, + "grad_norm": 13.125, + "learning_rate": 1.8311443013491576e-06, + "loss": 0.5121, + "step": 8188 + }, + { + "epoch": 0.697827013208351, + "grad_norm": 14.4375, + "learning_rate": 1.8309955249689305e-06, + "loss": 0.8198, + "step": 8189 + }, + { + "epoch": 0.697912228376651, + "grad_norm": 23.25, + "learning_rate": 1.8308467380895304e-06, + "loss": 0.5556, + "step": 8190 + }, + { + "epoch": 0.697997443544951, + "grad_norm": 15.625, + "learning_rate": 1.8306979407136452e-06, + "loss": 0.5531, + "step": 8191 + }, + { + "epoch": 0.6980826587132509, + "grad_norm": 13.9375, + "learning_rate": 1.8305491328439646e-06, + "loss": 0.6348, + "step": 8192 + }, + { + "epoch": 0.6981678738815509, + "grad_norm": 19.25, + "learning_rate": 1.8304003144831772e-06, + "loss": 0.8518, + "step": 8193 + }, + { + "epoch": 0.6982530890498508, + "grad_norm": 16.375, + "learning_rate": 1.8302514856339726e-06, + "loss": 0.7462, + "step": 8194 + }, + { + "epoch": 0.6983383042181508, + "grad_norm": 17.875, + "learning_rate": 1.83010264629904e-06, + "loss": 1.0599, + "step": 8195 + }, + { + "epoch": 0.6984235193864508, + "grad_norm": 16.625, + "learning_rate": 1.8299537964810693e-06, + "loss": 0.9519, + "step": 8196 + }, + { + "epoch": 0.6985087345547507, + "grad_norm": 13.625, + "learning_rate": 1.8298049361827503e-06, + "loss": 0.7515, + "step": 8197 + }, + { + "epoch": 0.6985939497230507, + "grad_norm": 16.125, + "learning_rate": 1.8296560654067732e-06, + "loss": 0.8538, + "step": 8198 + }, + { + "epoch": 0.6986791648913506, + "grad_norm": 15.625, + "learning_rate": 1.8295071841558278e-06, + "loss": 0.8301, + "step": 8199 + }, + { + "epoch": 0.6987643800596506, + "grad_norm": 14.6875, + "learning_rate": 1.8293582924326053e-06, + "loss": 0.8123, + "step": 8200 + }, + { + "epoch": 0.6988495952279505, + "grad_norm": 15.25, + "learning_rate": 1.8292093902397954e-06, + "loss": 0.6887, + "step": 8201 + }, + { + "epoch": 0.6989348103962505, + "grad_norm": 12.8125, + "learning_rate": 1.82906047758009e-06, + "loss": 0.6365, + "step": 8202 + }, + { + "epoch": 0.6990200255645505, + "grad_norm": 13.4375, + "learning_rate": 1.8289115544561787e-06, + "loss": 0.7295, + "step": 8203 + }, + { + "epoch": 0.6991052407328504, + "grad_norm": 14.0, + "learning_rate": 1.828762620870754e-06, + "loss": 0.7369, + "step": 8204 + }, + { + "epoch": 0.6991904559011504, + "grad_norm": 14.875, + "learning_rate": 1.8286136768265064e-06, + "loss": 0.8952, + "step": 8205 + }, + { + "epoch": 0.6992756710694503, + "grad_norm": 13.625, + "learning_rate": 1.828464722326128e-06, + "loss": 0.7115, + "step": 8206 + }, + { + "epoch": 0.6993608862377503, + "grad_norm": 19.375, + "learning_rate": 1.8283157573723106e-06, + "loss": 0.9904, + "step": 8207 + }, + { + "epoch": 0.6994461014060502, + "grad_norm": 16.875, + "learning_rate": 1.8281667819677457e-06, + "loss": 1.1654, + "step": 8208 + }, + { + "epoch": 0.6995313165743502, + "grad_norm": 13.0625, + "learning_rate": 1.8280177961151258e-06, + "loss": 0.6762, + "step": 8209 + }, + { + "epoch": 0.6996165317426501, + "grad_norm": 10.4375, + "learning_rate": 1.827868799817143e-06, + "loss": 0.3427, + "step": 8210 + }, + { + "epoch": 0.6997017469109501, + "grad_norm": 12.1875, + "learning_rate": 1.8277197930764898e-06, + "loss": 0.599, + "step": 8211 + }, + { + "epoch": 0.6997869620792501, + "grad_norm": 11.0, + "learning_rate": 1.8275707758958592e-06, + "loss": 0.3797, + "step": 8212 + }, + { + "epoch": 0.69987217724755, + "grad_norm": 13.8125, + "learning_rate": 1.8274217482779435e-06, + "loss": 0.6451, + "step": 8213 + }, + { + "epoch": 0.69995739241585, + "grad_norm": 12.4375, + "learning_rate": 1.8272727102254365e-06, + "loss": 0.7015, + "step": 8214 + }, + { + "epoch": 0.7000426075841499, + "grad_norm": 12.0625, + "learning_rate": 1.827123661741031e-06, + "loss": 0.3925, + "step": 8215 + }, + { + "epoch": 0.7001278227524499, + "grad_norm": 12.4375, + "learning_rate": 1.8269746028274203e-06, + "loss": 0.5469, + "step": 8216 + }, + { + "epoch": 0.7002130379207498, + "grad_norm": 12.625, + "learning_rate": 1.826825533487299e-06, + "loss": 0.6324, + "step": 8217 + }, + { + "epoch": 0.7002982530890498, + "grad_norm": 14.8125, + "learning_rate": 1.82667645372336e-06, + "loss": 0.4119, + "step": 8218 + }, + { + "epoch": 0.7003834682573498, + "grad_norm": 15.8125, + "learning_rate": 1.8265273635382973e-06, + "loss": 1.025, + "step": 8219 + }, + { + "epoch": 0.7004686834256497, + "grad_norm": 14.8125, + "learning_rate": 1.8263782629348056e-06, + "loss": 0.8116, + "step": 8220 + }, + { + "epoch": 0.7005538985939497, + "grad_norm": 15.375, + "learning_rate": 1.8262291519155793e-06, + "loss": 0.7729, + "step": 8221 + }, + { + "epoch": 0.7006391137622496, + "grad_norm": 14.875, + "learning_rate": 1.8260800304833123e-06, + "loss": 0.7839, + "step": 8222 + }, + { + "epoch": 0.7007243289305496, + "grad_norm": 17.0, + "learning_rate": 1.8259308986407006e-06, + "loss": 0.4612, + "step": 8223 + }, + { + "epoch": 0.7008095440988495, + "grad_norm": 18.0, + "learning_rate": 1.8257817563904378e-06, + "loss": 0.6319, + "step": 8224 + }, + { + "epoch": 0.7008947592671495, + "grad_norm": 15.6875, + "learning_rate": 1.8256326037352199e-06, + "loss": 0.9335, + "step": 8225 + }, + { + "epoch": 0.7009799744354495, + "grad_norm": 14.9375, + "learning_rate": 1.8254834406777416e-06, + "loss": 1.0858, + "step": 8226 + }, + { + "epoch": 0.7010651896037494, + "grad_norm": 12.375, + "learning_rate": 1.8253342672206997e-06, + "loss": 0.5482, + "step": 8227 + }, + { + "epoch": 0.7011504047720494, + "grad_norm": 16.875, + "learning_rate": 1.8251850833667882e-06, + "loss": 0.8393, + "step": 8228 + }, + { + "epoch": 0.7012356199403493, + "grad_norm": 14.125, + "learning_rate": 1.8250358891187042e-06, + "loss": 0.5882, + "step": 8229 + }, + { + "epoch": 0.7013208351086493, + "grad_norm": 13.875, + "learning_rate": 1.8248866844791435e-06, + "loss": 0.7213, + "step": 8230 + }, + { + "epoch": 0.7014060502769492, + "grad_norm": 12.625, + "learning_rate": 1.824737469450802e-06, + "loss": 0.5338, + "step": 8231 + }, + { + "epoch": 0.7014912654452492, + "grad_norm": 13.0, + "learning_rate": 1.824588244036377e-06, + "loss": 0.6128, + "step": 8232 + }, + { + "epoch": 0.7015764806135492, + "grad_norm": 16.125, + "learning_rate": 1.8244390082385644e-06, + "loss": 0.6144, + "step": 8233 + }, + { + "epoch": 0.7016616957818491, + "grad_norm": 10.875, + "learning_rate": 1.8242897620600612e-06, + "loss": 0.4473, + "step": 8234 + }, + { + "epoch": 0.7017469109501491, + "grad_norm": 15.625, + "learning_rate": 1.8241405055035644e-06, + "loss": 1.1708, + "step": 8235 + }, + { + "epoch": 0.701832126118449, + "grad_norm": 10.6875, + "learning_rate": 1.8239912385717718e-06, + "loss": 0.3221, + "step": 8236 + }, + { + "epoch": 0.701917341286749, + "grad_norm": 12.25, + "learning_rate": 1.8238419612673796e-06, + "loss": 0.516, + "step": 8237 + }, + { + "epoch": 0.702002556455049, + "grad_norm": 15.75, + "learning_rate": 1.8236926735930865e-06, + "loss": 0.5556, + "step": 8238 + }, + { + "epoch": 0.702087771623349, + "grad_norm": 18.25, + "learning_rate": 1.8235433755515903e-06, + "loss": 0.8166, + "step": 8239 + }, + { + "epoch": 0.702172986791649, + "grad_norm": 10.9375, + "learning_rate": 1.823394067145588e-06, + "loss": 0.453, + "step": 8240 + }, + { + "epoch": 0.7022582019599489, + "grad_norm": 17.25, + "learning_rate": 1.8232447483777783e-06, + "loss": 1.1542, + "step": 8241 + }, + { + "epoch": 0.7023434171282489, + "grad_norm": 13.3125, + "learning_rate": 1.8230954192508598e-06, + "loss": 0.5877, + "step": 8242 + }, + { + "epoch": 0.7024286322965488, + "grad_norm": 16.875, + "learning_rate": 1.8229460797675308e-06, + "loss": 0.8381, + "step": 8243 + }, + { + "epoch": 0.7025138474648488, + "grad_norm": 15.3125, + "learning_rate": 1.8227967299304899e-06, + "loss": 1.1072, + "step": 8244 + }, + { + "epoch": 0.7025990626331488, + "grad_norm": 15.4375, + "learning_rate": 1.822647369742436e-06, + "loss": 1.1704, + "step": 8245 + }, + { + "epoch": 0.7026842778014487, + "grad_norm": 12.375, + "learning_rate": 1.8224979992060685e-06, + "loss": 0.469, + "step": 8246 + }, + { + "epoch": 0.7027694929697487, + "grad_norm": 29.5, + "learning_rate": 1.8223486183240865e-06, + "loss": 1.0302, + "step": 8247 + }, + { + "epoch": 0.7028547081380486, + "grad_norm": 20.125, + "learning_rate": 1.8221992270991894e-06, + "loss": 1.0763, + "step": 8248 + }, + { + "epoch": 0.7029399233063486, + "grad_norm": 14.5, + "learning_rate": 1.8220498255340766e-06, + "loss": 0.8155, + "step": 8249 + }, + { + "epoch": 0.7030251384746485, + "grad_norm": 15.0, + "learning_rate": 1.8219004136314482e-06, + "loss": 0.5337, + "step": 8250 + }, + { + "epoch": 0.7031103536429485, + "grad_norm": 15.5, + "learning_rate": 1.8217509913940046e-06, + "loss": 0.909, + "step": 8251 + }, + { + "epoch": 0.7031955688112485, + "grad_norm": 12.6875, + "learning_rate": 1.8216015588244453e-06, + "loss": 0.6935, + "step": 8252 + }, + { + "epoch": 0.7032807839795484, + "grad_norm": 24.875, + "learning_rate": 1.8214521159254715e-06, + "loss": 1.37, + "step": 8253 + }, + { + "epoch": 0.7033659991478484, + "grad_norm": 14.0625, + "learning_rate": 1.8213026626997826e-06, + "loss": 0.8778, + "step": 8254 + }, + { + "epoch": 0.7034512143161483, + "grad_norm": 16.25, + "learning_rate": 1.8211531991500809e-06, + "loss": 0.467, + "step": 8255 + }, + { + "epoch": 0.7035364294844483, + "grad_norm": 14.25, + "learning_rate": 1.821003725279066e-06, + "loss": 0.8079, + "step": 8256 + }, + { + "epoch": 0.7036216446527482, + "grad_norm": 23.0, + "learning_rate": 1.8208542410894398e-06, + "loss": 0.7194, + "step": 8257 + }, + { + "epoch": 0.7037068598210482, + "grad_norm": 19.875, + "learning_rate": 1.8207047465839035e-06, + "loss": 0.723, + "step": 8258 + }, + { + "epoch": 0.7037920749893481, + "grad_norm": 14.0625, + "learning_rate": 1.8205552417651585e-06, + "loss": 0.6498, + "step": 8259 + }, + { + "epoch": 0.7038772901576481, + "grad_norm": 15.5, + "learning_rate": 1.8204057266359062e-06, + "loss": 0.7109, + "step": 8260 + }, + { + "epoch": 0.7039625053259481, + "grad_norm": 16.875, + "learning_rate": 1.8202562011988492e-06, + "loss": 0.6741, + "step": 8261 + }, + { + "epoch": 0.704047720494248, + "grad_norm": 19.0, + "learning_rate": 1.820106665456689e-06, + "loss": 0.6911, + "step": 8262 + }, + { + "epoch": 0.704132935662548, + "grad_norm": 10.9375, + "learning_rate": 1.8199571194121285e-06, + "loss": 0.3597, + "step": 8263 + }, + { + "epoch": 0.7042181508308479, + "grad_norm": 11.8125, + "learning_rate": 1.819807563067869e-06, + "loss": 0.483, + "step": 8264 + }, + { + "epoch": 0.7043033659991479, + "grad_norm": 24.125, + "learning_rate": 1.8196579964266145e-06, + "loss": 1.0289, + "step": 8265 + }, + { + "epoch": 0.7043885811674478, + "grad_norm": 19.875, + "learning_rate": 1.8195084194910668e-06, + "loss": 0.8748, + "step": 8266 + }, + { + "epoch": 0.7044737963357478, + "grad_norm": 22.625, + "learning_rate": 1.8193588322639294e-06, + "loss": 1.3506, + "step": 8267 + }, + { + "epoch": 0.7045590115040478, + "grad_norm": 17.375, + "learning_rate": 1.8192092347479057e-06, + "loss": 0.693, + "step": 8268 + }, + { + "epoch": 0.7046442266723477, + "grad_norm": 15.8125, + "learning_rate": 1.819059626945698e-06, + "loss": 0.977, + "step": 8269 + }, + { + "epoch": 0.7047294418406477, + "grad_norm": 15.875, + "learning_rate": 1.8189100088600111e-06, + "loss": 0.8157, + "step": 8270 + }, + { + "epoch": 0.7048146570089476, + "grad_norm": 12.6875, + "learning_rate": 1.8187603804935482e-06, + "loss": 0.6772, + "step": 8271 + }, + { + "epoch": 0.7048998721772476, + "grad_norm": 21.25, + "learning_rate": 1.8186107418490135e-06, + "loss": 0.9156, + "step": 8272 + }, + { + "epoch": 0.7049850873455475, + "grad_norm": 26.625, + "learning_rate": 1.8184610929291106e-06, + "loss": 1.1045, + "step": 8273 + }, + { + "epoch": 0.7050703025138475, + "grad_norm": 15.125, + "learning_rate": 1.818311433736544e-06, + "loss": 0.7433, + "step": 8274 + }, + { + "epoch": 0.7051555176821475, + "grad_norm": 14.3125, + "learning_rate": 1.8181617642740186e-06, + "loss": 0.8027, + "step": 8275 + }, + { + "epoch": 0.7052407328504474, + "grad_norm": 17.125, + "learning_rate": 1.818012084544239e-06, + "loss": 0.5, + "step": 8276 + }, + { + "epoch": 0.7053259480187474, + "grad_norm": 16.875, + "learning_rate": 1.8178623945499092e-06, + "loss": 0.8944, + "step": 8277 + }, + { + "epoch": 0.7054111631870473, + "grad_norm": 16.625, + "learning_rate": 1.8177126942937354e-06, + "loss": 0.704, + "step": 8278 + }, + { + "epoch": 0.7054963783553473, + "grad_norm": 12.5625, + "learning_rate": 1.817562983778422e-06, + "loss": 0.4785, + "step": 8279 + }, + { + "epoch": 0.7055815935236472, + "grad_norm": 14.6875, + "learning_rate": 1.8174132630066751e-06, + "loss": 0.8266, + "step": 8280 + }, + { + "epoch": 0.7056668086919472, + "grad_norm": 14.4375, + "learning_rate": 1.8172635319812e-06, + "loss": 0.5889, + "step": 8281 + }, + { + "epoch": 0.7057520238602472, + "grad_norm": 15.625, + "learning_rate": 1.817113790704702e-06, + "loss": 0.7331, + "step": 8282 + }, + { + "epoch": 0.7058372390285471, + "grad_norm": 20.375, + "learning_rate": 1.8169640391798876e-06, + "loss": 0.5778, + "step": 8283 + }, + { + "epoch": 0.7059224541968471, + "grad_norm": 17.375, + "learning_rate": 1.8168142774094633e-06, + "loss": 0.906, + "step": 8284 + }, + { + "epoch": 0.706007669365147, + "grad_norm": 14.9375, + "learning_rate": 1.8166645053961347e-06, + "loss": 0.6478, + "step": 8285 + }, + { + "epoch": 0.706092884533447, + "grad_norm": 12.25, + "learning_rate": 1.8165147231426088e-06, + "loss": 0.5336, + "step": 8286 + }, + { + "epoch": 0.7061780997017469, + "grad_norm": 9.1875, + "learning_rate": 1.8163649306515919e-06, + "loss": 0.3632, + "step": 8287 + }, + { + "epoch": 0.7062633148700469, + "grad_norm": 15.3125, + "learning_rate": 1.8162151279257918e-06, + "loss": 0.9746, + "step": 8288 + }, + { + "epoch": 0.7063485300383469, + "grad_norm": 19.625, + "learning_rate": 1.8160653149679145e-06, + "loss": 1.0229, + "step": 8289 + }, + { + "epoch": 0.7064337452066468, + "grad_norm": 25.0, + "learning_rate": 1.8159154917806675e-06, + "loss": 0.9772, + "step": 8290 + }, + { + "epoch": 0.7065189603749468, + "grad_norm": 14.25, + "learning_rate": 1.8157656583667587e-06, + "loss": 0.7552, + "step": 8291 + }, + { + "epoch": 0.7066041755432467, + "grad_norm": 14.875, + "learning_rate": 1.815615814728896e-06, + "loss": 0.8258, + "step": 8292 + }, + { + "epoch": 0.7066893907115467, + "grad_norm": 23.375, + "learning_rate": 1.8154659608697864e-06, + "loss": 0.8631, + "step": 8293 + }, + { + "epoch": 0.7067746058798466, + "grad_norm": 13.25, + "learning_rate": 1.815316096792138e-06, + "loss": 0.7681, + "step": 8294 + }, + { + "epoch": 0.7068598210481466, + "grad_norm": 14.3125, + "learning_rate": 1.8151662224986596e-06, + "loss": 0.9717, + "step": 8295 + }, + { + "epoch": 0.7069450362164466, + "grad_norm": 15.6875, + "learning_rate": 1.8150163379920592e-06, + "loss": 0.8529, + "step": 8296 + }, + { + "epoch": 0.7070302513847465, + "grad_norm": 13.75, + "learning_rate": 1.8148664432750455e-06, + "loss": 0.393, + "step": 8297 + }, + { + "epoch": 0.7071154665530465, + "grad_norm": 13.5, + "learning_rate": 1.8147165383503267e-06, + "loss": 0.6728, + "step": 8298 + }, + { + "epoch": 0.7072006817213464, + "grad_norm": 10.8125, + "learning_rate": 1.8145666232206128e-06, + "loss": 0.7041, + "step": 8299 + }, + { + "epoch": 0.7072858968896464, + "grad_norm": 13.25, + "learning_rate": 1.814416697888612e-06, + "loss": 0.5002, + "step": 8300 + }, + { + "epoch": 0.7073711120579463, + "grad_norm": 12.6875, + "learning_rate": 1.814266762357034e-06, + "loss": 0.4458, + "step": 8301 + }, + { + "epoch": 0.7074563272262463, + "grad_norm": 20.375, + "learning_rate": 1.8141168166285877e-06, + "loss": 0.8011, + "step": 8302 + }, + { + "epoch": 0.7075415423945463, + "grad_norm": 11.5625, + "learning_rate": 1.8139668607059841e-06, + "loss": 0.3866, + "step": 8303 + }, + { + "epoch": 0.7076267575628462, + "grad_norm": 18.75, + "learning_rate": 1.8138168945919316e-06, + "loss": 0.8114, + "step": 8304 + }, + { + "epoch": 0.7077119727311462, + "grad_norm": 22.0, + "learning_rate": 1.8136669182891408e-06, + "loss": 1.0237, + "step": 8305 + }, + { + "epoch": 0.7077971878994461, + "grad_norm": 13.5625, + "learning_rate": 1.8135169318003222e-06, + "loss": 0.4133, + "step": 8306 + }, + { + "epoch": 0.7078824030677461, + "grad_norm": 13.0, + "learning_rate": 1.813366935128186e-06, + "loss": 0.642, + "step": 8307 + }, + { + "epoch": 0.707967618236046, + "grad_norm": 11.25, + "learning_rate": 1.813216928275443e-06, + "loss": 0.2526, + "step": 8308 + }, + { + "epoch": 0.708052833404346, + "grad_norm": 14.25, + "learning_rate": 1.8130669112448035e-06, + "loss": 0.8389, + "step": 8309 + }, + { + "epoch": 0.708138048572646, + "grad_norm": 14.25, + "learning_rate": 1.8129168840389789e-06, + "loss": 0.9668, + "step": 8310 + }, + { + "epoch": 0.7082232637409459, + "grad_norm": 16.375, + "learning_rate": 1.8127668466606796e-06, + "loss": 0.8989, + "step": 8311 + }, + { + "epoch": 0.7083084789092459, + "grad_norm": 12.875, + "learning_rate": 1.8126167991126182e-06, + "loss": 0.7233, + "step": 8312 + }, + { + "epoch": 0.7083936940775458, + "grad_norm": 10.1875, + "learning_rate": 1.8124667413975047e-06, + "loss": 0.5178, + "step": 8313 + }, + { + "epoch": 0.7084789092458458, + "grad_norm": 15.5625, + "learning_rate": 1.8123166735180521e-06, + "loss": 0.7463, + "step": 8314 + }, + { + "epoch": 0.7085641244141457, + "grad_norm": 15.0, + "learning_rate": 1.8121665954769715e-06, + "loss": 0.8885, + "step": 8315 + }, + { + "epoch": 0.7086493395824457, + "grad_norm": 22.125, + "learning_rate": 1.8120165072769756e-06, + "loss": 1.2882, + "step": 8316 + }, + { + "epoch": 0.7087345547507456, + "grad_norm": 10.25, + "learning_rate": 1.811866408920776e-06, + "loss": 0.6725, + "step": 8317 + }, + { + "epoch": 0.7088197699190456, + "grad_norm": 16.875, + "learning_rate": 1.8117163004110854e-06, + "loss": 1.0656, + "step": 8318 + }, + { + "epoch": 0.7089049850873456, + "grad_norm": 20.0, + "learning_rate": 1.8115661817506164e-06, + "loss": 1.1641, + "step": 8319 + }, + { + "epoch": 0.7089902002556455, + "grad_norm": 24.375, + "learning_rate": 1.8114160529420818e-06, + "loss": 0.7925, + "step": 8320 + }, + { + "epoch": 0.7090754154239455, + "grad_norm": 18.25, + "learning_rate": 1.8112659139881946e-06, + "loss": 1.0362, + "step": 8321 + }, + { + "epoch": 0.7091606305922454, + "grad_norm": 15.75, + "learning_rate": 1.8111157648916677e-06, + "loss": 0.7768, + "step": 8322 + }, + { + "epoch": 0.7092458457605454, + "grad_norm": 17.75, + "learning_rate": 1.8109656056552153e-06, + "loss": 1.0236, + "step": 8323 + }, + { + "epoch": 0.7093310609288453, + "grad_norm": 10.625, + "learning_rate": 1.81081543628155e-06, + "loss": 0.5315, + "step": 8324 + }, + { + "epoch": 0.7094162760971453, + "grad_norm": 17.0, + "learning_rate": 1.8106652567733856e-06, + "loss": 0.7208, + "step": 8325 + }, + { + "epoch": 0.7095014912654453, + "grad_norm": 11.6875, + "learning_rate": 1.8105150671334363e-06, + "loss": 0.4416, + "step": 8326 + }, + { + "epoch": 0.7095867064337452, + "grad_norm": 13.5625, + "learning_rate": 1.8103648673644164e-06, + "loss": 0.5388, + "step": 8327 + }, + { + "epoch": 0.7096719216020452, + "grad_norm": 24.0, + "learning_rate": 1.8102146574690396e-06, + "loss": 1.357, + "step": 8328 + }, + { + "epoch": 0.7097571367703451, + "grad_norm": 13.4375, + "learning_rate": 1.8100644374500207e-06, + "loss": 0.8613, + "step": 8329 + }, + { + "epoch": 0.7098423519386451, + "grad_norm": 17.875, + "learning_rate": 1.809914207310074e-06, + "loss": 0.7143, + "step": 8330 + }, + { + "epoch": 0.709927567106945, + "grad_norm": 12.6875, + "learning_rate": 1.8097639670519145e-06, + "loss": 0.7625, + "step": 8331 + }, + { + "epoch": 0.710012782275245, + "grad_norm": 15.0, + "learning_rate": 1.8096137166782576e-06, + "loss": 0.7302, + "step": 8332 + }, + { + "epoch": 0.710097997443545, + "grad_norm": 14.9375, + "learning_rate": 1.8094634561918182e-06, + "loss": 0.981, + "step": 8333 + }, + { + "epoch": 0.7101832126118449, + "grad_norm": 9.5625, + "learning_rate": 1.8093131855953111e-06, + "loss": 0.3318, + "step": 8334 + }, + { + "epoch": 0.7102684277801449, + "grad_norm": 19.5, + "learning_rate": 1.8091629048914525e-06, + "loss": 0.983, + "step": 8335 + }, + { + "epoch": 0.7103536429484448, + "grad_norm": 17.0, + "learning_rate": 1.8090126140829577e-06, + "loss": 0.9012, + "step": 8336 + }, + { + "epoch": 0.7104388581167448, + "grad_norm": 15.0625, + "learning_rate": 1.808862313172543e-06, + "loss": 0.6816, + "step": 8337 + }, + { + "epoch": 0.7105240732850447, + "grad_norm": 18.875, + "learning_rate": 1.8087120021629245e-06, + "loss": 0.9793, + "step": 8338 + }, + { + "epoch": 0.7106092884533447, + "grad_norm": 15.3125, + "learning_rate": 1.8085616810568179e-06, + "loss": 0.8137, + "step": 8339 + }, + { + "epoch": 0.7106945036216447, + "grad_norm": 13.125, + "learning_rate": 1.8084113498569406e-06, + "loss": 0.4541, + "step": 8340 + }, + { + "epoch": 0.7107797187899446, + "grad_norm": 17.25, + "learning_rate": 1.8082610085660085e-06, + "loss": 0.8404, + "step": 8341 + }, + { + "epoch": 0.7108649339582446, + "grad_norm": 13.25, + "learning_rate": 1.8081106571867385e-06, + "loss": 0.8269, + "step": 8342 + }, + { + "epoch": 0.7109501491265445, + "grad_norm": 12.8125, + "learning_rate": 1.8079602957218474e-06, + "loss": 0.729, + "step": 8343 + }, + { + "epoch": 0.7110353642948445, + "grad_norm": 15.4375, + "learning_rate": 1.8078099241740533e-06, + "loss": 0.7775, + "step": 8344 + }, + { + "epoch": 0.7111205794631444, + "grad_norm": 13.5625, + "learning_rate": 1.8076595425460725e-06, + "loss": 0.7163, + "step": 8345 + }, + { + "epoch": 0.7112057946314444, + "grad_norm": 14.5, + "learning_rate": 1.807509150840623e-06, + "loss": 0.7853, + "step": 8346 + }, + { + "epoch": 0.7112910097997444, + "grad_norm": 14.6875, + "learning_rate": 1.8073587490604227e-06, + "loss": 0.5224, + "step": 8347 + }, + { + "epoch": 0.7113762249680443, + "grad_norm": 14.25, + "learning_rate": 1.8072083372081892e-06, + "loss": 0.9063, + "step": 8348 + }, + { + "epoch": 0.7114614401363443, + "grad_norm": 12.4375, + "learning_rate": 1.807057915286641e-06, + "loss": 0.581, + "step": 8349 + }, + { + "epoch": 0.7115466553046442, + "grad_norm": 12.75, + "learning_rate": 1.8069074832984958e-06, + "loss": 0.6699, + "step": 8350 + }, + { + "epoch": 0.7116318704729442, + "grad_norm": 14.0625, + "learning_rate": 1.8067570412464725e-06, + "loss": 0.5923, + "step": 8351 + }, + { + "epoch": 0.7117170856412441, + "grad_norm": 11.0625, + "learning_rate": 1.8066065891332896e-06, + "loss": 0.3533, + "step": 8352 + }, + { + "epoch": 0.7118023008095441, + "grad_norm": 9.8125, + "learning_rate": 1.806456126961666e-06, + "loss": 0.2979, + "step": 8353 + }, + { + "epoch": 0.711887515977844, + "grad_norm": 12.9375, + "learning_rate": 1.8063056547343203e-06, + "loss": 0.6823, + "step": 8354 + }, + { + "epoch": 0.711972731146144, + "grad_norm": 12.375, + "learning_rate": 1.8061551724539722e-06, + "loss": 0.5469, + "step": 8355 + }, + { + "epoch": 0.712057946314444, + "grad_norm": 10.5, + "learning_rate": 1.8060046801233411e-06, + "loss": 0.4253, + "step": 8356 + }, + { + "epoch": 0.7121431614827439, + "grad_norm": 14.0625, + "learning_rate": 1.8058541777451464e-06, + "loss": 0.659, + "step": 8357 + }, + { + "epoch": 0.7122283766510439, + "grad_norm": 14.375, + "learning_rate": 1.8057036653221071e-06, + "loss": 0.667, + "step": 8358 + }, + { + "epoch": 0.7123135918193438, + "grad_norm": 12.5625, + "learning_rate": 1.8055531428569443e-06, + "loss": 0.4437, + "step": 8359 + }, + { + "epoch": 0.7123988069876438, + "grad_norm": 13.1875, + "learning_rate": 1.8054026103523772e-06, + "loss": 0.3802, + "step": 8360 + }, + { + "epoch": 0.7124840221559438, + "grad_norm": 15.875, + "learning_rate": 1.8052520678111272e-06, + "loss": 0.4102, + "step": 8361 + }, + { + "epoch": 0.7125692373242437, + "grad_norm": 15.5, + "learning_rate": 1.805101515235913e-06, + "loss": 0.8885, + "step": 8362 + }, + { + "epoch": 0.7126544524925437, + "grad_norm": 14.5625, + "learning_rate": 1.8049509526294567e-06, + "loss": 0.8602, + "step": 8363 + }, + { + "epoch": 0.7127396676608436, + "grad_norm": 15.3125, + "learning_rate": 1.8048003799944786e-06, + "loss": 0.5404, + "step": 8364 + }, + { + "epoch": 0.7128248828291436, + "grad_norm": 15.9375, + "learning_rate": 1.8046497973336998e-06, + "loss": 0.834, + "step": 8365 + }, + { + "epoch": 0.7129100979974435, + "grad_norm": 12.375, + "learning_rate": 1.8044992046498413e-06, + "loss": 0.5426, + "step": 8366 + }, + { + "epoch": 0.7129953131657435, + "grad_norm": 15.125, + "learning_rate": 1.8043486019456247e-06, + "loss": 0.6189, + "step": 8367 + }, + { + "epoch": 0.7130805283340435, + "grad_norm": 15.0625, + "learning_rate": 1.8041979892237713e-06, + "loss": 0.6886, + "step": 8368 + }, + { + "epoch": 0.7131657435023434, + "grad_norm": 12.1875, + "learning_rate": 1.804047366487003e-06, + "loss": 0.5865, + "step": 8369 + }, + { + "epoch": 0.7132509586706434, + "grad_norm": 24.25, + "learning_rate": 1.8038967337380417e-06, + "loss": 1.3272, + "step": 8370 + }, + { + "epoch": 0.7133361738389433, + "grad_norm": 15.3125, + "learning_rate": 1.8037460909796096e-06, + "loss": 0.9567, + "step": 8371 + }, + { + "epoch": 0.7134213890072433, + "grad_norm": 17.125, + "learning_rate": 1.8035954382144282e-06, + "loss": 0.8599, + "step": 8372 + }, + { + "epoch": 0.7135066041755432, + "grad_norm": 12.0, + "learning_rate": 1.8034447754452211e-06, + "loss": 0.6604, + "step": 8373 + }, + { + "epoch": 0.7135918193438432, + "grad_norm": 30.375, + "learning_rate": 1.8032941026747102e-06, + "loss": 0.9519, + "step": 8374 + }, + { + "epoch": 0.7136770345121431, + "grad_norm": 17.625, + "learning_rate": 1.8031434199056184e-06, + "loss": 0.9951, + "step": 8375 + }, + { + "epoch": 0.7137622496804431, + "grad_norm": 23.125, + "learning_rate": 1.8029927271406688e-06, + "loss": 0.8564, + "step": 8376 + }, + { + "epoch": 0.7138474648487431, + "grad_norm": 15.5, + "learning_rate": 1.8028420243825848e-06, + "loss": 0.5792, + "step": 8377 + }, + { + "epoch": 0.713932680017043, + "grad_norm": 15.5625, + "learning_rate": 1.8026913116340893e-06, + "loss": 0.8336, + "step": 8378 + }, + { + "epoch": 0.714017895185343, + "grad_norm": 11.6875, + "learning_rate": 1.802540588897906e-06, + "loss": 0.6323, + "step": 8379 + }, + { + "epoch": 0.7141031103536429, + "grad_norm": 13.5625, + "learning_rate": 1.8023898561767586e-06, + "loss": 0.7817, + "step": 8380 + }, + { + "epoch": 0.7141883255219429, + "grad_norm": 20.625, + "learning_rate": 1.8022391134733713e-06, + "loss": 0.9023, + "step": 8381 + }, + { + "epoch": 0.7142735406902428, + "grad_norm": 10.5625, + "learning_rate": 1.8020883607904677e-06, + "loss": 0.3712, + "step": 8382 + }, + { + "epoch": 0.7143587558585428, + "grad_norm": 12.5, + "learning_rate": 1.8019375981307722e-06, + "loss": 0.6155, + "step": 8383 + }, + { + "epoch": 0.7144439710268428, + "grad_norm": 10.125, + "learning_rate": 1.8017868254970095e-06, + "loss": 0.4282, + "step": 8384 + }, + { + "epoch": 0.7145291861951427, + "grad_norm": 14.0, + "learning_rate": 1.8016360428919038e-06, + "loss": 0.5352, + "step": 8385 + }, + { + "epoch": 0.7146144013634427, + "grad_norm": 15.1875, + "learning_rate": 1.8014852503181802e-06, + "loss": 0.9668, + "step": 8386 + }, + { + "epoch": 0.7146996165317426, + "grad_norm": 17.125, + "learning_rate": 1.8013344477785632e-06, + "loss": 0.8365, + "step": 8387 + }, + { + "epoch": 0.7147848317000426, + "grad_norm": 12.8125, + "learning_rate": 1.8011836352757785e-06, + "loss": 0.5402, + "step": 8388 + }, + { + "epoch": 0.7148700468683425, + "grad_norm": 13.6875, + "learning_rate": 1.8010328128125515e-06, + "loss": 0.7868, + "step": 8389 + }, + { + "epoch": 0.7149552620366425, + "grad_norm": 14.5, + "learning_rate": 1.8008819803916073e-06, + "loss": 0.5965, + "step": 8390 + }, + { + "epoch": 0.7150404772049425, + "grad_norm": 14.6875, + "learning_rate": 1.8007311380156716e-06, + "loss": 0.4634, + "step": 8391 + }, + { + "epoch": 0.7151256923732424, + "grad_norm": 17.125, + "learning_rate": 1.8005802856874707e-06, + "loss": 0.8445, + "step": 8392 + }, + { + "epoch": 0.7152109075415424, + "grad_norm": 15.6875, + "learning_rate": 1.8004294234097304e-06, + "loss": 0.8743, + "step": 8393 + }, + { + "epoch": 0.7152961227098423, + "grad_norm": 14.625, + "learning_rate": 1.8002785511851766e-06, + "loss": 0.8426, + "step": 8394 + }, + { + "epoch": 0.7153813378781423, + "grad_norm": 21.625, + "learning_rate": 1.8001276690165364e-06, + "loss": 0.9245, + "step": 8395 + }, + { + "epoch": 0.7154665530464422, + "grad_norm": 15.25, + "learning_rate": 1.7999767769065357e-06, + "loss": 0.384, + "step": 8396 + }, + { + "epoch": 0.7155517682147422, + "grad_norm": 13.5625, + "learning_rate": 1.7998258748579022e-06, + "loss": 0.5564, + "step": 8397 + }, + { + "epoch": 0.7156369833830422, + "grad_norm": 15.875, + "learning_rate": 1.7996749628733618e-06, + "loss": 0.7272, + "step": 8398 + }, + { + "epoch": 0.7157221985513421, + "grad_norm": 17.375, + "learning_rate": 1.7995240409556421e-06, + "loss": 1.0192, + "step": 8399 + }, + { + "epoch": 0.7158074137196421, + "grad_norm": 13.8125, + "learning_rate": 1.7993731091074701e-06, + "loss": 0.9317, + "step": 8400 + }, + { + "epoch": 0.715892628887942, + "grad_norm": 21.5, + "learning_rate": 1.7992221673315743e-06, + "loss": 0.7983, + "step": 8401 + }, + { + "epoch": 0.715977844056242, + "grad_norm": 16.5, + "learning_rate": 1.7990712156306812e-06, + "loss": 1.0339, + "step": 8402 + }, + { + "epoch": 0.7160630592245419, + "grad_norm": 14.75, + "learning_rate": 1.798920254007519e-06, + "loss": 0.8863, + "step": 8403 + }, + { + "epoch": 0.7161482743928419, + "grad_norm": 8.1875, + "learning_rate": 1.7987692824648159e-06, + "loss": 0.4364, + "step": 8404 + }, + { + "epoch": 0.7162334895611419, + "grad_norm": 14.5625, + "learning_rate": 1.7986183010053001e-06, + "loss": 0.6309, + "step": 8405 + }, + { + "epoch": 0.7163187047294418, + "grad_norm": 11.8125, + "learning_rate": 1.7984673096317001e-06, + "loss": 0.4635, + "step": 8406 + }, + { + "epoch": 0.7164039198977418, + "grad_norm": 11.125, + "learning_rate": 1.798316308346744e-06, + "loss": 0.4399, + "step": 8407 + }, + { + "epoch": 0.7164891350660417, + "grad_norm": 16.375, + "learning_rate": 1.798165297153161e-06, + "loss": 0.9663, + "step": 8408 + }, + { + "epoch": 0.7165743502343417, + "grad_norm": 15.25, + "learning_rate": 1.7980142760536795e-06, + "loss": 0.8537, + "step": 8409 + }, + { + "epoch": 0.7166595654026416, + "grad_norm": 18.375, + "learning_rate": 1.7978632450510297e-06, + "loss": 0.8599, + "step": 8410 + }, + { + "epoch": 0.7167447805709416, + "grad_norm": 13.375, + "learning_rate": 1.7977122041479394e-06, + "loss": 0.7619, + "step": 8411 + }, + { + "epoch": 0.7168299957392416, + "grad_norm": 13.625, + "learning_rate": 1.7975611533471393e-06, + "loss": 0.682, + "step": 8412 + }, + { + "epoch": 0.7169152109075415, + "grad_norm": 10.375, + "learning_rate": 1.7974100926513579e-06, + "loss": 0.7359, + "step": 8413 + }, + { + "epoch": 0.7170004260758415, + "grad_norm": 13.5625, + "learning_rate": 1.7972590220633264e-06, + "loss": 0.6809, + "step": 8414 + }, + { + "epoch": 0.7170856412441414, + "grad_norm": 27.625, + "learning_rate": 1.7971079415857737e-06, + "loss": 1.25, + "step": 8415 + }, + { + "epoch": 0.7171708564124414, + "grad_norm": 15.375, + "learning_rate": 1.7969568512214305e-06, + "loss": 0.7435, + "step": 8416 + }, + { + "epoch": 0.7172560715807413, + "grad_norm": 20.875, + "learning_rate": 1.796805750973027e-06, + "loss": 1.1183, + "step": 8417 + }, + { + "epoch": 0.7173412867490413, + "grad_norm": 10.8125, + "learning_rate": 1.7966546408432936e-06, + "loss": 0.4151, + "step": 8418 + }, + { + "epoch": 0.7174265019173413, + "grad_norm": 23.5, + "learning_rate": 1.796503520834961e-06, + "loss": 0.8503, + "step": 8419 + }, + { + "epoch": 0.7175117170856412, + "grad_norm": 17.125, + "learning_rate": 1.7963523909507606e-06, + "loss": 0.9599, + "step": 8420 + }, + { + "epoch": 0.7175969322539412, + "grad_norm": 14.5625, + "learning_rate": 1.7962012511934226e-06, + "loss": 0.6236, + "step": 8421 + }, + { + "epoch": 0.7176821474222411, + "grad_norm": 18.125, + "learning_rate": 1.7960501015656793e-06, + "loss": 0.8873, + "step": 8422 + }, + { + "epoch": 0.7177673625905411, + "grad_norm": 11.875, + "learning_rate": 1.795898942070261e-06, + "loss": 0.3641, + "step": 8423 + }, + { + "epoch": 0.717852577758841, + "grad_norm": 11.6875, + "learning_rate": 1.7957477727099e-06, + "loss": 0.7945, + "step": 8424 + }, + { + "epoch": 0.717937792927141, + "grad_norm": 14.8125, + "learning_rate": 1.7955965934873284e-06, + "loss": 0.6206, + "step": 8425 + }, + { + "epoch": 0.718023008095441, + "grad_norm": 17.625, + "learning_rate": 1.7954454044052775e-06, + "loss": 0.9856, + "step": 8426 + }, + { + "epoch": 0.7181082232637409, + "grad_norm": 15.9375, + "learning_rate": 1.7952942054664795e-06, + "loss": 1.1367, + "step": 8427 + }, + { + "epoch": 0.7181934384320409, + "grad_norm": 13.625, + "learning_rate": 1.795142996673667e-06, + "loss": 0.6583, + "step": 8428 + }, + { + "epoch": 0.7182786536003408, + "grad_norm": 14.4375, + "learning_rate": 1.7949917780295722e-06, + "loss": 0.8533, + "step": 8429 + }, + { + "epoch": 0.7183638687686408, + "grad_norm": 19.5, + "learning_rate": 1.794840549536928e-06, + "loss": 1.093, + "step": 8430 + }, + { + "epoch": 0.7184490839369407, + "grad_norm": 21.625, + "learning_rate": 1.7946893111984675e-06, + "loss": 0.7753, + "step": 8431 + }, + { + "epoch": 0.7185342991052407, + "grad_norm": 19.5, + "learning_rate": 1.7945380630169231e-06, + "loss": 1.0423, + "step": 8432 + }, + { + "epoch": 0.7186195142735406, + "grad_norm": 15.5, + "learning_rate": 1.7943868049950286e-06, + "loss": 0.76, + "step": 8433 + }, + { + "epoch": 0.7187047294418406, + "grad_norm": 21.5, + "learning_rate": 1.794235537135517e-06, + "loss": 0.9214, + "step": 8434 + }, + { + "epoch": 0.7187899446101406, + "grad_norm": 10.375, + "learning_rate": 1.7940842594411217e-06, + "loss": 0.4538, + "step": 8435 + }, + { + "epoch": 0.7188751597784405, + "grad_norm": 14.4375, + "learning_rate": 1.7939329719145768e-06, + "loss": 0.7646, + "step": 8436 + }, + { + "epoch": 0.7189603749467405, + "grad_norm": 18.25, + "learning_rate": 1.7937816745586162e-06, + "loss": 0.8822, + "step": 8437 + }, + { + "epoch": 0.7190455901150404, + "grad_norm": 15.8125, + "learning_rate": 1.7936303673759742e-06, + "loss": 0.8798, + "step": 8438 + }, + { + "epoch": 0.7191308052833404, + "grad_norm": 12.0625, + "learning_rate": 1.7934790503693844e-06, + "loss": 0.4781, + "step": 8439 + }, + { + "epoch": 0.7192160204516403, + "grad_norm": 15.125, + "learning_rate": 1.793327723541582e-06, + "loss": 0.8587, + "step": 8440 + }, + { + "epoch": 0.7193012356199403, + "grad_norm": 13.1875, + "learning_rate": 1.7931763868953008e-06, + "loss": 0.695, + "step": 8441 + }, + { + "epoch": 0.7193864507882403, + "grad_norm": 16.125, + "learning_rate": 1.7930250404332766e-06, + "loss": 0.6579, + "step": 8442 + }, + { + "epoch": 0.7194716659565402, + "grad_norm": 13.0625, + "learning_rate": 1.7928736841582434e-06, + "loss": 0.5921, + "step": 8443 + }, + { + "epoch": 0.7195568811248402, + "grad_norm": 12.125, + "learning_rate": 1.7927223180729372e-06, + "loss": 0.5393, + "step": 8444 + }, + { + "epoch": 0.7196420962931401, + "grad_norm": 15.1875, + "learning_rate": 1.7925709421800928e-06, + "loss": 0.7309, + "step": 8445 + }, + { + "epoch": 0.7197273114614401, + "grad_norm": 13.5, + "learning_rate": 1.792419556482446e-06, + "loss": 0.3641, + "step": 8446 + }, + { + "epoch": 0.71981252662974, + "grad_norm": 14.75, + "learning_rate": 1.7922681609827322e-06, + "loss": 0.671, + "step": 8447 + }, + { + "epoch": 0.71989774179804, + "grad_norm": 15.4375, + "learning_rate": 1.7921167556836874e-06, + "loss": 0.8593, + "step": 8448 + }, + { + "epoch": 0.71998295696634, + "grad_norm": 19.875, + "learning_rate": 1.7919653405880478e-06, + "loss": 0.7353, + "step": 8449 + }, + { + "epoch": 0.7200681721346399, + "grad_norm": 17.0, + "learning_rate": 1.7918139156985499e-06, + "loss": 0.6623, + "step": 8450 + }, + { + "epoch": 0.7201533873029399, + "grad_norm": 16.5, + "learning_rate": 1.791662481017929e-06, + "loss": 1.1, + "step": 8451 + }, + { + "epoch": 0.7202386024712398, + "grad_norm": 15.0625, + "learning_rate": 1.7915110365489228e-06, + "loss": 0.324, + "step": 8452 + }, + { + "epoch": 0.7203238176395398, + "grad_norm": 12.8125, + "learning_rate": 1.7913595822942673e-06, + "loss": 0.5537, + "step": 8453 + }, + { + "epoch": 0.7204090328078397, + "grad_norm": 16.0, + "learning_rate": 1.7912081182567e-06, + "loss": 0.9242, + "step": 8454 + }, + { + "epoch": 0.7204942479761397, + "grad_norm": 13.4375, + "learning_rate": 1.7910566444389578e-06, + "loss": 0.7414, + "step": 8455 + }, + { + "epoch": 0.7205794631444397, + "grad_norm": 22.375, + "learning_rate": 1.790905160843778e-06, + "loss": 1.2834, + "step": 8456 + }, + { + "epoch": 0.7206646783127396, + "grad_norm": 16.375, + "learning_rate": 1.7907536674738978e-06, + "loss": 0.9382, + "step": 8457 + }, + { + "epoch": 0.7207498934810396, + "grad_norm": 23.5, + "learning_rate": 1.7906021643320554e-06, + "loss": 0.6758, + "step": 8458 + }, + { + "epoch": 0.7208351086493395, + "grad_norm": 14.6875, + "learning_rate": 1.790450651420988e-06, + "loss": 0.8636, + "step": 8459 + }, + { + "epoch": 0.7209203238176395, + "grad_norm": 15.125, + "learning_rate": 1.7902991287434337e-06, + "loss": 1.0673, + "step": 8460 + }, + { + "epoch": 0.7210055389859394, + "grad_norm": 15.375, + "learning_rate": 1.7901475963021315e-06, + "loss": 1.0023, + "step": 8461 + }, + { + "epoch": 0.7210907541542394, + "grad_norm": 16.25, + "learning_rate": 1.7899960540998185e-06, + "loss": 0.9036, + "step": 8462 + }, + { + "epoch": 0.7211759693225395, + "grad_norm": 12.875, + "learning_rate": 1.789844502139234e-06, + "loss": 0.8175, + "step": 8463 + }, + { + "epoch": 0.7212611844908394, + "grad_norm": 12.8125, + "learning_rate": 1.789692940423116e-06, + "loss": 0.325, + "step": 8464 + }, + { + "epoch": 0.7213463996591394, + "grad_norm": 20.875, + "learning_rate": 1.7895413689542045e-06, + "loss": 0.7589, + "step": 8465 + }, + { + "epoch": 0.7214316148274393, + "grad_norm": 15.5, + "learning_rate": 1.7893897877352378e-06, + "loss": 0.7837, + "step": 8466 + }, + { + "epoch": 0.7215168299957393, + "grad_norm": 10.5625, + "learning_rate": 1.7892381967689554e-06, + "loss": 0.3613, + "step": 8467 + }, + { + "epoch": 0.7216020451640393, + "grad_norm": 11.625, + "learning_rate": 1.7890865960580961e-06, + "loss": 0.5492, + "step": 8468 + }, + { + "epoch": 0.7216872603323392, + "grad_norm": 18.0, + "learning_rate": 1.7889349856054e-06, + "loss": 0.7181, + "step": 8469 + }, + { + "epoch": 0.7217724755006392, + "grad_norm": 13.5, + "learning_rate": 1.7887833654136073e-06, + "loss": 0.4474, + "step": 8470 + }, + { + "epoch": 0.7218576906689391, + "grad_norm": 14.4375, + "learning_rate": 1.7886317354854568e-06, + "loss": 0.5507, + "step": 8471 + }, + { + "epoch": 0.7219429058372391, + "grad_norm": 14.75, + "learning_rate": 1.7884800958236892e-06, + "loss": 0.9146, + "step": 8472 + }, + { + "epoch": 0.722028121005539, + "grad_norm": 18.0, + "learning_rate": 1.7883284464310453e-06, + "loss": 0.7204, + "step": 8473 + }, + { + "epoch": 0.722113336173839, + "grad_norm": 18.0, + "learning_rate": 1.7881767873102646e-06, + "loss": 0.8883, + "step": 8474 + }, + { + "epoch": 0.722198551342139, + "grad_norm": 22.0, + "learning_rate": 1.7880251184640878e-06, + "loss": 0.965, + "step": 8475 + }, + { + "epoch": 0.7222837665104389, + "grad_norm": 18.25, + "learning_rate": 1.7878734398952568e-06, + "loss": 0.8425, + "step": 8476 + }, + { + "epoch": 0.7223689816787389, + "grad_norm": 16.875, + "learning_rate": 1.7877217516065115e-06, + "loss": 0.7847, + "step": 8477 + }, + { + "epoch": 0.7224541968470388, + "grad_norm": 24.25, + "learning_rate": 1.7875700536005935e-06, + "loss": 1.0565, + "step": 8478 + }, + { + "epoch": 0.7225394120153388, + "grad_norm": 22.25, + "learning_rate": 1.7874183458802445e-06, + "loss": 1.0721, + "step": 8479 + }, + { + "epoch": 0.7226246271836387, + "grad_norm": 12.3125, + "learning_rate": 1.7872666284482052e-06, + "loss": 0.7625, + "step": 8480 + }, + { + "epoch": 0.7227098423519387, + "grad_norm": 9.875, + "learning_rate": 1.7871149013072175e-06, + "loss": 0.2838, + "step": 8481 + }, + { + "epoch": 0.7227950575202386, + "grad_norm": 14.375, + "learning_rate": 1.7869631644600236e-06, + "loss": 0.5975, + "step": 8482 + }, + { + "epoch": 0.7228802726885386, + "grad_norm": 10.5, + "learning_rate": 1.7868114179093655e-06, + "loss": 0.4271, + "step": 8483 + }, + { + "epoch": 0.7229654878568386, + "grad_norm": 17.0, + "learning_rate": 1.786659661657985e-06, + "loss": 0.7355, + "step": 8484 + }, + { + "epoch": 0.7230507030251385, + "grad_norm": 11.5625, + "learning_rate": 1.7865078957086245e-06, + "loss": 0.4958, + "step": 8485 + }, + { + "epoch": 0.7231359181934385, + "grad_norm": 16.875, + "learning_rate": 1.7863561200640275e-06, + "loss": 1.3118, + "step": 8486 + }, + { + "epoch": 0.7232211333617384, + "grad_norm": 12.8125, + "learning_rate": 1.786204334726936e-06, + "loss": 0.6518, + "step": 8487 + }, + { + "epoch": 0.7233063485300384, + "grad_norm": 13.4375, + "learning_rate": 1.7860525397000927e-06, + "loss": 0.7232, + "step": 8488 + }, + { + "epoch": 0.7233915636983383, + "grad_norm": 16.375, + "learning_rate": 1.7859007349862405e-06, + "loss": 0.8012, + "step": 8489 + }, + { + "epoch": 0.7234767788666383, + "grad_norm": 14.0625, + "learning_rate": 1.7857489205881239e-06, + "loss": 0.8907, + "step": 8490 + }, + { + "epoch": 0.7235619940349383, + "grad_norm": 13.8125, + "learning_rate": 1.7855970965084855e-06, + "loss": 0.5191, + "step": 8491 + }, + { + "epoch": 0.7236472092032382, + "grad_norm": 11.875, + "learning_rate": 1.7854452627500684e-06, + "loss": 0.5629, + "step": 8492 + }, + { + "epoch": 0.7237324243715382, + "grad_norm": 16.625, + "learning_rate": 1.7852934193156174e-06, + "loss": 0.7152, + "step": 8493 + }, + { + "epoch": 0.7238176395398381, + "grad_norm": 14.9375, + "learning_rate": 1.7851415662078757e-06, + "loss": 0.6132, + "step": 8494 + }, + { + "epoch": 0.7239028547081381, + "grad_norm": 15.625, + "learning_rate": 1.7849897034295885e-06, + "loss": 0.8485, + "step": 8495 + }, + { + "epoch": 0.723988069876438, + "grad_norm": 11.6875, + "learning_rate": 1.7848378309834987e-06, + "loss": 0.5086, + "step": 8496 + }, + { + "epoch": 0.724073285044738, + "grad_norm": 13.0, + "learning_rate": 1.784685948872352e-06, + "loss": 0.7946, + "step": 8497 + }, + { + "epoch": 0.724158500213038, + "grad_norm": 16.125, + "learning_rate": 1.784534057098892e-06, + "loss": 0.6018, + "step": 8498 + }, + { + "epoch": 0.7242437153813379, + "grad_norm": 10.4375, + "learning_rate": 1.7843821556658646e-06, + "loss": 0.496, + "step": 8499 + }, + { + "epoch": 0.7243289305496379, + "grad_norm": 15.5625, + "learning_rate": 1.784230244576014e-06, + "loss": 0.5848, + "step": 8500 + }, + { + "epoch": 0.7244141457179378, + "grad_norm": 19.625, + "learning_rate": 1.7840783238320856e-06, + "loss": 0.7122, + "step": 8501 + }, + { + "epoch": 0.7244993608862378, + "grad_norm": 17.0, + "learning_rate": 1.7839263934368247e-06, + "loss": 0.7983, + "step": 8502 + }, + { + "epoch": 0.7245845760545377, + "grad_norm": 28.125, + "learning_rate": 1.7837744533929775e-06, + "loss": 0.7489, + "step": 8503 + }, + { + "epoch": 0.7246697912228377, + "grad_norm": 21.625, + "learning_rate": 1.7836225037032887e-06, + "loss": 0.7452, + "step": 8504 + }, + { + "epoch": 0.7247550063911377, + "grad_norm": 13.3125, + "learning_rate": 1.7834705443705047e-06, + "loss": 0.7934, + "step": 8505 + }, + { + "epoch": 0.7248402215594376, + "grad_norm": 14.1875, + "learning_rate": 1.7833185753973716e-06, + "loss": 0.8602, + "step": 8506 + }, + { + "epoch": 0.7249254367277376, + "grad_norm": 19.875, + "learning_rate": 1.7831665967866358e-06, + "loss": 0.9688, + "step": 8507 + }, + { + "epoch": 0.7250106518960375, + "grad_norm": 15.375, + "learning_rate": 1.7830146085410433e-06, + "loss": 0.7071, + "step": 8508 + }, + { + "epoch": 0.7250958670643375, + "grad_norm": 16.375, + "learning_rate": 1.7828626106633403e-06, + "loss": 0.4915, + "step": 8509 + }, + { + "epoch": 0.7251810822326374, + "grad_norm": 11.0625, + "learning_rate": 1.7827106031562747e-06, + "loss": 0.5039, + "step": 8510 + }, + { + "epoch": 0.7252662974009374, + "grad_norm": 12.3125, + "learning_rate": 1.7825585860225927e-06, + "loss": 0.4425, + "step": 8511 + }, + { + "epoch": 0.7253515125692374, + "grad_norm": 21.5, + "learning_rate": 1.7824065592650417e-06, + "loss": 0.7248, + "step": 8512 + }, + { + "epoch": 0.7254367277375373, + "grad_norm": 10.875, + "learning_rate": 1.7822545228863684e-06, + "loss": 0.4146, + "step": 8513 + }, + { + "epoch": 0.7255219429058373, + "grad_norm": 14.6875, + "learning_rate": 1.7821024768893208e-06, + "loss": 0.801, + "step": 8514 + }, + { + "epoch": 0.7256071580741372, + "grad_norm": 18.875, + "learning_rate": 1.7819504212766463e-06, + "loss": 0.8424, + "step": 8515 + }, + { + "epoch": 0.7256923732424372, + "grad_norm": 13.125, + "learning_rate": 1.7817983560510932e-06, + "loss": 0.7092, + "step": 8516 + }, + { + "epoch": 0.7257775884107371, + "grad_norm": 10.875, + "learning_rate": 1.7816462812154086e-06, + "loss": 0.4323, + "step": 8517 + }, + { + "epoch": 0.7258628035790371, + "grad_norm": 19.125, + "learning_rate": 1.7814941967723413e-06, + "loss": 0.8228, + "step": 8518 + }, + { + "epoch": 0.725948018747337, + "grad_norm": 14.625, + "learning_rate": 1.7813421027246395e-06, + "loss": 0.4931, + "step": 8519 + }, + { + "epoch": 0.726033233915637, + "grad_norm": 13.8125, + "learning_rate": 1.7811899990750514e-06, + "loss": 0.6817, + "step": 8520 + }, + { + "epoch": 0.726118449083937, + "grad_norm": 17.5, + "learning_rate": 1.7810378858263256e-06, + "loss": 0.7145, + "step": 8521 + }, + { + "epoch": 0.7262036642522369, + "grad_norm": 9.5, + "learning_rate": 1.7808857629812114e-06, + "loss": 0.283, + "step": 8522 + }, + { + "epoch": 0.7262888794205369, + "grad_norm": 14.5, + "learning_rate": 1.780733630542458e-06, + "loss": 0.6382, + "step": 8523 + }, + { + "epoch": 0.7263740945888368, + "grad_norm": 15.5, + "learning_rate": 1.7805814885128136e-06, + "loss": 0.6895, + "step": 8524 + }, + { + "epoch": 0.7264593097571368, + "grad_norm": 21.5, + "learning_rate": 1.7804293368950286e-06, + "loss": 0.9316, + "step": 8525 + }, + { + "epoch": 0.7265445249254368, + "grad_norm": 14.875, + "learning_rate": 1.7802771756918519e-06, + "loss": 0.6778, + "step": 8526 + }, + { + "epoch": 0.7266297400937367, + "grad_norm": 13.875, + "learning_rate": 1.7801250049060336e-06, + "loss": 0.7596, + "step": 8527 + }, + { + "epoch": 0.7267149552620367, + "grad_norm": 15.8125, + "learning_rate": 1.779972824540323e-06, + "loss": 0.7848, + "step": 8528 + }, + { + "epoch": 0.7268001704303366, + "grad_norm": 26.625, + "learning_rate": 1.7798206345974712e-06, + "loss": 1.4553, + "step": 8529 + }, + { + "epoch": 0.7268853855986366, + "grad_norm": 12.375, + "learning_rate": 1.7796684350802268e-06, + "loss": 0.4869, + "step": 8530 + }, + { + "epoch": 0.7269706007669365, + "grad_norm": 12.875, + "learning_rate": 1.7795162259913421e-06, + "loss": 0.5979, + "step": 8531 + }, + { + "epoch": 0.7270558159352365, + "grad_norm": 19.5, + "learning_rate": 1.779364007333566e-06, + "loss": 0.6755, + "step": 8532 + }, + { + "epoch": 0.7271410311035365, + "grad_norm": 10.25, + "learning_rate": 1.7792117791096506e-06, + "loss": 0.4563, + "step": 8533 + }, + { + "epoch": 0.7272262462718364, + "grad_norm": 13.875, + "learning_rate": 1.7790595413223455e-06, + "loss": 0.5445, + "step": 8534 + }, + { + "epoch": 0.7273114614401364, + "grad_norm": 12.9375, + "learning_rate": 1.778907293974403e-06, + "loss": 0.606, + "step": 8535 + }, + { + "epoch": 0.7273966766084363, + "grad_norm": 17.875, + "learning_rate": 1.7787550370685735e-06, + "loss": 0.9824, + "step": 8536 + }, + { + "epoch": 0.7274818917767363, + "grad_norm": 14.25, + "learning_rate": 1.7786027706076092e-06, + "loss": 0.7652, + "step": 8537 + }, + { + "epoch": 0.7275671069450362, + "grad_norm": 13.625, + "learning_rate": 1.778450494594261e-06, + "loss": 0.5231, + "step": 8538 + }, + { + "epoch": 0.7276523221133362, + "grad_norm": 14.6875, + "learning_rate": 1.778298209031281e-06, + "loss": 0.7698, + "step": 8539 + }, + { + "epoch": 0.7277375372816361, + "grad_norm": 15.8125, + "learning_rate": 1.7781459139214213e-06, + "loss": 1.0011, + "step": 8540 + }, + { + "epoch": 0.7278227524499361, + "grad_norm": 17.25, + "learning_rate": 1.7779936092674339e-06, + "loss": 0.876, + "step": 8541 + }, + { + "epoch": 0.7279079676182361, + "grad_norm": 13.9375, + "learning_rate": 1.7778412950720704e-06, + "loss": 0.54, + "step": 8542 + }, + { + "epoch": 0.727993182786536, + "grad_norm": 15.3125, + "learning_rate": 1.7776889713380844e-06, + "loss": 0.5822, + "step": 8543 + }, + { + "epoch": 0.728078397954836, + "grad_norm": 13.4375, + "learning_rate": 1.777536638068228e-06, + "loss": 0.5462, + "step": 8544 + }, + { + "epoch": 0.7281636131231359, + "grad_norm": 20.5, + "learning_rate": 1.7773842952652539e-06, + "loss": 1.0861, + "step": 8545 + }, + { + "epoch": 0.7282488282914359, + "grad_norm": 19.375, + "learning_rate": 1.7772319429319153e-06, + "loss": 1.0924, + "step": 8546 + }, + { + "epoch": 0.7283340434597358, + "grad_norm": 9.625, + "learning_rate": 1.7770795810709653e-06, + "loss": 0.2995, + "step": 8547 + }, + { + "epoch": 0.7284192586280358, + "grad_norm": 13.5, + "learning_rate": 1.7769272096851578e-06, + "loss": 0.7393, + "step": 8548 + }, + { + "epoch": 0.7285044737963358, + "grad_norm": 10.8125, + "learning_rate": 1.776774828777245e-06, + "loss": 0.453, + "step": 8549 + }, + { + "epoch": 0.7285896889646357, + "grad_norm": 17.625, + "learning_rate": 1.7766224383499819e-06, + "loss": 0.7386, + "step": 8550 + }, + { + "epoch": 0.7286749041329357, + "grad_norm": 19.0, + "learning_rate": 1.7764700384061212e-06, + "loss": 0.9498, + "step": 8551 + }, + { + "epoch": 0.7287601193012356, + "grad_norm": 12.0625, + "learning_rate": 1.7763176289484182e-06, + "loss": 0.5819, + "step": 8552 + }, + { + "epoch": 0.7288453344695356, + "grad_norm": 12.5, + "learning_rate": 1.776165209979626e-06, + "loss": 0.6056, + "step": 8553 + }, + { + "epoch": 0.7289305496378355, + "grad_norm": 9.9375, + "learning_rate": 1.7760127815024994e-06, + "loss": 0.3769, + "step": 8554 + }, + { + "epoch": 0.7290157648061355, + "grad_norm": 13.3125, + "learning_rate": 1.7758603435197926e-06, + "loss": 0.6686, + "step": 8555 + }, + { + "epoch": 0.7291009799744355, + "grad_norm": 17.25, + "learning_rate": 1.7757078960342613e-06, + "loss": 0.8655, + "step": 8556 + }, + { + "epoch": 0.7291861951427354, + "grad_norm": 20.25, + "learning_rate": 1.775555439048659e-06, + "loss": 0.6658, + "step": 8557 + }, + { + "epoch": 0.7292714103110354, + "grad_norm": 16.375, + "learning_rate": 1.7754029725657417e-06, + "loss": 0.8874, + "step": 8558 + }, + { + "epoch": 0.7293566254793353, + "grad_norm": 14.625, + "learning_rate": 1.7752504965882643e-06, + "loss": 0.6466, + "step": 8559 + }, + { + "epoch": 0.7294418406476353, + "grad_norm": 12.0, + "learning_rate": 1.7750980111189826e-06, + "loss": 0.5399, + "step": 8560 + }, + { + "epoch": 0.7295270558159352, + "grad_norm": 10.9375, + "learning_rate": 1.7749455161606513e-06, + "loss": 0.4239, + "step": 8561 + }, + { + "epoch": 0.7296122709842352, + "grad_norm": 11.375, + "learning_rate": 1.7747930117160267e-06, + "loss": 0.2861, + "step": 8562 + }, + { + "epoch": 0.7296974861525352, + "grad_norm": 11.625, + "learning_rate": 1.7746404977878651e-06, + "loss": 0.419, + "step": 8563 + }, + { + "epoch": 0.7297827013208351, + "grad_norm": 14.5, + "learning_rate": 1.7744879743789218e-06, + "loss": 0.7734, + "step": 8564 + }, + { + "epoch": 0.7298679164891351, + "grad_norm": 14.375, + "learning_rate": 1.7743354414919534e-06, + "loss": 0.8239, + "step": 8565 + }, + { + "epoch": 0.729953131657435, + "grad_norm": 18.375, + "learning_rate": 1.7741828991297161e-06, + "loss": 0.8964, + "step": 8566 + }, + { + "epoch": 0.730038346825735, + "grad_norm": 12.75, + "learning_rate": 1.774030347294967e-06, + "loss": 0.5365, + "step": 8567 + }, + { + "epoch": 0.7301235619940349, + "grad_norm": 12.9375, + "learning_rate": 1.7738777859904626e-06, + "loss": 0.5052, + "step": 8568 + }, + { + "epoch": 0.7302087771623349, + "grad_norm": 16.625, + "learning_rate": 1.7737252152189599e-06, + "loss": 0.4295, + "step": 8569 + }, + { + "epoch": 0.7302939923306349, + "grad_norm": 12.8125, + "learning_rate": 1.7735726349832156e-06, + "loss": 0.4678, + "step": 8570 + }, + { + "epoch": 0.7303792074989348, + "grad_norm": 14.6875, + "learning_rate": 1.7734200452859876e-06, + "loss": 0.4154, + "step": 8571 + }, + { + "epoch": 0.7304644226672348, + "grad_norm": 11.375, + "learning_rate": 1.7732674461300328e-06, + "loss": 0.5816, + "step": 8572 + }, + { + "epoch": 0.7305496378355347, + "grad_norm": 15.1875, + "learning_rate": 1.7731148375181092e-06, + "loss": 0.4047, + "step": 8573 + }, + { + "epoch": 0.7306348530038347, + "grad_norm": 18.875, + "learning_rate": 1.7729622194529744e-06, + "loss": 1.0405, + "step": 8574 + }, + { + "epoch": 0.7307200681721346, + "grad_norm": 21.875, + "learning_rate": 1.7728095919373863e-06, + "loss": 0.8328, + "step": 8575 + }, + { + "epoch": 0.7308052833404346, + "grad_norm": 12.75, + "learning_rate": 1.7726569549741035e-06, + "loss": 0.5895, + "step": 8576 + }, + { + "epoch": 0.7308904985087346, + "grad_norm": 24.0, + "learning_rate": 1.7725043085658834e-06, + "loss": 0.8665, + "step": 8577 + }, + { + "epoch": 0.7309757136770345, + "grad_norm": 11.625, + "learning_rate": 1.7723516527154855e-06, + "loss": 0.6448, + "step": 8578 + }, + { + "epoch": 0.7310609288453345, + "grad_norm": 18.5, + "learning_rate": 1.772198987425668e-06, + "loss": 0.6823, + "step": 8579 + }, + { + "epoch": 0.7311461440136344, + "grad_norm": 15.5625, + "learning_rate": 1.7720463126991893e-06, + "loss": 0.8641, + "step": 8580 + }, + { + "epoch": 0.7312313591819344, + "grad_norm": 29.125, + "learning_rate": 1.7718936285388088e-06, + "loss": 0.8062, + "step": 8581 + }, + { + "epoch": 0.7313165743502343, + "grad_norm": 13.4375, + "learning_rate": 1.7717409349472858e-06, + "loss": 0.6037, + "step": 8582 + }, + { + "epoch": 0.7314017895185343, + "grad_norm": 16.75, + "learning_rate": 1.7715882319273794e-06, + "loss": 1.1658, + "step": 8583 + }, + { + "epoch": 0.7314870046868343, + "grad_norm": 20.75, + "learning_rate": 1.7714355194818494e-06, + "loss": 1.082, + "step": 8584 + }, + { + "epoch": 0.7315722198551342, + "grad_norm": 16.0, + "learning_rate": 1.771282797613455e-06, + "loss": 0.9383, + "step": 8585 + }, + { + "epoch": 0.7316574350234342, + "grad_norm": 16.5, + "learning_rate": 1.7711300663249564e-06, + "loss": 0.78, + "step": 8586 + }, + { + "epoch": 0.7317426501917341, + "grad_norm": 14.5, + "learning_rate": 1.7709773256191132e-06, + "loss": 0.757, + "step": 8587 + }, + { + "epoch": 0.7318278653600341, + "grad_norm": 17.75, + "learning_rate": 1.7708245754986861e-06, + "loss": 0.9254, + "step": 8588 + }, + { + "epoch": 0.731913080528334, + "grad_norm": 17.875, + "learning_rate": 1.770671815966435e-06, + "loss": 0.835, + "step": 8589 + }, + { + "epoch": 0.731998295696634, + "grad_norm": 13.1875, + "learning_rate": 1.7705190470251207e-06, + "loss": 0.7231, + "step": 8590 + }, + { + "epoch": 0.732083510864934, + "grad_norm": 19.5, + "learning_rate": 1.7703662686775036e-06, + "loss": 1.0814, + "step": 8591 + }, + { + "epoch": 0.7321687260332339, + "grad_norm": 21.75, + "learning_rate": 1.7702134809263449e-06, + "loss": 0.9058, + "step": 8592 + }, + { + "epoch": 0.7322539412015339, + "grad_norm": 19.875, + "learning_rate": 1.7700606837744056e-06, + "loss": 0.6174, + "step": 8593 + }, + { + "epoch": 0.7323391563698338, + "grad_norm": 14.75, + "learning_rate": 1.7699078772244467e-06, + "loss": 0.9005, + "step": 8594 + }, + { + "epoch": 0.7324243715381338, + "grad_norm": 11.25, + "learning_rate": 1.7697550612792296e-06, + "loss": 0.4177, + "step": 8595 + }, + { + "epoch": 0.7325095867064337, + "grad_norm": 11.125, + "learning_rate": 1.769602235941516e-06, + "loss": 0.4524, + "step": 8596 + }, + { + "epoch": 0.7325948018747337, + "grad_norm": 13.1875, + "learning_rate": 1.7694494012140676e-06, + "loss": 0.7261, + "step": 8597 + }, + { + "epoch": 0.7326800170430336, + "grad_norm": 13.0, + "learning_rate": 1.7692965570996456e-06, + "loss": 0.5186, + "step": 8598 + }, + { + "epoch": 0.7327652322113336, + "grad_norm": 18.625, + "learning_rate": 1.7691437036010135e-06, + "loss": 0.4464, + "step": 8599 + }, + { + "epoch": 0.7328504473796336, + "grad_norm": 22.125, + "learning_rate": 1.7689908407209321e-06, + "loss": 0.9657, + "step": 8600 + }, + { + "epoch": 0.7329356625479335, + "grad_norm": 14.4375, + "learning_rate": 1.7688379684621642e-06, + "loss": 0.6473, + "step": 8601 + }, + { + "epoch": 0.7330208777162335, + "grad_norm": 19.625, + "learning_rate": 1.7686850868274726e-06, + "loss": 0.6331, + "step": 8602 + }, + { + "epoch": 0.7331060928845334, + "grad_norm": 13.9375, + "learning_rate": 1.76853219581962e-06, + "loss": 0.5246, + "step": 8603 + }, + { + "epoch": 0.7331913080528334, + "grad_norm": 14.8125, + "learning_rate": 1.7683792954413694e-06, + "loss": 0.5925, + "step": 8604 + }, + { + "epoch": 0.7332765232211333, + "grad_norm": 15.625, + "learning_rate": 1.7682263856954835e-06, + "loss": 0.9027, + "step": 8605 + }, + { + "epoch": 0.7333617383894333, + "grad_norm": 21.875, + "learning_rate": 1.7680734665847254e-06, + "loss": 0.8009, + "step": 8606 + }, + { + "epoch": 0.7334469535577333, + "grad_norm": 15.75, + "learning_rate": 1.7679205381118592e-06, + "loss": 0.5548, + "step": 8607 + }, + { + "epoch": 0.7335321687260332, + "grad_norm": 17.75, + "learning_rate": 1.767767600279648e-06, + "loss": 1.0133, + "step": 8608 + }, + { + "epoch": 0.7336173838943332, + "grad_norm": 14.5, + "learning_rate": 1.7676146530908552e-06, + "loss": 0.9443, + "step": 8609 + }, + { + "epoch": 0.7337025990626331, + "grad_norm": 15.125, + "learning_rate": 1.7674616965482454e-06, + "loss": 0.7443, + "step": 8610 + }, + { + "epoch": 0.7337878142309331, + "grad_norm": 12.9375, + "learning_rate": 1.7673087306545825e-06, + "loss": 0.5754, + "step": 8611 + }, + { + "epoch": 0.733873029399233, + "grad_norm": 13.0625, + "learning_rate": 1.7671557554126303e-06, + "loss": 0.6166, + "step": 8612 + }, + { + "epoch": 0.733958244567533, + "grad_norm": 14.1875, + "learning_rate": 1.7670027708251542e-06, + "loss": 0.6266, + "step": 8613 + }, + { + "epoch": 0.734043459735833, + "grad_norm": 14.8125, + "learning_rate": 1.7668497768949177e-06, + "loss": 0.7187, + "step": 8614 + }, + { + "epoch": 0.7341286749041329, + "grad_norm": 11.8125, + "learning_rate": 1.766696773624686e-06, + "loss": 0.4585, + "step": 8615 + }, + { + "epoch": 0.7342138900724329, + "grad_norm": 19.375, + "learning_rate": 1.766543761017224e-06, + "loss": 0.441, + "step": 8616 + }, + { + "epoch": 0.7342991052407328, + "grad_norm": 18.375, + "learning_rate": 1.766390739075297e-06, + "loss": 0.8743, + "step": 8617 + }, + { + "epoch": 0.7343843204090328, + "grad_norm": 16.75, + "learning_rate": 1.76623770780167e-06, + "loss": 0.5088, + "step": 8618 + }, + { + "epoch": 0.7344695355773327, + "grad_norm": 23.25, + "learning_rate": 1.7660846671991083e-06, + "loss": 1.2049, + "step": 8619 + }, + { + "epoch": 0.7345547507456327, + "grad_norm": 21.0, + "learning_rate": 1.7659316172703777e-06, + "loss": 0.7994, + "step": 8620 + }, + { + "epoch": 0.7346399659139327, + "grad_norm": 15.5625, + "learning_rate": 1.7657785580182443e-06, + "loss": 0.7801, + "step": 8621 + }, + { + "epoch": 0.7347251810822326, + "grad_norm": 18.5, + "learning_rate": 1.7656254894454736e-06, + "loss": 0.7233, + "step": 8622 + }, + { + "epoch": 0.7348103962505326, + "grad_norm": 11.375, + "learning_rate": 1.7654724115548316e-06, + "loss": 0.3936, + "step": 8623 + }, + { + "epoch": 0.7348956114188325, + "grad_norm": 14.625, + "learning_rate": 1.7653193243490852e-06, + "loss": 1.1214, + "step": 8624 + }, + { + "epoch": 0.7349808265871325, + "grad_norm": 12.3125, + "learning_rate": 1.765166227831e-06, + "loss": 0.5206, + "step": 8625 + }, + { + "epoch": 0.7350660417554324, + "grad_norm": 13.625, + "learning_rate": 1.765013122003343e-06, + "loss": 0.733, + "step": 8626 + }, + { + "epoch": 0.7351512569237324, + "grad_norm": 12.6875, + "learning_rate": 1.7648600068688811e-06, + "loss": 0.7921, + "step": 8627 + }, + { + "epoch": 0.7352364720920324, + "grad_norm": 16.375, + "learning_rate": 1.7647068824303815e-06, + "loss": 0.8531, + "step": 8628 + }, + { + "epoch": 0.7353216872603323, + "grad_norm": 12.8125, + "learning_rate": 1.7645537486906107e-06, + "loss": 0.6251, + "step": 8629 + }, + { + "epoch": 0.7354069024286323, + "grad_norm": 17.375, + "learning_rate": 1.7644006056523358e-06, + "loss": 0.9287, + "step": 8630 + }, + { + "epoch": 0.7354921175969322, + "grad_norm": 12.75, + "learning_rate": 1.7642474533183252e-06, + "loss": 0.6049, + "step": 8631 + }, + { + "epoch": 0.7355773327652322, + "grad_norm": 19.375, + "learning_rate": 1.7640942916913457e-06, + "loss": 0.8854, + "step": 8632 + }, + { + "epoch": 0.7356625479335321, + "grad_norm": 11.0625, + "learning_rate": 1.7639411207741657e-06, + "loss": 0.3173, + "step": 8633 + }, + { + "epoch": 0.7357477631018321, + "grad_norm": 12.5, + "learning_rate": 1.7637879405695526e-06, + "loss": 0.4365, + "step": 8634 + }, + { + "epoch": 0.735832978270132, + "grad_norm": 20.625, + "learning_rate": 1.7636347510802751e-06, + "loss": 0.729, + "step": 8635 + }, + { + "epoch": 0.735918193438432, + "grad_norm": 13.625, + "learning_rate": 1.7634815523091007e-06, + "loss": 0.9007, + "step": 8636 + }, + { + "epoch": 0.736003408606732, + "grad_norm": 21.375, + "learning_rate": 1.7633283442587989e-06, + "loss": 0.8828, + "step": 8637 + }, + { + "epoch": 0.7360886237750319, + "grad_norm": 13.375, + "learning_rate": 1.763175126932137e-06, + "loss": 0.5454, + "step": 8638 + }, + { + "epoch": 0.7361738389433319, + "grad_norm": 16.625, + "learning_rate": 1.763021900331885e-06, + "loss": 0.8719, + "step": 8639 + }, + { + "epoch": 0.7362590541116318, + "grad_norm": 16.875, + "learning_rate": 1.7628686644608113e-06, + "loss": 0.5728, + "step": 8640 + }, + { + "epoch": 0.7363442692799318, + "grad_norm": 17.375, + "learning_rate": 1.762715419321685e-06, + "loss": 0.9567, + "step": 8641 + }, + { + "epoch": 0.7364294844482318, + "grad_norm": 16.875, + "learning_rate": 1.7625621649172755e-06, + "loss": 0.7863, + "step": 8642 + }, + { + "epoch": 0.7365146996165317, + "grad_norm": 13.3125, + "learning_rate": 1.7624089012503523e-06, + "loss": 0.7025, + "step": 8643 + }, + { + "epoch": 0.7365999147848317, + "grad_norm": 14.8125, + "learning_rate": 1.762255628323685e-06, + "loss": 0.7628, + "step": 8644 + }, + { + "epoch": 0.7366851299531316, + "grad_norm": 10.625, + "learning_rate": 1.7621023461400432e-06, + "loss": 0.3108, + "step": 8645 + }, + { + "epoch": 0.7367703451214316, + "grad_norm": 13.75, + "learning_rate": 1.7619490547021976e-06, + "loss": 0.6916, + "step": 8646 + }, + { + "epoch": 0.7368555602897315, + "grad_norm": 10.8125, + "learning_rate": 1.7617957540129171e-06, + "loss": 0.4654, + "step": 8647 + }, + { + "epoch": 0.7369407754580315, + "grad_norm": 13.9375, + "learning_rate": 1.761642444074973e-06, + "loss": 0.7203, + "step": 8648 + }, + { + "epoch": 0.7370259906263315, + "grad_norm": 10.1875, + "learning_rate": 1.7614891248911353e-06, + "loss": 0.4617, + "step": 8649 + }, + { + "epoch": 0.7371112057946314, + "grad_norm": 14.0, + "learning_rate": 1.7613357964641753e-06, + "loss": 0.5827, + "step": 8650 + }, + { + "epoch": 0.7371964209629314, + "grad_norm": 15.8125, + "learning_rate": 1.7611824587968624e-06, + "loss": 0.7434, + "step": 8651 + }, + { + "epoch": 0.7372816361312313, + "grad_norm": 12.6875, + "learning_rate": 1.7610291118919689e-06, + "loss": 0.3914, + "step": 8652 + }, + { + "epoch": 0.7373668512995313, + "grad_norm": 12.375, + "learning_rate": 1.7608757557522656e-06, + "loss": 0.6868, + "step": 8653 + }, + { + "epoch": 0.7374520664678312, + "grad_norm": 19.125, + "learning_rate": 1.7607223903805237e-06, + "loss": 0.6665, + "step": 8654 + }, + { + "epoch": 0.7375372816361312, + "grad_norm": 13.8125, + "learning_rate": 1.7605690157795141e-06, + "loss": 0.657, + "step": 8655 + }, + { + "epoch": 0.7376224968044311, + "grad_norm": 16.125, + "learning_rate": 1.7604156319520097e-06, + "loss": 1.1056, + "step": 8656 + }, + { + "epoch": 0.7377077119727311, + "grad_norm": 19.375, + "learning_rate": 1.7602622389007811e-06, + "loss": 0.8459, + "step": 8657 + }, + { + "epoch": 0.7377929271410311, + "grad_norm": 16.875, + "learning_rate": 1.760108836628601e-06, + "loss": 0.756, + "step": 8658 + }, + { + "epoch": 0.737878142309331, + "grad_norm": 12.125, + "learning_rate": 1.759955425138241e-06, + "loss": 0.5786, + "step": 8659 + }, + { + "epoch": 0.737963357477631, + "grad_norm": 16.375, + "learning_rate": 1.7598020044324738e-06, + "loss": 1.063, + "step": 8660 + }, + { + "epoch": 0.7380485726459309, + "grad_norm": 15.6875, + "learning_rate": 1.7596485745140718e-06, + "loss": 0.6755, + "step": 8661 + }, + { + "epoch": 0.7381337878142309, + "grad_norm": 15.375, + "learning_rate": 1.7594951353858069e-06, + "loss": 0.6532, + "step": 8662 + }, + { + "epoch": 0.7382190029825308, + "grad_norm": 15.25, + "learning_rate": 1.7593416870504535e-06, + "loss": 0.7708, + "step": 8663 + }, + { + "epoch": 0.7383042181508308, + "grad_norm": 12.9375, + "learning_rate": 1.759188229510783e-06, + "loss": 0.7735, + "step": 8664 + }, + { + "epoch": 0.7383894333191308, + "grad_norm": 16.125, + "learning_rate": 1.7590347627695698e-06, + "loss": 0.6911, + "step": 8665 + }, + { + "epoch": 0.7384746484874307, + "grad_norm": 12.875, + "learning_rate": 1.758881286829586e-06, + "loss": 0.6542, + "step": 8666 + }, + { + "epoch": 0.7385598636557307, + "grad_norm": 16.875, + "learning_rate": 1.758727801693606e-06, + "loss": 0.9329, + "step": 8667 + }, + { + "epoch": 0.7386450788240306, + "grad_norm": 17.25, + "learning_rate": 1.7585743073644027e-06, + "loss": 1.1321, + "step": 8668 + }, + { + "epoch": 0.7387302939923306, + "grad_norm": 13.9375, + "learning_rate": 1.7584208038447506e-06, + "loss": 0.5965, + "step": 8669 + }, + { + "epoch": 0.7388155091606305, + "grad_norm": 17.75, + "learning_rate": 1.758267291137423e-06, + "loss": 1.1661, + "step": 8670 + }, + { + "epoch": 0.7389007243289305, + "grad_norm": 12.375, + "learning_rate": 1.7581137692451945e-06, + "loss": 0.5061, + "step": 8671 + }, + { + "epoch": 0.7389859394972305, + "grad_norm": 16.125, + "learning_rate": 1.757960238170839e-06, + "loss": 0.3777, + "step": 8672 + }, + { + "epoch": 0.7390711546655304, + "grad_norm": 16.875, + "learning_rate": 1.7578066979171319e-06, + "loss": 1.1345, + "step": 8673 + }, + { + "epoch": 0.7391563698338304, + "grad_norm": 22.875, + "learning_rate": 1.7576531484868464e-06, + "loss": 0.7862, + "step": 8674 + }, + { + "epoch": 0.7392415850021303, + "grad_norm": 12.75, + "learning_rate": 1.7574995898827585e-06, + "loss": 0.734, + "step": 8675 + }, + { + "epoch": 0.7393268001704303, + "grad_norm": 13.375, + "learning_rate": 1.7573460221076422e-06, + "loss": 0.7169, + "step": 8676 + }, + { + "epoch": 0.7394120153387302, + "grad_norm": 17.75, + "learning_rate": 1.7571924451642735e-06, + "loss": 0.903, + "step": 8677 + }, + { + "epoch": 0.7394972305070302, + "grad_norm": 19.625, + "learning_rate": 1.7570388590554272e-06, + "loss": 1.1479, + "step": 8678 + }, + { + "epoch": 0.7395824456753302, + "grad_norm": 32.5, + "learning_rate": 1.756885263783879e-06, + "loss": 1.3808, + "step": 8679 + }, + { + "epoch": 0.7396676608436301, + "grad_norm": 16.375, + "learning_rate": 1.7567316593524038e-06, + "loss": 0.6421, + "step": 8680 + }, + { + "epoch": 0.7397528760119301, + "grad_norm": 10.375, + "learning_rate": 1.7565780457637784e-06, + "loss": 0.4352, + "step": 8681 + }, + { + "epoch": 0.73983809118023, + "grad_norm": 18.25, + "learning_rate": 1.7564244230207786e-06, + "loss": 0.9484, + "step": 8682 + }, + { + "epoch": 0.73992330634853, + "grad_norm": 18.125, + "learning_rate": 1.7562707911261796e-06, + "loss": 0.742, + "step": 8683 + }, + { + "epoch": 0.7400085215168299, + "grad_norm": 20.25, + "learning_rate": 1.7561171500827587e-06, + "loss": 1.0956, + "step": 8684 + }, + { + "epoch": 0.7400937366851299, + "grad_norm": 14.0, + "learning_rate": 1.7559634998932918e-06, + "loss": 0.6639, + "step": 8685 + }, + { + "epoch": 0.7401789518534299, + "grad_norm": 13.25, + "learning_rate": 1.755809840560556e-06, + "loss": 0.6339, + "step": 8686 + }, + { + "epoch": 0.7402641670217298, + "grad_norm": 11.0, + "learning_rate": 1.7556561720873272e-06, + "loss": 0.3729, + "step": 8687 + }, + { + "epoch": 0.7403493821900298, + "grad_norm": 11.4375, + "learning_rate": 1.7555024944763833e-06, + "loss": 0.6727, + "step": 8688 + }, + { + "epoch": 0.7404345973583298, + "grad_norm": 15.75, + "learning_rate": 1.7553488077305006e-06, + "loss": 0.7158, + "step": 8689 + }, + { + "epoch": 0.7405198125266298, + "grad_norm": 12.0625, + "learning_rate": 1.7551951118524574e-06, + "loss": 0.6316, + "step": 8690 + }, + { + "epoch": 0.7406050276949298, + "grad_norm": 15.875, + "learning_rate": 1.75504140684503e-06, + "loss": 1.0444, + "step": 8691 + }, + { + "epoch": 0.7406902428632297, + "grad_norm": 15.5, + "learning_rate": 1.7548876927109967e-06, + "loss": 0.5795, + "step": 8692 + }, + { + "epoch": 0.7407754580315297, + "grad_norm": 14.0, + "learning_rate": 1.7547339694531349e-06, + "loss": 0.7421, + "step": 8693 + }, + { + "epoch": 0.7408606731998296, + "grad_norm": 13.3125, + "learning_rate": 1.7545802370742233e-06, + "loss": 0.7057, + "step": 8694 + }, + { + "epoch": 0.7409458883681296, + "grad_norm": 12.0625, + "learning_rate": 1.7544264955770388e-06, + "loss": 0.491, + "step": 8695 + }, + { + "epoch": 0.7410311035364295, + "grad_norm": 41.75, + "learning_rate": 1.7542727449643605e-06, + "loss": 1.3983, + "step": 8696 + }, + { + "epoch": 0.7411163187047295, + "grad_norm": 33.5, + "learning_rate": 1.7541189852389664e-06, + "loss": 1.3734, + "step": 8697 + }, + { + "epoch": 0.7412015338730295, + "grad_norm": 17.5, + "learning_rate": 1.7539652164036359e-06, + "loss": 0.895, + "step": 8698 + }, + { + "epoch": 0.7412867490413294, + "grad_norm": 12.0, + "learning_rate": 1.7538114384611466e-06, + "loss": 0.7135, + "step": 8699 + }, + { + "epoch": 0.7413719642096294, + "grad_norm": 10.0, + "learning_rate": 1.7536576514142781e-06, + "loss": 0.6409, + "step": 8700 + }, + { + "epoch": 0.7414571793779293, + "grad_norm": 14.9375, + "learning_rate": 1.7535038552658097e-06, + "loss": 0.7891, + "step": 8701 + }, + { + "epoch": 0.7415423945462293, + "grad_norm": 18.0, + "learning_rate": 1.7533500500185198e-06, + "loss": 0.5417, + "step": 8702 + }, + { + "epoch": 0.7416276097145292, + "grad_norm": 15.5, + "learning_rate": 1.7531962356751886e-06, + "loss": 0.7711, + "step": 8703 + }, + { + "epoch": 0.7417128248828292, + "grad_norm": 18.625, + "learning_rate": 1.7530424122385953e-06, + "loss": 0.7421, + "step": 8704 + }, + { + "epoch": 0.7417980400511291, + "grad_norm": 24.75, + "learning_rate": 1.75288857971152e-06, + "loss": 0.8814, + "step": 8705 + }, + { + "epoch": 0.7418832552194291, + "grad_norm": 12.1875, + "learning_rate": 1.7527347380967424e-06, + "loss": 0.5617, + "step": 8706 + }, + { + "epoch": 0.7419684703877291, + "grad_norm": 14.625, + "learning_rate": 1.7525808873970424e-06, + "loss": 0.8583, + "step": 8707 + }, + { + "epoch": 0.742053685556029, + "grad_norm": 13.0625, + "learning_rate": 1.7524270276152003e-06, + "loss": 0.6855, + "step": 8708 + }, + { + "epoch": 0.742138900724329, + "grad_norm": 8.1875, + "learning_rate": 1.7522731587539967e-06, + "loss": 0.2737, + "step": 8709 + }, + { + "epoch": 0.7422241158926289, + "grad_norm": 12.8125, + "learning_rate": 1.7521192808162121e-06, + "loss": 0.4955, + "step": 8710 + }, + { + "epoch": 0.7423093310609289, + "grad_norm": 15.3125, + "learning_rate": 1.7519653938046271e-06, + "loss": 0.8397, + "step": 8711 + }, + { + "epoch": 0.7423945462292288, + "grad_norm": 14.1875, + "learning_rate": 1.7518114977220223e-06, + "loss": 1.0708, + "step": 8712 + }, + { + "epoch": 0.7424797613975288, + "grad_norm": 14.5, + "learning_rate": 1.75165759257118e-06, + "loss": 0.9256, + "step": 8713 + }, + { + "epoch": 0.7425649765658288, + "grad_norm": 12.5625, + "learning_rate": 1.75150367835488e-06, + "loss": 0.8408, + "step": 8714 + }, + { + "epoch": 0.7426501917341287, + "grad_norm": 20.0, + "learning_rate": 1.751349755075904e-06, + "loss": 1.176, + "step": 8715 + }, + { + "epoch": 0.7427354069024287, + "grad_norm": 19.875, + "learning_rate": 1.751195822737034e-06, + "loss": 0.818, + "step": 8716 + }, + { + "epoch": 0.7428206220707286, + "grad_norm": 24.125, + "learning_rate": 1.7510418813410513e-06, + "loss": 1.343, + "step": 8717 + }, + { + "epoch": 0.7429058372390286, + "grad_norm": 11.375, + "learning_rate": 1.7508879308907387e-06, + "loss": 0.3106, + "step": 8718 + }, + { + "epoch": 0.7429910524073285, + "grad_norm": 14.125, + "learning_rate": 1.7507339713888765e-06, + "loss": 0.6693, + "step": 8719 + }, + { + "epoch": 0.7430762675756285, + "grad_norm": 23.75, + "learning_rate": 1.7505800028382485e-06, + "loss": 1.0575, + "step": 8720 + }, + { + "epoch": 0.7431614827439285, + "grad_norm": 13.5625, + "learning_rate": 1.7504260252416363e-06, + "loss": 0.7606, + "step": 8721 + }, + { + "epoch": 0.7432466979122284, + "grad_norm": 16.5, + "learning_rate": 1.7502720386018233e-06, + "loss": 0.8169, + "step": 8722 + }, + { + "epoch": 0.7433319130805284, + "grad_norm": 19.125, + "learning_rate": 1.7501180429215905e-06, + "loss": 0.9029, + "step": 8723 + }, + { + "epoch": 0.7434171282488283, + "grad_norm": 17.875, + "learning_rate": 1.7499640382037222e-06, + "loss": 0.7621, + "step": 8724 + }, + { + "epoch": 0.7435023434171283, + "grad_norm": 15.0625, + "learning_rate": 1.7498100244510007e-06, + "loss": 0.9402, + "step": 8725 + }, + { + "epoch": 0.7435875585854282, + "grad_norm": 14.6875, + "learning_rate": 1.7496560016662102e-06, + "loss": 1.0549, + "step": 8726 + }, + { + "epoch": 0.7436727737537282, + "grad_norm": 17.125, + "learning_rate": 1.7495019698521327e-06, + "loss": 1.1288, + "step": 8727 + }, + { + "epoch": 0.7437579889220282, + "grad_norm": 16.75, + "learning_rate": 1.7493479290115523e-06, + "loss": 0.5801, + "step": 8728 + }, + { + "epoch": 0.7438432040903281, + "grad_norm": 16.625, + "learning_rate": 1.7491938791472527e-06, + "loss": 0.3154, + "step": 8729 + }, + { + "epoch": 0.7439284192586281, + "grad_norm": 11.25, + "learning_rate": 1.749039820262018e-06, + "loss": 0.5015, + "step": 8730 + }, + { + "epoch": 0.744013634426928, + "grad_norm": 17.0, + "learning_rate": 1.7488857523586317e-06, + "loss": 0.6858, + "step": 8731 + }, + { + "epoch": 0.744098849595228, + "grad_norm": 21.375, + "learning_rate": 1.748731675439878e-06, + "loss": 0.5715, + "step": 8732 + }, + { + "epoch": 0.7441840647635279, + "grad_norm": 16.75, + "learning_rate": 1.7485775895085416e-06, + "loss": 0.4485, + "step": 8733 + }, + { + "epoch": 0.7442692799318279, + "grad_norm": 12.1875, + "learning_rate": 1.7484234945674072e-06, + "loss": 0.5973, + "step": 8734 + }, + { + "epoch": 0.7443544951001279, + "grad_norm": 16.25, + "learning_rate": 1.7482693906192588e-06, + "loss": 0.495, + "step": 8735 + }, + { + "epoch": 0.7444397102684278, + "grad_norm": 11.625, + "learning_rate": 1.7481152776668812e-06, + "loss": 1.3128, + "step": 8736 + }, + { + "epoch": 0.7445249254367278, + "grad_norm": 14.75, + "learning_rate": 1.74796115571306e-06, + "loss": 0.7134, + "step": 8737 + }, + { + "epoch": 0.7446101406050277, + "grad_norm": 15.125, + "learning_rate": 1.7478070247605798e-06, + "loss": 0.6541, + "step": 8738 + }, + { + "epoch": 0.7446953557733277, + "grad_norm": 21.125, + "learning_rate": 1.747652884812226e-06, + "loss": 0.7929, + "step": 8739 + }, + { + "epoch": 0.7447805709416276, + "grad_norm": 35.25, + "learning_rate": 1.747498735870784e-06, + "loss": 1.3412, + "step": 8740 + }, + { + "epoch": 0.7448657861099276, + "grad_norm": 20.875, + "learning_rate": 1.7473445779390396e-06, + "loss": 1.2407, + "step": 8741 + }, + { + "epoch": 0.7449510012782276, + "grad_norm": 16.75, + "learning_rate": 1.747190411019779e-06, + "loss": 0.9036, + "step": 8742 + }, + { + "epoch": 0.7450362164465275, + "grad_norm": 18.25, + "learning_rate": 1.7470362351157875e-06, + "loss": 0.8357, + "step": 8743 + }, + { + "epoch": 0.7451214316148275, + "grad_norm": 18.625, + "learning_rate": 1.7468820502298511e-06, + "loss": 0.9005, + "step": 8744 + }, + { + "epoch": 0.7452066467831274, + "grad_norm": 12.625, + "learning_rate": 1.746727856364757e-06, + "loss": 0.5742, + "step": 8745 + }, + { + "epoch": 0.7452918619514274, + "grad_norm": 16.875, + "learning_rate": 1.7465736535232901e-06, + "loss": 0.7629, + "step": 8746 + }, + { + "epoch": 0.7453770771197273, + "grad_norm": 16.75, + "learning_rate": 1.746419441708239e-06, + "loss": 0.8716, + "step": 8747 + }, + { + "epoch": 0.7454622922880273, + "grad_norm": 14.9375, + "learning_rate": 1.7462652209223888e-06, + "loss": 0.5686, + "step": 8748 + }, + { + "epoch": 0.7455475074563273, + "grad_norm": 14.4375, + "learning_rate": 1.7461109911685272e-06, + "loss": 0.6032, + "step": 8749 + }, + { + "epoch": 0.7456327226246272, + "grad_norm": 11.0, + "learning_rate": 1.7459567524494409e-06, + "loss": 0.3629, + "step": 8750 + }, + { + "epoch": 0.7457179377929272, + "grad_norm": 14.1875, + "learning_rate": 1.7458025047679178e-06, + "loss": 0.6784, + "step": 8751 + }, + { + "epoch": 0.7458031529612271, + "grad_norm": 26.25, + "learning_rate": 1.7456482481267445e-06, + "loss": 0.576, + "step": 8752 + }, + { + "epoch": 0.7458883681295271, + "grad_norm": 25.0, + "learning_rate": 1.7454939825287087e-06, + "loss": 0.7948, + "step": 8753 + }, + { + "epoch": 0.745973583297827, + "grad_norm": 18.625, + "learning_rate": 1.7453397079765991e-06, + "loss": 0.8094, + "step": 8754 + }, + { + "epoch": 0.746058798466127, + "grad_norm": 15.0, + "learning_rate": 1.7451854244732025e-06, + "loss": 0.8838, + "step": 8755 + }, + { + "epoch": 0.746144013634427, + "grad_norm": 15.8125, + "learning_rate": 1.7450311320213074e-06, + "loss": 0.779, + "step": 8756 + }, + { + "epoch": 0.7462292288027269, + "grad_norm": 19.25, + "learning_rate": 1.7448768306237015e-06, + "loss": 0.7891, + "step": 8757 + }, + { + "epoch": 0.7463144439710269, + "grad_norm": 15.5625, + "learning_rate": 1.744722520283174e-06, + "loss": 0.9333, + "step": 8758 + }, + { + "epoch": 0.7463996591393268, + "grad_norm": 16.75, + "learning_rate": 1.7445682010025133e-06, + "loss": 0.6183, + "step": 8759 + }, + { + "epoch": 0.7464848743076268, + "grad_norm": 12.125, + "learning_rate": 1.7444138727845078e-06, + "loss": 0.3199, + "step": 8760 + }, + { + "epoch": 0.7465700894759267, + "grad_norm": 13.125, + "learning_rate": 1.7442595356319461e-06, + "loss": 0.4685, + "step": 8761 + }, + { + "epoch": 0.7466553046442267, + "grad_norm": 18.25, + "learning_rate": 1.744105189547618e-06, + "loss": 0.9534, + "step": 8762 + }, + { + "epoch": 0.7467405198125266, + "grad_norm": 14.5, + "learning_rate": 1.743950834534312e-06, + "loss": 0.7155, + "step": 8763 + }, + { + "epoch": 0.7468257349808266, + "grad_norm": 14.0625, + "learning_rate": 1.743796470594818e-06, + "loss": 0.6915, + "step": 8764 + }, + { + "epoch": 0.7469109501491266, + "grad_norm": 17.0, + "learning_rate": 1.7436420977319246e-06, + "loss": 0.9821, + "step": 8765 + }, + { + "epoch": 0.7469961653174265, + "grad_norm": 14.875, + "learning_rate": 1.7434877159484225e-06, + "loss": 0.763, + "step": 8766 + }, + { + "epoch": 0.7470813804857265, + "grad_norm": 17.625, + "learning_rate": 1.7433333252471013e-06, + "loss": 0.6521, + "step": 8767 + }, + { + "epoch": 0.7471665956540264, + "grad_norm": 14.875, + "learning_rate": 1.7431789256307505e-06, + "loss": 0.7515, + "step": 8768 + }, + { + "epoch": 0.7472518108223264, + "grad_norm": 12.0, + "learning_rate": 1.7430245171021608e-06, + "loss": 0.3938, + "step": 8769 + }, + { + "epoch": 0.7473370259906263, + "grad_norm": 15.4375, + "learning_rate": 1.742870099664122e-06, + "loss": 0.9272, + "step": 8770 + }, + { + "epoch": 0.7474222411589263, + "grad_norm": 12.6875, + "learning_rate": 1.7427156733194255e-06, + "loss": 0.4833, + "step": 8771 + }, + { + "epoch": 0.7475074563272263, + "grad_norm": 15.5, + "learning_rate": 1.742561238070861e-06, + "loss": 0.8576, + "step": 8772 + }, + { + "epoch": 0.7475926714955262, + "grad_norm": 18.375, + "learning_rate": 1.7424067939212195e-06, + "loss": 0.9216, + "step": 8773 + }, + { + "epoch": 0.7476778866638262, + "grad_norm": 13.4375, + "learning_rate": 1.742252340873292e-06, + "loss": 0.6485, + "step": 8774 + }, + { + "epoch": 0.7477631018321261, + "grad_norm": 12.8125, + "learning_rate": 1.7420978789298702e-06, + "loss": 0.8847, + "step": 8775 + }, + { + "epoch": 0.7478483170004261, + "grad_norm": 13.9375, + "learning_rate": 1.7419434080937442e-06, + "loss": 0.508, + "step": 8776 + }, + { + "epoch": 0.747933532168726, + "grad_norm": 18.0, + "learning_rate": 1.7417889283677066e-06, + "loss": 1.007, + "step": 8777 + }, + { + "epoch": 0.748018747337026, + "grad_norm": 10.25, + "learning_rate": 1.7416344397545482e-06, + "loss": 0.291, + "step": 8778 + }, + { + "epoch": 0.748103962505326, + "grad_norm": 15.0, + "learning_rate": 1.7414799422570616e-06, + "loss": 0.864, + "step": 8779 + }, + { + "epoch": 0.7481891776736259, + "grad_norm": 16.125, + "learning_rate": 1.7413254358780377e-06, + "loss": 0.9047, + "step": 8780 + }, + { + "epoch": 0.7482743928419259, + "grad_norm": 12.0625, + "learning_rate": 1.7411709206202692e-06, + "loss": 0.646, + "step": 8781 + }, + { + "epoch": 0.7483596080102258, + "grad_norm": 16.5, + "learning_rate": 1.7410163964865483e-06, + "loss": 0.7902, + "step": 8782 + }, + { + "epoch": 0.7484448231785258, + "grad_norm": 23.5, + "learning_rate": 1.7408618634796675e-06, + "loss": 0.8568, + "step": 8783 + }, + { + "epoch": 0.7485300383468257, + "grad_norm": 18.0, + "learning_rate": 1.7407073216024195e-06, + "loss": 0.8488, + "step": 8784 + }, + { + "epoch": 0.7486152535151257, + "grad_norm": 13.8125, + "learning_rate": 1.7405527708575962e-06, + "loss": 0.7358, + "step": 8785 + }, + { + "epoch": 0.7487004686834257, + "grad_norm": 16.625, + "learning_rate": 1.7403982112479912e-06, + "loss": 0.7834, + "step": 8786 + }, + { + "epoch": 0.7487856838517256, + "grad_norm": 14.875, + "learning_rate": 1.7402436427763974e-06, + "loss": 0.8336, + "step": 8787 + }, + { + "epoch": 0.7488708990200256, + "grad_norm": 19.125, + "learning_rate": 1.7400890654456083e-06, + "loss": 0.5317, + "step": 8788 + }, + { + "epoch": 0.7489561141883255, + "grad_norm": 10.0, + "learning_rate": 1.7399344792584163e-06, + "loss": 0.3257, + "step": 8789 + }, + { + "epoch": 0.7490413293566255, + "grad_norm": 14.375, + "learning_rate": 1.739779884217616e-06, + "loss": 0.3037, + "step": 8790 + }, + { + "epoch": 0.7491265445249254, + "grad_norm": 13.375, + "learning_rate": 1.739625280326001e-06, + "loss": 0.78, + "step": 8791 + }, + { + "epoch": 0.7492117596932254, + "grad_norm": 16.625, + "learning_rate": 1.7394706675863645e-06, + "loss": 0.5396, + "step": 8792 + }, + { + "epoch": 0.7492969748615254, + "grad_norm": 13.0, + "learning_rate": 1.7393160460015007e-06, + "loss": 0.6034, + "step": 8793 + }, + { + "epoch": 0.7493821900298253, + "grad_norm": 13.3125, + "learning_rate": 1.7391614155742042e-06, + "loss": 0.5717, + "step": 8794 + }, + { + "epoch": 0.7494674051981253, + "grad_norm": 15.5, + "learning_rate": 1.7390067763072688e-06, + "loss": 0.7804, + "step": 8795 + }, + { + "epoch": 0.7495526203664252, + "grad_norm": 25.0, + "learning_rate": 1.7388521282034895e-06, + "loss": 0.6993, + "step": 8796 + }, + { + "epoch": 0.7496378355347252, + "grad_norm": 14.0625, + "learning_rate": 1.73869747126566e-06, + "loss": 0.3596, + "step": 8797 + }, + { + "epoch": 0.7497230507030251, + "grad_norm": 13.5, + "learning_rate": 1.7385428054965764e-06, + "loss": 0.8027, + "step": 8798 + }, + { + "epoch": 0.7498082658713251, + "grad_norm": 16.25, + "learning_rate": 1.738388130899033e-06, + "loss": 0.7779, + "step": 8799 + }, + { + "epoch": 0.749893481039625, + "grad_norm": 33.25, + "learning_rate": 1.7382334474758248e-06, + "loss": 0.7611, + "step": 8800 + }, + { + "epoch": 0.749978696207925, + "grad_norm": 15.6875, + "learning_rate": 1.7380787552297474e-06, + "loss": 0.8858, + "step": 8801 + }, + { + "epoch": 0.750063911376225, + "grad_norm": 17.375, + "learning_rate": 1.7379240541635957e-06, + "loss": 0.8558, + "step": 8802 + }, + { + "epoch": 0.7501491265445249, + "grad_norm": 11.0625, + "learning_rate": 1.7377693442801664e-06, + "loss": 0.4349, + "step": 8803 + }, + { + "epoch": 0.7502343417128249, + "grad_norm": 12.4375, + "learning_rate": 1.7376146255822537e-06, + "loss": 0.4543, + "step": 8804 + }, + { + "epoch": 0.7503195568811248, + "grad_norm": 12.0, + "learning_rate": 1.7374598980726548e-06, + "loss": 0.6571, + "step": 8805 + }, + { + "epoch": 0.7504047720494248, + "grad_norm": 21.5, + "learning_rate": 1.7373051617541654e-06, + "loss": 1.2828, + "step": 8806 + }, + { + "epoch": 0.7504899872177248, + "grad_norm": 14.9375, + "learning_rate": 1.7371504166295818e-06, + "loss": 0.549, + "step": 8807 + }, + { + "epoch": 0.7505752023860247, + "grad_norm": 11.25, + "learning_rate": 1.7369956627017e-06, + "loss": 0.721, + "step": 8808 + }, + { + "epoch": 0.7506604175543247, + "grad_norm": 18.0, + "learning_rate": 1.7368408999733172e-06, + "loss": 0.8061, + "step": 8809 + }, + { + "epoch": 0.7507456327226246, + "grad_norm": 16.0, + "learning_rate": 1.7366861284472295e-06, + "loss": 0.3442, + "step": 8810 + }, + { + "epoch": 0.7508308478909246, + "grad_norm": 17.125, + "learning_rate": 1.7365313481262344e-06, + "loss": 0.8165, + "step": 8811 + }, + { + "epoch": 0.7509160630592245, + "grad_norm": 15.8125, + "learning_rate": 1.7363765590131282e-06, + "loss": 0.8678, + "step": 8812 + }, + { + "epoch": 0.7510012782275245, + "grad_norm": 16.75, + "learning_rate": 1.7362217611107087e-06, + "loss": 0.7003, + "step": 8813 + }, + { + "epoch": 0.7510864933958245, + "grad_norm": 17.375, + "learning_rate": 1.7360669544217728e-06, + "loss": 0.6057, + "step": 8814 + }, + { + "epoch": 0.7511717085641244, + "grad_norm": 23.625, + "learning_rate": 1.7359121389491184e-06, + "loss": 1.0507, + "step": 8815 + }, + { + "epoch": 0.7512569237324244, + "grad_norm": 15.5, + "learning_rate": 1.7357573146955434e-06, + "loss": 0.7692, + "step": 8816 + }, + { + "epoch": 0.7513421389007243, + "grad_norm": 14.5625, + "learning_rate": 1.7356024816638447e-06, + "loss": 0.6305, + "step": 8817 + }, + { + "epoch": 0.7514273540690243, + "grad_norm": 15.6875, + "learning_rate": 1.7354476398568213e-06, + "loss": 0.4532, + "step": 8818 + }, + { + "epoch": 0.7515125692373242, + "grad_norm": 22.5, + "learning_rate": 1.7352927892772707e-06, + "loss": 1.1847, + "step": 8819 + }, + { + "epoch": 0.7515977844056242, + "grad_norm": 15.25, + "learning_rate": 1.7351379299279913e-06, + "loss": 0.539, + "step": 8820 + }, + { + "epoch": 0.7516829995739241, + "grad_norm": 10.75, + "learning_rate": 1.7349830618117815e-06, + "loss": 0.3418, + "step": 8821 + }, + { + "epoch": 0.7517682147422241, + "grad_norm": 10.9375, + "learning_rate": 1.7348281849314405e-06, + "loss": 0.3841, + "step": 8822 + }, + { + "epoch": 0.7518534299105241, + "grad_norm": 20.375, + "learning_rate": 1.7346732992897663e-06, + "loss": 1.0008, + "step": 8823 + }, + { + "epoch": 0.751938645078824, + "grad_norm": 13.3125, + "learning_rate": 1.7345184048895588e-06, + "loss": 0.8183, + "step": 8824 + }, + { + "epoch": 0.752023860247124, + "grad_norm": 14.5625, + "learning_rate": 1.7343635017336158e-06, + "loss": 0.8955, + "step": 8825 + }, + { + "epoch": 0.7521090754154239, + "grad_norm": 18.25, + "learning_rate": 1.7342085898247377e-06, + "loss": 0.8762, + "step": 8826 + }, + { + "epoch": 0.7521942905837239, + "grad_norm": 10.0625, + "learning_rate": 1.734053669165723e-06, + "loss": 0.2592, + "step": 8827 + }, + { + "epoch": 0.7522795057520238, + "grad_norm": 14.125, + "learning_rate": 1.7338987397593727e-06, + "loss": 0.9215, + "step": 8828 + }, + { + "epoch": 0.7523647209203238, + "grad_norm": 20.125, + "learning_rate": 1.7337438016084845e-06, + "loss": 1.1696, + "step": 8829 + }, + { + "epoch": 0.7524499360886238, + "grad_norm": 23.125, + "learning_rate": 1.73358885471586e-06, + "loss": 1.0865, + "step": 8830 + }, + { + "epoch": 0.7525351512569237, + "grad_norm": 17.125, + "learning_rate": 1.7334338990842982e-06, + "loss": 1.0081, + "step": 8831 + }, + { + "epoch": 0.7526203664252237, + "grad_norm": 13.0, + "learning_rate": 1.7332789347166e-06, + "loss": 0.5814, + "step": 8832 + }, + { + "epoch": 0.7527055815935236, + "grad_norm": 20.5, + "learning_rate": 1.7331239616155655e-06, + "loss": 1.1821, + "step": 8833 + }, + { + "epoch": 0.7527907967618236, + "grad_norm": 21.0, + "learning_rate": 1.7329689797839953e-06, + "loss": 0.8885, + "step": 8834 + }, + { + "epoch": 0.7528760119301235, + "grad_norm": 49.75, + "learning_rate": 1.7328139892246894e-06, + "loss": 0.995, + "step": 8835 + }, + { + "epoch": 0.7529612270984235, + "grad_norm": 13.0, + "learning_rate": 1.7326589899404498e-06, + "loss": 0.4504, + "step": 8836 + }, + { + "epoch": 0.7530464422667235, + "grad_norm": 16.5, + "learning_rate": 1.7325039819340767e-06, + "loss": 0.9001, + "step": 8837 + }, + { + "epoch": 0.7531316574350234, + "grad_norm": 14.1875, + "learning_rate": 1.7323489652083714e-06, + "loss": 0.6195, + "step": 8838 + }, + { + "epoch": 0.7532168726033234, + "grad_norm": 16.0, + "learning_rate": 1.7321939397661352e-06, + "loss": 0.7212, + "step": 8839 + }, + { + "epoch": 0.7533020877716233, + "grad_norm": 17.875, + "learning_rate": 1.73203890561017e-06, + "loss": 0.9277, + "step": 8840 + }, + { + "epoch": 0.7533873029399233, + "grad_norm": 14.0, + "learning_rate": 1.7318838627432765e-06, + "loss": 0.7808, + "step": 8841 + }, + { + "epoch": 0.7534725181082232, + "grad_norm": 18.0, + "learning_rate": 1.7317288111682573e-06, + "loss": 0.7741, + "step": 8842 + }, + { + "epoch": 0.7535577332765232, + "grad_norm": 14.5, + "learning_rate": 1.731573750887914e-06, + "loss": 0.8483, + "step": 8843 + }, + { + "epoch": 0.7536429484448232, + "grad_norm": 11.75, + "learning_rate": 1.7314186819050492e-06, + "loss": 0.43, + "step": 8844 + }, + { + "epoch": 0.7537281636131231, + "grad_norm": 22.875, + "learning_rate": 1.7312636042224643e-06, + "loss": 1.0391, + "step": 8845 + }, + { + "epoch": 0.7538133787814231, + "grad_norm": 14.625, + "learning_rate": 1.731108517842962e-06, + "loss": 1.134, + "step": 8846 + }, + { + "epoch": 0.753898593949723, + "grad_norm": 10.625, + "learning_rate": 1.7309534227693453e-06, + "loss": 0.3056, + "step": 8847 + }, + { + "epoch": 0.753983809118023, + "grad_norm": 14.0625, + "learning_rate": 1.7307983190044164e-06, + "loss": 0.4403, + "step": 8848 + }, + { + "epoch": 0.7540690242863229, + "grad_norm": 12.0, + "learning_rate": 1.7306432065509784e-06, + "loss": 0.4834, + "step": 8849 + }, + { + "epoch": 0.7541542394546229, + "grad_norm": 13.8125, + "learning_rate": 1.7304880854118345e-06, + "loss": 0.8416, + "step": 8850 + }, + { + "epoch": 0.7542394546229229, + "grad_norm": 15.3125, + "learning_rate": 1.7303329555897873e-06, + "loss": 0.9184, + "step": 8851 + }, + { + "epoch": 0.7543246697912228, + "grad_norm": 14.875, + "learning_rate": 1.7301778170876413e-06, + "loss": 0.7101, + "step": 8852 + }, + { + "epoch": 0.7544098849595228, + "grad_norm": 12.375, + "learning_rate": 1.7300226699081984e-06, + "loss": 0.5973, + "step": 8853 + }, + { + "epoch": 0.7544951001278227, + "grad_norm": 16.5, + "learning_rate": 1.729867514054264e-06, + "loss": 1.0795, + "step": 8854 + }, + { + "epoch": 0.7545803152961227, + "grad_norm": 15.25, + "learning_rate": 1.7297123495286406e-06, + "loss": 0.6148, + "step": 8855 + }, + { + "epoch": 0.7546655304644226, + "grad_norm": 14.1875, + "learning_rate": 1.729557176334133e-06, + "loss": 0.5185, + "step": 8856 + }, + { + "epoch": 0.7547507456327226, + "grad_norm": 14.25, + "learning_rate": 1.7294019944735447e-06, + "loss": 0.7482, + "step": 8857 + }, + { + "epoch": 0.7548359608010226, + "grad_norm": 11.0625, + "learning_rate": 1.7292468039496804e-06, + "loss": 0.4201, + "step": 8858 + }, + { + "epoch": 0.7549211759693225, + "grad_norm": 19.125, + "learning_rate": 1.7290916047653442e-06, + "loss": 0.8541, + "step": 8859 + }, + { + "epoch": 0.7550063911376225, + "grad_norm": 14.5, + "learning_rate": 1.7289363969233413e-06, + "loss": 0.8045, + "step": 8860 + }, + { + "epoch": 0.7550916063059224, + "grad_norm": 20.5, + "learning_rate": 1.728781180426476e-06, + "loss": 0.9231, + "step": 8861 + }, + { + "epoch": 0.7551768214742224, + "grad_norm": 13.9375, + "learning_rate": 1.728625955277553e-06, + "loss": 0.7687, + "step": 8862 + }, + { + "epoch": 0.7552620366425223, + "grad_norm": 17.0, + "learning_rate": 1.7284707214793777e-06, + "loss": 1.1376, + "step": 8863 + }, + { + "epoch": 0.7553472518108223, + "grad_norm": 15.0625, + "learning_rate": 1.728315479034756e-06, + "loss": 0.4929, + "step": 8864 + }, + { + "epoch": 0.7554324669791223, + "grad_norm": 23.875, + "learning_rate": 1.728160227946492e-06, + "loss": 0.9484, + "step": 8865 + }, + { + "epoch": 0.7555176821474222, + "grad_norm": 11.8125, + "learning_rate": 1.728004968217392e-06, + "loss": 0.5762, + "step": 8866 + }, + { + "epoch": 0.7556028973157222, + "grad_norm": 14.875, + "learning_rate": 1.7278496998502614e-06, + "loss": 0.6943, + "step": 8867 + }, + { + "epoch": 0.7556881124840221, + "grad_norm": 12.375, + "learning_rate": 1.7276944228479067e-06, + "loss": 0.4078, + "step": 8868 + }, + { + "epoch": 0.7557733276523221, + "grad_norm": 15.6875, + "learning_rate": 1.7275391372131331e-06, + "loss": 0.6442, + "step": 8869 + }, + { + "epoch": 0.755858542820622, + "grad_norm": 15.75, + "learning_rate": 1.7273838429487468e-06, + "loss": 1.0687, + "step": 8870 + }, + { + "epoch": 0.755943757988922, + "grad_norm": 11.75, + "learning_rate": 1.727228540057555e-06, + "loss": 0.6581, + "step": 8871 + }, + { + "epoch": 0.756028973157222, + "grad_norm": 11.875, + "learning_rate": 1.7270732285423633e-06, + "loss": 0.4837, + "step": 8872 + }, + { + "epoch": 0.7561141883255219, + "grad_norm": 16.5, + "learning_rate": 1.7269179084059788e-06, + "loss": 0.6691, + "step": 8873 + }, + { + "epoch": 0.7561994034938219, + "grad_norm": 29.25, + "learning_rate": 1.7267625796512078e-06, + "loss": 0.9593, + "step": 8874 + }, + { + "epoch": 0.7562846186621218, + "grad_norm": 18.25, + "learning_rate": 1.7266072422808578e-06, + "loss": 0.6315, + "step": 8875 + }, + { + "epoch": 0.7563698338304218, + "grad_norm": 14.0625, + "learning_rate": 1.7264518962977356e-06, + "loss": 0.7552, + "step": 8876 + }, + { + "epoch": 0.7564550489987217, + "grad_norm": 14.9375, + "learning_rate": 1.7262965417046483e-06, + "loss": 0.6384, + "step": 8877 + }, + { + "epoch": 0.7565402641670217, + "grad_norm": 13.375, + "learning_rate": 1.7261411785044036e-06, + "loss": 0.5805, + "step": 8878 + }, + { + "epoch": 0.7566254793353216, + "grad_norm": 17.75, + "learning_rate": 1.7259858066998091e-06, + "loss": 1.0247, + "step": 8879 + }, + { + "epoch": 0.7567106945036216, + "grad_norm": 11.25, + "learning_rate": 1.7258304262936726e-06, + "loss": 0.3864, + "step": 8880 + }, + { + "epoch": 0.7567959096719216, + "grad_norm": 10.375, + "learning_rate": 1.725675037288802e-06, + "loss": 0.4187, + "step": 8881 + }, + { + "epoch": 0.7568811248402215, + "grad_norm": 11.875, + "learning_rate": 1.7255196396880044e-06, + "loss": 0.5016, + "step": 8882 + }, + { + "epoch": 0.7569663400085215, + "grad_norm": 13.4375, + "learning_rate": 1.7253642334940896e-06, + "loss": 0.7801, + "step": 8883 + }, + { + "epoch": 0.7570515551768214, + "grad_norm": 15.0625, + "learning_rate": 1.7252088187098642e-06, + "loss": 0.5035, + "step": 8884 + }, + { + "epoch": 0.7571367703451214, + "grad_norm": 18.0, + "learning_rate": 1.7250533953381381e-06, + "loss": 0.6229, + "step": 8885 + }, + { + "epoch": 0.7572219855134213, + "grad_norm": 15.8125, + "learning_rate": 1.7248979633817195e-06, + "loss": 0.6052, + "step": 8886 + }, + { + "epoch": 0.7573072006817213, + "grad_norm": 13.8125, + "learning_rate": 1.724742522843417e-06, + "loss": 0.7106, + "step": 8887 + }, + { + "epoch": 0.7573924158500213, + "grad_norm": 14.1875, + "learning_rate": 1.7245870737260398e-06, + "loss": 0.612, + "step": 8888 + }, + { + "epoch": 0.7574776310183212, + "grad_norm": 13.6875, + "learning_rate": 1.724431616032397e-06, + "loss": 0.842, + "step": 8889 + }, + { + "epoch": 0.7575628461866212, + "grad_norm": 21.75, + "learning_rate": 1.7242761497652976e-06, + "loss": 1.211, + "step": 8890 + }, + { + "epoch": 0.7576480613549211, + "grad_norm": 12.75, + "learning_rate": 1.7241206749275512e-06, + "loss": 0.6871, + "step": 8891 + }, + { + "epoch": 0.7577332765232211, + "grad_norm": 18.875, + "learning_rate": 1.723965191521968e-06, + "loss": 0.998, + "step": 8892 + }, + { + "epoch": 0.757818491691521, + "grad_norm": 11.75, + "learning_rate": 1.7238096995513567e-06, + "loss": 0.4797, + "step": 8893 + }, + { + "epoch": 0.757903706859821, + "grad_norm": 13.8125, + "learning_rate": 1.7236541990185279e-06, + "loss": 0.7879, + "step": 8894 + }, + { + "epoch": 0.757988922028121, + "grad_norm": 15.625, + "learning_rate": 1.7234986899262912e-06, + "loss": 0.8003, + "step": 8895 + }, + { + "epoch": 0.7580741371964209, + "grad_norm": 14.6875, + "learning_rate": 1.723343172277457e-06, + "loss": 0.606, + "step": 8896 + }, + { + "epoch": 0.7581593523647209, + "grad_norm": 18.625, + "learning_rate": 1.723187646074836e-06, + "loss": 1.0115, + "step": 8897 + }, + { + "epoch": 0.7582445675330208, + "grad_norm": 13.0625, + "learning_rate": 1.7230321113212383e-06, + "loss": 0.5035, + "step": 8898 + }, + { + "epoch": 0.7583297827013208, + "grad_norm": 16.5, + "learning_rate": 1.7228765680194745e-06, + "loss": 0.8066, + "step": 8899 + }, + { + "epoch": 0.7584149978696207, + "grad_norm": 12.25, + "learning_rate": 1.7227210161723556e-06, + "loss": 0.6151, + "step": 8900 + }, + { + "epoch": 0.7585002130379207, + "grad_norm": 14.4375, + "learning_rate": 1.722565455782693e-06, + "loss": 0.8543, + "step": 8901 + }, + { + "epoch": 0.7585854282062207, + "grad_norm": 19.25, + "learning_rate": 1.7224098868532968e-06, + "loss": 0.8868, + "step": 8902 + }, + { + "epoch": 0.7586706433745206, + "grad_norm": 15.875, + "learning_rate": 1.7222543093869793e-06, + "loss": 0.8081, + "step": 8903 + }, + { + "epoch": 0.7587558585428206, + "grad_norm": 13.0, + "learning_rate": 1.7220987233865517e-06, + "loss": 0.3198, + "step": 8904 + }, + { + "epoch": 0.7588410737111205, + "grad_norm": 19.75, + "learning_rate": 1.721943128854825e-06, + "loss": 0.8929, + "step": 8905 + }, + { + "epoch": 0.7589262888794205, + "grad_norm": 11.4375, + "learning_rate": 1.7217875257946115e-06, + "loss": 0.3741, + "step": 8906 + }, + { + "epoch": 0.7590115040477204, + "grad_norm": 17.125, + "learning_rate": 1.7216319142087232e-06, + "loss": 0.8314, + "step": 8907 + }, + { + "epoch": 0.7590967192160204, + "grad_norm": 31.5, + "learning_rate": 1.7214762940999716e-06, + "loss": 1.2375, + "step": 8908 + }, + { + "epoch": 0.7591819343843204, + "grad_norm": 15.5625, + "learning_rate": 1.7213206654711697e-06, + "loss": 0.9969, + "step": 8909 + }, + { + "epoch": 0.7592671495526203, + "grad_norm": 50.0, + "learning_rate": 1.721165028325129e-06, + "loss": 0.6129, + "step": 8910 + }, + { + "epoch": 0.7593523647209203, + "grad_norm": 14.75, + "learning_rate": 1.7210093826646628e-06, + "loss": 0.8879, + "step": 8911 + }, + { + "epoch": 0.7594375798892202, + "grad_norm": 16.75, + "learning_rate": 1.720853728492583e-06, + "loss": 0.8162, + "step": 8912 + }, + { + "epoch": 0.7595227950575202, + "grad_norm": 19.125, + "learning_rate": 1.7206980658117034e-06, + "loss": 0.967, + "step": 8913 + }, + { + "epoch": 0.7596080102258201, + "grad_norm": 16.875, + "learning_rate": 1.720542394624836e-06, + "loss": 0.6018, + "step": 8914 + }, + { + "epoch": 0.7596932253941202, + "grad_norm": 12.8125, + "learning_rate": 1.7203867149347946e-06, + "loss": 0.3263, + "step": 8915 + }, + { + "epoch": 0.7597784405624202, + "grad_norm": 9.75, + "learning_rate": 1.720231026744392e-06, + "loss": 0.203, + "step": 8916 + }, + { + "epoch": 0.7598636557307201, + "grad_norm": 12.75, + "learning_rate": 1.7200753300564422e-06, + "loss": 0.576, + "step": 8917 + }, + { + "epoch": 0.7599488708990201, + "grad_norm": 14.5, + "learning_rate": 1.7199196248737582e-06, + "loss": 0.7405, + "step": 8918 + }, + { + "epoch": 0.76003408606732, + "grad_norm": 16.125, + "learning_rate": 1.7197639111991543e-06, + "loss": 0.9682, + "step": 8919 + }, + { + "epoch": 0.76011930123562, + "grad_norm": 17.625, + "learning_rate": 1.7196081890354439e-06, + "loss": 0.8688, + "step": 8920 + }, + { + "epoch": 0.76020451640392, + "grad_norm": 12.0625, + "learning_rate": 1.7194524583854414e-06, + "loss": 0.5735, + "step": 8921 + }, + { + "epoch": 0.7602897315722199, + "grad_norm": 14.25, + "learning_rate": 1.719296719251961e-06, + "loss": 0.5425, + "step": 8922 + }, + { + "epoch": 0.7603749467405199, + "grad_norm": 13.75, + "learning_rate": 1.7191409716378165e-06, + "loss": 0.4827, + "step": 8923 + }, + { + "epoch": 0.7604601619088198, + "grad_norm": 15.875, + "learning_rate": 1.7189852155458236e-06, + "loss": 0.9541, + "step": 8924 + }, + { + "epoch": 0.7605453770771198, + "grad_norm": 18.625, + "learning_rate": 1.718829450978796e-06, + "loss": 0.9378, + "step": 8925 + }, + { + "epoch": 0.7606305922454197, + "grad_norm": 12.25, + "learning_rate": 1.7186736779395485e-06, + "loss": 0.5963, + "step": 8926 + }, + { + "epoch": 0.7607158074137197, + "grad_norm": 12.875, + "learning_rate": 1.7185178964308965e-06, + "loss": 0.5026, + "step": 8927 + }, + { + "epoch": 0.7608010225820196, + "grad_norm": 18.875, + "learning_rate": 1.7183621064556551e-06, + "loss": 0.6259, + "step": 8928 + }, + { + "epoch": 0.7608862377503196, + "grad_norm": 17.375, + "learning_rate": 1.7182063080166395e-06, + "loss": 0.9694, + "step": 8929 + }, + { + "epoch": 0.7609714529186196, + "grad_norm": 10.0625, + "learning_rate": 1.718050501116665e-06, + "loss": 0.4604, + "step": 8930 + }, + { + "epoch": 0.7610566680869195, + "grad_norm": 10.8125, + "learning_rate": 1.7178946857585472e-06, + "loss": 0.4042, + "step": 8931 + }, + { + "epoch": 0.7611418832552195, + "grad_norm": 12.5625, + "learning_rate": 1.717738861945102e-06, + "loss": 0.2951, + "step": 8932 + }, + { + "epoch": 0.7612270984235194, + "grad_norm": 14.8125, + "learning_rate": 1.7175830296791451e-06, + "loss": 1.141, + "step": 8933 + }, + { + "epoch": 0.7613123135918194, + "grad_norm": 14.0, + "learning_rate": 1.717427188963493e-06, + "loss": 0.809, + "step": 8934 + }, + { + "epoch": 0.7613975287601193, + "grad_norm": 14.25, + "learning_rate": 1.717271339800961e-06, + "loss": 0.8487, + "step": 8935 + }, + { + "epoch": 0.7614827439284193, + "grad_norm": 17.0, + "learning_rate": 1.7171154821943664e-06, + "loss": 0.6533, + "step": 8936 + }, + { + "epoch": 0.7615679590967193, + "grad_norm": 13.875, + "learning_rate": 1.7169596161465256e-06, + "loss": 0.6976, + "step": 8937 + }, + { + "epoch": 0.7616531742650192, + "grad_norm": 17.875, + "learning_rate": 1.7168037416602543e-06, + "loss": 0.9174, + "step": 8938 + }, + { + "epoch": 0.7617383894333192, + "grad_norm": 15.3125, + "learning_rate": 1.7166478587383707e-06, + "loss": 0.784, + "step": 8939 + }, + { + "epoch": 0.7618236046016191, + "grad_norm": 11.0625, + "learning_rate": 1.7164919673836902e-06, + "loss": 0.57, + "step": 8940 + }, + { + "epoch": 0.7619088197699191, + "grad_norm": 16.5, + "learning_rate": 1.7163360675990315e-06, + "loss": 0.8436, + "step": 8941 + }, + { + "epoch": 0.761994034938219, + "grad_norm": 14.9375, + "learning_rate": 1.7161801593872106e-06, + "loss": 0.4477, + "step": 8942 + }, + { + "epoch": 0.762079250106519, + "grad_norm": 15.375, + "learning_rate": 1.716024242751046e-06, + "loss": 0.8258, + "step": 8943 + }, + { + "epoch": 0.762164465274819, + "grad_norm": 14.375, + "learning_rate": 1.7158683176933542e-06, + "loss": 0.8524, + "step": 8944 + }, + { + "epoch": 0.7622496804431189, + "grad_norm": 19.5, + "learning_rate": 1.7157123842169537e-06, + "loss": 0.8037, + "step": 8945 + }, + { + "epoch": 0.7623348956114189, + "grad_norm": 12.1875, + "learning_rate": 1.7155564423246617e-06, + "loss": 0.4825, + "step": 8946 + }, + { + "epoch": 0.7624201107797188, + "grad_norm": 12.1875, + "learning_rate": 1.7154004920192973e-06, + "loss": 0.6762, + "step": 8947 + }, + { + "epoch": 0.7625053259480188, + "grad_norm": 13.9375, + "learning_rate": 1.7152445333036774e-06, + "loss": 0.4607, + "step": 8948 + }, + { + "epoch": 0.7625905411163187, + "grad_norm": 15.5625, + "learning_rate": 1.7150885661806212e-06, + "loss": 0.6954, + "step": 8949 + }, + { + "epoch": 0.7626757562846187, + "grad_norm": 13.4375, + "learning_rate": 1.7149325906529471e-06, + "loss": 0.5826, + "step": 8950 + }, + { + "epoch": 0.7627609714529187, + "grad_norm": 20.5, + "learning_rate": 1.7147766067234734e-06, + "loss": 1.1907, + "step": 8951 + }, + { + "epoch": 0.7628461866212186, + "grad_norm": 17.75, + "learning_rate": 1.714620614395019e-06, + "loss": 1.0698, + "step": 8952 + }, + { + "epoch": 0.7629314017895186, + "grad_norm": 21.125, + "learning_rate": 1.714464613670403e-06, + "loss": 0.8887, + "step": 8953 + }, + { + "epoch": 0.7630166169578185, + "grad_norm": 13.5625, + "learning_rate": 1.7143086045524443e-06, + "loss": 0.73, + "step": 8954 + }, + { + "epoch": 0.7631018321261185, + "grad_norm": 11.0, + "learning_rate": 1.7141525870439622e-06, + "loss": 0.4569, + "step": 8955 + }, + { + "epoch": 0.7631870472944184, + "grad_norm": 36.25, + "learning_rate": 1.713996561147776e-06, + "loss": 1.3506, + "step": 8956 + }, + { + "epoch": 0.7632722624627184, + "grad_norm": 11.3125, + "learning_rate": 1.7138405268667058e-06, + "loss": 0.5579, + "step": 8957 + }, + { + "epoch": 0.7633574776310184, + "grad_norm": 15.125, + "learning_rate": 1.7136844842035704e-06, + "loss": 0.8242, + "step": 8958 + }, + { + "epoch": 0.7634426927993183, + "grad_norm": 10.375, + "learning_rate": 1.71352843316119e-06, + "loss": 0.5989, + "step": 8959 + }, + { + "epoch": 0.7635279079676183, + "grad_norm": 16.375, + "learning_rate": 1.7133723737423852e-06, + "loss": 1.0877, + "step": 8960 + }, + { + "epoch": 0.7636131231359182, + "grad_norm": 18.25, + "learning_rate": 1.7132163059499751e-06, + "loss": 1.0761, + "step": 8961 + }, + { + "epoch": 0.7636983383042182, + "grad_norm": 15.125, + "learning_rate": 1.7130602297867814e-06, + "loss": 0.4392, + "step": 8962 + }, + { + "epoch": 0.7637835534725181, + "grad_norm": 15.3125, + "learning_rate": 1.7129041452556227e-06, + "loss": 0.837, + "step": 8963 + }, + { + "epoch": 0.7638687686408181, + "grad_norm": 14.6875, + "learning_rate": 1.7127480523593213e-06, + "loss": 0.6768, + "step": 8964 + }, + { + "epoch": 0.763953983809118, + "grad_norm": 17.375, + "learning_rate": 1.7125919511006966e-06, + "loss": 0.6845, + "step": 8965 + }, + { + "epoch": 0.764039198977418, + "grad_norm": 15.3125, + "learning_rate": 1.712435841482571e-06, + "loss": 0.3896, + "step": 8966 + }, + { + "epoch": 0.764124414145718, + "grad_norm": 30.25, + "learning_rate": 1.7122797235077642e-06, + "loss": 1.0668, + "step": 8967 + }, + { + "epoch": 0.7642096293140179, + "grad_norm": 10.625, + "learning_rate": 1.7121235971790982e-06, + "loss": 0.2162, + "step": 8968 + }, + { + "epoch": 0.7642948444823179, + "grad_norm": 12.1875, + "learning_rate": 1.7119674624993937e-06, + "loss": 0.6982, + "step": 8969 + }, + { + "epoch": 0.7643800596506178, + "grad_norm": 12.25, + "learning_rate": 1.711811319471473e-06, + "loss": 0.4853, + "step": 8970 + }, + { + "epoch": 0.7644652748189178, + "grad_norm": 10.5625, + "learning_rate": 1.7116551680981572e-06, + "loss": 0.339, + "step": 8971 + }, + { + "epoch": 0.7645504899872178, + "grad_norm": 12.3125, + "learning_rate": 1.7114990083822686e-06, + "loss": 0.7008, + "step": 8972 + }, + { + "epoch": 0.7646357051555177, + "grad_norm": 11.375, + "learning_rate": 1.7113428403266282e-06, + "loss": 0.7619, + "step": 8973 + }, + { + "epoch": 0.7647209203238177, + "grad_norm": 16.0, + "learning_rate": 1.711186663934059e-06, + "loss": 1.0166, + "step": 8974 + }, + { + "epoch": 0.7648061354921176, + "grad_norm": 13.625, + "learning_rate": 1.7110304792073835e-06, + "loss": 0.5381, + "step": 8975 + }, + { + "epoch": 0.7648913506604176, + "grad_norm": 15.0, + "learning_rate": 1.7108742861494233e-06, + "loss": 0.8544, + "step": 8976 + }, + { + "epoch": 0.7649765658287175, + "grad_norm": 20.375, + "learning_rate": 1.710718084763001e-06, + "loss": 0.9204, + "step": 8977 + }, + { + "epoch": 0.7650617809970175, + "grad_norm": 19.125, + "learning_rate": 1.71056187505094e-06, + "loss": 0.9626, + "step": 8978 + }, + { + "epoch": 0.7651469961653175, + "grad_norm": 16.0, + "learning_rate": 1.710405657016063e-06, + "loss": 0.7943, + "step": 8979 + }, + { + "epoch": 0.7652322113336174, + "grad_norm": 15.75, + "learning_rate": 1.7102494306611921e-06, + "loss": 0.9519, + "step": 8980 + }, + { + "epoch": 0.7653174265019174, + "grad_norm": 14.1875, + "learning_rate": 1.7100931959891517e-06, + "loss": 1.0013, + "step": 8981 + }, + { + "epoch": 0.7654026416702173, + "grad_norm": 19.375, + "learning_rate": 1.7099369530027645e-06, + "loss": 0.9115, + "step": 8982 + }, + { + "epoch": 0.7654878568385173, + "grad_norm": 13.625, + "learning_rate": 1.709780701704854e-06, + "loss": 0.7726, + "step": 8983 + }, + { + "epoch": 0.7655730720068172, + "grad_norm": 13.9375, + "learning_rate": 1.7096244420982438e-06, + "loss": 0.8881, + "step": 8984 + }, + { + "epoch": 0.7656582871751172, + "grad_norm": 9.875, + "learning_rate": 1.709468174185758e-06, + "loss": 0.3125, + "step": 8985 + }, + { + "epoch": 0.7657435023434171, + "grad_norm": 16.25, + "learning_rate": 1.70931189797022e-06, + "loss": 0.721, + "step": 8986 + }, + { + "epoch": 0.7658287175117171, + "grad_norm": 14.4375, + "learning_rate": 1.7091556134544547e-06, + "loss": 0.8042, + "step": 8987 + }, + { + "epoch": 0.7659139326800171, + "grad_norm": 17.25, + "learning_rate": 1.7089993206412848e-06, + "loss": 0.8803, + "step": 8988 + }, + { + "epoch": 0.765999147848317, + "grad_norm": 17.25, + "learning_rate": 1.7088430195335362e-06, + "loss": 0.9869, + "step": 8989 + }, + { + "epoch": 0.766084363016617, + "grad_norm": 13.9375, + "learning_rate": 1.7086867101340327e-06, + "loss": 0.6307, + "step": 8990 + }, + { + "epoch": 0.7661695781849169, + "grad_norm": 14.25, + "learning_rate": 1.708530392445599e-06, + "loss": 0.7213, + "step": 8991 + }, + { + "epoch": 0.7662547933532169, + "grad_norm": 19.75, + "learning_rate": 1.7083740664710602e-06, + "loss": 0.7864, + "step": 8992 + }, + { + "epoch": 0.7663400085215168, + "grad_norm": 12.875, + "learning_rate": 1.708217732213241e-06, + "loss": 0.5595, + "step": 8993 + }, + { + "epoch": 0.7664252236898168, + "grad_norm": 24.875, + "learning_rate": 1.7080613896749667e-06, + "loss": 0.6869, + "step": 8994 + }, + { + "epoch": 0.7665104388581168, + "grad_norm": 16.5, + "learning_rate": 1.7079050388590623e-06, + "loss": 0.8646, + "step": 8995 + }, + { + "epoch": 0.7665956540264167, + "grad_norm": 13.875, + "learning_rate": 1.7077486797683538e-06, + "loss": 0.6568, + "step": 8996 + }, + { + "epoch": 0.7666808691947167, + "grad_norm": 11.8125, + "learning_rate": 1.7075923124056657e-06, + "loss": 0.5769, + "step": 8997 + }, + { + "epoch": 0.7667660843630166, + "grad_norm": 14.0625, + "learning_rate": 1.7074359367738252e-06, + "loss": 0.6569, + "step": 8998 + }, + { + "epoch": 0.7668512995313166, + "grad_norm": 12.5625, + "learning_rate": 1.7072795528756566e-06, + "loss": 0.5376, + "step": 8999 + }, + { + "epoch": 0.7669365146996165, + "grad_norm": 12.0, + "learning_rate": 1.707123160713987e-06, + "loss": 0.5588, + "step": 9000 + }, + { + "epoch": 0.7670217298679165, + "grad_norm": 13.375, + "learning_rate": 1.7069667602916417e-06, + "loss": 0.5436, + "step": 9001 + }, + { + "epoch": 0.7671069450362165, + "grad_norm": 12.75, + "learning_rate": 1.7068103516114484e-06, + "loss": 0.6801, + "step": 9002 + }, + { + "epoch": 0.7671921602045164, + "grad_norm": 11.4375, + "learning_rate": 1.7066539346762322e-06, + "loss": 0.6156, + "step": 9003 + }, + { + "epoch": 0.7672773753728164, + "grad_norm": 17.875, + "learning_rate": 1.7064975094888202e-06, + "loss": 0.7216, + "step": 9004 + }, + { + "epoch": 0.7673625905411163, + "grad_norm": 15.3125, + "learning_rate": 1.7063410760520392e-06, + "loss": 0.8245, + "step": 9005 + }, + { + "epoch": 0.7674478057094163, + "grad_norm": 10.875, + "learning_rate": 1.706184634368716e-06, + "loss": 0.3248, + "step": 9006 + }, + { + "epoch": 0.7675330208777162, + "grad_norm": 12.75, + "learning_rate": 1.706028184441678e-06, + "loss": 0.4611, + "step": 9007 + }, + { + "epoch": 0.7676182360460162, + "grad_norm": 13.3125, + "learning_rate": 1.7058717262737517e-06, + "loss": 0.2809, + "step": 9008 + }, + { + "epoch": 0.7677034512143162, + "grad_norm": 24.125, + "learning_rate": 1.7057152598677651e-06, + "loss": 0.917, + "step": 9009 + }, + { + "epoch": 0.7677886663826161, + "grad_norm": 14.75, + "learning_rate": 1.7055587852265457e-06, + "loss": 0.7854, + "step": 9010 + }, + { + "epoch": 0.7678738815509161, + "grad_norm": 15.3125, + "learning_rate": 1.7054023023529208e-06, + "loss": 0.875, + "step": 9011 + }, + { + "epoch": 0.767959096719216, + "grad_norm": 18.75, + "learning_rate": 1.705245811249718e-06, + "loss": 0.7601, + "step": 9012 + }, + { + "epoch": 0.768044311887516, + "grad_norm": 13.6875, + "learning_rate": 1.7050893119197663e-06, + "loss": 0.4686, + "step": 9013 + }, + { + "epoch": 0.7681295270558159, + "grad_norm": 13.125, + "learning_rate": 1.7049328043658926e-06, + "loss": 0.5051, + "step": 9014 + }, + { + "epoch": 0.7682147422241159, + "grad_norm": 14.625, + "learning_rate": 1.704776288590926e-06, + "loss": 0.76, + "step": 9015 + }, + { + "epoch": 0.7682999573924159, + "grad_norm": 12.5, + "learning_rate": 1.7046197645976943e-06, + "loss": 0.3944, + "step": 9016 + }, + { + "epoch": 0.7683851725607158, + "grad_norm": 14.8125, + "learning_rate": 1.7044632323890261e-06, + "loss": 0.8577, + "step": 9017 + }, + { + "epoch": 0.7684703877290158, + "grad_norm": 23.75, + "learning_rate": 1.7043066919677503e-06, + "loss": 0.5575, + "step": 9018 + }, + { + "epoch": 0.7685556028973157, + "grad_norm": 13.75, + "learning_rate": 1.7041501433366965e-06, + "loss": 0.6031, + "step": 9019 + }, + { + "epoch": 0.7686408180656157, + "grad_norm": 13.25, + "learning_rate": 1.7039935864986923e-06, + "loss": 0.6532, + "step": 9020 + }, + { + "epoch": 0.7687260332339156, + "grad_norm": 12.6875, + "learning_rate": 1.7038370214565675e-06, + "loss": 0.6905, + "step": 9021 + }, + { + "epoch": 0.7688112484022156, + "grad_norm": 14.1875, + "learning_rate": 1.703680448213151e-06, + "loss": 0.698, + "step": 9022 + }, + { + "epoch": 0.7688964635705156, + "grad_norm": 14.8125, + "learning_rate": 1.7035238667712733e-06, + "loss": 0.548, + "step": 9023 + }, + { + "epoch": 0.7689816787388155, + "grad_norm": 11.8125, + "learning_rate": 1.703367277133763e-06, + "loss": 0.5977, + "step": 9024 + }, + { + "epoch": 0.7690668939071155, + "grad_norm": 14.375, + "learning_rate": 1.7032106793034501e-06, + "loss": 0.5938, + "step": 9025 + }, + { + "epoch": 0.7691521090754154, + "grad_norm": 12.625, + "learning_rate": 1.7030540732831644e-06, + "loss": 0.4421, + "step": 9026 + }, + { + "epoch": 0.7692373242437154, + "grad_norm": 14.375, + "learning_rate": 1.702897459075736e-06, + "loss": 0.9577, + "step": 9027 + }, + { + "epoch": 0.7693225394120153, + "grad_norm": 15.3125, + "learning_rate": 1.7027408366839953e-06, + "loss": 0.443, + "step": 9028 + }, + { + "epoch": 0.7694077545803153, + "grad_norm": 13.1875, + "learning_rate": 1.7025842061107723e-06, + "loss": 0.8565, + "step": 9029 + }, + { + "epoch": 0.7694929697486153, + "grad_norm": 15.125, + "learning_rate": 1.7024275673588975e-06, + "loss": 0.6184, + "step": 9030 + }, + { + "epoch": 0.7695781849169152, + "grad_norm": 15.0, + "learning_rate": 1.7022709204312018e-06, + "loss": 1.1135, + "step": 9031 + }, + { + "epoch": 0.7696634000852152, + "grad_norm": 14.125, + "learning_rate": 1.7021142653305156e-06, + "loss": 0.5914, + "step": 9032 + }, + { + "epoch": 0.7697486152535151, + "grad_norm": 16.25, + "learning_rate": 1.70195760205967e-06, + "loss": 0.6726, + "step": 9033 + }, + { + "epoch": 0.7698338304218151, + "grad_norm": 15.75, + "learning_rate": 1.7018009306214962e-06, + "loss": 0.8695, + "step": 9034 + }, + { + "epoch": 0.769919045590115, + "grad_norm": 13.75, + "learning_rate": 1.7016442510188251e-06, + "loss": 0.4837, + "step": 9035 + }, + { + "epoch": 0.770004260758415, + "grad_norm": 9.875, + "learning_rate": 1.7014875632544885e-06, + "loss": 0.1759, + "step": 9036 + }, + { + "epoch": 0.770089475926715, + "grad_norm": 13.75, + "learning_rate": 1.7013308673313173e-06, + "loss": 0.7079, + "step": 9037 + }, + { + "epoch": 0.7701746910950149, + "grad_norm": 17.625, + "learning_rate": 1.7011741632521436e-06, + "loss": 1.1155, + "step": 9038 + }, + { + "epoch": 0.7702599062633149, + "grad_norm": 17.0, + "learning_rate": 1.7010174510197992e-06, + "loss": 0.9273, + "step": 9039 + }, + { + "epoch": 0.7703451214316148, + "grad_norm": 14.25, + "learning_rate": 1.7008607306371155e-06, + "loss": 0.7976, + "step": 9040 + }, + { + "epoch": 0.7704303365999148, + "grad_norm": 14.0, + "learning_rate": 1.7007040021069254e-06, + "loss": 0.5912, + "step": 9041 + }, + { + "epoch": 0.7705155517682147, + "grad_norm": 16.75, + "learning_rate": 1.700547265432061e-06, + "loss": 0.5672, + "step": 9042 + }, + { + "epoch": 0.7706007669365147, + "grad_norm": 11.625, + "learning_rate": 1.700390520615354e-06, + "loss": 0.5232, + "step": 9043 + }, + { + "epoch": 0.7706859821048146, + "grad_norm": 19.0, + "learning_rate": 1.7002337676596373e-06, + "loss": 1.1698, + "step": 9044 + }, + { + "epoch": 0.7707711972731146, + "grad_norm": 18.75, + "learning_rate": 1.700077006567744e-06, + "loss": 0.8062, + "step": 9045 + }, + { + "epoch": 0.7708564124414146, + "grad_norm": 13.625, + "learning_rate": 1.6999202373425063e-06, + "loss": 0.5261, + "step": 9046 + }, + { + "epoch": 0.7709416276097145, + "grad_norm": 11.6875, + "learning_rate": 1.699763459986758e-06, + "loss": 0.5351, + "step": 9047 + }, + { + "epoch": 0.7710268427780145, + "grad_norm": 13.6875, + "learning_rate": 1.6996066745033312e-06, + "loss": 0.7916, + "step": 9048 + }, + { + "epoch": 0.7711120579463144, + "grad_norm": 16.0, + "learning_rate": 1.69944988089506e-06, + "loss": 0.8742, + "step": 9049 + }, + { + "epoch": 0.7711972731146144, + "grad_norm": 11.3125, + "learning_rate": 1.699293079164777e-06, + "loss": 0.4032, + "step": 9050 + }, + { + "epoch": 0.7712824882829143, + "grad_norm": 16.0, + "learning_rate": 1.699136269315317e-06, + "loss": 0.7968, + "step": 9051 + }, + { + "epoch": 0.7713677034512143, + "grad_norm": 11.0, + "learning_rate": 1.6989794513495125e-06, + "loss": 0.4347, + "step": 9052 + }, + { + "epoch": 0.7714529186195143, + "grad_norm": 11.875, + "learning_rate": 1.6988226252701982e-06, + "loss": 0.5018, + "step": 9053 + }, + { + "epoch": 0.7715381337878142, + "grad_norm": 10.5625, + "learning_rate": 1.698665791080207e-06, + "loss": 0.3569, + "step": 9054 + }, + { + "epoch": 0.7716233489561142, + "grad_norm": 17.75, + "learning_rate": 1.6985089487823745e-06, + "loss": 0.6398, + "step": 9055 + }, + { + "epoch": 0.7717085641244141, + "grad_norm": 16.5, + "learning_rate": 1.698352098379534e-06, + "loss": 0.7814, + "step": 9056 + }, + { + "epoch": 0.7717937792927141, + "grad_norm": 26.0, + "learning_rate": 1.6981952398745205e-06, + "loss": 1.3219, + "step": 9057 + }, + { + "epoch": 0.771878994461014, + "grad_norm": 17.875, + "learning_rate": 1.6980383732701678e-06, + "loss": 0.7834, + "step": 9058 + }, + { + "epoch": 0.771964209629314, + "grad_norm": 17.625, + "learning_rate": 1.6978814985693115e-06, + "loss": 0.7673, + "step": 9059 + }, + { + "epoch": 0.772049424797614, + "grad_norm": 24.5, + "learning_rate": 1.6977246157747862e-06, + "loss": 0.9748, + "step": 9060 + }, + { + "epoch": 0.7721346399659139, + "grad_norm": 9.9375, + "learning_rate": 1.6975677248894266e-06, + "loss": 0.4087, + "step": 9061 + }, + { + "epoch": 0.7722198551342139, + "grad_norm": 14.3125, + "learning_rate": 1.6974108259160683e-06, + "loss": 0.6819, + "step": 9062 + }, + { + "epoch": 0.7723050703025138, + "grad_norm": 27.625, + "learning_rate": 1.6972539188575465e-06, + "loss": 0.9034, + "step": 9063 + }, + { + "epoch": 0.7723902854708138, + "grad_norm": 23.25, + "learning_rate": 1.6970970037166967e-06, + "loss": 0.8258, + "step": 9064 + }, + { + "epoch": 0.7724755006391137, + "grad_norm": 18.875, + "learning_rate": 1.6969400804963542e-06, + "loss": 0.7505, + "step": 9065 + }, + { + "epoch": 0.7725607158074137, + "grad_norm": 11.4375, + "learning_rate": 1.6967831491993553e-06, + "loss": 0.5268, + "step": 9066 + }, + { + "epoch": 0.7726459309757137, + "grad_norm": 11.8125, + "learning_rate": 1.6966262098285355e-06, + "loss": 0.3593, + "step": 9067 + }, + { + "epoch": 0.7727311461440136, + "grad_norm": 16.875, + "learning_rate": 1.696469262386731e-06, + "loss": 0.776, + "step": 9068 + }, + { + "epoch": 0.7728163613123136, + "grad_norm": 17.125, + "learning_rate": 1.696312306876778e-06, + "loss": 1.0498, + "step": 9069 + }, + { + "epoch": 0.7729015764806135, + "grad_norm": 14.4375, + "learning_rate": 1.6961553433015127e-06, + "loss": 0.6401, + "step": 9070 + }, + { + "epoch": 0.7729867916489135, + "grad_norm": 11.125, + "learning_rate": 1.6959983716637717e-06, + "loss": 0.5161, + "step": 9071 + }, + { + "epoch": 0.7730720068172134, + "grad_norm": 13.125, + "learning_rate": 1.6958413919663918e-06, + "loss": 0.4393, + "step": 9072 + }, + { + "epoch": 0.7731572219855134, + "grad_norm": 13.75, + "learning_rate": 1.6956844042122094e-06, + "loss": 0.3946, + "step": 9073 + }, + { + "epoch": 0.7732424371538134, + "grad_norm": 19.25, + "learning_rate": 1.6955274084040618e-06, + "loss": 1.0158, + "step": 9074 + }, + { + "epoch": 0.7733276523221133, + "grad_norm": 15.0, + "learning_rate": 1.695370404544786e-06, + "loss": 0.787, + "step": 9075 + }, + { + "epoch": 0.7734128674904133, + "grad_norm": 17.0, + "learning_rate": 1.6952133926372188e-06, + "loss": 0.8103, + "step": 9076 + }, + { + "epoch": 0.7734980826587132, + "grad_norm": 30.0, + "learning_rate": 1.695056372684198e-06, + "loss": 0.9871, + "step": 9077 + }, + { + "epoch": 0.7735832978270132, + "grad_norm": 18.875, + "learning_rate": 1.694899344688561e-06, + "loss": 1.0234, + "step": 9078 + }, + { + "epoch": 0.7736685129953131, + "grad_norm": 29.875, + "learning_rate": 1.6947423086531458e-06, + "loss": 0.5055, + "step": 9079 + }, + { + "epoch": 0.7737537281636131, + "grad_norm": 16.25, + "learning_rate": 1.6945852645807895e-06, + "loss": 0.9303, + "step": 9080 + }, + { + "epoch": 0.773838943331913, + "grad_norm": 24.625, + "learning_rate": 1.6944282124743305e-06, + "loss": 0.8925, + "step": 9081 + }, + { + "epoch": 0.773924158500213, + "grad_norm": 24.0, + "learning_rate": 1.6942711523366069e-06, + "loss": 0.8611, + "step": 9082 + }, + { + "epoch": 0.774009373668513, + "grad_norm": 16.75, + "learning_rate": 1.694114084170457e-06, + "loss": 0.8173, + "step": 9083 + }, + { + "epoch": 0.7740945888368129, + "grad_norm": 16.25, + "learning_rate": 1.6939570079787184e-06, + "loss": 0.9825, + "step": 9084 + }, + { + "epoch": 0.7741798040051129, + "grad_norm": 14.75, + "learning_rate": 1.6937999237642308e-06, + "loss": 0.8613, + "step": 9085 + }, + { + "epoch": 0.7742650191734128, + "grad_norm": 10.0625, + "learning_rate": 1.693642831529832e-06, + "loss": 0.3232, + "step": 9086 + }, + { + "epoch": 0.7743502343417128, + "grad_norm": 20.375, + "learning_rate": 1.6934857312783613e-06, + "loss": 1.2607, + "step": 9087 + }, + { + "epoch": 0.7744354495100128, + "grad_norm": 14.125, + "learning_rate": 1.6933286230126575e-06, + "loss": 0.5303, + "step": 9088 + }, + { + "epoch": 0.7745206646783127, + "grad_norm": 12.9375, + "learning_rate": 1.6931715067355595e-06, + "loss": 0.4752, + "step": 9089 + }, + { + "epoch": 0.7746058798466127, + "grad_norm": 12.8125, + "learning_rate": 1.6930143824499068e-06, + "loss": 0.7418, + "step": 9090 + }, + { + "epoch": 0.7746910950149126, + "grad_norm": 17.5, + "learning_rate": 1.6928572501585386e-06, + "loss": 0.8885, + "step": 9091 + }, + { + "epoch": 0.7747763101832126, + "grad_norm": 13.75, + "learning_rate": 1.6927001098642948e-06, + "loss": 0.497, + "step": 9092 + }, + { + "epoch": 0.7748615253515125, + "grad_norm": 21.25, + "learning_rate": 1.6925429615700146e-06, + "loss": 1.0675, + "step": 9093 + }, + { + "epoch": 0.7749467405198125, + "grad_norm": 27.125, + "learning_rate": 1.6923858052785383e-06, + "loss": 1.0814, + "step": 9094 + }, + { + "epoch": 0.7750319556881125, + "grad_norm": 13.3125, + "learning_rate": 1.6922286409927052e-06, + "loss": 0.6301, + "step": 9095 + }, + { + "epoch": 0.7751171708564124, + "grad_norm": 13.625, + "learning_rate": 1.6920714687153567e-06, + "loss": 0.6973, + "step": 9096 + }, + { + "epoch": 0.7752023860247124, + "grad_norm": 17.625, + "learning_rate": 1.6919142884493315e-06, + "loss": 0.8307, + "step": 9097 + }, + { + "epoch": 0.7752876011930123, + "grad_norm": 15.625, + "learning_rate": 1.6917571001974712e-06, + "loss": 0.5827, + "step": 9098 + }, + { + "epoch": 0.7753728163613123, + "grad_norm": 18.5, + "learning_rate": 1.6915999039626153e-06, + "loss": 0.8677, + "step": 9099 + }, + { + "epoch": 0.7754580315296122, + "grad_norm": 14.125, + "learning_rate": 1.6914426997476057e-06, + "loss": 0.968, + "step": 9100 + }, + { + "epoch": 0.7755432466979122, + "grad_norm": 16.625, + "learning_rate": 1.6912854875552822e-06, + "loss": 0.6724, + "step": 9101 + }, + { + "epoch": 0.7756284618662121, + "grad_norm": 11.5, + "learning_rate": 1.6911282673884863e-06, + "loss": 0.521, + "step": 9102 + }, + { + "epoch": 0.7757136770345121, + "grad_norm": 14.9375, + "learning_rate": 1.6909710392500588e-06, + "loss": 0.7789, + "step": 9103 + }, + { + "epoch": 0.7757988922028121, + "grad_norm": 10.6875, + "learning_rate": 1.690813803142842e-06, + "loss": 0.3545, + "step": 9104 + }, + { + "epoch": 0.775884107371112, + "grad_norm": 16.625, + "learning_rate": 1.6906565590696755e-06, + "loss": 0.9103, + "step": 9105 + }, + { + "epoch": 0.775969322539412, + "grad_norm": 14.5625, + "learning_rate": 1.6904993070334024e-06, + "loss": 0.5881, + "step": 9106 + }, + { + "epoch": 0.7760545377077119, + "grad_norm": 15.875, + "learning_rate": 1.6903420470368636e-06, + "loss": 0.8179, + "step": 9107 + }, + { + "epoch": 0.7761397528760119, + "grad_norm": 9.5625, + "learning_rate": 1.6901847790829015e-06, + "loss": 0.2609, + "step": 9108 + }, + { + "epoch": 0.7762249680443118, + "grad_norm": 11.3125, + "learning_rate": 1.690027503174358e-06, + "loss": 0.4712, + "step": 9109 + }, + { + "epoch": 0.7763101832126118, + "grad_norm": 14.125, + "learning_rate": 1.6898702193140743e-06, + "loss": 0.7857, + "step": 9110 + }, + { + "epoch": 0.7763953983809118, + "grad_norm": 19.0, + "learning_rate": 1.689712927504894e-06, + "loss": 0.7153, + "step": 9111 + }, + { + "epoch": 0.7764806135492117, + "grad_norm": 14.5, + "learning_rate": 1.6895556277496588e-06, + "loss": 0.6683, + "step": 9112 + }, + { + "epoch": 0.7765658287175117, + "grad_norm": 17.5, + "learning_rate": 1.6893983200512115e-06, + "loss": 1.2218, + "step": 9113 + }, + { + "epoch": 0.7766510438858116, + "grad_norm": 14.375, + "learning_rate": 1.6892410044123942e-06, + "loss": 0.3115, + "step": 9114 + }, + { + "epoch": 0.7767362590541116, + "grad_norm": 18.125, + "learning_rate": 1.6890836808360509e-06, + "loss": 0.652, + "step": 9115 + }, + { + "epoch": 0.7768214742224115, + "grad_norm": 12.4375, + "learning_rate": 1.6889263493250235e-06, + "loss": 0.5894, + "step": 9116 + }, + { + "epoch": 0.7769066893907115, + "grad_norm": 13.75, + "learning_rate": 1.688769009882156e-06, + "loss": 0.7257, + "step": 9117 + }, + { + "epoch": 0.7769919045590115, + "grad_norm": 14.0625, + "learning_rate": 1.6886116625102907e-06, + "loss": 1.2324, + "step": 9118 + }, + { + "epoch": 0.7770771197273114, + "grad_norm": 14.25, + "learning_rate": 1.6884543072122721e-06, + "loss": 0.8572, + "step": 9119 + }, + { + "epoch": 0.7771623348956114, + "grad_norm": 13.375, + "learning_rate": 1.6882969439909432e-06, + "loss": 0.7263, + "step": 9120 + }, + { + "epoch": 0.7772475500639113, + "grad_norm": 15.5, + "learning_rate": 1.688139572849148e-06, + "loss": 0.6301, + "step": 9121 + }, + { + "epoch": 0.7773327652322113, + "grad_norm": 17.0, + "learning_rate": 1.6879821937897295e-06, + "loss": 0.6442, + "step": 9122 + }, + { + "epoch": 0.7774179804005112, + "grad_norm": 12.9375, + "learning_rate": 1.6878248068155328e-06, + "loss": 0.4774, + "step": 9123 + }, + { + "epoch": 0.7775031955688112, + "grad_norm": 14.4375, + "learning_rate": 1.6876674119294014e-06, + "loss": 0.9256, + "step": 9124 + }, + { + "epoch": 0.7775884107371112, + "grad_norm": 17.875, + "learning_rate": 1.6875100091341795e-06, + "loss": 0.9333, + "step": 9125 + }, + { + "epoch": 0.7776736259054111, + "grad_norm": 14.9375, + "learning_rate": 1.6873525984327122e-06, + "loss": 0.6986, + "step": 9126 + }, + { + "epoch": 0.7777588410737111, + "grad_norm": 12.25, + "learning_rate": 1.6871951798278435e-06, + "loss": 0.7764, + "step": 9127 + }, + { + "epoch": 0.777844056242011, + "grad_norm": 14.0625, + "learning_rate": 1.6870377533224186e-06, + "loss": 0.6542, + "step": 9128 + }, + { + "epoch": 0.777929271410311, + "grad_norm": 11.25, + "learning_rate": 1.6868803189192815e-06, + "loss": 0.4523, + "step": 9129 + }, + { + "epoch": 0.7780144865786109, + "grad_norm": 18.875, + "learning_rate": 1.686722876621278e-06, + "loss": 0.6665, + "step": 9130 + }, + { + "epoch": 0.7780997017469109, + "grad_norm": 14.4375, + "learning_rate": 1.6865654264312529e-06, + "loss": 0.4918, + "step": 9131 + }, + { + "epoch": 0.7781849169152109, + "grad_norm": 12.875, + "learning_rate": 1.6864079683520518e-06, + "loss": 0.5856, + "step": 9132 + }, + { + "epoch": 0.7782701320835108, + "grad_norm": 15.3125, + "learning_rate": 1.6862505023865194e-06, + "loss": 1.0167, + "step": 9133 + }, + { + "epoch": 0.7783553472518108, + "grad_norm": 10.875, + "learning_rate": 1.686093028537502e-06, + "loss": 0.3041, + "step": 9134 + }, + { + "epoch": 0.7784405624201107, + "grad_norm": 11.75, + "learning_rate": 1.6859355468078448e-06, + "loss": 0.5819, + "step": 9135 + }, + { + "epoch": 0.7785257775884107, + "grad_norm": 11.625, + "learning_rate": 1.6857780572003946e-06, + "loss": 0.4546, + "step": 9136 + }, + { + "epoch": 0.7786109927567106, + "grad_norm": 15.6875, + "learning_rate": 1.6856205597179959e-06, + "loss": 0.8451, + "step": 9137 + }, + { + "epoch": 0.7786962079250106, + "grad_norm": 14.3125, + "learning_rate": 1.685463054363496e-06, + "loss": 0.8406, + "step": 9138 + }, + { + "epoch": 0.7787814230933106, + "grad_norm": 16.875, + "learning_rate": 1.6853055411397406e-06, + "loss": 1.0408, + "step": 9139 + }, + { + "epoch": 0.7788666382616106, + "grad_norm": 15.875, + "learning_rate": 1.6851480200495765e-06, + "loss": 0.721, + "step": 9140 + }, + { + "epoch": 0.7789518534299106, + "grad_norm": 14.375, + "learning_rate": 1.6849904910958503e-06, + "loss": 0.603, + "step": 9141 + }, + { + "epoch": 0.7790370685982105, + "grad_norm": 11.6875, + "learning_rate": 1.6848329542814081e-06, + "loss": 0.4813, + "step": 9142 + }, + { + "epoch": 0.7791222837665105, + "grad_norm": 13.5, + "learning_rate": 1.6846754096090974e-06, + "loss": 0.7694, + "step": 9143 + }, + { + "epoch": 0.7792074989348104, + "grad_norm": 16.0, + "learning_rate": 1.684517857081765e-06, + "loss": 1.0108, + "step": 9144 + }, + { + "epoch": 0.7792927141031104, + "grad_norm": 11.125, + "learning_rate": 1.684360296702258e-06, + "loss": 0.5011, + "step": 9145 + }, + { + "epoch": 0.7793779292714104, + "grad_norm": 16.25, + "learning_rate": 1.6842027284734233e-06, + "loss": 0.801, + "step": 9146 + }, + { + "epoch": 0.7794631444397103, + "grad_norm": 18.25, + "learning_rate": 1.6840451523981089e-06, + "loss": 0.6285, + "step": 9147 + }, + { + "epoch": 0.7795483596080103, + "grad_norm": 12.9375, + "learning_rate": 1.6838875684791622e-06, + "loss": 0.5928, + "step": 9148 + }, + { + "epoch": 0.7796335747763102, + "grad_norm": 14.5625, + "learning_rate": 1.683729976719431e-06, + "loss": 0.7728, + "step": 9149 + }, + { + "epoch": 0.7797187899446102, + "grad_norm": 14.5, + "learning_rate": 1.6835723771217624e-06, + "loss": 0.4603, + "step": 9150 + }, + { + "epoch": 0.7798040051129101, + "grad_norm": 16.0, + "learning_rate": 1.6834147696890056e-06, + "loss": 0.6017, + "step": 9151 + }, + { + "epoch": 0.7798892202812101, + "grad_norm": 16.25, + "learning_rate": 1.6832571544240076e-06, + "loss": 1.0502, + "step": 9152 + }, + { + "epoch": 0.7799744354495101, + "grad_norm": 17.25, + "learning_rate": 1.6830995313296175e-06, + "loss": 0.7913, + "step": 9153 + }, + { + "epoch": 0.78005965061781, + "grad_norm": 12.9375, + "learning_rate": 1.682941900408683e-06, + "loss": 0.383, + "step": 9154 + }, + { + "epoch": 0.78014486578611, + "grad_norm": 14.8125, + "learning_rate": 1.6827842616640533e-06, + "loss": 0.6808, + "step": 9155 + }, + { + "epoch": 0.7802300809544099, + "grad_norm": 17.75, + "learning_rate": 1.6826266150985765e-06, + "loss": 0.631, + "step": 9156 + }, + { + "epoch": 0.7803152961227099, + "grad_norm": 12.5625, + "learning_rate": 1.6824689607151023e-06, + "loss": 0.8862, + "step": 9157 + }, + { + "epoch": 0.7804005112910098, + "grad_norm": 15.375, + "learning_rate": 1.6823112985164786e-06, + "loss": 0.7749, + "step": 9158 + }, + { + "epoch": 0.7804857264593098, + "grad_norm": 16.75, + "learning_rate": 1.682153628505555e-06, + "loss": 0.9177, + "step": 9159 + }, + { + "epoch": 0.7805709416276098, + "grad_norm": 16.125, + "learning_rate": 1.6819959506851807e-06, + "loss": 0.5646, + "step": 9160 + }, + { + "epoch": 0.7806561567959097, + "grad_norm": 15.25, + "learning_rate": 1.6818382650582055e-06, + "loss": 0.9146, + "step": 9161 + }, + { + "epoch": 0.7807413719642097, + "grad_norm": 13.75, + "learning_rate": 1.6816805716274787e-06, + "loss": 0.6513, + "step": 9162 + }, + { + "epoch": 0.7808265871325096, + "grad_norm": 11.125, + "learning_rate": 1.681522870395849e-06, + "loss": 0.5193, + "step": 9163 + }, + { + "epoch": 0.7809118023008096, + "grad_norm": 13.8125, + "learning_rate": 1.681365161366168e-06, + "loss": 0.7052, + "step": 9164 + }, + { + "epoch": 0.7809970174691095, + "grad_norm": 14.0, + "learning_rate": 1.6812074445412845e-06, + "loss": 0.4651, + "step": 9165 + }, + { + "epoch": 0.7810822326374095, + "grad_norm": 15.3125, + "learning_rate": 1.6810497199240491e-06, + "loss": 0.5505, + "step": 9166 + }, + { + "epoch": 0.7811674478057095, + "grad_norm": 16.25, + "learning_rate": 1.6808919875173113e-06, + "loss": 0.9399, + "step": 9167 + }, + { + "epoch": 0.7812526629740094, + "grad_norm": 19.875, + "learning_rate": 1.6807342473239224e-06, + "loss": 0.9514, + "step": 9168 + }, + { + "epoch": 0.7813378781423094, + "grad_norm": 15.0625, + "learning_rate": 1.6805764993467326e-06, + "loss": 0.6986, + "step": 9169 + }, + { + "epoch": 0.7814230933106093, + "grad_norm": 13.8125, + "learning_rate": 1.6804187435885921e-06, + "loss": 0.6275, + "step": 9170 + }, + { + "epoch": 0.7815083084789093, + "grad_norm": 14.375, + "learning_rate": 1.6802609800523523e-06, + "loss": 0.8237, + "step": 9171 + }, + { + "epoch": 0.7815935236472092, + "grad_norm": 19.0, + "learning_rate": 1.680103208740864e-06, + "loss": 0.9538, + "step": 9172 + }, + { + "epoch": 0.7816787388155092, + "grad_norm": 14.125, + "learning_rate": 1.6799454296569778e-06, + "loss": 0.7257, + "step": 9173 + }, + { + "epoch": 0.7817639539838092, + "grad_norm": 14.1875, + "learning_rate": 1.679787642803546e-06, + "loss": 0.8362, + "step": 9174 + }, + { + "epoch": 0.7818491691521091, + "grad_norm": 14.25, + "learning_rate": 1.6796298481834187e-06, + "loss": 0.9333, + "step": 9175 + }, + { + "epoch": 0.7819343843204091, + "grad_norm": 15.875, + "learning_rate": 1.6794720457994484e-06, + "loss": 0.8981, + "step": 9176 + }, + { + "epoch": 0.782019599488709, + "grad_norm": 12.0625, + "learning_rate": 1.679314235654486e-06, + "loss": 0.5872, + "step": 9177 + }, + { + "epoch": 0.782104814657009, + "grad_norm": 15.125, + "learning_rate": 1.6791564177513838e-06, + "loss": 0.7796, + "step": 9178 + }, + { + "epoch": 0.7821900298253089, + "grad_norm": 14.8125, + "learning_rate": 1.6789985920929937e-06, + "loss": 1.0153, + "step": 9179 + }, + { + "epoch": 0.7822752449936089, + "grad_norm": 13.9375, + "learning_rate": 1.6788407586821675e-06, + "loss": 0.58, + "step": 9180 + }, + { + "epoch": 0.7823604601619089, + "grad_norm": 11.8125, + "learning_rate": 1.6786829175217582e-06, + "loss": 0.4868, + "step": 9181 + }, + { + "epoch": 0.7824456753302088, + "grad_norm": 18.5, + "learning_rate": 1.6785250686146164e-06, + "loss": 0.8926, + "step": 9182 + }, + { + "epoch": 0.7825308904985088, + "grad_norm": 13.8125, + "learning_rate": 1.6783672119635966e-06, + "loss": 0.6253, + "step": 9183 + }, + { + "epoch": 0.7826161056668087, + "grad_norm": 16.25, + "learning_rate": 1.6782093475715499e-06, + "loss": 0.8603, + "step": 9184 + }, + { + "epoch": 0.7827013208351087, + "grad_norm": 16.25, + "learning_rate": 1.6780514754413306e-06, + "loss": 0.6022, + "step": 9185 + }, + { + "epoch": 0.7827865360034086, + "grad_norm": 16.375, + "learning_rate": 1.6778935955757897e-06, + "loss": 0.7856, + "step": 9186 + }, + { + "epoch": 0.7828717511717086, + "grad_norm": 13.1875, + "learning_rate": 1.6777357079777817e-06, + "loss": 0.552, + "step": 9187 + }, + { + "epoch": 0.7829569663400086, + "grad_norm": 17.375, + "learning_rate": 1.677577812650159e-06, + "loss": 0.562, + "step": 9188 + }, + { + "epoch": 0.7830421815083085, + "grad_norm": 13.0, + "learning_rate": 1.677419909595776e-06, + "loss": 0.5429, + "step": 9189 + }, + { + "epoch": 0.7831273966766085, + "grad_norm": 14.5625, + "learning_rate": 1.6772619988174846e-06, + "loss": 0.7473, + "step": 9190 + }, + { + "epoch": 0.7832126118449084, + "grad_norm": 16.125, + "learning_rate": 1.6771040803181398e-06, + "loss": 0.8336, + "step": 9191 + }, + { + "epoch": 0.7832978270132084, + "grad_norm": 17.5, + "learning_rate": 1.6769461541005944e-06, + "loss": 0.8144, + "step": 9192 + }, + { + "epoch": 0.7833830421815083, + "grad_norm": 13.4375, + "learning_rate": 1.6767882201677028e-06, + "loss": 0.5138, + "step": 9193 + }, + { + "epoch": 0.7834682573498083, + "grad_norm": 14.75, + "learning_rate": 1.676630278522319e-06, + "loss": 0.8737, + "step": 9194 + }, + { + "epoch": 0.7835534725181083, + "grad_norm": 14.8125, + "learning_rate": 1.676472329167297e-06, + "loss": 0.8462, + "step": 9195 + }, + { + "epoch": 0.7836386876864082, + "grad_norm": 14.25, + "learning_rate": 1.6763143721054912e-06, + "loss": 0.8616, + "step": 9196 + }, + { + "epoch": 0.7837239028547082, + "grad_norm": 12.3125, + "learning_rate": 1.676156407339756e-06, + "loss": 0.6099, + "step": 9197 + }, + { + "epoch": 0.7838091180230081, + "grad_norm": 14.1875, + "learning_rate": 1.6759984348729458e-06, + "loss": 0.9848, + "step": 9198 + }, + { + "epoch": 0.7838943331913081, + "grad_norm": 17.625, + "learning_rate": 1.6758404547079155e-06, + "loss": 0.7159, + "step": 9199 + }, + { + "epoch": 0.783979548359608, + "grad_norm": 14.875, + "learning_rate": 1.6756824668475202e-06, + "loss": 0.8403, + "step": 9200 + }, + { + "epoch": 0.784064763527908, + "grad_norm": 13.0, + "learning_rate": 1.6755244712946148e-06, + "loss": 0.7129, + "step": 9201 + }, + { + "epoch": 0.784149978696208, + "grad_norm": 15.125, + "learning_rate": 1.675366468052054e-06, + "loss": 0.5531, + "step": 9202 + }, + { + "epoch": 0.7842351938645079, + "grad_norm": 12.75, + "learning_rate": 1.6752084571226935e-06, + "loss": 0.6888, + "step": 9203 + }, + { + "epoch": 0.7843204090328079, + "grad_norm": 16.25, + "learning_rate": 1.6750504385093885e-06, + "loss": 0.8672, + "step": 9204 + }, + { + "epoch": 0.7844056242011078, + "grad_norm": 12.0, + "learning_rate": 1.6748924122149948e-06, + "loss": 0.5386, + "step": 9205 + }, + { + "epoch": 0.7844908393694078, + "grad_norm": 24.125, + "learning_rate": 1.6747343782423681e-06, + "loss": 0.9947, + "step": 9206 + }, + { + "epoch": 0.7845760545377077, + "grad_norm": 18.0, + "learning_rate": 1.6745763365943634e-06, + "loss": 0.7725, + "step": 9207 + }, + { + "epoch": 0.7846612697060077, + "grad_norm": 12.0625, + "learning_rate": 1.6744182872738379e-06, + "loss": 0.3866, + "step": 9208 + }, + { + "epoch": 0.7847464848743076, + "grad_norm": 13.8125, + "learning_rate": 1.6742602302836471e-06, + "loss": 0.7012, + "step": 9209 + }, + { + "epoch": 0.7848317000426076, + "grad_norm": 60.25, + "learning_rate": 1.6741021656266474e-06, + "loss": 0.9639, + "step": 9210 + }, + { + "epoch": 0.7849169152109076, + "grad_norm": 17.0, + "learning_rate": 1.673944093305695e-06, + "loss": 0.9644, + "step": 9211 + }, + { + "epoch": 0.7850021303792075, + "grad_norm": 18.5, + "learning_rate": 1.6737860133236466e-06, + "loss": 0.8649, + "step": 9212 + }, + { + "epoch": 0.7850873455475075, + "grad_norm": 17.5, + "learning_rate": 1.6736279256833588e-06, + "loss": 0.6805, + "step": 9213 + }, + { + "epoch": 0.7851725607158074, + "grad_norm": 15.625, + "learning_rate": 1.673469830387688e-06, + "loss": 1.0532, + "step": 9214 + }, + { + "epoch": 0.7852577758841074, + "grad_norm": 31.125, + "learning_rate": 1.673311727439492e-06, + "loss": 0.4484, + "step": 9215 + }, + { + "epoch": 0.7853429910524073, + "grad_norm": 14.125, + "learning_rate": 1.6731536168416268e-06, + "loss": 0.497, + "step": 9216 + }, + { + "epoch": 0.7854282062207073, + "grad_norm": 20.0, + "learning_rate": 1.6729954985969512e-06, + "loss": 1.251, + "step": 9217 + }, + { + "epoch": 0.7855134213890073, + "grad_norm": 14.125, + "learning_rate": 1.6728373727083209e-06, + "loss": 0.6152, + "step": 9218 + }, + { + "epoch": 0.7855986365573072, + "grad_norm": 11.5, + "learning_rate": 1.6726792391785942e-06, + "loss": 0.4704, + "step": 9219 + }, + { + "epoch": 0.7856838517256072, + "grad_norm": 16.75, + "learning_rate": 1.6725210980106285e-06, + "loss": 0.8961, + "step": 9220 + }, + { + "epoch": 0.7857690668939071, + "grad_norm": 15.0, + "learning_rate": 1.672362949207282e-06, + "loss": 0.497, + "step": 9221 + }, + { + "epoch": 0.7858542820622071, + "grad_norm": 16.75, + "learning_rate": 1.6722047927714123e-06, + "loss": 1.0312, + "step": 9222 + }, + { + "epoch": 0.785939497230507, + "grad_norm": 21.375, + "learning_rate": 1.672046628705877e-06, + "loss": 1.1299, + "step": 9223 + }, + { + "epoch": 0.786024712398807, + "grad_norm": 21.375, + "learning_rate": 1.671888457013535e-06, + "loss": 0.9332, + "step": 9224 + }, + { + "epoch": 0.786109927567107, + "grad_norm": 13.375, + "learning_rate": 1.6717302776972443e-06, + "loss": 0.4465, + "step": 9225 + }, + { + "epoch": 0.7861951427354069, + "grad_norm": 14.625, + "learning_rate": 1.6715720907598635e-06, + "loss": 0.6631, + "step": 9226 + }, + { + "epoch": 0.7862803579037069, + "grad_norm": 13.25, + "learning_rate": 1.6714138962042512e-06, + "loss": 0.6299, + "step": 9227 + }, + { + "epoch": 0.7863655730720068, + "grad_norm": 21.125, + "learning_rate": 1.6712556940332655e-06, + "loss": 0.7696, + "step": 9228 + }, + { + "epoch": 0.7864507882403068, + "grad_norm": 14.9375, + "learning_rate": 1.6710974842497663e-06, + "loss": 0.9325, + "step": 9229 + }, + { + "epoch": 0.7865360034086067, + "grad_norm": 11.8125, + "learning_rate": 1.670939266856612e-06, + "loss": 0.4064, + "step": 9230 + }, + { + "epoch": 0.7866212185769067, + "grad_norm": 16.625, + "learning_rate": 1.6707810418566617e-06, + "loss": 0.6951, + "step": 9231 + }, + { + "epoch": 0.7867064337452067, + "grad_norm": 28.875, + "learning_rate": 1.670622809252775e-06, + "loss": 1.1272, + "step": 9232 + }, + { + "epoch": 0.7867916489135066, + "grad_norm": 13.0625, + "learning_rate": 1.6704645690478108e-06, + "loss": 0.8552, + "step": 9233 + }, + { + "epoch": 0.7868768640818066, + "grad_norm": 15.4375, + "learning_rate": 1.6703063212446298e-06, + "loss": 0.8898, + "step": 9234 + }, + { + "epoch": 0.7869620792501065, + "grad_norm": 11.8125, + "learning_rate": 1.6701480658460903e-06, + "loss": 0.4441, + "step": 9235 + }, + { + "epoch": 0.7870472944184065, + "grad_norm": 15.125, + "learning_rate": 1.6699898028550528e-06, + "loss": 0.5961, + "step": 9236 + }, + { + "epoch": 0.7871325095867064, + "grad_norm": 14.125, + "learning_rate": 1.6698315322743774e-06, + "loss": 0.6096, + "step": 9237 + }, + { + "epoch": 0.7872177247550064, + "grad_norm": 13.5, + "learning_rate": 1.6696732541069244e-06, + "loss": 0.794, + "step": 9238 + }, + { + "epoch": 0.7873029399233064, + "grad_norm": 15.1875, + "learning_rate": 1.6695149683555528e-06, + "loss": 0.8106, + "step": 9239 + }, + { + "epoch": 0.7873881550916063, + "grad_norm": 12.5, + "learning_rate": 1.6693566750231246e-06, + "loss": 0.8082, + "step": 9240 + }, + { + "epoch": 0.7874733702599063, + "grad_norm": 11.5, + "learning_rate": 1.6691983741124992e-06, + "loss": 0.5338, + "step": 9241 + }, + { + "epoch": 0.7875585854282062, + "grad_norm": 16.0, + "learning_rate": 1.6690400656265382e-06, + "loss": 0.7448, + "step": 9242 + }, + { + "epoch": 0.7876438005965062, + "grad_norm": 20.5, + "learning_rate": 1.6688817495681012e-06, + "loss": 0.8265, + "step": 9243 + }, + { + "epoch": 0.7877290157648061, + "grad_norm": 16.125, + "learning_rate": 1.66872342594005e-06, + "loss": 0.8309, + "step": 9244 + }, + { + "epoch": 0.7878142309331061, + "grad_norm": 14.625, + "learning_rate": 1.6685650947452452e-06, + "loss": 0.7554, + "step": 9245 + }, + { + "epoch": 0.787899446101406, + "grad_norm": 21.75, + "learning_rate": 1.6684067559865486e-06, + "loss": 1.0237, + "step": 9246 + }, + { + "epoch": 0.787984661269706, + "grad_norm": 11.4375, + "learning_rate": 1.668248409666821e-06, + "loss": 0.4739, + "step": 9247 + }, + { + "epoch": 0.788069876438006, + "grad_norm": 18.0, + "learning_rate": 1.668090055788924e-06, + "loss": 0.8306, + "step": 9248 + }, + { + "epoch": 0.7881550916063059, + "grad_norm": 18.25, + "learning_rate": 1.6679316943557195e-06, + "loss": 0.801, + "step": 9249 + }, + { + "epoch": 0.7882403067746059, + "grad_norm": 12.5625, + "learning_rate": 1.667773325370069e-06, + "loss": 0.8462, + "step": 9250 + }, + { + "epoch": 0.7883255219429058, + "grad_norm": 35.5, + "learning_rate": 1.6676149488348341e-06, + "loss": 1.1651, + "step": 9251 + }, + { + "epoch": 0.7884107371112058, + "grad_norm": 17.25, + "learning_rate": 1.6674565647528773e-06, + "loss": 0.862, + "step": 9252 + }, + { + "epoch": 0.7884959522795058, + "grad_norm": 13.9375, + "learning_rate": 1.6672981731270608e-06, + "loss": 0.4958, + "step": 9253 + }, + { + "epoch": 0.7885811674478057, + "grad_norm": 15.6875, + "learning_rate": 1.6671397739602464e-06, + "loss": 0.927, + "step": 9254 + }, + { + "epoch": 0.7886663826161057, + "grad_norm": 29.0, + "learning_rate": 1.666981367255297e-06, + "loss": 0.8799, + "step": 9255 + }, + { + "epoch": 0.7887515977844056, + "grad_norm": 16.625, + "learning_rate": 1.6668229530150749e-06, + "loss": 0.6446, + "step": 9256 + }, + { + "epoch": 0.7888368129527056, + "grad_norm": 19.75, + "learning_rate": 1.666664531242443e-06, + "loss": 0.8819, + "step": 9257 + }, + { + "epoch": 0.7889220281210055, + "grad_norm": 11.6875, + "learning_rate": 1.666506101940264e-06, + "loss": 0.2891, + "step": 9258 + }, + { + "epoch": 0.7890072432893055, + "grad_norm": 10.1875, + "learning_rate": 1.6663476651114008e-06, + "loss": 0.4347, + "step": 9259 + }, + { + "epoch": 0.7890924584576055, + "grad_norm": 2112.0, + "learning_rate": 1.6661892207587165e-06, + "loss": 0.4986, + "step": 9260 + }, + { + "epoch": 0.7891776736259054, + "grad_norm": 14.4375, + "learning_rate": 1.6660307688850747e-06, + "loss": 0.7337, + "step": 9261 + }, + { + "epoch": 0.7892628887942054, + "grad_norm": 16.5, + "learning_rate": 1.6658723094933385e-06, + "loss": 0.7602, + "step": 9262 + }, + { + "epoch": 0.7893481039625053, + "grad_norm": 15.75, + "learning_rate": 1.665713842586371e-06, + "loss": 0.9238, + "step": 9263 + }, + { + "epoch": 0.7894333191308053, + "grad_norm": 17.75, + "learning_rate": 1.6655553681670373e-06, + "loss": 0.8818, + "step": 9264 + }, + { + "epoch": 0.7895185342991052, + "grad_norm": 12.875, + "learning_rate": 1.6653968862381998e-06, + "loss": 0.6838, + "step": 9265 + }, + { + "epoch": 0.7896037494674052, + "grad_norm": 15.5625, + "learning_rate": 1.6652383968027224e-06, + "loss": 0.9095, + "step": 9266 + }, + { + "epoch": 0.7896889646357051, + "grad_norm": 15.1875, + "learning_rate": 1.66507989986347e-06, + "loss": 1.1284, + "step": 9267 + }, + { + "epoch": 0.7897741798040051, + "grad_norm": 25.875, + "learning_rate": 1.6649213954233066e-06, + "loss": 0.8976, + "step": 9268 + }, + { + "epoch": 0.7898593949723051, + "grad_norm": 15.5, + "learning_rate": 1.664762883485096e-06, + "loss": 0.5452, + "step": 9269 + }, + { + "epoch": 0.789944610140605, + "grad_norm": 11.8125, + "learning_rate": 1.6646043640517034e-06, + "loss": 0.4223, + "step": 9270 + }, + { + "epoch": 0.790029825308905, + "grad_norm": 15.125, + "learning_rate": 1.6644458371259926e-06, + "loss": 0.6444, + "step": 9271 + }, + { + "epoch": 0.7901150404772049, + "grad_norm": 19.75, + "learning_rate": 1.6642873027108292e-06, + "loss": 1.3016, + "step": 9272 + }, + { + "epoch": 0.7902002556455049, + "grad_norm": 17.75, + "learning_rate": 1.664128760809077e-06, + "loss": 0.6226, + "step": 9273 + }, + { + "epoch": 0.7902854708138048, + "grad_norm": 16.0, + "learning_rate": 1.6639702114236024e-06, + "loss": 0.859, + "step": 9274 + }, + { + "epoch": 0.7903706859821048, + "grad_norm": 25.625, + "learning_rate": 1.663811654557269e-06, + "loss": 0.9279, + "step": 9275 + }, + { + "epoch": 0.7904559011504048, + "grad_norm": 14.625, + "learning_rate": 1.6636530902129436e-06, + "loss": 0.8733, + "step": 9276 + }, + { + "epoch": 0.7905411163187047, + "grad_norm": 15.5, + "learning_rate": 1.6634945183934905e-06, + "loss": 0.8732, + "step": 9277 + }, + { + "epoch": 0.7906263314870047, + "grad_norm": 17.375, + "learning_rate": 1.663335939101776e-06, + "loss": 0.9286, + "step": 9278 + }, + { + "epoch": 0.7907115466553046, + "grad_norm": 13.75, + "learning_rate": 1.6631773523406652e-06, + "loss": 0.6232, + "step": 9279 + }, + { + "epoch": 0.7907967618236046, + "grad_norm": 15.625, + "learning_rate": 1.6630187581130241e-06, + "loss": 0.8644, + "step": 9280 + }, + { + "epoch": 0.7908819769919045, + "grad_norm": 15.5, + "learning_rate": 1.6628601564217187e-06, + "loss": 0.5991, + "step": 9281 + }, + { + "epoch": 0.7909671921602045, + "grad_norm": 14.125, + "learning_rate": 1.6627015472696154e-06, + "loss": 0.6777, + "step": 9282 + }, + { + "epoch": 0.7910524073285045, + "grad_norm": 12.0, + "learning_rate": 1.66254293065958e-06, + "loss": 0.476, + "step": 9283 + }, + { + "epoch": 0.7911376224968044, + "grad_norm": 17.125, + "learning_rate": 1.6623843065944786e-06, + "loss": 0.8345, + "step": 9284 + }, + { + "epoch": 0.7912228376651044, + "grad_norm": 14.8125, + "learning_rate": 1.6622256750771787e-06, + "loss": 0.9714, + "step": 9285 + }, + { + "epoch": 0.7913080528334043, + "grad_norm": 20.25, + "learning_rate": 1.6620670361105458e-06, + "loss": 0.796, + "step": 9286 + }, + { + "epoch": 0.7913932680017043, + "grad_norm": 14.875, + "learning_rate": 1.6619083896974477e-06, + "loss": 0.8358, + "step": 9287 + }, + { + "epoch": 0.7914784831700042, + "grad_norm": 15.375, + "learning_rate": 1.6617497358407505e-06, + "loss": 0.6862, + "step": 9288 + }, + { + "epoch": 0.7915636983383042, + "grad_norm": 14.75, + "learning_rate": 1.6615910745433218e-06, + "loss": 0.7191, + "step": 9289 + }, + { + "epoch": 0.7916489135066042, + "grad_norm": 13.25, + "learning_rate": 1.6614324058080277e-06, + "loss": 0.853, + "step": 9290 + }, + { + "epoch": 0.7917341286749041, + "grad_norm": 13.625, + "learning_rate": 1.6612737296377374e-06, + "loss": 0.6146, + "step": 9291 + }, + { + "epoch": 0.7918193438432041, + "grad_norm": 18.375, + "learning_rate": 1.6611150460353164e-06, + "loss": 0.9618, + "step": 9292 + }, + { + "epoch": 0.791904559011504, + "grad_norm": 14.4375, + "learning_rate": 1.6609563550036336e-06, + "loss": 0.5117, + "step": 9293 + }, + { + "epoch": 0.791989774179804, + "grad_norm": 11.25, + "learning_rate": 1.6607976565455557e-06, + "loss": 0.5, + "step": 9294 + }, + { + "epoch": 0.7920749893481039, + "grad_norm": 10.125, + "learning_rate": 1.6606389506639517e-06, + "loss": 0.2575, + "step": 9295 + }, + { + "epoch": 0.7921602045164039, + "grad_norm": 13.1875, + "learning_rate": 1.6604802373616888e-06, + "loss": 0.6682, + "step": 9296 + }, + { + "epoch": 0.7922454196847039, + "grad_norm": 13.25, + "learning_rate": 1.660321516641635e-06, + "loss": 0.8294, + "step": 9297 + }, + { + "epoch": 0.7923306348530038, + "grad_norm": 12.0625, + "learning_rate": 1.6601627885066588e-06, + "loss": 0.4739, + "step": 9298 + }, + { + "epoch": 0.7924158500213038, + "grad_norm": 10.9375, + "learning_rate": 1.6600040529596287e-06, + "loss": 0.4657, + "step": 9299 + }, + { + "epoch": 0.7925010651896037, + "grad_norm": 10.875, + "learning_rate": 1.659845310003413e-06, + "loss": 0.298, + "step": 9300 + }, + { + "epoch": 0.7925862803579037, + "grad_norm": 12.625, + "learning_rate": 1.6596865596408802e-06, + "loss": 0.8404, + "step": 9301 + }, + { + "epoch": 0.7926714955262036, + "grad_norm": 15.4375, + "learning_rate": 1.6595278018748994e-06, + "loss": 0.3716, + "step": 9302 + }, + { + "epoch": 0.7927567106945036, + "grad_norm": 13.125, + "learning_rate": 1.6593690367083398e-06, + "loss": 0.6689, + "step": 9303 + }, + { + "epoch": 0.7928419258628036, + "grad_norm": 14.1875, + "learning_rate": 1.65921026414407e-06, + "loss": 0.8589, + "step": 9304 + }, + { + "epoch": 0.7929271410311035, + "grad_norm": 17.0, + "learning_rate": 1.6590514841849587e-06, + "loss": 0.8676, + "step": 9305 + }, + { + "epoch": 0.7930123561994035, + "grad_norm": 13.25, + "learning_rate": 1.658892696833876e-06, + "loss": 0.5505, + "step": 9306 + }, + { + "epoch": 0.7930975713677034, + "grad_norm": 19.75, + "learning_rate": 1.6587339020936912e-06, + "loss": 0.5966, + "step": 9307 + }, + { + "epoch": 0.7931827865360034, + "grad_norm": 14.875, + "learning_rate": 1.658575099967274e-06, + "loss": 0.7274, + "step": 9308 + }, + { + "epoch": 0.7932680017043033, + "grad_norm": 14.5, + "learning_rate": 1.6584162904574932e-06, + "loss": 0.8129, + "step": 9309 + }, + { + "epoch": 0.7933532168726033, + "grad_norm": 10.9375, + "learning_rate": 1.65825747356722e-06, + "loss": 0.4015, + "step": 9310 + }, + { + "epoch": 0.7934384320409033, + "grad_norm": 11.0, + "learning_rate": 1.6580986492993234e-06, + "loss": 0.4577, + "step": 9311 + }, + { + "epoch": 0.7935236472092032, + "grad_norm": 12.6875, + "learning_rate": 1.657939817656674e-06, + "loss": 0.677, + "step": 9312 + }, + { + "epoch": 0.7936088623775032, + "grad_norm": 14.5, + "learning_rate": 1.6577809786421417e-06, + "loss": 0.5482, + "step": 9313 + }, + { + "epoch": 0.7936940775458031, + "grad_norm": 17.25, + "learning_rate": 1.6576221322585973e-06, + "loss": 1.1866, + "step": 9314 + }, + { + "epoch": 0.7937792927141031, + "grad_norm": 12.6875, + "learning_rate": 1.6574632785089107e-06, + "loss": 0.5453, + "step": 9315 + }, + { + "epoch": 0.793864507882403, + "grad_norm": 12.5625, + "learning_rate": 1.657304417395953e-06, + "loss": 0.4564, + "step": 9316 + }, + { + "epoch": 0.793949723050703, + "grad_norm": 10.8125, + "learning_rate": 1.657145548922595e-06, + "loss": 0.3868, + "step": 9317 + }, + { + "epoch": 0.794034938219003, + "grad_norm": 11.625, + "learning_rate": 1.6569866730917074e-06, + "loss": 0.342, + "step": 9318 + }, + { + "epoch": 0.7941201533873029, + "grad_norm": 11.1875, + "learning_rate": 1.6568277899061618e-06, + "loss": 0.4057, + "step": 9319 + }, + { + "epoch": 0.7942053685556029, + "grad_norm": 21.5, + "learning_rate": 1.6566688993688285e-06, + "loss": 1.0049, + "step": 9320 + }, + { + "epoch": 0.7942905837239028, + "grad_norm": 14.6875, + "learning_rate": 1.6565100014825798e-06, + "loss": 0.5596, + "step": 9321 + }, + { + "epoch": 0.7943757988922028, + "grad_norm": 24.75, + "learning_rate": 1.6563510962502857e-06, + "loss": 0.9233, + "step": 9322 + }, + { + "epoch": 0.7944610140605027, + "grad_norm": 21.125, + "learning_rate": 1.6561921836748195e-06, + "loss": 0.673, + "step": 9323 + }, + { + "epoch": 0.7945462292288027, + "grad_norm": 22.375, + "learning_rate": 1.6560332637590514e-06, + "loss": 0.9202, + "step": 9324 + }, + { + "epoch": 0.7946314443971026, + "grad_norm": 23.0, + "learning_rate": 1.6558743365058546e-06, + "loss": 0.7403, + "step": 9325 + }, + { + "epoch": 0.7947166595654026, + "grad_norm": 15.125, + "learning_rate": 1.6557154019180998e-06, + "loss": 0.665, + "step": 9326 + }, + { + "epoch": 0.7948018747337026, + "grad_norm": 15.8125, + "learning_rate": 1.6555564599986603e-06, + "loss": 0.8816, + "step": 9327 + }, + { + "epoch": 0.7948870899020025, + "grad_norm": 13.1875, + "learning_rate": 1.6553975107504072e-06, + "loss": 0.707, + "step": 9328 + }, + { + "epoch": 0.7949723050703025, + "grad_norm": 13.25, + "learning_rate": 1.655238554176214e-06, + "loss": 0.5945, + "step": 9329 + }, + { + "epoch": 0.7950575202386024, + "grad_norm": 12.5625, + "learning_rate": 1.655079590278952e-06, + "loss": 0.6046, + "step": 9330 + }, + { + "epoch": 0.7951427354069024, + "grad_norm": 13.75, + "learning_rate": 1.6549206190614948e-06, + "loss": 0.8768, + "step": 9331 + }, + { + "epoch": 0.7952279505752023, + "grad_norm": 17.0, + "learning_rate": 1.6547616405267147e-06, + "loss": 0.5921, + "step": 9332 + }, + { + "epoch": 0.7953131657435023, + "grad_norm": 13.375, + "learning_rate": 1.6546026546774848e-06, + "loss": 0.5038, + "step": 9333 + }, + { + "epoch": 0.7953983809118023, + "grad_norm": 14.125, + "learning_rate": 1.6544436615166781e-06, + "loss": 0.6684, + "step": 9334 + }, + { + "epoch": 0.7954835960801022, + "grad_norm": 16.0, + "learning_rate": 1.654284661047168e-06, + "loss": 0.8307, + "step": 9335 + }, + { + "epoch": 0.7955688112484022, + "grad_norm": 13.375, + "learning_rate": 1.6541256532718272e-06, + "loss": 0.6982, + "step": 9336 + }, + { + "epoch": 0.7956540264167021, + "grad_norm": 16.25, + "learning_rate": 1.6539666381935293e-06, + "loss": 0.7893, + "step": 9337 + }, + { + "epoch": 0.7957392415850021, + "grad_norm": 13.9375, + "learning_rate": 1.6538076158151484e-06, + "loss": 0.6543, + "step": 9338 + }, + { + "epoch": 0.795824456753302, + "grad_norm": 13.0625, + "learning_rate": 1.6536485861395576e-06, + "loss": 0.4526, + "step": 9339 + }, + { + "epoch": 0.795909671921602, + "grad_norm": 17.125, + "learning_rate": 1.6534895491696313e-06, + "loss": 0.9724, + "step": 9340 + }, + { + "epoch": 0.795994887089902, + "grad_norm": 15.6875, + "learning_rate": 1.6533305049082427e-06, + "loss": 0.6917, + "step": 9341 + }, + { + "epoch": 0.7960801022582019, + "grad_norm": 10.875, + "learning_rate": 1.6531714533582664e-06, + "loss": 0.4756, + "step": 9342 + }, + { + "epoch": 0.7961653174265019, + "grad_norm": 12.9375, + "learning_rate": 1.6530123945225765e-06, + "loss": 0.6805, + "step": 9343 + }, + { + "epoch": 0.7962505325948018, + "grad_norm": 15.375, + "learning_rate": 1.6528533284040476e-06, + "loss": 0.8315, + "step": 9344 + }, + { + "epoch": 0.7963357477631018, + "grad_norm": 14.875, + "learning_rate": 1.6526942550055537e-06, + "loss": 0.7992, + "step": 9345 + }, + { + "epoch": 0.7964209629314017, + "grad_norm": 12.8125, + "learning_rate": 1.6525351743299697e-06, + "loss": 0.607, + "step": 9346 + }, + { + "epoch": 0.7965061780997017, + "grad_norm": 25.125, + "learning_rate": 1.6523760863801705e-06, + "loss": 1.0917, + "step": 9347 + }, + { + "epoch": 0.7965913932680017, + "grad_norm": 17.125, + "learning_rate": 1.6522169911590306e-06, + "loss": 1.0436, + "step": 9348 + }, + { + "epoch": 0.7966766084363016, + "grad_norm": 14.375, + "learning_rate": 1.6520578886694253e-06, + "loss": 0.8071, + "step": 9349 + }, + { + "epoch": 0.7967618236046016, + "grad_norm": 15.375, + "learning_rate": 1.6518987789142296e-06, + "loss": 0.7745, + "step": 9350 + }, + { + "epoch": 0.7968470387729015, + "grad_norm": 19.625, + "learning_rate": 1.6517396618963188e-06, + "loss": 0.8329, + "step": 9351 + }, + { + "epoch": 0.7969322539412015, + "grad_norm": 13.4375, + "learning_rate": 1.6515805376185685e-06, + "loss": 0.5778, + "step": 9352 + }, + { + "epoch": 0.7970174691095014, + "grad_norm": 12.9375, + "learning_rate": 1.6514214060838538e-06, + "loss": 0.7726, + "step": 9353 + }, + { + "epoch": 0.7971026842778014, + "grad_norm": 15.4375, + "learning_rate": 1.6512622672950506e-06, + "loss": 0.8389, + "step": 9354 + }, + { + "epoch": 0.7971878994461014, + "grad_norm": 19.0, + "learning_rate": 1.651103121255035e-06, + "loss": 1.2951, + "step": 9355 + }, + { + "epoch": 0.7972731146144013, + "grad_norm": 22.875, + "learning_rate": 1.6509439679666826e-06, + "loss": 1.0622, + "step": 9356 + }, + { + "epoch": 0.7973583297827013, + "grad_norm": 13.75, + "learning_rate": 1.6507848074328694e-06, + "loss": 0.865, + "step": 9357 + }, + { + "epoch": 0.7974435449510012, + "grad_norm": 11.9375, + "learning_rate": 1.6506256396564718e-06, + "loss": 0.5667, + "step": 9358 + }, + { + "epoch": 0.7975287601193012, + "grad_norm": 16.75, + "learning_rate": 1.6504664646403662e-06, + "loss": 0.6396, + "step": 9359 + }, + { + "epoch": 0.7976139752876011, + "grad_norm": 16.25, + "learning_rate": 1.6503072823874287e-06, + "loss": 0.7614, + "step": 9360 + }, + { + "epoch": 0.7976991904559011, + "grad_norm": 17.5, + "learning_rate": 1.6501480929005358e-06, + "loss": 1.3279, + "step": 9361 + }, + { + "epoch": 0.797784405624201, + "grad_norm": 12.0, + "learning_rate": 1.649988896182565e-06, + "loss": 0.7284, + "step": 9362 + }, + { + "epoch": 0.797869620792501, + "grad_norm": 20.0, + "learning_rate": 1.649829692236392e-06, + "loss": 0.9211, + "step": 9363 + }, + { + "epoch": 0.797954835960801, + "grad_norm": 12.125, + "learning_rate": 1.649670481064895e-06, + "loss": 0.6325, + "step": 9364 + }, + { + "epoch": 0.7980400511291009, + "grad_norm": 13.875, + "learning_rate": 1.6495112626709505e-06, + "loss": 0.7835, + "step": 9365 + }, + { + "epoch": 0.798125266297401, + "grad_norm": 19.0, + "learning_rate": 1.6493520370574354e-06, + "loss": 0.8174, + "step": 9366 + }, + { + "epoch": 0.798210481465701, + "grad_norm": 14.0625, + "learning_rate": 1.6491928042272276e-06, + "loss": 0.8003, + "step": 9367 + }, + { + "epoch": 0.7982956966340009, + "grad_norm": 20.5, + "learning_rate": 1.6490335641832045e-06, + "loss": 0.6849, + "step": 9368 + }, + { + "epoch": 0.7983809118023009, + "grad_norm": 14.125, + "learning_rate": 1.648874316928243e-06, + "loss": 0.8997, + "step": 9369 + }, + { + "epoch": 0.7984661269706008, + "grad_norm": 10.0, + "learning_rate": 1.648715062465222e-06, + "loss": 0.2884, + "step": 9370 + }, + { + "epoch": 0.7985513421389008, + "grad_norm": 14.0625, + "learning_rate": 1.648555800797019e-06, + "loss": 0.6575, + "step": 9371 + }, + { + "epoch": 0.7986365573072007, + "grad_norm": 12.1875, + "learning_rate": 1.6483965319265121e-06, + "loss": 0.7147, + "step": 9372 + }, + { + "epoch": 0.7987217724755007, + "grad_norm": 20.375, + "learning_rate": 1.6482372558565787e-06, + "loss": 1.119, + "step": 9373 + }, + { + "epoch": 0.7988069876438006, + "grad_norm": 12.125, + "learning_rate": 1.6480779725900979e-06, + "loss": 0.5244, + "step": 9374 + }, + { + "epoch": 0.7988922028121006, + "grad_norm": 16.0, + "learning_rate": 1.6479186821299475e-06, + "loss": 0.6104, + "step": 9375 + }, + { + "epoch": 0.7989774179804006, + "grad_norm": 11.9375, + "learning_rate": 1.647759384479007e-06, + "loss": 0.5298, + "step": 9376 + }, + { + "epoch": 0.7990626331487005, + "grad_norm": 13.0625, + "learning_rate": 1.647600079640154e-06, + "loss": 0.3543, + "step": 9377 + }, + { + "epoch": 0.7991478483170005, + "grad_norm": 10.5, + "learning_rate": 1.647440767616268e-06, + "loss": 0.4199, + "step": 9378 + }, + { + "epoch": 0.7992330634853004, + "grad_norm": 14.0, + "learning_rate": 1.6472814484102273e-06, + "loss": 0.7481, + "step": 9379 + }, + { + "epoch": 0.7993182786536004, + "grad_norm": 21.5, + "learning_rate": 1.6471221220249117e-06, + "loss": 0.8683, + "step": 9380 + }, + { + "epoch": 0.7994034938219003, + "grad_norm": 14.625, + "learning_rate": 1.6469627884631997e-06, + "loss": 0.6093, + "step": 9381 + }, + { + "epoch": 0.7994887089902003, + "grad_norm": 13.5, + "learning_rate": 1.6468034477279712e-06, + "loss": 0.7172, + "step": 9382 + }, + { + "epoch": 0.7995739241585003, + "grad_norm": 19.375, + "learning_rate": 1.646644099822105e-06, + "loss": 0.8839, + "step": 9383 + }, + { + "epoch": 0.7996591393268002, + "grad_norm": 14.375, + "learning_rate": 1.6464847447484817e-06, + "loss": 0.6429, + "step": 9384 + }, + { + "epoch": 0.7997443544951002, + "grad_norm": 17.5, + "learning_rate": 1.6463253825099799e-06, + "loss": 0.5567, + "step": 9385 + }, + { + "epoch": 0.7998295696634001, + "grad_norm": 14.3125, + "learning_rate": 1.6461660131094797e-06, + "loss": 0.5549, + "step": 9386 + }, + { + "epoch": 0.7999147848317001, + "grad_norm": 19.125, + "learning_rate": 1.6460066365498617e-06, + "loss": 1.2088, + "step": 9387 + }, + { + "epoch": 0.8, + "grad_norm": 16.25, + "learning_rate": 1.6458472528340054e-06, + "loss": 0.9374, + "step": 9388 + }, + { + "epoch": 0.8000852151683, + "grad_norm": 13.1875, + "learning_rate": 1.6456878619647914e-06, + "loss": 0.5813, + "step": 9389 + }, + { + "epoch": 0.8001704303366, + "grad_norm": 12.6875, + "learning_rate": 1.6455284639450996e-06, + "loss": 0.5815, + "step": 9390 + }, + { + "epoch": 0.8002556455048999, + "grad_norm": 14.1875, + "learning_rate": 1.645369058777811e-06, + "loss": 0.6624, + "step": 9391 + }, + { + "epoch": 0.8003408606731999, + "grad_norm": 14.6875, + "learning_rate": 1.6452096464658056e-06, + "loss": 0.5533, + "step": 9392 + }, + { + "epoch": 0.8004260758414998, + "grad_norm": 12.0625, + "learning_rate": 1.6450502270119645e-06, + "loss": 0.4451, + "step": 9393 + }, + { + "epoch": 0.8005112910097998, + "grad_norm": 11.5625, + "learning_rate": 1.6448908004191684e-06, + "loss": 0.5631, + "step": 9394 + }, + { + "epoch": 0.8005965061780997, + "grad_norm": 30.25, + "learning_rate": 1.644731366690299e-06, + "loss": 1.4518, + "step": 9395 + }, + { + "epoch": 0.8006817213463997, + "grad_norm": 12.0, + "learning_rate": 1.6445719258282367e-06, + "loss": 0.3933, + "step": 9396 + }, + { + "epoch": 0.8007669365146997, + "grad_norm": 18.125, + "learning_rate": 1.6444124778358628e-06, + "loss": 0.8528, + "step": 9397 + }, + { + "epoch": 0.8008521516829996, + "grad_norm": 18.5, + "learning_rate": 1.6442530227160585e-06, + "loss": 0.6588, + "step": 9398 + }, + { + "epoch": 0.8009373668512996, + "grad_norm": 13.375, + "learning_rate": 1.6440935604717059e-06, + "loss": 0.9586, + "step": 9399 + }, + { + "epoch": 0.8010225820195995, + "grad_norm": 20.125, + "learning_rate": 1.6439340911056868e-06, + "loss": 1.258, + "step": 9400 + }, + { + "epoch": 0.8011077971878995, + "grad_norm": 12.5625, + "learning_rate": 1.6437746146208816e-06, + "loss": 0.712, + "step": 9401 + }, + { + "epoch": 0.8011930123561994, + "grad_norm": 12.8125, + "learning_rate": 1.643615131020174e-06, + "loss": 0.7403, + "step": 9402 + }, + { + "epoch": 0.8012782275244994, + "grad_norm": 16.25, + "learning_rate": 1.6434556403064445e-06, + "loss": 0.5882, + "step": 9403 + }, + { + "epoch": 0.8013634426927994, + "grad_norm": 15.0, + "learning_rate": 1.6432961424825767e-06, + "loss": 0.7012, + "step": 9404 + }, + { + "epoch": 0.8014486578610993, + "grad_norm": 27.5, + "learning_rate": 1.6431366375514515e-06, + "loss": 1.2114, + "step": 9405 + }, + { + "epoch": 0.8015338730293993, + "grad_norm": 15.5625, + "learning_rate": 1.642977125515952e-06, + "loss": 0.8754, + "step": 9406 + }, + { + "epoch": 0.8016190881976992, + "grad_norm": 16.25, + "learning_rate": 1.6428176063789607e-06, + "loss": 0.8336, + "step": 9407 + }, + { + "epoch": 0.8017043033659992, + "grad_norm": 12.625, + "learning_rate": 1.6426580801433607e-06, + "loss": 0.7895, + "step": 9408 + }, + { + "epoch": 0.8017895185342991, + "grad_norm": 13.25, + "learning_rate": 1.642498546812034e-06, + "loss": 0.4936, + "step": 9409 + }, + { + "epoch": 0.8018747337025991, + "grad_norm": 16.75, + "learning_rate": 1.6423390063878636e-06, + "loss": 0.5878, + "step": 9410 + }, + { + "epoch": 0.801959948870899, + "grad_norm": 11.5625, + "learning_rate": 1.642179458873733e-06, + "loss": 0.6331, + "step": 9411 + }, + { + "epoch": 0.802045164039199, + "grad_norm": 18.75, + "learning_rate": 1.6420199042725258e-06, + "loss": 1.1841, + "step": 9412 + }, + { + "epoch": 0.802130379207499, + "grad_norm": 32.5, + "learning_rate": 1.6418603425871239e-06, + "loss": 1.244, + "step": 9413 + }, + { + "epoch": 0.8022155943757989, + "grad_norm": 11.25, + "learning_rate": 1.6417007738204118e-06, + "loss": 0.6814, + "step": 9414 + }, + { + "epoch": 0.8023008095440989, + "grad_norm": 9.25, + "learning_rate": 1.641541197975273e-06, + "loss": 0.2862, + "step": 9415 + }, + { + "epoch": 0.8023860247123988, + "grad_norm": 15.0, + "learning_rate": 1.6413816150545908e-06, + "loss": 0.767, + "step": 9416 + }, + { + "epoch": 0.8024712398806988, + "grad_norm": 13.4375, + "learning_rate": 1.6412220250612496e-06, + "loss": 0.752, + "step": 9417 + }, + { + "epoch": 0.8025564550489988, + "grad_norm": 15.6875, + "learning_rate": 1.6410624279981324e-06, + "loss": 0.7724, + "step": 9418 + }, + { + "epoch": 0.8026416702172987, + "grad_norm": 16.125, + "learning_rate": 1.6409028238681247e-06, + "loss": 0.9049, + "step": 9419 + }, + { + "epoch": 0.8027268853855987, + "grad_norm": 16.25, + "learning_rate": 1.6407432126741093e-06, + "loss": 0.984, + "step": 9420 + }, + { + "epoch": 0.8028121005538986, + "grad_norm": 16.25, + "learning_rate": 1.6405835944189713e-06, + "loss": 0.845, + "step": 9421 + }, + { + "epoch": 0.8028973157221986, + "grad_norm": 13.625, + "learning_rate": 1.6404239691055947e-06, + "loss": 0.8253, + "step": 9422 + }, + { + "epoch": 0.8029825308904985, + "grad_norm": 14.625, + "learning_rate": 1.6402643367368649e-06, + "loss": 0.2041, + "step": 9423 + }, + { + "epoch": 0.8030677460587985, + "grad_norm": 13.0, + "learning_rate": 1.6401046973156656e-06, + "loss": 0.4541, + "step": 9424 + }, + { + "epoch": 0.8031529612270984, + "grad_norm": 16.875, + "learning_rate": 1.6399450508448827e-06, + "loss": 0.6505, + "step": 9425 + }, + { + "epoch": 0.8032381763953984, + "grad_norm": 23.25, + "learning_rate": 1.6397853973274003e-06, + "loss": 1.4802, + "step": 9426 + }, + { + "epoch": 0.8033233915636984, + "grad_norm": 12.6875, + "learning_rate": 1.6396257367661037e-06, + "loss": 0.6071, + "step": 9427 + }, + { + "epoch": 0.8034086067319983, + "grad_norm": 16.625, + "learning_rate": 1.639466069163878e-06, + "loss": 0.6639, + "step": 9428 + }, + { + "epoch": 0.8034938219002983, + "grad_norm": 12.0625, + "learning_rate": 1.6393063945236095e-06, + "loss": 0.811, + "step": 9429 + }, + { + "epoch": 0.8035790370685982, + "grad_norm": 17.25, + "learning_rate": 1.6391467128481825e-06, + "loss": 0.6467, + "step": 9430 + }, + { + "epoch": 0.8036642522368982, + "grad_norm": 13.125, + "learning_rate": 1.6389870241404834e-06, + "loss": 0.6829, + "step": 9431 + }, + { + "epoch": 0.8037494674051981, + "grad_norm": 14.4375, + "learning_rate": 1.6388273284033972e-06, + "loss": 0.8062, + "step": 9432 + }, + { + "epoch": 0.8038346825734981, + "grad_norm": 19.125, + "learning_rate": 1.6386676256398104e-06, + "loss": 0.8652, + "step": 9433 + }, + { + "epoch": 0.8039198977417981, + "grad_norm": 14.125, + "learning_rate": 1.6385079158526088e-06, + "loss": 0.5597, + "step": 9434 + }, + { + "epoch": 0.804005112910098, + "grad_norm": 16.25, + "learning_rate": 1.6383481990446787e-06, + "loss": 0.4233, + "step": 9435 + }, + { + "epoch": 0.804090328078398, + "grad_norm": 9.8125, + "learning_rate": 1.6381884752189056e-06, + "loss": 0.252, + "step": 9436 + }, + { + "epoch": 0.8041755432466979, + "grad_norm": 16.875, + "learning_rate": 1.6380287443781768e-06, + "loss": 0.8427, + "step": 9437 + }, + { + "epoch": 0.8042607584149979, + "grad_norm": 12.5, + "learning_rate": 1.6378690065253784e-06, + "loss": 0.5465, + "step": 9438 + }, + { + "epoch": 0.8043459735832978, + "grad_norm": 35.0, + "learning_rate": 1.6377092616633967e-06, + "loss": 0.5623, + "step": 9439 + }, + { + "epoch": 0.8044311887515978, + "grad_norm": 15.25, + "learning_rate": 1.6375495097951188e-06, + "loss": 0.7402, + "step": 9440 + }, + { + "epoch": 0.8045164039198978, + "grad_norm": 14.75, + "learning_rate": 1.6373897509234318e-06, + "loss": 0.7543, + "step": 9441 + }, + { + "epoch": 0.8046016190881977, + "grad_norm": 20.25, + "learning_rate": 1.6372299850512224e-06, + "loss": 1.0142, + "step": 9442 + }, + { + "epoch": 0.8046868342564977, + "grad_norm": 15.0625, + "learning_rate": 1.6370702121813776e-06, + "loss": 1.0541, + "step": 9443 + }, + { + "epoch": 0.8047720494247976, + "grad_norm": 15.0, + "learning_rate": 1.6369104323167851e-06, + "loss": 0.525, + "step": 9444 + }, + { + "epoch": 0.8048572645930976, + "grad_norm": 15.6875, + "learning_rate": 1.6367506454603319e-06, + "loss": 0.7357, + "step": 9445 + }, + { + "epoch": 0.8049424797613975, + "grad_norm": 15.875, + "learning_rate": 1.6365908516149056e-06, + "loss": 0.8136, + "step": 9446 + }, + { + "epoch": 0.8050276949296975, + "grad_norm": 19.0, + "learning_rate": 1.6364310507833938e-06, + "loss": 0.7438, + "step": 9447 + }, + { + "epoch": 0.8051129100979975, + "grad_norm": 11.8125, + "learning_rate": 1.6362712429686844e-06, + "loss": 0.3985, + "step": 9448 + }, + { + "epoch": 0.8051981252662974, + "grad_norm": 16.25, + "learning_rate": 1.6361114281736653e-06, + "loss": 0.803, + "step": 9449 + }, + { + "epoch": 0.8052833404345974, + "grad_norm": 31.5, + "learning_rate": 1.6359516064012246e-06, + "loss": 0.6475, + "step": 9450 + }, + { + "epoch": 0.8053685556028973, + "grad_norm": 20.125, + "learning_rate": 1.6357917776542497e-06, + "loss": 0.9331, + "step": 9451 + }, + { + "epoch": 0.8054537707711973, + "grad_norm": 20.125, + "learning_rate": 1.6356319419356298e-06, + "loss": 1.2658, + "step": 9452 + }, + { + "epoch": 0.8055389859394972, + "grad_norm": 15.6875, + "learning_rate": 1.635472099248253e-06, + "loss": 0.6567, + "step": 9453 + }, + { + "epoch": 0.8056242011077972, + "grad_norm": 13.3125, + "learning_rate": 1.6353122495950075e-06, + "loss": 0.5096, + "step": 9454 + }, + { + "epoch": 0.8057094162760972, + "grad_norm": 18.125, + "learning_rate": 1.635152392978782e-06, + "loss": 0.808, + "step": 9455 + }, + { + "epoch": 0.8057946314443971, + "grad_norm": 15.125, + "learning_rate": 1.634992529402466e-06, + "loss": 0.9815, + "step": 9456 + }, + { + "epoch": 0.8058798466126971, + "grad_norm": 30.25, + "learning_rate": 1.634832658868948e-06, + "loss": 0.9481, + "step": 9457 + }, + { + "epoch": 0.805965061780997, + "grad_norm": 23.75, + "learning_rate": 1.6346727813811162e-06, + "loss": 1.0077, + "step": 9458 + }, + { + "epoch": 0.806050276949297, + "grad_norm": 10.75, + "learning_rate": 1.6345128969418608e-06, + "loss": 0.439, + "step": 9459 + }, + { + "epoch": 0.8061354921175969, + "grad_norm": 15.5, + "learning_rate": 1.6343530055540704e-06, + "loss": 0.8602, + "step": 9460 + }, + { + "epoch": 0.8062207072858969, + "grad_norm": 22.125, + "learning_rate": 1.6341931072206352e-06, + "loss": 0.7514, + "step": 9461 + }, + { + "epoch": 0.8063059224541969, + "grad_norm": 21.375, + "learning_rate": 1.6340332019444438e-06, + "loss": 0.935, + "step": 9462 + }, + { + "epoch": 0.8063911376224968, + "grad_norm": 15.6875, + "learning_rate": 1.6338732897283866e-06, + "loss": 1.0326, + "step": 9463 + }, + { + "epoch": 0.8064763527907968, + "grad_norm": 14.9375, + "learning_rate": 1.6337133705753523e-06, + "loss": 0.9387, + "step": 9464 + }, + { + "epoch": 0.8065615679590967, + "grad_norm": 10.375, + "learning_rate": 1.6335534444882326e-06, + "loss": 0.4275, + "step": 9465 + }, + { + "epoch": 0.8066467831273967, + "grad_norm": 20.0, + "learning_rate": 1.6333935114699156e-06, + "loss": 0.6612, + "step": 9466 + }, + { + "epoch": 0.8067319982956966, + "grad_norm": 11.8125, + "learning_rate": 1.6332335715232927e-06, + "loss": 0.5352, + "step": 9467 + }, + { + "epoch": 0.8068172134639966, + "grad_norm": 12.9375, + "learning_rate": 1.6330736246512535e-06, + "loss": 0.5707, + "step": 9468 + }, + { + "epoch": 0.8069024286322966, + "grad_norm": 24.5, + "learning_rate": 1.6329136708566892e-06, + "loss": 0.7378, + "step": 9469 + }, + { + "epoch": 0.8069876438005965, + "grad_norm": 11.75, + "learning_rate": 1.6327537101424894e-06, + "loss": 0.4521, + "step": 9470 + }, + { + "epoch": 0.8070728589688965, + "grad_norm": 13.25, + "learning_rate": 1.6325937425115454e-06, + "loss": 0.8363, + "step": 9471 + }, + { + "epoch": 0.8071580741371964, + "grad_norm": 11.8125, + "learning_rate": 1.6324337679667477e-06, + "loss": 0.4928, + "step": 9472 + }, + { + "epoch": 0.8072432893054964, + "grad_norm": 13.1875, + "learning_rate": 1.6322737865109872e-06, + "loss": 0.5882, + "step": 9473 + }, + { + "epoch": 0.8073285044737963, + "grad_norm": 11.875, + "learning_rate": 1.6321137981471551e-06, + "loss": 0.5267, + "step": 9474 + }, + { + "epoch": 0.8074137196420963, + "grad_norm": 14.25, + "learning_rate": 1.6319538028781423e-06, + "loss": 0.635, + "step": 9475 + }, + { + "epoch": 0.8074989348103963, + "grad_norm": 16.25, + "learning_rate": 1.63179380070684e-06, + "loss": 1.034, + "step": 9476 + }, + { + "epoch": 0.8075841499786962, + "grad_norm": 17.25, + "learning_rate": 1.6316337916361402e-06, + "loss": 1.1004, + "step": 9477 + }, + { + "epoch": 0.8076693651469962, + "grad_norm": 14.6875, + "learning_rate": 1.631473775668934e-06, + "loss": 0.5225, + "step": 9478 + }, + { + "epoch": 0.8077545803152961, + "grad_norm": 12.0, + "learning_rate": 1.6313137528081125e-06, + "loss": 0.6254, + "step": 9479 + }, + { + "epoch": 0.8078397954835961, + "grad_norm": 15.4375, + "learning_rate": 1.6311537230565686e-06, + "loss": 0.5595, + "step": 9480 + }, + { + "epoch": 0.807925010651896, + "grad_norm": 15.0, + "learning_rate": 1.6309936864171935e-06, + "loss": 0.4604, + "step": 9481 + }, + { + "epoch": 0.808010225820196, + "grad_norm": 18.875, + "learning_rate": 1.6308336428928794e-06, + "loss": 0.9487, + "step": 9482 + }, + { + "epoch": 0.808095440988496, + "grad_norm": 17.25, + "learning_rate": 1.6306735924865184e-06, + "loss": 0.5752, + "step": 9483 + }, + { + "epoch": 0.8081806561567959, + "grad_norm": 18.0, + "learning_rate": 1.6305135352010026e-06, + "loss": 0.9865, + "step": 9484 + }, + { + "epoch": 0.8082658713250959, + "grad_norm": 10.5, + "learning_rate": 1.630353471039225e-06, + "loss": 0.2656, + "step": 9485 + }, + { + "epoch": 0.8083510864933958, + "grad_norm": 12.9375, + "learning_rate": 1.6301934000040773e-06, + "loss": 0.8126, + "step": 9486 + }, + { + "epoch": 0.8084363016616958, + "grad_norm": 18.625, + "learning_rate": 1.6300333220984526e-06, + "loss": 1.0819, + "step": 9487 + }, + { + "epoch": 0.8085215168299957, + "grad_norm": 15.0625, + "learning_rate": 1.6298732373252436e-06, + "loss": 0.8976, + "step": 9488 + }, + { + "epoch": 0.8086067319982957, + "grad_norm": 11.9375, + "learning_rate": 1.6297131456873433e-06, + "loss": 0.6188, + "step": 9489 + }, + { + "epoch": 0.8086919471665956, + "grad_norm": 56.25, + "learning_rate": 1.6295530471876448e-06, + "loss": 0.7695, + "step": 9490 + }, + { + "epoch": 0.8087771623348956, + "grad_norm": 15.25, + "learning_rate": 1.629392941829041e-06, + "loss": 0.7134, + "step": 9491 + }, + { + "epoch": 0.8088623775031956, + "grad_norm": 24.375, + "learning_rate": 1.629232829614425e-06, + "loss": 0.8882, + "step": 9492 + }, + { + "epoch": 0.8089475926714955, + "grad_norm": 14.75, + "learning_rate": 1.6290727105466906e-06, + "loss": 0.7056, + "step": 9493 + }, + { + "epoch": 0.8090328078397955, + "grad_norm": 12.0, + "learning_rate": 1.6289125846287313e-06, + "loss": 0.5221, + "step": 9494 + }, + { + "epoch": 0.8091180230080954, + "grad_norm": 16.375, + "learning_rate": 1.6287524518634402e-06, + "loss": 0.6398, + "step": 9495 + }, + { + "epoch": 0.8092032381763954, + "grad_norm": 16.875, + "learning_rate": 1.6285923122537114e-06, + "loss": 0.8392, + "step": 9496 + }, + { + "epoch": 0.8092884533446953, + "grad_norm": 11.6875, + "learning_rate": 1.628432165802439e-06, + "loss": 0.4834, + "step": 9497 + }, + { + "epoch": 0.8093736685129953, + "grad_norm": 12.75, + "learning_rate": 1.628272012512517e-06, + "loss": 0.7347, + "step": 9498 + }, + { + "epoch": 0.8094588836812953, + "grad_norm": 16.0, + "learning_rate": 1.6281118523868395e-06, + "loss": 0.9386, + "step": 9499 + }, + { + "epoch": 0.8095440988495952, + "grad_norm": 17.125, + "learning_rate": 1.6279516854283e-06, + "loss": 0.8599, + "step": 9500 + }, + { + "epoch": 0.8096293140178952, + "grad_norm": 11.1875, + "learning_rate": 1.6277915116397938e-06, + "loss": 0.5091, + "step": 9501 + }, + { + "epoch": 0.8097145291861951, + "grad_norm": 18.75, + "learning_rate": 1.6276313310242151e-06, + "loss": 0.5912, + "step": 9502 + }, + { + "epoch": 0.8097997443544951, + "grad_norm": 15.0, + "learning_rate": 1.6274711435844582e-06, + "loss": 0.649, + "step": 9503 + }, + { + "epoch": 0.809884959522795, + "grad_norm": 10.625, + "learning_rate": 1.6273109493234184e-06, + "loss": 0.2716, + "step": 9504 + }, + { + "epoch": 0.809970174691095, + "grad_norm": 22.5, + "learning_rate": 1.6271507482439907e-06, + "loss": 0.3357, + "step": 9505 + }, + { + "epoch": 0.810055389859395, + "grad_norm": 10.75, + "learning_rate": 1.6269905403490693e-06, + "loss": 0.506, + "step": 9506 + }, + { + "epoch": 0.8101406050276949, + "grad_norm": 15.4375, + "learning_rate": 1.6268303256415496e-06, + "loss": 1.1093, + "step": 9507 + }, + { + "epoch": 0.8102258201959949, + "grad_norm": 12.4375, + "learning_rate": 1.6266701041243275e-06, + "loss": 0.4562, + "step": 9508 + }, + { + "epoch": 0.8103110353642948, + "grad_norm": 10.25, + "learning_rate": 1.6265098758002972e-06, + "loss": 0.2915, + "step": 9509 + }, + { + "epoch": 0.8103962505325948, + "grad_norm": 16.25, + "learning_rate": 1.6263496406723556e-06, + "loss": 0.7986, + "step": 9510 + }, + { + "epoch": 0.8104814657008947, + "grad_norm": 14.0625, + "learning_rate": 1.6261893987433968e-06, + "loss": 0.6524, + "step": 9511 + }, + { + "epoch": 0.8105666808691947, + "grad_norm": 10.75, + "learning_rate": 1.6260291500163177e-06, + "loss": 0.4812, + "step": 9512 + }, + { + "epoch": 0.8106518960374947, + "grad_norm": 25.875, + "learning_rate": 1.6258688944940138e-06, + "loss": 0.9877, + "step": 9513 + }, + { + "epoch": 0.8107371112057946, + "grad_norm": 18.5, + "learning_rate": 1.6257086321793812e-06, + "loss": 1.069, + "step": 9514 + }, + { + "epoch": 0.8108223263740946, + "grad_norm": 22.625, + "learning_rate": 1.625548363075315e-06, + "loss": 0.8805, + "step": 9515 + }, + { + "epoch": 0.8109075415423945, + "grad_norm": 12.75, + "learning_rate": 1.6253880871847127e-06, + "loss": 0.493, + "step": 9516 + }, + { + "epoch": 0.8109927567106945, + "grad_norm": 12.8125, + "learning_rate": 1.6252278045104703e-06, + "loss": 0.6482, + "step": 9517 + }, + { + "epoch": 0.8110779718789944, + "grad_norm": 14.125, + "learning_rate": 1.625067515055484e-06, + "loss": 0.3299, + "step": 9518 + }, + { + "epoch": 0.8111631870472944, + "grad_norm": 14.9375, + "learning_rate": 1.6249072188226505e-06, + "loss": 0.9892, + "step": 9519 + }, + { + "epoch": 0.8112484022155944, + "grad_norm": 18.625, + "learning_rate": 1.6247469158148668e-06, + "loss": 0.6329, + "step": 9520 + }, + { + "epoch": 0.8113336173838943, + "grad_norm": 15.4375, + "learning_rate": 1.624586606035029e-06, + "loss": 0.6345, + "step": 9521 + }, + { + "epoch": 0.8114188325521943, + "grad_norm": 25.625, + "learning_rate": 1.6244262894860352e-06, + "loss": 0.9274, + "step": 9522 + }, + { + "epoch": 0.8115040477204942, + "grad_norm": 17.25, + "learning_rate": 1.6242659661707816e-06, + "loss": 1.0147, + "step": 9523 + }, + { + "epoch": 0.8115892628887942, + "grad_norm": 22.375, + "learning_rate": 1.6241056360921654e-06, + "loss": 1.0602, + "step": 9524 + }, + { + "epoch": 0.8116744780570941, + "grad_norm": 23.375, + "learning_rate": 1.6239452992530843e-06, + "loss": 0.6488, + "step": 9525 + }, + { + "epoch": 0.8117596932253941, + "grad_norm": 23.0, + "learning_rate": 1.6237849556564355e-06, + "loss": 1.0389, + "step": 9526 + }, + { + "epoch": 0.811844908393694, + "grad_norm": 13.0625, + "learning_rate": 1.6236246053051172e-06, + "loss": 0.6787, + "step": 9527 + }, + { + "epoch": 0.811930123561994, + "grad_norm": 16.75, + "learning_rate": 1.6234642482020257e-06, + "loss": 0.5816, + "step": 9528 + }, + { + "epoch": 0.812015338730294, + "grad_norm": 19.0, + "learning_rate": 1.6233038843500604e-06, + "loss": 0.6062, + "step": 9529 + }, + { + "epoch": 0.8121005538985939, + "grad_norm": 30.25, + "learning_rate": 1.6231435137521184e-06, + "loss": 0.9707, + "step": 9530 + }, + { + "epoch": 0.8121857690668939, + "grad_norm": 14.75, + "learning_rate": 1.622983136411098e-06, + "loss": 0.7146, + "step": 9531 + }, + { + "epoch": 0.8122709842351938, + "grad_norm": 11.5, + "learning_rate": 1.6228227523298968e-06, + "loss": 0.4218, + "step": 9532 + }, + { + "epoch": 0.8123561994034938, + "grad_norm": 11.6875, + "learning_rate": 1.622662361511414e-06, + "loss": 0.395, + "step": 9533 + }, + { + "epoch": 0.8124414145717938, + "grad_norm": 14.4375, + "learning_rate": 1.6225019639585473e-06, + "loss": 0.7819, + "step": 9534 + }, + { + "epoch": 0.8125266297400937, + "grad_norm": 10.625, + "learning_rate": 1.6223415596741957e-06, + "loss": 0.3752, + "step": 9535 + }, + { + "epoch": 0.8126118449083937, + "grad_norm": 18.375, + "learning_rate": 1.6221811486612574e-06, + "loss": 1.1257, + "step": 9536 + }, + { + "epoch": 0.8126970600766936, + "grad_norm": 15.3125, + "learning_rate": 1.6220207309226318e-06, + "loss": 0.7075, + "step": 9537 + }, + { + "epoch": 0.8127822752449936, + "grad_norm": 10.4375, + "learning_rate": 1.6218603064612176e-06, + "loss": 0.2056, + "step": 9538 + }, + { + "epoch": 0.8128674904132935, + "grad_norm": 12.625, + "learning_rate": 1.6216998752799132e-06, + "loss": 0.6924, + "step": 9539 + }, + { + "epoch": 0.8129527055815935, + "grad_norm": 11.6875, + "learning_rate": 1.6215394373816187e-06, + "loss": 0.4378, + "step": 9540 + }, + { + "epoch": 0.8130379207498935, + "grad_norm": 17.25, + "learning_rate": 1.6213789927692328e-06, + "loss": 0.7156, + "step": 9541 + }, + { + "epoch": 0.8131231359181934, + "grad_norm": 18.5, + "learning_rate": 1.6212185414456554e-06, + "loss": 1.0168, + "step": 9542 + }, + { + "epoch": 0.8132083510864934, + "grad_norm": 12.5625, + "learning_rate": 1.621058083413785e-06, + "loss": 0.3976, + "step": 9543 + }, + { + "epoch": 0.8132935662547933, + "grad_norm": 12.6875, + "learning_rate": 1.6208976186765223e-06, + "loss": 0.7649, + "step": 9544 + }, + { + "epoch": 0.8133787814230933, + "grad_norm": 12.0, + "learning_rate": 1.6207371472367666e-06, + "loss": 0.4154, + "step": 9545 + }, + { + "epoch": 0.8134639965913932, + "grad_norm": 15.8125, + "learning_rate": 1.6205766690974185e-06, + "loss": 0.6174, + "step": 9546 + }, + { + "epoch": 0.8135492117596932, + "grad_norm": 22.0, + "learning_rate": 1.6204161842613764e-06, + "loss": 0.9451, + "step": 9547 + }, + { + "epoch": 0.8136344269279931, + "grad_norm": 13.0, + "learning_rate": 1.6202556927315419e-06, + "loss": 0.4744, + "step": 9548 + }, + { + "epoch": 0.8137196420962931, + "grad_norm": 21.125, + "learning_rate": 1.6200951945108143e-06, + "loss": 0.8855, + "step": 9549 + }, + { + "epoch": 0.8138048572645931, + "grad_norm": 19.5, + "learning_rate": 1.6199346896020948e-06, + "loss": 1.2526, + "step": 9550 + }, + { + "epoch": 0.813890072432893, + "grad_norm": 33.25, + "learning_rate": 1.6197741780082832e-06, + "loss": 1.0598, + "step": 9551 + }, + { + "epoch": 0.813975287601193, + "grad_norm": 12.8125, + "learning_rate": 1.6196136597322803e-06, + "loss": 0.7703, + "step": 9552 + }, + { + "epoch": 0.8140605027694929, + "grad_norm": 16.625, + "learning_rate": 1.619453134776987e-06, + "loss": 0.8769, + "step": 9553 + }, + { + "epoch": 0.8141457179377929, + "grad_norm": 11.125, + "learning_rate": 1.6192926031453041e-06, + "loss": 0.4337, + "step": 9554 + }, + { + "epoch": 0.8142309331060928, + "grad_norm": 34.5, + "learning_rate": 1.6191320648401325e-06, + "loss": 0.8267, + "step": 9555 + }, + { + "epoch": 0.8143161482743928, + "grad_norm": 14.625, + "learning_rate": 1.6189715198643732e-06, + "loss": 0.5283, + "step": 9556 + }, + { + "epoch": 0.8144013634426928, + "grad_norm": 23.75, + "learning_rate": 1.6188109682209274e-06, + "loss": 0.7032, + "step": 9557 + }, + { + "epoch": 0.8144865786109927, + "grad_norm": 16.125, + "learning_rate": 1.6186504099126971e-06, + "loss": 0.755, + "step": 9558 + }, + { + "epoch": 0.8145717937792927, + "grad_norm": 15.3125, + "learning_rate": 1.618489844942583e-06, + "loss": 0.9224, + "step": 9559 + }, + { + "epoch": 0.8146570089475926, + "grad_norm": 14.875, + "learning_rate": 1.618329273313486e-06, + "loss": 0.7357, + "step": 9560 + }, + { + "epoch": 0.8147422241158926, + "grad_norm": 12.5625, + "learning_rate": 1.6181686950283096e-06, + "loss": 0.7173, + "step": 9561 + }, + { + "epoch": 0.8148274392841925, + "grad_norm": 13.125, + "learning_rate": 1.6180081100899543e-06, + "loss": 0.6681, + "step": 9562 + }, + { + "epoch": 0.8149126544524925, + "grad_norm": 11.875, + "learning_rate": 1.6178475185013226e-06, + "loss": 0.5094, + "step": 9563 + }, + { + "epoch": 0.8149978696207925, + "grad_norm": 15.625, + "learning_rate": 1.617686920265316e-06, + "loss": 0.9409, + "step": 9564 + }, + { + "epoch": 0.8150830847890924, + "grad_norm": 9.5, + "learning_rate": 1.6175263153848372e-06, + "loss": 0.3378, + "step": 9565 + }, + { + "epoch": 0.8151682999573924, + "grad_norm": 22.25, + "learning_rate": 1.6173657038627881e-06, + "loss": 0.8048, + "step": 9566 + }, + { + "epoch": 0.8152535151256923, + "grad_norm": 12.875, + "learning_rate": 1.6172050857020722e-06, + "loss": 0.6447, + "step": 9567 + }, + { + "epoch": 0.8153387302939923, + "grad_norm": 14.9375, + "learning_rate": 1.61704446090559e-06, + "loss": 0.7769, + "step": 9568 + }, + { + "epoch": 0.8154239454622922, + "grad_norm": 16.125, + "learning_rate": 1.616883829476246e-06, + "loss": 0.6015, + "step": 9569 + }, + { + "epoch": 0.8155091606305922, + "grad_norm": 11.375, + "learning_rate": 1.6167231914169419e-06, + "loss": 0.3633, + "step": 9570 + }, + { + "epoch": 0.8155943757988922, + "grad_norm": 12.125, + "learning_rate": 1.6165625467305812e-06, + "loss": 0.4571, + "step": 9571 + }, + { + "epoch": 0.8156795909671921, + "grad_norm": 11.375, + "learning_rate": 1.6164018954200666e-06, + "loss": 0.4366, + "step": 9572 + }, + { + "epoch": 0.8157648061354921, + "grad_norm": 13.125, + "learning_rate": 1.616241237488301e-06, + "loss": 0.798, + "step": 9573 + }, + { + "epoch": 0.815850021303792, + "grad_norm": 18.75, + "learning_rate": 1.6160805729381884e-06, + "loss": 0.897, + "step": 9574 + }, + { + "epoch": 0.815935236472092, + "grad_norm": 19.625, + "learning_rate": 1.6159199017726316e-06, + "loss": 0.5949, + "step": 9575 + }, + { + "epoch": 0.8160204516403919, + "grad_norm": 18.0, + "learning_rate": 1.615759223994534e-06, + "loss": 0.6339, + "step": 9576 + }, + { + "epoch": 0.8161056668086919, + "grad_norm": 16.75, + "learning_rate": 1.6155985396067997e-06, + "loss": 0.9377, + "step": 9577 + }, + { + "epoch": 0.8161908819769919, + "grad_norm": 19.0, + "learning_rate": 1.615437848612332e-06, + "loss": 0.7673, + "step": 9578 + }, + { + "epoch": 0.8162760971452918, + "grad_norm": 19.875, + "learning_rate": 1.6152771510140352e-06, + "loss": 0.9419, + "step": 9579 + }, + { + "epoch": 0.8163613123135918, + "grad_norm": 18.25, + "learning_rate": 1.6151164468148128e-06, + "loss": 1.1502, + "step": 9580 + }, + { + "epoch": 0.8164465274818917, + "grad_norm": 19.25, + "learning_rate": 1.6149557360175689e-06, + "loss": 0.889, + "step": 9581 + }, + { + "epoch": 0.8165317426501917, + "grad_norm": 12.6875, + "learning_rate": 1.6147950186252078e-06, + "loss": 0.6464, + "step": 9582 + }, + { + "epoch": 0.8166169578184916, + "grad_norm": 18.875, + "learning_rate": 1.614634294640634e-06, + "loss": 0.6655, + "step": 9583 + }, + { + "epoch": 0.8167021729867916, + "grad_norm": 11.3125, + "learning_rate": 1.614473564066752e-06, + "loss": 0.5105, + "step": 9584 + }, + { + "epoch": 0.8167873881550916, + "grad_norm": 15.0625, + "learning_rate": 1.614312826906466e-06, + "loss": 0.6503, + "step": 9585 + }, + { + "epoch": 0.8168726033233915, + "grad_norm": 11.8125, + "learning_rate": 1.6141520831626808e-06, + "loss": 0.4852, + "step": 9586 + }, + { + "epoch": 0.8169578184916915, + "grad_norm": 16.375, + "learning_rate": 1.6139913328383013e-06, + "loss": 0.559, + "step": 9587 + }, + { + "epoch": 0.8170430336599914, + "grad_norm": 20.5, + "learning_rate": 1.6138305759362325e-06, + "loss": 1.1046, + "step": 9588 + }, + { + "epoch": 0.8171282488282914, + "grad_norm": 16.625, + "learning_rate": 1.6136698124593795e-06, + "loss": 1.0394, + "step": 9589 + }, + { + "epoch": 0.8172134639965913, + "grad_norm": 18.0, + "learning_rate": 1.613509042410647e-06, + "loss": 0.971, + "step": 9590 + }, + { + "epoch": 0.8172986791648913, + "grad_norm": 12.1875, + "learning_rate": 1.613348265792941e-06, + "loss": 0.6667, + "step": 9591 + }, + { + "epoch": 0.8173838943331914, + "grad_norm": 19.0, + "learning_rate": 1.6131874826091655e-06, + "loss": 0.9719, + "step": 9592 + }, + { + "epoch": 0.8174691095014913, + "grad_norm": 15.4375, + "learning_rate": 1.613026692862228e-06, + "loss": 0.813, + "step": 9593 + }, + { + "epoch": 0.8175543246697913, + "grad_norm": 12.4375, + "learning_rate": 1.6128658965550323e-06, + "loss": 0.5615, + "step": 9594 + }, + { + "epoch": 0.8176395398380912, + "grad_norm": 25.375, + "learning_rate": 1.612705093690486e-06, + "loss": 1.2454, + "step": 9595 + }, + { + "epoch": 0.8177247550063912, + "grad_norm": 18.0, + "learning_rate": 1.6125442842714928e-06, + "loss": 0.8158, + "step": 9596 + }, + { + "epoch": 0.8178099701746911, + "grad_norm": 19.5, + "learning_rate": 1.6123834683009603e-06, + "loss": 0.8649, + "step": 9597 + }, + { + "epoch": 0.8178951853429911, + "grad_norm": 15.0625, + "learning_rate": 1.6122226457817941e-06, + "loss": 1.1615, + "step": 9598 + }, + { + "epoch": 0.8179804005112911, + "grad_norm": 19.0, + "learning_rate": 1.612061816716901e-06, + "loss": 0.2495, + "step": 9599 + }, + { + "epoch": 0.818065615679591, + "grad_norm": 15.0, + "learning_rate": 1.611900981109186e-06, + "loss": 1.3293, + "step": 9600 + }, + { + "epoch": 0.818150830847891, + "grad_norm": 13.0, + "learning_rate": 1.6117401389615568e-06, + "loss": 0.5224, + "step": 9601 + }, + { + "epoch": 0.8182360460161909, + "grad_norm": 16.25, + "learning_rate": 1.611579290276919e-06, + "loss": 0.9867, + "step": 9602 + }, + { + "epoch": 0.8183212611844909, + "grad_norm": 17.5, + "learning_rate": 1.6114184350581802e-06, + "loss": 0.7392, + "step": 9603 + }, + { + "epoch": 0.8184064763527908, + "grad_norm": 13.5625, + "learning_rate": 1.611257573308247e-06, + "loss": 0.9066, + "step": 9604 + }, + { + "epoch": 0.8184916915210908, + "grad_norm": 13.5, + "learning_rate": 1.611096705030026e-06, + "loss": 0.6662, + "step": 9605 + }, + { + "epoch": 0.8185769066893908, + "grad_norm": 14.3125, + "learning_rate": 1.6109358302264245e-06, + "loss": 0.409, + "step": 9606 + }, + { + "epoch": 0.8186621218576907, + "grad_norm": 10.375, + "learning_rate": 1.6107749489003493e-06, + "loss": 0.5051, + "step": 9607 + }, + { + "epoch": 0.8187473370259907, + "grad_norm": 17.5, + "learning_rate": 1.6106140610547083e-06, + "loss": 0.8585, + "step": 9608 + }, + { + "epoch": 0.8188325521942906, + "grad_norm": 14.1875, + "learning_rate": 1.6104531666924084e-06, + "loss": 0.6428, + "step": 9609 + }, + { + "epoch": 0.8189177673625906, + "grad_norm": 21.875, + "learning_rate": 1.6102922658163575e-06, + "loss": 0.8343, + "step": 9610 + }, + { + "epoch": 0.8190029825308905, + "grad_norm": 16.75, + "learning_rate": 1.6101313584294632e-06, + "loss": 0.9247, + "step": 9611 + }, + { + "epoch": 0.8190881976991905, + "grad_norm": 13.875, + "learning_rate": 1.6099704445346328e-06, + "loss": 0.626, + "step": 9612 + }, + { + "epoch": 0.8191734128674905, + "grad_norm": 18.875, + "learning_rate": 1.6098095241347744e-06, + "loss": 1.1083, + "step": 9613 + }, + { + "epoch": 0.8192586280357904, + "grad_norm": 18.25, + "learning_rate": 1.6096485972327962e-06, + "loss": 0.761, + "step": 9614 + }, + { + "epoch": 0.8193438432040904, + "grad_norm": 10.8125, + "learning_rate": 1.6094876638316064e-06, + "loss": 0.4208, + "step": 9615 + }, + { + "epoch": 0.8194290583723903, + "grad_norm": 12.1875, + "learning_rate": 1.609326723934113e-06, + "loss": 0.603, + "step": 9616 + }, + { + "epoch": 0.8195142735406903, + "grad_norm": 14.0625, + "learning_rate": 1.6091657775432243e-06, + "loss": 0.7158, + "step": 9617 + }, + { + "epoch": 0.8195994887089902, + "grad_norm": 14.1875, + "learning_rate": 1.609004824661849e-06, + "loss": 0.6577, + "step": 9618 + }, + { + "epoch": 0.8196847038772902, + "grad_norm": 15.5625, + "learning_rate": 1.608843865292895e-06, + "loss": 0.8638, + "step": 9619 + }, + { + "epoch": 0.8197699190455902, + "grad_norm": 19.75, + "learning_rate": 1.608682899439272e-06, + "loss": 1.058, + "step": 9620 + }, + { + "epoch": 0.8198551342138901, + "grad_norm": 10.5625, + "learning_rate": 1.608521927103888e-06, + "loss": 0.3731, + "step": 9621 + }, + { + "epoch": 0.8199403493821901, + "grad_norm": 13.0625, + "learning_rate": 1.6083609482896524e-06, + "loss": 0.6416, + "step": 9622 + }, + { + "epoch": 0.82002556455049, + "grad_norm": 13.875, + "learning_rate": 1.6081999629994743e-06, + "loss": 0.564, + "step": 9623 + }, + { + "epoch": 0.82011077971879, + "grad_norm": 17.75, + "learning_rate": 1.608038971236262e-06, + "loss": 0.8689, + "step": 9624 + }, + { + "epoch": 0.8201959948870899, + "grad_norm": 16.375, + "learning_rate": 1.6078779730029262e-06, + "loss": 0.5513, + "step": 9625 + }, + { + "epoch": 0.8202812100553899, + "grad_norm": 12.5, + "learning_rate": 1.6077169683023749e-06, + "loss": 0.6399, + "step": 9626 + }, + { + "epoch": 0.8203664252236899, + "grad_norm": 17.25, + "learning_rate": 1.6075559571375188e-06, + "loss": 0.8342, + "step": 9627 + }, + { + "epoch": 0.8204516403919898, + "grad_norm": 13.75, + "learning_rate": 1.607394939511267e-06, + "loss": 0.7329, + "step": 9628 + }, + { + "epoch": 0.8205368555602898, + "grad_norm": 13.75, + "learning_rate": 1.607233915426529e-06, + "loss": 0.6268, + "step": 9629 + }, + { + "epoch": 0.8206220707285897, + "grad_norm": 11.625, + "learning_rate": 1.607072884886215e-06, + "loss": 0.3986, + "step": 9630 + }, + { + "epoch": 0.8207072858968897, + "grad_norm": 12.3125, + "learning_rate": 1.606911847893235e-06, + "loss": 0.6255, + "step": 9631 + }, + { + "epoch": 0.8207925010651896, + "grad_norm": 14.125, + "learning_rate": 1.606750804450499e-06, + "loss": 0.7942, + "step": 9632 + }, + { + "epoch": 0.8208777162334896, + "grad_norm": 25.875, + "learning_rate": 1.6065897545609172e-06, + "loss": 0.9177, + "step": 9633 + }, + { + "epoch": 0.8209629314017896, + "grad_norm": 25.125, + "learning_rate": 1.6064286982273997e-06, + "loss": 1.2602, + "step": 9634 + }, + { + "epoch": 0.8210481465700895, + "grad_norm": 20.125, + "learning_rate": 1.6062676354528575e-06, + "loss": 1.3172, + "step": 9635 + }, + { + "epoch": 0.8211333617383895, + "grad_norm": 14.5, + "learning_rate": 1.6061065662402008e-06, + "loss": 0.5658, + "step": 9636 + }, + { + "epoch": 0.8212185769066894, + "grad_norm": 15.5625, + "learning_rate": 1.6059454905923405e-06, + "loss": 0.7155, + "step": 9637 + }, + { + "epoch": 0.8213037920749894, + "grad_norm": 14.0625, + "learning_rate": 1.6057844085121868e-06, + "loss": 0.7629, + "step": 9638 + }, + { + "epoch": 0.8213890072432893, + "grad_norm": 17.0, + "learning_rate": 1.6056233200026517e-06, + "loss": 0.7474, + "step": 9639 + }, + { + "epoch": 0.8214742224115893, + "grad_norm": 11.75, + "learning_rate": 1.6054622250666452e-06, + "loss": 0.5844, + "step": 9640 + }, + { + "epoch": 0.8215594375798893, + "grad_norm": 15.4375, + "learning_rate": 1.6053011237070786e-06, + "loss": 0.9581, + "step": 9641 + }, + { + "epoch": 0.8216446527481892, + "grad_norm": 12.125, + "learning_rate": 1.6051400159268638e-06, + "loss": 0.5039, + "step": 9642 + }, + { + "epoch": 0.8217298679164892, + "grad_norm": 12.25, + "learning_rate": 1.604978901728912e-06, + "loss": 0.4849, + "step": 9643 + }, + { + "epoch": 0.8218150830847891, + "grad_norm": 11.9375, + "learning_rate": 1.6048177811161342e-06, + "loss": 0.6487, + "step": 9644 + }, + { + "epoch": 0.8219002982530891, + "grad_norm": 12.75, + "learning_rate": 1.604656654091442e-06, + "loss": 0.7016, + "step": 9645 + }, + { + "epoch": 0.821985513421389, + "grad_norm": 13.6875, + "learning_rate": 1.6044955206577479e-06, + "loss": 0.8086, + "step": 9646 + }, + { + "epoch": 0.822070728589689, + "grad_norm": 12.125, + "learning_rate": 1.6043343808179628e-06, + "loss": 0.5771, + "step": 9647 + }, + { + "epoch": 0.822155943757989, + "grad_norm": 15.4375, + "learning_rate": 1.604173234575e-06, + "loss": 0.9748, + "step": 9648 + }, + { + "epoch": 0.8222411589262889, + "grad_norm": 16.125, + "learning_rate": 1.6040120819317698e-06, + "loss": 0.7285, + "step": 9649 + }, + { + "epoch": 0.8223263740945889, + "grad_norm": 14.375, + "learning_rate": 1.6038509228911855e-06, + "loss": 0.7807, + "step": 9650 + }, + { + "epoch": 0.8224115892628888, + "grad_norm": 15.9375, + "learning_rate": 1.6036897574561594e-06, + "loss": 1.0198, + "step": 9651 + }, + { + "epoch": 0.8224968044311888, + "grad_norm": 20.625, + "learning_rate": 1.603528585629604e-06, + "loss": 0.9035, + "step": 9652 + }, + { + "epoch": 0.8225820195994887, + "grad_norm": 15.9375, + "learning_rate": 1.603367407414431e-06, + "loss": 0.7171, + "step": 9653 + }, + { + "epoch": 0.8226672347677887, + "grad_norm": 16.5, + "learning_rate": 1.6032062228135536e-06, + "loss": 0.6194, + "step": 9654 + }, + { + "epoch": 0.8227524499360886, + "grad_norm": 17.25, + "learning_rate": 1.6030450318298847e-06, + "loss": 0.886, + "step": 9655 + }, + { + "epoch": 0.8228376651043886, + "grad_norm": 27.75, + "learning_rate": 1.6028838344663371e-06, + "loss": 1.0366, + "step": 9656 + }, + { + "epoch": 0.8229228802726886, + "grad_norm": 12.5, + "learning_rate": 1.6027226307258239e-06, + "loss": 0.4763, + "step": 9657 + }, + { + "epoch": 0.8230080954409885, + "grad_norm": 15.125, + "learning_rate": 1.6025614206112583e-06, + "loss": 0.5358, + "step": 9658 + }, + { + "epoch": 0.8230933106092885, + "grad_norm": 17.875, + "learning_rate": 1.6024002041255527e-06, + "loss": 1.0267, + "step": 9659 + }, + { + "epoch": 0.8231785257775884, + "grad_norm": 11.5, + "learning_rate": 1.6022389812716214e-06, + "loss": 0.4523, + "step": 9660 + }, + { + "epoch": 0.8232637409458884, + "grad_norm": 24.625, + "learning_rate": 1.6020777520523776e-06, + "loss": 0.8703, + "step": 9661 + }, + { + "epoch": 0.8233489561141883, + "grad_norm": 12.9375, + "learning_rate": 1.6019165164707342e-06, + "loss": 0.1767, + "step": 9662 + }, + { + "epoch": 0.8234341712824883, + "grad_norm": 13.25, + "learning_rate": 1.6017552745296061e-06, + "loss": 0.6279, + "step": 9663 + }, + { + "epoch": 0.8235193864507883, + "grad_norm": 15.875, + "learning_rate": 1.6015940262319064e-06, + "loss": 0.7645, + "step": 9664 + }, + { + "epoch": 0.8236046016190882, + "grad_norm": 11.5, + "learning_rate": 1.601432771580549e-06, + "loss": 0.4265, + "step": 9665 + }, + { + "epoch": 0.8236898167873882, + "grad_norm": 18.0, + "learning_rate": 1.6012715105784477e-06, + "loss": 0.6059, + "step": 9666 + }, + { + "epoch": 0.8237750319556881, + "grad_norm": 19.625, + "learning_rate": 1.6011102432285172e-06, + "loss": 0.8687, + "step": 9667 + }, + { + "epoch": 0.8238602471239881, + "grad_norm": 17.625, + "learning_rate": 1.6009489695336717e-06, + "loss": 0.9237, + "step": 9668 + }, + { + "epoch": 0.823945462292288, + "grad_norm": 14.3125, + "learning_rate": 1.6007876894968252e-06, + "loss": 0.7254, + "step": 9669 + }, + { + "epoch": 0.824030677460588, + "grad_norm": 14.9375, + "learning_rate": 1.6006264031208923e-06, + "loss": 0.8134, + "step": 9670 + }, + { + "epoch": 0.824115892628888, + "grad_norm": 11.1875, + "learning_rate": 1.600465110408788e-06, + "loss": 0.571, + "step": 9671 + }, + { + "epoch": 0.8242011077971879, + "grad_norm": 17.25, + "learning_rate": 1.6003038113634268e-06, + "loss": 0.892, + "step": 9672 + }, + { + "epoch": 0.8242863229654879, + "grad_norm": 17.375, + "learning_rate": 1.6001425059877235e-06, + "loss": 0.7139, + "step": 9673 + }, + { + "epoch": 0.8243715381337878, + "grad_norm": 14.125, + "learning_rate": 1.5999811942845924e-06, + "loss": 0.6816, + "step": 9674 + }, + { + "epoch": 0.8244567533020878, + "grad_norm": 12.6875, + "learning_rate": 1.59981987625695e-06, + "loss": 0.4999, + "step": 9675 + }, + { + "epoch": 0.8245419684703877, + "grad_norm": 37.75, + "learning_rate": 1.5996585519077102e-06, + "loss": 1.0457, + "step": 9676 + }, + { + "epoch": 0.8246271836386877, + "grad_norm": 21.375, + "learning_rate": 1.5994972212397885e-06, + "loss": 0.7863, + "step": 9677 + }, + { + "epoch": 0.8247123988069877, + "grad_norm": 14.5, + "learning_rate": 1.599335884256101e-06, + "loss": 0.5265, + "step": 9678 + }, + { + "epoch": 0.8247976139752876, + "grad_norm": 17.875, + "learning_rate": 1.5991745409595628e-06, + "loss": 0.689, + "step": 9679 + }, + { + "epoch": 0.8248828291435876, + "grad_norm": 13.3125, + "learning_rate": 1.5990131913530897e-06, + "loss": 0.6493, + "step": 9680 + }, + { + "epoch": 0.8249680443118875, + "grad_norm": 14.75, + "learning_rate": 1.5988518354395965e-06, + "loss": 0.5115, + "step": 9681 + }, + { + "epoch": 0.8250532594801875, + "grad_norm": 14.125, + "learning_rate": 1.5986904732220005e-06, + "loss": 0.8284, + "step": 9682 + }, + { + "epoch": 0.8251384746484874, + "grad_norm": 17.25, + "learning_rate": 1.5985291047032166e-06, + "loss": 0.928, + "step": 9683 + }, + { + "epoch": 0.8252236898167874, + "grad_norm": 21.875, + "learning_rate": 1.5983677298861622e-06, + "loss": 0.8874, + "step": 9684 + }, + { + "epoch": 0.8253089049850874, + "grad_norm": 13.5, + "learning_rate": 1.5982063487737514e-06, + "loss": 0.565, + "step": 9685 + }, + { + "epoch": 0.8253941201533873, + "grad_norm": 11.875, + "learning_rate": 1.5980449613689025e-06, + "loss": 0.4842, + "step": 9686 + }, + { + "epoch": 0.8254793353216873, + "grad_norm": 14.3125, + "learning_rate": 1.5978835676745309e-06, + "loss": 0.8157, + "step": 9687 + }, + { + "epoch": 0.8255645504899872, + "grad_norm": 20.375, + "learning_rate": 1.5977221676935536e-06, + "loss": 0.4956, + "step": 9688 + }, + { + "epoch": 0.8256497656582872, + "grad_norm": 18.125, + "learning_rate": 1.597560761428887e-06, + "loss": 0.79, + "step": 9689 + }, + { + "epoch": 0.8257349808265871, + "grad_norm": 13.5, + "learning_rate": 1.5973993488834476e-06, + "loss": 0.619, + "step": 9690 + }, + { + "epoch": 0.8258201959948871, + "grad_norm": 12.125, + "learning_rate": 1.597237930060153e-06, + "loss": 0.6217, + "step": 9691 + }, + { + "epoch": 0.825905411163187, + "grad_norm": 12.875, + "learning_rate": 1.5970765049619197e-06, + "loss": 0.6151, + "step": 9692 + }, + { + "epoch": 0.825990626331487, + "grad_norm": 12.875, + "learning_rate": 1.596915073591665e-06, + "loss": 0.6642, + "step": 9693 + }, + { + "epoch": 0.826075841499787, + "grad_norm": 10.5625, + "learning_rate": 1.596753635952306e-06, + "loss": 0.3249, + "step": 9694 + }, + { + "epoch": 0.8261610566680869, + "grad_norm": 13.375, + "learning_rate": 1.59659219204676e-06, + "loss": 0.968, + "step": 9695 + }, + { + "epoch": 0.8262462718363869, + "grad_norm": 17.5, + "learning_rate": 1.5964307418779445e-06, + "loss": 1.0474, + "step": 9696 + }, + { + "epoch": 0.8263314870046868, + "grad_norm": 18.25, + "learning_rate": 1.5962692854487777e-06, + "loss": 0.9244, + "step": 9697 + }, + { + "epoch": 0.8264167021729868, + "grad_norm": 14.8125, + "learning_rate": 1.5961078227621762e-06, + "loss": 0.9517, + "step": 9698 + }, + { + "epoch": 0.8265019173412868, + "grad_norm": 20.125, + "learning_rate": 1.5959463538210584e-06, + "loss": 0.89, + "step": 9699 + }, + { + "epoch": 0.8265871325095867, + "grad_norm": 10.25, + "learning_rate": 1.595784878628342e-06, + "loss": 0.2895, + "step": 9700 + }, + { + "epoch": 0.8266723476778867, + "grad_norm": 17.0, + "learning_rate": 1.5956233971869455e-06, + "loss": 0.8022, + "step": 9701 + }, + { + "epoch": 0.8267575628461866, + "grad_norm": 12.75, + "learning_rate": 1.5954619094997865e-06, + "loss": 0.7376, + "step": 9702 + }, + { + "epoch": 0.8268427780144866, + "grad_norm": 20.5, + "learning_rate": 1.5953004155697835e-06, + "loss": 1.0007, + "step": 9703 + }, + { + "epoch": 0.8269279931827865, + "grad_norm": 12.8125, + "learning_rate": 1.5951389153998542e-06, + "loss": 0.7068, + "step": 9704 + }, + { + "epoch": 0.8270132083510865, + "grad_norm": 12.9375, + "learning_rate": 1.5949774089929185e-06, + "loss": 0.643, + "step": 9705 + }, + { + "epoch": 0.8270984235193864, + "grad_norm": 17.875, + "learning_rate": 1.5948158963518937e-06, + "loss": 1.1723, + "step": 9706 + }, + { + "epoch": 0.8271836386876864, + "grad_norm": 22.875, + "learning_rate": 1.5946543774796988e-06, + "loss": 0.9569, + "step": 9707 + }, + { + "epoch": 0.8272688538559864, + "grad_norm": 16.0, + "learning_rate": 1.5944928523792527e-06, + "loss": 0.6086, + "step": 9708 + }, + { + "epoch": 0.8273540690242863, + "grad_norm": 17.125, + "learning_rate": 1.5943313210534744e-06, + "loss": 0.9288, + "step": 9709 + }, + { + "epoch": 0.8274392841925863, + "grad_norm": 12.75, + "learning_rate": 1.594169783505283e-06, + "loss": 0.4042, + "step": 9710 + }, + { + "epoch": 0.8275244993608862, + "grad_norm": 13.0625, + "learning_rate": 1.5940082397375972e-06, + "loss": 0.7382, + "step": 9711 + }, + { + "epoch": 0.8276097145291862, + "grad_norm": 12.6875, + "learning_rate": 1.5938466897533373e-06, + "loss": 0.6501, + "step": 9712 + }, + { + "epoch": 0.8276949296974861, + "grad_norm": 14.125, + "learning_rate": 1.5936851335554212e-06, + "loss": 0.442, + "step": 9713 + }, + { + "epoch": 0.8277801448657861, + "grad_norm": 16.875, + "learning_rate": 1.5935235711467696e-06, + "loss": 0.79, + "step": 9714 + }, + { + "epoch": 0.8278653600340861, + "grad_norm": 14.8125, + "learning_rate": 1.5933620025303012e-06, + "loss": 0.7064, + "step": 9715 + }, + { + "epoch": 0.827950575202386, + "grad_norm": 16.125, + "learning_rate": 1.5932004277089365e-06, + "loss": 0.8353, + "step": 9716 + }, + { + "epoch": 0.828035790370686, + "grad_norm": 29.875, + "learning_rate": 1.593038846685595e-06, + "loss": 1.2232, + "step": 9717 + }, + { + "epoch": 0.8281210055389859, + "grad_norm": 11.875, + "learning_rate": 1.5928772594631964e-06, + "loss": 0.5594, + "step": 9718 + }, + { + "epoch": 0.8282062207072859, + "grad_norm": 19.875, + "learning_rate": 1.5927156660446608e-06, + "loss": 0.6795, + "step": 9719 + }, + { + "epoch": 0.8282914358755858, + "grad_norm": 13.0625, + "learning_rate": 1.5925540664329086e-06, + "loss": 0.5582, + "step": 9720 + }, + { + "epoch": 0.8283766510438858, + "grad_norm": 11.8125, + "learning_rate": 1.5923924606308601e-06, + "loss": 0.3808, + "step": 9721 + }, + { + "epoch": 0.8284618662121858, + "grad_norm": 14.5, + "learning_rate": 1.5922308486414357e-06, + "loss": 0.6694, + "step": 9722 + }, + { + "epoch": 0.8285470813804857, + "grad_norm": 17.5, + "learning_rate": 1.5920692304675554e-06, + "loss": 0.8276, + "step": 9723 + }, + { + "epoch": 0.8286322965487857, + "grad_norm": 32.25, + "learning_rate": 1.5919076061121402e-06, + "loss": 0.861, + "step": 9724 + }, + { + "epoch": 0.8287175117170856, + "grad_norm": 20.875, + "learning_rate": 1.591745975578111e-06, + "loss": 0.6563, + "step": 9725 + }, + { + "epoch": 0.8288027268853856, + "grad_norm": 25.875, + "learning_rate": 1.5915843388683878e-06, + "loss": 1.0286, + "step": 9726 + }, + { + "epoch": 0.8288879420536855, + "grad_norm": 33.75, + "learning_rate": 1.5914226959858925e-06, + "loss": 1.3663, + "step": 9727 + }, + { + "epoch": 0.8289731572219855, + "grad_norm": 13.625, + "learning_rate": 1.5912610469335458e-06, + "loss": 0.7798, + "step": 9728 + }, + { + "epoch": 0.8290583723902855, + "grad_norm": 16.25, + "learning_rate": 1.5910993917142687e-06, + "loss": 0.9191, + "step": 9729 + }, + { + "epoch": 0.8291435875585854, + "grad_norm": 19.125, + "learning_rate": 1.5909377303309823e-06, + "loss": 1.1412, + "step": 9730 + }, + { + "epoch": 0.8292288027268854, + "grad_norm": 12.6875, + "learning_rate": 1.5907760627866087e-06, + "loss": 0.4271, + "step": 9731 + }, + { + "epoch": 0.8293140178951853, + "grad_norm": 14.375, + "learning_rate": 1.5906143890840687e-06, + "loss": 0.4894, + "step": 9732 + }, + { + "epoch": 0.8293992330634853, + "grad_norm": 10.625, + "learning_rate": 1.5904527092262846e-06, + "loss": 1.027, + "step": 9733 + }, + { + "epoch": 0.8294844482317852, + "grad_norm": 18.125, + "learning_rate": 1.5902910232161772e-06, + "loss": 0.9958, + "step": 9734 + }, + { + "epoch": 0.8295696634000852, + "grad_norm": 21.125, + "learning_rate": 1.590129331056669e-06, + "loss": 0.8219, + "step": 9735 + }, + { + "epoch": 0.8296548785683852, + "grad_norm": 20.25, + "learning_rate": 1.5899676327506816e-06, + "loss": 0.572, + "step": 9736 + }, + { + "epoch": 0.8297400937366851, + "grad_norm": 18.5, + "learning_rate": 1.5898059283011377e-06, + "loss": 0.9582, + "step": 9737 + }, + { + "epoch": 0.8298253089049851, + "grad_norm": 13.9375, + "learning_rate": 1.5896442177109584e-06, + "loss": 0.5103, + "step": 9738 + }, + { + "epoch": 0.829910524073285, + "grad_norm": 17.875, + "learning_rate": 1.589482500983067e-06, + "loss": 0.7345, + "step": 9739 + }, + { + "epoch": 0.829995739241585, + "grad_norm": 11.75, + "learning_rate": 1.5893207781203851e-06, + "loss": 0.58, + "step": 9740 + }, + { + "epoch": 0.8300809544098849, + "grad_norm": 14.8125, + "learning_rate": 1.5891590491258358e-06, + "loss": 0.8765, + "step": 9741 + }, + { + "epoch": 0.8301661695781849, + "grad_norm": 14.25, + "learning_rate": 1.5889973140023414e-06, + "loss": 0.6009, + "step": 9742 + }, + { + "epoch": 0.8302513847464849, + "grad_norm": 12.9375, + "learning_rate": 1.5888355727528243e-06, + "loss": 0.7213, + "step": 9743 + }, + { + "epoch": 0.8303365999147848, + "grad_norm": 12.4375, + "learning_rate": 1.5886738253802078e-06, + "loss": 0.5879, + "step": 9744 + }, + { + "epoch": 0.8304218150830848, + "grad_norm": 15.75, + "learning_rate": 1.588512071887415e-06, + "loss": 0.9383, + "step": 9745 + }, + { + "epoch": 0.8305070302513847, + "grad_norm": 17.0, + "learning_rate": 1.5883503122773686e-06, + "loss": 0.7263, + "step": 9746 + }, + { + "epoch": 0.8305922454196847, + "grad_norm": 15.5625, + "learning_rate": 1.5881885465529914e-06, + "loss": 0.9516, + "step": 9747 + }, + { + "epoch": 0.8306774605879846, + "grad_norm": 17.375, + "learning_rate": 1.5880267747172076e-06, + "loss": 0.5601, + "step": 9748 + }, + { + "epoch": 0.8307626757562846, + "grad_norm": 13.625, + "learning_rate": 1.5878649967729397e-06, + "loss": 0.7941, + "step": 9749 + }, + { + "epoch": 0.8308478909245846, + "grad_norm": 15.5625, + "learning_rate": 1.587703212723112e-06, + "loss": 0.7843, + "step": 9750 + }, + { + "epoch": 0.8309331060928845, + "grad_norm": 10.6875, + "learning_rate": 1.5875414225706471e-06, + "loss": 0.4843, + "step": 9751 + }, + { + "epoch": 0.8310183212611845, + "grad_norm": 20.5, + "learning_rate": 1.5873796263184698e-06, + "loss": 0.7809, + "step": 9752 + }, + { + "epoch": 0.8311035364294844, + "grad_norm": 12.375, + "learning_rate": 1.5872178239695029e-06, + "loss": 0.5725, + "step": 9753 + }, + { + "epoch": 0.8311887515977844, + "grad_norm": 17.125, + "learning_rate": 1.5870560155266711e-06, + "loss": 0.9388, + "step": 9754 + }, + { + "epoch": 0.8312739667660843, + "grad_norm": 22.75, + "learning_rate": 1.5868942009928983e-06, + "loss": 1.2478, + "step": 9755 + }, + { + "epoch": 0.8313591819343843, + "grad_norm": 14.25, + "learning_rate": 1.5867323803711081e-06, + "loss": 0.7398, + "step": 9756 + }, + { + "epoch": 0.8314443971026843, + "grad_norm": 17.5, + "learning_rate": 1.5865705536642257e-06, + "loss": 0.8815, + "step": 9757 + }, + { + "epoch": 0.8315296122709842, + "grad_norm": 24.375, + "learning_rate": 1.586408720875175e-06, + "loss": 1.1097, + "step": 9758 + }, + { + "epoch": 0.8316148274392842, + "grad_norm": 11.5625, + "learning_rate": 1.5862468820068796e-06, + "loss": 0.5781, + "step": 9759 + }, + { + "epoch": 0.8317000426075841, + "grad_norm": 14.625, + "learning_rate": 1.5860850370622654e-06, + "loss": 0.3199, + "step": 9760 + }, + { + "epoch": 0.8317852577758841, + "grad_norm": 10.875, + "learning_rate": 1.5859231860442568e-06, + "loss": 0.5379, + "step": 9761 + }, + { + "epoch": 0.831870472944184, + "grad_norm": 14.0, + "learning_rate": 1.5857613289557782e-06, + "loss": 0.5074, + "step": 9762 + }, + { + "epoch": 0.831955688112484, + "grad_norm": 18.875, + "learning_rate": 1.5855994657997547e-06, + "loss": 0.9243, + "step": 9763 + }, + { + "epoch": 0.832040903280784, + "grad_norm": 15.75, + "learning_rate": 1.5854375965791112e-06, + "loss": 0.9017, + "step": 9764 + }, + { + "epoch": 0.8321261184490839, + "grad_norm": 21.25, + "learning_rate": 1.5852757212967737e-06, + "loss": 0.9408, + "step": 9765 + }, + { + "epoch": 0.8322113336173839, + "grad_norm": 12.3125, + "learning_rate": 1.5851138399556664e-06, + "loss": 0.4767, + "step": 9766 + }, + { + "epoch": 0.8322965487856838, + "grad_norm": 10.5, + "learning_rate": 1.584951952558715e-06, + "loss": 0.475, + "step": 9767 + }, + { + "epoch": 0.8323817639539838, + "grad_norm": 18.125, + "learning_rate": 1.5847900591088449e-06, + "loss": 1.1705, + "step": 9768 + }, + { + "epoch": 0.8324669791222837, + "grad_norm": 13.9375, + "learning_rate": 1.5846281596089822e-06, + "loss": 0.7003, + "step": 9769 + }, + { + "epoch": 0.8325521942905837, + "grad_norm": 13.5, + "learning_rate": 1.5844662540620517e-06, + "loss": 0.7125, + "step": 9770 + }, + { + "epoch": 0.8326374094588836, + "grad_norm": 11.3125, + "learning_rate": 1.58430434247098e-06, + "loss": 0.3365, + "step": 9771 + }, + { + "epoch": 0.8327226246271836, + "grad_norm": 12.5, + "learning_rate": 1.5841424248386922e-06, + "loss": 0.6143, + "step": 9772 + }, + { + "epoch": 0.8328078397954836, + "grad_norm": 14.1875, + "learning_rate": 1.5839805011681154e-06, + "loss": 0.6069, + "step": 9773 + }, + { + "epoch": 0.8328930549637835, + "grad_norm": 21.25, + "learning_rate": 1.5838185714621745e-06, + "loss": 0.653, + "step": 9774 + }, + { + "epoch": 0.8329782701320835, + "grad_norm": 17.375, + "learning_rate": 1.5836566357237968e-06, + "loss": 0.8936, + "step": 9775 + }, + { + "epoch": 0.8330634853003834, + "grad_norm": 14.75, + "learning_rate": 1.5834946939559079e-06, + "loss": 0.5792, + "step": 9776 + }, + { + "epoch": 0.8331487004686834, + "grad_norm": 16.25, + "learning_rate": 1.5833327461614345e-06, + "loss": 1.1301, + "step": 9777 + }, + { + "epoch": 0.8332339156369833, + "grad_norm": 15.3125, + "learning_rate": 1.5831707923433034e-06, + "loss": 0.8825, + "step": 9778 + }, + { + "epoch": 0.8333191308052833, + "grad_norm": 15.4375, + "learning_rate": 1.5830088325044407e-06, + "loss": 0.7664, + "step": 9779 + }, + { + "epoch": 0.8334043459735833, + "grad_norm": 9.5625, + "learning_rate": 1.5828468666477738e-06, + "loss": 0.2657, + "step": 9780 + }, + { + "epoch": 0.8334895611418832, + "grad_norm": 13.25, + "learning_rate": 1.5826848947762292e-06, + "loss": 0.5132, + "step": 9781 + }, + { + "epoch": 0.8335747763101832, + "grad_norm": 11.1875, + "learning_rate": 1.5825229168927344e-06, + "loss": 0.4001, + "step": 9782 + }, + { + "epoch": 0.8336599914784831, + "grad_norm": 14.875, + "learning_rate": 1.5823609330002155e-06, + "loss": 0.7679, + "step": 9783 + }, + { + "epoch": 0.8337452066467831, + "grad_norm": 11.0, + "learning_rate": 1.5821989431016007e-06, + "loss": 0.5385, + "step": 9784 + }, + { + "epoch": 0.833830421815083, + "grad_norm": 12.5, + "learning_rate": 1.5820369471998167e-06, + "loss": 0.5615, + "step": 9785 + }, + { + "epoch": 0.833915636983383, + "grad_norm": 13.0, + "learning_rate": 1.5818749452977917e-06, + "loss": 0.6283, + "step": 9786 + }, + { + "epoch": 0.834000852151683, + "grad_norm": 19.375, + "learning_rate": 1.5817129373984522e-06, + "loss": 0.2593, + "step": 9787 + }, + { + "epoch": 0.8340860673199829, + "grad_norm": 13.75, + "learning_rate": 1.5815509235047266e-06, + "loss": 0.6131, + "step": 9788 + }, + { + "epoch": 0.8341712824882829, + "grad_norm": 16.5, + "learning_rate": 1.581388903619542e-06, + "loss": 0.664, + "step": 9789 + }, + { + "epoch": 0.8342564976565828, + "grad_norm": 24.25, + "learning_rate": 1.5812268777458275e-06, + "loss": 0.8757, + "step": 9790 + }, + { + "epoch": 0.8343417128248828, + "grad_norm": 20.0, + "learning_rate": 1.5810648458865096e-06, + "loss": 1.0316, + "step": 9791 + }, + { + "epoch": 0.8344269279931827, + "grad_norm": 12.5, + "learning_rate": 1.580902808044517e-06, + "loss": 0.5827, + "step": 9792 + }, + { + "epoch": 0.8345121431614827, + "grad_norm": 13.875, + "learning_rate": 1.5807407642227782e-06, + "loss": 0.6134, + "step": 9793 + }, + { + "epoch": 0.8345973583297827, + "grad_norm": 15.1875, + "learning_rate": 1.580578714424221e-06, + "loss": 0.9598, + "step": 9794 + }, + { + "epoch": 0.8346825734980826, + "grad_norm": 15.875, + "learning_rate": 1.580416658651774e-06, + "loss": 0.9506, + "step": 9795 + }, + { + "epoch": 0.8347677886663826, + "grad_norm": 9.5625, + "learning_rate": 1.5802545969083657e-06, + "loss": 0.363, + "step": 9796 + }, + { + "epoch": 0.8348530038346825, + "grad_norm": 18.625, + "learning_rate": 1.5800925291969248e-06, + "loss": 0.8018, + "step": 9797 + }, + { + "epoch": 0.8349382190029825, + "grad_norm": 16.5, + "learning_rate": 1.5799304555203804e-06, + "loss": 1.0559, + "step": 9798 + }, + { + "epoch": 0.8350234341712824, + "grad_norm": 19.125, + "learning_rate": 1.5797683758816606e-06, + "loss": 0.8637, + "step": 9799 + }, + { + "epoch": 0.8351086493395824, + "grad_norm": 14.75, + "learning_rate": 1.5796062902836942e-06, + "loss": 0.7914, + "step": 9800 + }, + { + "epoch": 0.8351938645078824, + "grad_norm": 9.0625, + "learning_rate": 1.5794441987294113e-06, + "loss": 0.5058, + "step": 9801 + }, + { + "epoch": 0.8352790796761823, + "grad_norm": 14.9375, + "learning_rate": 1.5792821012217401e-06, + "loss": 0.6316, + "step": 9802 + }, + { + "epoch": 0.8353642948444823, + "grad_norm": 14.6875, + "learning_rate": 1.5791199977636104e-06, + "loss": 0.7757, + "step": 9803 + }, + { + "epoch": 0.8354495100127822, + "grad_norm": 13.0, + "learning_rate": 1.5789578883579512e-06, + "loss": 0.4614, + "step": 9804 + }, + { + "epoch": 0.8355347251810822, + "grad_norm": 10.4375, + "learning_rate": 1.5787957730076922e-06, + "loss": 0.3779, + "step": 9805 + }, + { + "epoch": 0.8356199403493821, + "grad_norm": 12.8125, + "learning_rate": 1.578633651715763e-06, + "loss": 0.5543, + "step": 9806 + }, + { + "epoch": 0.8357051555176821, + "grad_norm": 13.875, + "learning_rate": 1.5784715244850935e-06, + "loss": 0.5584, + "step": 9807 + }, + { + "epoch": 0.835790370685982, + "grad_norm": 16.5, + "learning_rate": 1.5783093913186126e-06, + "loss": 1.0975, + "step": 9808 + }, + { + "epoch": 0.835875585854282, + "grad_norm": 39.5, + "learning_rate": 1.5781472522192514e-06, + "loss": 0.9335, + "step": 9809 + }, + { + "epoch": 0.835960801022582, + "grad_norm": 15.6875, + "learning_rate": 1.5779851071899394e-06, + "loss": 0.8782, + "step": 9810 + }, + { + "epoch": 0.8360460161908819, + "grad_norm": 12.625, + "learning_rate": 1.5778229562336064e-06, + "loss": 0.6672, + "step": 9811 + }, + { + "epoch": 0.8361312313591819, + "grad_norm": 12.625, + "learning_rate": 1.577660799353183e-06, + "loss": 0.4586, + "step": 9812 + }, + { + "epoch": 0.8362164465274818, + "grad_norm": 13.75, + "learning_rate": 1.5774986365515998e-06, + "loss": 0.6113, + "step": 9813 + }, + { + "epoch": 0.8363016616957818, + "grad_norm": 14.8125, + "learning_rate": 1.5773364678317865e-06, + "loss": 0.5669, + "step": 9814 + }, + { + "epoch": 0.8363868768640818, + "grad_norm": 12.9375, + "learning_rate": 1.577174293196674e-06, + "loss": 0.7405, + "step": 9815 + }, + { + "epoch": 0.8364720920323817, + "grad_norm": 15.5625, + "learning_rate": 1.5770121126491934e-06, + "loss": 0.9952, + "step": 9816 + }, + { + "epoch": 0.8365573072006818, + "grad_norm": 18.375, + "learning_rate": 1.576849926192275e-06, + "loss": 0.8584, + "step": 9817 + }, + { + "epoch": 0.8366425223689817, + "grad_norm": 19.125, + "learning_rate": 1.5766877338288496e-06, + "loss": 0.9105, + "step": 9818 + }, + { + "epoch": 0.8367277375372817, + "grad_norm": 13.3125, + "learning_rate": 1.5765255355618484e-06, + "loss": 0.5223, + "step": 9819 + }, + { + "epoch": 0.8368129527055816, + "grad_norm": 22.25, + "learning_rate": 1.5763633313942025e-06, + "loss": 1.3057, + "step": 9820 + }, + { + "epoch": 0.8368981678738816, + "grad_norm": 17.75, + "learning_rate": 1.5762011213288427e-06, + "loss": 1.035, + "step": 9821 + }, + { + "epoch": 0.8369833830421816, + "grad_norm": 14.6875, + "learning_rate": 1.5760389053687014e-06, + "loss": 0.3214, + "step": 9822 + }, + { + "epoch": 0.8370685982104815, + "grad_norm": 15.5, + "learning_rate": 1.5758766835167083e-06, + "loss": 0.4658, + "step": 9823 + }, + { + "epoch": 0.8371538133787815, + "grad_norm": 12.1875, + "learning_rate": 1.5757144557757965e-06, + "loss": 0.5965, + "step": 9824 + }, + { + "epoch": 0.8372390285470814, + "grad_norm": 20.25, + "learning_rate": 1.5755522221488964e-06, + "loss": 0.8638, + "step": 9825 + }, + { + "epoch": 0.8373242437153814, + "grad_norm": 12.0625, + "learning_rate": 1.5753899826389407e-06, + "loss": 0.6263, + "step": 9826 + }, + { + "epoch": 0.8374094588836813, + "grad_norm": 13.0625, + "learning_rate": 1.5752277372488605e-06, + "loss": 0.6863, + "step": 9827 + }, + { + "epoch": 0.8374946740519813, + "grad_norm": 21.875, + "learning_rate": 1.5750654859815884e-06, + "loss": 0.6349, + "step": 9828 + }, + { + "epoch": 0.8375798892202813, + "grad_norm": 14.4375, + "learning_rate": 1.5749032288400557e-06, + "loss": 0.6991, + "step": 9829 + }, + { + "epoch": 0.8376651043885812, + "grad_norm": 14.3125, + "learning_rate": 1.5747409658271954e-06, + "loss": 0.9973, + "step": 9830 + }, + { + "epoch": 0.8377503195568812, + "grad_norm": 18.875, + "learning_rate": 1.574578696945939e-06, + "loss": 0.7255, + "step": 9831 + }, + { + "epoch": 0.8378355347251811, + "grad_norm": 14.4375, + "learning_rate": 1.574416422199219e-06, + "loss": 0.8038, + "step": 9832 + }, + { + "epoch": 0.8379207498934811, + "grad_norm": 19.875, + "learning_rate": 1.574254141589968e-06, + "loss": 1.0336, + "step": 9833 + }, + { + "epoch": 0.838005965061781, + "grad_norm": 15.625, + "learning_rate": 1.574091855121119e-06, + "loss": 0.7533, + "step": 9834 + }, + { + "epoch": 0.838091180230081, + "grad_norm": 14.3125, + "learning_rate": 1.5739295627956045e-06, + "loss": 0.5521, + "step": 9835 + }, + { + "epoch": 0.838176395398381, + "grad_norm": 23.5, + "learning_rate": 1.5737672646163564e-06, + "loss": 0.8693, + "step": 9836 + }, + { + "epoch": 0.8382616105666809, + "grad_norm": 12.8125, + "learning_rate": 1.5736049605863084e-06, + "loss": 0.8098, + "step": 9837 + }, + { + "epoch": 0.8383468257349809, + "grad_norm": 13.75, + "learning_rate": 1.5734426507083933e-06, + "loss": 0.6288, + "step": 9838 + }, + { + "epoch": 0.8384320409032808, + "grad_norm": 17.5, + "learning_rate": 1.573280334985545e-06, + "loss": 0.9859, + "step": 9839 + }, + { + "epoch": 0.8385172560715808, + "grad_norm": 17.875, + "learning_rate": 1.5731180134206953e-06, + "loss": 0.8724, + "step": 9840 + }, + { + "epoch": 0.8386024712398807, + "grad_norm": 17.625, + "learning_rate": 1.5729556860167782e-06, + "loss": 0.9944, + "step": 9841 + }, + { + "epoch": 0.8386876864081807, + "grad_norm": 15.6875, + "learning_rate": 1.572793352776727e-06, + "loss": 0.6638, + "step": 9842 + }, + { + "epoch": 0.8387729015764807, + "grad_norm": 18.75, + "learning_rate": 1.5726310137034756e-06, + "loss": 1.1067, + "step": 9843 + }, + { + "epoch": 0.8388581167447806, + "grad_norm": 13.0, + "learning_rate": 1.5724686687999572e-06, + "loss": 0.5252, + "step": 9844 + }, + { + "epoch": 0.8389433319130806, + "grad_norm": 30.75, + "learning_rate": 1.5723063180691058e-06, + "loss": 0.9885, + "step": 9845 + }, + { + "epoch": 0.8390285470813805, + "grad_norm": 25.125, + "learning_rate": 1.5721439615138547e-06, + "loss": 0.8916, + "step": 9846 + }, + { + "epoch": 0.8391137622496805, + "grad_norm": 16.875, + "learning_rate": 1.571981599137139e-06, + "loss": 1.0991, + "step": 9847 + }, + { + "epoch": 0.8391989774179804, + "grad_norm": 15.75, + "learning_rate": 1.5718192309418918e-06, + "loss": 1.0177, + "step": 9848 + }, + { + "epoch": 0.8392841925862804, + "grad_norm": 14.1875, + "learning_rate": 1.5716568569310469e-06, + "loss": 0.704, + "step": 9849 + }, + { + "epoch": 0.8393694077545804, + "grad_norm": 15.4375, + "learning_rate": 1.5714944771075396e-06, + "loss": 0.897, + "step": 9850 + }, + { + "epoch": 0.8394546229228803, + "grad_norm": 19.5, + "learning_rate": 1.571332091474304e-06, + "loss": 1.0276, + "step": 9851 + }, + { + "epoch": 0.8395398380911803, + "grad_norm": 17.25, + "learning_rate": 1.571169700034274e-06, + "loss": 0.7457, + "step": 9852 + }, + { + "epoch": 0.8396250532594802, + "grad_norm": 36.75, + "learning_rate": 1.5710073027903848e-06, + "loss": 0.8826, + "step": 9853 + }, + { + "epoch": 0.8397102684277802, + "grad_norm": 18.0, + "learning_rate": 1.5708448997455709e-06, + "loss": 0.9323, + "step": 9854 + }, + { + "epoch": 0.8397954835960801, + "grad_norm": 20.75, + "learning_rate": 1.5706824909027667e-06, + "loss": 0.9471, + "step": 9855 + }, + { + "epoch": 0.8398806987643801, + "grad_norm": 9.75, + "learning_rate": 1.5705200762649079e-06, + "loss": 0.3089, + "step": 9856 + }, + { + "epoch": 0.83996591393268, + "grad_norm": 19.375, + "learning_rate": 1.5703576558349285e-06, + "loss": 0.7622, + "step": 9857 + }, + { + "epoch": 0.84005112910098, + "grad_norm": 14.0625, + "learning_rate": 1.5701952296157643e-06, + "loss": 0.7318, + "step": 9858 + }, + { + "epoch": 0.84013634426928, + "grad_norm": 18.0, + "learning_rate": 1.5700327976103504e-06, + "loss": 0.9117, + "step": 9859 + }, + { + "epoch": 0.8402215594375799, + "grad_norm": 12.875, + "learning_rate": 1.5698703598216223e-06, + "loss": 0.6294, + "step": 9860 + }, + { + "epoch": 0.8403067746058799, + "grad_norm": 18.0, + "learning_rate": 1.5697079162525148e-06, + "loss": 0.6508, + "step": 9861 + }, + { + "epoch": 0.8403919897741798, + "grad_norm": 10.125, + "learning_rate": 1.5695454669059637e-06, + "loss": 0.2965, + "step": 9862 + }, + { + "epoch": 0.8404772049424798, + "grad_norm": 14.3125, + "learning_rate": 1.5693830117849048e-06, + "loss": 0.7872, + "step": 9863 + }, + { + "epoch": 0.8405624201107798, + "grad_norm": 15.0625, + "learning_rate": 1.5692205508922736e-06, + "loss": 1.0676, + "step": 9864 + }, + { + "epoch": 0.8406476352790797, + "grad_norm": 19.5, + "learning_rate": 1.569058084231006e-06, + "loss": 1.0584, + "step": 9865 + }, + { + "epoch": 0.8407328504473797, + "grad_norm": 18.375, + "learning_rate": 1.5688956118040382e-06, + "loss": 0.7414, + "step": 9866 + }, + { + "epoch": 0.8408180656156796, + "grad_norm": 10.25, + "learning_rate": 1.5687331336143063e-06, + "loss": 0.3077, + "step": 9867 + }, + { + "epoch": 0.8409032807839796, + "grad_norm": 19.625, + "learning_rate": 1.5685706496647454e-06, + "loss": 0.7172, + "step": 9868 + }, + { + "epoch": 0.8409884959522795, + "grad_norm": 16.125, + "learning_rate": 1.5684081599582932e-06, + "loss": 0.802, + "step": 9869 + }, + { + "epoch": 0.8410737111205795, + "grad_norm": 21.75, + "learning_rate": 1.568245664497885e-06, + "loss": 1.1287, + "step": 9870 + }, + { + "epoch": 0.8411589262888794, + "grad_norm": 15.9375, + "learning_rate": 1.568083163286458e-06, + "loss": 0.8503, + "step": 9871 + }, + { + "epoch": 0.8412441414571794, + "grad_norm": 15.9375, + "learning_rate": 1.5679206563269478e-06, + "loss": 0.4853, + "step": 9872 + }, + { + "epoch": 0.8413293566254794, + "grad_norm": 13.3125, + "learning_rate": 1.567758143622292e-06, + "loss": 0.6183, + "step": 9873 + }, + { + "epoch": 0.8414145717937793, + "grad_norm": 15.0625, + "learning_rate": 1.5675956251754268e-06, + "loss": 0.6364, + "step": 9874 + }, + { + "epoch": 0.8414997869620793, + "grad_norm": 17.625, + "learning_rate": 1.5674331009892896e-06, + "loss": 0.4606, + "step": 9875 + }, + { + "epoch": 0.8415850021303792, + "grad_norm": 12.8125, + "learning_rate": 1.5672705710668168e-06, + "loss": 0.7307, + "step": 9876 + }, + { + "epoch": 0.8416702172986792, + "grad_norm": 19.375, + "learning_rate": 1.567108035410946e-06, + "loss": 0.7943, + "step": 9877 + }, + { + "epoch": 0.8417554324669791, + "grad_norm": 16.125, + "learning_rate": 1.5669454940246138e-06, + "loss": 0.9012, + "step": 9878 + }, + { + "epoch": 0.8418406476352791, + "grad_norm": 19.5, + "learning_rate": 1.566782946910758e-06, + "loss": 1.0703, + "step": 9879 + }, + { + "epoch": 0.8419258628035791, + "grad_norm": 11.1875, + "learning_rate": 1.5666203940723157e-06, + "loss": 0.4871, + "step": 9880 + }, + { + "epoch": 0.842011077971879, + "grad_norm": 14.5, + "learning_rate": 1.5664578355122245e-06, + "loss": 0.7322, + "step": 9881 + }, + { + "epoch": 0.842096293140179, + "grad_norm": 15.1875, + "learning_rate": 1.566295271233422e-06, + "loss": 0.7332, + "step": 9882 + }, + { + "epoch": 0.8421815083084789, + "grad_norm": 12.5, + "learning_rate": 1.566132701238846e-06, + "loss": 0.7518, + "step": 9883 + }, + { + "epoch": 0.8422667234767789, + "grad_norm": 13.6875, + "learning_rate": 1.5659701255314345e-06, + "loss": 0.5796, + "step": 9884 + }, + { + "epoch": 0.8423519386450788, + "grad_norm": 12.375, + "learning_rate": 1.5658075441141243e-06, + "loss": 0.4041, + "step": 9885 + }, + { + "epoch": 0.8424371538133788, + "grad_norm": 15.6875, + "learning_rate": 1.5656449569898547e-06, + "loss": 0.768, + "step": 9886 + }, + { + "epoch": 0.8425223689816788, + "grad_norm": 14.5625, + "learning_rate": 1.5654823641615636e-06, + "loss": 0.672, + "step": 9887 + }, + { + "epoch": 0.8426075841499787, + "grad_norm": 12.375, + "learning_rate": 1.5653197656321886e-06, + "loss": 0.6673, + "step": 9888 + }, + { + "epoch": 0.8426927993182787, + "grad_norm": 12.0, + "learning_rate": 1.565157161404668e-06, + "loss": 0.6916, + "step": 9889 + }, + { + "epoch": 0.8427780144865786, + "grad_norm": 13.0625, + "learning_rate": 1.564994551481941e-06, + "loss": 0.6301, + "step": 9890 + }, + { + "epoch": 0.8428632296548786, + "grad_norm": 10.5625, + "learning_rate": 1.5648319358669458e-06, + "loss": 0.3635, + "step": 9891 + }, + { + "epoch": 0.8429484448231785, + "grad_norm": 14.875, + "learning_rate": 1.5646693145626208e-06, + "loss": 0.8302, + "step": 9892 + }, + { + "epoch": 0.8430336599914785, + "grad_norm": 9.8125, + "learning_rate": 1.5645066875719048e-06, + "loss": 0.2821, + "step": 9893 + }, + { + "epoch": 0.8431188751597785, + "grad_norm": 20.5, + "learning_rate": 1.5643440548977367e-06, + "loss": 1.1424, + "step": 9894 + }, + { + "epoch": 0.8432040903280784, + "grad_norm": 13.125, + "learning_rate": 1.5641814165430552e-06, + "loss": 0.8648, + "step": 9895 + }, + { + "epoch": 0.8432893054963784, + "grad_norm": 15.8125, + "learning_rate": 1.5640187725108e-06, + "loss": 0.8644, + "step": 9896 + }, + { + "epoch": 0.8433745206646783, + "grad_norm": 19.125, + "learning_rate": 1.56385612280391e-06, + "loss": 0.9258, + "step": 9897 + }, + { + "epoch": 0.8434597358329783, + "grad_norm": 13.75, + "learning_rate": 1.5636934674253237e-06, + "loss": 0.5752, + "step": 9898 + }, + { + "epoch": 0.8435449510012782, + "grad_norm": 20.125, + "learning_rate": 1.563530806377981e-06, + "loss": 0.4719, + "step": 9899 + }, + { + "epoch": 0.8436301661695782, + "grad_norm": 13.125, + "learning_rate": 1.5633681396648214e-06, + "loss": 0.343, + "step": 9900 + }, + { + "epoch": 0.8437153813378782, + "grad_norm": 14.125, + "learning_rate": 1.5632054672887848e-06, + "loss": 0.815, + "step": 9901 + }, + { + "epoch": 0.8438005965061781, + "grad_norm": 17.25, + "learning_rate": 1.56304278925281e-06, + "loss": 0.9937, + "step": 9902 + }, + { + "epoch": 0.8438858116744781, + "grad_norm": 14.0, + "learning_rate": 1.5628801055598376e-06, + "loss": 0.6641, + "step": 9903 + }, + { + "epoch": 0.843971026842778, + "grad_norm": 14.0, + "learning_rate": 1.5627174162128066e-06, + "loss": 0.9075, + "step": 9904 + }, + { + "epoch": 0.844056242011078, + "grad_norm": 10.25, + "learning_rate": 1.562554721214658e-06, + "loss": 0.5418, + "step": 9905 + }, + { + "epoch": 0.8441414571793779, + "grad_norm": 20.625, + "learning_rate": 1.5623920205683307e-06, + "loss": 0.8552, + "step": 9906 + }, + { + "epoch": 0.8442266723476779, + "grad_norm": 34.5, + "learning_rate": 1.5622293142767658e-06, + "loss": 0.9319, + "step": 9907 + }, + { + "epoch": 0.8443118875159779, + "grad_norm": 13.3125, + "learning_rate": 1.5620666023429035e-06, + "loss": 0.6535, + "step": 9908 + }, + { + "epoch": 0.8443971026842778, + "grad_norm": 17.375, + "learning_rate": 1.5619038847696835e-06, + "loss": 0.6715, + "step": 9909 + }, + { + "epoch": 0.8444823178525778, + "grad_norm": 18.625, + "learning_rate": 1.5617411615600465e-06, + "loss": 0.7355, + "step": 9910 + }, + { + "epoch": 0.8445675330208777, + "grad_norm": 17.5, + "learning_rate": 1.5615784327169337e-06, + "loss": 0.9041, + "step": 9911 + }, + { + "epoch": 0.8446527481891777, + "grad_norm": 13.4375, + "learning_rate": 1.5614156982432848e-06, + "loss": 0.6214, + "step": 9912 + }, + { + "epoch": 0.8447379633574776, + "grad_norm": 14.3125, + "learning_rate": 1.5612529581420415e-06, + "loss": 0.7267, + "step": 9913 + }, + { + "epoch": 0.8448231785257776, + "grad_norm": 11.375, + "learning_rate": 1.561090212416144e-06, + "loss": 0.565, + "step": 9914 + }, + { + "epoch": 0.8449083936940776, + "grad_norm": 13.75, + "learning_rate": 1.5609274610685338e-06, + "loss": 0.5954, + "step": 9915 + }, + { + "epoch": 0.8449936088623775, + "grad_norm": 14.375, + "learning_rate": 1.5607647041021516e-06, + "loss": 0.7366, + "step": 9916 + }, + { + "epoch": 0.8450788240306775, + "grad_norm": 13.4375, + "learning_rate": 1.5606019415199383e-06, + "loss": 0.7262, + "step": 9917 + }, + { + "epoch": 0.8451640391989774, + "grad_norm": 15.6875, + "learning_rate": 1.5604391733248361e-06, + "loss": 0.96, + "step": 9918 + }, + { + "epoch": 0.8452492543672774, + "grad_norm": 11.625, + "learning_rate": 1.5602763995197853e-06, + "loss": 0.6817, + "step": 9919 + }, + { + "epoch": 0.8453344695355773, + "grad_norm": 16.25, + "learning_rate": 1.5601136201077288e-06, + "loss": 0.9011, + "step": 9920 + }, + { + "epoch": 0.8454196847038773, + "grad_norm": 19.125, + "learning_rate": 1.5599508350916068e-06, + "loss": 0.7238, + "step": 9921 + }, + { + "epoch": 0.8455048998721773, + "grad_norm": 15.6875, + "learning_rate": 1.5597880444743617e-06, + "loss": 0.6033, + "step": 9922 + }, + { + "epoch": 0.8455901150404772, + "grad_norm": 14.0, + "learning_rate": 1.5596252482589348e-06, + "loss": 0.4132, + "step": 9923 + }, + { + "epoch": 0.8456753302087772, + "grad_norm": 12.0625, + "learning_rate": 1.5594624464482689e-06, + "loss": 0.4342, + "step": 9924 + }, + { + "epoch": 0.8457605453770771, + "grad_norm": 15.1875, + "learning_rate": 1.5592996390453047e-06, + "loss": 0.6257, + "step": 9925 + }, + { + "epoch": 0.8458457605453771, + "grad_norm": 12.9375, + "learning_rate": 1.5591368260529855e-06, + "loss": 0.7901, + "step": 9926 + }, + { + "epoch": 0.845930975713677, + "grad_norm": 12.625, + "learning_rate": 1.5589740074742525e-06, + "loss": 0.5534, + "step": 9927 + }, + { + "epoch": 0.846016190881977, + "grad_norm": 24.875, + "learning_rate": 1.5588111833120493e-06, + "loss": 0.8871, + "step": 9928 + }, + { + "epoch": 0.846101406050277, + "grad_norm": 16.625, + "learning_rate": 1.5586483535693166e-06, + "loss": 0.8796, + "step": 9929 + }, + { + "epoch": 0.8461866212185769, + "grad_norm": 17.125, + "learning_rate": 1.5584855182489986e-06, + "loss": 1.4636, + "step": 9930 + }, + { + "epoch": 0.8462718363868769, + "grad_norm": 12.9375, + "learning_rate": 1.5583226773540364e-06, + "loss": 0.6934, + "step": 9931 + }, + { + "epoch": 0.8463570515551768, + "grad_norm": 18.25, + "learning_rate": 1.5581598308873735e-06, + "loss": 1.0093, + "step": 9932 + }, + { + "epoch": 0.8464422667234768, + "grad_norm": 22.375, + "learning_rate": 1.5579969788519531e-06, + "loss": 0.9159, + "step": 9933 + }, + { + "epoch": 0.8465274818917767, + "grad_norm": 12.3125, + "learning_rate": 1.557834121250717e-06, + "loss": 0.6586, + "step": 9934 + }, + { + "epoch": 0.8466126970600767, + "grad_norm": 12.1875, + "learning_rate": 1.5576712580866088e-06, + "loss": 0.6704, + "step": 9935 + }, + { + "epoch": 0.8466979122283766, + "grad_norm": 15.0, + "learning_rate": 1.557508389362572e-06, + "loss": 0.9263, + "step": 9936 + }, + { + "epoch": 0.8467831273966766, + "grad_norm": 13.1875, + "learning_rate": 1.557345515081549e-06, + "loss": 0.893, + "step": 9937 + }, + { + "epoch": 0.8468683425649766, + "grad_norm": 15.1875, + "learning_rate": 1.5571826352464835e-06, + "loss": 0.722, + "step": 9938 + }, + { + "epoch": 0.8469535577332765, + "grad_norm": 25.125, + "learning_rate": 1.5570197498603193e-06, + "loss": 0.9335, + "step": 9939 + }, + { + "epoch": 0.8470387729015765, + "grad_norm": 20.625, + "learning_rate": 1.5568568589259991e-06, + "loss": 0.9344, + "step": 9940 + }, + { + "epoch": 0.8471239880698764, + "grad_norm": 19.0, + "learning_rate": 1.5566939624464672e-06, + "loss": 0.4741, + "step": 9941 + }, + { + "epoch": 0.8472092032381764, + "grad_norm": 11.5, + "learning_rate": 1.5565310604246664e-06, + "loss": 0.5916, + "step": 9942 + }, + { + "epoch": 0.8472944184064763, + "grad_norm": 15.0625, + "learning_rate": 1.5563681528635415e-06, + "loss": 0.8187, + "step": 9943 + }, + { + "epoch": 0.8473796335747763, + "grad_norm": 12.8125, + "learning_rate": 1.5562052397660363e-06, + "loss": 0.652, + "step": 9944 + }, + { + "epoch": 0.8474648487430763, + "grad_norm": 14.25, + "learning_rate": 1.5560423211350942e-06, + "loss": 0.5854, + "step": 9945 + }, + { + "epoch": 0.8475500639113762, + "grad_norm": 18.875, + "learning_rate": 1.5558793969736593e-06, + "loss": 0.8579, + "step": 9946 + }, + { + "epoch": 0.8476352790796762, + "grad_norm": 13.1875, + "learning_rate": 1.5557164672846765e-06, + "loss": 0.4096, + "step": 9947 + }, + { + "epoch": 0.8477204942479761, + "grad_norm": 15.0, + "learning_rate": 1.5555535320710895e-06, + "loss": 0.6412, + "step": 9948 + }, + { + "epoch": 0.8478057094162761, + "grad_norm": 16.375, + "learning_rate": 1.555390591335843e-06, + "loss": 0.624, + "step": 9949 + }, + { + "epoch": 0.847890924584576, + "grad_norm": 15.6875, + "learning_rate": 1.5552276450818816e-06, + "loss": 0.8301, + "step": 9950 + }, + { + "epoch": 0.847976139752876, + "grad_norm": 19.625, + "learning_rate": 1.5550646933121497e-06, + "loss": 1.0591, + "step": 9951 + }, + { + "epoch": 0.848061354921176, + "grad_norm": 11.375, + "learning_rate": 1.554901736029592e-06, + "loss": 0.3938, + "step": 9952 + }, + { + "epoch": 0.8481465700894759, + "grad_norm": 21.0, + "learning_rate": 1.554738773237153e-06, + "loss": 0.8778, + "step": 9953 + }, + { + "epoch": 0.8482317852577759, + "grad_norm": 14.3125, + "learning_rate": 1.5545758049377781e-06, + "loss": 0.5937, + "step": 9954 + }, + { + "epoch": 0.8483170004260758, + "grad_norm": 22.375, + "learning_rate": 1.554412831134412e-06, + "loss": 0.9817, + "step": 9955 + }, + { + "epoch": 0.8484022155943758, + "grad_norm": 23.375, + "learning_rate": 1.5542498518300006e-06, + "loss": 1.1974, + "step": 9956 + }, + { + "epoch": 0.8484874307626757, + "grad_norm": 11.375, + "learning_rate": 1.5540868670274879e-06, + "loss": 0.502, + "step": 9957 + }, + { + "epoch": 0.8485726459309757, + "grad_norm": 11.5, + "learning_rate": 1.5539238767298197e-06, + "loss": 0.3593, + "step": 9958 + }, + { + "epoch": 0.8486578610992757, + "grad_norm": 23.5, + "learning_rate": 1.5537608809399413e-06, + "loss": 1.1276, + "step": 9959 + }, + { + "epoch": 0.8487430762675756, + "grad_norm": 13.1875, + "learning_rate": 1.5535978796607989e-06, + "loss": 0.7329, + "step": 9960 + }, + { + "epoch": 0.8488282914358756, + "grad_norm": 9.25, + "learning_rate": 1.5534348728953368e-06, + "loss": 0.3222, + "step": 9961 + }, + { + "epoch": 0.8489135066041755, + "grad_norm": 15.0, + "learning_rate": 1.553271860646502e-06, + "loss": 0.8051, + "step": 9962 + }, + { + "epoch": 0.8489987217724755, + "grad_norm": 17.125, + "learning_rate": 1.5531088429172392e-06, + "loss": 0.8001, + "step": 9963 + }, + { + "epoch": 0.8490839369407754, + "grad_norm": 16.125, + "learning_rate": 1.5529458197104951e-06, + "loss": 0.7866, + "step": 9964 + }, + { + "epoch": 0.8491691521090754, + "grad_norm": 35.75, + "learning_rate": 1.5527827910292156e-06, + "loss": 0.697, + "step": 9965 + }, + { + "epoch": 0.8492543672773754, + "grad_norm": 32.0, + "learning_rate": 1.5526197568763465e-06, + "loss": 1.2927, + "step": 9966 + }, + { + "epoch": 0.8493395824456753, + "grad_norm": 12.0, + "learning_rate": 1.5524567172548339e-06, + "loss": 0.6935, + "step": 9967 + }, + { + "epoch": 0.8494247976139753, + "grad_norm": 13.0, + "learning_rate": 1.5522936721676243e-06, + "loss": 0.7174, + "step": 9968 + }, + { + "epoch": 0.8495100127822752, + "grad_norm": 15.875, + "learning_rate": 1.5521306216176644e-06, + "loss": 0.6199, + "step": 9969 + }, + { + "epoch": 0.8495952279505752, + "grad_norm": 15.6875, + "learning_rate": 1.5519675656078998e-06, + "loss": 0.7787, + "step": 9970 + }, + { + "epoch": 0.8496804431188751, + "grad_norm": 17.375, + "learning_rate": 1.551804504141278e-06, + "loss": 0.8508, + "step": 9971 + }, + { + "epoch": 0.8497656582871751, + "grad_norm": 15.1875, + "learning_rate": 1.551641437220745e-06, + "loss": 0.884, + "step": 9972 + }, + { + "epoch": 0.849850873455475, + "grad_norm": 16.625, + "learning_rate": 1.5514783648492485e-06, + "loss": 0.8434, + "step": 9973 + }, + { + "epoch": 0.849936088623775, + "grad_norm": 13.5, + "learning_rate": 1.551315287029734e-06, + "loss": 0.6282, + "step": 9974 + }, + { + "epoch": 0.850021303792075, + "grad_norm": 12.25, + "learning_rate": 1.5511522037651499e-06, + "loss": 0.6725, + "step": 9975 + }, + { + "epoch": 0.8501065189603749, + "grad_norm": 10.6875, + "learning_rate": 1.5509891150584422e-06, + "loss": 0.341, + "step": 9976 + }, + { + "epoch": 0.8501917341286749, + "grad_norm": 10.25, + "learning_rate": 1.5508260209125592e-06, + "loss": 0.3871, + "step": 9977 + }, + { + "epoch": 0.8502769492969748, + "grad_norm": 14.6875, + "learning_rate": 1.550662921330447e-06, + "loss": 0.6555, + "step": 9978 + }, + { + "epoch": 0.8503621644652748, + "grad_norm": 20.25, + "learning_rate": 1.5504998163150533e-06, + "loss": 1.0452, + "step": 9979 + }, + { + "epoch": 0.8504473796335748, + "grad_norm": 11.8125, + "learning_rate": 1.5503367058693259e-06, + "loss": 0.5478, + "step": 9980 + }, + { + "epoch": 0.8505325948018747, + "grad_norm": 18.25, + "learning_rate": 1.5501735899962122e-06, + "loss": 0.6974, + "step": 9981 + }, + { + "epoch": 0.8506178099701747, + "grad_norm": 13.625, + "learning_rate": 1.5500104686986604e-06, + "loss": 0.7311, + "step": 9982 + }, + { + "epoch": 0.8507030251384746, + "grad_norm": 35.25, + "learning_rate": 1.5498473419796172e-06, + "loss": 0.8109, + "step": 9983 + }, + { + "epoch": 0.8507882403067746, + "grad_norm": 19.25, + "learning_rate": 1.549684209842031e-06, + "loss": 0.922, + "step": 9984 + }, + { + "epoch": 0.8508734554750745, + "grad_norm": 13.9375, + "learning_rate": 1.5495210722888498e-06, + "loss": 0.7558, + "step": 9985 + }, + { + "epoch": 0.8509586706433745, + "grad_norm": 14.9375, + "learning_rate": 1.549357929323022e-06, + "loss": 0.8149, + "step": 9986 + }, + { + "epoch": 0.8510438858116744, + "grad_norm": 26.25, + "learning_rate": 1.549194780947495e-06, + "loss": 0.4548, + "step": 9987 + }, + { + "epoch": 0.8511291009799744, + "grad_norm": 23.125, + "learning_rate": 1.5490316271652176e-06, + "loss": 1.4007, + "step": 9988 + }, + { + "epoch": 0.8512143161482744, + "grad_norm": 16.875, + "learning_rate": 1.548868467979138e-06, + "loss": 0.6879, + "step": 9989 + }, + { + "epoch": 0.8512995313165743, + "grad_norm": 15.0, + "learning_rate": 1.5487053033922048e-06, + "loss": 0.7199, + "step": 9990 + }, + { + "epoch": 0.8513847464848743, + "grad_norm": 12.5, + "learning_rate": 1.5485421334073658e-06, + "loss": 0.7627, + "step": 9991 + }, + { + "epoch": 0.8514699616531742, + "grad_norm": 12.625, + "learning_rate": 1.5483789580275707e-06, + "loss": 0.903, + "step": 9992 + }, + { + "epoch": 0.8515551768214742, + "grad_norm": 15.375, + "learning_rate": 1.548215777255768e-06, + "loss": 1.1291, + "step": 9993 + }, + { + "epoch": 0.8516403919897741, + "grad_norm": 12.5625, + "learning_rate": 1.548052591094906e-06, + "loss": 0.6761, + "step": 9994 + }, + { + "epoch": 0.8517256071580741, + "grad_norm": 12.5625, + "learning_rate": 1.5478893995479338e-06, + "loss": 0.4795, + "step": 9995 + }, + { + "epoch": 0.8518108223263741, + "grad_norm": 18.5, + "learning_rate": 1.5477262026178008e-06, + "loss": 0.7059, + "step": 9996 + }, + { + "epoch": 0.851896037494674, + "grad_norm": 19.375, + "learning_rate": 1.5475630003074563e-06, + "loss": 0.8677, + "step": 9997 + }, + { + "epoch": 0.851981252662974, + "grad_norm": 10.1875, + "learning_rate": 1.5473997926198488e-06, + "loss": 0.3925, + "step": 9998 + }, + { + "epoch": 0.8520664678312739, + "grad_norm": 10.375, + "learning_rate": 1.5472365795579277e-06, + "loss": 0.421, + "step": 9999 + }, + { + "epoch": 0.8521516829995739, + "grad_norm": 12.5, + "learning_rate": 1.5470733611246432e-06, + "loss": 0.4204, + "step": 10000 + }, + { + "epoch": 0.8522368981678738, + "grad_norm": 13.4375, + "learning_rate": 1.5469101373229442e-06, + "loss": 0.6935, + "step": 10001 + }, + { + "epoch": 0.8523221133361738, + "grad_norm": 16.125, + "learning_rate": 1.54674690815578e-06, + "loss": 0.9133, + "step": 10002 + }, + { + "epoch": 0.8524073285044738, + "grad_norm": 13.0, + "learning_rate": 1.5465836736261015e-06, + "loss": 0.982, + "step": 10003 + }, + { + "epoch": 0.8524925436727737, + "grad_norm": 16.125, + "learning_rate": 1.546420433736857e-06, + "loss": 0.79, + "step": 10004 + }, + { + "epoch": 0.8525777588410737, + "grad_norm": 18.375, + "learning_rate": 1.5462571884909984e-06, + "loss": 0.2605, + "step": 10005 + }, + { + "epoch": 0.8526629740093736, + "grad_norm": 47.0, + "learning_rate": 1.5460939378914732e-06, + "loss": 1.1058, + "step": 10006 + }, + { + "epoch": 0.8527481891776736, + "grad_norm": 16.25, + "learning_rate": 1.5459306819412337e-06, + "loss": 0.7308, + "step": 10007 + }, + { + "epoch": 0.8528334043459735, + "grad_norm": 17.375, + "learning_rate": 1.5457674206432286e-06, + "loss": 0.2821, + "step": 10008 + }, + { + "epoch": 0.8529186195142735, + "grad_norm": 12.0, + "learning_rate": 1.5456041540004094e-06, + "loss": 0.7543, + "step": 10009 + }, + { + "epoch": 0.8530038346825735, + "grad_norm": 15.625, + "learning_rate": 1.545440882015725e-06, + "loss": 0.8549, + "step": 10010 + }, + { + "epoch": 0.8530890498508734, + "grad_norm": 15.125, + "learning_rate": 1.5452776046921275e-06, + "loss": 0.8032, + "step": 10011 + }, + { + "epoch": 0.8531742650191734, + "grad_norm": 11.625, + "learning_rate": 1.5451143220325664e-06, + "loss": 0.5344, + "step": 10012 + }, + { + "epoch": 0.8532594801874733, + "grad_norm": 16.125, + "learning_rate": 1.5449510340399932e-06, + "loss": 0.6638, + "step": 10013 + }, + { + "epoch": 0.8533446953557733, + "grad_norm": 16.25, + "learning_rate": 1.5447877407173577e-06, + "loss": 0.882, + "step": 10014 + }, + { + "epoch": 0.8534299105240732, + "grad_norm": 15.75, + "learning_rate": 1.5446244420676116e-06, + "loss": 0.8087, + "step": 10015 + }, + { + "epoch": 0.8535151256923732, + "grad_norm": 11.375, + "learning_rate": 1.5444611380937052e-06, + "loss": 0.4129, + "step": 10016 + }, + { + "epoch": 0.8536003408606732, + "grad_norm": 20.0, + "learning_rate": 1.5442978287985902e-06, + "loss": 0.9496, + "step": 10017 + }, + { + "epoch": 0.8536855560289731, + "grad_norm": 16.25, + "learning_rate": 1.5441345141852177e-06, + "loss": 1.0274, + "step": 10018 + }, + { + "epoch": 0.8537707711972731, + "grad_norm": 9.1875, + "learning_rate": 1.5439711942565384e-06, + "loss": 0.3175, + "step": 10019 + }, + { + "epoch": 0.853855986365573, + "grad_norm": 19.5, + "learning_rate": 1.5438078690155043e-06, + "loss": 0.7375, + "step": 10020 + }, + { + "epoch": 0.853941201533873, + "grad_norm": 18.375, + "learning_rate": 1.5436445384650663e-06, + "loss": 0.9234, + "step": 10021 + }, + { + "epoch": 0.8540264167021729, + "grad_norm": 15.1875, + "learning_rate": 1.5434812026081763e-06, + "loss": 0.6828, + "step": 10022 + }, + { + "epoch": 0.8541116318704729, + "grad_norm": 11.625, + "learning_rate": 1.5433178614477857e-06, + "loss": 0.4363, + "step": 10023 + }, + { + "epoch": 0.8541968470387729, + "grad_norm": 13.1875, + "learning_rate": 1.5431545149868466e-06, + "loss": 0.408, + "step": 10024 + }, + { + "epoch": 0.8542820622070728, + "grad_norm": 14.0, + "learning_rate": 1.5429911632283106e-06, + "loss": 0.7951, + "step": 10025 + }, + { + "epoch": 0.8543672773753728, + "grad_norm": 16.125, + "learning_rate": 1.5428278061751297e-06, + "loss": 0.826, + "step": 10026 + }, + { + "epoch": 0.8544524925436727, + "grad_norm": 19.625, + "learning_rate": 1.5426644438302557e-06, + "loss": 0.8248, + "step": 10027 + }, + { + "epoch": 0.8545377077119727, + "grad_norm": 14.4375, + "learning_rate": 1.5425010761966412e-06, + "loss": 0.9161, + "step": 10028 + }, + { + "epoch": 0.8546229228802726, + "grad_norm": 13.5625, + "learning_rate": 1.5423377032772376e-06, + "loss": 0.6557, + "step": 10029 + }, + { + "epoch": 0.8547081380485726, + "grad_norm": 20.625, + "learning_rate": 1.5421743250749984e-06, + "loss": 1.3825, + "step": 10030 + }, + { + "epoch": 0.8547933532168726, + "grad_norm": 11.5, + "learning_rate": 1.542010941592875e-06, + "loss": 0.2564, + "step": 10031 + }, + { + "epoch": 0.8548785683851725, + "grad_norm": 14.0, + "learning_rate": 1.5418475528338204e-06, + "loss": 0.552, + "step": 10032 + }, + { + "epoch": 0.8549637835534725, + "grad_norm": 13.25, + "learning_rate": 1.5416841588007863e-06, + "loss": 0.5562, + "step": 10033 + }, + { + "epoch": 0.8550489987217724, + "grad_norm": 14.5, + "learning_rate": 1.5415207594967272e-06, + "loss": 0.5113, + "step": 10034 + }, + { + "epoch": 0.8551342138900724, + "grad_norm": 12.5625, + "learning_rate": 1.5413573549245941e-06, + "loss": 0.6859, + "step": 10035 + }, + { + "epoch": 0.8552194290583723, + "grad_norm": 12.375, + "learning_rate": 1.541193945087341e-06, + "loss": 0.685, + "step": 10036 + }, + { + "epoch": 0.8553046442266723, + "grad_norm": 16.375, + "learning_rate": 1.5410305299879202e-06, + "loss": 0.8904, + "step": 10037 + }, + { + "epoch": 0.8553898593949723, + "grad_norm": 11.9375, + "learning_rate": 1.5408671096292855e-06, + "loss": 0.4668, + "step": 10038 + }, + { + "epoch": 0.8554750745632722, + "grad_norm": 13.5, + "learning_rate": 1.5407036840143893e-06, + "loss": 0.843, + "step": 10039 + }, + { + "epoch": 0.8555602897315722, + "grad_norm": 13.5, + "learning_rate": 1.5405402531461853e-06, + "loss": 0.4669, + "step": 10040 + }, + { + "epoch": 0.8556455048998721, + "grad_norm": 14.6875, + "learning_rate": 1.5403768170276268e-06, + "loss": 0.7578, + "step": 10041 + }, + { + "epoch": 0.8557307200681721, + "grad_norm": 19.5, + "learning_rate": 1.5402133756616672e-06, + "loss": 0.8379, + "step": 10042 + }, + { + "epoch": 0.8558159352364721, + "grad_norm": 14.5625, + "learning_rate": 1.54004992905126e-06, + "loss": 0.7742, + "step": 10043 + }, + { + "epoch": 0.8559011504047721, + "grad_norm": 14.9375, + "learning_rate": 1.5398864771993588e-06, + "loss": 1.0037, + "step": 10044 + }, + { + "epoch": 0.8559863655730721, + "grad_norm": 13.8125, + "learning_rate": 1.5397230201089178e-06, + "loss": 0.6259, + "step": 10045 + }, + { + "epoch": 0.856071580741372, + "grad_norm": 10.8125, + "learning_rate": 1.5395595577828903e-06, + "loss": 0.5707, + "step": 10046 + }, + { + "epoch": 0.856156795909672, + "grad_norm": 13.9375, + "learning_rate": 1.5393960902242306e-06, + "loss": 0.629, + "step": 10047 + }, + { + "epoch": 0.8562420110779719, + "grad_norm": 13.3125, + "learning_rate": 1.5392326174358921e-06, + "loss": 0.5655, + "step": 10048 + }, + { + "epoch": 0.8563272262462719, + "grad_norm": 15.6875, + "learning_rate": 1.5390691394208297e-06, + "loss": 0.1799, + "step": 10049 + }, + { + "epoch": 0.8564124414145718, + "grad_norm": 10.875, + "learning_rate": 1.5389056561819973e-06, + "loss": 0.4715, + "step": 10050 + }, + { + "epoch": 0.8564976565828718, + "grad_norm": 12.4375, + "learning_rate": 1.538742167722349e-06, + "loss": 0.4801, + "step": 10051 + }, + { + "epoch": 0.8565828717511718, + "grad_norm": 10.5625, + "learning_rate": 1.5385786740448395e-06, + "loss": 0.3631, + "step": 10052 + }, + { + "epoch": 0.8566680869194717, + "grad_norm": 16.625, + "learning_rate": 1.5384151751524228e-06, + "loss": 0.9585, + "step": 10053 + }, + { + "epoch": 0.8567533020877717, + "grad_norm": 13.5, + "learning_rate": 1.5382516710480544e-06, + "loss": 0.6098, + "step": 10054 + }, + { + "epoch": 0.8568385172560716, + "grad_norm": 19.875, + "learning_rate": 1.5380881617346879e-06, + "loss": 1.1762, + "step": 10055 + }, + { + "epoch": 0.8569237324243716, + "grad_norm": 21.25, + "learning_rate": 1.5379246472152792e-06, + "loss": 1.0837, + "step": 10056 + }, + { + "epoch": 0.8570089475926715, + "grad_norm": 12.625, + "learning_rate": 1.537761127492782e-06, + "loss": 0.6825, + "step": 10057 + }, + { + "epoch": 0.8570941627609715, + "grad_norm": 13.1875, + "learning_rate": 1.5375976025701526e-06, + "loss": 0.5408, + "step": 10058 + }, + { + "epoch": 0.8571793779292715, + "grad_norm": 17.375, + "learning_rate": 1.5374340724503448e-06, + "loss": 0.9366, + "step": 10059 + }, + { + "epoch": 0.8572645930975714, + "grad_norm": 17.125, + "learning_rate": 1.5372705371363145e-06, + "loss": 0.6548, + "step": 10060 + }, + { + "epoch": 0.8573498082658714, + "grad_norm": 13.9375, + "learning_rate": 1.5371069966310163e-06, + "loss": 0.6377, + "step": 10061 + }, + { + "epoch": 0.8574350234341713, + "grad_norm": 13.125, + "learning_rate": 1.5369434509374064e-06, + "loss": 0.8494, + "step": 10062 + }, + { + "epoch": 0.8575202386024713, + "grad_norm": 13.4375, + "learning_rate": 1.5367799000584395e-06, + "loss": 0.6824, + "step": 10063 + }, + { + "epoch": 0.8576054537707712, + "grad_norm": 12.25, + "learning_rate": 1.5366163439970718e-06, + "loss": 0.4818, + "step": 10064 + }, + { + "epoch": 0.8576906689390712, + "grad_norm": 16.625, + "learning_rate": 1.5364527827562582e-06, + "loss": 0.497, + "step": 10065 + }, + { + "epoch": 0.8577758841073712, + "grad_norm": 10.4375, + "learning_rate": 1.5362892163389553e-06, + "loss": 0.3624, + "step": 10066 + }, + { + "epoch": 0.8578610992756711, + "grad_norm": 18.375, + "learning_rate": 1.5361256447481176e-06, + "loss": 0.6826, + "step": 10067 + }, + { + "epoch": 0.8579463144439711, + "grad_norm": 13.3125, + "learning_rate": 1.5359620679867025e-06, + "loss": 0.7534, + "step": 10068 + }, + { + "epoch": 0.858031529612271, + "grad_norm": 20.25, + "learning_rate": 1.5357984860576647e-06, + "loss": 0.8449, + "step": 10069 + }, + { + "epoch": 0.858116744780571, + "grad_norm": 11.5625, + "learning_rate": 1.5356348989639611e-06, + "loss": 0.5237, + "step": 10070 + }, + { + "epoch": 0.8582019599488709, + "grad_norm": 14.6875, + "learning_rate": 1.5354713067085476e-06, + "loss": 0.6756, + "step": 10071 + }, + { + "epoch": 0.8582871751171709, + "grad_norm": 11.5, + "learning_rate": 1.5353077092943804e-06, + "loss": 0.6356, + "step": 10072 + }, + { + "epoch": 0.8583723902854709, + "grad_norm": 14.125, + "learning_rate": 1.5351441067244163e-06, + "loss": 0.6461, + "step": 10073 + }, + { + "epoch": 0.8584576054537708, + "grad_norm": 16.375, + "learning_rate": 1.5349804990016115e-06, + "loss": 0.8606, + "step": 10074 + }, + { + "epoch": 0.8585428206220708, + "grad_norm": 11.5625, + "learning_rate": 1.5348168861289222e-06, + "loss": 0.1709, + "step": 10075 + }, + { + "epoch": 0.8586280357903707, + "grad_norm": 10.6875, + "learning_rate": 1.5346532681093054e-06, + "loss": 0.3805, + "step": 10076 + }, + { + "epoch": 0.8587132509586707, + "grad_norm": 13.25, + "learning_rate": 1.534489644945718e-06, + "loss": 0.543, + "step": 10077 + }, + { + "epoch": 0.8587984661269706, + "grad_norm": 10.375, + "learning_rate": 1.5343260166411165e-06, + "loss": 0.4412, + "step": 10078 + }, + { + "epoch": 0.8588836812952706, + "grad_norm": 24.125, + "learning_rate": 1.5341623831984582e-06, + "loss": 0.7599, + "step": 10079 + }, + { + "epoch": 0.8589688964635706, + "grad_norm": 10.75, + "learning_rate": 1.5339987446206993e-06, + "loss": 0.9472, + "step": 10080 + }, + { + "epoch": 0.8590541116318705, + "grad_norm": 15.5625, + "learning_rate": 1.5338351009107979e-06, + "loss": 0.9006, + "step": 10081 + }, + { + "epoch": 0.8591393268001705, + "grad_norm": 15.0, + "learning_rate": 1.533671452071711e-06, + "loss": 0.8534, + "step": 10082 + }, + { + "epoch": 0.8592245419684704, + "grad_norm": 12.5625, + "learning_rate": 1.5335077981063955e-06, + "loss": 0.5402, + "step": 10083 + }, + { + "epoch": 0.8593097571367704, + "grad_norm": 13.5625, + "learning_rate": 1.5333441390178088e-06, + "loss": 0.7779, + "step": 10084 + }, + { + "epoch": 0.8593949723050703, + "grad_norm": 14.6875, + "learning_rate": 1.5331804748089092e-06, + "loss": 0.8625, + "step": 10085 + }, + { + "epoch": 0.8594801874733703, + "grad_norm": 11.5625, + "learning_rate": 1.533016805482653e-06, + "loss": 0.4704, + "step": 10086 + }, + { + "epoch": 0.8595654026416703, + "grad_norm": 13.0, + "learning_rate": 1.5328531310419985e-06, + "loss": 0.5451, + "step": 10087 + }, + { + "epoch": 0.8596506178099702, + "grad_norm": 14.375, + "learning_rate": 1.532689451489904e-06, + "loss": 0.5704, + "step": 10088 + }, + { + "epoch": 0.8597358329782702, + "grad_norm": 14.625, + "learning_rate": 1.5325257668293264e-06, + "loss": 0.7277, + "step": 10089 + }, + { + "epoch": 0.8598210481465701, + "grad_norm": 21.5, + "learning_rate": 1.5323620770632248e-06, + "loss": 0.7453, + "step": 10090 + }, + { + "epoch": 0.8599062633148701, + "grad_norm": 16.375, + "learning_rate": 1.5321983821945558e-06, + "loss": 0.7724, + "step": 10091 + }, + { + "epoch": 0.85999147848317, + "grad_norm": 17.5, + "learning_rate": 1.5320346822262786e-06, + "loss": 0.9587, + "step": 10092 + }, + { + "epoch": 0.86007669365147, + "grad_norm": 10.1875, + "learning_rate": 1.5318709771613508e-06, + "loss": 0.297, + "step": 10093 + }, + { + "epoch": 0.86016190881977, + "grad_norm": 14.375, + "learning_rate": 1.5317072670027316e-06, + "loss": 0.7796, + "step": 10094 + }, + { + "epoch": 0.8602471239880699, + "grad_norm": 16.75, + "learning_rate": 1.531543551753378e-06, + "loss": 0.967, + "step": 10095 + }, + { + "epoch": 0.8603323391563699, + "grad_norm": 12.875, + "learning_rate": 1.5313798314162498e-06, + "loss": 0.5251, + "step": 10096 + }, + { + "epoch": 0.8604175543246698, + "grad_norm": 11.0625, + "learning_rate": 1.5312161059943048e-06, + "loss": 0.4213, + "step": 10097 + }, + { + "epoch": 0.8605027694929698, + "grad_norm": 18.0, + "learning_rate": 1.5310523754905024e-06, + "loss": 0.9128, + "step": 10098 + }, + { + "epoch": 0.8605879846612697, + "grad_norm": 12.375, + "learning_rate": 1.5308886399078008e-06, + "loss": 0.4332, + "step": 10099 + }, + { + "epoch": 0.8606731998295697, + "grad_norm": 14.0, + "learning_rate": 1.5307248992491588e-06, + "loss": 0.6996, + "step": 10100 + }, + { + "epoch": 0.8607584149978696, + "grad_norm": 10.625, + "learning_rate": 1.5305611535175357e-06, + "loss": 0.5253, + "step": 10101 + }, + { + "epoch": 0.8608436301661696, + "grad_norm": 16.375, + "learning_rate": 1.5303974027158905e-06, + "loss": 0.7075, + "step": 10102 + }, + { + "epoch": 0.8609288453344696, + "grad_norm": 11.5, + "learning_rate": 1.5302336468471823e-06, + "loss": 0.5526, + "step": 10103 + }, + { + "epoch": 0.8610140605027695, + "grad_norm": 12.9375, + "learning_rate": 1.53006988591437e-06, + "loss": 0.8887, + "step": 10104 + }, + { + "epoch": 0.8610992756710695, + "grad_norm": 15.3125, + "learning_rate": 1.5299061199204136e-06, + "loss": 0.6814, + "step": 10105 + }, + { + "epoch": 0.8611844908393694, + "grad_norm": 15.125, + "learning_rate": 1.5297423488682721e-06, + "loss": 0.9171, + "step": 10106 + }, + { + "epoch": 0.8612697060076694, + "grad_norm": 18.125, + "learning_rate": 1.529578572760905e-06, + "loss": 1.0592, + "step": 10107 + }, + { + "epoch": 0.8613549211759693, + "grad_norm": 10.0625, + "learning_rate": 1.5294147916012717e-06, + "loss": 0.2861, + "step": 10108 + }, + { + "epoch": 0.8614401363442693, + "grad_norm": 21.375, + "learning_rate": 1.5292510053923327e-06, + "loss": 1.1863, + "step": 10109 + }, + { + "epoch": 0.8615253515125693, + "grad_norm": 15.625, + "learning_rate": 1.5290872141370466e-06, + "loss": 0.9118, + "step": 10110 + }, + { + "epoch": 0.8616105666808692, + "grad_norm": 13.875, + "learning_rate": 1.5289234178383744e-06, + "loss": 0.5765, + "step": 10111 + }, + { + "epoch": 0.8616957818491692, + "grad_norm": 24.5, + "learning_rate": 1.5287596164992752e-06, + "loss": 1.1467, + "step": 10112 + }, + { + "epoch": 0.8617809970174691, + "grad_norm": 19.5, + "learning_rate": 1.5285958101227098e-06, + "loss": 1.1988, + "step": 10113 + }, + { + "epoch": 0.8618662121857691, + "grad_norm": 43.75, + "learning_rate": 1.5284319987116371e-06, + "loss": 0.4794, + "step": 10114 + }, + { + "epoch": 0.861951427354069, + "grad_norm": 23.375, + "learning_rate": 1.5282681822690194e-06, + "loss": 0.6724, + "step": 10115 + }, + { + "epoch": 0.862036642522369, + "grad_norm": 8.875, + "learning_rate": 1.528104360797815e-06, + "loss": 0.3069, + "step": 10116 + }, + { + "epoch": 0.862121857690669, + "grad_norm": 14.4375, + "learning_rate": 1.5279405343009854e-06, + "loss": 0.7261, + "step": 10117 + }, + { + "epoch": 0.8622070728589689, + "grad_norm": 14.5, + "learning_rate": 1.5277767027814908e-06, + "loss": 0.8507, + "step": 10118 + }, + { + "epoch": 0.8622922880272689, + "grad_norm": 12.1875, + "learning_rate": 1.5276128662422917e-06, + "loss": 0.4278, + "step": 10119 + }, + { + "epoch": 0.8623775031955688, + "grad_norm": 16.75, + "learning_rate": 1.5274490246863495e-06, + "loss": 1.0737, + "step": 10120 + }, + { + "epoch": 0.8624627183638688, + "grad_norm": 10.0, + "learning_rate": 1.527285178116624e-06, + "loss": 0.453, + "step": 10121 + }, + { + "epoch": 0.8625479335321687, + "grad_norm": 14.3125, + "learning_rate": 1.5271213265360766e-06, + "loss": 0.6308, + "step": 10122 + }, + { + "epoch": 0.8626331487004687, + "grad_norm": 17.25, + "learning_rate": 1.526957469947668e-06, + "loss": 0.6245, + "step": 10123 + }, + { + "epoch": 0.8627183638687687, + "grad_norm": 14.6875, + "learning_rate": 1.5267936083543597e-06, + "loss": 0.5867, + "step": 10124 + }, + { + "epoch": 0.8628035790370686, + "grad_norm": 12.75, + "learning_rate": 1.5266297417591121e-06, + "loss": 0.5257, + "step": 10125 + }, + { + "epoch": 0.8628887942053686, + "grad_norm": 12.0, + "learning_rate": 1.5264658701648874e-06, + "loss": 0.4022, + "step": 10126 + }, + { + "epoch": 0.8629740093736685, + "grad_norm": 15.1875, + "learning_rate": 1.5263019935746462e-06, + "loss": 0.7267, + "step": 10127 + }, + { + "epoch": 0.8630592245419685, + "grad_norm": 14.5625, + "learning_rate": 1.5261381119913504e-06, + "loss": 0.8053, + "step": 10128 + }, + { + "epoch": 0.8631444397102684, + "grad_norm": 13.5, + "learning_rate": 1.5259742254179607e-06, + "loss": 0.7908, + "step": 10129 + }, + { + "epoch": 0.8632296548785684, + "grad_norm": 22.625, + "learning_rate": 1.5258103338574396e-06, + "loss": 0.8839, + "step": 10130 + }, + { + "epoch": 0.8633148700468684, + "grad_norm": 14.8125, + "learning_rate": 1.5256464373127483e-06, + "loss": 0.6898, + "step": 10131 + }, + { + "epoch": 0.8634000852151683, + "grad_norm": 12.875, + "learning_rate": 1.5254825357868488e-06, + "loss": 0.5738, + "step": 10132 + }, + { + "epoch": 0.8634853003834683, + "grad_norm": 14.375, + "learning_rate": 1.5253186292827027e-06, + "loss": 0.7751, + "step": 10133 + }, + { + "epoch": 0.8635705155517682, + "grad_norm": 14.25, + "learning_rate": 1.5251547178032722e-06, + "loss": 0.634, + "step": 10134 + }, + { + "epoch": 0.8636557307200682, + "grad_norm": 15.875, + "learning_rate": 1.5249908013515193e-06, + "loss": 0.4709, + "step": 10135 + }, + { + "epoch": 0.8637409458883681, + "grad_norm": 19.5, + "learning_rate": 1.5248268799304062e-06, + "loss": 0.9359, + "step": 10136 + }, + { + "epoch": 0.8638261610566681, + "grad_norm": 17.5, + "learning_rate": 1.5246629535428947e-06, + "loss": 0.6998, + "step": 10137 + }, + { + "epoch": 0.863911376224968, + "grad_norm": 11.25, + "learning_rate": 1.5244990221919479e-06, + "loss": 0.4152, + "step": 10138 + }, + { + "epoch": 0.863996591393268, + "grad_norm": 13.9375, + "learning_rate": 1.5243350858805272e-06, + "loss": 0.5428, + "step": 10139 + }, + { + "epoch": 0.864081806561568, + "grad_norm": 17.125, + "learning_rate": 1.5241711446115958e-06, + "loss": 0.8772, + "step": 10140 + }, + { + "epoch": 0.8641670217298679, + "grad_norm": 18.0, + "learning_rate": 1.5240071983881164e-06, + "loss": 0.9802, + "step": 10141 + }, + { + "epoch": 0.8642522368981679, + "grad_norm": 17.625, + "learning_rate": 1.523843247213051e-06, + "loss": 0.8427, + "step": 10142 + }, + { + "epoch": 0.8643374520664678, + "grad_norm": 14.5625, + "learning_rate": 1.5236792910893633e-06, + "loss": 0.8581, + "step": 10143 + }, + { + "epoch": 0.8644226672347678, + "grad_norm": 18.625, + "learning_rate": 1.523515330020015e-06, + "loss": 0.812, + "step": 10144 + }, + { + "epoch": 0.8645078824030678, + "grad_norm": 14.5625, + "learning_rate": 1.52335136400797e-06, + "loss": 0.858, + "step": 10145 + }, + { + "epoch": 0.8645930975713677, + "grad_norm": 13.0, + "learning_rate": 1.5231873930561909e-06, + "loss": 0.5582, + "step": 10146 + }, + { + "epoch": 0.8646783127396677, + "grad_norm": 14.125, + "learning_rate": 1.5230234171676412e-06, + "loss": 0.7779, + "step": 10147 + }, + { + "epoch": 0.8647635279079676, + "grad_norm": 21.375, + "learning_rate": 1.5228594363452831e-06, + "loss": 0.6991, + "step": 10148 + }, + { + "epoch": 0.8648487430762676, + "grad_norm": 12.875, + "learning_rate": 1.522695450592081e-06, + "loss": 0.5775, + "step": 10149 + }, + { + "epoch": 0.8649339582445675, + "grad_norm": 18.625, + "learning_rate": 1.5225314599109978e-06, + "loss": 0.8759, + "step": 10150 + }, + { + "epoch": 0.8650191734128675, + "grad_norm": 17.125, + "learning_rate": 1.5223674643049975e-06, + "loss": 0.8011, + "step": 10151 + }, + { + "epoch": 0.8651043885811674, + "grad_norm": 9.875, + "learning_rate": 1.5222034637770424e-06, + "loss": 0.2941, + "step": 10152 + }, + { + "epoch": 0.8651896037494674, + "grad_norm": 17.75, + "learning_rate": 1.5220394583300976e-06, + "loss": 0.5878, + "step": 10153 + }, + { + "epoch": 0.8652748189177674, + "grad_norm": 13.625, + "learning_rate": 1.521875447967126e-06, + "loss": 0.8522, + "step": 10154 + }, + { + "epoch": 0.8653600340860673, + "grad_norm": 11.25, + "learning_rate": 1.5217114326910918e-06, + "loss": 0.5354, + "step": 10155 + }, + { + "epoch": 0.8654452492543673, + "grad_norm": 16.875, + "learning_rate": 1.521547412504959e-06, + "loss": 0.8246, + "step": 10156 + }, + { + "epoch": 0.8655304644226672, + "grad_norm": 12.75, + "learning_rate": 1.5213833874116908e-06, + "loss": 0.6896, + "step": 10157 + }, + { + "epoch": 0.8656156795909672, + "grad_norm": 14.625, + "learning_rate": 1.5212193574142522e-06, + "loss": 0.7967, + "step": 10158 + }, + { + "epoch": 0.8657008947592671, + "grad_norm": 15.75, + "learning_rate": 1.5210553225156072e-06, + "loss": 0.6133, + "step": 10159 + }, + { + "epoch": 0.8657861099275671, + "grad_norm": 17.375, + "learning_rate": 1.52089128271872e-06, + "loss": 0.8595, + "step": 10160 + }, + { + "epoch": 0.8658713250958671, + "grad_norm": 16.625, + "learning_rate": 1.5207272380265545e-06, + "loss": 1.028, + "step": 10161 + }, + { + "epoch": 0.865956540264167, + "grad_norm": 21.0, + "learning_rate": 1.520563188442076e-06, + "loss": 0.7477, + "step": 10162 + }, + { + "epoch": 0.866041755432467, + "grad_norm": 18.5, + "learning_rate": 1.5203991339682487e-06, + "loss": 1.1012, + "step": 10163 + }, + { + "epoch": 0.8661269706007669, + "grad_norm": 19.5, + "learning_rate": 1.520235074608037e-06, + "loss": 0.7039, + "step": 10164 + }, + { + "epoch": 0.8662121857690669, + "grad_norm": 15.875, + "learning_rate": 1.5200710103644054e-06, + "loss": 1.0405, + "step": 10165 + }, + { + "epoch": 0.8662974009373668, + "grad_norm": 19.0, + "learning_rate": 1.5199069412403193e-06, + "loss": 1.412, + "step": 10166 + }, + { + "epoch": 0.8663826161056668, + "grad_norm": 11.75, + "learning_rate": 1.5197428672387432e-06, + "loss": 0.4241, + "step": 10167 + }, + { + "epoch": 0.8664678312739668, + "grad_norm": 15.625, + "learning_rate": 1.519578788362643e-06, + "loss": 0.8056, + "step": 10168 + }, + { + "epoch": 0.8665530464422667, + "grad_norm": 13.875, + "learning_rate": 1.5194147046149821e-06, + "loss": 1.0708, + "step": 10169 + }, + { + "epoch": 0.8666382616105667, + "grad_norm": 17.875, + "learning_rate": 1.519250615998727e-06, + "loss": 0.7453, + "step": 10170 + }, + { + "epoch": 0.8667234767788666, + "grad_norm": 10.6875, + "learning_rate": 1.5190865225168422e-06, + "loss": 0.3066, + "step": 10171 + }, + { + "epoch": 0.8668086919471666, + "grad_norm": 11.0625, + "learning_rate": 1.5189224241722935e-06, + "loss": 0.4125, + "step": 10172 + }, + { + "epoch": 0.8668939071154665, + "grad_norm": 14.0625, + "learning_rate": 1.5187583209680464e-06, + "loss": 0.6304, + "step": 10173 + }, + { + "epoch": 0.8669791222837665, + "grad_norm": 12.125, + "learning_rate": 1.5185942129070657e-06, + "loss": 0.9896, + "step": 10174 + }, + { + "epoch": 0.8670643374520665, + "grad_norm": 17.0, + "learning_rate": 1.518430099992318e-06, + "loss": 1.027, + "step": 10175 + }, + { + "epoch": 0.8671495526203664, + "grad_norm": 32.0, + "learning_rate": 1.5182659822267685e-06, + "loss": 0.6872, + "step": 10176 + }, + { + "epoch": 0.8672347677886664, + "grad_norm": 12.25, + "learning_rate": 1.5181018596133825e-06, + "loss": 0.5029, + "step": 10177 + }, + { + "epoch": 0.8673199829569663, + "grad_norm": 12.0, + "learning_rate": 1.5179377321551264e-06, + "loss": 0.5507, + "step": 10178 + }, + { + "epoch": 0.8674051981252663, + "grad_norm": 16.625, + "learning_rate": 1.5177735998549664e-06, + "loss": 0.9249, + "step": 10179 + }, + { + "epoch": 0.8674904132935662, + "grad_norm": 15.4375, + "learning_rate": 1.517609462715868e-06, + "loss": 0.6407, + "step": 10180 + }, + { + "epoch": 0.8675756284618662, + "grad_norm": 16.875, + "learning_rate": 1.5174453207407974e-06, + "loss": 0.9847, + "step": 10181 + }, + { + "epoch": 0.8676608436301662, + "grad_norm": 14.8125, + "learning_rate": 1.517281173932721e-06, + "loss": 0.8747, + "step": 10182 + }, + { + "epoch": 0.8677460587984661, + "grad_norm": 13.3125, + "learning_rate": 1.517117022294605e-06, + "loss": 0.8621, + "step": 10183 + }, + { + "epoch": 0.8678312739667661, + "grad_norm": 12.9375, + "learning_rate": 1.5169528658294158e-06, + "loss": 0.5167, + "step": 10184 + }, + { + "epoch": 0.867916489135066, + "grad_norm": 13.0, + "learning_rate": 1.51678870454012e-06, + "loss": 0.4209, + "step": 10185 + }, + { + "epoch": 0.868001704303366, + "grad_norm": 13.125, + "learning_rate": 1.5166245384296838e-06, + "loss": 0.7921, + "step": 10186 + }, + { + "epoch": 0.8680869194716659, + "grad_norm": 15.5, + "learning_rate": 1.5164603675010741e-06, + "loss": 0.6952, + "step": 10187 + }, + { + "epoch": 0.8681721346399659, + "grad_norm": 16.0, + "learning_rate": 1.5162961917572577e-06, + "loss": 0.8782, + "step": 10188 + }, + { + "epoch": 0.8682573498082659, + "grad_norm": 18.625, + "learning_rate": 1.5161320112012012e-06, + "loss": 0.8164, + "step": 10189 + }, + { + "epoch": 0.8683425649765658, + "grad_norm": 10.625, + "learning_rate": 1.5159678258358718e-06, + "loss": 0.6189, + "step": 10190 + }, + { + "epoch": 0.8684277801448658, + "grad_norm": 17.125, + "learning_rate": 1.5158036356642364e-06, + "loss": 0.7026, + "step": 10191 + }, + { + "epoch": 0.8685129953131657, + "grad_norm": 13.4375, + "learning_rate": 1.515639440689262e-06, + "loss": 0.6142, + "step": 10192 + }, + { + "epoch": 0.8685982104814657, + "grad_norm": 12.5, + "learning_rate": 1.5154752409139156e-06, + "loss": 0.6408, + "step": 10193 + }, + { + "epoch": 0.8686834256497656, + "grad_norm": 11.875, + "learning_rate": 1.5153110363411647e-06, + "loss": 0.3683, + "step": 10194 + }, + { + "epoch": 0.8687686408180656, + "grad_norm": 16.125, + "learning_rate": 1.5151468269739764e-06, + "loss": 0.8836, + "step": 10195 + }, + { + "epoch": 0.8688538559863656, + "grad_norm": 14.625, + "learning_rate": 1.514982612815319e-06, + "loss": 0.5157, + "step": 10196 + }, + { + "epoch": 0.8689390711546655, + "grad_norm": 12.125, + "learning_rate": 1.5148183938681586e-06, + "loss": 0.6051, + "step": 10197 + }, + { + "epoch": 0.8690242863229655, + "grad_norm": 12.875, + "learning_rate": 1.514654170135464e-06, + "loss": 0.6138, + "step": 10198 + }, + { + "epoch": 0.8691095014912654, + "grad_norm": 11.25, + "learning_rate": 1.5144899416202019e-06, + "loss": 0.4368, + "step": 10199 + }, + { + "epoch": 0.8691947166595654, + "grad_norm": 12.625, + "learning_rate": 1.5143257083253413e-06, + "loss": 0.6164, + "step": 10200 + }, + { + "epoch": 0.8692799318278653, + "grad_norm": 11.9375, + "learning_rate": 1.5141614702538488e-06, + "loss": 0.4718, + "step": 10201 + }, + { + "epoch": 0.8693651469961653, + "grad_norm": 11.9375, + "learning_rate": 1.5139972274086928e-06, + "loss": 0.4012, + "step": 10202 + }, + { + "epoch": 0.8694503621644653, + "grad_norm": 14.625, + "learning_rate": 1.5138329797928416e-06, + "loss": 0.8543, + "step": 10203 + }, + { + "epoch": 0.8695355773327652, + "grad_norm": 18.875, + "learning_rate": 1.5136687274092632e-06, + "loss": 1.0437, + "step": 10204 + }, + { + "epoch": 0.8696207925010652, + "grad_norm": 13.1875, + "learning_rate": 1.5135044702609259e-06, + "loss": 0.4893, + "step": 10205 + }, + { + "epoch": 0.8697060076693651, + "grad_norm": 13.625, + "learning_rate": 1.5133402083507978e-06, + "loss": 0.8573, + "step": 10206 + }, + { + "epoch": 0.8697912228376651, + "grad_norm": 12.625, + "learning_rate": 1.513175941681847e-06, + "loss": 0.6931, + "step": 10207 + }, + { + "epoch": 0.869876438005965, + "grad_norm": 14.6875, + "learning_rate": 1.5130116702570427e-06, + "loss": 0.6333, + "step": 10208 + }, + { + "epoch": 0.869961653174265, + "grad_norm": 11.5625, + "learning_rate": 1.5128473940793531e-06, + "loss": 0.4928, + "step": 10209 + }, + { + "epoch": 0.870046868342565, + "grad_norm": 25.5, + "learning_rate": 1.5126831131517464e-06, + "loss": 0.9812, + "step": 10210 + }, + { + "epoch": 0.8701320835108649, + "grad_norm": 16.0, + "learning_rate": 1.5125188274771918e-06, + "loss": 0.8645, + "step": 10211 + }, + { + "epoch": 0.8702172986791649, + "grad_norm": 15.0625, + "learning_rate": 1.5123545370586581e-06, + "loss": 0.7024, + "step": 10212 + }, + { + "epoch": 0.8703025138474648, + "grad_norm": 12.5625, + "learning_rate": 1.5121902418991142e-06, + "loss": 0.6007, + "step": 10213 + }, + { + "epoch": 0.8703877290157648, + "grad_norm": 13.375, + "learning_rate": 1.5120259420015286e-06, + "loss": 0.6857, + "step": 10214 + }, + { + "epoch": 0.8704729441840647, + "grad_norm": 15.6875, + "learning_rate": 1.511861637368871e-06, + "loss": 0.6279, + "step": 10215 + }, + { + "epoch": 0.8705581593523647, + "grad_norm": 16.125, + "learning_rate": 1.5116973280041103e-06, + "loss": 1.0007, + "step": 10216 + }, + { + "epoch": 0.8706433745206646, + "grad_norm": 15.0, + "learning_rate": 1.5115330139102157e-06, + "loss": 0.9044, + "step": 10217 + }, + { + "epoch": 0.8707285896889646, + "grad_norm": 14.5, + "learning_rate": 1.5113686950901566e-06, + "loss": 0.5351, + "step": 10218 + }, + { + "epoch": 0.8708138048572646, + "grad_norm": 13.375, + "learning_rate": 1.5112043715469022e-06, + "loss": 0.7688, + "step": 10219 + }, + { + "epoch": 0.8708990200255645, + "grad_norm": 14.3125, + "learning_rate": 1.5110400432834224e-06, + "loss": 1.0671, + "step": 10220 + }, + { + "epoch": 0.8709842351938645, + "grad_norm": 15.9375, + "learning_rate": 1.5108757103026864e-06, + "loss": 0.7939, + "step": 10221 + }, + { + "epoch": 0.8710694503621644, + "grad_norm": 12.3125, + "learning_rate": 1.510711372607664e-06, + "loss": 0.5829, + "step": 10222 + }, + { + "epoch": 0.8711546655304644, + "grad_norm": 13.6875, + "learning_rate": 1.5105470302013248e-06, + "loss": 0.7145, + "step": 10223 + }, + { + "epoch": 0.8712398806987643, + "grad_norm": 11.6875, + "learning_rate": 1.510382683086639e-06, + "loss": 0.3441, + "step": 10224 + }, + { + "epoch": 0.8713250958670643, + "grad_norm": 17.75, + "learning_rate": 1.5102183312665757e-06, + "loss": 0.8196, + "step": 10225 + }, + { + "epoch": 0.8714103110353643, + "grad_norm": 17.75, + "learning_rate": 1.510053974744106e-06, + "loss": 0.9774, + "step": 10226 + }, + { + "epoch": 0.8714955262036642, + "grad_norm": 10.375, + "learning_rate": 1.509889613522199e-06, + "loss": 0.3194, + "step": 10227 + }, + { + "epoch": 0.8715807413719642, + "grad_norm": 10.25, + "learning_rate": 1.5097252476038262e-06, + "loss": 0.2893, + "step": 10228 + }, + { + "epoch": 0.8716659565402641, + "grad_norm": 13.625, + "learning_rate": 1.5095608769919566e-06, + "loss": 0.8017, + "step": 10229 + }, + { + "epoch": 0.8717511717085641, + "grad_norm": 16.5, + "learning_rate": 1.5093965016895606e-06, + "loss": 0.8652, + "step": 10230 + }, + { + "epoch": 0.871836386876864, + "grad_norm": 12.8125, + "learning_rate": 1.5092321216996092e-06, + "loss": 0.4625, + "step": 10231 + }, + { + "epoch": 0.871921602045164, + "grad_norm": 16.5, + "learning_rate": 1.5090677370250729e-06, + "loss": 0.4071, + "step": 10232 + }, + { + "epoch": 0.872006817213464, + "grad_norm": 17.0, + "learning_rate": 1.5089033476689219e-06, + "loss": 0.9149, + "step": 10233 + }, + { + "epoch": 0.8720920323817639, + "grad_norm": 12.9375, + "learning_rate": 1.508738953634127e-06, + "loss": 1.024, + "step": 10234 + }, + { + "epoch": 0.8721772475500639, + "grad_norm": 15.0, + "learning_rate": 1.5085745549236588e-06, + "loss": 0.8391, + "step": 10235 + }, + { + "epoch": 0.8722624627183638, + "grad_norm": 14.9375, + "learning_rate": 1.5084101515404889e-06, + "loss": 0.5198, + "step": 10236 + }, + { + "epoch": 0.8723476778866638, + "grad_norm": 18.375, + "learning_rate": 1.5082457434875874e-06, + "loss": 0.8274, + "step": 10237 + }, + { + "epoch": 0.8724328930549637, + "grad_norm": 16.125, + "learning_rate": 1.508081330767926e-06, + "loss": 0.9801, + "step": 10238 + }, + { + "epoch": 0.8725181082232637, + "grad_norm": 13.3125, + "learning_rate": 1.5079169133844749e-06, + "loss": 0.591, + "step": 10239 + }, + { + "epoch": 0.8726033233915637, + "grad_norm": 15.9375, + "learning_rate": 1.507752491340206e-06, + "loss": 0.6914, + "step": 10240 + }, + { + "epoch": 0.8726885385598636, + "grad_norm": 13.9375, + "learning_rate": 1.5075880646380905e-06, + "loss": 1.0586, + "step": 10241 + }, + { + "epoch": 0.8727737537281636, + "grad_norm": 11.6875, + "learning_rate": 1.5074236332810993e-06, + "loss": 0.3778, + "step": 10242 + }, + { + "epoch": 0.8728589688964635, + "grad_norm": 18.25, + "learning_rate": 1.5072591972722047e-06, + "loss": 0.8658, + "step": 10243 + }, + { + "epoch": 0.8729441840647635, + "grad_norm": 16.625, + "learning_rate": 1.5070947566143775e-06, + "loss": 0.8338, + "step": 10244 + }, + { + "epoch": 0.8730293992330634, + "grad_norm": 12.375, + "learning_rate": 1.5069303113105895e-06, + "loss": 0.8188, + "step": 10245 + }, + { + "epoch": 0.8731146144013634, + "grad_norm": 13.1875, + "learning_rate": 1.5067658613638124e-06, + "loss": 0.5875, + "step": 10246 + }, + { + "epoch": 0.8731998295696634, + "grad_norm": 15.125, + "learning_rate": 1.506601406777018e-06, + "loss": 0.9176, + "step": 10247 + }, + { + "epoch": 0.8732850447379633, + "grad_norm": 21.0, + "learning_rate": 1.506436947553178e-06, + "loss": 0.6382, + "step": 10248 + }, + { + "epoch": 0.8733702599062633, + "grad_norm": 14.8125, + "learning_rate": 1.5062724836952649e-06, + "loss": 0.6083, + "step": 10249 + }, + { + "epoch": 0.8734554750745632, + "grad_norm": 14.875, + "learning_rate": 1.5061080152062498e-06, + "loss": 0.6199, + "step": 10250 + }, + { + "epoch": 0.8735406902428632, + "grad_norm": 17.875, + "learning_rate": 1.5059435420891058e-06, + "loss": 0.6179, + "step": 10251 + }, + { + "epoch": 0.8736259054111631, + "grad_norm": 15.625, + "learning_rate": 1.5057790643468042e-06, + "loss": 0.7413, + "step": 10252 + }, + { + "epoch": 0.8737111205794631, + "grad_norm": 16.625, + "learning_rate": 1.5056145819823182e-06, + "loss": 0.6481, + "step": 10253 + }, + { + "epoch": 0.873796335747763, + "grad_norm": 13.4375, + "learning_rate": 1.5054500949986193e-06, + "loss": 0.5957, + "step": 10254 + }, + { + "epoch": 0.873881550916063, + "grad_norm": 17.5, + "learning_rate": 1.5052856033986804e-06, + "loss": 1.1521, + "step": 10255 + }, + { + "epoch": 0.873966766084363, + "grad_norm": 13.375, + "learning_rate": 1.5051211071854738e-06, + "loss": 0.636, + "step": 10256 + }, + { + "epoch": 0.8740519812526629, + "grad_norm": 15.125, + "learning_rate": 1.5049566063619725e-06, + "loss": 0.7734, + "step": 10257 + }, + { + "epoch": 0.8741371964209629, + "grad_norm": 16.375, + "learning_rate": 1.504792100931149e-06, + "loss": 0.6933, + "step": 10258 + }, + { + "epoch": 0.8742224115892628, + "grad_norm": 14.5, + "learning_rate": 1.5046275908959755e-06, + "loss": 0.5732, + "step": 10259 + }, + { + "epoch": 0.8743076267575628, + "grad_norm": 13.875, + "learning_rate": 1.5044630762594257e-06, + "loss": 0.6248, + "step": 10260 + }, + { + "epoch": 0.8743928419258628, + "grad_norm": 15.875, + "learning_rate": 1.5042985570244722e-06, + "loss": 0.9475, + "step": 10261 + }, + { + "epoch": 0.8744780570941627, + "grad_norm": 14.25, + "learning_rate": 1.5041340331940885e-06, + "loss": 0.9059, + "step": 10262 + }, + { + "epoch": 0.8745632722624627, + "grad_norm": 21.5, + "learning_rate": 1.5039695047712466e-06, + "loss": 0.7358, + "step": 10263 + }, + { + "epoch": 0.8746484874307626, + "grad_norm": 17.375, + "learning_rate": 1.5038049717589206e-06, + "loss": 1.0387, + "step": 10264 + }, + { + "epoch": 0.8747337025990626, + "grad_norm": 14.1875, + "learning_rate": 1.5036404341600835e-06, + "loss": 0.5354, + "step": 10265 + }, + { + "epoch": 0.8748189177673625, + "grad_norm": 15.3125, + "learning_rate": 1.503475891977709e-06, + "loss": 0.5545, + "step": 10266 + }, + { + "epoch": 0.8749041329356625, + "grad_norm": 12.9375, + "learning_rate": 1.50331134521477e-06, + "loss": 0.688, + "step": 10267 + }, + { + "epoch": 0.8749893481039624, + "grad_norm": 11.125, + "learning_rate": 1.5031467938742404e-06, + "loss": 0.4511, + "step": 10268 + }, + { + "epoch": 0.8750745632722625, + "grad_norm": 14.4375, + "learning_rate": 1.5029822379590936e-06, + "loss": 0.4503, + "step": 10269 + }, + { + "epoch": 0.8751597784405625, + "grad_norm": 16.75, + "learning_rate": 1.5028176774723033e-06, + "loss": 0.9331, + "step": 10270 + }, + { + "epoch": 0.8752449936088624, + "grad_norm": 17.875, + "learning_rate": 1.5026531124168431e-06, + "loss": 0.8639, + "step": 10271 + }, + { + "epoch": 0.8753302087771624, + "grad_norm": 14.5625, + "learning_rate": 1.5024885427956877e-06, + "loss": 0.7644, + "step": 10272 + }, + { + "epoch": 0.8754154239454623, + "grad_norm": 12.75, + "learning_rate": 1.50232396861181e-06, + "loss": 0.4953, + "step": 10273 + }, + { + "epoch": 0.8755006391137623, + "grad_norm": 24.375, + "learning_rate": 1.5021593898681844e-06, + "loss": 0.9405, + "step": 10274 + }, + { + "epoch": 0.8755858542820623, + "grad_norm": 12.8125, + "learning_rate": 1.5019948065677852e-06, + "loss": 0.5444, + "step": 10275 + }, + { + "epoch": 0.8756710694503622, + "grad_norm": 12.375, + "learning_rate": 1.5018302187135866e-06, + "loss": 0.5449, + "step": 10276 + }, + { + "epoch": 0.8757562846186622, + "grad_norm": 13.375, + "learning_rate": 1.5016656263085627e-06, + "loss": 0.608, + "step": 10277 + }, + { + "epoch": 0.8758414997869621, + "grad_norm": 25.375, + "learning_rate": 1.5015010293556875e-06, + "loss": 1.1416, + "step": 10278 + }, + { + "epoch": 0.8759267149552621, + "grad_norm": 12.75, + "learning_rate": 1.5013364278579362e-06, + "loss": 0.5875, + "step": 10279 + }, + { + "epoch": 0.876011930123562, + "grad_norm": 11.5, + "learning_rate": 1.5011718218182824e-06, + "loss": 0.338, + "step": 10280 + }, + { + "epoch": 0.876097145291862, + "grad_norm": 17.0, + "learning_rate": 1.501007211239702e-06, + "loss": 0.7929, + "step": 10281 + }, + { + "epoch": 0.876182360460162, + "grad_norm": 20.75, + "learning_rate": 1.5008425961251682e-06, + "loss": 0.8831, + "step": 10282 + }, + { + "epoch": 0.8762675756284619, + "grad_norm": 18.375, + "learning_rate": 1.5006779764776565e-06, + "loss": 1.1731, + "step": 10283 + }, + { + "epoch": 0.8763527907967619, + "grad_norm": 18.5, + "learning_rate": 1.500513352300142e-06, + "loss": 0.9694, + "step": 10284 + }, + { + "epoch": 0.8764380059650618, + "grad_norm": 14.375, + "learning_rate": 1.5003487235955994e-06, + "loss": 0.9148, + "step": 10285 + }, + { + "epoch": 0.8765232211333618, + "grad_norm": 18.125, + "learning_rate": 1.5001840903670034e-06, + "loss": 0.9419, + "step": 10286 + }, + { + "epoch": 0.8766084363016617, + "grad_norm": 14.3125, + "learning_rate": 1.5000194526173291e-06, + "loss": 0.7039, + "step": 10287 + }, + { + "epoch": 0.8766936514699617, + "grad_norm": 14.25, + "learning_rate": 1.4998548103495519e-06, + "loss": 0.6037, + "step": 10288 + }, + { + "epoch": 0.8767788666382617, + "grad_norm": 17.0, + "learning_rate": 1.4996901635666475e-06, + "loss": 0.8228, + "step": 10289 + }, + { + "epoch": 0.8768640818065616, + "grad_norm": 13.8125, + "learning_rate": 1.4995255122715905e-06, + "loss": 0.6379, + "step": 10290 + }, + { + "epoch": 0.8769492969748616, + "grad_norm": 12.9375, + "learning_rate": 1.4993608564673567e-06, + "loss": 0.5076, + "step": 10291 + }, + { + "epoch": 0.8770345121431615, + "grad_norm": 14.0625, + "learning_rate": 1.499196196156921e-06, + "loss": 0.6565, + "step": 10292 + }, + { + "epoch": 0.8771197273114615, + "grad_norm": 15.6875, + "learning_rate": 1.49903153134326e-06, + "loss": 0.9863, + "step": 10293 + }, + { + "epoch": 0.8772049424797614, + "grad_norm": 15.6875, + "learning_rate": 1.4988668620293487e-06, + "loss": 0.9442, + "step": 10294 + }, + { + "epoch": 0.8772901576480614, + "grad_norm": 12.5625, + "learning_rate": 1.4987021882181625e-06, + "loss": 0.5681, + "step": 10295 + }, + { + "epoch": 0.8773753728163614, + "grad_norm": 12.5, + "learning_rate": 1.4985375099126781e-06, + "loss": 0.5828, + "step": 10296 + }, + { + "epoch": 0.8774605879846613, + "grad_norm": 13.875, + "learning_rate": 1.4983728271158709e-06, + "loss": 0.6703, + "step": 10297 + }, + { + "epoch": 0.8775458031529613, + "grad_norm": 18.0, + "learning_rate": 1.498208139830717e-06, + "loss": 0.68, + "step": 10298 + }, + { + "epoch": 0.8776310183212612, + "grad_norm": 12.125, + "learning_rate": 1.4980434480601922e-06, + "loss": 0.7133, + "step": 10299 + }, + { + "epoch": 0.8777162334895612, + "grad_norm": 17.625, + "learning_rate": 1.4978787518072732e-06, + "loss": 0.7827, + "step": 10300 + }, + { + "epoch": 0.8778014486578611, + "grad_norm": 15.75, + "learning_rate": 1.4977140510749354e-06, + "loss": 0.6571, + "step": 10301 + }, + { + "epoch": 0.8778866638261611, + "grad_norm": 14.625, + "learning_rate": 1.4975493458661564e-06, + "loss": 0.7491, + "step": 10302 + }, + { + "epoch": 0.877971878994461, + "grad_norm": 15.0625, + "learning_rate": 1.497384636183911e-06, + "loss": 0.5453, + "step": 10303 + }, + { + "epoch": 0.878057094162761, + "grad_norm": 15.3125, + "learning_rate": 1.4972199220311766e-06, + "loss": 0.9451, + "step": 10304 + }, + { + "epoch": 0.878142309331061, + "grad_norm": 12.75, + "learning_rate": 1.4970552034109298e-06, + "loss": 0.6084, + "step": 10305 + }, + { + "epoch": 0.8782275244993609, + "grad_norm": 15.1875, + "learning_rate": 1.4968904803261475e-06, + "loss": 0.8222, + "step": 10306 + }, + { + "epoch": 0.8783127396676609, + "grad_norm": 19.625, + "learning_rate": 1.4967257527798052e-06, + "loss": 0.845, + "step": 10307 + }, + { + "epoch": 0.8783979548359608, + "grad_norm": 12.0625, + "learning_rate": 1.4965610207748806e-06, + "loss": 0.2869, + "step": 10308 + }, + { + "epoch": 0.8784831700042608, + "grad_norm": 15.9375, + "learning_rate": 1.4963962843143503e-06, + "loss": 1.0366, + "step": 10309 + }, + { + "epoch": 0.8785683851725608, + "grad_norm": 15.5625, + "learning_rate": 1.4962315434011918e-06, + "loss": 0.705, + "step": 10310 + }, + { + "epoch": 0.8786536003408607, + "grad_norm": 15.25, + "learning_rate": 1.4960667980383816e-06, + "loss": 0.4886, + "step": 10311 + }, + { + "epoch": 0.8787388155091607, + "grad_norm": 18.375, + "learning_rate": 1.4959020482288966e-06, + "loss": 0.8341, + "step": 10312 + }, + { + "epoch": 0.8788240306774606, + "grad_norm": 11.8125, + "learning_rate": 1.4957372939757148e-06, + "loss": 0.5601, + "step": 10313 + }, + { + "epoch": 0.8789092458457606, + "grad_norm": 14.625, + "learning_rate": 1.4955725352818127e-06, + "loss": 0.8358, + "step": 10314 + }, + { + "epoch": 0.8789944610140605, + "grad_norm": 13.3125, + "learning_rate": 1.495407772150168e-06, + "loss": 0.76, + "step": 10315 + }, + { + "epoch": 0.8790796761823605, + "grad_norm": 12.3125, + "learning_rate": 1.495243004583758e-06, + "loss": 0.7222, + "step": 10316 + }, + { + "epoch": 0.8791648913506604, + "grad_norm": 13.3125, + "learning_rate": 1.4950782325855603e-06, + "loss": 0.8994, + "step": 10317 + }, + { + "epoch": 0.8792501065189604, + "grad_norm": 15.5, + "learning_rate": 1.4949134561585527e-06, + "loss": 0.936, + "step": 10318 + }, + { + "epoch": 0.8793353216872604, + "grad_norm": 14.3125, + "learning_rate": 1.4947486753057125e-06, + "loss": 0.5492, + "step": 10319 + }, + { + "epoch": 0.8794205368555603, + "grad_norm": 12.9375, + "learning_rate": 1.4945838900300174e-06, + "loss": 0.5649, + "step": 10320 + }, + { + "epoch": 0.8795057520238603, + "grad_norm": 12.75, + "learning_rate": 1.4944191003344457e-06, + "loss": 0.7471, + "step": 10321 + }, + { + "epoch": 0.8795909671921602, + "grad_norm": 11.75, + "learning_rate": 1.4942543062219752e-06, + "loss": 0.6514, + "step": 10322 + }, + { + "epoch": 0.8796761823604602, + "grad_norm": 18.25, + "learning_rate": 1.4940895076955838e-06, + "loss": 0.8471, + "step": 10323 + }, + { + "epoch": 0.8797613975287601, + "grad_norm": 16.75, + "learning_rate": 1.4939247047582491e-06, + "loss": 0.8896, + "step": 10324 + }, + { + "epoch": 0.8798466126970601, + "grad_norm": 16.375, + "learning_rate": 1.4937598974129503e-06, + "loss": 1.0708, + "step": 10325 + }, + { + "epoch": 0.8799318278653601, + "grad_norm": 13.75, + "learning_rate": 1.4935950856626646e-06, + "loss": 0.781, + "step": 10326 + }, + { + "epoch": 0.88001704303366, + "grad_norm": 17.0, + "learning_rate": 1.4934302695103708e-06, + "loss": 0.5999, + "step": 10327 + }, + { + "epoch": 0.88010225820196, + "grad_norm": 13.3125, + "learning_rate": 1.4932654489590473e-06, + "loss": 0.496, + "step": 10328 + }, + { + "epoch": 0.8801874733702599, + "grad_norm": 12.375, + "learning_rate": 1.4931006240116726e-06, + "loss": 0.5902, + "step": 10329 + }, + { + "epoch": 0.8802726885385599, + "grad_norm": 12.125, + "learning_rate": 1.4929357946712253e-06, + "loss": 0.3672, + "step": 10330 + }, + { + "epoch": 0.8803579037068598, + "grad_norm": 12.0625, + "learning_rate": 1.4927709609406835e-06, + "loss": 0.6805, + "step": 10331 + }, + { + "epoch": 0.8804431188751598, + "grad_norm": 19.0, + "learning_rate": 1.4926061228230264e-06, + "loss": 1.146, + "step": 10332 + }, + { + "epoch": 0.8805283340434598, + "grad_norm": 12.375, + "learning_rate": 1.4924412803212327e-06, + "loss": 0.5561, + "step": 10333 + }, + { + "epoch": 0.8806135492117597, + "grad_norm": 16.0, + "learning_rate": 1.4922764334382816e-06, + "loss": 0.8491, + "step": 10334 + }, + { + "epoch": 0.8806987643800597, + "grad_norm": 10.375, + "learning_rate": 1.4921115821771514e-06, + "loss": 0.3205, + "step": 10335 + }, + { + "epoch": 0.8807839795483596, + "grad_norm": 14.875, + "learning_rate": 1.4919467265408216e-06, + "loss": 0.6154, + "step": 10336 + }, + { + "epoch": 0.8808691947166596, + "grad_norm": 16.125, + "learning_rate": 1.4917818665322708e-06, + "loss": 0.8127, + "step": 10337 + }, + { + "epoch": 0.8809544098849595, + "grad_norm": 19.25, + "learning_rate": 1.4916170021544792e-06, + "loss": 0.6141, + "step": 10338 + }, + { + "epoch": 0.8810396250532595, + "grad_norm": 19.75, + "learning_rate": 1.491452133410425e-06, + "loss": 1.012, + "step": 10339 + }, + { + "epoch": 0.8811248402215595, + "grad_norm": 16.875, + "learning_rate": 1.491287260303088e-06, + "loss": 0.8337, + "step": 10340 + }, + { + "epoch": 0.8812100553898594, + "grad_norm": 14.4375, + "learning_rate": 1.4911223828354473e-06, + "loss": 1.0503, + "step": 10341 + }, + { + "epoch": 0.8812952705581594, + "grad_norm": 14.1875, + "learning_rate": 1.4909575010104832e-06, + "loss": 0.8615, + "step": 10342 + }, + { + "epoch": 0.8813804857264593, + "grad_norm": 23.25, + "learning_rate": 1.4907926148311747e-06, + "loss": 0.7213, + "step": 10343 + }, + { + "epoch": 0.8814657008947593, + "grad_norm": 17.125, + "learning_rate": 1.4906277243005012e-06, + "loss": 0.7568, + "step": 10344 + }, + { + "epoch": 0.8815509160630592, + "grad_norm": 23.25, + "learning_rate": 1.4904628294214426e-06, + "loss": 1.1367, + "step": 10345 + }, + { + "epoch": 0.8816361312313592, + "grad_norm": 19.25, + "learning_rate": 1.4902979301969795e-06, + "loss": 0.9534, + "step": 10346 + }, + { + "epoch": 0.8817213463996592, + "grad_norm": 19.25, + "learning_rate": 1.4901330266300905e-06, + "loss": 1.0428, + "step": 10347 + }, + { + "epoch": 0.8818065615679591, + "grad_norm": 12.9375, + "learning_rate": 1.4899681187237567e-06, + "loss": 0.5236, + "step": 10348 + }, + { + "epoch": 0.8818917767362591, + "grad_norm": 25.125, + "learning_rate": 1.4898032064809575e-06, + "loss": 0.7212, + "step": 10349 + }, + { + "epoch": 0.881976991904559, + "grad_norm": 22.5, + "learning_rate": 1.4896382899046733e-06, + "loss": 1.0081, + "step": 10350 + }, + { + "epoch": 0.882062207072859, + "grad_norm": 39.0, + "learning_rate": 1.489473368997884e-06, + "loss": 0.8309, + "step": 10351 + }, + { + "epoch": 0.8821474222411589, + "grad_norm": 14.625, + "learning_rate": 1.4893084437635702e-06, + "loss": 0.6007, + "step": 10352 + }, + { + "epoch": 0.8822326374094589, + "grad_norm": 12.25, + "learning_rate": 1.4891435142047123e-06, + "loss": 0.553, + "step": 10353 + }, + { + "epoch": 0.8823178525777589, + "grad_norm": 17.375, + "learning_rate": 1.4889785803242907e-06, + "loss": 0.8502, + "step": 10354 + }, + { + "epoch": 0.8824030677460588, + "grad_norm": 15.5, + "learning_rate": 1.4888136421252857e-06, + "loss": 0.7555, + "step": 10355 + }, + { + "epoch": 0.8824882829143588, + "grad_norm": 19.625, + "learning_rate": 1.4886486996106781e-06, + "loss": 0.8821, + "step": 10356 + }, + { + "epoch": 0.8825734980826587, + "grad_norm": 13.5625, + "learning_rate": 1.4884837527834485e-06, + "loss": 0.4516, + "step": 10357 + }, + { + "epoch": 0.8826587132509587, + "grad_norm": 17.75, + "learning_rate": 1.4883188016465776e-06, + "loss": 0.7758, + "step": 10358 + }, + { + "epoch": 0.8827439284192586, + "grad_norm": 16.875, + "learning_rate": 1.4881538462030466e-06, + "loss": 0.9124, + "step": 10359 + }, + { + "epoch": 0.8828291435875586, + "grad_norm": 11.3125, + "learning_rate": 1.4879888864558356e-06, + "loss": 0.56, + "step": 10360 + }, + { + "epoch": 0.8829143587558586, + "grad_norm": 13.1875, + "learning_rate": 1.4878239224079265e-06, + "loss": 0.7051, + "step": 10361 + }, + { + "epoch": 0.8829995739241585, + "grad_norm": 11.8125, + "learning_rate": 1.4876589540622998e-06, + "loss": 0.6376, + "step": 10362 + }, + { + "epoch": 0.8830847890924585, + "grad_norm": 12.125, + "learning_rate": 1.487493981421937e-06, + "loss": 0.6026, + "step": 10363 + }, + { + "epoch": 0.8831700042607584, + "grad_norm": 18.125, + "learning_rate": 1.487329004489819e-06, + "loss": 0.9058, + "step": 10364 + }, + { + "epoch": 0.8832552194290584, + "grad_norm": 20.375, + "learning_rate": 1.487164023268927e-06, + "loss": 0.748, + "step": 10365 + }, + { + "epoch": 0.8833404345973583, + "grad_norm": 13.25, + "learning_rate": 1.4869990377622432e-06, + "loss": 0.6584, + "step": 10366 + }, + { + "epoch": 0.8834256497656583, + "grad_norm": 29.375, + "learning_rate": 1.4868340479727477e-06, + "loss": 1.1751, + "step": 10367 + }, + { + "epoch": 0.8835108649339583, + "grad_norm": 20.375, + "learning_rate": 1.4866690539034235e-06, + "loss": 0.9223, + "step": 10368 + }, + { + "epoch": 0.8835960801022582, + "grad_norm": 12.5625, + "learning_rate": 1.486504055557251e-06, + "loss": 0.6066, + "step": 10369 + }, + { + "epoch": 0.8836812952705582, + "grad_norm": 15.8125, + "learning_rate": 1.4863390529372126e-06, + "loss": 0.8881, + "step": 10370 + }, + { + "epoch": 0.8837665104388581, + "grad_norm": 18.0, + "learning_rate": 1.48617404604629e-06, + "loss": 1.0259, + "step": 10371 + }, + { + "epoch": 0.8838517256071581, + "grad_norm": 18.875, + "learning_rate": 1.4860090348874648e-06, + "loss": 0.8101, + "step": 10372 + }, + { + "epoch": 0.883936940775458, + "grad_norm": 19.375, + "learning_rate": 1.485844019463719e-06, + "loss": 0.7419, + "step": 10373 + }, + { + "epoch": 0.884022155943758, + "grad_norm": 11.4375, + "learning_rate": 1.4856789997780347e-06, + "loss": 0.482, + "step": 10374 + }, + { + "epoch": 0.884107371112058, + "grad_norm": 14.0, + "learning_rate": 1.4855139758333936e-06, + "loss": 0.4415, + "step": 10375 + }, + { + "epoch": 0.8841925862803579, + "grad_norm": 12.875, + "learning_rate": 1.4853489476327785e-06, + "loss": 0.7432, + "step": 10376 + }, + { + "epoch": 0.8842778014486579, + "grad_norm": 16.0, + "learning_rate": 1.4851839151791707e-06, + "loss": 0.6941, + "step": 10377 + }, + { + "epoch": 0.8843630166169578, + "grad_norm": 13.125, + "learning_rate": 1.4850188784755538e-06, + "loss": 0.6133, + "step": 10378 + }, + { + "epoch": 0.8844482317852578, + "grad_norm": 22.25, + "learning_rate": 1.4848538375249088e-06, + "loss": 0.6276, + "step": 10379 + }, + { + "epoch": 0.8845334469535577, + "grad_norm": 11.1875, + "learning_rate": 1.484688792330219e-06, + "loss": 0.3604, + "step": 10380 + }, + { + "epoch": 0.8846186621218577, + "grad_norm": 19.375, + "learning_rate": 1.484523742894467e-06, + "loss": 0.4631, + "step": 10381 + }, + { + "epoch": 0.8847038772901576, + "grad_norm": 12.625, + "learning_rate": 1.4843586892206346e-06, + "loss": 0.5561, + "step": 10382 + }, + { + "epoch": 0.8847890924584576, + "grad_norm": 15.8125, + "learning_rate": 1.4841936313117056e-06, + "loss": 0.5532, + "step": 10383 + }, + { + "epoch": 0.8848743076267576, + "grad_norm": 13.9375, + "learning_rate": 1.4840285691706618e-06, + "loss": 0.8556, + "step": 10384 + }, + { + "epoch": 0.8849595227950575, + "grad_norm": 16.625, + "learning_rate": 1.4838635028004866e-06, + "loss": 0.7676, + "step": 10385 + }, + { + "epoch": 0.8850447379633575, + "grad_norm": 12.4375, + "learning_rate": 1.4836984322041626e-06, + "loss": 0.5567, + "step": 10386 + }, + { + "epoch": 0.8851299531316574, + "grad_norm": 17.0, + "learning_rate": 1.4835333573846733e-06, + "loss": 1.0572, + "step": 10387 + }, + { + "epoch": 0.8852151682999574, + "grad_norm": 11.0, + "learning_rate": 1.4833682783450007e-06, + "loss": 0.4456, + "step": 10388 + }, + { + "epoch": 0.8853003834682573, + "grad_norm": 11.125, + "learning_rate": 1.4832031950881292e-06, + "loss": 0.4612, + "step": 10389 + }, + { + "epoch": 0.8853855986365573, + "grad_norm": 16.25, + "learning_rate": 1.4830381076170411e-06, + "loss": 0.7626, + "step": 10390 + }, + { + "epoch": 0.8854708138048573, + "grad_norm": 16.375, + "learning_rate": 1.4828730159347204e-06, + "loss": 0.5575, + "step": 10391 + }, + { + "epoch": 0.8855560289731572, + "grad_norm": 13.9375, + "learning_rate": 1.4827079200441498e-06, + "loss": 0.8416, + "step": 10392 + }, + { + "epoch": 0.8856412441414572, + "grad_norm": 11.8125, + "learning_rate": 1.4825428199483135e-06, + "loss": 0.6601, + "step": 10393 + }, + { + "epoch": 0.8857264593097571, + "grad_norm": 11.9375, + "learning_rate": 1.4823777156501943e-06, + "loss": 0.6939, + "step": 10394 + }, + { + "epoch": 0.8858116744780571, + "grad_norm": 17.125, + "learning_rate": 1.482212607152776e-06, + "loss": 0.943, + "step": 10395 + }, + { + "epoch": 0.885896889646357, + "grad_norm": 13.4375, + "learning_rate": 1.4820474944590426e-06, + "loss": 0.6323, + "step": 10396 + }, + { + "epoch": 0.885982104814657, + "grad_norm": 14.25, + "learning_rate": 1.4818823775719777e-06, + "loss": 0.815, + "step": 10397 + }, + { + "epoch": 0.886067319982957, + "grad_norm": 17.5, + "learning_rate": 1.4817172564945648e-06, + "loss": 0.7762, + "step": 10398 + }, + { + "epoch": 0.8861525351512569, + "grad_norm": 17.875, + "learning_rate": 1.4815521312297882e-06, + "loss": 0.71, + "step": 10399 + }, + { + "epoch": 0.8862377503195569, + "grad_norm": 16.625, + "learning_rate": 1.481387001780632e-06, + "loss": 0.7437, + "step": 10400 + }, + { + "epoch": 0.8863229654878568, + "grad_norm": 19.25, + "learning_rate": 1.4812218681500794e-06, + "loss": 0.9264, + "step": 10401 + }, + { + "epoch": 0.8864081806561568, + "grad_norm": 18.875, + "learning_rate": 1.4810567303411155e-06, + "loss": 0.8544, + "step": 10402 + }, + { + "epoch": 0.8864933958244567, + "grad_norm": 17.375, + "learning_rate": 1.4808915883567244e-06, + "loss": 1.0451, + "step": 10403 + }, + { + "epoch": 0.8865786109927567, + "grad_norm": 15.125, + "learning_rate": 1.4807264421998899e-06, + "loss": 0.851, + "step": 10404 + }, + { + "epoch": 0.8866638261610567, + "grad_norm": 16.375, + "learning_rate": 1.4805612918735963e-06, + "loss": 0.9114, + "step": 10405 + }, + { + "epoch": 0.8867490413293566, + "grad_norm": 18.75, + "learning_rate": 1.4803961373808285e-06, + "loss": 0.7644, + "step": 10406 + }, + { + "epoch": 0.8868342564976566, + "grad_norm": 14.375, + "learning_rate": 1.4802309787245708e-06, + "loss": 0.5669, + "step": 10407 + }, + { + "epoch": 0.8869194716659565, + "grad_norm": 11.3125, + "learning_rate": 1.480065815907808e-06, + "loss": 0.4577, + "step": 10408 + }, + { + "epoch": 0.8870046868342565, + "grad_norm": 17.875, + "learning_rate": 1.4799006489335244e-06, + "loss": 1.0599, + "step": 10409 + }, + { + "epoch": 0.8870899020025564, + "grad_norm": 10.0, + "learning_rate": 1.4797354778047048e-06, + "loss": 0.3546, + "step": 10410 + }, + { + "epoch": 0.8871751171708564, + "grad_norm": 22.5, + "learning_rate": 1.4795703025243344e-06, + "loss": 1.0385, + "step": 10411 + }, + { + "epoch": 0.8872603323391564, + "grad_norm": 11.0, + "learning_rate": 1.4794051230953973e-06, + "loss": 0.5235, + "step": 10412 + }, + { + "epoch": 0.8873455475074563, + "grad_norm": 13.75, + "learning_rate": 1.4792399395208792e-06, + "loss": 0.7556, + "step": 10413 + }, + { + "epoch": 0.8874307626757563, + "grad_norm": 13.375, + "learning_rate": 1.4790747518037652e-06, + "loss": 0.7138, + "step": 10414 + }, + { + "epoch": 0.8875159778440562, + "grad_norm": 14.875, + "learning_rate": 1.4789095599470396e-06, + "loss": 0.9096, + "step": 10415 + }, + { + "epoch": 0.8876011930123562, + "grad_norm": 17.125, + "learning_rate": 1.478744363953688e-06, + "loss": 0.8506, + "step": 10416 + }, + { + "epoch": 0.8876864081806561, + "grad_norm": 13.625, + "learning_rate": 1.4785791638266961e-06, + "loss": 0.5188, + "step": 10417 + }, + { + "epoch": 0.8877716233489561, + "grad_norm": 15.8125, + "learning_rate": 1.4784139595690486e-06, + "loss": 0.7408, + "step": 10418 + }, + { + "epoch": 0.887856838517256, + "grad_norm": 20.875, + "learning_rate": 1.4782487511837314e-06, + "loss": 0.9588, + "step": 10419 + }, + { + "epoch": 0.887942053685556, + "grad_norm": 13.4375, + "learning_rate": 1.4780835386737297e-06, + "loss": 0.7271, + "step": 10420 + }, + { + "epoch": 0.888027268853856, + "grad_norm": 12.75, + "learning_rate": 1.477918322042029e-06, + "loss": 0.6001, + "step": 10421 + }, + { + "epoch": 0.8881124840221559, + "grad_norm": 13.0, + "learning_rate": 1.4777531012916147e-06, + "loss": 0.5802, + "step": 10422 + }, + { + "epoch": 0.8881976991904559, + "grad_norm": 11.1875, + "learning_rate": 1.4775878764254737e-06, + "loss": 0.4183, + "step": 10423 + }, + { + "epoch": 0.8882829143587558, + "grad_norm": 28.375, + "learning_rate": 1.4774226474465902e-06, + "loss": 1.2556, + "step": 10424 + }, + { + "epoch": 0.8883681295270558, + "grad_norm": 14.5, + "learning_rate": 1.477257414357951e-06, + "loss": 0.9505, + "step": 10425 + }, + { + "epoch": 0.8884533446953558, + "grad_norm": 13.625, + "learning_rate": 1.4770921771625415e-06, + "loss": 0.6335, + "step": 10426 + }, + { + "epoch": 0.8885385598636557, + "grad_norm": 14.1875, + "learning_rate": 1.4769269358633485e-06, + "loss": 0.6803, + "step": 10427 + }, + { + "epoch": 0.8886237750319557, + "grad_norm": 13.625, + "learning_rate": 1.4767616904633572e-06, + "loss": 0.6328, + "step": 10428 + }, + { + "epoch": 0.8887089902002556, + "grad_norm": 19.75, + "learning_rate": 1.4765964409655544e-06, + "loss": 0.8183, + "step": 10429 + }, + { + "epoch": 0.8887942053685556, + "grad_norm": 18.75, + "learning_rate": 1.4764311873729259e-06, + "loss": 0.8082, + "step": 10430 + }, + { + "epoch": 0.8888794205368555, + "grad_norm": 12.5625, + "learning_rate": 1.4762659296884581e-06, + "loss": 0.6917, + "step": 10431 + }, + { + "epoch": 0.8889646357051555, + "grad_norm": 17.375, + "learning_rate": 1.4761006679151377e-06, + "loss": 0.4813, + "step": 10432 + }, + { + "epoch": 0.8890498508734554, + "grad_norm": 17.0, + "learning_rate": 1.4759354020559501e-06, + "loss": 0.9395, + "step": 10433 + }, + { + "epoch": 0.8891350660417554, + "grad_norm": 13.375, + "learning_rate": 1.4757701321138834e-06, + "loss": 0.4614, + "step": 10434 + }, + { + "epoch": 0.8892202812100554, + "grad_norm": 19.875, + "learning_rate": 1.475604858091923e-06, + "loss": 0.5981, + "step": 10435 + }, + { + "epoch": 0.8893054963783553, + "grad_norm": 11.75, + "learning_rate": 1.4754395799930565e-06, + "loss": 0.4949, + "step": 10436 + }, + { + "epoch": 0.8893907115466553, + "grad_norm": 15.75, + "learning_rate": 1.4752742978202694e-06, + "loss": 0.7965, + "step": 10437 + }, + { + "epoch": 0.8894759267149552, + "grad_norm": 16.125, + "learning_rate": 1.4751090115765495e-06, + "loss": 0.8804, + "step": 10438 + }, + { + "epoch": 0.8895611418832552, + "grad_norm": 15.3125, + "learning_rate": 1.4749437212648832e-06, + "loss": 0.8737, + "step": 10439 + }, + { + "epoch": 0.8896463570515551, + "grad_norm": 14.9375, + "learning_rate": 1.4747784268882584e-06, + "loss": 0.5247, + "step": 10440 + }, + { + "epoch": 0.8897315722198551, + "grad_norm": 8.9375, + "learning_rate": 1.4746131284496606e-06, + "loss": 0.2444, + "step": 10441 + }, + { + "epoch": 0.8898167873881551, + "grad_norm": 17.5, + "learning_rate": 1.4744478259520777e-06, + "loss": 0.4327, + "step": 10442 + }, + { + "epoch": 0.889902002556455, + "grad_norm": 19.375, + "learning_rate": 1.4742825193984973e-06, + "loss": 0.9235, + "step": 10443 + }, + { + "epoch": 0.889987217724755, + "grad_norm": 16.25, + "learning_rate": 1.474117208791906e-06, + "loss": 0.8009, + "step": 10444 + }, + { + "epoch": 0.8900724328930549, + "grad_norm": 12.25, + "learning_rate": 1.4739518941352915e-06, + "loss": 0.5349, + "step": 10445 + }, + { + "epoch": 0.8901576480613549, + "grad_norm": 14.4375, + "learning_rate": 1.473786575431641e-06, + "loss": 0.7068, + "step": 10446 + }, + { + "epoch": 0.8902428632296548, + "grad_norm": 21.625, + "learning_rate": 1.4736212526839418e-06, + "loss": 0.6457, + "step": 10447 + }, + { + "epoch": 0.8903280783979548, + "grad_norm": 35.5, + "learning_rate": 1.4734559258951822e-06, + "loss": 1.0247, + "step": 10448 + }, + { + "epoch": 0.8904132935662548, + "grad_norm": 14.0625, + "learning_rate": 1.473290595068349e-06, + "loss": 0.7623, + "step": 10449 + }, + { + "epoch": 0.8904985087345547, + "grad_norm": 10.75, + "learning_rate": 1.4731252602064303e-06, + "loss": 0.3704, + "step": 10450 + }, + { + "epoch": 0.8905837239028547, + "grad_norm": 15.3125, + "learning_rate": 1.4729599213124138e-06, + "loss": 0.8652, + "step": 10451 + }, + { + "epoch": 0.8906689390711546, + "grad_norm": 14.9375, + "learning_rate": 1.4727945783892872e-06, + "loss": 1.0799, + "step": 10452 + }, + { + "epoch": 0.8907541542394546, + "grad_norm": 14.9375, + "learning_rate": 1.4726292314400386e-06, + "loss": 0.5508, + "step": 10453 + }, + { + "epoch": 0.8908393694077545, + "grad_norm": 16.5, + "learning_rate": 1.4724638804676557e-06, + "loss": 0.9287, + "step": 10454 + }, + { + "epoch": 0.8909245845760545, + "grad_norm": 14.5, + "learning_rate": 1.4722985254751274e-06, + "loss": 0.7868, + "step": 10455 + }, + { + "epoch": 0.8910097997443545, + "grad_norm": 17.0, + "learning_rate": 1.4721331664654407e-06, + "loss": 0.8529, + "step": 10456 + }, + { + "epoch": 0.8910950149126544, + "grad_norm": 13.5625, + "learning_rate": 1.4719678034415846e-06, + "loss": 0.6699, + "step": 10457 + }, + { + "epoch": 0.8911802300809544, + "grad_norm": 15.4375, + "learning_rate": 1.4718024364065468e-06, + "loss": 0.7054, + "step": 10458 + }, + { + "epoch": 0.8912654452492543, + "grad_norm": 13.3125, + "learning_rate": 1.471637065363316e-06, + "loss": 0.6142, + "step": 10459 + }, + { + "epoch": 0.8913506604175543, + "grad_norm": 14.8125, + "learning_rate": 1.4714716903148807e-06, + "loss": 0.7312, + "step": 10460 + }, + { + "epoch": 0.8914358755858542, + "grad_norm": 19.25, + "learning_rate": 1.4713063112642289e-06, + "loss": 0.4958, + "step": 10461 + }, + { + "epoch": 0.8915210907541542, + "grad_norm": 24.5, + "learning_rate": 1.4711409282143495e-06, + "loss": 0.79, + "step": 10462 + }, + { + "epoch": 0.8916063059224542, + "grad_norm": 13.6875, + "learning_rate": 1.4709755411682316e-06, + "loss": 0.6834, + "step": 10463 + }, + { + "epoch": 0.8916915210907541, + "grad_norm": 14.875, + "learning_rate": 1.4708101501288635e-06, + "loss": 0.8516, + "step": 10464 + }, + { + "epoch": 0.8917767362590541, + "grad_norm": 14.25, + "learning_rate": 1.4706447550992333e-06, + "loss": 0.4842, + "step": 10465 + }, + { + "epoch": 0.891861951427354, + "grad_norm": 16.75, + "learning_rate": 1.470479356082331e-06, + "loss": 0.7353, + "step": 10466 + }, + { + "epoch": 0.891947166595654, + "grad_norm": 13.25, + "learning_rate": 1.470313953081145e-06, + "loss": 0.6872, + "step": 10467 + }, + { + "epoch": 0.8920323817639539, + "grad_norm": 21.5, + "learning_rate": 1.4701485460986646e-06, + "loss": 1.0796, + "step": 10468 + }, + { + "epoch": 0.8921175969322539, + "grad_norm": 15.1875, + "learning_rate": 1.469983135137878e-06, + "loss": 0.8546, + "step": 10469 + }, + { + "epoch": 0.8922028121005539, + "grad_norm": 15.125, + "learning_rate": 1.4698177202017751e-06, + "loss": 1.0181, + "step": 10470 + }, + { + "epoch": 0.8922880272688538, + "grad_norm": 15.0, + "learning_rate": 1.469652301293345e-06, + "loss": 0.5778, + "step": 10471 + }, + { + "epoch": 0.8923732424371538, + "grad_norm": 12.0625, + "learning_rate": 1.4694868784155776e-06, + "loss": 0.793, + "step": 10472 + }, + { + "epoch": 0.8924584576054537, + "grad_norm": 15.4375, + "learning_rate": 1.4693214515714604e-06, + "loss": 0.9842, + "step": 10473 + }, + { + "epoch": 0.8925436727737537, + "grad_norm": 16.5, + "learning_rate": 1.469156020763985e-06, + "loss": 0.9131, + "step": 10474 + }, + { + "epoch": 0.8926288879420536, + "grad_norm": 16.5, + "learning_rate": 1.4689905859961392e-06, + "loss": 1.0074, + "step": 10475 + }, + { + "epoch": 0.8927141031103536, + "grad_norm": 9.0625, + "learning_rate": 1.4688251472709138e-06, + "loss": 0.2624, + "step": 10476 + }, + { + "epoch": 0.8927993182786536, + "grad_norm": 15.125, + "learning_rate": 1.4686597045912978e-06, + "loss": 0.7598, + "step": 10477 + }, + { + "epoch": 0.8928845334469535, + "grad_norm": 13.9375, + "learning_rate": 1.4684942579602809e-06, + "loss": 0.5933, + "step": 10478 + }, + { + "epoch": 0.8929697486152535, + "grad_norm": 13.375, + "learning_rate": 1.4683288073808533e-06, + "loss": 0.6608, + "step": 10479 + }, + { + "epoch": 0.8930549637835534, + "grad_norm": 17.75, + "learning_rate": 1.4681633528560043e-06, + "loss": 0.7335, + "step": 10480 + }, + { + "epoch": 0.8931401789518534, + "grad_norm": 19.125, + "learning_rate": 1.4679978943887245e-06, + "loss": 1.1301, + "step": 10481 + }, + { + "epoch": 0.8932253941201533, + "grad_norm": 13.3125, + "learning_rate": 1.4678324319820032e-06, + "loss": 0.8879, + "step": 10482 + }, + { + "epoch": 0.8933106092884533, + "grad_norm": 16.875, + "learning_rate": 1.4676669656388308e-06, + "loss": 0.793, + "step": 10483 + }, + { + "epoch": 0.8933958244567533, + "grad_norm": 17.75, + "learning_rate": 1.467501495362198e-06, + "loss": 0.6204, + "step": 10484 + }, + { + "epoch": 0.8934810396250532, + "grad_norm": 12.9375, + "learning_rate": 1.467336021155094e-06, + "loss": 0.6806, + "step": 10485 + }, + { + "epoch": 0.8935662547933532, + "grad_norm": 19.125, + "learning_rate": 1.4671705430205093e-06, + "loss": 1.0269, + "step": 10486 + }, + { + "epoch": 0.8936514699616531, + "grad_norm": 13.9375, + "learning_rate": 1.467005060961435e-06, + "loss": 0.7571, + "step": 10487 + }, + { + "epoch": 0.8937366851299531, + "grad_norm": 14.0625, + "learning_rate": 1.4668395749808613e-06, + "loss": 0.6106, + "step": 10488 + }, + { + "epoch": 0.893821900298253, + "grad_norm": 15.25, + "learning_rate": 1.4666740850817778e-06, + "loss": 0.9196, + "step": 10489 + }, + { + "epoch": 0.893907115466553, + "grad_norm": 16.5, + "learning_rate": 1.4665085912671756e-06, + "loss": 0.5685, + "step": 10490 + }, + { + "epoch": 0.893992330634853, + "grad_norm": 17.0, + "learning_rate": 1.4663430935400457e-06, + "loss": 0.5427, + "step": 10491 + }, + { + "epoch": 0.8940775458031529, + "grad_norm": 28.5, + "learning_rate": 1.4661775919033788e-06, + "loss": 0.7841, + "step": 10492 + }, + { + "epoch": 0.8941627609714529, + "grad_norm": 11.1875, + "learning_rate": 1.4660120863601654e-06, + "loss": 0.5337, + "step": 10493 + }, + { + "epoch": 0.8942479761397529, + "grad_norm": 14.1875, + "learning_rate": 1.4658465769133958e-06, + "loss": 0.7457, + "step": 10494 + }, + { + "epoch": 0.8943331913080529, + "grad_norm": 26.625, + "learning_rate": 1.465681063566062e-06, + "loss": 1.1096, + "step": 10495 + }, + { + "epoch": 0.8944184064763528, + "grad_norm": 12.0625, + "learning_rate": 1.4655155463211545e-06, + "loss": 0.5296, + "step": 10496 + }, + { + "epoch": 0.8945036216446528, + "grad_norm": 19.25, + "learning_rate": 1.465350025181664e-06, + "loss": 0.5372, + "step": 10497 + }, + { + "epoch": 0.8945888368129528, + "grad_norm": 14.4375, + "learning_rate": 1.4651845001505824e-06, + "loss": 0.5976, + "step": 10498 + }, + { + "epoch": 0.8946740519812527, + "grad_norm": 14.875, + "learning_rate": 1.4650189712309004e-06, + "loss": 0.5368, + "step": 10499 + }, + { + "epoch": 0.8947592671495527, + "grad_norm": 16.5, + "learning_rate": 1.4648534384256093e-06, + "loss": 0.8193, + "step": 10500 + }, + { + "epoch": 0.8948444823178526, + "grad_norm": 11.25, + "learning_rate": 1.4646879017377005e-06, + "loss": 0.4957, + "step": 10501 + }, + { + "epoch": 0.8949296974861526, + "grad_norm": 14.0625, + "learning_rate": 1.4645223611701656e-06, + "loss": 0.6787, + "step": 10502 + }, + { + "epoch": 0.8950149126544525, + "grad_norm": 15.0, + "learning_rate": 1.4643568167259955e-06, + "loss": 0.5931, + "step": 10503 + }, + { + "epoch": 0.8951001278227525, + "grad_norm": 14.8125, + "learning_rate": 1.4641912684081826e-06, + "loss": 0.747, + "step": 10504 + }, + { + "epoch": 0.8951853429910525, + "grad_norm": 11.375, + "learning_rate": 1.464025716219718e-06, + "loss": 0.6261, + "step": 10505 + }, + { + "epoch": 0.8952705581593524, + "grad_norm": 14.875, + "learning_rate": 1.4638601601635936e-06, + "loss": 0.3864, + "step": 10506 + }, + { + "epoch": 0.8953557733276524, + "grad_norm": 19.25, + "learning_rate": 1.4636946002428007e-06, + "loss": 1.0076, + "step": 10507 + }, + { + "epoch": 0.8954409884959523, + "grad_norm": 18.0, + "learning_rate": 1.4635290364603316e-06, + "loss": 0.7457, + "step": 10508 + }, + { + "epoch": 0.8955262036642523, + "grad_norm": 15.0, + "learning_rate": 1.4633634688191784e-06, + "loss": 0.763, + "step": 10509 + }, + { + "epoch": 0.8956114188325522, + "grad_norm": 14.25, + "learning_rate": 1.4631978973223326e-06, + "loss": 0.7147, + "step": 10510 + }, + { + "epoch": 0.8956966340008522, + "grad_norm": 13.8125, + "learning_rate": 1.463032321972786e-06, + "loss": 0.9861, + "step": 10511 + }, + { + "epoch": 0.8957818491691522, + "grad_norm": 16.5, + "learning_rate": 1.4628667427735315e-06, + "loss": 0.8539, + "step": 10512 + }, + { + "epoch": 0.8958670643374521, + "grad_norm": 17.5, + "learning_rate": 1.4627011597275611e-06, + "loss": 0.7863, + "step": 10513 + }, + { + "epoch": 0.8959522795057521, + "grad_norm": 13.25, + "learning_rate": 1.4625355728378666e-06, + "loss": 0.6729, + "step": 10514 + }, + { + "epoch": 0.896037494674052, + "grad_norm": 16.375, + "learning_rate": 1.4623699821074405e-06, + "loss": 0.8672, + "step": 10515 + }, + { + "epoch": 0.896122709842352, + "grad_norm": 13.8125, + "learning_rate": 1.4622043875392755e-06, + "loss": 0.4958, + "step": 10516 + }, + { + "epoch": 0.8962079250106519, + "grad_norm": 15.25, + "learning_rate": 1.4620387891363636e-06, + "loss": 0.8627, + "step": 10517 + }, + { + "epoch": 0.8962931401789519, + "grad_norm": 18.25, + "learning_rate": 1.4618731869016976e-06, + "loss": 0.667, + "step": 10518 + }, + { + "epoch": 0.8963783553472519, + "grad_norm": 19.875, + "learning_rate": 1.4617075808382703e-06, + "loss": 1.2579, + "step": 10519 + }, + { + "epoch": 0.8964635705155518, + "grad_norm": 13.0625, + "learning_rate": 1.4615419709490736e-06, + "loss": 0.5814, + "step": 10520 + }, + { + "epoch": 0.8965487856838518, + "grad_norm": 13.625, + "learning_rate": 1.4613763572371015e-06, + "loss": 0.8294, + "step": 10521 + }, + { + "epoch": 0.8966340008521517, + "grad_norm": 12.25, + "learning_rate": 1.4612107397053455e-06, + "loss": 0.6992, + "step": 10522 + }, + { + "epoch": 0.8967192160204517, + "grad_norm": 13.375, + "learning_rate": 1.4610451183567991e-06, + "loss": 0.6981, + "step": 10523 + }, + { + "epoch": 0.8968044311887516, + "grad_norm": 10.3125, + "learning_rate": 1.4608794931944554e-06, + "loss": 0.2513, + "step": 10524 + }, + { + "epoch": 0.8968896463570516, + "grad_norm": 15.5, + "learning_rate": 1.4607138642213072e-06, + "loss": 0.7937, + "step": 10525 + }, + { + "epoch": 0.8969748615253516, + "grad_norm": 14.9375, + "learning_rate": 1.4605482314403474e-06, + "loss": 0.7128, + "step": 10526 + }, + { + "epoch": 0.8970600766936515, + "grad_norm": 17.5, + "learning_rate": 1.4603825948545694e-06, + "loss": 0.9293, + "step": 10527 + }, + { + "epoch": 0.8971452918619515, + "grad_norm": 19.125, + "learning_rate": 1.4602169544669662e-06, + "loss": 1.0681, + "step": 10528 + }, + { + "epoch": 0.8972305070302514, + "grad_norm": 17.75, + "learning_rate": 1.460051310280532e-06, + "loss": 0.8572, + "step": 10529 + }, + { + "epoch": 0.8973157221985514, + "grad_norm": 21.0, + "learning_rate": 1.4598856622982586e-06, + "loss": 0.9751, + "step": 10530 + }, + { + "epoch": 0.8974009373668513, + "grad_norm": 15.0625, + "learning_rate": 1.4597200105231404e-06, + "loss": 0.7405, + "step": 10531 + }, + { + "epoch": 0.8974861525351513, + "grad_norm": 17.25, + "learning_rate": 1.4595543549581706e-06, + "loss": 0.7768, + "step": 10532 + }, + { + "epoch": 0.8975713677034513, + "grad_norm": 11.4375, + "learning_rate": 1.459388695606343e-06, + "loss": 0.5918, + "step": 10533 + }, + { + "epoch": 0.8976565828717512, + "grad_norm": 13.9375, + "learning_rate": 1.4592230324706514e-06, + "loss": 0.4957, + "step": 10534 + }, + { + "epoch": 0.8977417980400512, + "grad_norm": 18.125, + "learning_rate": 1.4590573655540887e-06, + "loss": 0.9917, + "step": 10535 + }, + { + "epoch": 0.8978270132083511, + "grad_norm": 18.5, + "learning_rate": 1.4588916948596496e-06, + "loss": 0.942, + "step": 10536 + }, + { + "epoch": 0.8979122283766511, + "grad_norm": 16.75, + "learning_rate": 1.4587260203903274e-06, + "loss": 0.9032, + "step": 10537 + }, + { + "epoch": 0.897997443544951, + "grad_norm": 14.875, + "learning_rate": 1.4585603421491162e-06, + "loss": 0.7249, + "step": 10538 + }, + { + "epoch": 0.898082658713251, + "grad_norm": 13.6875, + "learning_rate": 1.4583946601390095e-06, + "loss": 0.7009, + "step": 10539 + }, + { + "epoch": 0.898167873881551, + "grad_norm": 17.0, + "learning_rate": 1.458228974363002e-06, + "loss": 0.9629, + "step": 10540 + }, + { + "epoch": 0.8982530890498509, + "grad_norm": 14.5625, + "learning_rate": 1.4580632848240877e-06, + "loss": 0.4967, + "step": 10541 + }, + { + "epoch": 0.8983383042181509, + "grad_norm": 13.8125, + "learning_rate": 1.4578975915252607e-06, + "loss": 0.6625, + "step": 10542 + }, + { + "epoch": 0.8984235193864508, + "grad_norm": 15.375, + "learning_rate": 1.4577318944695145e-06, + "loss": 0.949, + "step": 10543 + }, + { + "epoch": 0.8985087345547508, + "grad_norm": 12.625, + "learning_rate": 1.4575661936598447e-06, + "loss": 0.5023, + "step": 10544 + }, + { + "epoch": 0.8985939497230507, + "grad_norm": 16.75, + "learning_rate": 1.4574004890992451e-06, + "loss": 0.7043, + "step": 10545 + }, + { + "epoch": 0.8986791648913507, + "grad_norm": 20.125, + "learning_rate": 1.45723478079071e-06, + "loss": 0.6524, + "step": 10546 + }, + { + "epoch": 0.8987643800596506, + "grad_norm": 16.0, + "learning_rate": 1.4570690687372338e-06, + "loss": 0.7655, + "step": 10547 + }, + { + "epoch": 0.8988495952279506, + "grad_norm": 25.25, + "learning_rate": 1.4569033529418113e-06, + "loss": 1.3418, + "step": 10548 + }, + { + "epoch": 0.8989348103962506, + "grad_norm": 13.3125, + "learning_rate": 1.4567376334074372e-06, + "loss": 0.6286, + "step": 10549 + }, + { + "epoch": 0.8990200255645505, + "grad_norm": 11.75, + "learning_rate": 1.4565719101371062e-06, + "loss": 0.4998, + "step": 10550 + }, + { + "epoch": 0.8991052407328505, + "grad_norm": 17.0, + "learning_rate": 1.4564061831338132e-06, + "loss": 0.8654, + "step": 10551 + }, + { + "epoch": 0.8991904559011504, + "grad_norm": 18.25, + "learning_rate": 1.4562404524005527e-06, + "loss": 0.9501, + "step": 10552 + }, + { + "epoch": 0.8992756710694504, + "grad_norm": 12.875, + "learning_rate": 1.4560747179403203e-06, + "loss": 0.5746, + "step": 10553 + }, + { + "epoch": 0.8993608862377503, + "grad_norm": 18.0, + "learning_rate": 1.4559089797561099e-06, + "loss": 1.26, + "step": 10554 + }, + { + "epoch": 0.8994461014060503, + "grad_norm": 16.5, + "learning_rate": 1.4557432378509176e-06, + "loss": 0.4765, + "step": 10555 + }, + { + "epoch": 0.8995313165743503, + "grad_norm": 15.75, + "learning_rate": 1.4555774922277377e-06, + "loss": 0.3654, + "step": 10556 + }, + { + "epoch": 0.8996165317426502, + "grad_norm": 14.8125, + "learning_rate": 1.4554117428895664e-06, + "loss": 0.679, + "step": 10557 + }, + { + "epoch": 0.8997017469109502, + "grad_norm": 12.1875, + "learning_rate": 1.4552459898393977e-06, + "loss": 0.4775, + "step": 10558 + }, + { + "epoch": 0.8997869620792501, + "grad_norm": 12.9375, + "learning_rate": 1.455080233080228e-06, + "loss": 0.3762, + "step": 10559 + }, + { + "epoch": 0.8998721772475501, + "grad_norm": 16.875, + "learning_rate": 1.4549144726150522e-06, + "loss": 0.7119, + "step": 10560 + }, + { + "epoch": 0.89995739241585, + "grad_norm": 19.625, + "learning_rate": 1.4547487084468659e-06, + "loss": 0.9674, + "step": 10561 + }, + { + "epoch": 0.90004260758415, + "grad_norm": 12.6875, + "learning_rate": 1.4545829405786643e-06, + "loss": 0.7943, + "step": 10562 + }, + { + "epoch": 0.90012782275245, + "grad_norm": 12.125, + "learning_rate": 1.4544171690134434e-06, + "loss": 0.5507, + "step": 10563 + }, + { + "epoch": 0.9002130379207499, + "grad_norm": 11.375, + "learning_rate": 1.4542513937541983e-06, + "loss": 0.3777, + "step": 10564 + }, + { + "epoch": 0.9002982530890499, + "grad_norm": 17.375, + "learning_rate": 1.4540856148039256e-06, + "loss": 0.6855, + "step": 10565 + }, + { + "epoch": 0.9003834682573498, + "grad_norm": 14.4375, + "learning_rate": 1.4539198321656207e-06, + "loss": 0.6312, + "step": 10566 + }, + { + "epoch": 0.9004686834256498, + "grad_norm": 16.5, + "learning_rate": 1.4537540458422788e-06, + "loss": 0.8699, + "step": 10567 + }, + { + "epoch": 0.9005538985939497, + "grad_norm": 14.1875, + "learning_rate": 1.453588255836897e-06, + "loss": 0.7147, + "step": 10568 + }, + { + "epoch": 0.9006391137622497, + "grad_norm": 13.5625, + "learning_rate": 1.4534224621524703e-06, + "loss": 0.6076, + "step": 10569 + }, + { + "epoch": 0.9007243289305497, + "grad_norm": 17.125, + "learning_rate": 1.4532566647919956e-06, + "loss": 1.1227, + "step": 10570 + }, + { + "epoch": 0.9008095440988496, + "grad_norm": 20.125, + "learning_rate": 1.453090863758468e-06, + "loss": 0.8747, + "step": 10571 + }, + { + "epoch": 0.9008947592671496, + "grad_norm": 16.125, + "learning_rate": 1.4529250590548846e-06, + "loss": 0.6937, + "step": 10572 + }, + { + "epoch": 0.9009799744354495, + "grad_norm": 15.75, + "learning_rate": 1.4527592506842411e-06, + "loss": 1.0556, + "step": 10573 + }, + { + "epoch": 0.9010651896037495, + "grad_norm": 14.5625, + "learning_rate": 1.4525934386495343e-06, + "loss": 0.666, + "step": 10574 + }, + { + "epoch": 0.9011504047720494, + "grad_norm": 18.25, + "learning_rate": 1.4524276229537605e-06, + "loss": 0.8782, + "step": 10575 + }, + { + "epoch": 0.9012356199403494, + "grad_norm": 14.8125, + "learning_rate": 1.4522618035999156e-06, + "loss": 0.5147, + "step": 10576 + }, + { + "epoch": 0.9013208351086494, + "grad_norm": 15.4375, + "learning_rate": 1.4520959805909966e-06, + "loss": 0.8113, + "step": 10577 + }, + { + "epoch": 0.9014060502769493, + "grad_norm": 16.5, + "learning_rate": 1.4519301539300005e-06, + "loss": 0.3953, + "step": 10578 + }, + { + "epoch": 0.9014912654452493, + "grad_norm": 12.6875, + "learning_rate": 1.4517643236199228e-06, + "loss": 0.5396, + "step": 10579 + }, + { + "epoch": 0.9015764806135492, + "grad_norm": 13.75, + "learning_rate": 1.4515984896637612e-06, + "loss": 0.7874, + "step": 10580 + }, + { + "epoch": 0.9016616957818492, + "grad_norm": 12.6875, + "learning_rate": 1.4514326520645116e-06, + "loss": 0.6227, + "step": 10581 + }, + { + "epoch": 0.9017469109501491, + "grad_norm": 14.625, + "learning_rate": 1.4512668108251722e-06, + "loss": 0.928, + "step": 10582 + }, + { + "epoch": 0.9018321261184491, + "grad_norm": 13.125, + "learning_rate": 1.4511009659487391e-06, + "loss": 0.4735, + "step": 10583 + }, + { + "epoch": 0.901917341286749, + "grad_norm": 9.9375, + "learning_rate": 1.450935117438209e-06, + "loss": 0.3723, + "step": 10584 + }, + { + "epoch": 0.902002556455049, + "grad_norm": 13.25, + "learning_rate": 1.450769265296579e-06, + "loss": 0.6483, + "step": 10585 + }, + { + "epoch": 0.902087771623349, + "grad_norm": 15.875, + "learning_rate": 1.450603409526847e-06, + "loss": 0.6936, + "step": 10586 + }, + { + "epoch": 0.9021729867916489, + "grad_norm": 24.0, + "learning_rate": 1.4504375501320098e-06, + "loss": 0.9486, + "step": 10587 + }, + { + "epoch": 0.9022582019599489, + "grad_norm": 10.625, + "learning_rate": 1.4502716871150637e-06, + "loss": 0.3607, + "step": 10588 + }, + { + "epoch": 0.9023434171282488, + "grad_norm": 12.0625, + "learning_rate": 1.4501058204790075e-06, + "loss": 0.8044, + "step": 10589 + }, + { + "epoch": 0.9024286322965488, + "grad_norm": 16.0, + "learning_rate": 1.449939950226838e-06, + "loss": 0.7381, + "step": 10590 + }, + { + "epoch": 0.9025138474648488, + "grad_norm": 12.1875, + "learning_rate": 1.4497740763615523e-06, + "loss": 0.5229, + "step": 10591 + }, + { + "epoch": 0.9025990626331487, + "grad_norm": 11.0, + "learning_rate": 1.449608198886148e-06, + "loss": 0.4299, + "step": 10592 + }, + { + "epoch": 0.9026842778014487, + "grad_norm": 20.0, + "learning_rate": 1.4494423178036229e-06, + "loss": 1.2262, + "step": 10593 + }, + { + "epoch": 0.9027694929697486, + "grad_norm": 14.875, + "learning_rate": 1.4492764331169748e-06, + "loss": 0.942, + "step": 10594 + }, + { + "epoch": 0.9028547081380486, + "grad_norm": 12.1875, + "learning_rate": 1.4491105448292012e-06, + "loss": 0.611, + "step": 10595 + }, + { + "epoch": 0.9029399233063485, + "grad_norm": 12.75, + "learning_rate": 1.4489446529432993e-06, + "loss": 0.5748, + "step": 10596 + }, + { + "epoch": 0.9030251384746485, + "grad_norm": 12.875, + "learning_rate": 1.4487787574622677e-06, + "loss": 0.4048, + "step": 10597 + }, + { + "epoch": 0.9031103536429484, + "grad_norm": 16.5, + "learning_rate": 1.4486128583891043e-06, + "loss": 0.9708, + "step": 10598 + }, + { + "epoch": 0.9031955688112484, + "grad_norm": 14.5, + "learning_rate": 1.448446955726807e-06, + "loss": 0.7211, + "step": 10599 + }, + { + "epoch": 0.9032807839795484, + "grad_norm": 15.9375, + "learning_rate": 1.448281049478373e-06, + "loss": 0.6689, + "step": 10600 + }, + { + "epoch": 0.9033659991478483, + "grad_norm": 13.3125, + "learning_rate": 1.4481151396468015e-06, + "loss": 0.5666, + "step": 10601 + }, + { + "epoch": 0.9034512143161483, + "grad_norm": 11.3125, + "learning_rate": 1.4479492262350902e-06, + "loss": 0.3463, + "step": 10602 + }, + { + "epoch": 0.9035364294844482, + "grad_norm": 11.1875, + "learning_rate": 1.447783309246237e-06, + "loss": 0.2918, + "step": 10603 + }, + { + "epoch": 0.9036216446527482, + "grad_norm": 15.0625, + "learning_rate": 1.447617388683241e-06, + "loss": 0.5017, + "step": 10604 + }, + { + "epoch": 0.9037068598210481, + "grad_norm": 14.5, + "learning_rate": 1.4474514645490995e-06, + "loss": 0.6685, + "step": 10605 + }, + { + "epoch": 0.9037920749893481, + "grad_norm": 16.5, + "learning_rate": 1.4472855368468122e-06, + "loss": 0.793, + "step": 10606 + }, + { + "epoch": 0.9038772901576481, + "grad_norm": 14.125, + "learning_rate": 1.4471196055793762e-06, + "loss": 0.5274, + "step": 10607 + }, + { + "epoch": 0.903962505325948, + "grad_norm": 18.625, + "learning_rate": 1.4469536707497912e-06, + "loss": 0.7363, + "step": 10608 + }, + { + "epoch": 0.904047720494248, + "grad_norm": 17.0, + "learning_rate": 1.4467877323610546e-06, + "loss": 0.8751, + "step": 10609 + }, + { + "epoch": 0.9041329356625479, + "grad_norm": 12.625, + "learning_rate": 1.4466217904161667e-06, + "loss": 0.7177, + "step": 10610 + }, + { + "epoch": 0.9042181508308479, + "grad_norm": 14.1875, + "learning_rate": 1.4464558449181248e-06, + "loss": 0.6765, + "step": 10611 + }, + { + "epoch": 0.9043033659991478, + "grad_norm": 15.5625, + "learning_rate": 1.4462898958699284e-06, + "loss": 0.7753, + "step": 10612 + }, + { + "epoch": 0.9043885811674478, + "grad_norm": 15.25, + "learning_rate": 1.4461239432745758e-06, + "loss": 0.6275, + "step": 10613 + }, + { + "epoch": 0.9044737963357478, + "grad_norm": 13.8125, + "learning_rate": 1.445957987135067e-06, + "loss": 0.5641, + "step": 10614 + }, + { + "epoch": 0.9045590115040477, + "grad_norm": 15.375, + "learning_rate": 1.4457920274544e-06, + "loss": 0.9015, + "step": 10615 + }, + { + "epoch": 0.9046442266723477, + "grad_norm": 13.6875, + "learning_rate": 1.4456260642355741e-06, + "loss": 0.7694, + "step": 10616 + }, + { + "epoch": 0.9047294418406476, + "grad_norm": 15.4375, + "learning_rate": 1.4454600974815884e-06, + "loss": 0.8692, + "step": 10617 + }, + { + "epoch": 0.9048146570089476, + "grad_norm": 16.375, + "learning_rate": 1.4452941271954424e-06, + "loss": 0.7682, + "step": 10618 + }, + { + "epoch": 0.9048998721772475, + "grad_norm": 13.125, + "learning_rate": 1.4451281533801351e-06, + "loss": 0.8098, + "step": 10619 + }, + { + "epoch": 0.9049850873455475, + "grad_norm": 12.5, + "learning_rate": 1.4449621760386657e-06, + "loss": 0.6211, + "step": 10620 + }, + { + "epoch": 0.9050703025138475, + "grad_norm": 13.3125, + "learning_rate": 1.444796195174034e-06, + "loss": 0.518, + "step": 10621 + }, + { + "epoch": 0.9051555176821474, + "grad_norm": 14.375, + "learning_rate": 1.444630210789239e-06, + "loss": 0.6669, + "step": 10622 + }, + { + "epoch": 0.9052407328504474, + "grad_norm": 12.5625, + "learning_rate": 1.4444642228872807e-06, + "loss": 0.5608, + "step": 10623 + }, + { + "epoch": 0.9053259480187473, + "grad_norm": 18.875, + "learning_rate": 1.444298231471158e-06, + "loss": 0.6591, + "step": 10624 + }, + { + "epoch": 0.9054111631870473, + "grad_norm": 33.5, + "learning_rate": 1.4441322365438709e-06, + "loss": 0.9929, + "step": 10625 + }, + { + "epoch": 0.9054963783553472, + "grad_norm": 13.625, + "learning_rate": 1.4439662381084191e-06, + "loss": 0.598, + "step": 10626 + }, + { + "epoch": 0.9055815935236472, + "grad_norm": 16.125, + "learning_rate": 1.4438002361678024e-06, + "loss": 0.9534, + "step": 10627 + }, + { + "epoch": 0.9056668086919472, + "grad_norm": 19.375, + "learning_rate": 1.4436342307250203e-06, + "loss": 1.1651, + "step": 10628 + }, + { + "epoch": 0.9057520238602471, + "grad_norm": 16.75, + "learning_rate": 1.4434682217830731e-06, + "loss": 0.8231, + "step": 10629 + }, + { + "epoch": 0.9058372390285471, + "grad_norm": 19.0, + "learning_rate": 1.4433022093449608e-06, + "loss": 1.0404, + "step": 10630 + }, + { + "epoch": 0.905922454196847, + "grad_norm": 13.25, + "learning_rate": 1.4431361934136832e-06, + "loss": 0.7809, + "step": 10631 + }, + { + "epoch": 0.906007669365147, + "grad_norm": 14.625, + "learning_rate": 1.4429701739922397e-06, + "loss": 0.7415, + "step": 10632 + }, + { + "epoch": 0.9060928845334469, + "grad_norm": 11.8125, + "learning_rate": 1.442804151083632e-06, + "loss": 0.4831, + "step": 10633 + }, + { + "epoch": 0.9061780997017469, + "grad_norm": 17.875, + "learning_rate": 1.442638124690859e-06, + "loss": 0.7217, + "step": 10634 + }, + { + "epoch": 0.9062633148700469, + "grad_norm": 15.9375, + "learning_rate": 1.442472094816921e-06, + "loss": 1.0018, + "step": 10635 + }, + { + "epoch": 0.9063485300383468, + "grad_norm": 16.0, + "learning_rate": 1.442306061464819e-06, + "loss": 0.9099, + "step": 10636 + }, + { + "epoch": 0.9064337452066468, + "grad_norm": 11.5625, + "learning_rate": 1.4421400246375533e-06, + "loss": 0.4625, + "step": 10637 + }, + { + "epoch": 0.9065189603749467, + "grad_norm": 15.25, + "learning_rate": 1.4419739843381239e-06, + "loss": 0.905, + "step": 10638 + }, + { + "epoch": 0.9066041755432467, + "grad_norm": 15.75, + "learning_rate": 1.4418079405695317e-06, + "loss": 0.9072, + "step": 10639 + }, + { + "epoch": 0.9066893907115466, + "grad_norm": 14.0, + "learning_rate": 1.4416418933347772e-06, + "loss": 0.5375, + "step": 10640 + }, + { + "epoch": 0.9067746058798466, + "grad_norm": 13.5, + "learning_rate": 1.441475842636861e-06, + "loss": 0.5738, + "step": 10641 + }, + { + "epoch": 0.9068598210481466, + "grad_norm": 20.0, + "learning_rate": 1.4413097884787835e-06, + "loss": 0.976, + "step": 10642 + }, + { + "epoch": 0.9069450362164465, + "grad_norm": 24.75, + "learning_rate": 1.441143730863546e-06, + "loss": 0.9979, + "step": 10643 + }, + { + "epoch": 0.9070302513847465, + "grad_norm": 37.75, + "learning_rate": 1.4409776697941491e-06, + "loss": 0.8972, + "step": 10644 + }, + { + "epoch": 0.9071154665530464, + "grad_norm": 16.0, + "learning_rate": 1.4408116052735932e-06, + "loss": 0.8906, + "step": 10645 + }, + { + "epoch": 0.9072006817213464, + "grad_norm": 11.125, + "learning_rate": 1.44064553730488e-06, + "loss": 0.4891, + "step": 10646 + }, + { + "epoch": 0.9072858968896463, + "grad_norm": 11.375, + "learning_rate": 1.4404794658910104e-06, + "loss": 0.4495, + "step": 10647 + }, + { + "epoch": 0.9073711120579463, + "grad_norm": 12.5625, + "learning_rate": 1.4403133910349854e-06, + "loss": 0.6476, + "step": 10648 + }, + { + "epoch": 0.9074563272262463, + "grad_norm": 10.875, + "learning_rate": 1.4401473127398054e-06, + "loss": 0.3401, + "step": 10649 + }, + { + "epoch": 0.9075415423945462, + "grad_norm": 14.5625, + "learning_rate": 1.4399812310084728e-06, + "loss": 1.1216, + "step": 10650 + }, + { + "epoch": 0.9076267575628462, + "grad_norm": 14.8125, + "learning_rate": 1.4398151458439884e-06, + "loss": 0.6491, + "step": 10651 + }, + { + "epoch": 0.9077119727311461, + "grad_norm": 20.125, + "learning_rate": 1.4396490572493535e-06, + "loss": 0.4412, + "step": 10652 + }, + { + "epoch": 0.9077971878994461, + "grad_norm": 11.375, + "learning_rate": 1.4394829652275688e-06, + "loss": 0.4415, + "step": 10653 + }, + { + "epoch": 0.907882403067746, + "grad_norm": 11.9375, + "learning_rate": 1.4393168697816367e-06, + "loss": 0.6512, + "step": 10654 + }, + { + "epoch": 0.907967618236046, + "grad_norm": 14.25, + "learning_rate": 1.439150770914559e-06, + "loss": 0.9739, + "step": 10655 + }, + { + "epoch": 0.908052833404346, + "grad_norm": 13.125, + "learning_rate": 1.4389846686293357e-06, + "loss": 0.51, + "step": 10656 + }, + { + "epoch": 0.9081380485726459, + "grad_norm": 12.5, + "learning_rate": 1.4388185629289697e-06, + "loss": 0.7369, + "step": 10657 + }, + { + "epoch": 0.9082232637409459, + "grad_norm": 20.375, + "learning_rate": 1.4386524538164626e-06, + "loss": 0.856, + "step": 10658 + }, + { + "epoch": 0.9083084789092458, + "grad_norm": 14.1875, + "learning_rate": 1.4384863412948161e-06, + "loss": 0.669, + "step": 10659 + }, + { + "epoch": 0.9083936940775458, + "grad_norm": 10.0625, + "learning_rate": 1.4383202253670315e-06, + "loss": 0.3547, + "step": 10660 + }, + { + "epoch": 0.9084789092458457, + "grad_norm": 12.625, + "learning_rate": 1.4381541060361111e-06, + "loss": 0.5269, + "step": 10661 + }, + { + "epoch": 0.9085641244141457, + "grad_norm": 13.0, + "learning_rate": 1.4379879833050568e-06, + "loss": 0.4384, + "step": 10662 + }, + { + "epoch": 0.9086493395824456, + "grad_norm": 17.375, + "learning_rate": 1.4378218571768709e-06, + "loss": 0.6667, + "step": 10663 + }, + { + "epoch": 0.9087345547507456, + "grad_norm": 13.125, + "learning_rate": 1.437655727654555e-06, + "loss": 0.564, + "step": 10664 + }, + { + "epoch": 0.9088197699190456, + "grad_norm": 14.4375, + "learning_rate": 1.4374895947411114e-06, + "loss": 1.0638, + "step": 10665 + }, + { + "epoch": 0.9089049850873455, + "grad_norm": 11.8125, + "learning_rate": 1.437323458439542e-06, + "loss": 0.6386, + "step": 10666 + }, + { + "epoch": 0.9089902002556455, + "grad_norm": 15.6875, + "learning_rate": 1.43715731875285e-06, + "loss": 0.5608, + "step": 10667 + }, + { + "epoch": 0.9090754154239454, + "grad_norm": 15.5625, + "learning_rate": 1.4369911756840365e-06, + "loss": 0.8517, + "step": 10668 + }, + { + "epoch": 0.9091606305922454, + "grad_norm": 16.125, + "learning_rate": 1.4368250292361047e-06, + "loss": 0.8857, + "step": 10669 + }, + { + "epoch": 0.9092458457605453, + "grad_norm": 13.8125, + "learning_rate": 1.436658879412056e-06, + "loss": 0.6905, + "step": 10670 + }, + { + "epoch": 0.9093310609288453, + "grad_norm": 14.3125, + "learning_rate": 1.4364927262148948e-06, + "loss": 0.6195, + "step": 10671 + }, + { + "epoch": 0.9094162760971453, + "grad_norm": 22.0, + "learning_rate": 1.436326569647622e-06, + "loss": 0.9549, + "step": 10672 + }, + { + "epoch": 0.9095014912654452, + "grad_norm": 18.5, + "learning_rate": 1.4361604097132402e-06, + "loss": 0.5507, + "step": 10673 + }, + { + "epoch": 0.9095867064337452, + "grad_norm": 45.0, + "learning_rate": 1.4359942464147533e-06, + "loss": 0.9017, + "step": 10674 + }, + { + "epoch": 0.9096719216020451, + "grad_norm": 14.0, + "learning_rate": 1.4358280797551633e-06, + "loss": 0.7672, + "step": 10675 + }, + { + "epoch": 0.9097571367703451, + "grad_norm": 18.0, + "learning_rate": 1.435661909737473e-06, + "loss": 0.7525, + "step": 10676 + }, + { + "epoch": 0.909842351938645, + "grad_norm": 13.9375, + "learning_rate": 1.4354957363646848e-06, + "loss": 0.757, + "step": 10677 + }, + { + "epoch": 0.909927567106945, + "grad_norm": 17.0, + "learning_rate": 1.4353295596398025e-06, + "loss": 0.9639, + "step": 10678 + }, + { + "epoch": 0.910012782275245, + "grad_norm": 12.5, + "learning_rate": 1.4351633795658287e-06, + "loss": 0.2647, + "step": 10679 + }, + { + "epoch": 0.9100979974435449, + "grad_norm": 31.875, + "learning_rate": 1.4349971961457666e-06, + "loss": 0.8639, + "step": 10680 + }, + { + "epoch": 0.9101832126118449, + "grad_norm": 12.875, + "learning_rate": 1.4348310093826187e-06, + "loss": 0.3345, + "step": 10681 + }, + { + "epoch": 0.9102684277801448, + "grad_norm": 18.75, + "learning_rate": 1.434664819279389e-06, + "loss": 1.0251, + "step": 10682 + }, + { + "epoch": 0.9103536429484448, + "grad_norm": 20.5, + "learning_rate": 1.4344986258390803e-06, + "loss": 0.8824, + "step": 10683 + }, + { + "epoch": 0.9104388581167447, + "grad_norm": 17.625, + "learning_rate": 1.4343324290646957e-06, + "loss": 1.1389, + "step": 10684 + }, + { + "epoch": 0.9105240732850447, + "grad_norm": 14.0, + "learning_rate": 1.4341662289592386e-06, + "loss": 0.8872, + "step": 10685 + }, + { + "epoch": 0.9106092884533447, + "grad_norm": 12.0, + "learning_rate": 1.4340000255257128e-06, + "loss": 0.3582, + "step": 10686 + }, + { + "epoch": 0.9106945036216446, + "grad_norm": 20.25, + "learning_rate": 1.4338338187671215e-06, + "loss": 0.8789, + "step": 10687 + }, + { + "epoch": 0.9107797187899446, + "grad_norm": 29.5, + "learning_rate": 1.433667608686468e-06, + "loss": 1.428, + "step": 10688 + }, + { + "epoch": 0.9108649339582445, + "grad_norm": 24.75, + "learning_rate": 1.4335013952867566e-06, + "loss": 0.9267, + "step": 10689 + }, + { + "epoch": 0.9109501491265445, + "grad_norm": 11.3125, + "learning_rate": 1.4333351785709898e-06, + "loss": 0.4721, + "step": 10690 + }, + { + "epoch": 0.9110353642948444, + "grad_norm": 9.4375, + "learning_rate": 1.4331689585421726e-06, + "loss": 0.3674, + "step": 10691 + }, + { + "epoch": 0.9111205794631444, + "grad_norm": 10.1875, + "learning_rate": 1.433002735203308e-06, + "loss": 0.2565, + "step": 10692 + }, + { + "epoch": 0.9112057946314444, + "grad_norm": 10.0625, + "learning_rate": 1.4328365085573997e-06, + "loss": 0.5713, + "step": 10693 + }, + { + "epoch": 0.9112910097997443, + "grad_norm": 16.625, + "learning_rate": 1.4326702786074518e-06, + "loss": 0.9657, + "step": 10694 + }, + { + "epoch": 0.9113762249680443, + "grad_norm": 25.5, + "learning_rate": 1.4325040453564687e-06, + "loss": 1.1145, + "step": 10695 + }, + { + "epoch": 0.9114614401363442, + "grad_norm": 13.3125, + "learning_rate": 1.4323378088074536e-06, + "loss": 0.7592, + "step": 10696 + }, + { + "epoch": 0.9115466553046442, + "grad_norm": 12.9375, + "learning_rate": 1.432171568963411e-06, + "loss": 0.5014, + "step": 10697 + }, + { + "epoch": 0.9116318704729441, + "grad_norm": 20.875, + "learning_rate": 1.432005325827345e-06, + "loss": 1.0508, + "step": 10698 + }, + { + "epoch": 0.9117170856412441, + "grad_norm": 11.5, + "learning_rate": 1.4318390794022602e-06, + "loss": 0.4104, + "step": 10699 + }, + { + "epoch": 0.911802300809544, + "grad_norm": 12.4375, + "learning_rate": 1.4316728296911598e-06, + "loss": 0.5074, + "step": 10700 + }, + { + "epoch": 0.911887515977844, + "grad_norm": 19.25, + "learning_rate": 1.4315065766970487e-06, + "loss": 1.1956, + "step": 10701 + }, + { + "epoch": 0.911972731146144, + "grad_norm": 12.875, + "learning_rate": 1.4313403204229313e-06, + "loss": 0.6013, + "step": 10702 + }, + { + "epoch": 0.9120579463144439, + "grad_norm": 24.5, + "learning_rate": 1.431174060871812e-06, + "loss": 0.9155, + "step": 10703 + }, + { + "epoch": 0.9121431614827439, + "grad_norm": 17.0, + "learning_rate": 1.4310077980466953e-06, + "loss": 1.2339, + "step": 10704 + }, + { + "epoch": 0.9122283766510438, + "grad_norm": 16.375, + "learning_rate": 1.4308415319505856e-06, + "loss": 0.8029, + "step": 10705 + }, + { + "epoch": 0.9123135918193438, + "grad_norm": 11.625, + "learning_rate": 1.4306752625864875e-06, + "loss": 0.4112, + "step": 10706 + }, + { + "epoch": 0.9123988069876438, + "grad_norm": 13.25, + "learning_rate": 1.430508989957406e-06, + "loss": 0.7702, + "step": 10707 + }, + { + "epoch": 0.9124840221559437, + "grad_norm": 17.125, + "learning_rate": 1.4303427140663454e-06, + "loss": 1.094, + "step": 10708 + }, + { + "epoch": 0.9125692373242437, + "grad_norm": 14.8125, + "learning_rate": 1.4301764349163102e-06, + "loss": 0.706, + "step": 10709 + }, + { + "epoch": 0.9126544524925436, + "grad_norm": 15.75, + "learning_rate": 1.4300101525103063e-06, + "loss": 0.6238, + "step": 10710 + }, + { + "epoch": 0.9127396676608436, + "grad_norm": 16.5, + "learning_rate": 1.4298438668513373e-06, + "loss": 0.6509, + "step": 10711 + }, + { + "epoch": 0.9128248828291435, + "grad_norm": 18.5, + "learning_rate": 1.4296775779424096e-06, + "loss": 0.8594, + "step": 10712 + }, + { + "epoch": 0.9129100979974435, + "grad_norm": 19.75, + "learning_rate": 1.4295112857865268e-06, + "loss": 1.0079, + "step": 10713 + }, + { + "epoch": 0.9129953131657434, + "grad_norm": 15.8125, + "learning_rate": 1.4293449903866946e-06, + "loss": 0.8087, + "step": 10714 + }, + { + "epoch": 0.9130805283340434, + "grad_norm": 23.25, + "learning_rate": 1.429178691745918e-06, + "loss": 1.0736, + "step": 10715 + }, + { + "epoch": 0.9131657435023434, + "grad_norm": 15.9375, + "learning_rate": 1.4290123898672026e-06, + "loss": 0.7095, + "step": 10716 + }, + { + "epoch": 0.9132509586706433, + "grad_norm": 14.9375, + "learning_rate": 1.428846084753553e-06, + "loss": 0.7069, + "step": 10717 + }, + { + "epoch": 0.9133361738389433, + "grad_norm": 17.0, + "learning_rate": 1.428679776407975e-06, + "loss": 0.8269, + "step": 10718 + }, + { + "epoch": 0.9134213890072432, + "grad_norm": 15.8125, + "learning_rate": 1.4285134648334733e-06, + "loss": 0.7666, + "step": 10719 + }, + { + "epoch": 0.9135066041755433, + "grad_norm": 15.3125, + "learning_rate": 1.4283471500330543e-06, + "loss": 0.6026, + "step": 10720 + }, + { + "epoch": 0.9135918193438433, + "grad_norm": 15.4375, + "learning_rate": 1.428180832009723e-06, + "loss": 0.971, + "step": 10721 + }, + { + "epoch": 0.9136770345121432, + "grad_norm": 16.125, + "learning_rate": 1.4280145107664848e-06, + "loss": 0.7324, + "step": 10722 + }, + { + "epoch": 0.9137622496804432, + "grad_norm": 14.3125, + "learning_rate": 1.4278481863063447e-06, + "loss": 0.8911, + "step": 10723 + }, + { + "epoch": 0.9138474648487431, + "grad_norm": 14.875, + "learning_rate": 1.4276818586323098e-06, + "loss": 0.8858, + "step": 10724 + }, + { + "epoch": 0.9139326800170431, + "grad_norm": 11.875, + "learning_rate": 1.427515527747385e-06, + "loss": 0.4187, + "step": 10725 + }, + { + "epoch": 0.914017895185343, + "grad_norm": 77.5, + "learning_rate": 1.4273491936545757e-06, + "loss": 0.636, + "step": 10726 + }, + { + "epoch": 0.914103110353643, + "grad_norm": 11.625, + "learning_rate": 1.4271828563568884e-06, + "loss": 0.3706, + "step": 10727 + }, + { + "epoch": 0.914188325521943, + "grad_norm": 15.375, + "learning_rate": 1.427016515857329e-06, + "loss": 0.9849, + "step": 10728 + }, + { + "epoch": 0.9142735406902429, + "grad_norm": 11.125, + "learning_rate": 1.4268501721589029e-06, + "loss": 0.5107, + "step": 10729 + }, + { + "epoch": 0.9143587558585429, + "grad_norm": 13.4375, + "learning_rate": 1.4266838252646162e-06, + "loss": 0.566, + "step": 10730 + }, + { + "epoch": 0.9144439710268428, + "grad_norm": 14.0, + "learning_rate": 1.4265174751774752e-06, + "loss": 1.0135, + "step": 10731 + }, + { + "epoch": 0.9145291861951428, + "grad_norm": 13.125, + "learning_rate": 1.426351121900486e-06, + "loss": 0.5943, + "step": 10732 + }, + { + "epoch": 0.9146144013634427, + "grad_norm": 20.0, + "learning_rate": 1.4261847654366547e-06, + "loss": 1.2011, + "step": 10733 + }, + { + "epoch": 0.9146996165317427, + "grad_norm": 13.0, + "learning_rate": 1.4260184057889872e-06, + "loss": 0.452, + "step": 10734 + }, + { + "epoch": 0.9147848317000427, + "grad_norm": 11.375, + "learning_rate": 1.4258520429604905e-06, + "loss": 0.9703, + "step": 10735 + }, + { + "epoch": 0.9148700468683426, + "grad_norm": 13.4375, + "learning_rate": 1.4256856769541704e-06, + "loss": 0.4704, + "step": 10736 + }, + { + "epoch": 0.9149552620366426, + "grad_norm": 13.4375, + "learning_rate": 1.4255193077730336e-06, + "loss": 0.6168, + "step": 10737 + }, + { + "epoch": 0.9150404772049425, + "grad_norm": 14.375, + "learning_rate": 1.4253529354200862e-06, + "loss": 0.4867, + "step": 10738 + }, + { + "epoch": 0.9151256923732425, + "grad_norm": 21.5, + "learning_rate": 1.425186559898335e-06, + "loss": 0.7618, + "step": 10739 + }, + { + "epoch": 0.9152109075415424, + "grad_norm": 25.75, + "learning_rate": 1.4250201812107866e-06, + "loss": 0.843, + "step": 10740 + }, + { + "epoch": 0.9152961227098424, + "grad_norm": 16.0, + "learning_rate": 1.4248537993604472e-06, + "loss": 0.5797, + "step": 10741 + }, + { + "epoch": 0.9153813378781424, + "grad_norm": 31.5, + "learning_rate": 1.4246874143503241e-06, + "loss": 0.8654, + "step": 10742 + }, + { + "epoch": 0.9154665530464423, + "grad_norm": 39.75, + "learning_rate": 1.4245210261834236e-06, + "loss": 1.0468, + "step": 10743 + }, + { + "epoch": 0.9155517682147423, + "grad_norm": 13.9375, + "learning_rate": 1.4243546348627529e-06, + "loss": 0.8127, + "step": 10744 + }, + { + "epoch": 0.9156369833830422, + "grad_norm": 13.875, + "learning_rate": 1.4241882403913184e-06, + "loss": 0.5697, + "step": 10745 + }, + { + "epoch": 0.9157221985513422, + "grad_norm": 12.1875, + "learning_rate": 1.4240218427721276e-06, + "loss": 0.4631, + "step": 10746 + }, + { + "epoch": 0.9158074137196421, + "grad_norm": 10.375, + "learning_rate": 1.4238554420081865e-06, + "loss": 0.6388, + "step": 10747 + }, + { + "epoch": 0.9158926288879421, + "grad_norm": 15.875, + "learning_rate": 1.4236890381025033e-06, + "loss": 0.7782, + "step": 10748 + }, + { + "epoch": 0.915977844056242, + "grad_norm": 11.3125, + "learning_rate": 1.4235226310580843e-06, + "loss": 0.2891, + "step": 10749 + }, + { + "epoch": 0.916063059224542, + "grad_norm": 13.875, + "learning_rate": 1.423356220877937e-06, + "loss": 0.6451, + "step": 10750 + }, + { + "epoch": 0.916148274392842, + "grad_norm": 12.875, + "learning_rate": 1.423189807565068e-06, + "loss": 0.6233, + "step": 10751 + }, + { + "epoch": 0.9162334895611419, + "grad_norm": 11.625, + "learning_rate": 1.4230233911224858e-06, + "loss": 0.5776, + "step": 10752 + }, + { + "epoch": 0.9163187047294419, + "grad_norm": 17.375, + "learning_rate": 1.4228569715531962e-06, + "loss": 0.8446, + "step": 10753 + }, + { + "epoch": 0.9164039198977418, + "grad_norm": 19.125, + "learning_rate": 1.4226905488602077e-06, + "loss": 0.8129, + "step": 10754 + }, + { + "epoch": 0.9164891350660418, + "grad_norm": 12.25, + "learning_rate": 1.422524123046527e-06, + "loss": 0.4956, + "step": 10755 + }, + { + "epoch": 0.9165743502343418, + "grad_norm": 14.25, + "learning_rate": 1.4223576941151624e-06, + "loss": 0.7465, + "step": 10756 + }, + { + "epoch": 0.9166595654026417, + "grad_norm": 13.25, + "learning_rate": 1.4221912620691206e-06, + "loss": 0.6157, + "step": 10757 + }, + { + "epoch": 0.9167447805709417, + "grad_norm": 16.75, + "learning_rate": 1.4220248269114093e-06, + "loss": 0.7655, + "step": 10758 + }, + { + "epoch": 0.9168299957392416, + "grad_norm": 18.375, + "learning_rate": 1.421858388645037e-06, + "loss": 0.9627, + "step": 10759 + }, + { + "epoch": 0.9169152109075416, + "grad_norm": 14.5, + "learning_rate": 1.4216919472730104e-06, + "loss": 0.6259, + "step": 10760 + }, + { + "epoch": 0.9170004260758415, + "grad_norm": 17.125, + "learning_rate": 1.421525502798338e-06, + "loss": 0.8324, + "step": 10761 + }, + { + "epoch": 0.9170856412441415, + "grad_norm": 13.9375, + "learning_rate": 1.4213590552240267e-06, + "loss": 0.8552, + "step": 10762 + }, + { + "epoch": 0.9171708564124414, + "grad_norm": 12.0, + "learning_rate": 1.4211926045530855e-06, + "loss": 0.4953, + "step": 10763 + }, + { + "epoch": 0.9172560715807414, + "grad_norm": 14.5, + "learning_rate": 1.4210261507885214e-06, + "loss": 0.6579, + "step": 10764 + }, + { + "epoch": 0.9173412867490414, + "grad_norm": 16.0, + "learning_rate": 1.4208596939333433e-06, + "loss": 0.6735, + "step": 10765 + }, + { + "epoch": 0.9174265019173413, + "grad_norm": 14.5625, + "learning_rate": 1.4206932339905582e-06, + "loss": 0.6644, + "step": 10766 + }, + { + "epoch": 0.9175117170856413, + "grad_norm": 16.875, + "learning_rate": 1.4205267709631748e-06, + "loss": 0.7312, + "step": 10767 + }, + { + "epoch": 0.9175969322539412, + "grad_norm": 15.3125, + "learning_rate": 1.4203603048542014e-06, + "loss": 0.6361, + "step": 10768 + }, + { + "epoch": 0.9176821474222412, + "grad_norm": 14.375, + "learning_rate": 1.4201938356666462e-06, + "loss": 0.6674, + "step": 10769 + }, + { + "epoch": 0.9177673625905411, + "grad_norm": 15.5, + "learning_rate": 1.4200273634035166e-06, + "loss": 0.9394, + "step": 10770 + }, + { + "epoch": 0.9178525777588411, + "grad_norm": 10.625, + "learning_rate": 1.4198608880678219e-06, + "loss": 0.934, + "step": 10771 + }, + { + "epoch": 0.9179377929271411, + "grad_norm": 11.75, + "learning_rate": 1.4196944096625698e-06, + "loss": 0.4415, + "step": 10772 + }, + { + "epoch": 0.918023008095441, + "grad_norm": 15.3125, + "learning_rate": 1.4195279281907696e-06, + "loss": 0.7838, + "step": 10773 + }, + { + "epoch": 0.918108223263741, + "grad_norm": 27.625, + "learning_rate": 1.4193614436554293e-06, + "loss": 0.7849, + "step": 10774 + }, + { + "epoch": 0.9181934384320409, + "grad_norm": 10.625, + "learning_rate": 1.419194956059557e-06, + "loss": 0.4729, + "step": 10775 + }, + { + "epoch": 0.9182786536003409, + "grad_norm": 13.625, + "learning_rate": 1.4190284654061617e-06, + "loss": 0.7227, + "step": 10776 + }, + { + "epoch": 0.9183638687686408, + "grad_norm": 16.625, + "learning_rate": 1.4188619716982523e-06, + "loss": 0.8909, + "step": 10777 + }, + { + "epoch": 0.9184490839369408, + "grad_norm": 14.375, + "learning_rate": 1.4186954749388373e-06, + "loss": 0.6256, + "step": 10778 + }, + { + "epoch": 0.9185342991052408, + "grad_norm": 14.25, + "learning_rate": 1.4185289751309253e-06, + "loss": 0.7952, + "step": 10779 + }, + { + "epoch": 0.9186195142735407, + "grad_norm": 11.9375, + "learning_rate": 1.4183624722775251e-06, + "loss": 0.2722, + "step": 10780 + }, + { + "epoch": 0.9187047294418407, + "grad_norm": 14.0, + "learning_rate": 1.4181959663816462e-06, + "loss": 0.4637, + "step": 10781 + }, + { + "epoch": 0.9187899446101406, + "grad_norm": 14.75, + "learning_rate": 1.4180294574462966e-06, + "loss": 0.9549, + "step": 10782 + }, + { + "epoch": 0.9188751597784406, + "grad_norm": 18.125, + "learning_rate": 1.4178629454744858e-06, + "loss": 1.1805, + "step": 10783 + }, + { + "epoch": 0.9189603749467405, + "grad_norm": 22.375, + "learning_rate": 1.4176964304692232e-06, + "loss": 0.6905, + "step": 10784 + }, + { + "epoch": 0.9190455901150405, + "grad_norm": 14.3125, + "learning_rate": 1.417529912433517e-06, + "loss": 0.6286, + "step": 10785 + }, + { + "epoch": 0.9191308052833405, + "grad_norm": 16.0, + "learning_rate": 1.417363391370377e-06, + "loss": 1.0318, + "step": 10786 + }, + { + "epoch": 0.9192160204516404, + "grad_norm": 12.6875, + "learning_rate": 1.4171968672828122e-06, + "loss": 0.7355, + "step": 10787 + }, + { + "epoch": 0.9193012356199404, + "grad_norm": 20.375, + "learning_rate": 1.4170303401738316e-06, + "loss": 0.7971, + "step": 10788 + }, + { + "epoch": 0.9193864507882403, + "grad_norm": 9.1875, + "learning_rate": 1.4168638100464455e-06, + "loss": 0.2897, + "step": 10789 + }, + { + "epoch": 0.9194716659565403, + "grad_norm": 25.0, + "learning_rate": 1.416697276903662e-06, + "loss": 1.0276, + "step": 10790 + }, + { + "epoch": 0.9195568811248402, + "grad_norm": 13.125, + "learning_rate": 1.416530740748491e-06, + "loss": 0.7483, + "step": 10791 + }, + { + "epoch": 0.9196420962931402, + "grad_norm": 13.3125, + "learning_rate": 1.4163642015839423e-06, + "loss": 0.4395, + "step": 10792 + }, + { + "epoch": 0.9197273114614402, + "grad_norm": 19.5, + "learning_rate": 1.4161976594130253e-06, + "loss": 0.7301, + "step": 10793 + }, + { + "epoch": 0.9198125266297401, + "grad_norm": 20.5, + "learning_rate": 1.416031114238749e-06, + "loss": 1.1139, + "step": 10794 + }, + { + "epoch": 0.9198977417980401, + "grad_norm": 12.0625, + "learning_rate": 1.415864566064124e-06, + "loss": 0.4253, + "step": 10795 + }, + { + "epoch": 0.91998295696634, + "grad_norm": 12.0, + "learning_rate": 1.4156980148921589e-06, + "loss": 0.511, + "step": 10796 + }, + { + "epoch": 0.92006817213464, + "grad_norm": 13.625, + "learning_rate": 1.415531460725865e-06, + "loss": 0.7488, + "step": 10797 + }, + { + "epoch": 0.9201533873029399, + "grad_norm": 17.0, + "learning_rate": 1.4153649035682504e-06, + "loss": 0.8851, + "step": 10798 + }, + { + "epoch": 0.9202386024712399, + "grad_norm": 18.125, + "learning_rate": 1.415198343422326e-06, + "loss": 0.9737, + "step": 10799 + }, + { + "epoch": 0.9203238176395399, + "grad_norm": 16.375, + "learning_rate": 1.415031780291101e-06, + "loss": 0.764, + "step": 10800 + }, + { + "epoch": 0.9204090328078398, + "grad_norm": 10.125, + "learning_rate": 1.4148652141775864e-06, + "loss": 0.4451, + "step": 10801 + }, + { + "epoch": 0.9204942479761398, + "grad_norm": 19.375, + "learning_rate": 1.4146986450847912e-06, + "loss": 0.9216, + "step": 10802 + }, + { + "epoch": 0.9205794631444397, + "grad_norm": 12.875, + "learning_rate": 1.4145320730157262e-06, + "loss": 0.6393, + "step": 10803 + }, + { + "epoch": 0.9206646783127397, + "grad_norm": 13.5, + "learning_rate": 1.4143654979734006e-06, + "loss": 0.8285, + "step": 10804 + }, + { + "epoch": 0.9207498934810396, + "grad_norm": 15.875, + "learning_rate": 1.4141989199608256e-06, + "loss": 0.7608, + "step": 10805 + }, + { + "epoch": 0.9208351086493396, + "grad_norm": 20.75, + "learning_rate": 1.414032338981011e-06, + "loss": 0.9688, + "step": 10806 + }, + { + "epoch": 0.9209203238176396, + "grad_norm": 13.75, + "learning_rate": 1.4138657550369672e-06, + "loss": 0.357, + "step": 10807 + }, + { + "epoch": 0.9210055389859395, + "grad_norm": 25.375, + "learning_rate": 1.413699168131704e-06, + "loss": 1.0686, + "step": 10808 + }, + { + "epoch": 0.9210907541542395, + "grad_norm": 20.75, + "learning_rate": 1.4135325782682324e-06, + "loss": 0.8688, + "step": 10809 + }, + { + "epoch": 0.9211759693225394, + "grad_norm": 14.125, + "learning_rate": 1.413365985449563e-06, + "loss": 0.6263, + "step": 10810 + }, + { + "epoch": 0.9212611844908394, + "grad_norm": 13.6875, + "learning_rate": 1.4131993896787056e-06, + "loss": 0.5667, + "step": 10811 + }, + { + "epoch": 0.9213463996591393, + "grad_norm": 11.375, + "learning_rate": 1.413032790958671e-06, + "loss": 0.8596, + "step": 10812 + }, + { + "epoch": 0.9214316148274393, + "grad_norm": 17.5, + "learning_rate": 1.4128661892924702e-06, + "loss": 1.1177, + "step": 10813 + }, + { + "epoch": 0.9215168299957393, + "grad_norm": 12.5, + "learning_rate": 1.4126995846831137e-06, + "loss": 0.4265, + "step": 10814 + }, + { + "epoch": 0.9216020451640392, + "grad_norm": 14.3125, + "learning_rate": 1.4125329771336115e-06, + "loss": 0.8378, + "step": 10815 + }, + { + "epoch": 0.9216872603323392, + "grad_norm": 20.375, + "learning_rate": 1.4123663666469756e-06, + "loss": 0.9733, + "step": 10816 + }, + { + "epoch": 0.9217724755006391, + "grad_norm": 12.0, + "learning_rate": 1.412199753226216e-06, + "loss": 0.6782, + "step": 10817 + }, + { + "epoch": 0.9218576906689391, + "grad_norm": 13.5, + "learning_rate": 1.412033136874344e-06, + "loss": 0.9881, + "step": 10818 + }, + { + "epoch": 0.921942905837239, + "grad_norm": 15.1875, + "learning_rate": 1.4118665175943696e-06, + "loss": 0.7595, + "step": 10819 + }, + { + "epoch": 0.922028121005539, + "grad_norm": 11.75, + "learning_rate": 1.4116998953893051e-06, + "loss": 0.705, + "step": 10820 + }, + { + "epoch": 0.922113336173839, + "grad_norm": 16.375, + "learning_rate": 1.4115332702621607e-06, + "loss": 0.9031, + "step": 10821 + }, + { + "epoch": 0.9221985513421389, + "grad_norm": 14.5, + "learning_rate": 1.4113666422159477e-06, + "loss": 0.5239, + "step": 10822 + }, + { + "epoch": 0.9222837665104389, + "grad_norm": 21.0, + "learning_rate": 1.411200011253677e-06, + "loss": 0.977, + "step": 10823 + }, + { + "epoch": 0.9223689816787388, + "grad_norm": 14.9375, + "learning_rate": 1.4110333773783603e-06, + "loss": 0.9753, + "step": 10824 + }, + { + "epoch": 0.9224541968470388, + "grad_norm": 10.3125, + "learning_rate": 1.4108667405930082e-06, + "loss": 0.3109, + "step": 10825 + }, + { + "epoch": 0.9225394120153387, + "grad_norm": 13.125, + "learning_rate": 1.4107001009006324e-06, + "loss": 0.544, + "step": 10826 + }, + { + "epoch": 0.9226246271836387, + "grad_norm": 16.5, + "learning_rate": 1.4105334583042444e-06, + "loss": 0.8261, + "step": 10827 + }, + { + "epoch": 0.9227098423519386, + "grad_norm": 11.1875, + "learning_rate": 1.4103668128068548e-06, + "loss": 0.352, + "step": 10828 + }, + { + "epoch": 0.9227950575202386, + "grad_norm": 13.0, + "learning_rate": 1.4102001644114766e-06, + "loss": 0.3697, + "step": 10829 + }, + { + "epoch": 0.9228802726885386, + "grad_norm": 18.375, + "learning_rate": 1.4100335131211195e-06, + "loss": 0.9283, + "step": 10830 + }, + { + "epoch": 0.9229654878568385, + "grad_norm": 17.5, + "learning_rate": 1.4098668589387966e-06, + "loss": 0.8075, + "step": 10831 + }, + { + "epoch": 0.9230507030251385, + "grad_norm": 14.4375, + "learning_rate": 1.409700201867518e-06, + "loss": 0.5521, + "step": 10832 + }, + { + "epoch": 0.9231359181934384, + "grad_norm": 13.625, + "learning_rate": 1.409533541910297e-06, + "loss": 0.2999, + "step": 10833 + }, + { + "epoch": 0.9232211333617384, + "grad_norm": 11.25, + "learning_rate": 1.4093668790701437e-06, + "loss": 0.5054, + "step": 10834 + }, + { + "epoch": 0.9233063485300383, + "grad_norm": 16.0, + "learning_rate": 1.4092002133500713e-06, + "loss": 0.7719, + "step": 10835 + }, + { + "epoch": 0.9233915636983383, + "grad_norm": 12.5, + "learning_rate": 1.4090335447530903e-06, + "loss": 0.4164, + "step": 10836 + }, + { + "epoch": 0.9234767788666383, + "grad_norm": 20.25, + "learning_rate": 1.408866873282214e-06, + "loss": 0.7232, + "step": 10837 + }, + { + "epoch": 0.9235619940349382, + "grad_norm": 14.0625, + "learning_rate": 1.4087001989404529e-06, + "loss": 0.6379, + "step": 10838 + }, + { + "epoch": 0.9236472092032382, + "grad_norm": 13.625, + "learning_rate": 1.40853352173082e-06, + "loss": 0.664, + "step": 10839 + }, + { + "epoch": 0.9237324243715381, + "grad_norm": 21.5, + "learning_rate": 1.4083668416563265e-06, + "loss": 1.0119, + "step": 10840 + }, + { + "epoch": 0.9238176395398381, + "grad_norm": 28.75, + "learning_rate": 1.4082001587199853e-06, + "loss": 0.9788, + "step": 10841 + }, + { + "epoch": 0.923902854708138, + "grad_norm": 16.125, + "learning_rate": 1.408033472924808e-06, + "loss": 0.7658, + "step": 10842 + }, + { + "epoch": 0.923988069876438, + "grad_norm": 14.6875, + "learning_rate": 1.407866784273807e-06, + "loss": 0.6037, + "step": 10843 + }, + { + "epoch": 0.924073285044738, + "grad_norm": 13.3125, + "learning_rate": 1.4077000927699945e-06, + "loss": 0.4422, + "step": 10844 + }, + { + "epoch": 0.9241585002130379, + "grad_norm": 13.625, + "learning_rate": 1.4075333984163824e-06, + "loss": 0.6748, + "step": 10845 + }, + { + "epoch": 0.9242437153813379, + "grad_norm": 15.5, + "learning_rate": 1.407366701215984e-06, + "loss": 0.3819, + "step": 10846 + }, + { + "epoch": 0.9243289305496378, + "grad_norm": 14.625, + "learning_rate": 1.4072000011718106e-06, + "loss": 0.6826, + "step": 10847 + }, + { + "epoch": 0.9244141457179378, + "grad_norm": 14.625, + "learning_rate": 1.4070332982868754e-06, + "loss": 0.7509, + "step": 10848 + }, + { + "epoch": 0.9244993608862377, + "grad_norm": 19.5, + "learning_rate": 1.40686659256419e-06, + "loss": 0.2675, + "step": 10849 + }, + { + "epoch": 0.9245845760545377, + "grad_norm": 10.25, + "learning_rate": 1.4066998840067682e-06, + "loss": 0.3213, + "step": 10850 + }, + { + "epoch": 0.9246697912228377, + "grad_norm": 15.8125, + "learning_rate": 1.4065331726176218e-06, + "loss": 0.6964, + "step": 10851 + }, + { + "epoch": 0.9247550063911376, + "grad_norm": 12.25, + "learning_rate": 1.4063664583997632e-06, + "loss": 0.4816, + "step": 10852 + }, + { + "epoch": 0.9248402215594376, + "grad_norm": 18.75, + "learning_rate": 1.4061997413562057e-06, + "loss": 0.7428, + "step": 10853 + }, + { + "epoch": 0.9249254367277375, + "grad_norm": 15.9375, + "learning_rate": 1.406033021489962e-06, + "loss": 0.9001, + "step": 10854 + }, + { + "epoch": 0.9250106518960375, + "grad_norm": 18.25, + "learning_rate": 1.4058662988040444e-06, + "loss": 0.7291, + "step": 10855 + }, + { + "epoch": 0.9250958670643374, + "grad_norm": 18.875, + "learning_rate": 1.4056995733014664e-06, + "loss": 0.8586, + "step": 10856 + }, + { + "epoch": 0.9251810822326374, + "grad_norm": 14.8125, + "learning_rate": 1.4055328449852398e-06, + "loss": 0.7642, + "step": 10857 + }, + { + "epoch": 0.9252662974009374, + "grad_norm": 13.0, + "learning_rate": 1.405366113858379e-06, + "loss": 0.592, + "step": 10858 + }, + { + "epoch": 0.9253515125692373, + "grad_norm": 15.4375, + "learning_rate": 1.4051993799238962e-06, + "loss": 0.6228, + "step": 10859 + }, + { + "epoch": 0.9254367277375373, + "grad_norm": 12.0625, + "learning_rate": 1.405032643184804e-06, + "loss": 0.4445, + "step": 10860 + }, + { + "epoch": 0.9255219429058372, + "grad_norm": 20.25, + "learning_rate": 1.4048659036441166e-06, + "loss": 0.9376, + "step": 10861 + }, + { + "epoch": 0.9256071580741372, + "grad_norm": 17.5, + "learning_rate": 1.4046991613048466e-06, + "loss": 0.8778, + "step": 10862 + }, + { + "epoch": 0.9256923732424371, + "grad_norm": 11.1875, + "learning_rate": 1.4045324161700073e-06, + "loss": 0.3592, + "step": 10863 + }, + { + "epoch": 0.9257775884107371, + "grad_norm": 13.875, + "learning_rate": 1.4043656682426112e-06, + "loss": 0.4612, + "step": 10864 + }, + { + "epoch": 0.925862803579037, + "grad_norm": 16.625, + "learning_rate": 1.4041989175256727e-06, + "loss": 0.5827, + "step": 10865 + }, + { + "epoch": 0.925948018747337, + "grad_norm": 10.8125, + "learning_rate": 1.4040321640222046e-06, + "loss": 0.2933, + "step": 10866 + }, + { + "epoch": 0.926033233915637, + "grad_norm": 16.875, + "learning_rate": 1.4038654077352206e-06, + "loss": 0.8013, + "step": 10867 + }, + { + "epoch": 0.9261184490839369, + "grad_norm": 17.625, + "learning_rate": 1.4036986486677336e-06, + "loss": 0.8841, + "step": 10868 + }, + { + "epoch": 0.9262036642522369, + "grad_norm": 16.25, + "learning_rate": 1.4035318868227577e-06, + "loss": 0.6938, + "step": 10869 + }, + { + "epoch": 0.9262888794205368, + "grad_norm": 14.4375, + "learning_rate": 1.4033651222033062e-06, + "loss": 1.1186, + "step": 10870 + }, + { + "epoch": 0.9263740945888368, + "grad_norm": 12.25, + "learning_rate": 1.4031983548123928e-06, + "loss": 0.717, + "step": 10871 + }, + { + "epoch": 0.9264593097571368, + "grad_norm": 26.5, + "learning_rate": 1.403031584653031e-06, + "loss": 0.4043, + "step": 10872 + }, + { + "epoch": 0.9265445249254367, + "grad_norm": 11.0625, + "learning_rate": 1.4028648117282345e-06, + "loss": 0.3551, + "step": 10873 + }, + { + "epoch": 0.9266297400937367, + "grad_norm": 21.0, + "learning_rate": 1.4026980360410172e-06, + "loss": 0.9203, + "step": 10874 + }, + { + "epoch": 0.9267149552620366, + "grad_norm": 14.0, + "learning_rate": 1.4025312575943928e-06, + "loss": 0.8876, + "step": 10875 + }, + { + "epoch": 0.9268001704303366, + "grad_norm": 19.0, + "learning_rate": 1.4023644763913752e-06, + "loss": 0.9873, + "step": 10876 + }, + { + "epoch": 0.9268853855986365, + "grad_norm": 11.4375, + "learning_rate": 1.4021976924349783e-06, + "loss": 0.553, + "step": 10877 + }, + { + "epoch": 0.9269706007669365, + "grad_norm": 15.1875, + "learning_rate": 1.402030905728216e-06, + "loss": 0.5433, + "step": 10878 + }, + { + "epoch": 0.9270558159352364, + "grad_norm": 14.375, + "learning_rate": 1.4018641162741023e-06, + "loss": 0.9289, + "step": 10879 + }, + { + "epoch": 0.9271410311035364, + "grad_norm": 13.3125, + "learning_rate": 1.4016973240756516e-06, + "loss": 0.6297, + "step": 10880 + }, + { + "epoch": 0.9272262462718364, + "grad_norm": 13.25, + "learning_rate": 1.4015305291358774e-06, + "loss": 0.6875, + "step": 10881 + }, + { + "epoch": 0.9273114614401363, + "grad_norm": 17.0, + "learning_rate": 1.4013637314577947e-06, + "loss": 0.9845, + "step": 10882 + }, + { + "epoch": 0.9273966766084363, + "grad_norm": 28.375, + "learning_rate": 1.4011969310444165e-06, + "loss": 1.8797, + "step": 10883 + }, + { + "epoch": 0.9274818917767362, + "grad_norm": 13.75, + "learning_rate": 1.4010301278987582e-06, + "loss": 0.6481, + "step": 10884 + }, + { + "epoch": 0.9275671069450362, + "grad_norm": 22.875, + "learning_rate": 1.400863322023833e-06, + "loss": 1.3978, + "step": 10885 + }, + { + "epoch": 0.9276523221133361, + "grad_norm": 17.125, + "learning_rate": 1.4006965134226567e-06, + "loss": 1.0267, + "step": 10886 + }, + { + "epoch": 0.9277375372816361, + "grad_norm": 11.125, + "learning_rate": 1.4005297020982422e-06, + "loss": 0.6056, + "step": 10887 + }, + { + "epoch": 0.9278227524499361, + "grad_norm": 13.5, + "learning_rate": 1.400362888053605e-06, + "loss": 0.8652, + "step": 10888 + }, + { + "epoch": 0.927907967618236, + "grad_norm": 19.625, + "learning_rate": 1.4001960712917586e-06, + "loss": 0.9383, + "step": 10889 + }, + { + "epoch": 0.927993182786536, + "grad_norm": 13.4375, + "learning_rate": 1.4000292518157188e-06, + "loss": 0.5327, + "step": 10890 + }, + { + "epoch": 0.9280783979548359, + "grad_norm": 17.25, + "learning_rate": 1.3998624296284996e-06, + "loss": 0.7557, + "step": 10891 + }, + { + "epoch": 0.9281636131231359, + "grad_norm": 15.25, + "learning_rate": 1.3996956047331153e-06, + "loss": 0.7795, + "step": 10892 + }, + { + "epoch": 0.9282488282914358, + "grad_norm": 12.875, + "learning_rate": 1.3995287771325808e-06, + "loss": 0.5039, + "step": 10893 + }, + { + "epoch": 0.9283340434597358, + "grad_norm": 16.5, + "learning_rate": 1.3993619468299112e-06, + "loss": 0.6502, + "step": 10894 + }, + { + "epoch": 0.9284192586280358, + "grad_norm": 14.4375, + "learning_rate": 1.3991951138281209e-06, + "loss": 0.8852, + "step": 10895 + }, + { + "epoch": 0.9285044737963357, + "grad_norm": 14.0625, + "learning_rate": 1.3990282781302247e-06, + "loss": 0.5781, + "step": 10896 + }, + { + "epoch": 0.9285896889646357, + "grad_norm": 14.625, + "learning_rate": 1.398861439739238e-06, + "loss": 0.4403, + "step": 10897 + }, + { + "epoch": 0.9286749041329356, + "grad_norm": 13.1875, + "learning_rate": 1.3986945986581752e-06, + "loss": 0.6279, + "step": 10898 + }, + { + "epoch": 0.9287601193012356, + "grad_norm": 12.4375, + "learning_rate": 1.3985277548900518e-06, + "loss": 0.5392, + "step": 10899 + }, + { + "epoch": 0.9288453344695355, + "grad_norm": 11.5, + "learning_rate": 1.3983609084378818e-06, + "loss": 0.4736, + "step": 10900 + }, + { + "epoch": 0.9289305496378355, + "grad_norm": 16.625, + "learning_rate": 1.3981940593046817e-06, + "loss": 0.7413, + "step": 10901 + }, + { + "epoch": 0.9290157648061355, + "grad_norm": 13.6875, + "learning_rate": 1.3980272074934656e-06, + "loss": 0.7464, + "step": 10902 + }, + { + "epoch": 0.9291009799744354, + "grad_norm": 13.625, + "learning_rate": 1.397860353007249e-06, + "loss": 0.7442, + "step": 10903 + }, + { + "epoch": 0.9291861951427354, + "grad_norm": 18.75, + "learning_rate": 1.397693495849047e-06, + "loss": 0.9192, + "step": 10904 + }, + { + "epoch": 0.9292714103110353, + "grad_norm": 15.1875, + "learning_rate": 1.3975266360218753e-06, + "loss": 0.8816, + "step": 10905 + }, + { + "epoch": 0.9293566254793353, + "grad_norm": 10.875, + "learning_rate": 1.397359773528749e-06, + "loss": 0.5126, + "step": 10906 + }, + { + "epoch": 0.9294418406476352, + "grad_norm": 14.5, + "learning_rate": 1.3971929083726831e-06, + "loss": 0.5135, + "step": 10907 + }, + { + "epoch": 0.9295270558159352, + "grad_norm": 12.0, + "learning_rate": 1.3970260405566935e-06, + "loss": 0.6412, + "step": 10908 + }, + { + "epoch": 0.9296122709842352, + "grad_norm": 13.5, + "learning_rate": 1.3968591700837955e-06, + "loss": 0.432, + "step": 10909 + }, + { + "epoch": 0.9296974861525351, + "grad_norm": 21.75, + "learning_rate": 1.3966922969570044e-06, + "loss": 0.8647, + "step": 10910 + }, + { + "epoch": 0.9297827013208351, + "grad_norm": 12.5, + "learning_rate": 1.3965254211793364e-06, + "loss": 0.5473, + "step": 10911 + }, + { + "epoch": 0.929867916489135, + "grad_norm": 13.875, + "learning_rate": 1.3963585427538064e-06, + "loss": 0.5646, + "step": 10912 + }, + { + "epoch": 0.929953131657435, + "grad_norm": 20.125, + "learning_rate": 1.3961916616834304e-06, + "loss": 1.0701, + "step": 10913 + }, + { + "epoch": 0.9300383468257349, + "grad_norm": 13.5625, + "learning_rate": 1.3960247779712241e-06, + "loss": 0.6506, + "step": 10914 + }, + { + "epoch": 0.9301235619940349, + "grad_norm": 19.875, + "learning_rate": 1.3958578916202031e-06, + "loss": 0.794, + "step": 10915 + }, + { + "epoch": 0.9302087771623349, + "grad_norm": 28.75, + "learning_rate": 1.3956910026333837e-06, + "loss": 1.3003, + "step": 10916 + }, + { + "epoch": 0.9302939923306348, + "grad_norm": 16.625, + "learning_rate": 1.395524111013781e-06, + "loss": 0.8453, + "step": 10917 + }, + { + "epoch": 0.9303792074989348, + "grad_norm": 15.875, + "learning_rate": 1.3953572167644115e-06, + "loss": 0.7645, + "step": 10918 + }, + { + "epoch": 0.9304644226672347, + "grad_norm": 17.75, + "learning_rate": 1.3951903198882908e-06, + "loss": 0.9429, + "step": 10919 + }, + { + "epoch": 0.9305496378355347, + "grad_norm": 11.9375, + "learning_rate": 1.3950234203884351e-06, + "loss": 0.7103, + "step": 10920 + }, + { + "epoch": 0.9306348530038346, + "grad_norm": 13.25, + "learning_rate": 1.3948565182678602e-06, + "loss": 0.4285, + "step": 10921 + }, + { + "epoch": 0.9307200681721346, + "grad_norm": 12.0, + "learning_rate": 1.3946896135295827e-06, + "loss": 0.4695, + "step": 10922 + }, + { + "epoch": 0.9308052833404346, + "grad_norm": 14.5625, + "learning_rate": 1.3945227061766182e-06, + "loss": 0.9451, + "step": 10923 + }, + { + "epoch": 0.9308904985087345, + "grad_norm": 13.625, + "learning_rate": 1.394355796211983e-06, + "loss": 0.6517, + "step": 10924 + }, + { + "epoch": 0.9309757136770345, + "grad_norm": 10.6875, + "learning_rate": 1.3941888836386933e-06, + "loss": 0.4722, + "step": 10925 + }, + { + "epoch": 0.9310609288453344, + "grad_norm": 14.8125, + "learning_rate": 1.3940219684597656e-06, + "loss": 0.632, + "step": 10926 + }, + { + "epoch": 0.9311461440136344, + "grad_norm": 14.75, + "learning_rate": 1.3938550506782162e-06, + "loss": 0.9764, + "step": 10927 + }, + { + "epoch": 0.9312313591819343, + "grad_norm": 18.375, + "learning_rate": 1.3936881302970608e-06, + "loss": 0.9851, + "step": 10928 + }, + { + "epoch": 0.9313165743502343, + "grad_norm": 17.0, + "learning_rate": 1.3935212073193173e-06, + "loss": 1.0548, + "step": 10929 + }, + { + "epoch": 0.9314017895185343, + "grad_norm": 19.0, + "learning_rate": 1.393354281748001e-06, + "loss": 0.9416, + "step": 10930 + }, + { + "epoch": 0.9314870046868342, + "grad_norm": 19.75, + "learning_rate": 1.3931873535861284e-06, + "loss": 1.1657, + "step": 10931 + }, + { + "epoch": 0.9315722198551342, + "grad_norm": 15.25, + "learning_rate": 1.393020422836716e-06, + "loss": 0.824, + "step": 10932 + }, + { + "epoch": 0.9316574350234341, + "grad_norm": 16.25, + "learning_rate": 1.3928534895027816e-06, + "loss": 0.95, + "step": 10933 + }, + { + "epoch": 0.9317426501917341, + "grad_norm": 13.5, + "learning_rate": 1.3926865535873401e-06, + "loss": 0.7342, + "step": 10934 + }, + { + "epoch": 0.931827865360034, + "grad_norm": 14.625, + "learning_rate": 1.39251961509341e-06, + "loss": 0.7263, + "step": 10935 + }, + { + "epoch": 0.931913080528334, + "grad_norm": 18.375, + "learning_rate": 1.3923526740240067e-06, + "loss": 0.6572, + "step": 10936 + }, + { + "epoch": 0.931998295696634, + "grad_norm": 14.9375, + "learning_rate": 1.3921857303821477e-06, + "loss": 0.4874, + "step": 10937 + }, + { + "epoch": 0.9320835108649339, + "grad_norm": 12.5, + "learning_rate": 1.3920187841708488e-06, + "loss": 0.5382, + "step": 10938 + }, + { + "epoch": 0.9321687260332339, + "grad_norm": 15.1875, + "learning_rate": 1.3918518353931288e-06, + "loss": 0.8582, + "step": 10939 + }, + { + "epoch": 0.9322539412015338, + "grad_norm": 11.5625, + "learning_rate": 1.3916848840520025e-06, + "loss": 0.4581, + "step": 10940 + }, + { + "epoch": 0.9323391563698338, + "grad_norm": 13.75, + "learning_rate": 1.3915179301504883e-06, + "loss": 0.4752, + "step": 10941 + }, + { + "epoch": 0.9324243715381337, + "grad_norm": 21.25, + "learning_rate": 1.3913509736916025e-06, + "loss": 1.1669, + "step": 10942 + }, + { + "epoch": 0.9325095867064337, + "grad_norm": 10.4375, + "learning_rate": 1.3911840146783628e-06, + "loss": 0.593, + "step": 10943 + }, + { + "epoch": 0.9325948018747336, + "grad_norm": 19.625, + "learning_rate": 1.3910170531137862e-06, + "loss": 1.05, + "step": 10944 + }, + { + "epoch": 0.9326800170430336, + "grad_norm": 16.875, + "learning_rate": 1.3908500890008894e-06, + "loss": 0.5251, + "step": 10945 + }, + { + "epoch": 0.9327652322113337, + "grad_norm": 16.875, + "learning_rate": 1.3906831223426895e-06, + "loss": 0.8289, + "step": 10946 + }, + { + "epoch": 0.9328504473796336, + "grad_norm": 14.875, + "learning_rate": 1.3905161531422047e-06, + "loss": 0.9262, + "step": 10947 + }, + { + "epoch": 0.9329356625479336, + "grad_norm": 11.1875, + "learning_rate": 1.3903491814024515e-06, + "loss": 0.4337, + "step": 10948 + }, + { + "epoch": 0.9330208777162335, + "grad_norm": 15.1875, + "learning_rate": 1.390182207126447e-06, + "loss": 0.5801, + "step": 10949 + }, + { + "epoch": 0.9331060928845335, + "grad_norm": 24.375, + "learning_rate": 1.3900152303172096e-06, + "loss": 1.1195, + "step": 10950 + }, + { + "epoch": 0.9331913080528335, + "grad_norm": 26.75, + "learning_rate": 1.389848250977756e-06, + "loss": 0.784, + "step": 10951 + }, + { + "epoch": 0.9332765232211334, + "grad_norm": 15.1875, + "learning_rate": 1.389681269111104e-06, + "loss": 0.9963, + "step": 10952 + }, + { + "epoch": 0.9333617383894334, + "grad_norm": 13.125, + "learning_rate": 1.3895142847202706e-06, + "loss": 0.5923, + "step": 10953 + }, + { + "epoch": 0.9334469535577333, + "grad_norm": 10.875, + "learning_rate": 1.3893472978082741e-06, + "loss": 0.3826, + "step": 10954 + }, + { + "epoch": 0.9335321687260333, + "grad_norm": 12.125, + "learning_rate": 1.389180308378132e-06, + "loss": 0.6673, + "step": 10955 + }, + { + "epoch": 0.9336173838943332, + "grad_norm": 19.5, + "learning_rate": 1.3890133164328618e-06, + "loss": 1.3411, + "step": 10956 + }, + { + "epoch": 0.9337025990626332, + "grad_norm": 13.125, + "learning_rate": 1.3888463219754806e-06, + "loss": 0.6956, + "step": 10957 + }, + { + "epoch": 0.9337878142309332, + "grad_norm": 16.125, + "learning_rate": 1.3886793250090071e-06, + "loss": 0.6626, + "step": 10958 + }, + { + "epoch": 0.9338730293992331, + "grad_norm": 15.875, + "learning_rate": 1.3885123255364588e-06, + "loss": 0.9476, + "step": 10959 + }, + { + "epoch": 0.9339582445675331, + "grad_norm": 14.0625, + "learning_rate": 1.3883453235608532e-06, + "loss": 0.5006, + "step": 10960 + }, + { + "epoch": 0.934043459735833, + "grad_norm": 11.6875, + "learning_rate": 1.3881783190852089e-06, + "loss": 0.452, + "step": 10961 + }, + { + "epoch": 0.934128674904133, + "grad_norm": 12.875, + "learning_rate": 1.3880113121125432e-06, + "loss": 0.5628, + "step": 10962 + }, + { + "epoch": 0.9342138900724329, + "grad_norm": 14.1875, + "learning_rate": 1.3878443026458745e-06, + "loss": 0.699, + "step": 10963 + }, + { + "epoch": 0.9342991052407329, + "grad_norm": 13.6875, + "learning_rate": 1.3876772906882203e-06, + "loss": 0.5204, + "step": 10964 + }, + { + "epoch": 0.9343843204090329, + "grad_norm": 20.125, + "learning_rate": 1.3875102762425994e-06, + "loss": 1.1444, + "step": 10965 + }, + { + "epoch": 0.9344695355773328, + "grad_norm": 16.75, + "learning_rate": 1.3873432593120289e-06, + "loss": 0.7788, + "step": 10966 + }, + { + "epoch": 0.9345547507456328, + "grad_norm": 12.0, + "learning_rate": 1.3871762398995286e-06, + "loss": 0.8329, + "step": 10967 + }, + { + "epoch": 0.9346399659139327, + "grad_norm": 14.125, + "learning_rate": 1.3870092180081148e-06, + "loss": 0.5616, + "step": 10968 + }, + { + "epoch": 0.9347251810822327, + "grad_norm": 14.4375, + "learning_rate": 1.3868421936408072e-06, + "loss": 0.8149, + "step": 10969 + }, + { + "epoch": 0.9348103962505326, + "grad_norm": 14.5, + "learning_rate": 1.386675166800623e-06, + "loss": 0.8236, + "step": 10970 + }, + { + "epoch": 0.9348956114188326, + "grad_norm": 15.875, + "learning_rate": 1.386508137490582e-06, + "loss": 0.7371, + "step": 10971 + }, + { + "epoch": 0.9349808265871326, + "grad_norm": 18.25, + "learning_rate": 1.3863411057137007e-06, + "loss": 1.2087, + "step": 10972 + }, + { + "epoch": 0.9350660417554325, + "grad_norm": 18.5, + "learning_rate": 1.3861740714729991e-06, + "loss": 0.7364, + "step": 10973 + }, + { + "epoch": 0.9351512569237325, + "grad_norm": 28.625, + "learning_rate": 1.3860070347714948e-06, + "loss": 0.7654, + "step": 10974 + }, + { + "epoch": 0.9352364720920324, + "grad_norm": 23.375, + "learning_rate": 1.3858399956122067e-06, + "loss": 0.9621, + "step": 10975 + }, + { + "epoch": 0.9353216872603324, + "grad_norm": 14.0, + "learning_rate": 1.3856729539981533e-06, + "loss": 0.8979, + "step": 10976 + }, + { + "epoch": 0.9354069024286323, + "grad_norm": 17.625, + "learning_rate": 1.3855059099323534e-06, + "loss": 0.9275, + "step": 10977 + }, + { + "epoch": 0.9354921175969323, + "grad_norm": 21.125, + "learning_rate": 1.3853388634178252e-06, + "loss": 0.8401, + "step": 10978 + }, + { + "epoch": 0.9355773327652323, + "grad_norm": 17.125, + "learning_rate": 1.3851718144575877e-06, + "loss": 0.5417, + "step": 10979 + }, + { + "epoch": 0.9356625479335322, + "grad_norm": 19.5, + "learning_rate": 1.3850047630546598e-06, + "loss": 0.9337, + "step": 10980 + }, + { + "epoch": 0.9357477631018322, + "grad_norm": 17.75, + "learning_rate": 1.3848377092120595e-06, + "loss": 0.7117, + "step": 10981 + }, + { + "epoch": 0.9358329782701321, + "grad_norm": 12.6875, + "learning_rate": 1.3846706529328068e-06, + "loss": 0.4003, + "step": 10982 + }, + { + "epoch": 0.9359181934384321, + "grad_norm": 17.625, + "learning_rate": 1.3845035942199196e-06, + "loss": 0.9912, + "step": 10983 + }, + { + "epoch": 0.936003408606732, + "grad_norm": 69.0, + "learning_rate": 1.3843365330764177e-06, + "loss": 1.1518, + "step": 10984 + }, + { + "epoch": 0.936088623775032, + "grad_norm": 17.25, + "learning_rate": 1.384169469505319e-06, + "loss": 0.9458, + "step": 10985 + }, + { + "epoch": 0.936173838943332, + "grad_norm": 16.75, + "learning_rate": 1.3840024035096434e-06, + "loss": 0.8911, + "step": 10986 + }, + { + "epoch": 0.9362590541116319, + "grad_norm": 17.375, + "learning_rate": 1.3838353350924093e-06, + "loss": 0.9467, + "step": 10987 + }, + { + "epoch": 0.9363442692799319, + "grad_norm": 13.75, + "learning_rate": 1.3836682642566369e-06, + "loss": 0.6473, + "step": 10988 + }, + { + "epoch": 0.9364294844482318, + "grad_norm": 15.0, + "learning_rate": 1.3835011910053437e-06, + "loss": 1.0352, + "step": 10989 + }, + { + "epoch": 0.9365146996165318, + "grad_norm": 13.9375, + "learning_rate": 1.38333411534155e-06, + "loss": 0.5928, + "step": 10990 + }, + { + "epoch": 0.9365999147848317, + "grad_norm": 12.25, + "learning_rate": 1.3831670372682745e-06, + "loss": 0.6214, + "step": 10991 + }, + { + "epoch": 0.9366851299531317, + "grad_norm": 12.3125, + "learning_rate": 1.3829999567885375e-06, + "loss": 0.4847, + "step": 10992 + }, + { + "epoch": 0.9367703451214316, + "grad_norm": 20.875, + "learning_rate": 1.3828328739053567e-06, + "loss": 1.021, + "step": 10993 + }, + { + "epoch": 0.9368555602897316, + "grad_norm": 22.125, + "learning_rate": 1.3826657886217526e-06, + "loss": 1.0854, + "step": 10994 + }, + { + "epoch": 0.9369407754580316, + "grad_norm": 12.0, + "learning_rate": 1.3824987009407443e-06, + "loss": 0.4298, + "step": 10995 + }, + { + "epoch": 0.9370259906263315, + "grad_norm": 11.5625, + "learning_rate": 1.382331610865351e-06, + "loss": 0.519, + "step": 10996 + }, + { + "epoch": 0.9371112057946315, + "grad_norm": 14.625, + "learning_rate": 1.3821645183985929e-06, + "loss": 0.493, + "step": 10997 + }, + { + "epoch": 0.9371964209629314, + "grad_norm": 12.625, + "learning_rate": 1.3819974235434889e-06, + "loss": 0.454, + "step": 10998 + }, + { + "epoch": 0.9372816361312314, + "grad_norm": 14.125, + "learning_rate": 1.3818303263030586e-06, + "loss": 0.9092, + "step": 10999 + }, + { + "epoch": 0.9373668512995313, + "grad_norm": 17.375, + "learning_rate": 1.3816632266803217e-06, + "loss": 1.0118, + "step": 11000 + }, + { + "epoch": 0.9374520664678313, + "grad_norm": 12.75, + "learning_rate": 1.3814961246782982e-06, + "loss": 0.5042, + "step": 11001 + }, + { + "epoch": 0.9375372816361313, + "grad_norm": 14.8125, + "learning_rate": 1.381329020300007e-06, + "loss": 0.6875, + "step": 11002 + }, + { + "epoch": 0.9376224968044312, + "grad_norm": 17.0, + "learning_rate": 1.3811619135484689e-06, + "loss": 0.7561, + "step": 11003 + }, + { + "epoch": 0.9377077119727312, + "grad_norm": 18.0, + "learning_rate": 1.3809948044267028e-06, + "loss": 0.8927, + "step": 11004 + }, + { + "epoch": 0.9377929271410311, + "grad_norm": 32.75, + "learning_rate": 1.380827692937729e-06, + "loss": 0.9235, + "step": 11005 + }, + { + "epoch": 0.9378781423093311, + "grad_norm": 14.75, + "learning_rate": 1.380660579084567e-06, + "loss": 0.7896, + "step": 11006 + }, + { + "epoch": 0.937963357477631, + "grad_norm": 25.0, + "learning_rate": 1.3804934628702372e-06, + "loss": 1.4595, + "step": 11007 + }, + { + "epoch": 0.938048572645931, + "grad_norm": 12.875, + "learning_rate": 1.3803263442977592e-06, + "loss": 0.5708, + "step": 11008 + }, + { + "epoch": 0.938133787814231, + "grad_norm": 14.0, + "learning_rate": 1.3801592233701533e-06, + "loss": 0.9006, + "step": 11009 + }, + { + "epoch": 0.9382190029825309, + "grad_norm": 12.4375, + "learning_rate": 1.3799921000904386e-06, + "loss": 0.4296, + "step": 11010 + }, + { + "epoch": 0.9383042181508309, + "grad_norm": 19.25, + "learning_rate": 1.3798249744616368e-06, + "loss": 1.0577, + "step": 11011 + }, + { + "epoch": 0.9383894333191308, + "grad_norm": 16.625, + "learning_rate": 1.3796578464867668e-06, + "loss": 0.8821, + "step": 11012 + }, + { + "epoch": 0.9384746484874308, + "grad_norm": 15.75, + "learning_rate": 1.379490716168849e-06, + "loss": 0.6517, + "step": 11013 + }, + { + "epoch": 0.9385598636557307, + "grad_norm": 15.75, + "learning_rate": 1.379323583510904e-06, + "loss": 0.9636, + "step": 11014 + }, + { + "epoch": 0.9386450788240307, + "grad_norm": 17.125, + "learning_rate": 1.379156448515952e-06, + "loss": 0.7771, + "step": 11015 + }, + { + "epoch": 0.9387302939923307, + "grad_norm": 19.0, + "learning_rate": 1.378989311187013e-06, + "loss": 1.2333, + "step": 11016 + }, + { + "epoch": 0.9388155091606306, + "grad_norm": 14.75, + "learning_rate": 1.3788221715271068e-06, + "loss": 0.6809, + "step": 11017 + }, + { + "epoch": 0.9389007243289306, + "grad_norm": 18.625, + "learning_rate": 1.3786550295392551e-06, + "loss": 0.6074, + "step": 11018 + }, + { + "epoch": 0.9389859394972305, + "grad_norm": 11.6875, + "learning_rate": 1.3784878852264771e-06, + "loss": 0.53, + "step": 11019 + }, + { + "epoch": 0.9390711546655305, + "grad_norm": 16.25, + "learning_rate": 1.3783207385917947e-06, + "loss": 0.7652, + "step": 11020 + }, + { + "epoch": 0.9391563698338304, + "grad_norm": 21.875, + "learning_rate": 1.3781535896382265e-06, + "loss": 0.9202, + "step": 11021 + }, + { + "epoch": 0.9392415850021304, + "grad_norm": 11.9375, + "learning_rate": 1.3779864383687947e-06, + "loss": 0.4956, + "step": 11022 + }, + { + "epoch": 0.9393268001704304, + "grad_norm": 11.5, + "learning_rate": 1.3778192847865188e-06, + "loss": 0.4734, + "step": 11023 + }, + { + "epoch": 0.9394120153387303, + "grad_norm": 13.0625, + "learning_rate": 1.3776521288944206e-06, + "loss": 0.7378, + "step": 11024 + }, + { + "epoch": 0.9394972305070303, + "grad_norm": 11.1875, + "learning_rate": 1.3774849706955193e-06, + "loss": 0.3357, + "step": 11025 + }, + { + "epoch": 0.9395824456753302, + "grad_norm": 15.0625, + "learning_rate": 1.377317810192837e-06, + "loss": 0.6553, + "step": 11026 + }, + { + "epoch": 0.9396676608436302, + "grad_norm": 15.8125, + "learning_rate": 1.3771506473893933e-06, + "loss": 0.8324, + "step": 11027 + }, + { + "epoch": 0.9397528760119301, + "grad_norm": 18.75, + "learning_rate": 1.3769834822882096e-06, + "loss": 1.2539, + "step": 11028 + }, + { + "epoch": 0.9398380911802301, + "grad_norm": 17.125, + "learning_rate": 1.3768163148923073e-06, + "loss": 0.9988, + "step": 11029 + }, + { + "epoch": 0.93992330634853, + "grad_norm": 20.0, + "learning_rate": 1.3766491452047062e-06, + "loss": 1.0352, + "step": 11030 + }, + { + "epoch": 0.94000852151683, + "grad_norm": 15.1875, + "learning_rate": 1.3764819732284276e-06, + "loss": 0.6993, + "step": 11031 + }, + { + "epoch": 0.94009373668513, + "grad_norm": 13.25, + "learning_rate": 1.3763147989664927e-06, + "loss": 0.5443, + "step": 11032 + }, + { + "epoch": 0.9401789518534299, + "grad_norm": 13.375, + "learning_rate": 1.3761476224219226e-06, + "loss": 0.6295, + "step": 11033 + }, + { + "epoch": 0.9402641670217299, + "grad_norm": 19.375, + "learning_rate": 1.3759804435977375e-06, + "loss": 1.2088, + "step": 11034 + }, + { + "epoch": 0.9403493821900298, + "grad_norm": 17.375, + "learning_rate": 1.3758132624969594e-06, + "loss": 1.0608, + "step": 11035 + }, + { + "epoch": 0.9404345973583298, + "grad_norm": 11.5625, + "learning_rate": 1.3756460791226095e-06, + "loss": 0.444, + "step": 11036 + }, + { + "epoch": 0.9405198125266298, + "grad_norm": 13.4375, + "learning_rate": 1.3754788934777084e-06, + "loss": 0.8103, + "step": 11037 + }, + { + "epoch": 0.9406050276949297, + "grad_norm": 15.4375, + "learning_rate": 1.3753117055652774e-06, + "loss": 0.4821, + "step": 11038 + }, + { + "epoch": 0.9406902428632297, + "grad_norm": 17.375, + "learning_rate": 1.3751445153883377e-06, + "loss": 0.9522, + "step": 11039 + }, + { + "epoch": 0.9407754580315296, + "grad_norm": 15.4375, + "learning_rate": 1.374977322949911e-06, + "loss": 0.8709, + "step": 11040 + }, + { + "epoch": 0.9408606731998296, + "grad_norm": 14.75, + "learning_rate": 1.3748101282530185e-06, + "loss": 0.6893, + "step": 11041 + }, + { + "epoch": 0.9409458883681295, + "grad_norm": 15.125, + "learning_rate": 1.3746429313006812e-06, + "loss": 0.8334, + "step": 11042 + }, + { + "epoch": 0.9410311035364295, + "grad_norm": 14.75, + "learning_rate": 1.374475732095921e-06, + "loss": 0.5598, + "step": 11043 + }, + { + "epoch": 0.9411163187047294, + "grad_norm": 14.25, + "learning_rate": 1.374308530641759e-06, + "loss": 0.4919, + "step": 11044 + }, + { + "epoch": 0.9412015338730294, + "grad_norm": 24.75, + "learning_rate": 1.3741413269412172e-06, + "loss": 1.2341, + "step": 11045 + }, + { + "epoch": 0.9412867490413294, + "grad_norm": 16.375, + "learning_rate": 1.373974120997316e-06, + "loss": 0.765, + "step": 11046 + }, + { + "epoch": 0.9413719642096293, + "grad_norm": 18.0, + "learning_rate": 1.3738069128130787e-06, + "loss": 0.7991, + "step": 11047 + }, + { + "epoch": 0.9414571793779293, + "grad_norm": 11.4375, + "learning_rate": 1.3736397023915254e-06, + "loss": 0.468, + "step": 11048 + }, + { + "epoch": 0.9415423945462292, + "grad_norm": 16.875, + "learning_rate": 1.3734724897356785e-06, + "loss": 0.8174, + "step": 11049 + }, + { + "epoch": 0.9416276097145292, + "grad_norm": 13.625, + "learning_rate": 1.3733052748485598e-06, + "loss": 0.5951, + "step": 11050 + }, + { + "epoch": 0.9417128248828291, + "grad_norm": 17.625, + "learning_rate": 1.3731380577331901e-06, + "loss": 0.9719, + "step": 11051 + }, + { + "epoch": 0.9417980400511291, + "grad_norm": 12.0, + "learning_rate": 1.3729708383925925e-06, + "loss": 0.5042, + "step": 11052 + }, + { + "epoch": 0.9418832552194291, + "grad_norm": 14.5, + "learning_rate": 1.3728036168297882e-06, + "loss": 0.8386, + "step": 11053 + }, + { + "epoch": 0.941968470387729, + "grad_norm": 14.0625, + "learning_rate": 1.372636393047799e-06, + "loss": 0.6466, + "step": 11054 + }, + { + "epoch": 0.942053685556029, + "grad_norm": 15.5625, + "learning_rate": 1.3724691670496466e-06, + "loss": 0.757, + "step": 11055 + }, + { + "epoch": 0.9421389007243289, + "grad_norm": 16.25, + "learning_rate": 1.3723019388383532e-06, + "loss": 0.8277, + "step": 11056 + }, + { + "epoch": 0.9422241158926289, + "grad_norm": 17.875, + "learning_rate": 1.3721347084169412e-06, + "loss": 0.9954, + "step": 11057 + }, + { + "epoch": 0.9423093310609288, + "grad_norm": 13.375, + "learning_rate": 1.3719674757884318e-06, + "loss": 0.7858, + "step": 11058 + }, + { + "epoch": 0.9423945462292288, + "grad_norm": 19.5, + "learning_rate": 1.3718002409558476e-06, + "loss": 0.9745, + "step": 11059 + }, + { + "epoch": 0.9424797613975288, + "grad_norm": 19.125, + "learning_rate": 1.3716330039222106e-06, + "loss": 1.0373, + "step": 11060 + }, + { + "epoch": 0.9425649765658287, + "grad_norm": 13.9375, + "learning_rate": 1.3714657646905428e-06, + "loss": 0.7585, + "step": 11061 + }, + { + "epoch": 0.9426501917341287, + "grad_norm": 15.375, + "learning_rate": 1.3712985232638668e-06, + "loss": 0.8033, + "step": 11062 + }, + { + "epoch": 0.9427354069024286, + "grad_norm": 23.125, + "learning_rate": 1.371131279645204e-06, + "loss": 0.8723, + "step": 11063 + }, + { + "epoch": 0.9428206220707286, + "grad_norm": 14.125, + "learning_rate": 1.3709640338375775e-06, + "loss": 0.6407, + "step": 11064 + }, + { + "epoch": 0.9429058372390285, + "grad_norm": 19.25, + "learning_rate": 1.3707967858440092e-06, + "loss": 0.8283, + "step": 11065 + }, + { + "epoch": 0.9429910524073285, + "grad_norm": 14.4375, + "learning_rate": 1.3706295356675214e-06, + "loss": 0.8207, + "step": 11066 + }, + { + "epoch": 0.9430762675756285, + "grad_norm": 13.6875, + "learning_rate": 1.3704622833111367e-06, + "loss": 0.8268, + "step": 11067 + }, + { + "epoch": 0.9431614827439284, + "grad_norm": 16.0, + "learning_rate": 1.3702950287778772e-06, + "loss": 0.8447, + "step": 11068 + }, + { + "epoch": 0.9432466979122284, + "grad_norm": 12.875, + "learning_rate": 1.3701277720707662e-06, + "loss": 0.6439, + "step": 11069 + }, + { + "epoch": 0.9433319130805283, + "grad_norm": 12.0, + "learning_rate": 1.3699605131928247e-06, + "loss": 0.5469, + "step": 11070 + }, + { + "epoch": 0.9434171282488283, + "grad_norm": 17.125, + "learning_rate": 1.3697932521470766e-06, + "loss": 0.7244, + "step": 11071 + }, + { + "epoch": 0.9435023434171282, + "grad_norm": 22.0, + "learning_rate": 1.3696259889365438e-06, + "loss": 0.9343, + "step": 11072 + }, + { + "epoch": 0.9435875585854282, + "grad_norm": 14.3125, + "learning_rate": 1.3694587235642496e-06, + "loss": 0.6349, + "step": 11073 + }, + { + "epoch": 0.9436727737537282, + "grad_norm": 12.9375, + "learning_rate": 1.3692914560332154e-06, + "loss": 0.5881, + "step": 11074 + }, + { + "epoch": 0.9437579889220281, + "grad_norm": 15.6875, + "learning_rate": 1.3691241863464653e-06, + "loss": 0.8899, + "step": 11075 + }, + { + "epoch": 0.9438432040903281, + "grad_norm": 16.625, + "learning_rate": 1.3689569145070205e-06, + "loss": 0.7084, + "step": 11076 + }, + { + "epoch": 0.943928419258628, + "grad_norm": 12.5, + "learning_rate": 1.3687896405179058e-06, + "loss": 0.6732, + "step": 11077 + }, + { + "epoch": 0.944013634426928, + "grad_norm": 13.3125, + "learning_rate": 1.368622364382142e-06, + "loss": 0.5679, + "step": 11078 + }, + { + "epoch": 0.9440988495952279, + "grad_norm": 13.8125, + "learning_rate": 1.368455086102753e-06, + "loss": 0.6421, + "step": 11079 + }, + { + "epoch": 0.9441840647635279, + "grad_norm": 17.25, + "learning_rate": 1.368287805682761e-06, + "loss": 0.83, + "step": 11080 + }, + { + "epoch": 0.9442692799318279, + "grad_norm": 13.5, + "learning_rate": 1.36812052312519e-06, + "loss": 0.4536, + "step": 11081 + }, + { + "epoch": 0.9443544951001278, + "grad_norm": 14.25, + "learning_rate": 1.3679532384330624e-06, + "loss": 0.7111, + "step": 11082 + }, + { + "epoch": 0.9444397102684278, + "grad_norm": 14.375, + "learning_rate": 1.3677859516094008e-06, + "loss": 0.5433, + "step": 11083 + }, + { + "epoch": 0.9445249254367277, + "grad_norm": 13.25, + "learning_rate": 1.3676186626572291e-06, + "loss": 0.7989, + "step": 11084 + }, + { + "epoch": 0.9446101406050277, + "grad_norm": 12.5, + "learning_rate": 1.3674513715795695e-06, + "loss": 0.5892, + "step": 11085 + }, + { + "epoch": 0.9446953557733276, + "grad_norm": 21.0, + "learning_rate": 1.3672840783794458e-06, + "loss": 0.8751, + "step": 11086 + }, + { + "epoch": 0.9447805709416276, + "grad_norm": 17.125, + "learning_rate": 1.3671167830598806e-06, + "loss": 0.9412, + "step": 11087 + }, + { + "epoch": 0.9448657861099276, + "grad_norm": 12.1875, + "learning_rate": 1.3669494856238974e-06, + "loss": 0.5189, + "step": 11088 + }, + { + "epoch": 0.9449510012782275, + "grad_norm": 12.625, + "learning_rate": 1.3667821860745196e-06, + "loss": 0.4717, + "step": 11089 + }, + { + "epoch": 0.9450362164465275, + "grad_norm": 26.125, + "learning_rate": 1.36661488441477e-06, + "loss": 0.9389, + "step": 11090 + }, + { + "epoch": 0.9451214316148274, + "grad_norm": 12.6875, + "learning_rate": 1.366447580647672e-06, + "loss": 0.5008, + "step": 11091 + }, + { + "epoch": 0.9452066467831274, + "grad_norm": 19.875, + "learning_rate": 1.3662802747762495e-06, + "loss": 0.9154, + "step": 11092 + }, + { + "epoch": 0.9452918619514273, + "grad_norm": 21.625, + "learning_rate": 1.3661129668035255e-06, + "loss": 1.2268, + "step": 11093 + }, + { + "epoch": 0.9453770771197273, + "grad_norm": 28.375, + "learning_rate": 1.3659456567325236e-06, + "loss": 1.0453, + "step": 11094 + }, + { + "epoch": 0.9454622922880273, + "grad_norm": 9.75, + "learning_rate": 1.3657783445662664e-06, + "loss": 1.2264, + "step": 11095 + }, + { + "epoch": 0.9455475074563272, + "grad_norm": 13.6875, + "learning_rate": 1.3656110303077788e-06, + "loss": 0.6336, + "step": 11096 + }, + { + "epoch": 0.9456327226246272, + "grad_norm": 14.0, + "learning_rate": 1.3654437139600834e-06, + "loss": 0.6112, + "step": 11097 + }, + { + "epoch": 0.9457179377929271, + "grad_norm": 13.875, + "learning_rate": 1.3652763955262039e-06, + "loss": 0.5786, + "step": 11098 + }, + { + "epoch": 0.9458031529612271, + "grad_norm": 14.3125, + "learning_rate": 1.365109075009164e-06, + "loss": 0.6727, + "step": 11099 + }, + { + "epoch": 0.945888368129527, + "grad_norm": 17.125, + "learning_rate": 1.3649417524119874e-06, + "loss": 0.8454, + "step": 11100 + }, + { + "epoch": 0.945973583297827, + "grad_norm": 18.5, + "learning_rate": 1.3647744277376979e-06, + "loss": 0.7791, + "step": 11101 + }, + { + "epoch": 0.946058798466127, + "grad_norm": 15.125, + "learning_rate": 1.3646071009893188e-06, + "loss": 0.6159, + "step": 11102 + }, + { + "epoch": 0.9461440136344269, + "grad_norm": 16.125, + "learning_rate": 1.3644397721698742e-06, + "loss": 0.9871, + "step": 11103 + }, + { + "epoch": 0.9462292288027269, + "grad_norm": 10.625, + "learning_rate": 1.364272441282388e-06, + "loss": 0.4288, + "step": 11104 + }, + { + "epoch": 0.9463144439710268, + "grad_norm": 13.5, + "learning_rate": 1.364105108329884e-06, + "loss": 0.5338, + "step": 11105 + }, + { + "epoch": 0.9463996591393268, + "grad_norm": 14.4375, + "learning_rate": 1.3639377733153856e-06, + "loss": 0.6392, + "step": 11106 + }, + { + "epoch": 0.9464848743076267, + "grad_norm": 14.375, + "learning_rate": 1.3637704362419171e-06, + "loss": 0.6271, + "step": 11107 + }, + { + "epoch": 0.9465700894759267, + "grad_norm": 16.25, + "learning_rate": 1.3636030971125026e-06, + "loss": 0.896, + "step": 11108 + }, + { + "epoch": 0.9466553046442266, + "grad_norm": 25.625, + "learning_rate": 1.3634357559301662e-06, + "loss": 1.2843, + "step": 11109 + }, + { + "epoch": 0.9467405198125266, + "grad_norm": 14.25, + "learning_rate": 1.363268412697931e-06, + "loss": 0.7364, + "step": 11110 + }, + { + "epoch": 0.9468257349808266, + "grad_norm": 12.125, + "learning_rate": 1.3631010674188219e-06, + "loss": 0.4524, + "step": 11111 + }, + { + "epoch": 0.9469109501491265, + "grad_norm": 13.0625, + "learning_rate": 1.3629337200958626e-06, + "loss": 0.505, + "step": 11112 + }, + { + "epoch": 0.9469961653174265, + "grad_norm": 13.0, + "learning_rate": 1.3627663707320775e-06, + "loss": 0.6783, + "step": 11113 + }, + { + "epoch": 0.9470813804857264, + "grad_norm": 17.5, + "learning_rate": 1.3625990193304911e-06, + "loss": 0.5561, + "step": 11114 + }, + { + "epoch": 0.9471665956540264, + "grad_norm": 21.75, + "learning_rate": 1.362431665894127e-06, + "loss": 0.8594, + "step": 11115 + }, + { + "epoch": 0.9472518108223263, + "grad_norm": 19.25, + "learning_rate": 1.3622643104260095e-06, + "loss": 1.1498, + "step": 11116 + }, + { + "epoch": 0.9473370259906263, + "grad_norm": 9.6875, + "learning_rate": 1.362096952929163e-06, + "loss": 0.2352, + "step": 11117 + }, + { + "epoch": 0.9474222411589263, + "grad_norm": 15.3125, + "learning_rate": 1.3619295934066117e-06, + "loss": 0.8051, + "step": 11118 + }, + { + "epoch": 0.9475074563272262, + "grad_norm": 17.375, + "learning_rate": 1.3617622318613806e-06, + "loss": 0.8366, + "step": 11119 + }, + { + "epoch": 0.9475926714955262, + "grad_norm": 14.1875, + "learning_rate": 1.361594868296493e-06, + "loss": 0.6228, + "step": 11120 + }, + { + "epoch": 0.9476778866638261, + "grad_norm": 11.125, + "learning_rate": 1.361427502714974e-06, + "loss": 0.423, + "step": 11121 + }, + { + "epoch": 0.9477631018321261, + "grad_norm": 23.875, + "learning_rate": 1.3612601351198485e-06, + "loss": 1.1474, + "step": 11122 + }, + { + "epoch": 0.947848317000426, + "grad_norm": 14.625, + "learning_rate": 1.3610927655141402e-06, + "loss": 0.6743, + "step": 11123 + }, + { + "epoch": 0.947933532168726, + "grad_norm": 10.4375, + "learning_rate": 1.3609253939008738e-06, + "loss": 0.4045, + "step": 11124 + }, + { + "epoch": 0.948018747337026, + "grad_norm": 15.0625, + "learning_rate": 1.3607580202830739e-06, + "loss": 0.4785, + "step": 11125 + }, + { + "epoch": 0.9481039625053259, + "grad_norm": 18.5, + "learning_rate": 1.3605906446637656e-06, + "loss": 0.9845, + "step": 11126 + }, + { + "epoch": 0.9481891776736259, + "grad_norm": 12.5, + "learning_rate": 1.3604232670459727e-06, + "loss": 0.6442, + "step": 11127 + }, + { + "epoch": 0.9482743928419258, + "grad_norm": 13.9375, + "learning_rate": 1.3602558874327206e-06, + "loss": 0.8672, + "step": 11128 + }, + { + "epoch": 0.9483596080102258, + "grad_norm": 19.125, + "learning_rate": 1.3600885058270335e-06, + "loss": 0.7797, + "step": 11129 + }, + { + "epoch": 0.9484448231785257, + "grad_norm": 11.25, + "learning_rate": 1.359921122231937e-06, + "loss": 0.543, + "step": 11130 + }, + { + "epoch": 0.9485300383468257, + "grad_norm": 13.4375, + "learning_rate": 1.3597537366504546e-06, + "loss": 0.5551, + "step": 11131 + }, + { + "epoch": 0.9486152535151257, + "grad_norm": 11.5, + "learning_rate": 1.359586349085612e-06, + "loss": 0.6685, + "step": 11132 + }, + { + "epoch": 0.9487004686834256, + "grad_norm": 17.0, + "learning_rate": 1.3594189595404338e-06, + "loss": 0.7485, + "step": 11133 + }, + { + "epoch": 0.9487856838517256, + "grad_norm": 15.4375, + "learning_rate": 1.3592515680179452e-06, + "loss": 1.0117, + "step": 11134 + }, + { + "epoch": 0.9488708990200255, + "grad_norm": 14.5625, + "learning_rate": 1.3590841745211707e-06, + "loss": 0.7559, + "step": 11135 + }, + { + "epoch": 0.9489561141883255, + "grad_norm": 14.125, + "learning_rate": 1.3589167790531354e-06, + "loss": 0.996, + "step": 11136 + }, + { + "epoch": 0.9490413293566254, + "grad_norm": 14.25, + "learning_rate": 1.3587493816168648e-06, + "loss": 0.4351, + "step": 11137 + }, + { + "epoch": 0.9491265445249254, + "grad_norm": 12.5, + "learning_rate": 1.3585819822153834e-06, + "loss": 0.5481, + "step": 11138 + }, + { + "epoch": 0.9492117596932254, + "grad_norm": 11.625, + "learning_rate": 1.3584145808517163e-06, + "loss": 0.5167, + "step": 11139 + }, + { + "epoch": 0.9492969748615253, + "grad_norm": 23.5, + "learning_rate": 1.3582471775288883e-06, + "loss": 0.8805, + "step": 11140 + }, + { + "epoch": 0.9493821900298253, + "grad_norm": 12.625, + "learning_rate": 1.3580797722499256e-06, + "loss": 0.8834, + "step": 11141 + }, + { + "epoch": 0.9494674051981252, + "grad_norm": 15.6875, + "learning_rate": 1.3579123650178525e-06, + "loss": 1.0889, + "step": 11142 + }, + { + "epoch": 0.9495526203664252, + "grad_norm": 13.9375, + "learning_rate": 1.3577449558356944e-06, + "loss": 0.5831, + "step": 11143 + }, + { + "epoch": 0.9496378355347251, + "grad_norm": 12.1875, + "learning_rate": 1.3575775447064765e-06, + "loss": 0.5942, + "step": 11144 + }, + { + "epoch": 0.9497230507030251, + "grad_norm": 17.125, + "learning_rate": 1.3574101316332244e-06, + "loss": 0.4823, + "step": 11145 + }, + { + "epoch": 0.949808265871325, + "grad_norm": 11.5, + "learning_rate": 1.357242716618963e-06, + "loss": 0.6503, + "step": 11146 + }, + { + "epoch": 0.949893481039625, + "grad_norm": 13.5, + "learning_rate": 1.357075299666718e-06, + "loss": 0.8248, + "step": 11147 + }, + { + "epoch": 0.949978696207925, + "grad_norm": 23.75, + "learning_rate": 1.3569078807795148e-06, + "loss": 1.0501, + "step": 11148 + }, + { + "epoch": 0.9500639113762249, + "grad_norm": 16.125, + "learning_rate": 1.3567404599603784e-06, + "loss": 0.724, + "step": 11149 + }, + { + "epoch": 0.9501491265445249, + "grad_norm": 14.5625, + "learning_rate": 1.3565730372123348e-06, + "loss": 0.8842, + "step": 11150 + }, + { + "epoch": 0.9502343417128248, + "grad_norm": 16.0, + "learning_rate": 1.356405612538409e-06, + "loss": 1.0217, + "step": 11151 + }, + { + "epoch": 0.9503195568811248, + "grad_norm": 18.125, + "learning_rate": 1.3562381859416268e-06, + "loss": 0.6596, + "step": 11152 + }, + { + "epoch": 0.9504047720494248, + "grad_norm": 18.5, + "learning_rate": 1.3560707574250137e-06, + "loss": 1.1156, + "step": 11153 + }, + { + "epoch": 0.9504899872177247, + "grad_norm": 12.6875, + "learning_rate": 1.3559033269915959e-06, + "loss": 0.5993, + "step": 11154 + }, + { + "epoch": 0.9505752023860247, + "grad_norm": 13.75, + "learning_rate": 1.355735894644398e-06, + "loss": 0.6029, + "step": 11155 + }, + { + "epoch": 0.9506604175543246, + "grad_norm": 15.5, + "learning_rate": 1.3555684603864464e-06, + "loss": 0.8165, + "step": 11156 + }, + { + "epoch": 0.9507456327226246, + "grad_norm": 13.1875, + "learning_rate": 1.355401024220766e-06, + "loss": 0.7239, + "step": 11157 + }, + { + "epoch": 0.9508308478909245, + "grad_norm": 17.375, + "learning_rate": 1.355233586150384e-06, + "loss": 1.0024, + "step": 11158 + }, + { + "epoch": 0.9509160630592245, + "grad_norm": 21.875, + "learning_rate": 1.3550661461783245e-06, + "loss": 0.6146, + "step": 11159 + }, + { + "epoch": 0.9510012782275244, + "grad_norm": 11.5625, + "learning_rate": 1.3548987043076145e-06, + "loss": 0.5588, + "step": 11160 + }, + { + "epoch": 0.9510864933958244, + "grad_norm": 12.125, + "learning_rate": 1.3547312605412793e-06, + "loss": 0.5698, + "step": 11161 + }, + { + "epoch": 0.9511717085641244, + "grad_norm": 14.6875, + "learning_rate": 1.3545638148823453e-06, + "loss": 0.8529, + "step": 11162 + }, + { + "epoch": 0.9512569237324243, + "grad_norm": 19.25, + "learning_rate": 1.3543963673338372e-06, + "loss": 0.442, + "step": 11163 + }, + { + "epoch": 0.9513421389007243, + "grad_norm": 17.375, + "learning_rate": 1.3542289178987825e-06, + "loss": 0.8318, + "step": 11164 + }, + { + "epoch": 0.9514273540690242, + "grad_norm": 25.5, + "learning_rate": 1.354061466580206e-06, + "loss": 0.491, + "step": 11165 + }, + { + "epoch": 0.9515125692373242, + "grad_norm": 40.0, + "learning_rate": 1.3538940133811344e-06, + "loss": 0.9467, + "step": 11166 + }, + { + "epoch": 0.9515977844056241, + "grad_norm": 12.8125, + "learning_rate": 1.3537265583045933e-06, + "loss": 0.5424, + "step": 11167 + }, + { + "epoch": 0.9516829995739241, + "grad_norm": 15.0625, + "learning_rate": 1.353559101353609e-06, + "loss": 0.8793, + "step": 11168 + }, + { + "epoch": 0.9517682147422241, + "grad_norm": 14.75, + "learning_rate": 1.3533916425312076e-06, + "loss": 0.8525, + "step": 11169 + }, + { + "epoch": 0.951853429910524, + "grad_norm": 17.75, + "learning_rate": 1.3532241818404157e-06, + "loss": 1.0974, + "step": 11170 + }, + { + "epoch": 0.9519386450788241, + "grad_norm": 17.625, + "learning_rate": 1.3530567192842586e-06, + "loss": 0.9203, + "step": 11171 + }, + { + "epoch": 0.952023860247124, + "grad_norm": 14.8125, + "learning_rate": 1.352889254865763e-06, + "loss": 0.9527, + "step": 11172 + }, + { + "epoch": 0.952109075415424, + "grad_norm": 26.0, + "learning_rate": 1.352721788587955e-06, + "loss": 1.0209, + "step": 11173 + }, + { + "epoch": 0.952194290583724, + "grad_norm": 14.0625, + "learning_rate": 1.3525543204538613e-06, + "loss": 0.681, + "step": 11174 + }, + { + "epoch": 0.9522795057520239, + "grad_norm": 14.9375, + "learning_rate": 1.3523868504665077e-06, + "loss": 0.6908, + "step": 11175 + }, + { + "epoch": 0.9523647209203239, + "grad_norm": 25.375, + "learning_rate": 1.3522193786289206e-06, + "loss": 0.9578, + "step": 11176 + }, + { + "epoch": 0.9524499360886238, + "grad_norm": 17.125, + "learning_rate": 1.3520519049441268e-06, + "loss": 0.9875, + "step": 11177 + }, + { + "epoch": 0.9525351512569238, + "grad_norm": 17.375, + "learning_rate": 1.3518844294151523e-06, + "loss": 0.8847, + "step": 11178 + }, + { + "epoch": 0.9526203664252237, + "grad_norm": 13.9375, + "learning_rate": 1.351716952045024e-06, + "loss": 0.7185, + "step": 11179 + }, + { + "epoch": 0.9527055815935237, + "grad_norm": 20.0, + "learning_rate": 1.3515494728367677e-06, + "loss": 0.797, + "step": 11180 + }, + { + "epoch": 0.9527907967618237, + "grad_norm": 14.0625, + "learning_rate": 1.3513819917934108e-06, + "loss": 0.7569, + "step": 11181 + }, + { + "epoch": 0.9528760119301236, + "grad_norm": 11.9375, + "learning_rate": 1.3512145089179787e-06, + "loss": 0.5086, + "step": 11182 + }, + { + "epoch": 0.9529612270984236, + "grad_norm": 13.125, + "learning_rate": 1.351047024213499e-06, + "loss": 0.7267, + "step": 11183 + }, + { + "epoch": 0.9530464422667235, + "grad_norm": 12.1875, + "learning_rate": 1.350879537682998e-06, + "loss": 0.673, + "step": 11184 + }, + { + "epoch": 0.9531316574350235, + "grad_norm": 23.5, + "learning_rate": 1.3507120493295026e-06, + "loss": 0.9373, + "step": 11185 + }, + { + "epoch": 0.9532168726033234, + "grad_norm": 11.0, + "learning_rate": 1.3505445591560387e-06, + "loss": 0.4915, + "step": 11186 + }, + { + "epoch": 0.9533020877716234, + "grad_norm": 20.0, + "learning_rate": 1.3503770671656336e-06, + "loss": 0.9672, + "step": 11187 + }, + { + "epoch": 0.9533873029399234, + "grad_norm": 17.625, + "learning_rate": 1.350209573361314e-06, + "loss": 1.1056, + "step": 11188 + }, + { + "epoch": 0.9534725181082233, + "grad_norm": 20.125, + "learning_rate": 1.3500420777461065e-06, + "loss": 0.5954, + "step": 11189 + }, + { + "epoch": 0.9535577332765233, + "grad_norm": 14.5625, + "learning_rate": 1.3498745803230383e-06, + "loss": 0.9035, + "step": 11190 + }, + { + "epoch": 0.9536429484448232, + "grad_norm": 20.25, + "learning_rate": 1.3497070810951363e-06, + "loss": 0.8066, + "step": 11191 + }, + { + "epoch": 0.9537281636131232, + "grad_norm": 14.5, + "learning_rate": 1.3495395800654265e-06, + "loss": 0.9601, + "step": 11192 + }, + { + "epoch": 0.9538133787814231, + "grad_norm": 17.625, + "learning_rate": 1.3493720772369365e-06, + "loss": 1.07, + "step": 11193 + }, + { + "epoch": 0.9538985939497231, + "grad_norm": 21.875, + "learning_rate": 1.3492045726126934e-06, + "loss": 0.6457, + "step": 11194 + }, + { + "epoch": 0.953983809118023, + "grad_norm": 12.5625, + "learning_rate": 1.3490370661957238e-06, + "loss": 0.5155, + "step": 11195 + }, + { + "epoch": 0.954069024286323, + "grad_norm": 14.4375, + "learning_rate": 1.3488695579890549e-06, + "loss": 0.9245, + "step": 11196 + }, + { + "epoch": 0.954154239454623, + "grad_norm": 14.5625, + "learning_rate": 1.3487020479957133e-06, + "loss": 0.5774, + "step": 11197 + }, + { + "epoch": 0.9542394546229229, + "grad_norm": 14.5, + "learning_rate": 1.3485345362187268e-06, + "loss": 0.9729, + "step": 11198 + }, + { + "epoch": 0.9543246697912229, + "grad_norm": 13.3125, + "learning_rate": 1.348367022661122e-06, + "loss": 0.8422, + "step": 11199 + }, + { + "epoch": 0.9544098849595228, + "grad_norm": 11.3125, + "learning_rate": 1.3481995073259263e-06, + "loss": 0.3893, + "step": 11200 + }, + { + "epoch": 0.9544951001278228, + "grad_norm": 12.625, + "learning_rate": 1.3480319902161665e-06, + "loss": 0.5194, + "step": 11201 + }, + { + "epoch": 0.9545803152961228, + "grad_norm": 15.6875, + "learning_rate": 1.3478644713348705e-06, + "loss": 1.1465, + "step": 11202 + }, + { + "epoch": 0.9546655304644227, + "grad_norm": 12.625, + "learning_rate": 1.3476969506850648e-06, + "loss": 0.6907, + "step": 11203 + }, + { + "epoch": 0.9547507456327227, + "grad_norm": 12.8125, + "learning_rate": 1.3475294282697769e-06, + "loss": 0.7228, + "step": 11204 + }, + { + "epoch": 0.9548359608010226, + "grad_norm": 21.75, + "learning_rate": 1.347361904092034e-06, + "loss": 0.8344, + "step": 11205 + }, + { + "epoch": 0.9549211759693226, + "grad_norm": 11.5625, + "learning_rate": 1.3471943781548638e-06, + "loss": 0.4806, + "step": 11206 + }, + { + "epoch": 0.9550063911376225, + "grad_norm": 14.375, + "learning_rate": 1.3470268504612937e-06, + "loss": 0.526, + "step": 11207 + }, + { + "epoch": 0.9550916063059225, + "grad_norm": 14.5625, + "learning_rate": 1.3468593210143508e-06, + "loss": 0.4478, + "step": 11208 + }, + { + "epoch": 0.9551768214742224, + "grad_norm": 37.5, + "learning_rate": 1.3466917898170623e-06, + "loss": 1.4774, + "step": 11209 + }, + { + "epoch": 0.9552620366425224, + "grad_norm": 20.75, + "learning_rate": 1.3465242568724556e-06, + "loss": 0.9637, + "step": 11210 + }, + { + "epoch": 0.9553472518108224, + "grad_norm": 11.375, + "learning_rate": 1.3463567221835593e-06, + "loss": 0.5761, + "step": 11211 + }, + { + "epoch": 0.9554324669791223, + "grad_norm": 15.625, + "learning_rate": 1.3461891857533998e-06, + "loss": 0.9901, + "step": 11212 + }, + { + "epoch": 0.9555176821474223, + "grad_norm": 16.375, + "learning_rate": 1.346021647585005e-06, + "loss": 1.0439, + "step": 11213 + }, + { + "epoch": 0.9556028973157222, + "grad_norm": 16.125, + "learning_rate": 1.3458541076814022e-06, + "loss": 0.927, + "step": 11214 + }, + { + "epoch": 0.9556881124840222, + "grad_norm": 13.0, + "learning_rate": 1.34568656604562e-06, + "loss": 0.6717, + "step": 11215 + }, + { + "epoch": 0.9557733276523221, + "grad_norm": 18.375, + "learning_rate": 1.3455190226806846e-06, + "loss": 1.05, + "step": 11216 + }, + { + "epoch": 0.9558585428206221, + "grad_norm": 16.625, + "learning_rate": 1.3453514775896248e-06, + "loss": 0.7356, + "step": 11217 + }, + { + "epoch": 0.9559437579889221, + "grad_norm": 15.6875, + "learning_rate": 1.3451839307754673e-06, + "loss": 1.0657, + "step": 11218 + }, + { + "epoch": 0.956028973157222, + "grad_norm": 12.1875, + "learning_rate": 1.3450163822412413e-06, + "loss": 0.4628, + "step": 11219 + }, + { + "epoch": 0.956114188325522, + "grad_norm": 17.375, + "learning_rate": 1.3448488319899734e-06, + "loss": 0.6879, + "step": 11220 + }, + { + "epoch": 0.9561994034938219, + "grad_norm": 13.25, + "learning_rate": 1.3446812800246913e-06, + "loss": 0.4198, + "step": 11221 + }, + { + "epoch": 0.9562846186621219, + "grad_norm": 13.5, + "learning_rate": 1.344513726348424e-06, + "loss": 0.5079, + "step": 11222 + }, + { + "epoch": 0.9563698338304218, + "grad_norm": 18.75, + "learning_rate": 1.3443461709641983e-06, + "loss": 0.6333, + "step": 11223 + }, + { + "epoch": 0.9564550489987218, + "grad_norm": 19.25, + "learning_rate": 1.3441786138750424e-06, + "loss": 0.7613, + "step": 11224 + }, + { + "epoch": 0.9565402641670218, + "grad_norm": 44.25, + "learning_rate": 1.344011055083984e-06, + "loss": 1.1897, + "step": 11225 + }, + { + "epoch": 0.9566254793353217, + "grad_norm": 12.0, + "learning_rate": 1.3438434945940515e-06, + "loss": 0.7008, + "step": 11226 + }, + { + "epoch": 0.9567106945036217, + "grad_norm": 12.8125, + "learning_rate": 1.343675932408273e-06, + "loss": 0.9346, + "step": 11227 + }, + { + "epoch": 0.9567959096719216, + "grad_norm": 17.875, + "learning_rate": 1.343508368529676e-06, + "loss": 0.7028, + "step": 11228 + }, + { + "epoch": 0.9568811248402216, + "grad_norm": 14.5625, + "learning_rate": 1.3433408029612885e-06, + "loss": 0.8707, + "step": 11229 + }, + { + "epoch": 0.9569663400085215, + "grad_norm": 22.25, + "learning_rate": 1.3431732357061389e-06, + "loss": 1.4393, + "step": 11230 + }, + { + "epoch": 0.9570515551768215, + "grad_norm": 11.125, + "learning_rate": 1.3430056667672555e-06, + "loss": 0.3387, + "step": 11231 + }, + { + "epoch": 0.9571367703451215, + "grad_norm": 18.375, + "learning_rate": 1.3428380961476662e-06, + "loss": 0.5938, + "step": 11232 + }, + { + "epoch": 0.9572219855134214, + "grad_norm": 19.25, + "learning_rate": 1.3426705238503986e-06, + "loss": 0.9805, + "step": 11233 + }, + { + "epoch": 0.9573072006817214, + "grad_norm": 18.875, + "learning_rate": 1.3425029498784819e-06, + "loss": 1.0721, + "step": 11234 + }, + { + "epoch": 0.9573924158500213, + "grad_norm": 10.3125, + "learning_rate": 1.3423353742349441e-06, + "loss": 0.355, + "step": 11235 + }, + { + "epoch": 0.9574776310183213, + "grad_norm": 17.625, + "learning_rate": 1.3421677969228124e-06, + "loss": 0.7045, + "step": 11236 + }, + { + "epoch": 0.9575628461866212, + "grad_norm": 15.1875, + "learning_rate": 1.3420002179451167e-06, + "loss": 0.6337, + "step": 11237 + }, + { + "epoch": 0.9576480613549212, + "grad_norm": 22.0, + "learning_rate": 1.341832637304884e-06, + "loss": 0.9191, + "step": 11238 + }, + { + "epoch": 0.9577332765232212, + "grad_norm": 21.0, + "learning_rate": 1.3416650550051438e-06, + "loss": 1.2782, + "step": 11239 + }, + { + "epoch": 0.9578184916915211, + "grad_norm": 17.25, + "learning_rate": 1.3414974710489234e-06, + "loss": 0.7834, + "step": 11240 + }, + { + "epoch": 0.9579037068598211, + "grad_norm": 13.0, + "learning_rate": 1.341329885439252e-06, + "loss": 0.7518, + "step": 11241 + }, + { + "epoch": 0.957988922028121, + "grad_norm": 10.75, + "learning_rate": 1.3411622981791574e-06, + "loss": 0.3337, + "step": 11242 + }, + { + "epoch": 0.958074137196421, + "grad_norm": 16.5, + "learning_rate": 1.340994709271669e-06, + "loss": 0.6011, + "step": 11243 + }, + { + "epoch": 0.9581593523647209, + "grad_norm": 14.0, + "learning_rate": 1.3408271187198142e-06, + "loss": 0.5868, + "step": 11244 + }, + { + "epoch": 0.9582445675330209, + "grad_norm": 9.5, + "learning_rate": 1.340659526526622e-06, + "loss": 0.4471, + "step": 11245 + }, + { + "epoch": 0.9583297827013209, + "grad_norm": 14.4375, + "learning_rate": 1.340491932695121e-06, + "loss": 0.4826, + "step": 11246 + }, + { + "epoch": 0.9584149978696208, + "grad_norm": 12.375, + "learning_rate": 1.34032433722834e-06, + "loss": 0.6259, + "step": 11247 + }, + { + "epoch": 0.9585002130379208, + "grad_norm": 16.375, + "learning_rate": 1.3401567401293075e-06, + "loss": 0.9066, + "step": 11248 + }, + { + "epoch": 0.9585854282062207, + "grad_norm": 26.0, + "learning_rate": 1.3399891414010519e-06, + "loss": 1.1396, + "step": 11249 + }, + { + "epoch": 0.9586706433745207, + "grad_norm": 18.75, + "learning_rate": 1.3398215410466018e-06, + "loss": 1.0114, + "step": 11250 + }, + { + "epoch": 0.9587558585428206, + "grad_norm": 15.375, + "learning_rate": 1.3396539390689863e-06, + "loss": 0.9949, + "step": 11251 + }, + { + "epoch": 0.9588410737111206, + "grad_norm": 13.6875, + "learning_rate": 1.339486335471234e-06, + "loss": 0.6976, + "step": 11252 + }, + { + "epoch": 0.9589262888794206, + "grad_norm": 18.75, + "learning_rate": 1.3393187302563734e-06, + "loss": 0.698, + "step": 11253 + }, + { + "epoch": 0.9590115040477205, + "grad_norm": 12.3125, + "learning_rate": 1.3391511234274337e-06, + "loss": 0.8465, + "step": 11254 + }, + { + "epoch": 0.9590967192160205, + "grad_norm": 14.75, + "learning_rate": 1.3389835149874438e-06, + "loss": 0.7536, + "step": 11255 + }, + { + "epoch": 0.9591819343843204, + "grad_norm": 17.25, + "learning_rate": 1.3388159049394323e-06, + "loss": 1.0718, + "step": 11256 + }, + { + "epoch": 0.9592671495526204, + "grad_norm": 12.3125, + "learning_rate": 1.3386482932864275e-06, + "loss": 0.4184, + "step": 11257 + }, + { + "epoch": 0.9593523647209203, + "grad_norm": 31.125, + "learning_rate": 1.3384806800314592e-06, + "loss": 0.7978, + "step": 11258 + }, + { + "epoch": 0.9594375798892203, + "grad_norm": 14.0625, + "learning_rate": 1.3383130651775561e-06, + "loss": 0.6583, + "step": 11259 + }, + { + "epoch": 0.9595227950575203, + "grad_norm": 15.625, + "learning_rate": 1.3381454487277473e-06, + "loss": 0.4511, + "step": 11260 + }, + { + "epoch": 0.9596080102258202, + "grad_norm": 17.75, + "learning_rate": 1.3379778306850614e-06, + "loss": 0.8283, + "step": 11261 + }, + { + "epoch": 0.9596932253941202, + "grad_norm": 11.9375, + "learning_rate": 1.3378102110525277e-06, + "loss": 0.3871, + "step": 11262 + }, + { + "epoch": 0.9597784405624201, + "grad_norm": 13.625, + "learning_rate": 1.337642589833175e-06, + "loss": 0.4974, + "step": 11263 + }, + { + "epoch": 0.9598636557307201, + "grad_norm": 13.6875, + "learning_rate": 1.337474967030033e-06, + "loss": 0.8338, + "step": 11264 + }, + { + "epoch": 0.95994887089902, + "grad_norm": 15.875, + "learning_rate": 1.33730734264613e-06, + "loss": 0.714, + "step": 11265 + }, + { + "epoch": 0.96003408606732, + "grad_norm": 14.8125, + "learning_rate": 1.3371397166844955e-06, + "loss": 1.0014, + "step": 11266 + }, + { + "epoch": 0.96011930123562, + "grad_norm": 20.875, + "learning_rate": 1.3369720891481585e-06, + "loss": 1.0857, + "step": 11267 + }, + { + "epoch": 0.9602045164039199, + "grad_norm": 15.3125, + "learning_rate": 1.3368044600401489e-06, + "loss": 0.6234, + "step": 11268 + }, + { + "epoch": 0.9602897315722199, + "grad_norm": 13.125, + "learning_rate": 1.3366368293634954e-06, + "loss": 0.5609, + "step": 11269 + }, + { + "epoch": 0.9603749467405198, + "grad_norm": 12.9375, + "learning_rate": 1.3364691971212271e-06, + "loss": 0.5431, + "step": 11270 + }, + { + "epoch": 0.9604601619088198, + "grad_norm": 14.5625, + "learning_rate": 1.336301563316373e-06, + "loss": 0.5444, + "step": 11271 + }, + { + "epoch": 0.9605453770771197, + "grad_norm": 12.25, + "learning_rate": 1.3361339279519635e-06, + "loss": 0.6693, + "step": 11272 + }, + { + "epoch": 0.9606305922454197, + "grad_norm": 17.625, + "learning_rate": 1.3359662910310273e-06, + "loss": 1.0053, + "step": 11273 + }, + { + "epoch": 0.9607158074137196, + "grad_norm": 15.0, + "learning_rate": 1.3357986525565934e-06, + "loss": 0.8832, + "step": 11274 + }, + { + "epoch": 0.9608010225820196, + "grad_norm": 13.375, + "learning_rate": 1.3356310125316917e-06, + "loss": 0.6233, + "step": 11275 + }, + { + "epoch": 0.9608862377503196, + "grad_norm": 13.25, + "learning_rate": 1.3354633709593517e-06, + "loss": 0.5333, + "step": 11276 + }, + { + "epoch": 0.9609714529186195, + "grad_norm": 13.1875, + "learning_rate": 1.3352957278426026e-06, + "loss": 0.7211, + "step": 11277 + }, + { + "epoch": 0.9610566680869195, + "grad_norm": 14.375, + "learning_rate": 1.3351280831844737e-06, + "loss": 0.5842, + "step": 11278 + }, + { + "epoch": 0.9611418832552194, + "grad_norm": 10.3125, + "learning_rate": 1.334960436987995e-06, + "loss": 0.3551, + "step": 11279 + }, + { + "epoch": 0.9612270984235194, + "grad_norm": 11.8125, + "learning_rate": 1.3347927892561956e-06, + "loss": 0.4117, + "step": 11280 + }, + { + "epoch": 0.9613123135918193, + "grad_norm": 17.0, + "learning_rate": 1.3346251399921053e-06, + "loss": 1.1344, + "step": 11281 + }, + { + "epoch": 0.9613975287601193, + "grad_norm": 17.375, + "learning_rate": 1.3344574891987535e-06, + "loss": 0.9437, + "step": 11282 + }, + { + "epoch": 0.9614827439284193, + "grad_norm": 20.625, + "learning_rate": 1.33428983687917e-06, + "loss": 1.0353, + "step": 11283 + }, + { + "epoch": 0.9615679590967192, + "grad_norm": 16.875, + "learning_rate": 1.3341221830363845e-06, + "loss": 0.8744, + "step": 11284 + }, + { + "epoch": 0.9616531742650192, + "grad_norm": 11.9375, + "learning_rate": 1.3339545276734267e-06, + "loss": 0.3047, + "step": 11285 + }, + { + "epoch": 0.9617383894333191, + "grad_norm": 19.875, + "learning_rate": 1.3337868707933258e-06, + "loss": 0.8173, + "step": 11286 + }, + { + "epoch": 0.9618236046016191, + "grad_norm": 10.375, + "learning_rate": 1.333619212399112e-06, + "loss": 0.4057, + "step": 11287 + }, + { + "epoch": 0.961908819769919, + "grad_norm": 12.25, + "learning_rate": 1.3334515524938152e-06, + "loss": 0.4672, + "step": 11288 + }, + { + "epoch": 0.961994034938219, + "grad_norm": 14.4375, + "learning_rate": 1.3332838910804644e-06, + "loss": 0.5211, + "step": 11289 + }, + { + "epoch": 0.962079250106519, + "grad_norm": 17.375, + "learning_rate": 1.3331162281620905e-06, + "loss": 0.4687, + "step": 11290 + }, + { + "epoch": 0.9621644652748189, + "grad_norm": 12.8125, + "learning_rate": 1.3329485637417223e-06, + "loss": 0.6322, + "step": 11291 + }, + { + "epoch": 0.9622496804431189, + "grad_norm": 12.8125, + "learning_rate": 1.3327808978223905e-06, + "loss": 0.7581, + "step": 11292 + }, + { + "epoch": 0.9623348956114188, + "grad_norm": 16.125, + "learning_rate": 1.3326132304071244e-06, + "loss": 0.846, + "step": 11293 + }, + { + "epoch": 0.9624201107797188, + "grad_norm": 12.125, + "learning_rate": 1.3324455614989542e-06, + "loss": 0.534, + "step": 11294 + }, + { + "epoch": 0.9625053259480187, + "grad_norm": 14.5, + "learning_rate": 1.3322778911009099e-06, + "loss": 0.693, + "step": 11295 + }, + { + "epoch": 0.9625905411163187, + "grad_norm": 14.125, + "learning_rate": 1.3321102192160216e-06, + "loss": 0.5275, + "step": 11296 + }, + { + "epoch": 0.9626757562846187, + "grad_norm": 18.25, + "learning_rate": 1.3319425458473188e-06, + "loss": 1.0927, + "step": 11297 + }, + { + "epoch": 0.9627609714529186, + "grad_norm": 34.25, + "learning_rate": 1.3317748709978317e-06, + "loss": 0.9722, + "step": 11298 + }, + { + "epoch": 0.9628461866212186, + "grad_norm": 16.0, + "learning_rate": 1.3316071946705902e-06, + "loss": 1.2774, + "step": 11299 + }, + { + "epoch": 0.9629314017895185, + "grad_norm": 13.375, + "learning_rate": 1.3314395168686256e-06, + "loss": 0.7155, + "step": 11300 + }, + { + "epoch": 0.9630166169578185, + "grad_norm": 13.3125, + "learning_rate": 1.331271837594966e-06, + "loss": 0.7423, + "step": 11301 + }, + { + "epoch": 0.9631018321261184, + "grad_norm": 15.1875, + "learning_rate": 1.3311041568526432e-06, + "loss": 0.8511, + "step": 11302 + }, + { + "epoch": 0.9631870472944184, + "grad_norm": 18.375, + "learning_rate": 1.3309364746446863e-06, + "loss": 1.085, + "step": 11303 + }, + { + "epoch": 0.9632722624627184, + "grad_norm": 12.0, + "learning_rate": 1.3307687909741263e-06, + "loss": 0.4792, + "step": 11304 + }, + { + "epoch": 0.9633574776310183, + "grad_norm": 13.0, + "learning_rate": 1.3306011058439926e-06, + "loss": 0.7825, + "step": 11305 + }, + { + "epoch": 0.9634426927993183, + "grad_norm": 14.4375, + "learning_rate": 1.3304334192573163e-06, + "loss": 0.6797, + "step": 11306 + }, + { + "epoch": 0.9635279079676182, + "grad_norm": 13.5, + "learning_rate": 1.3302657312171268e-06, + "loss": 0.649, + "step": 11307 + }, + { + "epoch": 0.9636131231359182, + "grad_norm": 14.5625, + "learning_rate": 1.3300980417264552e-06, + "loss": 0.5308, + "step": 11308 + }, + { + "epoch": 0.9636983383042181, + "grad_norm": 14.5, + "learning_rate": 1.3299303507883316e-06, + "loss": 0.7404, + "step": 11309 + }, + { + "epoch": 0.9637835534725181, + "grad_norm": 15.8125, + "learning_rate": 1.3297626584057853e-06, + "loss": 0.9554, + "step": 11310 + }, + { + "epoch": 0.963868768640818, + "grad_norm": 21.5, + "learning_rate": 1.3295949645818484e-06, + "loss": 0.9217, + "step": 11311 + }, + { + "epoch": 0.963953983809118, + "grad_norm": 15.8125, + "learning_rate": 1.3294272693195504e-06, + "loss": 0.9075, + "step": 11312 + }, + { + "epoch": 0.964039198977418, + "grad_norm": 16.75, + "learning_rate": 1.3292595726219215e-06, + "loss": 0.753, + "step": 11313 + }, + { + "epoch": 0.9641244141457179, + "grad_norm": 18.375, + "learning_rate": 1.3290918744919926e-06, + "loss": 0.8299, + "step": 11314 + }, + { + "epoch": 0.9642096293140179, + "grad_norm": 15.5625, + "learning_rate": 1.3289241749327939e-06, + "loss": 0.7705, + "step": 11315 + }, + { + "epoch": 0.9642948444823178, + "grad_norm": 14.5, + "learning_rate": 1.3287564739473558e-06, + "loss": 0.4943, + "step": 11316 + }, + { + "epoch": 0.9643800596506178, + "grad_norm": 11.5625, + "learning_rate": 1.3285887715387095e-06, + "loss": 0.2655, + "step": 11317 + }, + { + "epoch": 0.9644652748189178, + "grad_norm": 16.0, + "learning_rate": 1.3284210677098847e-06, + "loss": 0.8507, + "step": 11318 + }, + { + "epoch": 0.9645504899872177, + "grad_norm": 12.75, + "learning_rate": 1.3282533624639124e-06, + "loss": 0.6636, + "step": 11319 + }, + { + "epoch": 0.9646357051555177, + "grad_norm": 14.8125, + "learning_rate": 1.328085655803823e-06, + "loss": 0.8428, + "step": 11320 + }, + { + "epoch": 0.9647209203238176, + "grad_norm": 14.8125, + "learning_rate": 1.3279179477326475e-06, + "loss": 0.9147, + "step": 11321 + }, + { + "epoch": 0.9648061354921176, + "grad_norm": 14.5625, + "learning_rate": 1.3277502382534163e-06, + "loss": 0.7329, + "step": 11322 + }, + { + "epoch": 0.9648913506604175, + "grad_norm": 14.4375, + "learning_rate": 1.3275825273691603e-06, + "loss": 0.5594, + "step": 11323 + }, + { + "epoch": 0.9649765658287175, + "grad_norm": 10.75, + "learning_rate": 1.3274148150829098e-06, + "loss": 0.5454, + "step": 11324 + }, + { + "epoch": 0.9650617809970174, + "grad_norm": 20.5, + "learning_rate": 1.3272471013976956e-06, + "loss": 1.0577, + "step": 11325 + }, + { + "epoch": 0.9651469961653174, + "grad_norm": 19.25, + "learning_rate": 1.3270793863165489e-06, + "loss": 0.8825, + "step": 11326 + }, + { + "epoch": 0.9652322113336174, + "grad_norm": 13.8125, + "learning_rate": 1.3269116698424997e-06, + "loss": 0.6828, + "step": 11327 + }, + { + "epoch": 0.9653174265019173, + "grad_norm": 18.375, + "learning_rate": 1.32674395197858e-06, + "loss": 0.714, + "step": 11328 + }, + { + "epoch": 0.9654026416702173, + "grad_norm": 15.125, + "learning_rate": 1.3265762327278195e-06, + "loss": 0.682, + "step": 11329 + }, + { + "epoch": 0.9654878568385172, + "grad_norm": 12.3125, + "learning_rate": 1.3264085120932495e-06, + "loss": 0.6552, + "step": 11330 + }, + { + "epoch": 0.9655730720068172, + "grad_norm": 13.75, + "learning_rate": 1.3262407900779006e-06, + "loss": 0.4089, + "step": 11331 + }, + { + "epoch": 0.9656582871751171, + "grad_norm": 14.9375, + "learning_rate": 1.3260730666848043e-06, + "loss": 0.5968, + "step": 11332 + }, + { + "epoch": 0.9657435023434171, + "grad_norm": 12.4375, + "learning_rate": 1.3259053419169912e-06, + "loss": 0.6452, + "step": 11333 + }, + { + "epoch": 0.9658287175117171, + "grad_norm": 12.0625, + "learning_rate": 1.3257376157774924e-06, + "loss": 0.3799, + "step": 11334 + }, + { + "epoch": 0.965913932680017, + "grad_norm": 18.625, + "learning_rate": 1.3255698882693382e-06, + "loss": 1.0815, + "step": 11335 + }, + { + "epoch": 0.965999147848317, + "grad_norm": 16.125, + "learning_rate": 1.3254021593955605e-06, + "loss": 1.021, + "step": 11336 + }, + { + "epoch": 0.9660843630166169, + "grad_norm": 17.0, + "learning_rate": 1.3252344291591898e-06, + "loss": 0.9381, + "step": 11337 + }, + { + "epoch": 0.9661695781849169, + "grad_norm": 11.5, + "learning_rate": 1.3250666975632577e-06, + "loss": 0.4586, + "step": 11338 + }, + { + "epoch": 0.9662547933532168, + "grad_norm": 21.125, + "learning_rate": 1.3248989646107943e-06, + "loss": 1.1198, + "step": 11339 + }, + { + "epoch": 0.9663400085215168, + "grad_norm": 13.75, + "learning_rate": 1.3247312303048315e-06, + "loss": 0.5648, + "step": 11340 + }, + { + "epoch": 0.9664252236898168, + "grad_norm": 14.75, + "learning_rate": 1.3245634946484005e-06, + "loss": 0.7224, + "step": 11341 + }, + { + "epoch": 0.9665104388581167, + "grad_norm": 16.125, + "learning_rate": 1.3243957576445314e-06, + "loss": 0.7446, + "step": 11342 + }, + { + "epoch": 0.9665956540264167, + "grad_norm": 12.125, + "learning_rate": 1.324228019296257e-06, + "loss": 0.5621, + "step": 11343 + }, + { + "epoch": 0.9666808691947166, + "grad_norm": 13.5625, + "learning_rate": 1.324060279606607e-06, + "loss": 0.7114, + "step": 11344 + }, + { + "epoch": 0.9667660843630166, + "grad_norm": 14.0625, + "learning_rate": 1.3238925385786141e-06, + "loss": 0.8529, + "step": 11345 + }, + { + "epoch": 0.9668512995313165, + "grad_norm": 13.0, + "learning_rate": 1.323724796215308e-06, + "loss": 0.6608, + "step": 11346 + }, + { + "epoch": 0.9669365146996165, + "grad_norm": 11.8125, + "learning_rate": 1.3235570525197212e-06, + "loss": 0.3913, + "step": 11347 + }, + { + "epoch": 0.9670217298679165, + "grad_norm": 25.0, + "learning_rate": 1.3233893074948838e-06, + "loss": 0.9876, + "step": 11348 + }, + { + "epoch": 0.9671069450362164, + "grad_norm": 21.125, + "learning_rate": 1.3232215611438288e-06, + "loss": 0.9292, + "step": 11349 + }, + { + "epoch": 0.9671921602045164, + "grad_norm": 13.375, + "learning_rate": 1.323053813469586e-06, + "loss": 0.692, + "step": 11350 + }, + { + "epoch": 0.9672773753728163, + "grad_norm": 13.3125, + "learning_rate": 1.322886064475187e-06, + "loss": 0.8455, + "step": 11351 + }, + { + "epoch": 0.9673625905411163, + "grad_norm": 14.125, + "learning_rate": 1.3227183141636637e-06, + "loss": 0.7569, + "step": 11352 + }, + { + "epoch": 0.9674478057094162, + "grad_norm": 12.75, + "learning_rate": 1.3225505625380481e-06, + "loss": 0.6274, + "step": 11353 + }, + { + "epoch": 0.9675330208777162, + "grad_norm": 16.125, + "learning_rate": 1.32238280960137e-06, + "loss": 0.6565, + "step": 11354 + }, + { + "epoch": 0.9676182360460162, + "grad_norm": 13.3125, + "learning_rate": 1.322215055356662e-06, + "loss": 0.7698, + "step": 11355 + }, + { + "epoch": 0.9677034512143161, + "grad_norm": 17.0, + "learning_rate": 1.3220472998069553e-06, + "loss": 0.9246, + "step": 11356 + }, + { + "epoch": 0.9677886663826161, + "grad_norm": 12.5625, + "learning_rate": 1.3218795429552813e-06, + "loss": 0.6619, + "step": 11357 + }, + { + "epoch": 0.967873881550916, + "grad_norm": 15.0625, + "learning_rate": 1.3217117848046721e-06, + "loss": 0.9119, + "step": 11358 + }, + { + "epoch": 0.967959096719216, + "grad_norm": 14.9375, + "learning_rate": 1.3215440253581585e-06, + "loss": 0.796, + "step": 11359 + }, + { + "epoch": 0.9680443118875159, + "grad_norm": 15.75, + "learning_rate": 1.3213762646187724e-06, + "loss": 0.7491, + "step": 11360 + }, + { + "epoch": 0.9681295270558159, + "grad_norm": 13.375, + "learning_rate": 1.3212085025895458e-06, + "loss": 0.674, + "step": 11361 + }, + { + "epoch": 0.9682147422241159, + "grad_norm": 17.125, + "learning_rate": 1.3210407392735097e-06, + "loss": 0.5801, + "step": 11362 + }, + { + "epoch": 0.9682999573924158, + "grad_norm": 20.375, + "learning_rate": 1.3208729746736958e-06, + "loss": 0.9625, + "step": 11363 + }, + { + "epoch": 0.9683851725607158, + "grad_norm": 18.25, + "learning_rate": 1.3207052087931362e-06, + "loss": 0.8111, + "step": 11364 + }, + { + "epoch": 0.9684703877290157, + "grad_norm": 11.375, + "learning_rate": 1.3205374416348627e-06, + "loss": 0.5216, + "step": 11365 + }, + { + "epoch": 0.9685556028973157, + "grad_norm": 14.5, + "learning_rate": 1.3203696732019065e-06, + "loss": 0.7154, + "step": 11366 + }, + { + "epoch": 0.9686408180656156, + "grad_norm": 19.75, + "learning_rate": 1.320201903497299e-06, + "loss": 0.8924, + "step": 11367 + }, + { + "epoch": 0.9687260332339156, + "grad_norm": 11.1875, + "learning_rate": 1.320034132524073e-06, + "loss": 0.4112, + "step": 11368 + }, + { + "epoch": 0.9688112484022156, + "grad_norm": 11.875, + "learning_rate": 1.31986636028526e-06, + "loss": 0.5414, + "step": 11369 + }, + { + "epoch": 0.9688964635705155, + "grad_norm": 16.0, + "learning_rate": 1.3196985867838913e-06, + "loss": 0.7155, + "step": 11370 + }, + { + "epoch": 0.9689816787388155, + "grad_norm": 11.9375, + "learning_rate": 1.3195308120229988e-06, + "loss": 0.3473, + "step": 11371 + }, + { + "epoch": 0.9690668939071154, + "grad_norm": 15.0625, + "learning_rate": 1.319363036005615e-06, + "loss": 0.8222, + "step": 11372 + }, + { + "epoch": 0.9691521090754154, + "grad_norm": 20.375, + "learning_rate": 1.3191952587347714e-06, + "loss": 0.978, + "step": 11373 + }, + { + "epoch": 0.9692373242437153, + "grad_norm": 18.0, + "learning_rate": 1.3190274802134995e-06, + "loss": 0.6053, + "step": 11374 + }, + { + "epoch": 0.9693225394120153, + "grad_norm": 22.125, + "learning_rate": 1.318859700444832e-06, + "loss": 0.9879, + "step": 11375 + }, + { + "epoch": 0.9694077545803153, + "grad_norm": 14.5625, + "learning_rate": 1.3186919194318002e-06, + "loss": 0.8097, + "step": 11376 + }, + { + "epoch": 0.9694929697486152, + "grad_norm": 14.625, + "learning_rate": 1.3185241371774368e-06, + "loss": 1.3046, + "step": 11377 + }, + { + "epoch": 0.9695781849169152, + "grad_norm": 15.0625, + "learning_rate": 1.3183563536847729e-06, + "loss": 0.7367, + "step": 11378 + }, + { + "epoch": 0.9696634000852151, + "grad_norm": 20.5, + "learning_rate": 1.318188568956841e-06, + "loss": 0.8635, + "step": 11379 + }, + { + "epoch": 0.9697486152535151, + "grad_norm": 16.0, + "learning_rate": 1.318020782996673e-06, + "loss": 0.8515, + "step": 11380 + }, + { + "epoch": 0.969833830421815, + "grad_norm": 14.625, + "learning_rate": 1.3178529958073014e-06, + "loss": 0.8373, + "step": 11381 + }, + { + "epoch": 0.969919045590115, + "grad_norm": 13.75, + "learning_rate": 1.317685207391758e-06, + "loss": 0.7394, + "step": 11382 + }, + { + "epoch": 0.970004260758415, + "grad_norm": 11.75, + "learning_rate": 1.3175174177530744e-06, + "loss": 0.5973, + "step": 11383 + }, + { + "epoch": 0.9700894759267149, + "grad_norm": 15.0, + "learning_rate": 1.3173496268942835e-06, + "loss": 0.6504, + "step": 11384 + }, + { + "epoch": 0.9701746910950149, + "grad_norm": 15.375, + "learning_rate": 1.3171818348184168e-06, + "loss": 0.7858, + "step": 11385 + }, + { + "epoch": 0.9702599062633148, + "grad_norm": 17.5, + "learning_rate": 1.3170140415285071e-06, + "loss": 0.843, + "step": 11386 + }, + { + "epoch": 0.9703451214316148, + "grad_norm": 13.8125, + "learning_rate": 1.3168462470275864e-06, + "loss": 0.494, + "step": 11387 + }, + { + "epoch": 0.9704303365999147, + "grad_norm": 12.1875, + "learning_rate": 1.316678451318686e-06, + "loss": 0.4649, + "step": 11388 + }, + { + "epoch": 0.9705155517682147, + "grad_norm": 15.6875, + "learning_rate": 1.3165106544048397e-06, + "loss": 0.7248, + "step": 11389 + }, + { + "epoch": 0.9706007669365146, + "grad_norm": 19.25, + "learning_rate": 1.316342856289079e-06, + "loss": 0.9645, + "step": 11390 + }, + { + "epoch": 0.9706859821048146, + "grad_norm": 10.25, + "learning_rate": 1.3161750569744357e-06, + "loss": 0.6843, + "step": 11391 + }, + { + "epoch": 0.9707711972731146, + "grad_norm": 13.3125, + "learning_rate": 1.3160072564639429e-06, + "loss": 0.4417, + "step": 11392 + }, + { + "epoch": 0.9708564124414145, + "grad_norm": 20.375, + "learning_rate": 1.3158394547606326e-06, + "loss": 0.8439, + "step": 11393 + }, + { + "epoch": 0.9709416276097145, + "grad_norm": 11.375, + "learning_rate": 1.3156716518675372e-06, + "loss": 0.4711, + "step": 11394 + }, + { + "epoch": 0.9710268427780144, + "grad_norm": 11.4375, + "learning_rate": 1.3155038477876886e-06, + "loss": 0.4872, + "step": 11395 + }, + { + "epoch": 0.9711120579463144, + "grad_norm": 14.6875, + "learning_rate": 1.31533604252412e-06, + "loss": 0.7636, + "step": 11396 + }, + { + "epoch": 0.9711972731146145, + "grad_norm": 14.5, + "learning_rate": 1.3151682360798632e-06, + "loss": 0.5719, + "step": 11397 + }, + { + "epoch": 0.9712824882829144, + "grad_norm": 13.125, + "learning_rate": 1.3150004284579513e-06, + "loss": 0.402, + "step": 11398 + }, + { + "epoch": 0.9713677034512144, + "grad_norm": 15.1875, + "learning_rate": 1.3148326196614158e-06, + "loss": 0.7881, + "step": 11399 + }, + { + "epoch": 0.9714529186195143, + "grad_norm": 17.5, + "learning_rate": 1.31466480969329e-06, + "loss": 0.8346, + "step": 11400 + }, + { + "epoch": 0.9715381337878143, + "grad_norm": 15.125, + "learning_rate": 1.3144969985566057e-06, + "loss": 0.9679, + "step": 11401 + }, + { + "epoch": 0.9716233489561142, + "grad_norm": 25.625, + "learning_rate": 1.314329186254396e-06, + "loss": 0.7881, + "step": 11402 + }, + { + "epoch": 0.9717085641244142, + "grad_norm": 12.6875, + "learning_rate": 1.3141613727896932e-06, + "loss": 0.4283, + "step": 11403 + }, + { + "epoch": 0.9717937792927142, + "grad_norm": 21.0, + "learning_rate": 1.3139935581655297e-06, + "loss": 1.1608, + "step": 11404 + }, + { + "epoch": 0.9718789944610141, + "grad_norm": 12.8125, + "learning_rate": 1.3138257423849382e-06, + "loss": 0.7783, + "step": 11405 + }, + { + "epoch": 0.9719642096293141, + "grad_norm": 20.875, + "learning_rate": 1.3136579254509518e-06, + "loss": 0.8851, + "step": 11406 + }, + { + "epoch": 0.972049424797614, + "grad_norm": 15.4375, + "learning_rate": 1.3134901073666022e-06, + "loss": 0.5869, + "step": 11407 + }, + { + "epoch": 0.972134639965914, + "grad_norm": 12.875, + "learning_rate": 1.313322288134923e-06, + "loss": 0.7989, + "step": 11408 + }, + { + "epoch": 0.9722198551342139, + "grad_norm": 14.75, + "learning_rate": 1.313154467758946e-06, + "loss": 0.6269, + "step": 11409 + }, + { + "epoch": 0.9723050703025139, + "grad_norm": 25.375, + "learning_rate": 1.3129866462417047e-06, + "loss": 1.1742, + "step": 11410 + }, + { + "epoch": 0.9723902854708139, + "grad_norm": 13.125, + "learning_rate": 1.3128188235862308e-06, + "loss": 0.5059, + "step": 11411 + }, + { + "epoch": 0.9724755006391138, + "grad_norm": 19.875, + "learning_rate": 1.3126509997955579e-06, + "loss": 1.0356, + "step": 11412 + }, + { + "epoch": 0.9725607158074138, + "grad_norm": 14.9375, + "learning_rate": 1.3124831748727183e-06, + "loss": 0.7453, + "step": 11413 + }, + { + "epoch": 0.9726459309757137, + "grad_norm": 14.875, + "learning_rate": 1.312315348820745e-06, + "loss": 0.6555, + "step": 11414 + }, + { + "epoch": 0.9727311461440137, + "grad_norm": 14.9375, + "learning_rate": 1.3121475216426708e-06, + "loss": 0.7213, + "step": 11415 + }, + { + "epoch": 0.9728163613123136, + "grad_norm": 14.3125, + "learning_rate": 1.311979693341528e-06, + "loss": 0.7463, + "step": 11416 + }, + { + "epoch": 0.9729015764806136, + "grad_norm": 11.9375, + "learning_rate": 1.31181186392035e-06, + "loss": 0.66, + "step": 11417 + }, + { + "epoch": 0.9729867916489136, + "grad_norm": 15.6875, + "learning_rate": 1.3116440333821699e-06, + "loss": 0.9052, + "step": 11418 + }, + { + "epoch": 0.9730720068172135, + "grad_norm": 16.5, + "learning_rate": 1.3114762017300194e-06, + "loss": 0.9488, + "step": 11419 + }, + { + "epoch": 0.9731572219855135, + "grad_norm": 14.5, + "learning_rate": 1.3113083689669325e-06, + "loss": 0.7577, + "step": 11420 + }, + { + "epoch": 0.9732424371538134, + "grad_norm": 11.125, + "learning_rate": 1.3111405350959417e-06, + "loss": 0.3497, + "step": 11421 + }, + { + "epoch": 0.9733276523221134, + "grad_norm": 16.5, + "learning_rate": 1.3109727001200801e-06, + "loss": 0.622, + "step": 11422 + }, + { + "epoch": 0.9734128674904133, + "grad_norm": 14.5, + "learning_rate": 1.3108048640423803e-06, + "loss": 0.5931, + "step": 11423 + }, + { + "epoch": 0.9734980826587133, + "grad_norm": 17.125, + "learning_rate": 1.3106370268658752e-06, + "loss": 0.8682, + "step": 11424 + }, + { + "epoch": 0.9735832978270132, + "grad_norm": 13.1875, + "learning_rate": 1.3104691885935986e-06, + "loss": 0.5822, + "step": 11425 + }, + { + "epoch": 0.9736685129953132, + "grad_norm": 27.25, + "learning_rate": 1.3103013492285827e-06, + "loss": 1.0992, + "step": 11426 + }, + { + "epoch": 0.9737537281636132, + "grad_norm": 13.5625, + "learning_rate": 1.3101335087738607e-06, + "loss": 0.7458, + "step": 11427 + }, + { + "epoch": 0.9738389433319131, + "grad_norm": 11.5625, + "learning_rate": 1.3099656672324659e-06, + "loss": 0.6046, + "step": 11428 + }, + { + "epoch": 0.9739241585002131, + "grad_norm": 15.4375, + "learning_rate": 1.309797824607431e-06, + "loss": 0.9732, + "step": 11429 + }, + { + "epoch": 0.974009373668513, + "grad_norm": 13.0625, + "learning_rate": 1.3096299809017896e-06, + "loss": 0.8161, + "step": 11430 + }, + { + "epoch": 0.974094588836813, + "grad_norm": 16.375, + "learning_rate": 1.3094621361185738e-06, + "loss": 0.5617, + "step": 11431 + }, + { + "epoch": 0.974179804005113, + "grad_norm": 12.4375, + "learning_rate": 1.309294290260818e-06, + "loss": 0.6842, + "step": 11432 + }, + { + "epoch": 0.9742650191734129, + "grad_norm": 15.8125, + "learning_rate": 1.3091264433315545e-06, + "loss": 0.8921, + "step": 11433 + }, + { + "epoch": 0.9743502343417129, + "grad_norm": 11.75, + "learning_rate": 1.308958595333817e-06, + "loss": 0.4538, + "step": 11434 + }, + { + "epoch": 0.9744354495100128, + "grad_norm": 13.25, + "learning_rate": 1.3087907462706384e-06, + "loss": 0.5017, + "step": 11435 + }, + { + "epoch": 0.9745206646783128, + "grad_norm": 13.5625, + "learning_rate": 1.3086228961450517e-06, + "loss": 0.928, + "step": 11436 + }, + { + "epoch": 0.9746058798466127, + "grad_norm": 18.875, + "learning_rate": 1.3084550449600903e-06, + "loss": 0.677, + "step": 11437 + }, + { + "epoch": 0.9746910950149127, + "grad_norm": 16.0, + "learning_rate": 1.3082871927187877e-06, + "loss": 0.6583, + "step": 11438 + }, + { + "epoch": 0.9747763101832126, + "grad_norm": 13.8125, + "learning_rate": 1.3081193394241765e-06, + "loss": 0.7183, + "step": 11439 + }, + { + "epoch": 0.9748615253515126, + "grad_norm": 16.5, + "learning_rate": 1.3079514850792906e-06, + "loss": 0.9364, + "step": 11440 + }, + { + "epoch": 0.9749467405198126, + "grad_norm": 11.5, + "learning_rate": 1.307783629687163e-06, + "loss": 0.369, + "step": 11441 + }, + { + "epoch": 0.9750319556881125, + "grad_norm": 15.0, + "learning_rate": 1.3076157732508271e-06, + "loss": 0.5203, + "step": 11442 + }, + { + "epoch": 0.9751171708564125, + "grad_norm": 21.625, + "learning_rate": 1.3074479157733165e-06, + "loss": 0.7108, + "step": 11443 + }, + { + "epoch": 0.9752023860247124, + "grad_norm": 14.0, + "learning_rate": 1.3072800572576638e-06, + "loss": 0.8176, + "step": 11444 + }, + { + "epoch": 0.9752876011930124, + "grad_norm": 18.75, + "learning_rate": 1.3071121977069035e-06, + "loss": 0.8551, + "step": 11445 + }, + { + "epoch": 0.9753728163613123, + "grad_norm": 11.5625, + "learning_rate": 1.306944337124068e-06, + "loss": 0.492, + "step": 11446 + }, + { + "epoch": 0.9754580315296123, + "grad_norm": 11.625, + "learning_rate": 1.3067764755121911e-06, + "loss": 0.4828, + "step": 11447 + }, + { + "epoch": 0.9755432466979123, + "grad_norm": 13.0, + "learning_rate": 1.3066086128743062e-06, + "loss": 0.5425, + "step": 11448 + }, + { + "epoch": 0.9756284618662122, + "grad_norm": 16.375, + "learning_rate": 1.3064407492134468e-06, + "loss": 0.8659, + "step": 11449 + }, + { + "epoch": 0.9757136770345122, + "grad_norm": 12.3125, + "learning_rate": 1.306272884532646e-06, + "loss": 0.4714, + "step": 11450 + }, + { + "epoch": 0.9757988922028121, + "grad_norm": 16.125, + "learning_rate": 1.3061050188349382e-06, + "loss": 0.7596, + "step": 11451 + }, + { + "epoch": 0.9758841073711121, + "grad_norm": 14.0, + "learning_rate": 1.3059371521233557e-06, + "loss": 0.8078, + "step": 11452 + }, + { + "epoch": 0.975969322539412, + "grad_norm": 11.3125, + "learning_rate": 1.3057692844009331e-06, + "loss": 0.4329, + "step": 11453 + }, + { + "epoch": 0.976054537707712, + "grad_norm": 14.0, + "learning_rate": 1.3056014156707031e-06, + "loss": 0.8311, + "step": 11454 + }, + { + "epoch": 0.976139752876012, + "grad_norm": 14.875, + "learning_rate": 1.3054335459356998e-06, + "loss": 0.486, + "step": 11455 + }, + { + "epoch": 0.9762249680443119, + "grad_norm": 13.625, + "learning_rate": 1.3052656751989565e-06, + "loss": 0.5052, + "step": 11456 + }, + { + "epoch": 0.9763101832126119, + "grad_norm": 22.125, + "learning_rate": 1.305097803463507e-06, + "loss": 0.9126, + "step": 11457 + }, + { + "epoch": 0.9763953983809118, + "grad_norm": 13.6875, + "learning_rate": 1.3049299307323842e-06, + "loss": 0.7523, + "step": 11458 + }, + { + "epoch": 0.9764806135492118, + "grad_norm": 23.125, + "learning_rate": 1.304762057008623e-06, + "loss": 0.6702, + "step": 11459 + }, + { + "epoch": 0.9765658287175117, + "grad_norm": 16.375, + "learning_rate": 1.3045941822952562e-06, + "loss": 0.7718, + "step": 11460 + }, + { + "epoch": 0.9766510438858117, + "grad_norm": 14.375, + "learning_rate": 1.3044263065953173e-06, + "loss": 0.8307, + "step": 11461 + }, + { + "epoch": 0.9767362590541117, + "grad_norm": 15.6875, + "learning_rate": 1.3042584299118407e-06, + "loss": 0.5395, + "step": 11462 + }, + { + "epoch": 0.9768214742224116, + "grad_norm": 11.3125, + "learning_rate": 1.3040905522478598e-06, + "loss": 0.4624, + "step": 11463 + }, + { + "epoch": 0.9769066893907116, + "grad_norm": 15.0625, + "learning_rate": 1.303922673606408e-06, + "loss": 0.7776, + "step": 11464 + }, + { + "epoch": 0.9769919045590115, + "grad_norm": 13.3125, + "learning_rate": 1.303754793990519e-06, + "loss": 0.6031, + "step": 11465 + }, + { + "epoch": 0.9770771197273115, + "grad_norm": 20.0, + "learning_rate": 1.3035869134032272e-06, + "loss": 1.0215, + "step": 11466 + }, + { + "epoch": 0.9771623348956114, + "grad_norm": 20.375, + "learning_rate": 1.303419031847566e-06, + "loss": 0.8326, + "step": 11467 + }, + { + "epoch": 0.9772475500639114, + "grad_norm": 16.75, + "learning_rate": 1.303251149326569e-06, + "loss": 0.9276, + "step": 11468 + }, + { + "epoch": 0.9773327652322114, + "grad_norm": 14.375, + "learning_rate": 1.3030832658432702e-06, + "loss": 0.7195, + "step": 11469 + }, + { + "epoch": 0.9774179804005113, + "grad_norm": 14.0, + "learning_rate": 1.3029153814007034e-06, + "loss": 0.6745, + "step": 11470 + }, + { + "epoch": 0.9775031955688113, + "grad_norm": 10.375, + "learning_rate": 1.3027474960019026e-06, + "loss": 0.3661, + "step": 11471 + }, + { + "epoch": 0.9775884107371112, + "grad_norm": 15.1875, + "learning_rate": 1.3025796096499016e-06, + "loss": 0.5098, + "step": 11472 + }, + { + "epoch": 0.9776736259054112, + "grad_norm": 14.9375, + "learning_rate": 1.3024117223477337e-06, + "loss": 0.7546, + "step": 11473 + }, + { + "epoch": 0.9777588410737111, + "grad_norm": 19.375, + "learning_rate": 1.3022438340984337e-06, + "loss": 0.9314, + "step": 11474 + }, + { + "epoch": 0.9778440562420111, + "grad_norm": 13.3125, + "learning_rate": 1.302075944905035e-06, + "loss": 0.6504, + "step": 11475 + }, + { + "epoch": 0.977929271410311, + "grad_norm": 14.6875, + "learning_rate": 1.3019080547705713e-06, + "loss": 0.478, + "step": 11476 + }, + { + "epoch": 0.978014486578611, + "grad_norm": 13.75, + "learning_rate": 1.3017401636980772e-06, + "loss": 0.4074, + "step": 11477 + }, + { + "epoch": 0.978099701746911, + "grad_norm": 15.75, + "learning_rate": 1.3015722716905862e-06, + "loss": 1.022, + "step": 11478 + }, + { + "epoch": 0.9781849169152109, + "grad_norm": 11.3125, + "learning_rate": 1.3014043787511327e-06, + "loss": 0.4357, + "step": 11479 + }, + { + "epoch": 0.9782701320835109, + "grad_norm": 13.5, + "learning_rate": 1.3012364848827497e-06, + "loss": 0.9168, + "step": 11480 + }, + { + "epoch": 0.9783553472518108, + "grad_norm": 15.125, + "learning_rate": 1.3010685900884723e-06, + "loss": 0.853, + "step": 11481 + }, + { + "epoch": 0.9784405624201108, + "grad_norm": 26.375, + "learning_rate": 1.300900694371334e-06, + "loss": 0.9446, + "step": 11482 + }, + { + "epoch": 0.9785257775884108, + "grad_norm": 18.625, + "learning_rate": 1.3007327977343695e-06, + "loss": 0.9474, + "step": 11483 + }, + { + "epoch": 0.9786109927567107, + "grad_norm": 18.625, + "learning_rate": 1.3005649001806118e-06, + "loss": 0.8374, + "step": 11484 + }, + { + "epoch": 0.9786962079250107, + "grad_norm": 14.5625, + "learning_rate": 1.3003970017130958e-06, + "loss": 0.4643, + "step": 11485 + }, + { + "epoch": 0.9787814230933106, + "grad_norm": 17.75, + "learning_rate": 1.3002291023348545e-06, + "loss": 0.6401, + "step": 11486 + }, + { + "epoch": 0.9788666382616106, + "grad_norm": 11.875, + "learning_rate": 1.300061202048924e-06, + "loss": 0.5251, + "step": 11487 + }, + { + "epoch": 0.9789518534299105, + "grad_norm": 20.0, + "learning_rate": 1.2998933008583365e-06, + "loss": 1.1831, + "step": 11488 + }, + { + "epoch": 0.9790370685982105, + "grad_norm": 14.5625, + "learning_rate": 1.2997253987661272e-06, + "loss": 0.9766, + "step": 11489 + }, + { + "epoch": 0.9791222837665104, + "grad_norm": 22.375, + "learning_rate": 1.2995574957753294e-06, + "loss": 1.1832, + "step": 11490 + }, + { + "epoch": 0.9792074989348104, + "grad_norm": 18.375, + "learning_rate": 1.2993895918889783e-06, + "loss": 0.8554, + "step": 11491 + }, + { + "epoch": 0.9792927141031104, + "grad_norm": 16.25, + "learning_rate": 1.2992216871101077e-06, + "loss": 0.9373, + "step": 11492 + }, + { + "epoch": 0.9793779292714103, + "grad_norm": 13.125, + "learning_rate": 1.2990537814417513e-06, + "loss": 0.8313, + "step": 11493 + }, + { + "epoch": 0.9794631444397103, + "grad_norm": 14.0, + "learning_rate": 1.298885874886944e-06, + "loss": 0.8489, + "step": 11494 + }, + { + "epoch": 0.9795483596080102, + "grad_norm": 15.375, + "learning_rate": 1.29871796744872e-06, + "loss": 0.978, + "step": 11495 + }, + { + "epoch": 0.9796335747763102, + "grad_norm": 13.5625, + "learning_rate": 1.298550059130113e-06, + "loss": 0.6944, + "step": 11496 + }, + { + "epoch": 0.9797187899446101, + "grad_norm": 15.75, + "learning_rate": 1.2983821499341576e-06, + "loss": 1.0332, + "step": 11497 + }, + { + "epoch": 0.9798040051129101, + "grad_norm": 23.875, + "learning_rate": 1.298214239863888e-06, + "loss": 1.2074, + "step": 11498 + }, + { + "epoch": 0.9798892202812101, + "grad_norm": 38.75, + "learning_rate": 1.298046328922339e-06, + "loss": 1.1794, + "step": 11499 + }, + { + "epoch": 0.97997443544951, + "grad_norm": 11.0, + "learning_rate": 1.2978784171125445e-06, + "loss": 0.5342, + "step": 11500 + }, + { + "epoch": 0.98005965061781, + "grad_norm": 17.875, + "learning_rate": 1.2977105044375385e-06, + "loss": 0.836, + "step": 11501 + }, + { + "epoch": 0.9801448657861099, + "grad_norm": 16.375, + "learning_rate": 1.2975425909003558e-06, + "loss": 0.6023, + "step": 11502 + }, + { + "epoch": 0.9802300809544099, + "grad_norm": 13.375, + "learning_rate": 1.297374676504031e-06, + "loss": 0.4707, + "step": 11503 + }, + { + "epoch": 0.9803152961227098, + "grad_norm": 12.0, + "learning_rate": 1.297206761251598e-06, + "loss": 0.471, + "step": 11504 + }, + { + "epoch": 0.9804005112910098, + "grad_norm": 16.25, + "learning_rate": 1.2970388451460912e-06, + "loss": 0.6333, + "step": 11505 + }, + { + "epoch": 0.9804857264593098, + "grad_norm": 14.1875, + "learning_rate": 1.296870928190545e-06, + "loss": 0.7306, + "step": 11506 + }, + { + "epoch": 0.9805709416276097, + "grad_norm": 19.25, + "learning_rate": 1.2967030103879946e-06, + "loss": 1.1319, + "step": 11507 + }, + { + "epoch": 0.9806561567959097, + "grad_norm": 13.6875, + "learning_rate": 1.2965350917414735e-06, + "loss": 0.795, + "step": 11508 + }, + { + "epoch": 0.9807413719642096, + "grad_norm": 14.8125, + "learning_rate": 1.2963671722540164e-06, + "loss": 0.8793, + "step": 11509 + }, + { + "epoch": 0.9808265871325096, + "grad_norm": 11.875, + "learning_rate": 1.2961992519286578e-06, + "loss": 0.3831, + "step": 11510 + }, + { + "epoch": 0.9809118023008095, + "grad_norm": 14.0, + "learning_rate": 1.2960313307684324e-06, + "loss": 0.8376, + "step": 11511 + }, + { + "epoch": 0.9809970174691095, + "grad_norm": 16.125, + "learning_rate": 1.2958634087763747e-06, + "loss": 0.8669, + "step": 11512 + }, + { + "epoch": 0.9810822326374095, + "grad_norm": 17.5, + "learning_rate": 1.2956954859555188e-06, + "loss": 0.8526, + "step": 11513 + }, + { + "epoch": 0.9811674478057094, + "grad_norm": 14.5, + "learning_rate": 1.2955275623088995e-06, + "loss": 0.6765, + "step": 11514 + }, + { + "epoch": 0.9812526629740094, + "grad_norm": 16.875, + "learning_rate": 1.2953596378395517e-06, + "loss": 1.0282, + "step": 11515 + }, + { + "epoch": 0.9813378781423093, + "grad_norm": 19.0, + "learning_rate": 1.2951917125505095e-06, + "loss": 0.6992, + "step": 11516 + }, + { + "epoch": 0.9814230933106093, + "grad_norm": 15.625, + "learning_rate": 1.2950237864448077e-06, + "loss": 0.7495, + "step": 11517 + }, + { + "epoch": 0.9815083084789092, + "grad_norm": 13.0, + "learning_rate": 1.2948558595254803e-06, + "loss": 0.7088, + "step": 11518 + }, + { + "epoch": 0.9815935236472092, + "grad_norm": 12.8125, + "learning_rate": 1.294687931795563e-06, + "loss": 0.5023, + "step": 11519 + }, + { + "epoch": 0.9816787388155092, + "grad_norm": 15.0, + "learning_rate": 1.2945200032580898e-06, + "loss": 0.812, + "step": 11520 + }, + { + "epoch": 0.9817639539838091, + "grad_norm": 16.375, + "learning_rate": 1.2943520739160953e-06, + "loss": 0.9937, + "step": 11521 + }, + { + "epoch": 0.9818491691521091, + "grad_norm": 18.5, + "learning_rate": 1.294184143772614e-06, + "loss": 1.0094, + "step": 11522 + }, + { + "epoch": 0.981934384320409, + "grad_norm": 13.5625, + "learning_rate": 1.294016212830681e-06, + "loss": 0.5905, + "step": 11523 + }, + { + "epoch": 0.982019599488709, + "grad_norm": 14.5, + "learning_rate": 1.2938482810933307e-06, + "loss": 0.8264, + "step": 11524 + }, + { + "epoch": 0.9821048146570089, + "grad_norm": 11.25, + "learning_rate": 1.293680348563598e-06, + "loss": 0.3662, + "step": 11525 + }, + { + "epoch": 0.9821900298253089, + "grad_norm": 16.375, + "learning_rate": 1.2935124152445173e-06, + "loss": 0.5752, + "step": 11526 + }, + { + "epoch": 0.9822752449936089, + "grad_norm": 13.9375, + "learning_rate": 1.2933444811391235e-06, + "loss": 0.698, + "step": 11527 + }, + { + "epoch": 0.9823604601619088, + "grad_norm": 12.25, + "learning_rate": 1.2931765462504513e-06, + "loss": 0.5559, + "step": 11528 + }, + { + "epoch": 0.9824456753302088, + "grad_norm": 17.375, + "learning_rate": 1.2930086105815354e-06, + "loss": 0.9016, + "step": 11529 + }, + { + "epoch": 0.9825308904985087, + "grad_norm": 13.25, + "learning_rate": 1.292840674135411e-06, + "loss": 0.4944, + "step": 11530 + }, + { + "epoch": 0.9826161056668087, + "grad_norm": 12.4375, + "learning_rate": 1.292672736915112e-06, + "loss": 0.6946, + "step": 11531 + }, + { + "epoch": 0.9827013208351086, + "grad_norm": 11.875, + "learning_rate": 1.2925047989236748e-06, + "loss": 0.4064, + "step": 11532 + }, + { + "epoch": 0.9827865360034086, + "grad_norm": 13.9375, + "learning_rate": 1.292336860164132e-06, + "loss": 0.6902, + "step": 11533 + }, + { + "epoch": 0.9828717511717086, + "grad_norm": 17.0, + "learning_rate": 1.2921689206395203e-06, + "loss": 0.7277, + "step": 11534 + }, + { + "epoch": 0.9829569663400085, + "grad_norm": 20.75, + "learning_rate": 1.2920009803528733e-06, + "loss": 1.0278, + "step": 11535 + }, + { + "epoch": 0.9830421815083085, + "grad_norm": 20.875, + "learning_rate": 1.2918330393072269e-06, + "loss": 1.1821, + "step": 11536 + }, + { + "epoch": 0.9831273966766084, + "grad_norm": 11.25, + "learning_rate": 1.2916650975056147e-06, + "loss": 0.4462, + "step": 11537 + }, + { + "epoch": 0.9832126118449084, + "grad_norm": 18.0, + "learning_rate": 1.291497154951073e-06, + "loss": 1.034, + "step": 11538 + }, + { + "epoch": 0.9832978270132083, + "grad_norm": 14.5, + "learning_rate": 1.2913292116466353e-06, + "loss": 0.6812, + "step": 11539 + }, + { + "epoch": 0.9833830421815083, + "grad_norm": 10.3125, + "learning_rate": 1.2911612675953378e-06, + "loss": 0.3512, + "step": 11540 + }, + { + "epoch": 0.9834682573498083, + "grad_norm": 12.5625, + "learning_rate": 1.2909933228002144e-06, + "loss": 0.5021, + "step": 11541 + }, + { + "epoch": 0.9835534725181082, + "grad_norm": 12.9375, + "learning_rate": 1.2908253772643006e-06, + "loss": 0.4651, + "step": 11542 + }, + { + "epoch": 0.9836386876864082, + "grad_norm": 13.625, + "learning_rate": 1.2906574309906312e-06, + "loss": 0.6579, + "step": 11543 + }, + { + "epoch": 0.9837239028547081, + "grad_norm": 17.0, + "learning_rate": 1.290489483982241e-06, + "loss": 0.7014, + "step": 11544 + }, + { + "epoch": 0.9838091180230081, + "grad_norm": 19.625, + "learning_rate": 1.2903215362421654e-06, + "loss": 1.1158, + "step": 11545 + }, + { + "epoch": 0.983894333191308, + "grad_norm": 15.5625, + "learning_rate": 1.2901535877734385e-06, + "loss": 0.8292, + "step": 11546 + }, + { + "epoch": 0.983979548359608, + "grad_norm": 16.375, + "learning_rate": 1.2899856385790964e-06, + "loss": 0.6922, + "step": 11547 + }, + { + "epoch": 0.984064763527908, + "grad_norm": 18.75, + "learning_rate": 1.2898176886621736e-06, + "loss": 1.3498, + "step": 11548 + }, + { + "epoch": 0.9841499786962079, + "grad_norm": 19.125, + "learning_rate": 1.289649738025705e-06, + "loss": 0.9013, + "step": 11549 + }, + { + "epoch": 0.9842351938645079, + "grad_norm": 17.75, + "learning_rate": 1.2894817866727257e-06, + "loss": 0.8277, + "step": 11550 + }, + { + "epoch": 0.9843204090328078, + "grad_norm": 13.375, + "learning_rate": 1.289313834606271e-06, + "loss": 0.5901, + "step": 11551 + }, + { + "epoch": 0.9844056242011078, + "grad_norm": 17.5, + "learning_rate": 1.2891458818293757e-06, + "loss": 0.7029, + "step": 11552 + }, + { + "epoch": 0.9844908393694077, + "grad_norm": 18.5, + "learning_rate": 1.288977928345075e-06, + "loss": 1.0259, + "step": 11553 + }, + { + "epoch": 0.9845760545377077, + "grad_norm": 14.875, + "learning_rate": 1.2888099741564036e-06, + "loss": 0.6933, + "step": 11554 + }, + { + "epoch": 0.9846612697060076, + "grad_norm": 17.625, + "learning_rate": 1.2886420192663973e-06, + "loss": 1.0109, + "step": 11555 + }, + { + "epoch": 0.9847464848743076, + "grad_norm": 19.0, + "learning_rate": 1.288474063678091e-06, + "loss": 0.9019, + "step": 11556 + }, + { + "epoch": 0.9848317000426076, + "grad_norm": 16.875, + "learning_rate": 1.2883061073945194e-06, + "loss": 0.7381, + "step": 11557 + }, + { + "epoch": 0.9849169152109075, + "grad_norm": 29.125, + "learning_rate": 1.2881381504187177e-06, + "loss": 1.1664, + "step": 11558 + }, + { + "epoch": 0.9850021303792075, + "grad_norm": 12.25, + "learning_rate": 1.2879701927537218e-06, + "loss": 0.617, + "step": 11559 + }, + { + "epoch": 0.9850873455475074, + "grad_norm": 15.3125, + "learning_rate": 1.287802234402566e-06, + "loss": 0.6309, + "step": 11560 + }, + { + "epoch": 0.9851725607158074, + "grad_norm": 11.25, + "learning_rate": 1.2876342753682856e-06, + "loss": 0.299, + "step": 11561 + }, + { + "epoch": 0.9852577758841073, + "grad_norm": 14.75, + "learning_rate": 1.2874663156539163e-06, + "loss": 0.4828, + "step": 11562 + }, + { + "epoch": 0.9853429910524073, + "grad_norm": 12.9375, + "learning_rate": 1.2872983552624932e-06, + "loss": 0.2755, + "step": 11563 + }, + { + "epoch": 0.9854282062207073, + "grad_norm": 24.125, + "learning_rate": 1.287130394197051e-06, + "loss": 1.0892, + "step": 11564 + }, + { + "epoch": 0.9855134213890072, + "grad_norm": 15.75, + "learning_rate": 1.2869624324606252e-06, + "loss": 0.3957, + "step": 11565 + }, + { + "epoch": 0.9855986365573072, + "grad_norm": 12.0, + "learning_rate": 1.2867944700562512e-06, + "loss": 0.58, + "step": 11566 + }, + { + "epoch": 0.9856838517256071, + "grad_norm": 18.0, + "learning_rate": 1.286626506986964e-06, + "loss": 0.8879, + "step": 11567 + }, + { + "epoch": 0.9857690668939071, + "grad_norm": 15.8125, + "learning_rate": 1.2864585432557995e-06, + "loss": 0.8339, + "step": 11568 + }, + { + "epoch": 0.985854282062207, + "grad_norm": 13.3125, + "learning_rate": 1.2862905788657918e-06, + "loss": 0.5279, + "step": 11569 + }, + { + "epoch": 0.985939497230507, + "grad_norm": 54.25, + "learning_rate": 1.2861226138199773e-06, + "loss": 0.5374, + "step": 11570 + }, + { + "epoch": 0.986024712398807, + "grad_norm": 14.4375, + "learning_rate": 1.2859546481213905e-06, + "loss": 0.7737, + "step": 11571 + }, + { + "epoch": 0.9861099275671069, + "grad_norm": 16.625, + "learning_rate": 1.2857866817730674e-06, + "loss": 0.7286, + "step": 11572 + }, + { + "epoch": 0.9861951427354069, + "grad_norm": 15.0625, + "learning_rate": 1.2856187147780427e-06, + "loss": 0.8106, + "step": 11573 + }, + { + "epoch": 0.9862803579037068, + "grad_norm": 16.125, + "learning_rate": 1.2854507471393521e-06, + "loss": 0.7582, + "step": 11574 + }, + { + "epoch": 0.9863655730720068, + "grad_norm": 14.75, + "learning_rate": 1.2852827788600309e-06, + "loss": 0.5743, + "step": 11575 + }, + { + "epoch": 0.9864507882403067, + "grad_norm": 17.625, + "learning_rate": 1.2851148099431142e-06, + "loss": 0.6944, + "step": 11576 + }, + { + "epoch": 0.9865360034086067, + "grad_norm": 10.75, + "learning_rate": 1.284946840391638e-06, + "loss": 0.4466, + "step": 11577 + }, + { + "epoch": 0.9866212185769067, + "grad_norm": 13.1875, + "learning_rate": 1.2847788702086369e-06, + "loss": 0.5045, + "step": 11578 + }, + { + "epoch": 0.9867064337452066, + "grad_norm": 17.875, + "learning_rate": 1.2846108993971464e-06, + "loss": 0.5855, + "step": 11579 + }, + { + "epoch": 0.9867916489135066, + "grad_norm": 14.4375, + "learning_rate": 1.2844429279602028e-06, + "loss": 0.913, + "step": 11580 + }, + { + "epoch": 0.9868768640818065, + "grad_norm": 16.75, + "learning_rate": 1.2842749559008405e-06, + "loss": 0.6198, + "step": 11581 + }, + { + "epoch": 0.9869620792501065, + "grad_norm": 15.25, + "learning_rate": 1.2841069832220952e-06, + "loss": 0.7165, + "step": 11582 + }, + { + "epoch": 0.9870472944184064, + "grad_norm": 12.3125, + "learning_rate": 1.2839390099270024e-06, + "loss": 0.624, + "step": 11583 + }, + { + "epoch": 0.9871325095867064, + "grad_norm": 18.125, + "learning_rate": 1.2837710360185975e-06, + "loss": 1.0329, + "step": 11584 + }, + { + "epoch": 0.9872177247550064, + "grad_norm": 14.1875, + "learning_rate": 1.2836030614999168e-06, + "loss": 0.7766, + "step": 11585 + }, + { + "epoch": 0.9873029399233063, + "grad_norm": 14.25, + "learning_rate": 1.2834350863739941e-06, + "loss": 0.319, + "step": 11586 + }, + { + "epoch": 0.9873881550916063, + "grad_norm": 18.5, + "learning_rate": 1.283267110643866e-06, + "loss": 0.7883, + "step": 11587 + }, + { + "epoch": 0.9874733702599062, + "grad_norm": 25.75, + "learning_rate": 1.2830991343125675e-06, + "loss": 0.6929, + "step": 11588 + }, + { + "epoch": 0.9875585854282062, + "grad_norm": 15.375, + "learning_rate": 1.2829311573831351e-06, + "loss": 0.7721, + "step": 11589 + }, + { + "epoch": 0.9876438005965061, + "grad_norm": 13.25, + "learning_rate": 1.2827631798586032e-06, + "loss": 0.4744, + "step": 11590 + }, + { + "epoch": 0.9877290157648061, + "grad_norm": 15.6875, + "learning_rate": 1.2825952017420077e-06, + "loss": 0.911, + "step": 11591 + }, + { + "epoch": 0.987814230933106, + "grad_norm": 12.125, + "learning_rate": 1.2824272230363838e-06, + "loss": 0.5891, + "step": 11592 + }, + { + "epoch": 0.987899446101406, + "grad_norm": 14.5625, + "learning_rate": 1.2822592437447678e-06, + "loss": 0.5944, + "step": 11593 + }, + { + "epoch": 0.987984661269706, + "grad_norm": 17.5, + "learning_rate": 1.2820912638701946e-06, + "loss": 0.986, + "step": 11594 + }, + { + "epoch": 0.9880698764380059, + "grad_norm": 15.125, + "learning_rate": 1.2819232834157003e-06, + "loss": 0.8919, + "step": 11595 + }, + { + "epoch": 0.9881550916063059, + "grad_norm": 17.375, + "learning_rate": 1.2817553023843196e-06, + "loss": 0.9322, + "step": 11596 + }, + { + "epoch": 0.9882403067746058, + "grad_norm": 12.875, + "learning_rate": 1.2815873207790891e-06, + "loss": 0.4923, + "step": 11597 + }, + { + "epoch": 0.9883255219429058, + "grad_norm": 21.125, + "learning_rate": 1.2814193386030438e-06, + "loss": 0.9952, + "step": 11598 + }, + { + "epoch": 0.9884107371112058, + "grad_norm": 16.375, + "learning_rate": 1.2812513558592194e-06, + "loss": 1.0879, + "step": 11599 + }, + { + "epoch": 0.9884959522795057, + "grad_norm": 20.125, + "learning_rate": 1.2810833725506516e-06, + "loss": 0.7975, + "step": 11600 + }, + { + "epoch": 0.9885811674478057, + "grad_norm": 23.875, + "learning_rate": 1.2809153886803763e-06, + "loss": 1.2124, + "step": 11601 + }, + { + "epoch": 0.9886663826161056, + "grad_norm": 12.5, + "learning_rate": 1.2807474042514287e-06, + "loss": 0.2909, + "step": 11602 + }, + { + "epoch": 0.9887515977844056, + "grad_norm": 17.25, + "learning_rate": 1.2805794192668442e-06, + "loss": 0.5975, + "step": 11603 + }, + { + "epoch": 0.9888368129527055, + "grad_norm": 15.375, + "learning_rate": 1.280411433729659e-06, + "loss": 0.7363, + "step": 11604 + }, + { + "epoch": 0.9889220281210055, + "grad_norm": 17.625, + "learning_rate": 1.280243447642909e-06, + "loss": 0.9198, + "step": 11605 + }, + { + "epoch": 0.9890072432893054, + "grad_norm": 15.9375, + "learning_rate": 1.2800754610096293e-06, + "loss": 0.5472, + "step": 11606 + }, + { + "epoch": 0.9890924584576054, + "grad_norm": 14.5, + "learning_rate": 1.2799074738328556e-06, + "loss": 0.9422, + "step": 11607 + }, + { + "epoch": 0.9891776736259054, + "grad_norm": 14.6875, + "learning_rate": 1.279739486115624e-06, + "loss": 0.758, + "step": 11608 + }, + { + "epoch": 0.9892628887942053, + "grad_norm": 12.25, + "learning_rate": 1.2795714978609699e-06, + "loss": 0.5117, + "step": 11609 + }, + { + "epoch": 0.9893481039625053, + "grad_norm": 12.9375, + "learning_rate": 1.2794035090719293e-06, + "loss": 0.4344, + "step": 11610 + }, + { + "epoch": 0.9894333191308052, + "grad_norm": 20.75, + "learning_rate": 1.2792355197515371e-06, + "loss": 0.8544, + "step": 11611 + }, + { + "epoch": 0.9895185342991052, + "grad_norm": 18.75, + "learning_rate": 1.2790675299028302e-06, + "loss": 0.9502, + "step": 11612 + }, + { + "epoch": 0.9896037494674051, + "grad_norm": 13.3125, + "learning_rate": 1.2788995395288436e-06, + "loss": 0.3178, + "step": 11613 + }, + { + "epoch": 0.9896889646357051, + "grad_norm": 13.9375, + "learning_rate": 1.2787315486326131e-06, + "loss": 0.653, + "step": 11614 + }, + { + "epoch": 0.9897741798040051, + "grad_norm": 19.125, + "learning_rate": 1.278563557217175e-06, + "loss": 0.8924, + "step": 11615 + }, + { + "epoch": 0.989859394972305, + "grad_norm": 14.4375, + "learning_rate": 1.2783955652855645e-06, + "loss": 0.6874, + "step": 11616 + }, + { + "epoch": 0.989944610140605, + "grad_norm": 16.75, + "learning_rate": 1.2782275728408178e-06, + "loss": 0.9195, + "step": 11617 + }, + { + "epoch": 0.9900298253089049, + "grad_norm": 12.625, + "learning_rate": 1.27805957988597e-06, + "loss": 0.3683, + "step": 11618 + }, + { + "epoch": 0.9901150404772049, + "grad_norm": 14.875, + "learning_rate": 1.277891586424058e-06, + "loss": 0.895, + "step": 11619 + }, + { + "epoch": 0.9902002556455048, + "grad_norm": 11.5625, + "learning_rate": 1.2777235924581164e-06, + "loss": 0.4849, + "step": 11620 + }, + { + "epoch": 0.9902854708138048, + "grad_norm": 13.25, + "learning_rate": 1.277555597991182e-06, + "loss": 0.5929, + "step": 11621 + }, + { + "epoch": 0.9903706859821048, + "grad_norm": 17.5, + "learning_rate": 1.2773876030262899e-06, + "loss": 0.846, + "step": 11622 + }, + { + "epoch": 0.9904559011504048, + "grad_norm": 14.1875, + "learning_rate": 1.2772196075664767e-06, + "loss": 0.773, + "step": 11623 + }, + { + "epoch": 0.9905411163187048, + "grad_norm": 15.375, + "learning_rate": 1.277051611614777e-06, + "loss": 0.9419, + "step": 11624 + }, + { + "epoch": 0.9906263314870047, + "grad_norm": 16.0, + "learning_rate": 1.2768836151742287e-06, + "loss": 0.9098, + "step": 11625 + }, + { + "epoch": 0.9907115466553047, + "grad_norm": 17.125, + "learning_rate": 1.2767156182478655e-06, + "loss": 0.92, + "step": 11626 + }, + { + "epoch": 0.9907967618236047, + "grad_norm": 20.5, + "learning_rate": 1.2765476208387247e-06, + "loss": 0.9697, + "step": 11627 + }, + { + "epoch": 0.9908819769919046, + "grad_norm": 13.8125, + "learning_rate": 1.2763796229498413e-06, + "loss": 0.782, + "step": 11628 + }, + { + "epoch": 0.9909671921602046, + "grad_norm": 14.6875, + "learning_rate": 1.2762116245842518e-06, + "loss": 0.7664, + "step": 11629 + }, + { + "epoch": 0.9910524073285045, + "grad_norm": 13.9375, + "learning_rate": 1.276043625744992e-06, + "loss": 0.64, + "step": 11630 + }, + { + "epoch": 0.9911376224968045, + "grad_norm": 23.75, + "learning_rate": 1.2758756264350978e-06, + "loss": 1.0513, + "step": 11631 + }, + { + "epoch": 0.9912228376651044, + "grad_norm": 18.75, + "learning_rate": 1.2757076266576042e-06, + "loss": 0.7448, + "step": 11632 + }, + { + "epoch": 0.9913080528334044, + "grad_norm": 31.5, + "learning_rate": 1.275539626415549e-06, + "loss": 1.2298, + "step": 11633 + }, + { + "epoch": 0.9913932680017044, + "grad_norm": 13.6875, + "learning_rate": 1.2753716257119663e-06, + "loss": 0.4897, + "step": 11634 + }, + { + "epoch": 0.9914784831700043, + "grad_norm": 21.25, + "learning_rate": 1.275203624549893e-06, + "loss": 0.9425, + "step": 11635 + }, + { + "epoch": 0.9915636983383043, + "grad_norm": 14.5, + "learning_rate": 1.275035622932365e-06, + "loss": 0.6666, + "step": 11636 + }, + { + "epoch": 0.9916489135066042, + "grad_norm": 15.6875, + "learning_rate": 1.2748676208624184e-06, + "loss": 1.1565, + "step": 11637 + }, + { + "epoch": 0.9917341286749042, + "grad_norm": 10.9375, + "learning_rate": 1.2746996183430885e-06, + "loss": 0.3345, + "step": 11638 + }, + { + "epoch": 0.9918193438432041, + "grad_norm": 18.5, + "learning_rate": 1.2745316153774115e-06, + "loss": 0.8908, + "step": 11639 + }, + { + "epoch": 0.9919045590115041, + "grad_norm": 16.25, + "learning_rate": 1.2743636119684238e-06, + "loss": 0.9383, + "step": 11640 + }, + { + "epoch": 0.991989774179804, + "grad_norm": 16.75, + "learning_rate": 1.2741956081191612e-06, + "loss": 0.9483, + "step": 11641 + }, + { + "epoch": 0.992074989348104, + "grad_norm": 20.125, + "learning_rate": 1.2740276038326595e-06, + "loss": 0.8267, + "step": 11642 + }, + { + "epoch": 0.992160204516404, + "grad_norm": 14.0625, + "learning_rate": 1.2738595991119548e-06, + "loss": 0.7569, + "step": 11643 + }, + { + "epoch": 0.9922454196847039, + "grad_norm": 12.8125, + "learning_rate": 1.2736915939600833e-06, + "loss": 0.5399, + "step": 11644 + }, + { + "epoch": 0.9923306348530039, + "grad_norm": 11.375, + "learning_rate": 1.2735235883800808e-06, + "loss": 0.4138, + "step": 11645 + }, + { + "epoch": 0.9924158500213038, + "grad_norm": 13.3125, + "learning_rate": 1.2733555823749832e-06, + "loss": 0.5622, + "step": 11646 + }, + { + "epoch": 0.9925010651896038, + "grad_norm": 14.625, + "learning_rate": 1.2731875759478272e-06, + "loss": 0.8255, + "step": 11647 + }, + { + "epoch": 0.9925862803579037, + "grad_norm": 16.375, + "learning_rate": 1.273019569101648e-06, + "loss": 0.4036, + "step": 11648 + }, + { + "epoch": 0.9926714955262037, + "grad_norm": 22.25, + "learning_rate": 1.2728515618394824e-06, + "loss": 0.6206, + "step": 11649 + }, + { + "epoch": 0.9927567106945037, + "grad_norm": 13.125, + "learning_rate": 1.2726835541643656e-06, + "loss": 0.6498, + "step": 11650 + }, + { + "epoch": 0.9928419258628036, + "grad_norm": 13.0625, + "learning_rate": 1.2725155460793345e-06, + "loss": 0.3856, + "step": 11651 + }, + { + "epoch": 0.9929271410311036, + "grad_norm": 13.875, + "learning_rate": 1.2723475375874244e-06, + "loss": 0.4691, + "step": 11652 + }, + { + "epoch": 0.9930123561994035, + "grad_norm": 25.5, + "learning_rate": 1.2721795286916724e-06, + "loss": 0.7642, + "step": 11653 + }, + { + "epoch": 0.9930975713677035, + "grad_norm": 14.125, + "learning_rate": 1.2720115193951138e-06, + "loss": 0.5637, + "step": 11654 + }, + { + "epoch": 0.9931827865360034, + "grad_norm": 21.25, + "learning_rate": 1.271843509700785e-06, + "loss": 1.1254, + "step": 11655 + }, + { + "epoch": 0.9932680017043034, + "grad_norm": 15.9375, + "learning_rate": 1.2716754996117217e-06, + "loss": 0.7872, + "step": 11656 + }, + { + "epoch": 0.9933532168726034, + "grad_norm": 14.8125, + "learning_rate": 1.2715074891309606e-06, + "loss": 0.8337, + "step": 11657 + }, + { + "epoch": 0.9934384320409033, + "grad_norm": 14.375, + "learning_rate": 1.2713394782615376e-06, + "loss": 0.6906, + "step": 11658 + }, + { + "epoch": 0.9935236472092033, + "grad_norm": 19.125, + "learning_rate": 1.2711714670064886e-06, + "loss": 0.6924, + "step": 11659 + }, + { + "epoch": 0.9936088623775032, + "grad_norm": 14.0625, + "learning_rate": 1.27100345536885e-06, + "loss": 0.4877, + "step": 11660 + }, + { + "epoch": 0.9936940775458032, + "grad_norm": 14.375, + "learning_rate": 1.270835443351658e-06, + "loss": 0.8702, + "step": 11661 + }, + { + "epoch": 0.9937792927141031, + "grad_norm": 13.875, + "learning_rate": 1.2706674309579483e-06, + "loss": 0.5319, + "step": 11662 + }, + { + "epoch": 0.9938645078824031, + "grad_norm": 14.75, + "learning_rate": 1.2704994181907577e-06, + "loss": 0.8544, + "step": 11663 + }, + { + "epoch": 0.9939497230507031, + "grad_norm": 9.5, + "learning_rate": 1.2703314050531213e-06, + "loss": 0.2873, + "step": 11664 + }, + { + "epoch": 0.994034938219003, + "grad_norm": 16.75, + "learning_rate": 1.2701633915480765e-06, + "loss": 0.8987, + "step": 11665 + }, + { + "epoch": 0.994120153387303, + "grad_norm": 16.375, + "learning_rate": 1.2699953776786589e-06, + "loss": 1.0721, + "step": 11666 + }, + { + "epoch": 0.9942053685556029, + "grad_norm": 13.5625, + "learning_rate": 1.2698273634479044e-06, + "loss": 0.9995, + "step": 11667 + }, + { + "epoch": 0.9942905837239029, + "grad_norm": 15.4375, + "learning_rate": 1.26965934885885e-06, + "loss": 1.0857, + "step": 11668 + }, + { + "epoch": 0.9943757988922028, + "grad_norm": 13.3125, + "learning_rate": 1.269491333914531e-06, + "loss": 0.7042, + "step": 11669 + }, + { + "epoch": 0.9944610140605028, + "grad_norm": 17.375, + "learning_rate": 1.2693233186179844e-06, + "loss": 1.1651, + "step": 11670 + }, + { + "epoch": 0.9945462292288028, + "grad_norm": 13.625, + "learning_rate": 1.2691553029722455e-06, + "loss": 0.5201, + "step": 11671 + }, + { + "epoch": 0.9946314443971027, + "grad_norm": 19.375, + "learning_rate": 1.2689872869803514e-06, + "loss": 0.8782, + "step": 11672 + }, + { + "epoch": 0.9947166595654027, + "grad_norm": 13.9375, + "learning_rate": 1.2688192706453375e-06, + "loss": 0.7475, + "step": 11673 + }, + { + "epoch": 0.9948018747337026, + "grad_norm": 14.0, + "learning_rate": 1.2686512539702412e-06, + "loss": 0.5983, + "step": 11674 + }, + { + "epoch": 0.9948870899020026, + "grad_norm": 18.375, + "learning_rate": 1.2684832369580974e-06, + "loss": 0.536, + "step": 11675 + }, + { + "epoch": 0.9949723050703025, + "grad_norm": 12.5, + "learning_rate": 1.268315219611943e-06, + "loss": 0.5895, + "step": 11676 + }, + { + "epoch": 0.9950575202386025, + "grad_norm": 14.0, + "learning_rate": 1.268147201934814e-06, + "loss": 0.7835, + "step": 11677 + }, + { + "epoch": 0.9951427354069025, + "grad_norm": 10.125, + "learning_rate": 1.2679791839297476e-06, + "loss": 0.3089, + "step": 11678 + }, + { + "epoch": 0.9952279505752024, + "grad_norm": 13.375, + "learning_rate": 1.2678111655997784e-06, + "loss": 0.4106, + "step": 11679 + }, + { + "epoch": 0.9953131657435024, + "grad_norm": 13.5, + "learning_rate": 1.267643146947944e-06, + "loss": 0.6568, + "step": 11680 + }, + { + "epoch": 0.9953983809118023, + "grad_norm": 13.375, + "learning_rate": 1.2674751279772797e-06, + "loss": 0.426, + "step": 11681 + }, + { + "epoch": 0.9954835960801023, + "grad_norm": 14.125, + "learning_rate": 1.2673071086908226e-06, + "loss": 0.7429, + "step": 11682 + }, + { + "epoch": 0.9955688112484022, + "grad_norm": 13.75, + "learning_rate": 1.2671390890916089e-06, + "loss": 0.5406, + "step": 11683 + }, + { + "epoch": 0.9956540264167022, + "grad_norm": 17.625, + "learning_rate": 1.266971069182674e-06, + "loss": 1.0032, + "step": 11684 + }, + { + "epoch": 0.9957392415850022, + "grad_norm": 15.625, + "learning_rate": 1.2668030489670551e-06, + "loss": 0.6928, + "step": 11685 + }, + { + "epoch": 0.9958244567533021, + "grad_norm": 13.875, + "learning_rate": 1.2666350284477885e-06, + "loss": 0.5017, + "step": 11686 + }, + { + "epoch": 0.9959096719216021, + "grad_norm": 16.375, + "learning_rate": 1.2664670076279101e-06, + "loss": 0.8655, + "step": 11687 + }, + { + "epoch": 0.995994887089902, + "grad_norm": 16.125, + "learning_rate": 1.2662989865104561e-06, + "loss": 0.7047, + "step": 11688 + }, + { + "epoch": 0.996080102258202, + "grad_norm": 12.375, + "learning_rate": 1.2661309650984632e-06, + "loss": 0.5351, + "step": 11689 + }, + { + "epoch": 0.9961653174265019, + "grad_norm": 19.375, + "learning_rate": 1.2659629433949674e-06, + "loss": 0.8598, + "step": 11690 + }, + { + "epoch": 0.9962505325948019, + "grad_norm": 10.75, + "learning_rate": 1.2657949214030055e-06, + "loss": 0.4937, + "step": 11691 + }, + { + "epoch": 0.9963357477631019, + "grad_norm": 13.1875, + "learning_rate": 1.265626899125613e-06, + "loss": 0.3935, + "step": 11692 + }, + { + "epoch": 0.9964209629314018, + "grad_norm": 14.5, + "learning_rate": 1.2654588765658268e-06, + "loss": 0.7739, + "step": 11693 + }, + { + "epoch": 0.9965061780997018, + "grad_norm": 19.75, + "learning_rate": 1.2652908537266833e-06, + "loss": 0.8091, + "step": 11694 + }, + { + "epoch": 0.9965913932680017, + "grad_norm": 9.875, + "learning_rate": 1.265122830611219e-06, + "loss": 0.3855, + "step": 11695 + }, + { + "epoch": 0.9966766084363017, + "grad_norm": 21.875, + "learning_rate": 1.2649548072224694e-06, + "loss": 1.1789, + "step": 11696 + }, + { + "epoch": 0.9967618236046016, + "grad_norm": 24.125, + "learning_rate": 1.2647867835634717e-06, + "loss": 0.8682, + "step": 11697 + }, + { + "epoch": 0.9968470387729016, + "grad_norm": 13.0625, + "learning_rate": 1.264618759637262e-06, + "loss": 0.5722, + "step": 11698 + }, + { + "epoch": 0.9969322539412016, + "grad_norm": 13.3125, + "learning_rate": 1.2644507354468762e-06, + "loss": 0.8484, + "step": 11699 + }, + { + "epoch": 0.9970174691095015, + "grad_norm": 15.0, + "learning_rate": 1.2642827109953515e-06, + "loss": 0.9121, + "step": 11700 + }, + { + "epoch": 0.9971026842778015, + "grad_norm": 14.875, + "learning_rate": 1.2641146862857237e-06, + "loss": 1.1333, + "step": 11701 + }, + { + "epoch": 0.9971878994461014, + "grad_norm": 11.5, + "learning_rate": 1.2639466613210293e-06, + "loss": 0.5709, + "step": 11702 + }, + { + "epoch": 0.9972731146144014, + "grad_norm": 12.0625, + "learning_rate": 1.2637786361043048e-06, + "loss": 0.6844, + "step": 11703 + }, + { + "epoch": 0.9973583297827013, + "grad_norm": 12.6875, + "learning_rate": 1.2636106106385864e-06, + "loss": 0.6668, + "step": 11704 + }, + { + "epoch": 0.9974435449510013, + "grad_norm": 19.25, + "learning_rate": 1.2634425849269104e-06, + "loss": 0.8497, + "step": 11705 + }, + { + "epoch": 0.9975287601193012, + "grad_norm": 25.25, + "learning_rate": 1.2632745589723139e-06, + "loss": 0.8703, + "step": 11706 + }, + { + "epoch": 0.9976139752876012, + "grad_norm": 11.875, + "learning_rate": 1.2631065327778323e-06, + "loss": 0.4848, + "step": 11707 + }, + { + "epoch": 0.9976991904559012, + "grad_norm": 16.375, + "learning_rate": 1.2629385063465027e-06, + "loss": 0.8481, + "step": 11708 + }, + { + "epoch": 0.9977844056242011, + "grad_norm": 13.5625, + "learning_rate": 1.262770479681361e-06, + "loss": 0.5407, + "step": 11709 + }, + { + "epoch": 0.9978696207925011, + "grad_norm": 14.375, + "learning_rate": 1.2626024527854443e-06, + "loss": 0.5804, + "step": 11710 + }, + { + "epoch": 0.997954835960801, + "grad_norm": 15.0625, + "learning_rate": 1.2624344256617886e-06, + "loss": 0.7516, + "step": 11711 + }, + { + "epoch": 0.998040051129101, + "grad_norm": 13.75, + "learning_rate": 1.26226639831343e-06, + "loss": 0.731, + "step": 11712 + }, + { + "epoch": 0.998125266297401, + "grad_norm": 20.25, + "learning_rate": 1.262098370743405e-06, + "loss": 0.8451, + "step": 11713 + }, + { + "epoch": 0.9982104814657009, + "grad_norm": 17.75, + "learning_rate": 1.2619303429547509e-06, + "loss": 0.8531, + "step": 11714 + }, + { + "epoch": 0.9982956966340009, + "grad_norm": 14.6875, + "learning_rate": 1.2617623149505035e-06, + "loss": 0.8063, + "step": 11715 + }, + { + "epoch": 0.9983809118023008, + "grad_norm": 14.6875, + "learning_rate": 1.2615942867336988e-06, + "loss": 0.5685, + "step": 11716 + }, + { + "epoch": 0.9984661269706008, + "grad_norm": 16.75, + "learning_rate": 1.261426258307374e-06, + "loss": 0.8099, + "step": 11717 + }, + { + "epoch": 0.9985513421389007, + "grad_norm": 13.75, + "learning_rate": 1.261258229674565e-06, + "loss": 0.7717, + "step": 11718 + }, + { + "epoch": 0.9986365573072007, + "grad_norm": 13.9375, + "learning_rate": 1.2610902008383086e-06, + "loss": 0.5017, + "step": 11719 + }, + { + "epoch": 0.9987217724755006, + "grad_norm": 16.875, + "learning_rate": 1.260922171801641e-06, + "loss": 0.5945, + "step": 11720 + }, + { + "epoch": 0.9988069876438006, + "grad_norm": 13.5625, + "learning_rate": 1.260754142567599e-06, + "loss": 0.7152, + "step": 11721 + }, + { + "epoch": 0.9988922028121006, + "grad_norm": 17.375, + "learning_rate": 1.2605861131392183e-06, + "loss": 1.0023, + "step": 11722 + }, + { + "epoch": 0.9989774179804005, + "grad_norm": 12.6875, + "learning_rate": 1.2604180835195369e-06, + "loss": 0.6268, + "step": 11723 + }, + { + "epoch": 0.9990626331487005, + "grad_norm": 16.125, + "learning_rate": 1.2602500537115892e-06, + "loss": 0.8231, + "step": 11724 + }, + { + "epoch": 0.9991478483170004, + "grad_norm": 17.25, + "learning_rate": 1.2600820237184132e-06, + "loss": 0.8764, + "step": 11725 + }, + { + "epoch": 0.9992330634853004, + "grad_norm": 11.3125, + "learning_rate": 1.2599139935430445e-06, + "loss": 0.5251, + "step": 11726 + }, + { + "epoch": 0.9993182786536003, + "grad_norm": 14.4375, + "learning_rate": 1.259745963188521e-06, + "loss": 0.8632, + "step": 11727 + }, + { + "epoch": 0.9994034938219003, + "grad_norm": 12.875, + "learning_rate": 1.2595779326578771e-06, + "loss": 0.6063, + "step": 11728 + }, + { + "epoch": 0.9994887089902003, + "grad_norm": 13.0625, + "learning_rate": 1.2594099019541508e-06, + "loss": 0.5859, + "step": 11729 + }, + { + "epoch": 0.9995739241585002, + "grad_norm": 11.25, + "learning_rate": 1.2592418710803778e-06, + "loss": 1.142, + "step": 11730 + }, + { + "epoch": 0.9996591393268002, + "grad_norm": 14.625, + "learning_rate": 1.2590738400395955e-06, + "loss": 0.8674, + "step": 11731 + }, + { + "epoch": 0.9997443544951001, + "grad_norm": 31.875, + "learning_rate": 1.258905808834839e-06, + "loss": 0.779, + "step": 11732 + }, + { + "epoch": 0.9998295696634001, + "grad_norm": 17.5, + "learning_rate": 1.2587377774691462e-06, + "loss": 0.5853, + "step": 11733 + }, + { + "epoch": 0.9999147848317, + "grad_norm": 19.875, + "learning_rate": 1.2585697459455526e-06, + "loss": 0.8478, + "step": 11734 + }, + { + "epoch": 1.0, + "grad_norm": 12.9375, + "learning_rate": 1.258401714267095e-06, + "loss": 0.5969, + "step": 11735 + }, + { + "epoch": 1.0, + "eval_loss": 0.9627697467803955, + "eval_runtime": 221.5463, + "eval_samples_per_second": 7.646, + "eval_steps_per_second": 7.646, + "step": 11735 + }, + { + "epoch": 1.0000852151683, + "grad_norm": 11.625, + "learning_rate": 1.2582336824368105e-06, + "loss": 0.5151, + "step": 11736 + }, + { + "epoch": 1.0001704303366, + "grad_norm": 13.3125, + "learning_rate": 1.2580656504577347e-06, + "loss": 0.5121, + "step": 11737 + }, + { + "epoch": 1.0002556455048999, + "grad_norm": 14.9375, + "learning_rate": 1.2578976183329044e-06, + "loss": 0.7862, + "step": 11738 + }, + { + "epoch": 1.0003408606731998, + "grad_norm": 15.8125, + "learning_rate": 1.2577295860653566e-06, + "loss": 0.638, + "step": 11739 + }, + { + "epoch": 1.0004260758414998, + "grad_norm": 11.6875, + "learning_rate": 1.257561553658127e-06, + "loss": 0.3454, + "step": 11740 + }, + { + "epoch": 1.0005112910097997, + "grad_norm": 14.5, + "learning_rate": 1.2573935211142526e-06, + "loss": 0.8697, + "step": 11741 + }, + { + "epoch": 1.0005965061780997, + "grad_norm": 22.5, + "learning_rate": 1.25722548843677e-06, + "loss": 0.7993, + "step": 11742 + }, + { + "epoch": 1.0006817213463997, + "grad_norm": 9.8125, + "learning_rate": 1.2570574556287154e-06, + "loss": 0.4092, + "step": 11743 + }, + { + "epoch": 1.0007669365146996, + "grad_norm": 13.25, + "learning_rate": 1.2568894226931257e-06, + "loss": 0.5428, + "step": 11744 + }, + { + "epoch": 1.0008521516829996, + "grad_norm": 11.25, + "learning_rate": 1.2567213896330368e-06, + "loss": 0.6371, + "step": 11745 + }, + { + "epoch": 1.0009373668512995, + "grad_norm": 13.375, + "learning_rate": 1.256553356451486e-06, + "loss": 0.5713, + "step": 11746 + }, + { + "epoch": 1.0010225820195995, + "grad_norm": 10.25, + "learning_rate": 1.2563853231515094e-06, + "loss": 0.3222, + "step": 11747 + }, + { + "epoch": 1.0011077971878994, + "grad_norm": 12.1875, + "learning_rate": 1.256217289736144e-06, + "loss": 0.3801, + "step": 11748 + }, + { + "epoch": 1.0011930123561994, + "grad_norm": 10.875, + "learning_rate": 1.256049256208425e-06, + "loss": 0.3724, + "step": 11749 + }, + { + "epoch": 1.0012782275244994, + "grad_norm": 13.0625, + "learning_rate": 1.2558812225713904e-06, + "loss": 0.3376, + "step": 11750 + }, + { + "epoch": 1.0013634426927993, + "grad_norm": 15.4375, + "learning_rate": 1.2557131888280762e-06, + "loss": 0.4657, + "step": 11751 + }, + { + "epoch": 1.0014486578610993, + "grad_norm": 12.125, + "learning_rate": 1.2555451549815187e-06, + "loss": 0.6529, + "step": 11752 + }, + { + "epoch": 1.0015338730293992, + "grad_norm": 17.375, + "learning_rate": 1.2553771210347548e-06, + "loss": 0.9263, + "step": 11753 + }, + { + "epoch": 1.0016190881976992, + "grad_norm": 10.5, + "learning_rate": 1.2552090869908206e-06, + "loss": 0.3706, + "step": 11754 + }, + { + "epoch": 1.0017043033659991, + "grad_norm": 11.3125, + "learning_rate": 1.2550410528527537e-06, + "loss": 0.4044, + "step": 11755 + }, + { + "epoch": 1.001789518534299, + "grad_norm": 10.5625, + "learning_rate": 1.2548730186235895e-06, + "loss": 0.4461, + "step": 11756 + }, + { + "epoch": 1.001874733702599, + "grad_norm": 16.75, + "learning_rate": 1.254704984306365e-06, + "loss": 0.6447, + "step": 11757 + }, + { + "epoch": 1.001959948870899, + "grad_norm": 12.375, + "learning_rate": 1.2545369499041163e-06, + "loss": 0.4562, + "step": 11758 + }, + { + "epoch": 1.002045164039199, + "grad_norm": 15.875, + "learning_rate": 1.254368915419881e-06, + "loss": 0.6444, + "step": 11759 + }, + { + "epoch": 1.002130379207499, + "grad_norm": 14.5625, + "learning_rate": 1.2542008808566946e-06, + "loss": 0.4757, + "step": 11760 + }, + { + "epoch": 1.0022155943757989, + "grad_norm": 12.25, + "learning_rate": 1.2540328462175944e-06, + "loss": 0.3719, + "step": 11761 + }, + { + "epoch": 1.0023008095440988, + "grad_norm": 16.125, + "learning_rate": 1.253864811505616e-06, + "loss": 0.663, + "step": 11762 + }, + { + "epoch": 1.0023860247123988, + "grad_norm": 13.0, + "learning_rate": 1.2536967767237974e-06, + "loss": 0.6091, + "step": 11763 + }, + { + "epoch": 1.0024712398806988, + "grad_norm": 12.3125, + "learning_rate": 1.2535287418751734e-06, + "loss": 0.3965, + "step": 11764 + }, + { + "epoch": 1.0025564550489987, + "grad_norm": 15.0625, + "learning_rate": 1.2533607069627822e-06, + "loss": 0.4776, + "step": 11765 + }, + { + "epoch": 1.0026416702172987, + "grad_norm": 16.625, + "learning_rate": 1.2531926719896592e-06, + "loss": 0.489, + "step": 11766 + }, + { + "epoch": 1.0027268853855986, + "grad_norm": 10.4375, + "learning_rate": 1.2530246369588416e-06, + "loss": 0.2656, + "step": 11767 + }, + { + "epoch": 1.0028121005538986, + "grad_norm": 16.5, + "learning_rate": 1.2528566018733658e-06, + "loss": 0.6037, + "step": 11768 + }, + { + "epoch": 1.0028973157221985, + "grad_norm": 12.875, + "learning_rate": 1.2526885667362685e-06, + "loss": 0.4394, + "step": 11769 + }, + { + "epoch": 1.0029825308904985, + "grad_norm": 12.0625, + "learning_rate": 1.252520531550586e-06, + "loss": 0.4246, + "step": 11770 + }, + { + "epoch": 1.0030677460587984, + "grad_norm": 20.125, + "learning_rate": 1.2523524963193548e-06, + "loss": 0.9519, + "step": 11771 + }, + { + "epoch": 1.0031529612270984, + "grad_norm": 14.625, + "learning_rate": 1.2521844610456118e-06, + "loss": 0.7885, + "step": 11772 + }, + { + "epoch": 1.0032381763953984, + "grad_norm": 14.6875, + "learning_rate": 1.252016425732393e-06, + "loss": 0.5641, + "step": 11773 + }, + { + "epoch": 1.0033233915636983, + "grad_norm": 22.875, + "learning_rate": 1.2518483903827359e-06, + "loss": 0.3061, + "step": 11774 + }, + { + "epoch": 1.0034086067319983, + "grad_norm": 25.75, + "learning_rate": 1.2516803549996762e-06, + "loss": 1.3185, + "step": 11775 + }, + { + "epoch": 1.0034938219002982, + "grad_norm": 22.125, + "learning_rate": 1.251512319586251e-06, + "loss": 0.9861, + "step": 11776 + }, + { + "epoch": 1.0035790370685982, + "grad_norm": 15.1875, + "learning_rate": 1.2513442841454962e-06, + "loss": 0.5692, + "step": 11777 + }, + { + "epoch": 1.0036642522368981, + "grad_norm": 16.875, + "learning_rate": 1.2511762486804493e-06, + "loss": 0.4114, + "step": 11778 + }, + { + "epoch": 1.003749467405198, + "grad_norm": 14.3125, + "learning_rate": 1.2510082131941465e-06, + "loss": 0.7378, + "step": 11779 + }, + { + "epoch": 1.003834682573498, + "grad_norm": 19.375, + "learning_rate": 1.250840177689624e-06, + "loss": 0.4243, + "step": 11780 + }, + { + "epoch": 1.003919897741798, + "grad_norm": 12.8125, + "learning_rate": 1.2506721421699185e-06, + "loss": 0.5027, + "step": 11781 + }, + { + "epoch": 1.004005112910098, + "grad_norm": 14.875, + "learning_rate": 1.250504106638067e-06, + "loss": 0.6991, + "step": 11782 + }, + { + "epoch": 1.004090328078398, + "grad_norm": 13.625, + "learning_rate": 1.2503360710971058e-06, + "loss": 0.514, + "step": 11783 + }, + { + "epoch": 1.0041755432466979, + "grad_norm": 13.375, + "learning_rate": 1.2501680355500711e-06, + "loss": 0.723, + "step": 11784 + }, + { + "epoch": 1.0042607584149978, + "grad_norm": 19.125, + "learning_rate": 1.25e-06, + "loss": 0.7626, + "step": 11785 + }, + { + "epoch": 1.0043459735832978, + "grad_norm": 12.9375, + "learning_rate": 1.2498319644499293e-06, + "loss": 0.4792, + "step": 11786 + }, + { + "epoch": 1.0044311887515978, + "grad_norm": 11.75, + "learning_rate": 1.2496639289028947e-06, + "loss": 0.36, + "step": 11787 + }, + { + "epoch": 1.0045164039198977, + "grad_norm": 21.625, + "learning_rate": 1.2494958933619333e-06, + "loss": 1.3249, + "step": 11788 + }, + { + "epoch": 1.0046016190881977, + "grad_norm": 25.25, + "learning_rate": 1.2493278578300817e-06, + "loss": 1.2345, + "step": 11789 + }, + { + "epoch": 1.0046868342564976, + "grad_norm": 16.875, + "learning_rate": 1.2491598223103766e-06, + "loss": 0.7527, + "step": 11790 + }, + { + "epoch": 1.0047720494247976, + "grad_norm": 10.75, + "learning_rate": 1.248991786805854e-06, + "loss": 0.4175, + "step": 11791 + }, + { + "epoch": 1.0048572645930975, + "grad_norm": 12.625, + "learning_rate": 1.248823751319551e-06, + "loss": 0.3917, + "step": 11792 + }, + { + "epoch": 1.0049424797613975, + "grad_norm": 14.375, + "learning_rate": 1.248655715854504e-06, + "loss": 0.7347, + "step": 11793 + }, + { + "epoch": 1.0050276949296975, + "grad_norm": 19.25, + "learning_rate": 1.2484876804137493e-06, + "loss": 0.9084, + "step": 11794 + }, + { + "epoch": 1.0051129100979974, + "grad_norm": 13.5, + "learning_rate": 1.2483196450003242e-06, + "loss": 0.6526, + "step": 11795 + }, + { + "epoch": 1.0051981252662974, + "grad_norm": 17.5, + "learning_rate": 1.2481516096172643e-06, + "loss": 0.8381, + "step": 11796 + }, + { + "epoch": 1.0052833404345973, + "grad_norm": 14.25, + "learning_rate": 1.247983574267607e-06, + "loss": 0.5441, + "step": 11797 + }, + { + "epoch": 1.0053685556028973, + "grad_norm": 16.5, + "learning_rate": 1.2478155389543888e-06, + "loss": 0.4717, + "step": 11798 + }, + { + "epoch": 1.0054537707711972, + "grad_norm": 16.25, + "learning_rate": 1.2476475036806456e-06, + "loss": 0.7018, + "step": 11799 + }, + { + "epoch": 1.0055389859394972, + "grad_norm": 10.5, + "learning_rate": 1.2474794684494147e-06, + "loss": 0.3482, + "step": 11800 + }, + { + "epoch": 1.0056242011077972, + "grad_norm": 16.875, + "learning_rate": 1.2473114332637322e-06, + "loss": 0.7476, + "step": 11801 + }, + { + "epoch": 1.0057094162760971, + "grad_norm": 14.4375, + "learning_rate": 1.2471433981266344e-06, + "loss": 0.7372, + "step": 11802 + }, + { + "epoch": 1.005794631444397, + "grad_norm": 10.1875, + "learning_rate": 1.2469753630411586e-06, + "loss": 0.3906, + "step": 11803 + }, + { + "epoch": 1.005879846612697, + "grad_norm": 16.125, + "learning_rate": 1.2468073280103408e-06, + "loss": 0.4751, + "step": 11804 + }, + { + "epoch": 1.005965061780997, + "grad_norm": 14.5625, + "learning_rate": 1.2466392930372182e-06, + "loss": 0.6205, + "step": 11805 + }, + { + "epoch": 1.006050276949297, + "grad_norm": 15.5625, + "learning_rate": 1.246471258124827e-06, + "loss": 0.6441, + "step": 11806 + }, + { + "epoch": 1.006135492117597, + "grad_norm": 20.75, + "learning_rate": 1.2463032232762032e-06, + "loss": 0.7897, + "step": 11807 + }, + { + "epoch": 1.0062207072858969, + "grad_norm": 12.9375, + "learning_rate": 1.2461351884943843e-06, + "loss": 0.5583, + "step": 11808 + }, + { + "epoch": 1.0063059224541968, + "grad_norm": 12.4375, + "learning_rate": 1.2459671537824062e-06, + "loss": 0.4413, + "step": 11809 + }, + { + "epoch": 1.0063911376224968, + "grad_norm": 15.4375, + "learning_rate": 1.2457991191433058e-06, + "loss": 0.6705, + "step": 11810 + }, + { + "epoch": 1.0064763527907967, + "grad_norm": 17.25, + "learning_rate": 1.2456310845801192e-06, + "loss": 0.8274, + "step": 11811 + }, + { + "epoch": 1.0065615679590967, + "grad_norm": 11.9375, + "learning_rate": 1.2454630500958837e-06, + "loss": 0.3353, + "step": 11812 + }, + { + "epoch": 1.0066467831273966, + "grad_norm": 12.4375, + "learning_rate": 1.2452950156936353e-06, + "loss": 0.4705, + "step": 11813 + }, + { + "epoch": 1.0067319982956966, + "grad_norm": 17.25, + "learning_rate": 1.245126981376411e-06, + "loss": 0.4269, + "step": 11814 + }, + { + "epoch": 1.0068172134639966, + "grad_norm": 12.1875, + "learning_rate": 1.244958947147247e-06, + "loss": 0.4694, + "step": 11815 + }, + { + "epoch": 1.0069024286322965, + "grad_norm": 14.125, + "learning_rate": 1.2447909130091796e-06, + "loss": 0.457, + "step": 11816 + }, + { + "epoch": 1.0069876438005965, + "grad_norm": 12.6875, + "learning_rate": 1.2446228789652456e-06, + "loss": 0.635, + "step": 11817 + }, + { + "epoch": 1.0070728589688964, + "grad_norm": 10.625, + "learning_rate": 1.2444548450184817e-06, + "loss": 0.3132, + "step": 11818 + }, + { + "epoch": 1.0071580741371964, + "grad_norm": 22.125, + "learning_rate": 1.2442868111719242e-06, + "loss": 1.086, + "step": 11819 + }, + { + "epoch": 1.0072432893054963, + "grad_norm": 20.875, + "learning_rate": 1.24411877742861e-06, + "loss": 1.3773, + "step": 11820 + }, + { + "epoch": 1.0073285044737963, + "grad_norm": 15.3125, + "learning_rate": 1.2439507437915752e-06, + "loss": 0.6197, + "step": 11821 + }, + { + "epoch": 1.0074137196420963, + "grad_norm": 18.25, + "learning_rate": 1.243782710263857e-06, + "loss": 1.0505, + "step": 11822 + }, + { + "epoch": 1.0074989348103962, + "grad_norm": 16.25, + "learning_rate": 1.243614676848491e-06, + "loss": 0.6746, + "step": 11823 + }, + { + "epoch": 1.0075841499786962, + "grad_norm": 12.375, + "learning_rate": 1.2434466435485143e-06, + "loss": 0.6761, + "step": 11824 + }, + { + "epoch": 1.0076693651469961, + "grad_norm": 22.75, + "learning_rate": 1.2432786103669634e-06, + "loss": 0.9316, + "step": 11825 + }, + { + "epoch": 1.007754580315296, + "grad_norm": 13.75, + "learning_rate": 1.2431105773068745e-06, + "loss": 0.6794, + "step": 11826 + }, + { + "epoch": 1.007839795483596, + "grad_norm": 12.6875, + "learning_rate": 1.2429425443712848e-06, + "loss": 0.7683, + "step": 11827 + }, + { + "epoch": 1.007925010651896, + "grad_norm": 15.0, + "learning_rate": 1.24277451156323e-06, + "loss": 0.5239, + "step": 11828 + }, + { + "epoch": 1.008010225820196, + "grad_norm": 17.0, + "learning_rate": 1.2426064788857474e-06, + "loss": 1.1397, + "step": 11829 + }, + { + "epoch": 1.008095440988496, + "grad_norm": 23.25, + "learning_rate": 1.2424384463418734e-06, + "loss": 0.888, + "step": 11830 + }, + { + "epoch": 1.0081806561567959, + "grad_norm": 14.5, + "learning_rate": 1.242270413934644e-06, + "loss": 0.6674, + "step": 11831 + }, + { + "epoch": 1.0082658713250958, + "grad_norm": 12.75, + "learning_rate": 1.2421023816670958e-06, + "loss": 0.4278, + "step": 11832 + }, + { + "epoch": 1.0083510864933958, + "grad_norm": 12.875, + "learning_rate": 1.2419343495422657e-06, + "loss": 0.3403, + "step": 11833 + }, + { + "epoch": 1.0084363016616957, + "grad_norm": 17.5, + "learning_rate": 1.2417663175631899e-06, + "loss": 1.0834, + "step": 11834 + }, + { + "epoch": 1.0085215168299957, + "grad_norm": 15.875, + "learning_rate": 1.2415982857329051e-06, + "loss": 0.4746, + "step": 11835 + }, + { + "epoch": 1.0086067319982956, + "grad_norm": 13.6875, + "learning_rate": 1.2414302540544474e-06, + "loss": 0.2932, + "step": 11836 + }, + { + "epoch": 1.0086919471665956, + "grad_norm": 17.0, + "learning_rate": 1.2412622225308542e-06, + "loss": 0.6587, + "step": 11837 + }, + { + "epoch": 1.0087771623348956, + "grad_norm": 17.5, + "learning_rate": 1.2410941911651613e-06, + "loss": 0.7996, + "step": 11838 + }, + { + "epoch": 1.0088623775031955, + "grad_norm": 16.5, + "learning_rate": 1.240926159960405e-06, + "loss": 0.7228, + "step": 11839 + }, + { + "epoch": 1.0089475926714955, + "grad_norm": 16.375, + "learning_rate": 1.2407581289196224e-06, + "loss": 0.8024, + "step": 11840 + }, + { + "epoch": 1.0090328078397954, + "grad_norm": 16.375, + "learning_rate": 1.2405900980458496e-06, + "loss": 0.5653, + "step": 11841 + }, + { + "epoch": 1.0091180230080954, + "grad_norm": 9.9375, + "learning_rate": 1.240422067342123e-06, + "loss": 0.4692, + "step": 11842 + }, + { + "epoch": 1.0092032381763953, + "grad_norm": 9.6875, + "learning_rate": 1.2402540368114795e-06, + "loss": 0.3533, + "step": 11843 + }, + { + "epoch": 1.0092884533446953, + "grad_norm": 14.125, + "learning_rate": 1.2400860064569553e-06, + "loss": 0.5151, + "step": 11844 + }, + { + "epoch": 1.0093736685129953, + "grad_norm": 11.5625, + "learning_rate": 1.2399179762815868e-06, + "loss": 0.3319, + "step": 11845 + }, + { + "epoch": 1.0094588836812952, + "grad_norm": 13.25, + "learning_rate": 1.2397499462884114e-06, + "loss": 0.6069, + "step": 11846 + }, + { + "epoch": 1.0095440988495952, + "grad_norm": 13.125, + "learning_rate": 1.2395819164804638e-06, + "loss": 0.8084, + "step": 11847 + }, + { + "epoch": 1.0096293140178951, + "grad_norm": 16.25, + "learning_rate": 1.2394138868607821e-06, + "loss": 0.6835, + "step": 11848 + }, + { + "epoch": 1.009714529186195, + "grad_norm": 16.0, + "learning_rate": 1.2392458574324016e-06, + "loss": 0.8352, + "step": 11849 + }, + { + "epoch": 1.009799744354495, + "grad_norm": 14.5625, + "learning_rate": 1.2390778281983594e-06, + "loss": 0.6048, + "step": 11850 + }, + { + "epoch": 1.009884959522795, + "grad_norm": 16.875, + "learning_rate": 1.2389097991616918e-06, + "loss": 0.6834, + "step": 11851 + }, + { + "epoch": 1.009970174691095, + "grad_norm": 14.1875, + "learning_rate": 1.2387417703254354e-06, + "loss": 0.5094, + "step": 11852 + }, + { + "epoch": 1.010055389859395, + "grad_norm": 19.875, + "learning_rate": 1.2385737416926263e-06, + "loss": 0.5048, + "step": 11853 + }, + { + "epoch": 1.0101406050276949, + "grad_norm": 13.9375, + "learning_rate": 1.2384057132663014e-06, + "loss": 0.7713, + "step": 11854 + }, + { + "epoch": 1.0102258201959948, + "grad_norm": 21.25, + "learning_rate": 1.2382376850494973e-06, + "loss": 0.5077, + "step": 11855 + }, + { + "epoch": 1.0103110353642948, + "grad_norm": 14.6875, + "learning_rate": 1.2380696570452493e-06, + "loss": 0.5841, + "step": 11856 + }, + { + "epoch": 1.0103962505325947, + "grad_norm": 13.9375, + "learning_rate": 1.2379016292565952e-06, + "loss": 0.632, + "step": 11857 + }, + { + "epoch": 1.0104814657008947, + "grad_norm": 15.9375, + "learning_rate": 1.2377336016865704e-06, + "loss": 0.5922, + "step": 11858 + }, + { + "epoch": 1.0105666808691947, + "grad_norm": 21.75, + "learning_rate": 1.237565574338212e-06, + "loss": 0.6537, + "step": 11859 + }, + { + "epoch": 1.0106518960374946, + "grad_norm": 11.5, + "learning_rate": 1.2373975472145559e-06, + "loss": 0.3533, + "step": 11860 + }, + { + "epoch": 1.0107371112057946, + "grad_norm": 11.625, + "learning_rate": 1.237229520318639e-06, + "loss": 0.3931, + "step": 11861 + }, + { + "epoch": 1.0108223263740945, + "grad_norm": 20.25, + "learning_rate": 1.2370614936534973e-06, + "loss": 0.7817, + "step": 11862 + }, + { + "epoch": 1.0109075415423945, + "grad_norm": 16.125, + "learning_rate": 1.2368934672221684e-06, + "loss": 0.8911, + "step": 11863 + }, + { + "epoch": 1.0109927567106944, + "grad_norm": 16.625, + "learning_rate": 1.2367254410276868e-06, + "loss": 0.855, + "step": 11864 + }, + { + "epoch": 1.0110779718789944, + "grad_norm": 11.125, + "learning_rate": 1.23655741507309e-06, + "loss": 0.2178, + "step": 11865 + }, + { + "epoch": 1.0111631870472944, + "grad_norm": 26.375, + "learning_rate": 1.236389389361414e-06, + "loss": 0.6537, + "step": 11866 + }, + { + "epoch": 1.0112484022155943, + "grad_norm": 10.4375, + "learning_rate": 1.2362213638956958e-06, + "loss": 0.2725, + "step": 11867 + }, + { + "epoch": 1.0113336173838943, + "grad_norm": 18.5, + "learning_rate": 1.2360533386789709e-06, + "loss": 0.908, + "step": 11868 + }, + { + "epoch": 1.0114188325521942, + "grad_norm": 18.25, + "learning_rate": 1.2358853137142767e-06, + "loss": 0.6545, + "step": 11869 + }, + { + "epoch": 1.0115040477204942, + "grad_norm": 17.625, + "learning_rate": 1.2357172890046487e-06, + "loss": 0.6677, + "step": 11870 + }, + { + "epoch": 1.0115892628887941, + "grad_norm": 15.625, + "learning_rate": 1.2355492645531242e-06, + "loss": 0.6883, + "step": 11871 + }, + { + "epoch": 1.011674478057094, + "grad_norm": 10.8125, + "learning_rate": 1.2353812403627389e-06, + "loss": 0.4726, + "step": 11872 + }, + { + "epoch": 1.011759693225394, + "grad_norm": 12.125, + "learning_rate": 1.2352132164365287e-06, + "loss": 0.434, + "step": 11873 + }, + { + "epoch": 1.011844908393694, + "grad_norm": 12.5625, + "learning_rate": 1.235045192777531e-06, + "loss": 0.4478, + "step": 11874 + }, + { + "epoch": 1.011930123561994, + "grad_norm": 17.625, + "learning_rate": 1.2348771693887815e-06, + "loss": 0.789, + "step": 11875 + }, + { + "epoch": 1.012015338730294, + "grad_norm": 15.25, + "learning_rate": 1.234709146273317e-06, + "loss": 0.6791, + "step": 11876 + }, + { + "epoch": 1.0121005538985939, + "grad_norm": 14.25, + "learning_rate": 1.2345411234341732e-06, + "loss": 0.6555, + "step": 11877 + }, + { + "epoch": 1.0121857690668938, + "grad_norm": 11.4375, + "learning_rate": 1.2343731008743873e-06, + "loss": 0.4536, + "step": 11878 + }, + { + "epoch": 1.0122709842351938, + "grad_norm": 16.5, + "learning_rate": 1.2342050785969954e-06, + "loss": 0.3965, + "step": 11879 + }, + { + "epoch": 1.0123561994034938, + "grad_norm": 10.9375, + "learning_rate": 1.2340370566050332e-06, + "loss": 0.3549, + "step": 11880 + }, + { + "epoch": 1.0124414145717937, + "grad_norm": 13.8125, + "learning_rate": 1.2338690349015373e-06, + "loss": 0.6124, + "step": 11881 + }, + { + "epoch": 1.0125266297400937, + "grad_norm": 18.75, + "learning_rate": 1.2337010134895445e-06, + "loss": 0.5992, + "step": 11882 + }, + { + "epoch": 1.0126118449083936, + "grad_norm": 14.25, + "learning_rate": 1.2335329923720903e-06, + "loss": 0.4274, + "step": 11883 + }, + { + "epoch": 1.0126970600766936, + "grad_norm": 12.875, + "learning_rate": 1.233364971552212e-06, + "loss": 0.4246, + "step": 11884 + }, + { + "epoch": 1.0127822752449935, + "grad_norm": 12.5625, + "learning_rate": 1.2331969510329447e-06, + "loss": 0.4619, + "step": 11885 + }, + { + "epoch": 1.0128674904132935, + "grad_norm": 18.25, + "learning_rate": 1.233028930817326e-06, + "loss": 0.5692, + "step": 11886 + }, + { + "epoch": 1.0129527055815934, + "grad_norm": 17.5, + "learning_rate": 1.2328609109083918e-06, + "loss": 0.62, + "step": 11887 + }, + { + "epoch": 1.0130379207498934, + "grad_norm": 12.1875, + "learning_rate": 1.2326928913091776e-06, + "loss": 0.2991, + "step": 11888 + }, + { + "epoch": 1.0131231359181934, + "grad_norm": 16.25, + "learning_rate": 1.2325248720227207e-06, + "loss": 0.6146, + "step": 11889 + }, + { + "epoch": 1.0132083510864933, + "grad_norm": 12.9375, + "learning_rate": 1.2323568530520564e-06, + "loss": 0.5334, + "step": 11890 + }, + { + "epoch": 1.0132935662547933, + "grad_norm": 14.0, + "learning_rate": 1.232188834400222e-06, + "loss": 0.4535, + "step": 11891 + }, + { + "epoch": 1.0133787814230932, + "grad_norm": 23.625, + "learning_rate": 1.2320208160702528e-06, + "loss": 0.5553, + "step": 11892 + }, + { + "epoch": 1.0134639965913932, + "grad_norm": 11.6875, + "learning_rate": 1.2318527980651859e-06, + "loss": 0.6068, + "step": 11893 + }, + { + "epoch": 1.0135492117596931, + "grad_norm": 17.5, + "learning_rate": 1.2316847803880574e-06, + "loss": 0.4843, + "step": 11894 + }, + { + "epoch": 1.013634426927993, + "grad_norm": 12.8125, + "learning_rate": 1.2315167630419035e-06, + "loss": 0.5045, + "step": 11895 + }, + { + "epoch": 1.013719642096293, + "grad_norm": 15.625, + "learning_rate": 1.2313487460297594e-06, + "loss": 0.9105, + "step": 11896 + }, + { + "epoch": 1.013804857264593, + "grad_norm": 12.0, + "learning_rate": 1.2311807293546627e-06, + "loss": 0.2828, + "step": 11897 + }, + { + "epoch": 1.013890072432893, + "grad_norm": 36.25, + "learning_rate": 1.2310127130196488e-06, + "loss": 1.1786, + "step": 11898 + }, + { + "epoch": 1.013975287601193, + "grad_norm": 22.125, + "learning_rate": 1.2308446970277547e-06, + "loss": 0.8472, + "step": 11899 + }, + { + "epoch": 1.0140605027694929, + "grad_norm": 22.5, + "learning_rate": 1.2306766813820158e-06, + "loss": 0.6123, + "step": 11900 + }, + { + "epoch": 1.0141457179377928, + "grad_norm": 25.625, + "learning_rate": 1.2305086660854692e-06, + "loss": 1.007, + "step": 11901 + }, + { + "epoch": 1.0142309331060928, + "grad_norm": 14.375, + "learning_rate": 1.23034065114115e-06, + "loss": 0.6128, + "step": 11902 + }, + { + "epoch": 1.0143161482743928, + "grad_norm": 12.75, + "learning_rate": 1.230172636552096e-06, + "loss": 0.3913, + "step": 11903 + }, + { + "epoch": 1.0144013634426927, + "grad_norm": 16.125, + "learning_rate": 1.2300046223213418e-06, + "loss": 0.622, + "step": 11904 + }, + { + "epoch": 1.0144865786109927, + "grad_norm": 16.375, + "learning_rate": 1.2298366084519237e-06, + "loss": 0.7441, + "step": 11905 + }, + { + "epoch": 1.0145717937792926, + "grad_norm": 15.5, + "learning_rate": 1.229668594946879e-06, + "loss": 0.822, + "step": 11906 + }, + { + "epoch": 1.0146570089475926, + "grad_norm": 12.625, + "learning_rate": 1.229500581809243e-06, + "loss": 0.4293, + "step": 11907 + }, + { + "epoch": 1.0147422241158925, + "grad_norm": 20.75, + "learning_rate": 1.229332569042052e-06, + "loss": 0.8744, + "step": 11908 + }, + { + "epoch": 1.0148274392841925, + "grad_norm": 16.25, + "learning_rate": 1.2291645566483425e-06, + "loss": 1.002, + "step": 11909 + }, + { + "epoch": 1.0149126544524925, + "grad_norm": 22.625, + "learning_rate": 1.22899654463115e-06, + "loss": 0.524, + "step": 11910 + }, + { + "epoch": 1.0149978696207924, + "grad_norm": 15.6875, + "learning_rate": 1.2288285329935118e-06, + "loss": 0.77, + "step": 11911 + }, + { + "epoch": 1.0150830847890924, + "grad_norm": 12.9375, + "learning_rate": 1.228660521738463e-06, + "loss": 0.3541, + "step": 11912 + }, + { + "epoch": 1.0151682999573923, + "grad_norm": 12.25, + "learning_rate": 1.2284925108690398e-06, + "loss": 0.4796, + "step": 11913 + }, + { + "epoch": 1.0152535151256923, + "grad_norm": 19.375, + "learning_rate": 1.2283245003882787e-06, + "loss": 0.6963, + "step": 11914 + }, + { + "epoch": 1.0153387302939922, + "grad_norm": 21.0, + "learning_rate": 1.2281564902992153e-06, + "loss": 1.0846, + "step": 11915 + }, + { + "epoch": 1.0154239454622922, + "grad_norm": 14.125, + "learning_rate": 1.2279884806048866e-06, + "loss": 0.5684, + "step": 11916 + }, + { + "epoch": 1.0155091606305922, + "grad_norm": 14.9375, + "learning_rate": 1.2278204713083278e-06, + "loss": 0.6135, + "step": 11917 + }, + { + "epoch": 1.0155943757988921, + "grad_norm": 18.25, + "learning_rate": 1.2276524624125756e-06, + "loss": 0.6737, + "step": 11918 + }, + { + "epoch": 1.015679590967192, + "grad_norm": 19.875, + "learning_rate": 1.2274844539206663e-06, + "loss": 0.8726, + "step": 11919 + }, + { + "epoch": 1.015764806135492, + "grad_norm": 19.0, + "learning_rate": 1.2273164458356348e-06, + "loss": 0.6975, + "step": 11920 + }, + { + "epoch": 1.015850021303792, + "grad_norm": 12.8125, + "learning_rate": 1.2271484381605183e-06, + "loss": 0.448, + "step": 11921 + }, + { + "epoch": 1.015935236472092, + "grad_norm": 15.5, + "learning_rate": 1.2269804308983521e-06, + "loss": 0.7315, + "step": 11922 + }, + { + "epoch": 1.016020451640392, + "grad_norm": 14.4375, + "learning_rate": 1.2268124240521732e-06, + "loss": 0.4692, + "step": 11923 + }, + { + "epoch": 1.0161056668086919, + "grad_norm": 14.875, + "learning_rate": 1.226644417625017e-06, + "loss": 0.7028, + "step": 11924 + }, + { + "epoch": 1.0161908819769918, + "grad_norm": 14.5, + "learning_rate": 1.2264764116199194e-06, + "loss": 0.3643, + "step": 11925 + }, + { + "epoch": 1.0162760971452918, + "grad_norm": 16.0, + "learning_rate": 1.226308406039917e-06, + "loss": 0.6644, + "step": 11926 + }, + { + "epoch": 1.0163613123135917, + "grad_norm": 12.875, + "learning_rate": 1.2261404008880452e-06, + "loss": 0.5277, + "step": 11927 + }, + { + "epoch": 1.0164465274818917, + "grad_norm": 16.625, + "learning_rate": 1.225972396167341e-06, + "loss": 0.5956, + "step": 11928 + }, + { + "epoch": 1.0165317426501916, + "grad_norm": 19.0, + "learning_rate": 1.2258043918808394e-06, + "loss": 0.9156, + "step": 11929 + }, + { + "epoch": 1.0166169578184916, + "grad_norm": 12.25, + "learning_rate": 1.2256363880315766e-06, + "loss": 0.3284, + "step": 11930 + }, + { + "epoch": 1.0167021729867916, + "grad_norm": 17.125, + "learning_rate": 1.225468384622589e-06, + "loss": 0.6787, + "step": 11931 + }, + { + "epoch": 1.0167873881550915, + "grad_norm": 20.875, + "learning_rate": 1.225300381656912e-06, + "loss": 0.5145, + "step": 11932 + }, + { + "epoch": 1.0168726033233915, + "grad_norm": 13.25, + "learning_rate": 1.225132379137582e-06, + "loss": 0.5158, + "step": 11933 + }, + { + "epoch": 1.0169578184916914, + "grad_norm": 13.6875, + "learning_rate": 1.224964377067635e-06, + "loss": 0.2742, + "step": 11934 + }, + { + "epoch": 1.0170430336599914, + "grad_norm": 14.375, + "learning_rate": 1.224796375450107e-06, + "loss": 0.5291, + "step": 11935 + }, + { + "epoch": 1.0171282488282913, + "grad_norm": 11.0, + "learning_rate": 1.224628374288034e-06, + "loss": 0.5255, + "step": 11936 + }, + { + "epoch": 1.0172134639965913, + "grad_norm": 17.125, + "learning_rate": 1.2244603735844515e-06, + "loss": 0.542, + "step": 11937 + }, + { + "epoch": 1.0172986791648913, + "grad_norm": 12.125, + "learning_rate": 1.224292373342396e-06, + "loss": 0.4438, + "step": 11938 + }, + { + "epoch": 1.0173838943331912, + "grad_norm": 10.625, + "learning_rate": 1.2241243735649029e-06, + "loss": 0.36, + "step": 11939 + }, + { + "epoch": 1.0174691095014912, + "grad_norm": 18.875, + "learning_rate": 1.2239563742550084e-06, + "loss": 0.6099, + "step": 11940 + }, + { + "epoch": 1.0175543246697911, + "grad_norm": 19.75, + "learning_rate": 1.2237883754157485e-06, + "loss": 0.6168, + "step": 11941 + }, + { + "epoch": 1.017639539838091, + "grad_norm": 11.1875, + "learning_rate": 1.223620377050159e-06, + "loss": 0.4309, + "step": 11942 + }, + { + "epoch": 1.017724755006391, + "grad_norm": 24.25, + "learning_rate": 1.2234523791612757e-06, + "loss": 0.697, + "step": 11943 + }, + { + "epoch": 1.017809970174691, + "grad_norm": 17.75, + "learning_rate": 1.223284381752135e-06, + "loss": 0.6615, + "step": 11944 + }, + { + "epoch": 1.017895185342991, + "grad_norm": 16.5, + "learning_rate": 1.223116384825772e-06, + "loss": 0.6214, + "step": 11945 + }, + { + "epoch": 1.017980400511291, + "grad_norm": 17.25, + "learning_rate": 1.2229483883852231e-06, + "loss": 0.685, + "step": 11946 + }, + { + "epoch": 1.0180656156795909, + "grad_norm": 14.8125, + "learning_rate": 1.2227803924335237e-06, + "loss": 0.582, + "step": 11947 + }, + { + "epoch": 1.0181508308478908, + "grad_norm": 15.25, + "learning_rate": 1.2226123969737103e-06, + "loss": 0.4938, + "step": 11948 + }, + { + "epoch": 1.0182360460161908, + "grad_norm": 13.25, + "learning_rate": 1.2224444020088183e-06, + "loss": 0.2816, + "step": 11949 + }, + { + "epoch": 1.0183212611844907, + "grad_norm": 15.8125, + "learning_rate": 1.2222764075418838e-06, + "loss": 0.6537, + "step": 11950 + }, + { + "epoch": 1.0184064763527907, + "grad_norm": 12.6875, + "learning_rate": 1.2221084135759422e-06, + "loss": 0.4018, + "step": 11951 + }, + { + "epoch": 1.0184916915210906, + "grad_norm": 16.75, + "learning_rate": 1.2219404201140303e-06, + "loss": 0.6995, + "step": 11952 + }, + { + "epoch": 1.0185769066893906, + "grad_norm": 15.5625, + "learning_rate": 1.2217724271591828e-06, + "loss": 0.5302, + "step": 11953 + }, + { + "epoch": 1.0186621218576906, + "grad_norm": 15.125, + "learning_rate": 1.2216044347144357e-06, + "loss": 0.6748, + "step": 11954 + }, + { + "epoch": 1.0187473370259905, + "grad_norm": 21.25, + "learning_rate": 1.2214364427828254e-06, + "loss": 0.8775, + "step": 11955 + }, + { + "epoch": 1.0188325521942905, + "grad_norm": 22.875, + "learning_rate": 1.2212684513673873e-06, + "loss": 0.6695, + "step": 11956 + }, + { + "epoch": 1.0189177673625904, + "grad_norm": 17.75, + "learning_rate": 1.2211004604711566e-06, + "loss": 0.5567, + "step": 11957 + }, + { + "epoch": 1.0190029825308904, + "grad_norm": 16.25, + "learning_rate": 1.2209324700971702e-06, + "loss": 0.5106, + "step": 11958 + }, + { + "epoch": 1.0190881976991903, + "grad_norm": 18.5, + "learning_rate": 1.2207644802484629e-06, + "loss": 0.6221, + "step": 11959 + }, + { + "epoch": 1.0191734128674903, + "grad_norm": 13.5, + "learning_rate": 1.2205964909280715e-06, + "loss": 0.4123, + "step": 11960 + }, + { + "epoch": 1.0192586280357903, + "grad_norm": 16.5, + "learning_rate": 1.2204285021390308e-06, + "loss": 1.0866, + "step": 11961 + }, + { + "epoch": 1.0193438432040902, + "grad_norm": 15.625, + "learning_rate": 1.2202605138843763e-06, + "loss": 0.6068, + "step": 11962 + }, + { + "epoch": 1.0194290583723902, + "grad_norm": 15.25, + "learning_rate": 1.2200925261671448e-06, + "loss": 0.5363, + "step": 11963 + }, + { + "epoch": 1.0195142735406901, + "grad_norm": 15.75, + "learning_rate": 1.219924538990371e-06, + "loss": 0.5703, + "step": 11964 + }, + { + "epoch": 1.0195994887089903, + "grad_norm": 12.125, + "learning_rate": 1.2197565523570914e-06, + "loss": 0.3931, + "step": 11965 + }, + { + "epoch": 1.0196847038772903, + "grad_norm": 16.125, + "learning_rate": 1.2195885662703407e-06, + "loss": 0.608, + "step": 11966 + }, + { + "epoch": 1.0197699190455902, + "grad_norm": 18.625, + "learning_rate": 1.219420580733156e-06, + "loss": 0.5983, + "step": 11967 + }, + { + "epoch": 1.0198551342138902, + "grad_norm": 12.3125, + "learning_rate": 1.219252595748572e-06, + "loss": 0.5137, + "step": 11968 + }, + { + "epoch": 1.0199403493821901, + "grad_norm": 12.125, + "learning_rate": 1.2190846113196242e-06, + "loss": 0.4368, + "step": 11969 + }, + { + "epoch": 1.02002556455049, + "grad_norm": 21.625, + "learning_rate": 1.2189166274493488e-06, + "loss": 0.7287, + "step": 11970 + }, + { + "epoch": 1.02011077971879, + "grad_norm": 25.375, + "learning_rate": 1.218748644140781e-06, + "loss": 0.7125, + "step": 11971 + }, + { + "epoch": 1.02019599488709, + "grad_norm": 13.0625, + "learning_rate": 1.2185806613969564e-06, + "loss": 0.4505, + "step": 11972 + }, + { + "epoch": 1.02028121005539, + "grad_norm": 12.3125, + "learning_rate": 1.2184126792209113e-06, + "loss": 0.4416, + "step": 11973 + }, + { + "epoch": 1.02036642522369, + "grad_norm": 12.8125, + "learning_rate": 1.2182446976156804e-06, + "loss": 0.4423, + "step": 11974 + }, + { + "epoch": 1.0204516403919899, + "grad_norm": 20.75, + "learning_rate": 1.2180767165843001e-06, + "loss": 0.838, + "step": 11975 + }, + { + "epoch": 1.0205368555602898, + "grad_norm": 18.75, + "learning_rate": 1.217908736129806e-06, + "loss": 0.5747, + "step": 11976 + }, + { + "epoch": 1.0206220707285898, + "grad_norm": 21.5, + "learning_rate": 1.2177407562552327e-06, + "loss": 0.7355, + "step": 11977 + }, + { + "epoch": 1.0207072858968897, + "grad_norm": 20.875, + "learning_rate": 1.2175727769636166e-06, + "loss": 0.655, + "step": 11978 + }, + { + "epoch": 1.0207925010651897, + "grad_norm": 11.4375, + "learning_rate": 1.2174047982579927e-06, + "loss": 0.392, + "step": 11979 + }, + { + "epoch": 1.0208777162334897, + "grad_norm": 16.375, + "learning_rate": 1.2172368201413972e-06, + "loss": 0.605, + "step": 11980 + }, + { + "epoch": 1.0209629314017896, + "grad_norm": 14.5625, + "learning_rate": 1.217068842616865e-06, + "loss": 0.5198, + "step": 11981 + }, + { + "epoch": 1.0210481465700896, + "grad_norm": 16.0, + "learning_rate": 1.2169008656874323e-06, + "loss": 0.5874, + "step": 11982 + }, + { + "epoch": 1.0211333617383895, + "grad_norm": 12.8125, + "learning_rate": 1.216732889356134e-06, + "loss": 0.7363, + "step": 11983 + }, + { + "epoch": 1.0212185769066895, + "grad_norm": 19.625, + "learning_rate": 1.2165649136260063e-06, + "loss": 0.5927, + "step": 11984 + }, + { + "epoch": 1.0213037920749894, + "grad_norm": 16.25, + "learning_rate": 1.216396938500084e-06, + "loss": 0.8948, + "step": 11985 + }, + { + "epoch": 1.0213890072432894, + "grad_norm": 14.75, + "learning_rate": 1.2162289639814027e-06, + "loss": 0.7612, + "step": 11986 + }, + { + "epoch": 1.0214742224115894, + "grad_norm": 14.5, + "learning_rate": 1.216060990072998e-06, + "loss": 0.3882, + "step": 11987 + }, + { + "epoch": 1.0215594375798893, + "grad_norm": 11.3125, + "learning_rate": 1.2158930167779053e-06, + "loss": 0.5435, + "step": 11988 + }, + { + "epoch": 1.0216446527481893, + "grad_norm": 15.4375, + "learning_rate": 1.2157250440991599e-06, + "loss": 0.6417, + "step": 11989 + }, + { + "epoch": 1.0217298679164892, + "grad_norm": 12.9375, + "learning_rate": 1.2155570720397978e-06, + "loss": 0.4316, + "step": 11990 + }, + { + "epoch": 1.0218150830847892, + "grad_norm": 12.9375, + "learning_rate": 1.2153891006028536e-06, + "loss": 0.7241, + "step": 11991 + }, + { + "epoch": 1.0219002982530891, + "grad_norm": 14.75, + "learning_rate": 1.2152211297913633e-06, + "loss": 0.6106, + "step": 11992 + }, + { + "epoch": 1.021985513421389, + "grad_norm": 14.625, + "learning_rate": 1.2150531596083628e-06, + "loss": 0.6398, + "step": 11993 + }, + { + "epoch": 1.022070728589689, + "grad_norm": 15.3125, + "learning_rate": 1.2148851900568862e-06, + "loss": 0.7378, + "step": 11994 + }, + { + "epoch": 1.022155943757989, + "grad_norm": 15.0, + "learning_rate": 1.2147172211399698e-06, + "loss": 0.4501, + "step": 11995 + }, + { + "epoch": 1.022241158926289, + "grad_norm": 17.25, + "learning_rate": 1.2145492528606483e-06, + "loss": 0.4264, + "step": 11996 + }, + { + "epoch": 1.022326374094589, + "grad_norm": 17.375, + "learning_rate": 1.2143812852219577e-06, + "loss": 0.5939, + "step": 11997 + }, + { + "epoch": 1.0224115892628889, + "grad_norm": 15.125, + "learning_rate": 1.2142133182269328e-06, + "loss": 0.7487, + "step": 11998 + }, + { + "epoch": 1.0224968044311888, + "grad_norm": 18.75, + "learning_rate": 1.2140453518786097e-06, + "loss": 0.9839, + "step": 11999 + }, + { + "epoch": 1.0225820195994888, + "grad_norm": 12.375, + "learning_rate": 1.2138773861800227e-06, + "loss": 0.434, + "step": 12000 + }, + { + "epoch": 1.0226672347677888, + "grad_norm": 14.3125, + "learning_rate": 1.2137094211342087e-06, + "loss": 0.5554, + "step": 12001 + }, + { + "epoch": 1.0227524499360887, + "grad_norm": 14.9375, + "learning_rate": 1.2135414567442012e-06, + "loss": 0.5954, + "step": 12002 + }, + { + "epoch": 1.0228376651043887, + "grad_norm": 18.875, + "learning_rate": 1.2133734930130365e-06, + "loss": 0.6646, + "step": 12003 + }, + { + "epoch": 1.0229228802726886, + "grad_norm": 9.5, + "learning_rate": 1.213205529943749e-06, + "loss": 0.4746, + "step": 12004 + }, + { + "epoch": 1.0230080954409886, + "grad_norm": 15.9375, + "learning_rate": 1.2130375675393752e-06, + "loss": 0.8507, + "step": 12005 + }, + { + "epoch": 1.0230933106092885, + "grad_norm": 23.25, + "learning_rate": 1.2128696058029494e-06, + "loss": 0.6255, + "step": 12006 + }, + { + "epoch": 1.0231785257775885, + "grad_norm": 26.375, + "learning_rate": 1.2127016447375072e-06, + "loss": 0.7, + "step": 12007 + }, + { + "epoch": 1.0232637409458885, + "grad_norm": 13.0625, + "learning_rate": 1.2125336843460837e-06, + "loss": 0.4802, + "step": 12008 + }, + { + "epoch": 1.0233489561141884, + "grad_norm": 17.375, + "learning_rate": 1.2123657246317146e-06, + "loss": 0.8062, + "step": 12009 + }, + { + "epoch": 1.0234341712824884, + "grad_norm": 17.5, + "learning_rate": 1.2121977655974346e-06, + "loss": 0.8518, + "step": 12010 + }, + { + "epoch": 1.0235193864507883, + "grad_norm": 13.25, + "learning_rate": 1.2120298072462787e-06, + "loss": 0.4189, + "step": 12011 + }, + { + "epoch": 1.0236046016190883, + "grad_norm": 11.25, + "learning_rate": 1.2118618495812827e-06, + "loss": 0.3746, + "step": 12012 + }, + { + "epoch": 1.0236898167873882, + "grad_norm": 16.25, + "learning_rate": 1.211693892605481e-06, + "loss": 0.524, + "step": 12013 + }, + { + "epoch": 1.0237750319556882, + "grad_norm": 15.625, + "learning_rate": 1.2115259363219096e-06, + "loss": 0.6629, + "step": 12014 + }, + { + "epoch": 1.0238602471239882, + "grad_norm": 19.25, + "learning_rate": 1.2113579807336027e-06, + "loss": 0.852, + "step": 12015 + }, + { + "epoch": 1.0239454622922881, + "grad_norm": 11.5, + "learning_rate": 1.2111900258435964e-06, + "loss": 0.3769, + "step": 12016 + }, + { + "epoch": 1.024030677460588, + "grad_norm": 17.375, + "learning_rate": 1.2110220716549257e-06, + "loss": 1.0195, + "step": 12017 + }, + { + "epoch": 1.024115892628888, + "grad_norm": 12.4375, + "learning_rate": 1.210854118170625e-06, + "loss": 0.2603, + "step": 12018 + }, + { + "epoch": 1.024201107797188, + "grad_norm": 16.875, + "learning_rate": 1.2106861653937293e-06, + "loss": 0.501, + "step": 12019 + }, + { + "epoch": 1.024286322965488, + "grad_norm": 14.5625, + "learning_rate": 1.2105182133272747e-06, + "loss": 0.6349, + "step": 12020 + }, + { + "epoch": 1.024371538133788, + "grad_norm": 18.125, + "learning_rate": 1.2103502619742954e-06, + "loss": 0.7254, + "step": 12021 + }, + { + "epoch": 1.0244567533020879, + "grad_norm": 18.375, + "learning_rate": 1.2101823113378268e-06, + "loss": 0.8706, + "step": 12022 + }, + { + "epoch": 1.0245419684703878, + "grad_norm": 18.0, + "learning_rate": 1.2100143614209038e-06, + "loss": 0.6695, + "step": 12023 + }, + { + "epoch": 1.0246271836386878, + "grad_norm": 20.5, + "learning_rate": 1.2098464122265617e-06, + "loss": 0.5533, + "step": 12024 + }, + { + "epoch": 1.0247123988069877, + "grad_norm": 24.0, + "learning_rate": 1.2096784637578355e-06, + "loss": 0.8595, + "step": 12025 + }, + { + "epoch": 1.0247976139752877, + "grad_norm": 11.0625, + "learning_rate": 1.2095105160177594e-06, + "loss": 0.4157, + "step": 12026 + }, + { + "epoch": 1.0248828291435876, + "grad_norm": 17.625, + "learning_rate": 1.2093425690093694e-06, + "loss": 0.6626, + "step": 12027 + }, + { + "epoch": 1.0249680443118876, + "grad_norm": 15.5, + "learning_rate": 1.2091746227356996e-06, + "loss": 0.5582, + "step": 12028 + }, + { + "epoch": 1.0250532594801876, + "grad_norm": 15.25, + "learning_rate": 1.2090066771997858e-06, + "loss": 0.6928, + "step": 12029 + }, + { + "epoch": 1.0251384746484875, + "grad_norm": 14.8125, + "learning_rate": 1.2088387324046624e-06, + "loss": 0.6567, + "step": 12030 + }, + { + "epoch": 1.0252236898167875, + "grad_norm": 12.6875, + "learning_rate": 1.2086707883533649e-06, + "loss": 0.3253, + "step": 12031 + }, + { + "epoch": 1.0253089049850874, + "grad_norm": 22.5, + "learning_rate": 1.2085028450489272e-06, + "loss": 0.8854, + "step": 12032 + }, + { + "epoch": 1.0253941201533874, + "grad_norm": 20.875, + "learning_rate": 1.2083349024943857e-06, + "loss": 0.8109, + "step": 12033 + }, + { + "epoch": 1.0254793353216873, + "grad_norm": 13.75, + "learning_rate": 1.2081669606927738e-06, + "loss": 0.6547, + "step": 12034 + }, + { + "epoch": 1.0255645504899873, + "grad_norm": 13.0625, + "learning_rate": 1.2079990196471273e-06, + "loss": 0.4259, + "step": 12035 + }, + { + "epoch": 1.0256497656582872, + "grad_norm": 31.875, + "learning_rate": 1.2078310793604802e-06, + "loss": 0.5704, + "step": 12036 + }, + { + "epoch": 1.0257349808265872, + "grad_norm": 17.5, + "learning_rate": 1.2076631398358683e-06, + "loss": 0.4021, + "step": 12037 + }, + { + "epoch": 1.0258201959948872, + "grad_norm": 14.625, + "learning_rate": 1.2074952010763259e-06, + "loss": 0.4289, + "step": 12038 + }, + { + "epoch": 1.0259054111631871, + "grad_norm": 29.625, + "learning_rate": 1.207327263084888e-06, + "loss": 0.837, + "step": 12039 + }, + { + "epoch": 1.025990626331487, + "grad_norm": 21.75, + "learning_rate": 1.207159325864589e-06, + "loss": 0.8295, + "step": 12040 + }, + { + "epoch": 1.026075841499787, + "grad_norm": 9.1875, + "learning_rate": 1.2069913894184648e-06, + "loss": 0.5827, + "step": 12041 + }, + { + "epoch": 1.026161056668087, + "grad_norm": 16.875, + "learning_rate": 1.2068234537495493e-06, + "loss": 0.7938, + "step": 12042 + }, + { + "epoch": 1.026246271836387, + "grad_norm": 16.875, + "learning_rate": 1.2066555188608769e-06, + "loss": 0.6356, + "step": 12043 + }, + { + "epoch": 1.026331487004687, + "grad_norm": 16.625, + "learning_rate": 1.2064875847554834e-06, + "loss": 0.7821, + "step": 12044 + }, + { + "epoch": 1.0264167021729869, + "grad_norm": 16.75, + "learning_rate": 1.2063196514364024e-06, + "loss": 0.6765, + "step": 12045 + }, + { + "epoch": 1.0265019173412868, + "grad_norm": 12.625, + "learning_rate": 1.2061517189066695e-06, + "loss": 0.3162, + "step": 12046 + }, + { + "epoch": 1.0265871325095868, + "grad_norm": 13.8125, + "learning_rate": 1.2059837871693193e-06, + "loss": 0.9344, + "step": 12047 + }, + { + "epoch": 1.0266723476778867, + "grad_norm": 17.125, + "learning_rate": 1.205815856227386e-06, + "loss": 0.8398, + "step": 12048 + }, + { + "epoch": 1.0267575628461867, + "grad_norm": 12.0625, + "learning_rate": 1.2056479260839053e-06, + "loss": 0.4291, + "step": 12049 + }, + { + "epoch": 1.0268427780144866, + "grad_norm": 25.75, + "learning_rate": 1.2054799967419109e-06, + "loss": 0.518, + "step": 12050 + }, + { + "epoch": 1.0269279931827866, + "grad_norm": 13.8125, + "learning_rate": 1.2053120682044372e-06, + "loss": 0.839, + "step": 12051 + }, + { + "epoch": 1.0270132083510866, + "grad_norm": 12.5625, + "learning_rate": 1.2051441404745199e-06, + "loss": 0.4096, + "step": 12052 + }, + { + "epoch": 1.0270984235193865, + "grad_norm": 12.3125, + "learning_rate": 1.2049762135551925e-06, + "loss": 0.4311, + "step": 12053 + }, + { + "epoch": 1.0271836386876865, + "grad_norm": 15.875, + "learning_rate": 1.204808287449491e-06, + "loss": 0.8613, + "step": 12054 + }, + { + "epoch": 1.0272688538559864, + "grad_norm": 15.5, + "learning_rate": 1.2046403621604483e-06, + "loss": 0.9077, + "step": 12055 + }, + { + "epoch": 1.0273540690242864, + "grad_norm": 15.125, + "learning_rate": 1.2044724376911005e-06, + "loss": 0.5185, + "step": 12056 + }, + { + "epoch": 1.0274392841925863, + "grad_norm": 15.125, + "learning_rate": 1.2043045140444812e-06, + "loss": 0.741, + "step": 12057 + }, + { + "epoch": 1.0275244993608863, + "grad_norm": 16.375, + "learning_rate": 1.2041365912236257e-06, + "loss": 0.5084, + "step": 12058 + }, + { + "epoch": 1.0276097145291863, + "grad_norm": 14.4375, + "learning_rate": 1.203968669231568e-06, + "loss": 0.7012, + "step": 12059 + }, + { + "epoch": 1.0276949296974862, + "grad_norm": 19.375, + "learning_rate": 1.2038007480713424e-06, + "loss": 0.7373, + "step": 12060 + }, + { + "epoch": 1.0277801448657862, + "grad_norm": 13.9375, + "learning_rate": 1.2036328277459842e-06, + "loss": 0.8328, + "step": 12061 + }, + { + "epoch": 1.0278653600340861, + "grad_norm": 12.4375, + "learning_rate": 1.203464908258527e-06, + "loss": 0.5911, + "step": 12062 + }, + { + "epoch": 1.027950575202386, + "grad_norm": 17.375, + "learning_rate": 1.2032969896120058e-06, + "loss": 1.1189, + "step": 12063 + }, + { + "epoch": 1.028035790370686, + "grad_norm": 14.75, + "learning_rate": 1.2031290718094552e-06, + "loss": 0.3555, + "step": 12064 + }, + { + "epoch": 1.028121005538986, + "grad_norm": 11.25, + "learning_rate": 1.202961154853909e-06, + "loss": 0.5411, + "step": 12065 + }, + { + "epoch": 1.028206220707286, + "grad_norm": 9.4375, + "learning_rate": 1.2027932387484025e-06, + "loss": 0.4445, + "step": 12066 + }, + { + "epoch": 1.028291435875586, + "grad_norm": 13.4375, + "learning_rate": 1.2026253234959695e-06, + "loss": 0.4197, + "step": 12067 + }, + { + "epoch": 1.0283766510438859, + "grad_norm": 13.5, + "learning_rate": 1.2024574090996444e-06, + "loss": 0.4373, + "step": 12068 + }, + { + "epoch": 1.0284618662121858, + "grad_norm": 13.4375, + "learning_rate": 1.2022894955624619e-06, + "loss": 0.5399, + "step": 12069 + }, + { + "epoch": 1.0285470813804858, + "grad_norm": 13.25, + "learning_rate": 1.202121582887456e-06, + "loss": 0.7038, + "step": 12070 + }, + { + "epoch": 1.0286322965487857, + "grad_norm": 17.0, + "learning_rate": 1.2019536710776611e-06, + "loss": 0.6916, + "step": 12071 + }, + { + "epoch": 1.0287175117170857, + "grad_norm": 13.5625, + "learning_rate": 1.201785760136112e-06, + "loss": 0.2329, + "step": 12072 + }, + { + "epoch": 1.0288027268853857, + "grad_norm": 17.0, + "learning_rate": 1.2016178500658426e-06, + "loss": 0.4703, + "step": 12073 + }, + { + "epoch": 1.0288879420536856, + "grad_norm": 33.5, + "learning_rate": 1.2014499408698877e-06, + "loss": 0.6961, + "step": 12074 + }, + { + "epoch": 1.0289731572219856, + "grad_norm": 10.5625, + "learning_rate": 1.2012820325512805e-06, + "loss": 0.3699, + "step": 12075 + }, + { + "epoch": 1.0290583723902855, + "grad_norm": 17.25, + "learning_rate": 1.2011141251130564e-06, + "loss": 0.7795, + "step": 12076 + }, + { + "epoch": 1.0291435875585855, + "grad_norm": 12.375, + "learning_rate": 1.2009462185582489e-06, + "loss": 0.2929, + "step": 12077 + }, + { + "epoch": 1.0292288027268854, + "grad_norm": 12.0, + "learning_rate": 1.200778312889893e-06, + "loss": 0.5457, + "step": 12078 + }, + { + "epoch": 1.0293140178951854, + "grad_norm": 17.0, + "learning_rate": 1.200610408111022e-06, + "loss": 0.8433, + "step": 12079 + }, + { + "epoch": 1.0293992330634854, + "grad_norm": 13.375, + "learning_rate": 1.2004425042246706e-06, + "loss": 0.5004, + "step": 12080 + }, + { + "epoch": 1.0294844482317853, + "grad_norm": 14.0, + "learning_rate": 1.2002746012338733e-06, + "loss": 0.5252, + "step": 12081 + }, + { + "epoch": 1.0295696634000853, + "grad_norm": 15.875, + "learning_rate": 1.2001066991416642e-06, + "loss": 0.3842, + "step": 12082 + }, + { + "epoch": 1.0296548785683852, + "grad_norm": 12.5625, + "learning_rate": 1.1999387979510767e-06, + "loss": 0.3632, + "step": 12083 + }, + { + "epoch": 1.0297400937366852, + "grad_norm": 21.625, + "learning_rate": 1.1997708976651457e-06, + "loss": 0.9221, + "step": 12084 + }, + { + "epoch": 1.0298253089049851, + "grad_norm": 17.625, + "learning_rate": 1.1996029982869049e-06, + "loss": 0.7283, + "step": 12085 + }, + { + "epoch": 1.029910524073285, + "grad_norm": 26.625, + "learning_rate": 1.1994350998193886e-06, + "loss": 0.7366, + "step": 12086 + }, + { + "epoch": 1.029995739241585, + "grad_norm": 20.625, + "learning_rate": 1.199267202265631e-06, + "loss": 0.8069, + "step": 12087 + }, + { + "epoch": 1.030080954409885, + "grad_norm": 11.625, + "learning_rate": 1.1990993056286662e-06, + "loss": 0.2936, + "step": 12088 + }, + { + "epoch": 1.030166169578185, + "grad_norm": 12.9375, + "learning_rate": 1.1989314099115277e-06, + "loss": 0.4712, + "step": 12089 + }, + { + "epoch": 1.030251384746485, + "grad_norm": 19.625, + "learning_rate": 1.1987635151172507e-06, + "loss": 0.8653, + "step": 12090 + }, + { + "epoch": 1.0303365999147849, + "grad_norm": 11.75, + "learning_rate": 1.1985956212488682e-06, + "loss": 0.4539, + "step": 12091 + }, + { + "epoch": 1.0304218150830848, + "grad_norm": 21.0, + "learning_rate": 1.1984277283094142e-06, + "loss": 0.5299, + "step": 12092 + }, + { + "epoch": 1.0305070302513848, + "grad_norm": 17.75, + "learning_rate": 1.1982598363019232e-06, + "loss": 0.4305, + "step": 12093 + }, + { + "epoch": 1.0305922454196847, + "grad_norm": 12.5, + "learning_rate": 1.198091945229429e-06, + "loss": 0.5654, + "step": 12094 + }, + { + "epoch": 1.0306774605879847, + "grad_norm": 16.75, + "learning_rate": 1.1979240550949654e-06, + "loss": 0.7586, + "step": 12095 + }, + { + "epoch": 1.0307626757562847, + "grad_norm": 15.875, + "learning_rate": 1.1977561659015667e-06, + "loss": 0.6773, + "step": 12096 + }, + { + "epoch": 1.0308478909245846, + "grad_norm": 13.9375, + "learning_rate": 1.1975882776522663e-06, + "loss": 0.6426, + "step": 12097 + }, + { + "epoch": 1.0309331060928846, + "grad_norm": 46.5, + "learning_rate": 1.197420390350099e-06, + "loss": 0.7543, + "step": 12098 + }, + { + "epoch": 1.0310183212611845, + "grad_norm": 9.75, + "learning_rate": 1.197252503998098e-06, + "loss": 0.2529, + "step": 12099 + }, + { + "epoch": 1.0311035364294845, + "grad_norm": 13.75, + "learning_rate": 1.197084618599297e-06, + "loss": 0.5301, + "step": 12100 + }, + { + "epoch": 1.0311887515977844, + "grad_norm": 12.6875, + "learning_rate": 1.1969167341567302e-06, + "loss": 0.3899, + "step": 12101 + }, + { + "epoch": 1.0312739667660844, + "grad_norm": 12.5, + "learning_rate": 1.1967488506734313e-06, + "loss": 0.4672, + "step": 12102 + }, + { + "epoch": 1.0313591819343844, + "grad_norm": 14.0, + "learning_rate": 1.1965809681524345e-06, + "loss": 0.4629, + "step": 12103 + }, + { + "epoch": 1.0314443971026843, + "grad_norm": 12.8125, + "learning_rate": 1.1964130865967728e-06, + "loss": 0.2669, + "step": 12104 + }, + { + "epoch": 1.0315296122709843, + "grad_norm": 15.0625, + "learning_rate": 1.196245206009481e-06, + "loss": 0.5727, + "step": 12105 + }, + { + "epoch": 1.0316148274392842, + "grad_norm": 17.0, + "learning_rate": 1.1960773263935927e-06, + "loss": 1.0707, + "step": 12106 + }, + { + "epoch": 1.0317000426075842, + "grad_norm": 13.5, + "learning_rate": 1.1959094477521408e-06, + "loss": 0.3832, + "step": 12107 + }, + { + "epoch": 1.0317852577758841, + "grad_norm": 15.1875, + "learning_rate": 1.1957415700881597e-06, + "loss": 0.7825, + "step": 12108 + }, + { + "epoch": 1.031870472944184, + "grad_norm": 16.0, + "learning_rate": 1.195573693404683e-06, + "loss": 0.7548, + "step": 12109 + }, + { + "epoch": 1.031955688112484, + "grad_norm": 13.625, + "learning_rate": 1.1954058177047442e-06, + "loss": 0.6043, + "step": 12110 + }, + { + "epoch": 1.032040903280784, + "grad_norm": 13.875, + "learning_rate": 1.1952379429913773e-06, + "loss": 0.6998, + "step": 12111 + }, + { + "epoch": 1.032126118449084, + "grad_norm": 14.375, + "learning_rate": 1.1950700692676158e-06, + "loss": 0.7099, + "step": 12112 + }, + { + "epoch": 1.032211333617384, + "grad_norm": 17.25, + "learning_rate": 1.1949021965364934e-06, + "loss": 0.6449, + "step": 12113 + }, + { + "epoch": 1.0322965487856839, + "grad_norm": 18.5, + "learning_rate": 1.1947343248010441e-06, + "loss": 0.8241, + "step": 12114 + }, + { + "epoch": 1.0323817639539838, + "grad_norm": 16.125, + "learning_rate": 1.1945664540643006e-06, + "loss": 0.6415, + "step": 12115 + }, + { + "epoch": 1.0324669791222838, + "grad_norm": 14.8125, + "learning_rate": 1.1943985843292973e-06, + "loss": 0.743, + "step": 12116 + }, + { + "epoch": 1.0325521942905838, + "grad_norm": 14.75, + "learning_rate": 1.1942307155990673e-06, + "loss": 0.6418, + "step": 12117 + }, + { + "epoch": 1.0326374094588837, + "grad_norm": 24.25, + "learning_rate": 1.1940628478766445e-06, + "loss": 0.7129, + "step": 12118 + }, + { + "epoch": 1.0327226246271837, + "grad_norm": 17.0, + "learning_rate": 1.1938949811650622e-06, + "loss": 0.7839, + "step": 12119 + }, + { + "epoch": 1.0328078397954836, + "grad_norm": 18.0, + "learning_rate": 1.193727115467354e-06, + "loss": 0.7728, + "step": 12120 + }, + { + "epoch": 1.0328930549637836, + "grad_norm": 16.5, + "learning_rate": 1.1935592507865534e-06, + "loss": 0.5396, + "step": 12121 + }, + { + "epoch": 1.0329782701320835, + "grad_norm": 11.5625, + "learning_rate": 1.193391387125694e-06, + "loss": 0.3635, + "step": 12122 + }, + { + "epoch": 1.0330634853003835, + "grad_norm": 12.875, + "learning_rate": 1.1932235244878093e-06, + "loss": 0.5104, + "step": 12123 + }, + { + "epoch": 1.0331487004686835, + "grad_norm": 14.625, + "learning_rate": 1.1930556628759326e-06, + "loss": 0.64, + "step": 12124 + }, + { + "epoch": 1.0332339156369834, + "grad_norm": 12.0625, + "learning_rate": 1.192887802293097e-06, + "loss": 0.5729, + "step": 12125 + }, + { + "epoch": 1.0333191308052834, + "grad_norm": 16.375, + "learning_rate": 1.1927199427423364e-06, + "loss": 0.777, + "step": 12126 + }, + { + "epoch": 1.0334043459735833, + "grad_norm": 18.75, + "learning_rate": 1.1925520842266839e-06, + "loss": 0.5323, + "step": 12127 + }, + { + "epoch": 1.0334895611418833, + "grad_norm": 16.875, + "learning_rate": 1.192384226749173e-06, + "loss": 0.9187, + "step": 12128 + }, + { + "epoch": 1.0335747763101832, + "grad_norm": 19.5, + "learning_rate": 1.192216370312837e-06, + "loss": 0.7986, + "step": 12129 + }, + { + "epoch": 1.0336599914784832, + "grad_norm": 24.0, + "learning_rate": 1.1920485149207098e-06, + "loss": 1.1853, + "step": 12130 + }, + { + "epoch": 1.0337452066467832, + "grad_norm": 11.1875, + "learning_rate": 1.1918806605758241e-06, + "loss": 0.4111, + "step": 12131 + }, + { + "epoch": 1.0338304218150831, + "grad_norm": 18.125, + "learning_rate": 1.191712807281213e-06, + "loss": 0.7398, + "step": 12132 + }, + { + "epoch": 1.033915636983383, + "grad_norm": 11.9375, + "learning_rate": 1.1915449550399101e-06, + "loss": 1.1856, + "step": 12133 + }, + { + "epoch": 1.034000852151683, + "grad_norm": 15.125, + "learning_rate": 1.1913771038549487e-06, + "loss": 0.6455, + "step": 12134 + }, + { + "epoch": 1.034086067319983, + "grad_norm": 17.75, + "learning_rate": 1.191209253729362e-06, + "loss": 0.5236, + "step": 12135 + }, + { + "epoch": 1.034171282488283, + "grad_norm": 14.8125, + "learning_rate": 1.1910414046661831e-06, + "loss": 0.7039, + "step": 12136 + }, + { + "epoch": 1.034256497656583, + "grad_norm": 18.375, + "learning_rate": 1.1908735566684455e-06, + "loss": 0.3722, + "step": 12137 + }, + { + "epoch": 1.0343417128248829, + "grad_norm": 18.25, + "learning_rate": 1.1907057097391819e-06, + "loss": 0.441, + "step": 12138 + }, + { + "epoch": 1.0344269279931828, + "grad_norm": 16.625, + "learning_rate": 1.1905378638814266e-06, + "loss": 0.5087, + "step": 12139 + }, + { + "epoch": 1.0345121431614828, + "grad_norm": 15.0, + "learning_rate": 1.190370019098211e-06, + "loss": 0.7126, + "step": 12140 + }, + { + "epoch": 1.0345973583297827, + "grad_norm": 12.875, + "learning_rate": 1.1902021753925695e-06, + "loss": 0.3717, + "step": 12141 + }, + { + "epoch": 1.0346825734980827, + "grad_norm": 17.25, + "learning_rate": 1.1900343327675345e-06, + "loss": 0.8439, + "step": 12142 + }, + { + "epoch": 1.0347677886663826, + "grad_norm": 16.75, + "learning_rate": 1.1898664912261397e-06, + "loss": 0.7514, + "step": 12143 + }, + { + "epoch": 1.0348530038346826, + "grad_norm": 14.125, + "learning_rate": 1.1896986507714175e-06, + "loss": 0.5308, + "step": 12144 + }, + { + "epoch": 1.0349382190029826, + "grad_norm": 14.25, + "learning_rate": 1.1895308114064018e-06, + "loss": 0.6424, + "step": 12145 + }, + { + "epoch": 1.0350234341712825, + "grad_norm": 15.6875, + "learning_rate": 1.1893629731341247e-06, + "loss": 0.6105, + "step": 12146 + }, + { + "epoch": 1.0351086493395825, + "grad_norm": 15.4375, + "learning_rate": 1.1891951359576203e-06, + "loss": 0.4787, + "step": 12147 + }, + { + "epoch": 1.0351938645078824, + "grad_norm": 12.375, + "learning_rate": 1.1890272998799205e-06, + "loss": 0.5066, + "step": 12148 + }, + { + "epoch": 1.0352790796761824, + "grad_norm": 14.75, + "learning_rate": 1.1888594649040585e-06, + "loss": 0.5076, + "step": 12149 + }, + { + "epoch": 1.0353642948444823, + "grad_norm": 19.125, + "learning_rate": 1.188691631033068e-06, + "loss": 0.9105, + "step": 12150 + }, + { + "epoch": 1.0354495100127823, + "grad_norm": 13.5625, + "learning_rate": 1.1885237982699808e-06, + "loss": 0.3469, + "step": 12151 + }, + { + "epoch": 1.0355347251810822, + "grad_norm": 14.25, + "learning_rate": 1.1883559666178308e-06, + "loss": 0.6317, + "step": 12152 + }, + { + "epoch": 1.0356199403493822, + "grad_norm": 12.875, + "learning_rate": 1.18818813607965e-06, + "loss": 0.5086, + "step": 12153 + }, + { + "epoch": 1.0357051555176822, + "grad_norm": 13.5, + "learning_rate": 1.188020306658472e-06, + "loss": 0.7064, + "step": 12154 + }, + { + "epoch": 1.0357903706859821, + "grad_norm": 17.5, + "learning_rate": 1.1878524783573298e-06, + "loss": 0.7342, + "step": 12155 + }, + { + "epoch": 1.035875585854282, + "grad_norm": 13.375, + "learning_rate": 1.1876846511792557e-06, + "loss": 0.548, + "step": 12156 + }, + { + "epoch": 1.035960801022582, + "grad_norm": 16.125, + "learning_rate": 1.1875168251272822e-06, + "loss": 0.8826, + "step": 12157 + }, + { + "epoch": 1.036046016190882, + "grad_norm": 15.75, + "learning_rate": 1.1873490002044428e-06, + "loss": 0.6361, + "step": 12158 + }, + { + "epoch": 1.036131231359182, + "grad_norm": 13.8125, + "learning_rate": 1.1871811764137694e-06, + "loss": 0.6078, + "step": 12159 + }, + { + "epoch": 1.036216446527482, + "grad_norm": 20.5, + "learning_rate": 1.187013353758296e-06, + "loss": 0.7864, + "step": 12160 + }, + { + "epoch": 1.0363016616957819, + "grad_norm": 16.125, + "learning_rate": 1.1868455322410542e-06, + "loss": 0.8146, + "step": 12161 + }, + { + "epoch": 1.0363868768640818, + "grad_norm": 11.625, + "learning_rate": 1.1866777118650772e-06, + "loss": 0.3335, + "step": 12162 + }, + { + "epoch": 1.0364720920323818, + "grad_norm": 12.0, + "learning_rate": 1.1865098926333982e-06, + "loss": 0.5209, + "step": 12163 + }, + { + "epoch": 1.0365573072006817, + "grad_norm": 11.375, + "learning_rate": 1.1863420745490486e-06, + "loss": 0.3578, + "step": 12164 + }, + { + "epoch": 1.0366425223689817, + "grad_norm": 17.5, + "learning_rate": 1.186174257615062e-06, + "loss": 0.7336, + "step": 12165 + }, + { + "epoch": 1.0367277375372816, + "grad_norm": 16.5, + "learning_rate": 1.1860064418344705e-06, + "loss": 0.8736, + "step": 12166 + }, + { + "epoch": 1.0368129527055816, + "grad_norm": 13.0625, + "learning_rate": 1.1858386272103072e-06, + "loss": 0.5575, + "step": 12167 + }, + { + "epoch": 1.0368981678738816, + "grad_norm": 12.875, + "learning_rate": 1.1856708137456042e-06, + "loss": 0.6453, + "step": 12168 + }, + { + "epoch": 1.0369833830421815, + "grad_norm": 14.25, + "learning_rate": 1.1855030014433945e-06, + "loss": 0.7769, + "step": 12169 + }, + { + "epoch": 1.0370685982104815, + "grad_norm": 16.375, + "learning_rate": 1.1853351903067103e-06, + "loss": 0.6374, + "step": 12170 + }, + { + "epoch": 1.0371538133787814, + "grad_norm": 12.9375, + "learning_rate": 1.1851673803385848e-06, + "loss": 0.3573, + "step": 12171 + }, + { + "epoch": 1.0372390285470814, + "grad_norm": 20.25, + "learning_rate": 1.1849995715420493e-06, + "loss": 0.6214, + "step": 12172 + }, + { + "epoch": 1.0373242437153813, + "grad_norm": 12.625, + "learning_rate": 1.1848317639201372e-06, + "loss": 0.3044, + "step": 12173 + }, + { + "epoch": 1.0374094588836813, + "grad_norm": 13.0, + "learning_rate": 1.1846639574758803e-06, + "loss": 0.6743, + "step": 12174 + }, + { + "epoch": 1.0374946740519813, + "grad_norm": 13.9375, + "learning_rate": 1.1844961522123116e-06, + "loss": 0.5411, + "step": 12175 + }, + { + "epoch": 1.0375798892202812, + "grad_norm": 15.625, + "learning_rate": 1.1843283481324632e-06, + "loss": 0.81, + "step": 12176 + }, + { + "epoch": 1.0376651043885812, + "grad_norm": 12.1875, + "learning_rate": 1.1841605452393676e-06, + "loss": 0.4303, + "step": 12177 + }, + { + "epoch": 1.0377503195568811, + "grad_norm": 14.6875, + "learning_rate": 1.1839927435360571e-06, + "loss": 0.6501, + "step": 12178 + }, + { + "epoch": 1.037835534725181, + "grad_norm": 15.6875, + "learning_rate": 1.1838249430255647e-06, + "loss": 0.38, + "step": 12179 + }, + { + "epoch": 1.037920749893481, + "grad_norm": 13.125, + "learning_rate": 1.1836571437109216e-06, + "loss": 0.5126, + "step": 12180 + }, + { + "epoch": 1.038005965061781, + "grad_norm": 20.125, + "learning_rate": 1.1834893455951605e-06, + "loss": 0.7334, + "step": 12181 + }, + { + "epoch": 1.038091180230081, + "grad_norm": 17.25, + "learning_rate": 1.1833215486813141e-06, + "loss": 0.6822, + "step": 12182 + }, + { + "epoch": 1.038176395398381, + "grad_norm": 11.6875, + "learning_rate": 1.183153752972414e-06, + "loss": 0.279, + "step": 12183 + }, + { + "epoch": 1.0382616105666809, + "grad_norm": 16.625, + "learning_rate": 1.1829859584714933e-06, + "loss": 0.7268, + "step": 12184 + }, + { + "epoch": 1.0383468257349808, + "grad_norm": 21.0, + "learning_rate": 1.1828181651815832e-06, + "loss": 0.9824, + "step": 12185 + }, + { + "epoch": 1.0384320409032808, + "grad_norm": 15.1875, + "learning_rate": 1.1826503731057167e-06, + "loss": 0.6139, + "step": 12186 + }, + { + "epoch": 1.0385172560715807, + "grad_norm": 20.125, + "learning_rate": 1.182482582246926e-06, + "loss": 0.6879, + "step": 12187 + }, + { + "epoch": 1.0386024712398807, + "grad_norm": 14.5625, + "learning_rate": 1.1823147926082427e-06, + "loss": 0.5429, + "step": 12188 + }, + { + "epoch": 1.0386876864081807, + "grad_norm": 13.4375, + "learning_rate": 1.182147004192699e-06, + "loss": 0.402, + "step": 12189 + }, + { + "epoch": 1.0387729015764806, + "grad_norm": 17.25, + "learning_rate": 1.1819792170033273e-06, + "loss": 0.8406, + "step": 12190 + }, + { + "epoch": 1.0388581167447806, + "grad_norm": 13.0625, + "learning_rate": 1.1818114310431592e-06, + "loss": 0.5648, + "step": 12191 + }, + { + "epoch": 1.0389433319130805, + "grad_norm": 11.3125, + "learning_rate": 1.1816436463152275e-06, + "loss": 0.3962, + "step": 12192 + }, + { + "epoch": 1.0390285470813805, + "grad_norm": 19.875, + "learning_rate": 1.1814758628225636e-06, + "loss": 0.6288, + "step": 12193 + }, + { + "epoch": 1.0391137622496804, + "grad_norm": 12.0625, + "learning_rate": 1.1813080805682e-06, + "loss": 0.4618, + "step": 12194 + }, + { + "epoch": 1.0391989774179804, + "grad_norm": 15.125, + "learning_rate": 1.181140299555168e-06, + "loss": 0.3864, + "step": 12195 + }, + { + "epoch": 1.0392841925862804, + "grad_norm": 21.125, + "learning_rate": 1.180972519786501e-06, + "loss": 0.9002, + "step": 12196 + }, + { + "epoch": 1.0393694077545803, + "grad_norm": 21.25, + "learning_rate": 1.1808047412652292e-06, + "loss": 1.0315, + "step": 12197 + }, + { + "epoch": 1.0394546229228803, + "grad_norm": 11.0625, + "learning_rate": 1.1806369639943855e-06, + "loss": 0.3812, + "step": 12198 + }, + { + "epoch": 1.0395398380911802, + "grad_norm": 14.5, + "learning_rate": 1.1804691879770016e-06, + "loss": 0.7292, + "step": 12199 + }, + { + "epoch": 1.0396250532594802, + "grad_norm": 15.4375, + "learning_rate": 1.1803014132161091e-06, + "loss": 0.8382, + "step": 12200 + }, + { + "epoch": 1.0397102684277801, + "grad_norm": 16.25, + "learning_rate": 1.1801336397147407e-06, + "loss": 0.5333, + "step": 12201 + }, + { + "epoch": 1.03979548359608, + "grad_norm": 14.4375, + "learning_rate": 1.1799658674759274e-06, + "loss": 0.5389, + "step": 12202 + }, + { + "epoch": 1.03988069876438, + "grad_norm": 13.4375, + "learning_rate": 1.1797980965027011e-06, + "loss": 0.4826, + "step": 12203 + }, + { + "epoch": 1.03996591393268, + "grad_norm": 25.625, + "learning_rate": 1.1796303267980943e-06, + "loss": 0.9351, + "step": 12204 + }, + { + "epoch": 1.04005112910098, + "grad_norm": 18.125, + "learning_rate": 1.1794625583651381e-06, + "loss": 0.4966, + "step": 12205 + }, + { + "epoch": 1.04013634426928, + "grad_norm": 12.9375, + "learning_rate": 1.179294791206864e-06, + "loss": 0.6054, + "step": 12206 + }, + { + "epoch": 1.0402215594375799, + "grad_norm": 12.6875, + "learning_rate": 1.1791270253263046e-06, + "loss": 0.4565, + "step": 12207 + }, + { + "epoch": 1.0403067746058798, + "grad_norm": 15.5, + "learning_rate": 1.1789592607264907e-06, + "loss": 0.702, + "step": 12208 + }, + { + "epoch": 1.0403919897741798, + "grad_norm": 16.25, + "learning_rate": 1.1787914974104546e-06, + "loss": 0.5439, + "step": 12209 + }, + { + "epoch": 1.0404772049424797, + "grad_norm": 12.4375, + "learning_rate": 1.1786237353812276e-06, + "loss": 0.3849, + "step": 12210 + }, + { + "epoch": 1.0405624201107797, + "grad_norm": 14.5625, + "learning_rate": 1.1784559746418417e-06, + "loss": 0.6931, + "step": 12211 + }, + { + "epoch": 1.0406476352790797, + "grad_norm": 17.75, + "learning_rate": 1.1782882151953285e-06, + "loss": 0.5247, + "step": 12212 + }, + { + "epoch": 1.0407328504473796, + "grad_norm": 17.75, + "learning_rate": 1.1781204570447189e-06, + "loss": 0.7136, + "step": 12213 + }, + { + "epoch": 1.0408180656156796, + "grad_norm": 14.4375, + "learning_rate": 1.177952700193045e-06, + "loss": 0.5578, + "step": 12214 + }, + { + "epoch": 1.0409032807839795, + "grad_norm": 14.125, + "learning_rate": 1.1777849446433382e-06, + "loss": 0.4661, + "step": 12215 + }, + { + "epoch": 1.0409884959522795, + "grad_norm": 17.25, + "learning_rate": 1.1776171903986305e-06, + "loss": 0.6947, + "step": 12216 + }, + { + "epoch": 1.0410737111205794, + "grad_norm": 14.0, + "learning_rate": 1.1774494374619525e-06, + "loss": 0.5227, + "step": 12217 + }, + { + "epoch": 1.0411589262888794, + "grad_norm": 14.8125, + "learning_rate": 1.1772816858363363e-06, + "loss": 0.3847, + "step": 12218 + }, + { + "epoch": 1.0412441414571794, + "grad_norm": 13.4375, + "learning_rate": 1.1771139355248131e-06, + "loss": 0.4344, + "step": 12219 + }, + { + "epoch": 1.0413293566254793, + "grad_norm": 16.375, + "learning_rate": 1.176946186530415e-06, + "loss": 0.9283, + "step": 12220 + }, + { + "epoch": 1.0414145717937793, + "grad_norm": 13.8125, + "learning_rate": 1.1767784388561718e-06, + "loss": 0.5963, + "step": 12221 + }, + { + "epoch": 1.0414997869620792, + "grad_norm": 19.75, + "learning_rate": 1.1766106925051164e-06, + "loss": 0.8079, + "step": 12222 + }, + { + "epoch": 1.0415850021303792, + "grad_norm": 16.5, + "learning_rate": 1.1764429474802792e-06, + "loss": 0.6416, + "step": 12223 + }, + { + "epoch": 1.0416702172986791, + "grad_norm": 15.9375, + "learning_rate": 1.1762752037846922e-06, + "loss": 0.615, + "step": 12224 + }, + { + "epoch": 1.041755432466979, + "grad_norm": 16.5, + "learning_rate": 1.176107461421386e-06, + "loss": 0.5623, + "step": 12225 + }, + { + "epoch": 1.041840647635279, + "grad_norm": 15.125, + "learning_rate": 1.175939720393393e-06, + "loss": 0.7099, + "step": 12226 + }, + { + "epoch": 1.041925862803579, + "grad_norm": 13.125, + "learning_rate": 1.1757719807037432e-06, + "loss": 0.4801, + "step": 12227 + }, + { + "epoch": 1.042011077971879, + "grad_norm": 14.5625, + "learning_rate": 1.1756042423554688e-06, + "loss": 0.4208, + "step": 12228 + }, + { + "epoch": 1.042096293140179, + "grad_norm": 13.125, + "learning_rate": 1.1754365053516002e-06, + "loss": 0.2615, + "step": 12229 + }, + { + "epoch": 1.0421815083084789, + "grad_norm": 15.1875, + "learning_rate": 1.1752687696951687e-06, + "loss": 0.5503, + "step": 12230 + }, + { + "epoch": 1.0422667234767788, + "grad_norm": 10.9375, + "learning_rate": 1.1751010353892062e-06, + "loss": 0.3954, + "step": 12231 + }, + { + "epoch": 1.0423519386450788, + "grad_norm": 11.5625, + "learning_rate": 1.1749333024367427e-06, + "loss": 0.4969, + "step": 12232 + }, + { + "epoch": 1.0424371538133788, + "grad_norm": 13.625, + "learning_rate": 1.1747655708408104e-06, + "loss": 0.6721, + "step": 12233 + }, + { + "epoch": 1.0425223689816787, + "grad_norm": 13.6875, + "learning_rate": 1.17459784060444e-06, + "loss": 0.6874, + "step": 12234 + }, + { + "epoch": 1.0426075841499787, + "grad_norm": 9.3125, + "learning_rate": 1.174430111730662e-06, + "loss": 0.3838, + "step": 12235 + }, + { + "epoch": 1.0426927993182786, + "grad_norm": 14.3125, + "learning_rate": 1.1742623842225082e-06, + "loss": 0.4641, + "step": 12236 + }, + { + "epoch": 1.0427780144865786, + "grad_norm": 14.75, + "learning_rate": 1.1740946580830094e-06, + "loss": 0.5698, + "step": 12237 + }, + { + "epoch": 1.0428632296548785, + "grad_norm": 15.8125, + "learning_rate": 1.173926933315196e-06, + "loss": 0.5265, + "step": 12238 + }, + { + "epoch": 1.0429484448231785, + "grad_norm": 19.75, + "learning_rate": 1.1737592099220998e-06, + "loss": 0.6866, + "step": 12239 + }, + { + "epoch": 1.0430336599914785, + "grad_norm": 11.5, + "learning_rate": 1.173591487906751e-06, + "loss": 0.3103, + "step": 12240 + }, + { + "epoch": 1.0431188751597784, + "grad_norm": 14.6875, + "learning_rate": 1.173423767272181e-06, + "loss": 0.5651, + "step": 12241 + }, + { + "epoch": 1.0432040903280784, + "grad_norm": 38.25, + "learning_rate": 1.1732560480214203e-06, + "loss": 0.9726, + "step": 12242 + }, + { + "epoch": 1.0432893054963783, + "grad_norm": 17.75, + "learning_rate": 1.1730883301575003e-06, + "loss": 0.5436, + "step": 12243 + }, + { + "epoch": 1.0433745206646783, + "grad_norm": 13.5625, + "learning_rate": 1.1729206136834518e-06, + "loss": 0.639, + "step": 12244 + }, + { + "epoch": 1.0434597358329782, + "grad_norm": 17.375, + "learning_rate": 1.1727528986023046e-06, + "loss": 0.4507, + "step": 12245 + }, + { + "epoch": 1.0435449510012782, + "grad_norm": 14.6875, + "learning_rate": 1.1725851849170909e-06, + "loss": 0.6545, + "step": 12246 + }, + { + "epoch": 1.0436301661695782, + "grad_norm": 11.25, + "learning_rate": 1.1724174726308401e-06, + "loss": 0.2853, + "step": 12247 + }, + { + "epoch": 1.0437153813378781, + "grad_norm": 15.9375, + "learning_rate": 1.1722497617465841e-06, + "loss": 0.6409, + "step": 12248 + }, + { + "epoch": 1.043800596506178, + "grad_norm": 12.875, + "learning_rate": 1.172082052267353e-06, + "loss": 0.5506, + "step": 12249 + }, + { + "epoch": 1.043885811674478, + "grad_norm": 16.0, + "learning_rate": 1.1719143441961772e-06, + "loss": 0.654, + "step": 12250 + }, + { + "epoch": 1.043971026842778, + "grad_norm": 17.25, + "learning_rate": 1.171746637536088e-06, + "loss": 0.6704, + "step": 12251 + }, + { + "epoch": 1.044056242011078, + "grad_norm": 13.875, + "learning_rate": 1.1715789322901161e-06, + "loss": 0.5007, + "step": 12252 + }, + { + "epoch": 1.044141457179378, + "grad_norm": 15.625, + "learning_rate": 1.1714112284612912e-06, + "loss": 0.4469, + "step": 12253 + }, + { + "epoch": 1.0442266723476779, + "grad_norm": 17.375, + "learning_rate": 1.1712435260526446e-06, + "loss": 0.8604, + "step": 12254 + }, + { + "epoch": 1.0443118875159778, + "grad_norm": 12.6875, + "learning_rate": 1.1710758250672065e-06, + "loss": 0.4008, + "step": 12255 + }, + { + "epoch": 1.0443971026842778, + "grad_norm": 28.75, + "learning_rate": 1.1709081255080078e-06, + "loss": 0.674, + "step": 12256 + }, + { + "epoch": 1.0444823178525777, + "grad_norm": 20.125, + "learning_rate": 1.1707404273780787e-06, + "loss": 1.0972, + "step": 12257 + }, + { + "epoch": 1.0445675330208777, + "grad_norm": 16.75, + "learning_rate": 1.17057273068045e-06, + "loss": 0.6659, + "step": 12258 + }, + { + "epoch": 1.0446527481891776, + "grad_norm": 18.25, + "learning_rate": 1.1704050354181516e-06, + "loss": 0.8329, + "step": 12259 + }, + { + "epoch": 1.0447379633574776, + "grad_norm": 16.25, + "learning_rate": 1.1702373415942144e-06, + "loss": 0.6349, + "step": 12260 + }, + { + "epoch": 1.0448231785257776, + "grad_norm": 18.875, + "learning_rate": 1.1700696492116692e-06, + "loss": 0.837, + "step": 12261 + }, + { + "epoch": 1.0449083936940775, + "grad_norm": 14.8125, + "learning_rate": 1.169901958273545e-06, + "loss": 0.707, + "step": 12262 + }, + { + "epoch": 1.0449936088623775, + "grad_norm": 10.8125, + "learning_rate": 1.1697342687828734e-06, + "loss": 0.2384, + "step": 12263 + }, + { + "epoch": 1.0450788240306774, + "grad_norm": 13.5625, + "learning_rate": 1.1695665807426843e-06, + "loss": 0.484, + "step": 12264 + }, + { + "epoch": 1.0451640391989774, + "grad_norm": 20.75, + "learning_rate": 1.1693988941560076e-06, + "loss": 0.542, + "step": 12265 + }, + { + "epoch": 1.0452492543672773, + "grad_norm": 16.375, + "learning_rate": 1.1692312090258743e-06, + "loss": 0.9995, + "step": 12266 + }, + { + "epoch": 1.0453344695355773, + "grad_norm": 17.125, + "learning_rate": 1.169063525355314e-06, + "loss": 0.7402, + "step": 12267 + }, + { + "epoch": 1.0454196847038772, + "grad_norm": 13.125, + "learning_rate": 1.1688958431473573e-06, + "loss": 0.5055, + "step": 12268 + }, + { + "epoch": 1.0455048998721772, + "grad_norm": 12.6875, + "learning_rate": 1.1687281624050344e-06, + "loss": 0.4756, + "step": 12269 + }, + { + "epoch": 1.0455901150404772, + "grad_norm": 10.5625, + "learning_rate": 1.168560483131375e-06, + "loss": 0.4452, + "step": 12270 + }, + { + "epoch": 1.0456753302087771, + "grad_norm": 20.75, + "learning_rate": 1.16839280532941e-06, + "loss": 0.7191, + "step": 12271 + }, + { + "epoch": 1.045760545377077, + "grad_norm": 18.75, + "learning_rate": 1.1682251290021687e-06, + "loss": 0.9346, + "step": 12272 + }, + { + "epoch": 1.045845760545377, + "grad_norm": 12.0625, + "learning_rate": 1.1680574541526818e-06, + "loss": 0.3446, + "step": 12273 + }, + { + "epoch": 1.045930975713677, + "grad_norm": 11.625, + "learning_rate": 1.1678897807839786e-06, + "loss": 0.3199, + "step": 12274 + }, + { + "epoch": 1.046016190881977, + "grad_norm": 21.0, + "learning_rate": 1.1677221088990901e-06, + "loss": 0.7114, + "step": 12275 + }, + { + "epoch": 1.046101406050277, + "grad_norm": 15.0625, + "learning_rate": 1.1675544385010456e-06, + "loss": 0.5859, + "step": 12276 + }, + { + "epoch": 1.0461866212185769, + "grad_norm": 15.1875, + "learning_rate": 1.1673867695928758e-06, + "loss": 0.5342, + "step": 12277 + }, + { + "epoch": 1.0462718363868768, + "grad_norm": 18.25, + "learning_rate": 1.16721910217761e-06, + "loss": 0.8177, + "step": 12278 + }, + { + "epoch": 1.0463570515551768, + "grad_norm": 11.375, + "learning_rate": 1.1670514362582782e-06, + "loss": 0.5343, + "step": 12279 + }, + { + "epoch": 1.0464422667234767, + "grad_norm": 15.6875, + "learning_rate": 1.16688377183791e-06, + "loss": 0.8078, + "step": 12280 + }, + { + "epoch": 1.0465274818917767, + "grad_norm": 17.125, + "learning_rate": 1.1667161089195358e-06, + "loss": 0.6942, + "step": 12281 + }, + { + "epoch": 1.0466126970600766, + "grad_norm": 18.375, + "learning_rate": 1.1665484475061853e-06, + "loss": 0.6339, + "step": 12282 + }, + { + "epoch": 1.0466979122283766, + "grad_norm": 14.3125, + "learning_rate": 1.1663807876008882e-06, + "loss": 0.6903, + "step": 12283 + }, + { + "epoch": 1.0467831273966766, + "grad_norm": 12.875, + "learning_rate": 1.1662131292066744e-06, + "loss": 0.4029, + "step": 12284 + }, + { + "epoch": 1.0468683425649765, + "grad_norm": 16.5, + "learning_rate": 1.166045472326574e-06, + "loss": 0.718, + "step": 12285 + }, + { + "epoch": 1.0469535577332765, + "grad_norm": 16.875, + "learning_rate": 1.165877816963616e-06, + "loss": 0.6453, + "step": 12286 + }, + { + "epoch": 1.0470387729015764, + "grad_norm": 17.0, + "learning_rate": 1.1657101631208301e-06, + "loss": 0.7366, + "step": 12287 + }, + { + "epoch": 1.0471239880698764, + "grad_norm": 18.25, + "learning_rate": 1.1655425108012467e-06, + "loss": 0.736, + "step": 12288 + }, + { + "epoch": 1.0472092032381763, + "grad_norm": 18.875, + "learning_rate": 1.165374860007895e-06, + "loss": 1.0333, + "step": 12289 + }, + { + "epoch": 1.0472944184064763, + "grad_norm": 14.625, + "learning_rate": 1.1652072107438046e-06, + "loss": 0.5516, + "step": 12290 + }, + { + "epoch": 1.0473796335747763, + "grad_norm": 25.625, + "learning_rate": 1.1650395630120051e-06, + "loss": 1.1468, + "step": 12291 + }, + { + "epoch": 1.0474648487430762, + "grad_norm": 19.875, + "learning_rate": 1.1648719168155263e-06, + "loss": 0.8665, + "step": 12292 + }, + { + "epoch": 1.0475500639113762, + "grad_norm": 15.1875, + "learning_rate": 1.164704272157398e-06, + "loss": 0.5032, + "step": 12293 + }, + { + "epoch": 1.0476352790796761, + "grad_norm": 21.625, + "learning_rate": 1.164536629040649e-06, + "loss": 0.7211, + "step": 12294 + }, + { + "epoch": 1.047720494247976, + "grad_norm": 15.75, + "learning_rate": 1.1643689874683085e-06, + "loss": 0.8516, + "step": 12295 + }, + { + "epoch": 1.047805709416276, + "grad_norm": 17.375, + "learning_rate": 1.1642013474434068e-06, + "loss": 0.7846, + "step": 12296 + }, + { + "epoch": 1.047890924584576, + "grad_norm": 20.125, + "learning_rate": 1.1640337089689732e-06, + "loss": 0.5703, + "step": 12297 + }, + { + "epoch": 1.047976139752876, + "grad_norm": 13.5625, + "learning_rate": 1.1638660720480367e-06, + "loss": 0.6434, + "step": 12298 + }, + { + "epoch": 1.048061354921176, + "grad_norm": 12.5, + "learning_rate": 1.1636984366836268e-06, + "loss": 0.4529, + "step": 12299 + }, + { + "epoch": 1.0481465700894759, + "grad_norm": 16.625, + "learning_rate": 1.1635308028787733e-06, + "loss": 0.4214, + "step": 12300 + }, + { + "epoch": 1.0482317852577758, + "grad_norm": 11.5, + "learning_rate": 1.1633631706365054e-06, + "loss": 0.3778, + "step": 12301 + }, + { + "epoch": 1.0483170004260758, + "grad_norm": 21.125, + "learning_rate": 1.1631955399598513e-06, + "loss": 0.7536, + "step": 12302 + }, + { + "epoch": 1.0484022155943757, + "grad_norm": 20.0, + "learning_rate": 1.1630279108518417e-06, + "loss": 0.8039, + "step": 12303 + }, + { + "epoch": 1.0484874307626757, + "grad_norm": 19.75, + "learning_rate": 1.1628602833155047e-06, + "loss": 0.8201, + "step": 12304 + }, + { + "epoch": 1.0485726459309757, + "grad_norm": 14.9375, + "learning_rate": 1.1626926573538705e-06, + "loss": 0.6493, + "step": 12305 + }, + { + "epoch": 1.0486578610992756, + "grad_norm": 15.8125, + "learning_rate": 1.1625250329699673e-06, + "loss": 0.5935, + "step": 12306 + }, + { + "epoch": 1.0487430762675756, + "grad_norm": 16.625, + "learning_rate": 1.1623574101668252e-06, + "loss": 0.6973, + "step": 12307 + }, + { + "epoch": 1.0488282914358755, + "grad_norm": 14.5, + "learning_rate": 1.1621897889474725e-06, + "loss": 0.42, + "step": 12308 + }, + { + "epoch": 1.0489135066041755, + "grad_norm": 21.25, + "learning_rate": 1.1620221693149392e-06, + "loss": 0.8247, + "step": 12309 + }, + { + "epoch": 1.0489987217724754, + "grad_norm": 11.875, + "learning_rate": 1.1618545512722532e-06, + "loss": 0.4533, + "step": 12310 + }, + { + "epoch": 1.0490839369407754, + "grad_norm": 19.625, + "learning_rate": 1.1616869348224443e-06, + "loss": 0.8493, + "step": 12311 + }, + { + "epoch": 1.0491691521090754, + "grad_norm": 9.5625, + "learning_rate": 1.161519319968541e-06, + "loss": 0.1809, + "step": 12312 + }, + { + "epoch": 1.0492543672773753, + "grad_norm": 20.375, + "learning_rate": 1.161351706713573e-06, + "loss": 0.841, + "step": 12313 + }, + { + "epoch": 1.0493395824456753, + "grad_norm": 15.3125, + "learning_rate": 1.1611840950605682e-06, + "loss": 0.391, + "step": 12314 + }, + { + "epoch": 1.0494247976139752, + "grad_norm": 19.125, + "learning_rate": 1.1610164850125566e-06, + "loss": 0.7404, + "step": 12315 + }, + { + "epoch": 1.0495100127822752, + "grad_norm": 20.5, + "learning_rate": 1.1608488765725661e-06, + "loss": 0.9264, + "step": 12316 + }, + { + "epoch": 1.0495952279505751, + "grad_norm": 20.375, + "learning_rate": 1.1606812697436268e-06, + "loss": 0.6435, + "step": 12317 + }, + { + "epoch": 1.049680443118875, + "grad_norm": 9.75, + "learning_rate": 1.1605136645287666e-06, + "loss": 0.3014, + "step": 12318 + }, + { + "epoch": 1.049765658287175, + "grad_norm": 14.8125, + "learning_rate": 1.160346060931014e-06, + "loss": 0.6051, + "step": 12319 + }, + { + "epoch": 1.049850873455475, + "grad_norm": 12.4375, + "learning_rate": 1.1601784589533986e-06, + "loss": 0.4976, + "step": 12320 + }, + { + "epoch": 1.049936088623775, + "grad_norm": 15.8125, + "learning_rate": 1.1600108585989485e-06, + "loss": 0.8374, + "step": 12321 + }, + { + "epoch": 1.050021303792075, + "grad_norm": 17.75, + "learning_rate": 1.159843259870693e-06, + "loss": 0.7385, + "step": 12322 + }, + { + "epoch": 1.0501065189603749, + "grad_norm": 14.125, + "learning_rate": 1.15967566277166e-06, + "loss": 0.8023, + "step": 12323 + }, + { + "epoch": 1.0501917341286748, + "grad_norm": 16.0, + "learning_rate": 1.1595080673048791e-06, + "loss": 0.6686, + "step": 12324 + }, + { + "epoch": 1.0502769492969748, + "grad_norm": 15.8125, + "learning_rate": 1.1593404734733784e-06, + "loss": 0.3566, + "step": 12325 + }, + { + "epoch": 1.0503621644652748, + "grad_norm": 13.75, + "learning_rate": 1.1591728812801865e-06, + "loss": 0.3083, + "step": 12326 + }, + { + "epoch": 1.0504473796335747, + "grad_norm": 28.5, + "learning_rate": 1.1590052907283316e-06, + "loss": 0.6044, + "step": 12327 + }, + { + "epoch": 1.0505325948018747, + "grad_norm": 16.25, + "learning_rate": 1.1588377018208428e-06, + "loss": 0.5091, + "step": 12328 + }, + { + "epoch": 1.0506178099701746, + "grad_norm": 21.0, + "learning_rate": 1.1586701145607483e-06, + "loss": 1.0696, + "step": 12329 + }, + { + "epoch": 1.0507030251384746, + "grad_norm": 15.6875, + "learning_rate": 1.1585025289510768e-06, + "loss": 0.63, + "step": 12330 + }, + { + "epoch": 1.0507882403067745, + "grad_norm": 14.625, + "learning_rate": 1.1583349449948564e-06, + "loss": 0.5416, + "step": 12331 + }, + { + "epoch": 1.0508734554750745, + "grad_norm": 20.625, + "learning_rate": 1.158167362695116e-06, + "loss": 0.7185, + "step": 12332 + }, + { + "epoch": 1.0509586706433744, + "grad_norm": 11.4375, + "learning_rate": 1.1579997820548835e-06, + "loss": 0.3981, + "step": 12333 + }, + { + "epoch": 1.0510438858116744, + "grad_norm": 16.375, + "learning_rate": 1.1578322030771878e-06, + "loss": 0.7342, + "step": 12334 + }, + { + "epoch": 1.0511291009799744, + "grad_norm": 17.875, + "learning_rate": 1.1576646257650567e-06, + "loss": 0.7714, + "step": 12335 + }, + { + "epoch": 1.0512143161482743, + "grad_norm": 12.0, + "learning_rate": 1.1574970501215183e-06, + "loss": 0.3909, + "step": 12336 + }, + { + "epoch": 1.0512995313165743, + "grad_norm": 13.75, + "learning_rate": 1.1573294761496016e-06, + "loss": 0.6491, + "step": 12337 + }, + { + "epoch": 1.0513847464848742, + "grad_norm": 19.0, + "learning_rate": 1.1571619038523342e-06, + "loss": 0.967, + "step": 12338 + }, + { + "epoch": 1.0514699616531742, + "grad_norm": 19.75, + "learning_rate": 1.156994333232745e-06, + "loss": 0.6683, + "step": 12339 + }, + { + "epoch": 1.0515551768214741, + "grad_norm": 48.25, + "learning_rate": 1.1568267642938611e-06, + "loss": 0.7705, + "step": 12340 + }, + { + "epoch": 1.051640391989774, + "grad_norm": 12.6875, + "learning_rate": 1.1566591970387117e-06, + "loss": 0.2951, + "step": 12341 + }, + { + "epoch": 1.051725607158074, + "grad_norm": 14.8125, + "learning_rate": 1.1564916314703247e-06, + "loss": 0.3566, + "step": 12342 + }, + { + "epoch": 1.051810822326374, + "grad_norm": 14.875, + "learning_rate": 1.1563240675917277e-06, + "loss": 0.7374, + "step": 12343 + }, + { + "epoch": 1.051896037494674, + "grad_norm": 14.8125, + "learning_rate": 1.1561565054059487e-06, + "loss": 0.807, + "step": 12344 + }, + { + "epoch": 1.051981252662974, + "grad_norm": 17.375, + "learning_rate": 1.1559889449160163e-06, + "loss": 0.7581, + "step": 12345 + }, + { + "epoch": 1.0520664678312739, + "grad_norm": 13.625, + "learning_rate": 1.155821386124958e-06, + "loss": 0.3939, + "step": 12346 + }, + { + "epoch": 1.0521516829995738, + "grad_norm": 19.375, + "learning_rate": 1.155653829035802e-06, + "loss": 0.8974, + "step": 12347 + }, + { + "epoch": 1.0522368981678738, + "grad_norm": 15.625, + "learning_rate": 1.1554862736515761e-06, + "loss": 0.6698, + "step": 12348 + }, + { + "epoch": 1.0523221133361738, + "grad_norm": 14.375, + "learning_rate": 1.1553187199753087e-06, + "loss": 0.6102, + "step": 12349 + }, + { + "epoch": 1.0524073285044737, + "grad_norm": 15.5, + "learning_rate": 1.1551511680100272e-06, + "loss": 0.3598, + "step": 12350 + }, + { + "epoch": 1.0524925436727737, + "grad_norm": 43.5, + "learning_rate": 1.1549836177587591e-06, + "loss": 0.8307, + "step": 12351 + }, + { + "epoch": 1.0525777588410736, + "grad_norm": 16.25, + "learning_rate": 1.1548160692245329e-06, + "loss": 0.4264, + "step": 12352 + }, + { + "epoch": 1.0526629740093736, + "grad_norm": 17.125, + "learning_rate": 1.1546485224103756e-06, + "loss": 0.6326, + "step": 12353 + }, + { + "epoch": 1.0527481891776735, + "grad_norm": 13.5625, + "learning_rate": 1.1544809773193158e-06, + "loss": 0.4014, + "step": 12354 + }, + { + "epoch": 1.0528334043459735, + "grad_norm": 19.125, + "learning_rate": 1.1543134339543804e-06, + "loss": 0.4165, + "step": 12355 + }, + { + "epoch": 1.0529186195142735, + "grad_norm": 17.75, + "learning_rate": 1.1541458923185978e-06, + "loss": 0.9075, + "step": 12356 + }, + { + "epoch": 1.0530038346825734, + "grad_norm": 11.5, + "learning_rate": 1.1539783524149952e-06, + "loss": 0.3407, + "step": 12357 + }, + { + "epoch": 1.0530890498508734, + "grad_norm": 18.125, + "learning_rate": 1.1538108142466009e-06, + "loss": 0.6585, + "step": 12358 + }, + { + "epoch": 1.0531742650191733, + "grad_norm": 14.4375, + "learning_rate": 1.1536432778164411e-06, + "loss": 0.647, + "step": 12359 + }, + { + "epoch": 1.0532594801874733, + "grad_norm": 14.375, + "learning_rate": 1.1534757431275446e-06, + "loss": 0.6065, + "step": 12360 + }, + { + "epoch": 1.0533446953557732, + "grad_norm": 16.0, + "learning_rate": 1.1533082101829381e-06, + "loss": 0.8037, + "step": 12361 + }, + { + "epoch": 1.0534299105240732, + "grad_norm": 15.5625, + "learning_rate": 1.1531406789856497e-06, + "loss": 0.7653, + "step": 12362 + }, + { + "epoch": 1.0535151256923732, + "grad_norm": 15.4375, + "learning_rate": 1.1529731495387065e-06, + "loss": 0.4392, + "step": 12363 + }, + { + "epoch": 1.0536003408606731, + "grad_norm": 11.5, + "learning_rate": 1.1528056218451362e-06, + "loss": 0.336, + "step": 12364 + }, + { + "epoch": 1.053685556028973, + "grad_norm": 14.6875, + "learning_rate": 1.1526380959079657e-06, + "loss": 0.5924, + "step": 12365 + }, + { + "epoch": 1.053770771197273, + "grad_norm": 18.0, + "learning_rate": 1.1524705717302236e-06, + "loss": 0.877, + "step": 12366 + }, + { + "epoch": 1.053855986365573, + "grad_norm": 13.5, + "learning_rate": 1.1523030493149359e-06, + "loss": 0.4274, + "step": 12367 + }, + { + "epoch": 1.053941201533873, + "grad_norm": 13.0625, + "learning_rate": 1.15213552866513e-06, + "loss": 0.6436, + "step": 12368 + }, + { + "epoch": 1.054026416702173, + "grad_norm": 14.4375, + "learning_rate": 1.1519680097838337e-06, + "loss": 0.667, + "step": 12369 + }, + { + "epoch": 1.0541116318704729, + "grad_norm": 15.9375, + "learning_rate": 1.151800492674074e-06, + "loss": 0.5561, + "step": 12370 + }, + { + "epoch": 1.0541968470387728, + "grad_norm": 13.3125, + "learning_rate": 1.1516329773388782e-06, + "loss": 0.4434, + "step": 12371 + }, + { + "epoch": 1.0542820622070728, + "grad_norm": 16.75, + "learning_rate": 1.1514654637812736e-06, + "loss": 0.7159, + "step": 12372 + }, + { + "epoch": 1.0543672773753727, + "grad_norm": 20.5, + "learning_rate": 1.1512979520042867e-06, + "loss": 0.654, + "step": 12373 + }, + { + "epoch": 1.0544524925436727, + "grad_norm": 20.25, + "learning_rate": 1.1511304420109458e-06, + "loss": 0.7388, + "step": 12374 + }, + { + "epoch": 1.0545377077119726, + "grad_norm": 13.25, + "learning_rate": 1.1509629338042768e-06, + "loss": 0.4155, + "step": 12375 + }, + { + "epoch": 1.0546229228802726, + "grad_norm": 15.0, + "learning_rate": 1.150795427387307e-06, + "loss": 0.7324, + "step": 12376 + }, + { + "epoch": 1.0547081380485726, + "grad_norm": 19.75, + "learning_rate": 1.1506279227630637e-06, + "loss": 0.7354, + "step": 12377 + }, + { + "epoch": 1.0547933532168725, + "grad_norm": 11.875, + "learning_rate": 1.1504604199345737e-06, + "loss": 0.4438, + "step": 12378 + }, + { + "epoch": 1.0548785683851725, + "grad_norm": 20.375, + "learning_rate": 1.1502929189048643e-06, + "loss": 0.552, + "step": 12379 + }, + { + "epoch": 1.0549637835534724, + "grad_norm": 12.25, + "learning_rate": 1.1501254196769616e-06, + "loss": 0.5927, + "step": 12380 + }, + { + "epoch": 1.0550489987217724, + "grad_norm": 15.5, + "learning_rate": 1.1499579222538935e-06, + "loss": 0.594, + "step": 12381 + }, + { + "epoch": 1.0551342138900723, + "grad_norm": 15.25, + "learning_rate": 1.1497904266386863e-06, + "loss": 0.4722, + "step": 12382 + }, + { + "epoch": 1.0552194290583723, + "grad_norm": 13.9375, + "learning_rate": 1.1496229328343666e-06, + "loss": 0.741, + "step": 12383 + }, + { + "epoch": 1.0553046442266723, + "grad_norm": 13.3125, + "learning_rate": 1.1494554408439617e-06, + "loss": 0.5476, + "step": 12384 + }, + { + "epoch": 1.0553898593949722, + "grad_norm": 15.0625, + "learning_rate": 1.1492879506704978e-06, + "loss": 0.3974, + "step": 12385 + }, + { + "epoch": 1.0554750745632722, + "grad_norm": 24.25, + "learning_rate": 1.1491204623170021e-06, + "loss": 0.9395, + "step": 12386 + }, + { + "epoch": 1.0555602897315721, + "grad_norm": 19.125, + "learning_rate": 1.1489529757865011e-06, + "loss": 0.6363, + "step": 12387 + }, + { + "epoch": 1.055645504899872, + "grad_norm": 19.25, + "learning_rate": 1.1487854910820213e-06, + "loss": 0.8712, + "step": 12388 + }, + { + "epoch": 1.055730720068172, + "grad_norm": 22.125, + "learning_rate": 1.1486180082065896e-06, + "loss": 0.7786, + "step": 12389 + }, + { + "epoch": 1.055815935236472, + "grad_norm": 18.125, + "learning_rate": 1.1484505271632327e-06, + "loss": 0.5247, + "step": 12390 + }, + { + "epoch": 1.055901150404772, + "grad_norm": 13.0625, + "learning_rate": 1.1482830479549764e-06, + "loss": 0.4275, + "step": 12391 + }, + { + "epoch": 1.055986365573072, + "grad_norm": 13.1875, + "learning_rate": 1.148115570584848e-06, + "loss": 0.4442, + "step": 12392 + }, + { + "epoch": 1.0560715807413719, + "grad_norm": 14.75, + "learning_rate": 1.1479480950558734e-06, + "loss": 0.7148, + "step": 12393 + }, + { + "epoch": 1.0561567959096718, + "grad_norm": 11.6875, + "learning_rate": 1.1477806213710796e-06, + "loss": 0.4649, + "step": 12394 + }, + { + "epoch": 1.0562420110779718, + "grad_norm": 13.4375, + "learning_rate": 1.1476131495334925e-06, + "loss": 0.524, + "step": 12395 + }, + { + "epoch": 1.0563272262462717, + "grad_norm": 17.5, + "learning_rate": 1.1474456795461391e-06, + "loss": 0.8097, + "step": 12396 + }, + { + "epoch": 1.0564124414145717, + "grad_norm": 25.0, + "learning_rate": 1.147278211412045e-06, + "loss": 1.0126, + "step": 12397 + }, + { + "epoch": 1.0564976565828716, + "grad_norm": 11.25, + "learning_rate": 1.147110745134237e-06, + "loss": 0.374, + "step": 12398 + }, + { + "epoch": 1.0565828717511716, + "grad_norm": 14.75, + "learning_rate": 1.146943280715742e-06, + "loss": 0.5695, + "step": 12399 + }, + { + "epoch": 1.0566680869194716, + "grad_norm": 14.0625, + "learning_rate": 1.1467758181595847e-06, + "loss": 0.4398, + "step": 12400 + }, + { + "epoch": 1.0567533020877715, + "grad_norm": 13.75, + "learning_rate": 1.1466083574687926e-06, + "loss": 0.4198, + "step": 12401 + }, + { + "epoch": 1.0568385172560715, + "grad_norm": 11.75, + "learning_rate": 1.1464408986463915e-06, + "loss": 0.4795, + "step": 12402 + }, + { + "epoch": 1.0569237324243714, + "grad_norm": 19.0, + "learning_rate": 1.1462734416954071e-06, + "loss": 0.5306, + "step": 12403 + }, + { + "epoch": 1.0570089475926714, + "grad_norm": 21.75, + "learning_rate": 1.146105986618866e-06, + "loss": 0.819, + "step": 12404 + }, + { + "epoch": 1.0570941627609716, + "grad_norm": 10.625, + "learning_rate": 1.1459385334197942e-06, + "loss": 0.3365, + "step": 12405 + }, + { + "epoch": 1.0571793779292715, + "grad_norm": 16.625, + "learning_rate": 1.145771082101218e-06, + "loss": 0.6668, + "step": 12406 + }, + { + "epoch": 1.0572645930975715, + "grad_norm": 14.3125, + "learning_rate": 1.1456036326661634e-06, + "loss": 0.7222, + "step": 12407 + }, + { + "epoch": 1.0573498082658714, + "grad_norm": 12.6875, + "learning_rate": 1.1454361851176554e-06, + "loss": 0.4683, + "step": 12408 + }, + { + "epoch": 1.0574350234341714, + "grad_norm": 21.375, + "learning_rate": 1.1452687394587211e-06, + "loss": 0.7266, + "step": 12409 + }, + { + "epoch": 1.0575202386024714, + "grad_norm": 22.875, + "learning_rate": 1.1451012956923857e-06, + "loss": 1.0329, + "step": 12410 + }, + { + "epoch": 1.0576054537707713, + "grad_norm": 11.5, + "learning_rate": 1.1449338538216757e-06, + "loss": 0.3147, + "step": 12411 + }, + { + "epoch": 1.0576906689390713, + "grad_norm": 12.625, + "learning_rate": 1.1447664138496162e-06, + "loss": 0.4077, + "step": 12412 + }, + { + "epoch": 1.0577758841073712, + "grad_norm": 12.5, + "learning_rate": 1.144598975779234e-06, + "loss": 0.5454, + "step": 12413 + }, + { + "epoch": 1.0578610992756712, + "grad_norm": 16.375, + "learning_rate": 1.1444315396135538e-06, + "loss": 0.6366, + "step": 12414 + }, + { + "epoch": 1.0579463144439711, + "grad_norm": 17.0, + "learning_rate": 1.1442641053556023e-06, + "loss": 0.6623, + "step": 12415 + }, + { + "epoch": 1.058031529612271, + "grad_norm": 17.75, + "learning_rate": 1.1440966730084046e-06, + "loss": 0.7245, + "step": 12416 + }, + { + "epoch": 1.058116744780571, + "grad_norm": 14.25, + "learning_rate": 1.1439292425749867e-06, + "loss": 0.6474, + "step": 12417 + }, + { + "epoch": 1.058201959948871, + "grad_norm": 15.5625, + "learning_rate": 1.1437618140583734e-06, + "loss": 0.9257, + "step": 12418 + }, + { + "epoch": 1.058287175117171, + "grad_norm": 16.875, + "learning_rate": 1.1435943874615915e-06, + "loss": 0.7231, + "step": 12419 + }, + { + "epoch": 1.058372390285471, + "grad_norm": 17.125, + "learning_rate": 1.1434269627876656e-06, + "loss": 0.4903, + "step": 12420 + }, + { + "epoch": 1.0584576054537709, + "grad_norm": 17.375, + "learning_rate": 1.143259540039622e-06, + "loss": 0.6428, + "step": 12421 + }, + { + "epoch": 1.0585428206220708, + "grad_norm": 13.5625, + "learning_rate": 1.1430921192204854e-06, + "loss": 0.4689, + "step": 12422 + }, + { + "epoch": 1.0586280357903708, + "grad_norm": 19.375, + "learning_rate": 1.1429247003332824e-06, + "loss": 0.5604, + "step": 12423 + }, + { + "epoch": 1.0587132509586707, + "grad_norm": 17.5, + "learning_rate": 1.1427572833810376e-06, + "loss": 0.7459, + "step": 12424 + }, + { + "epoch": 1.0587984661269707, + "grad_norm": 11.5, + "learning_rate": 1.142589868366776e-06, + "loss": 0.3855, + "step": 12425 + }, + { + "epoch": 1.0588836812952707, + "grad_norm": 18.375, + "learning_rate": 1.142422455293524e-06, + "loss": 0.5321, + "step": 12426 + }, + { + "epoch": 1.0589688964635706, + "grad_norm": 12.625, + "learning_rate": 1.142255044164306e-06, + "loss": 0.3933, + "step": 12427 + }, + { + "epoch": 1.0590541116318706, + "grad_norm": 11.875, + "learning_rate": 1.142087634982148e-06, + "loss": 0.5832, + "step": 12428 + }, + { + "epoch": 1.0591393268001705, + "grad_norm": 17.25, + "learning_rate": 1.1419202277500746e-06, + "loss": 0.8307, + "step": 12429 + }, + { + "epoch": 1.0592245419684705, + "grad_norm": 14.3125, + "learning_rate": 1.1417528224711117e-06, + "loss": 0.3145, + "step": 12430 + }, + { + "epoch": 1.0593097571367704, + "grad_norm": 20.125, + "learning_rate": 1.1415854191482845e-06, + "loss": 0.3717, + "step": 12431 + }, + { + "epoch": 1.0593949723050704, + "grad_norm": 18.875, + "learning_rate": 1.1414180177846174e-06, + "loss": 0.7545, + "step": 12432 + }, + { + "epoch": 1.0594801874733704, + "grad_norm": 12.9375, + "learning_rate": 1.1412506183831357e-06, + "loss": 0.2973, + "step": 12433 + }, + { + "epoch": 1.0595654026416703, + "grad_norm": 16.75, + "learning_rate": 1.141083220946865e-06, + "loss": 0.9628, + "step": 12434 + }, + { + "epoch": 1.0596506178099703, + "grad_norm": 13.8125, + "learning_rate": 1.1409158254788295e-06, + "loss": 0.5443, + "step": 12435 + }, + { + "epoch": 1.0597358329782702, + "grad_norm": 11.625, + "learning_rate": 1.1407484319820553e-06, + "loss": 0.4232, + "step": 12436 + }, + { + "epoch": 1.0598210481465702, + "grad_norm": 10.25, + "learning_rate": 1.1405810404595664e-06, + "loss": 0.2284, + "step": 12437 + }, + { + "epoch": 1.0599062633148701, + "grad_norm": 19.375, + "learning_rate": 1.1404136509143883e-06, + "loss": 0.6412, + "step": 12438 + }, + { + "epoch": 1.05999147848317, + "grad_norm": 15.0625, + "learning_rate": 1.1402462633495463e-06, + "loss": 0.6547, + "step": 12439 + }, + { + "epoch": 1.06007669365147, + "grad_norm": 16.25, + "learning_rate": 1.1400788777680639e-06, + "loss": 0.5748, + "step": 12440 + }, + { + "epoch": 1.06016190881977, + "grad_norm": 12.5625, + "learning_rate": 1.139911494172967e-06, + "loss": 0.5351, + "step": 12441 + }, + { + "epoch": 1.06024712398807, + "grad_norm": 18.75, + "learning_rate": 1.1397441125672798e-06, + "loss": 0.6128, + "step": 12442 + }, + { + "epoch": 1.06033233915637, + "grad_norm": 12.4375, + "learning_rate": 1.1395767329540275e-06, + "loss": 0.5185, + "step": 12443 + }, + { + "epoch": 1.0604175543246699, + "grad_norm": 16.875, + "learning_rate": 1.1394093553362348e-06, + "loss": 0.877, + "step": 12444 + }, + { + "epoch": 1.0605027694929698, + "grad_norm": 14.625, + "learning_rate": 1.1392419797169263e-06, + "loss": 0.6787, + "step": 12445 + }, + { + "epoch": 1.0605879846612698, + "grad_norm": 19.5, + "learning_rate": 1.1390746060991264e-06, + "loss": 0.5324, + "step": 12446 + }, + { + "epoch": 1.0606731998295698, + "grad_norm": 39.25, + "learning_rate": 1.1389072344858606e-06, + "loss": 1.2895, + "step": 12447 + }, + { + "epoch": 1.0607584149978697, + "grad_norm": 17.75, + "learning_rate": 1.1387398648801521e-06, + "loss": 0.7786, + "step": 12448 + }, + { + "epoch": 1.0608436301661697, + "grad_norm": 21.625, + "learning_rate": 1.1385724972850262e-06, + "loss": 0.6146, + "step": 12449 + }, + { + "epoch": 1.0609288453344696, + "grad_norm": 23.125, + "learning_rate": 1.1384051317035073e-06, + "loss": 0.9894, + "step": 12450 + }, + { + "epoch": 1.0610140605027696, + "grad_norm": 15.875, + "learning_rate": 1.13823776813862e-06, + "loss": 0.6607, + "step": 12451 + }, + { + "epoch": 1.0610992756710695, + "grad_norm": 16.625, + "learning_rate": 1.1380704065933885e-06, + "loss": 0.7828, + "step": 12452 + }, + { + "epoch": 1.0611844908393695, + "grad_norm": 15.0, + "learning_rate": 1.1379030470708375e-06, + "loss": 0.5605, + "step": 12453 + }, + { + "epoch": 1.0612697060076695, + "grad_norm": 18.375, + "learning_rate": 1.1377356895739907e-06, + "loss": 0.6024, + "step": 12454 + }, + { + "epoch": 1.0613549211759694, + "grad_norm": 16.75, + "learning_rate": 1.1375683341058736e-06, + "loss": 0.3908, + "step": 12455 + }, + { + "epoch": 1.0614401363442694, + "grad_norm": 17.125, + "learning_rate": 1.1374009806695095e-06, + "loss": 0.7532, + "step": 12456 + }, + { + "epoch": 1.0615253515125693, + "grad_norm": 14.0, + "learning_rate": 1.1372336292679227e-06, + "loss": 0.7757, + "step": 12457 + }, + { + "epoch": 1.0616105666808693, + "grad_norm": 15.0, + "learning_rate": 1.1370662799041378e-06, + "loss": 0.6946, + "step": 12458 + }, + { + "epoch": 1.0616957818491692, + "grad_norm": 14.6875, + "learning_rate": 1.1368989325811783e-06, + "loss": 0.5821, + "step": 12459 + }, + { + "epoch": 1.0617809970174692, + "grad_norm": 14.875, + "learning_rate": 1.1367315873020695e-06, + "loss": 0.4802, + "step": 12460 + }, + { + "epoch": 1.0618662121857692, + "grad_norm": 27.125, + "learning_rate": 1.1365642440698344e-06, + "loss": 1.3032, + "step": 12461 + }, + { + "epoch": 1.061951427354069, + "grad_norm": 15.0, + "learning_rate": 1.1363969028874976e-06, + "loss": 0.7396, + "step": 12462 + }, + { + "epoch": 1.062036642522369, + "grad_norm": 16.375, + "learning_rate": 1.1362295637580828e-06, + "loss": 0.754, + "step": 12463 + }, + { + "epoch": 1.062121857690669, + "grad_norm": 14.75, + "learning_rate": 1.136062226684615e-06, + "loss": 0.7202, + "step": 12464 + }, + { + "epoch": 1.062207072858969, + "grad_norm": 14.5625, + "learning_rate": 1.1358948916701166e-06, + "loss": 0.7034, + "step": 12465 + }, + { + "epoch": 1.062292288027269, + "grad_norm": 12.5625, + "learning_rate": 1.1357275587176125e-06, + "loss": 0.5123, + "step": 12466 + }, + { + "epoch": 1.062377503195569, + "grad_norm": 21.5, + "learning_rate": 1.135560227830126e-06, + "loss": 0.6569, + "step": 12467 + }, + { + "epoch": 1.0624627183638689, + "grad_norm": 26.625, + "learning_rate": 1.1353928990106817e-06, + "loss": 0.7784, + "step": 12468 + }, + { + "epoch": 1.0625479335321688, + "grad_norm": 12.75, + "learning_rate": 1.1352255722623026e-06, + "loss": 0.4901, + "step": 12469 + }, + { + "epoch": 1.0626331487004688, + "grad_norm": 18.375, + "learning_rate": 1.1350582475880128e-06, + "loss": 0.6289, + "step": 12470 + }, + { + "epoch": 1.0627183638687687, + "grad_norm": 13.3125, + "learning_rate": 1.134890924990836e-06, + "loss": 0.4797, + "step": 12471 + }, + { + "epoch": 1.0628035790370687, + "grad_norm": 15.625, + "learning_rate": 1.1347236044737965e-06, + "loss": 0.7833, + "step": 12472 + }, + { + "epoch": 1.0628887942053686, + "grad_norm": 19.25, + "learning_rate": 1.1345562860399172e-06, + "loss": 0.8292, + "step": 12473 + }, + { + "epoch": 1.0629740093736686, + "grad_norm": 15.3125, + "learning_rate": 1.1343889696922216e-06, + "loss": 0.8105, + "step": 12474 + }, + { + "epoch": 1.0630592245419686, + "grad_norm": 14.1875, + "learning_rate": 1.1342216554337338e-06, + "loss": 0.5428, + "step": 12475 + }, + { + "epoch": 1.0631444397102685, + "grad_norm": 18.0, + "learning_rate": 1.1340543432674769e-06, + "loss": 0.7002, + "step": 12476 + }, + { + "epoch": 1.0632296548785685, + "grad_norm": 12.8125, + "learning_rate": 1.1338870331964747e-06, + "loss": 0.7705, + "step": 12477 + }, + { + "epoch": 1.0633148700468684, + "grad_norm": 17.375, + "learning_rate": 1.1337197252237504e-06, + "loss": 0.4863, + "step": 12478 + }, + { + "epoch": 1.0634000852151684, + "grad_norm": 12.25, + "learning_rate": 1.133552419352328e-06, + "loss": 0.3831, + "step": 12479 + }, + { + "epoch": 1.0634853003834683, + "grad_norm": 14.25, + "learning_rate": 1.1333851155852307e-06, + "loss": 0.516, + "step": 12480 + }, + { + "epoch": 1.0635705155517683, + "grad_norm": 22.75, + "learning_rate": 1.133217813925481e-06, + "loss": 1.1684, + "step": 12481 + }, + { + "epoch": 1.0636557307200682, + "grad_norm": 57.25, + "learning_rate": 1.133050514376103e-06, + "loss": 0.7768, + "step": 12482 + }, + { + "epoch": 1.0637409458883682, + "grad_norm": 14.125, + "learning_rate": 1.1328832169401198e-06, + "loss": 0.4273, + "step": 12483 + }, + { + "epoch": 1.0638261610566682, + "grad_norm": 14.4375, + "learning_rate": 1.1327159216205546e-06, + "loss": 0.7931, + "step": 12484 + }, + { + "epoch": 1.0639113762249681, + "grad_norm": 14.8125, + "learning_rate": 1.132548628420431e-06, + "loss": 0.801, + "step": 12485 + }, + { + "epoch": 1.063996591393268, + "grad_norm": 16.125, + "learning_rate": 1.132381337342771e-06, + "loss": 0.4861, + "step": 12486 + }, + { + "epoch": 1.064081806561568, + "grad_norm": 11.8125, + "learning_rate": 1.1322140483905992e-06, + "loss": 0.5341, + "step": 12487 + }, + { + "epoch": 1.064167021729868, + "grad_norm": 21.375, + "learning_rate": 1.132046761566938e-06, + "loss": 0.9495, + "step": 12488 + }, + { + "epoch": 1.064252236898168, + "grad_norm": 21.625, + "learning_rate": 1.1318794768748103e-06, + "loss": 1.1944, + "step": 12489 + }, + { + "epoch": 1.064337452066468, + "grad_norm": 14.75, + "learning_rate": 1.1317121943172391e-06, + "loss": 0.7163, + "step": 12490 + }, + { + "epoch": 1.0644226672347679, + "grad_norm": 11.75, + "learning_rate": 1.1315449138972475e-06, + "loss": 0.4529, + "step": 12491 + }, + { + "epoch": 1.0645078824030678, + "grad_norm": 15.375, + "learning_rate": 1.1313776356178586e-06, + "loss": 0.5153, + "step": 12492 + }, + { + "epoch": 1.0645930975713678, + "grad_norm": 10.875, + "learning_rate": 1.1312103594820948e-06, + "loss": 0.3603, + "step": 12493 + }, + { + "epoch": 1.0646783127396677, + "grad_norm": 25.375, + "learning_rate": 1.1310430854929793e-06, + "loss": 1.2229, + "step": 12494 + }, + { + "epoch": 1.0647635279079677, + "grad_norm": 15.75, + "learning_rate": 1.130875813653535e-06, + "loss": 0.8113, + "step": 12495 + }, + { + "epoch": 1.0648487430762676, + "grad_norm": 12.375, + "learning_rate": 1.1307085439667852e-06, + "loss": 0.5185, + "step": 12496 + }, + { + "epoch": 1.0649339582445676, + "grad_norm": 12.9375, + "learning_rate": 1.130541276435751e-06, + "loss": 0.5143, + "step": 12497 + }, + { + "epoch": 1.0650191734128676, + "grad_norm": 11.875, + "learning_rate": 1.1303740110634564e-06, + "loss": 0.411, + "step": 12498 + }, + { + "epoch": 1.0651043885811675, + "grad_norm": 14.875, + "learning_rate": 1.1302067478529237e-06, + "loss": 0.7828, + "step": 12499 + }, + { + "epoch": 1.0651896037494675, + "grad_norm": 12.8125, + "learning_rate": 1.1300394868071755e-06, + "loss": 0.3977, + "step": 12500 + }, + { + "epoch": 1.0652748189177674, + "grad_norm": 11.25, + "learning_rate": 1.1298722279292342e-06, + "loss": 0.3172, + "step": 12501 + }, + { + "epoch": 1.0653600340860674, + "grad_norm": 16.375, + "learning_rate": 1.1297049712221228e-06, + "loss": 0.4312, + "step": 12502 + }, + { + "epoch": 1.0654452492543673, + "grad_norm": 19.125, + "learning_rate": 1.1295377166888635e-06, + "loss": 0.8106, + "step": 12503 + }, + { + "epoch": 1.0655304644226673, + "grad_norm": 17.0, + "learning_rate": 1.129370464332479e-06, + "loss": 0.8247, + "step": 12504 + }, + { + "epoch": 1.0656156795909673, + "grad_norm": 18.125, + "learning_rate": 1.1292032141559914e-06, + "loss": 0.6249, + "step": 12505 + }, + { + "epoch": 1.0657008947592672, + "grad_norm": 12.8125, + "learning_rate": 1.129035966162423e-06, + "loss": 0.2157, + "step": 12506 + }, + { + "epoch": 1.0657861099275672, + "grad_norm": 8.75, + "learning_rate": 1.1288687203547964e-06, + "loss": 0.1909, + "step": 12507 + }, + { + "epoch": 1.0658713250958671, + "grad_norm": 13.8125, + "learning_rate": 1.1287014767361338e-06, + "loss": 0.745, + "step": 12508 + }, + { + "epoch": 1.065956540264167, + "grad_norm": 20.25, + "learning_rate": 1.1285342353094574e-06, + "loss": 0.787, + "step": 12509 + }, + { + "epoch": 1.066041755432467, + "grad_norm": 12.5625, + "learning_rate": 1.1283669960777896e-06, + "loss": 0.5329, + "step": 12510 + }, + { + "epoch": 1.066126970600767, + "grad_norm": 12.3125, + "learning_rate": 1.1281997590441526e-06, + "loss": 0.5704, + "step": 12511 + }, + { + "epoch": 1.066212185769067, + "grad_norm": 18.0, + "learning_rate": 1.1280325242115688e-06, + "loss": 0.5973, + "step": 12512 + }, + { + "epoch": 1.066297400937367, + "grad_norm": 12.25, + "learning_rate": 1.1278652915830596e-06, + "loss": 0.3941, + "step": 12513 + }, + { + "epoch": 1.0663826161056669, + "grad_norm": 11.125, + "learning_rate": 1.1276980611616472e-06, + "loss": 0.3744, + "step": 12514 + }, + { + "epoch": 1.0664678312739668, + "grad_norm": 14.75, + "learning_rate": 1.127530832950354e-06, + "loss": 0.6654, + "step": 12515 + }, + { + "epoch": 1.0665530464422668, + "grad_norm": 14.0, + "learning_rate": 1.1273636069522015e-06, + "loss": 0.6663, + "step": 12516 + }, + { + "epoch": 1.0666382616105667, + "grad_norm": 12.6875, + "learning_rate": 1.1271963831702122e-06, + "loss": 0.626, + "step": 12517 + }, + { + "epoch": 1.0667234767788667, + "grad_norm": 16.5, + "learning_rate": 1.1270291616074075e-06, + "loss": 0.8603, + "step": 12518 + }, + { + "epoch": 1.0668086919471667, + "grad_norm": 13.625, + "learning_rate": 1.1268619422668098e-06, + "loss": 0.6728, + "step": 12519 + }, + { + "epoch": 1.0668939071154666, + "grad_norm": 18.0, + "learning_rate": 1.126694725151441e-06, + "loss": 0.7052, + "step": 12520 + }, + { + "epoch": 1.0669791222837666, + "grad_norm": 13.875, + "learning_rate": 1.126527510264322e-06, + "loss": 0.485, + "step": 12521 + }, + { + "epoch": 1.0670643374520665, + "grad_norm": 17.0, + "learning_rate": 1.126360297608475e-06, + "loss": 0.9502, + "step": 12522 + }, + { + "epoch": 1.0671495526203665, + "grad_norm": 13.875, + "learning_rate": 1.1261930871869217e-06, + "loss": 0.4682, + "step": 12523 + }, + { + "epoch": 1.0672347677886664, + "grad_norm": 19.875, + "learning_rate": 1.126025879002684e-06, + "loss": 0.4903, + "step": 12524 + }, + { + "epoch": 1.0673199829569664, + "grad_norm": 12.1875, + "learning_rate": 1.1258586730587832e-06, + "loss": 0.5351, + "step": 12525 + }, + { + "epoch": 1.0674051981252664, + "grad_norm": 16.75, + "learning_rate": 1.125691469358241e-06, + "loss": 0.8411, + "step": 12526 + }, + { + "epoch": 1.0674904132935663, + "grad_norm": 22.375, + "learning_rate": 1.1255242679040795e-06, + "loss": 0.7882, + "step": 12527 + }, + { + "epoch": 1.0675756284618663, + "grad_norm": 19.75, + "learning_rate": 1.1253570686993188e-06, + "loss": 0.7725, + "step": 12528 + }, + { + "epoch": 1.0676608436301662, + "grad_norm": 14.1875, + "learning_rate": 1.125189871746982e-06, + "loss": 0.4929, + "step": 12529 + }, + { + "epoch": 1.0677460587984662, + "grad_norm": 14.25, + "learning_rate": 1.1250226770500895e-06, + "loss": 0.6896, + "step": 12530 + }, + { + "epoch": 1.0678312739667661, + "grad_norm": 28.0, + "learning_rate": 1.1248554846116625e-06, + "loss": 0.6349, + "step": 12531 + }, + { + "epoch": 1.067916489135066, + "grad_norm": 12.6875, + "learning_rate": 1.1246882944347232e-06, + "loss": 0.4214, + "step": 12532 + }, + { + "epoch": 1.068001704303366, + "grad_norm": 14.9375, + "learning_rate": 1.124521106522292e-06, + "loss": 0.4369, + "step": 12533 + }, + { + "epoch": 1.068086919471666, + "grad_norm": 17.75, + "learning_rate": 1.124353920877391e-06, + "loss": 0.5722, + "step": 12534 + }, + { + "epoch": 1.068172134639966, + "grad_norm": 12.125, + "learning_rate": 1.1241867375030404e-06, + "loss": 0.5572, + "step": 12535 + }, + { + "epoch": 1.068257349808266, + "grad_norm": 17.0, + "learning_rate": 1.1240195564022625e-06, + "loss": 0.6553, + "step": 12536 + }, + { + "epoch": 1.0683425649765659, + "grad_norm": 15.5625, + "learning_rate": 1.123852377578078e-06, + "loss": 0.6276, + "step": 12537 + }, + { + "epoch": 1.0684277801448658, + "grad_norm": 12.375, + "learning_rate": 1.1236852010335077e-06, + "loss": 0.4641, + "step": 12538 + }, + { + "epoch": 1.0685129953131658, + "grad_norm": 16.875, + "learning_rate": 1.1235180267715728e-06, + "loss": 1.0508, + "step": 12539 + }, + { + "epoch": 1.0685982104814657, + "grad_norm": 12.0, + "learning_rate": 1.1233508547952942e-06, + "loss": 0.3245, + "step": 12540 + }, + { + "epoch": 1.0686834256497657, + "grad_norm": 19.375, + "learning_rate": 1.1231836851076931e-06, + "loss": 0.6954, + "step": 12541 + }, + { + "epoch": 1.0687686408180657, + "grad_norm": 13.5625, + "learning_rate": 1.1230165177117906e-06, + "loss": 0.7253, + "step": 12542 + }, + { + "epoch": 1.0688538559863656, + "grad_norm": 14.125, + "learning_rate": 1.122849352610607e-06, + "loss": 0.5273, + "step": 12543 + }, + { + "epoch": 1.0689390711546656, + "grad_norm": 22.75, + "learning_rate": 1.1226821898071635e-06, + "loss": 0.4173, + "step": 12544 + }, + { + "epoch": 1.0690242863229655, + "grad_norm": 13.9375, + "learning_rate": 1.1225150293044813e-06, + "loss": 0.7599, + "step": 12545 + }, + { + "epoch": 1.0691095014912655, + "grad_norm": 15.4375, + "learning_rate": 1.12234787110558e-06, + "loss": 0.7337, + "step": 12546 + }, + { + "epoch": 1.0691947166595654, + "grad_norm": 21.125, + "learning_rate": 1.1221807152134816e-06, + "loss": 1.0787, + "step": 12547 + }, + { + "epoch": 1.0692799318278654, + "grad_norm": 11.8125, + "learning_rate": 1.1220135616312057e-06, + "loss": 0.5756, + "step": 12548 + }, + { + "epoch": 1.0693651469961654, + "grad_norm": 12.75, + "learning_rate": 1.1218464103617737e-06, + "loss": 0.486, + "step": 12549 + }, + { + "epoch": 1.0694503621644653, + "grad_norm": 16.5, + "learning_rate": 1.121679261408206e-06, + "loss": 0.6116, + "step": 12550 + }, + { + "epoch": 1.0695355773327653, + "grad_norm": 33.0, + "learning_rate": 1.1215121147735229e-06, + "loss": 0.7586, + "step": 12551 + }, + { + "epoch": 1.0696207925010652, + "grad_norm": 11.1875, + "learning_rate": 1.121344970460745e-06, + "loss": 0.3002, + "step": 12552 + }, + { + "epoch": 1.0697060076693652, + "grad_norm": 12.625, + "learning_rate": 1.1211778284728936e-06, + "loss": 0.3824, + "step": 12553 + }, + { + "epoch": 1.0697912228376651, + "grad_norm": 12.875, + "learning_rate": 1.121010688812988e-06, + "loss": 0.3992, + "step": 12554 + }, + { + "epoch": 1.069876438005965, + "grad_norm": 14.5625, + "learning_rate": 1.1208435514840485e-06, + "loss": 0.4742, + "step": 12555 + }, + { + "epoch": 1.069961653174265, + "grad_norm": 15.0, + "learning_rate": 1.1206764164890962e-06, + "loss": 0.6216, + "step": 12556 + }, + { + "epoch": 1.070046868342565, + "grad_norm": 10.3125, + "learning_rate": 1.1205092838311513e-06, + "loss": 0.2655, + "step": 12557 + }, + { + "epoch": 1.070132083510865, + "grad_norm": 16.5, + "learning_rate": 1.1203421535132334e-06, + "loss": 0.7743, + "step": 12558 + }, + { + "epoch": 1.070217298679165, + "grad_norm": 14.9375, + "learning_rate": 1.1201750255383636e-06, + "loss": 0.4042, + "step": 12559 + }, + { + "epoch": 1.0703025138474649, + "grad_norm": 14.0625, + "learning_rate": 1.1200078999095612e-06, + "loss": 0.5265, + "step": 12560 + }, + { + "epoch": 1.0703877290157648, + "grad_norm": 19.5, + "learning_rate": 1.1198407766298475e-06, + "loss": 0.851, + "step": 12561 + }, + { + "epoch": 1.0704729441840648, + "grad_norm": 10.75, + "learning_rate": 1.1196736557022414e-06, + "loss": 0.3177, + "step": 12562 + }, + { + "epoch": 1.0705581593523648, + "grad_norm": 18.375, + "learning_rate": 1.1195065371297633e-06, + "loss": 0.5534, + "step": 12563 + }, + { + "epoch": 1.0706433745206647, + "grad_norm": 15.875, + "learning_rate": 1.1193394209154335e-06, + "loss": 0.5832, + "step": 12564 + }, + { + "epoch": 1.0707285896889647, + "grad_norm": 14.625, + "learning_rate": 1.1191723070622714e-06, + "loss": 0.4795, + "step": 12565 + }, + { + "epoch": 1.0708138048572646, + "grad_norm": 16.875, + "learning_rate": 1.1190051955732974e-06, + "loss": 0.6782, + "step": 12566 + }, + { + "epoch": 1.0708990200255646, + "grad_norm": 28.625, + "learning_rate": 1.1188380864515313e-06, + "loss": 1.1923, + "step": 12567 + }, + { + "epoch": 1.0709842351938645, + "grad_norm": 17.625, + "learning_rate": 1.118670979699993e-06, + "loss": 0.5508, + "step": 12568 + }, + { + "epoch": 1.0710694503621645, + "grad_norm": 13.375, + "learning_rate": 1.1185038753217024e-06, + "loss": 0.4533, + "step": 12569 + }, + { + "epoch": 1.0711546655304645, + "grad_norm": 14.25, + "learning_rate": 1.1183367733196785e-06, + "loss": 0.39, + "step": 12570 + }, + { + "epoch": 1.0712398806987644, + "grad_norm": 16.125, + "learning_rate": 1.1181696736969418e-06, + "loss": 0.7198, + "step": 12571 + }, + { + "epoch": 1.0713250958670644, + "grad_norm": 16.0, + "learning_rate": 1.1180025764565117e-06, + "loss": 0.7945, + "step": 12572 + }, + { + "epoch": 1.0714103110353643, + "grad_norm": 21.5, + "learning_rate": 1.1178354816014073e-06, + "loss": 0.7167, + "step": 12573 + }, + { + "epoch": 1.0714955262036643, + "grad_norm": 10.875, + "learning_rate": 1.1176683891346491e-06, + "loss": 0.3108, + "step": 12574 + }, + { + "epoch": 1.0715807413719642, + "grad_norm": 17.75, + "learning_rate": 1.117501299059256e-06, + "loss": 0.4911, + "step": 12575 + }, + { + "epoch": 1.0716659565402642, + "grad_norm": 16.625, + "learning_rate": 1.1173342113782476e-06, + "loss": 0.652, + "step": 12576 + }, + { + "epoch": 1.0717511717085642, + "grad_norm": 18.0, + "learning_rate": 1.117167126094644e-06, + "loss": 0.5362, + "step": 12577 + }, + { + "epoch": 1.0718363868768641, + "grad_norm": 14.9375, + "learning_rate": 1.1170000432114634e-06, + "loss": 0.6891, + "step": 12578 + }, + { + "epoch": 1.071921602045164, + "grad_norm": 12.8125, + "learning_rate": 1.116832962731726e-06, + "loss": 0.5162, + "step": 12579 + }, + { + "epoch": 1.072006817213464, + "grad_norm": 12.625, + "learning_rate": 1.1166658846584505e-06, + "loss": 0.4009, + "step": 12580 + }, + { + "epoch": 1.072092032381764, + "grad_norm": 15.3125, + "learning_rate": 1.1164988089946568e-06, + "loss": 0.7076, + "step": 12581 + }, + { + "epoch": 1.072177247550064, + "grad_norm": 13.625, + "learning_rate": 1.1163317357433638e-06, + "loss": 0.5016, + "step": 12582 + }, + { + "epoch": 1.072262462718364, + "grad_norm": 12.25, + "learning_rate": 1.1161646649075909e-06, + "loss": 0.4133, + "step": 12583 + }, + { + "epoch": 1.0723476778866639, + "grad_norm": 28.875, + "learning_rate": 1.1159975964903566e-06, + "loss": 1.1796, + "step": 12584 + }, + { + "epoch": 1.0724328930549638, + "grad_norm": 16.75, + "learning_rate": 1.1158305304946814e-06, + "loss": 0.8258, + "step": 12585 + }, + { + "epoch": 1.0725181082232638, + "grad_norm": 16.875, + "learning_rate": 1.115663466923583e-06, + "loss": 0.5508, + "step": 12586 + }, + { + "epoch": 1.0726033233915637, + "grad_norm": 8.8125, + "learning_rate": 1.1154964057800806e-06, + "loss": 0.8667, + "step": 12587 + }, + { + "epoch": 1.0726885385598637, + "grad_norm": 13.3125, + "learning_rate": 1.1153293470671936e-06, + "loss": 0.4107, + "step": 12588 + }, + { + "epoch": 1.0727737537281636, + "grad_norm": 17.375, + "learning_rate": 1.1151622907879407e-06, + "loss": 0.6532, + "step": 12589 + }, + { + "epoch": 1.0728589688964636, + "grad_norm": 20.375, + "learning_rate": 1.1149952369453404e-06, + "loss": 0.7967, + "step": 12590 + }, + { + "epoch": 1.0729441840647636, + "grad_norm": 13.875, + "learning_rate": 1.1148281855424125e-06, + "loss": 0.795, + "step": 12591 + }, + { + "epoch": 1.0730293992330635, + "grad_norm": 24.75, + "learning_rate": 1.114661136582175e-06, + "loss": 1.0112, + "step": 12592 + }, + { + "epoch": 1.0731146144013635, + "grad_norm": 19.375, + "learning_rate": 1.114494090067647e-06, + "loss": 0.7698, + "step": 12593 + }, + { + "epoch": 1.0731998295696634, + "grad_norm": 13.625, + "learning_rate": 1.114327046001847e-06, + "loss": 0.3142, + "step": 12594 + }, + { + "epoch": 1.0732850447379634, + "grad_norm": 14.5, + "learning_rate": 1.1141600043877935e-06, + "loss": 0.6812, + "step": 12595 + }, + { + "epoch": 1.0733702599062633, + "grad_norm": 18.5, + "learning_rate": 1.1139929652285056e-06, + "loss": 0.4265, + "step": 12596 + }, + { + "epoch": 1.0734554750745633, + "grad_norm": 15.4375, + "learning_rate": 1.113825928527001e-06, + "loss": 0.4827, + "step": 12597 + }, + { + "epoch": 1.0735406902428632, + "grad_norm": 16.375, + "learning_rate": 1.1136588942862995e-06, + "loss": 0.5722, + "step": 12598 + }, + { + "epoch": 1.0736259054111632, + "grad_norm": 17.375, + "learning_rate": 1.1134918625094185e-06, + "loss": 0.522, + "step": 12599 + }, + { + "epoch": 1.0737111205794632, + "grad_norm": 18.875, + "learning_rate": 1.1133248331993771e-06, + "loss": 0.898, + "step": 12600 + }, + { + "epoch": 1.0737963357477631, + "grad_norm": 16.5, + "learning_rate": 1.113157806359193e-06, + "loss": 0.6983, + "step": 12601 + }, + { + "epoch": 1.073881550916063, + "grad_norm": 17.625, + "learning_rate": 1.1129907819918856e-06, + "loss": 0.7055, + "step": 12602 + }, + { + "epoch": 1.073966766084363, + "grad_norm": 17.125, + "learning_rate": 1.112823760100472e-06, + "loss": 0.8291, + "step": 12603 + }, + { + "epoch": 1.074051981252663, + "grad_norm": 23.375, + "learning_rate": 1.1126567406879713e-06, + "loss": 1.2439, + "step": 12604 + }, + { + "epoch": 1.074137196420963, + "grad_norm": 17.25, + "learning_rate": 1.112489723757401e-06, + "loss": 0.5468, + "step": 12605 + }, + { + "epoch": 1.074222411589263, + "grad_norm": 15.625, + "learning_rate": 1.11232270931178e-06, + "loss": 0.6623, + "step": 12606 + }, + { + "epoch": 1.0743076267575629, + "grad_norm": 19.0, + "learning_rate": 1.1121556973541259e-06, + "loss": 0.7655, + "step": 12607 + }, + { + "epoch": 1.0743928419258628, + "grad_norm": 14.0625, + "learning_rate": 1.111988687887457e-06, + "loss": 0.7381, + "step": 12608 + }, + { + "epoch": 1.0744780570941628, + "grad_norm": 21.375, + "learning_rate": 1.1118216809147913e-06, + "loss": 0.6058, + "step": 12609 + }, + { + "epoch": 1.0745632722624627, + "grad_norm": 16.25, + "learning_rate": 1.111654676439147e-06, + "loss": 0.8145, + "step": 12610 + }, + { + "epoch": 1.0746484874307627, + "grad_norm": 12.4375, + "learning_rate": 1.1114876744635416e-06, + "loss": 0.3852, + "step": 12611 + }, + { + "epoch": 1.0747337025990626, + "grad_norm": 16.875, + "learning_rate": 1.111320674990993e-06, + "loss": 0.5855, + "step": 12612 + }, + { + "epoch": 1.0748189177673626, + "grad_norm": 16.125, + "learning_rate": 1.1111536780245196e-06, + "loss": 0.4772, + "step": 12613 + }, + { + "epoch": 1.0749041329356626, + "grad_norm": 10.875, + "learning_rate": 1.1109866835671386e-06, + "loss": 0.2645, + "step": 12614 + }, + { + "epoch": 1.0749893481039625, + "grad_norm": 15.0, + "learning_rate": 1.1108196916218683e-06, + "loss": 0.5453, + "step": 12615 + }, + { + "epoch": 1.0750745632722625, + "grad_norm": 15.375, + "learning_rate": 1.1106527021917259e-06, + "loss": 0.6427, + "step": 12616 + }, + { + "epoch": 1.0751597784405624, + "grad_norm": 17.625, + "learning_rate": 1.1104857152797294e-06, + "loss": 0.7207, + "step": 12617 + }, + { + "epoch": 1.0752449936088624, + "grad_norm": 19.875, + "learning_rate": 1.1103187308888967e-06, + "loss": 0.5573, + "step": 12618 + }, + { + "epoch": 1.0753302087771623, + "grad_norm": 11.5625, + "learning_rate": 1.1101517490222447e-06, + "loss": 0.3856, + "step": 12619 + }, + { + "epoch": 1.0754154239454623, + "grad_norm": 18.5, + "learning_rate": 1.1099847696827908e-06, + "loss": 0.4324, + "step": 12620 + }, + { + "epoch": 1.0755006391137623, + "grad_norm": 10.375, + "learning_rate": 1.1098177928735534e-06, + "loss": 0.2635, + "step": 12621 + }, + { + "epoch": 1.0755858542820622, + "grad_norm": 10.3125, + "learning_rate": 1.1096508185975492e-06, + "loss": 0.2998, + "step": 12622 + }, + { + "epoch": 1.0756710694503622, + "grad_norm": 13.6875, + "learning_rate": 1.109483846857796e-06, + "loss": 0.5418, + "step": 12623 + }, + { + "epoch": 1.0757562846186621, + "grad_norm": 14.6875, + "learning_rate": 1.1093168776573105e-06, + "loss": 0.6295, + "step": 12624 + }, + { + "epoch": 1.075841499786962, + "grad_norm": 20.375, + "learning_rate": 1.109149910999111e-06, + "loss": 0.7588, + "step": 12625 + }, + { + "epoch": 1.075926714955262, + "grad_norm": 13.4375, + "learning_rate": 1.1089829468862146e-06, + "loss": 0.5768, + "step": 12626 + }, + { + "epoch": 1.076011930123562, + "grad_norm": 14.875, + "learning_rate": 1.1088159853216374e-06, + "loss": 0.9127, + "step": 12627 + }, + { + "epoch": 1.076097145291862, + "grad_norm": 14.4375, + "learning_rate": 1.1086490263083977e-06, + "loss": 0.5684, + "step": 12628 + }, + { + "epoch": 1.076182360460162, + "grad_norm": 17.25, + "learning_rate": 1.1084820698495121e-06, + "loss": 0.7395, + "step": 12629 + }, + { + "epoch": 1.0762675756284619, + "grad_norm": 21.375, + "learning_rate": 1.1083151159479977e-06, + "loss": 0.5173, + "step": 12630 + }, + { + "epoch": 1.0763527907967618, + "grad_norm": 14.125, + "learning_rate": 1.1081481646068718e-06, + "loss": 0.4354, + "step": 12631 + }, + { + "epoch": 1.0764380059650618, + "grad_norm": 16.125, + "learning_rate": 1.1079812158291512e-06, + "loss": 0.5181, + "step": 12632 + }, + { + "epoch": 1.0765232211333617, + "grad_norm": 13.125, + "learning_rate": 1.1078142696178525e-06, + "loss": 0.557, + "step": 12633 + }, + { + "epoch": 1.0766084363016617, + "grad_norm": 12.3125, + "learning_rate": 1.107647325975994e-06, + "loss": 0.4769, + "step": 12634 + }, + { + "epoch": 1.0766936514699617, + "grad_norm": 15.3125, + "learning_rate": 1.1074803849065906e-06, + "loss": 0.8167, + "step": 12635 + }, + { + "epoch": 1.0767788666382616, + "grad_norm": 13.1875, + "learning_rate": 1.10731344641266e-06, + "loss": 0.4167, + "step": 12636 + }, + { + "epoch": 1.0768640818065616, + "grad_norm": 12.875, + "learning_rate": 1.1071465104972188e-06, + "loss": 0.3861, + "step": 12637 + }, + { + "epoch": 1.0769492969748615, + "grad_norm": 14.375, + "learning_rate": 1.1069795771632841e-06, + "loss": 0.4314, + "step": 12638 + }, + { + "epoch": 1.0770345121431615, + "grad_norm": 11.5, + "learning_rate": 1.106812646413872e-06, + "loss": 0.3936, + "step": 12639 + }, + { + "epoch": 1.0771197273114614, + "grad_norm": 16.875, + "learning_rate": 1.1066457182519997e-06, + "loss": 0.6304, + "step": 12640 + }, + { + "epoch": 1.0772049424797614, + "grad_norm": 12.5, + "learning_rate": 1.106478792680683e-06, + "loss": 0.4781, + "step": 12641 + }, + { + "epoch": 1.0772901576480614, + "grad_norm": 15.8125, + "learning_rate": 1.1063118697029394e-06, + "loss": 0.6092, + "step": 12642 + }, + { + "epoch": 1.0773753728163613, + "grad_norm": 41.75, + "learning_rate": 1.1061449493217844e-06, + "loss": 0.6386, + "step": 12643 + }, + { + "epoch": 1.0774605879846613, + "grad_norm": 22.375, + "learning_rate": 1.1059780315402348e-06, + "loss": 0.8606, + "step": 12644 + }, + { + "epoch": 1.0775458031529612, + "grad_norm": 10.5, + "learning_rate": 1.1058111163613071e-06, + "loss": 0.2788, + "step": 12645 + }, + { + "epoch": 1.0776310183212612, + "grad_norm": 16.375, + "learning_rate": 1.1056442037880174e-06, + "loss": 0.6694, + "step": 12646 + }, + { + "epoch": 1.0777162334895611, + "grad_norm": 14.1875, + "learning_rate": 1.1054772938233825e-06, + "loss": 0.6951, + "step": 12647 + }, + { + "epoch": 1.077801448657861, + "grad_norm": 16.75, + "learning_rate": 1.1053103864704175e-06, + "loss": 0.7417, + "step": 12648 + }, + { + "epoch": 1.077886663826161, + "grad_norm": 14.8125, + "learning_rate": 1.10514348173214e-06, + "loss": 0.4677, + "step": 12649 + }, + { + "epoch": 1.077971878994461, + "grad_norm": 16.125, + "learning_rate": 1.1049765796115655e-06, + "loss": 0.8182, + "step": 12650 + }, + { + "epoch": 1.078057094162761, + "grad_norm": 16.125, + "learning_rate": 1.1048096801117098e-06, + "loss": 0.7342, + "step": 12651 + }, + { + "epoch": 1.078142309331061, + "grad_norm": 19.875, + "learning_rate": 1.104642783235589e-06, + "loss": 0.5371, + "step": 12652 + }, + { + "epoch": 1.0782275244993609, + "grad_norm": 14.9375, + "learning_rate": 1.1044758889862195e-06, + "loss": 0.5905, + "step": 12653 + }, + { + "epoch": 1.0783127396676608, + "grad_norm": 12.9375, + "learning_rate": 1.1043089973666167e-06, + "loss": 0.4751, + "step": 12654 + }, + { + "epoch": 1.0783979548359608, + "grad_norm": 14.625, + "learning_rate": 1.104142108379797e-06, + "loss": 0.5421, + "step": 12655 + }, + { + "epoch": 1.0784831700042607, + "grad_norm": 15.625, + "learning_rate": 1.103975222028776e-06, + "loss": 0.7648, + "step": 12656 + }, + { + "epoch": 1.0785683851725607, + "grad_norm": 16.5, + "learning_rate": 1.1038083383165698e-06, + "loss": 0.635, + "step": 12657 + }, + { + "epoch": 1.0786536003408607, + "grad_norm": 19.625, + "learning_rate": 1.1036414572461942e-06, + "loss": 1.1928, + "step": 12658 + }, + { + "epoch": 1.0787388155091606, + "grad_norm": 15.4375, + "learning_rate": 1.1034745788206642e-06, + "loss": 0.6254, + "step": 12659 + }, + { + "epoch": 1.0788240306774606, + "grad_norm": 18.0, + "learning_rate": 1.1033077030429958e-06, + "loss": 0.4726, + "step": 12660 + }, + { + "epoch": 1.0789092458457605, + "grad_norm": 19.25, + "learning_rate": 1.1031408299162049e-06, + "loss": 0.8441, + "step": 12661 + }, + { + "epoch": 1.0789944610140605, + "grad_norm": 37.0, + "learning_rate": 1.102973959443307e-06, + "loss": 0.841, + "step": 12662 + }, + { + "epoch": 1.0790796761823604, + "grad_norm": 14.9375, + "learning_rate": 1.102807091627317e-06, + "loss": 0.5267, + "step": 12663 + }, + { + "epoch": 1.0791648913506604, + "grad_norm": 14.625, + "learning_rate": 1.1026402264712515e-06, + "loss": 0.8202, + "step": 12664 + }, + { + "epoch": 1.0792501065189604, + "grad_norm": 14.75, + "learning_rate": 1.102473363978125e-06, + "loss": 0.3713, + "step": 12665 + }, + { + "epoch": 1.0793353216872603, + "grad_norm": 15.0, + "learning_rate": 1.102306504150953e-06, + "loss": 0.7345, + "step": 12666 + }, + { + "epoch": 1.0794205368555603, + "grad_norm": 15.25, + "learning_rate": 1.1021396469927517e-06, + "loss": 0.8014, + "step": 12667 + }, + { + "epoch": 1.0795057520238602, + "grad_norm": 19.5, + "learning_rate": 1.1019727925065352e-06, + "loss": 0.4512, + "step": 12668 + }, + { + "epoch": 1.0795909671921602, + "grad_norm": 16.875, + "learning_rate": 1.101805940695319e-06, + "loss": 0.7712, + "step": 12669 + }, + { + "epoch": 1.0796761823604601, + "grad_norm": 12.125, + "learning_rate": 1.1016390915621186e-06, + "loss": 0.4908, + "step": 12670 + }, + { + "epoch": 1.07976139752876, + "grad_norm": 15.0625, + "learning_rate": 1.1014722451099489e-06, + "loss": 0.6974, + "step": 12671 + }, + { + "epoch": 1.07984661269706, + "grad_norm": 15.125, + "learning_rate": 1.101305401341825e-06, + "loss": 0.529, + "step": 12672 + }, + { + "epoch": 1.07993182786536, + "grad_norm": 19.75, + "learning_rate": 1.101138560260762e-06, + "loss": 0.5152, + "step": 12673 + }, + { + "epoch": 1.08001704303366, + "grad_norm": 19.25, + "learning_rate": 1.1009717218697753e-06, + "loss": 0.799, + "step": 12674 + }, + { + "epoch": 1.08010225820196, + "grad_norm": 13.0625, + "learning_rate": 1.1008048861718798e-06, + "loss": 0.3283, + "step": 12675 + }, + { + "epoch": 1.0801874733702599, + "grad_norm": 15.75, + "learning_rate": 1.100638053170089e-06, + "loss": 0.7773, + "step": 12676 + }, + { + "epoch": 1.0802726885385598, + "grad_norm": 20.125, + "learning_rate": 1.1004712228674196e-06, + "loss": 0.6056, + "step": 12677 + }, + { + "epoch": 1.0803579037068598, + "grad_norm": 22.125, + "learning_rate": 1.100304395266885e-06, + "loss": 1.0682, + "step": 12678 + }, + { + "epoch": 1.0804431188751598, + "grad_norm": 11.1875, + "learning_rate": 1.1001375703715008e-06, + "loss": 0.2952, + "step": 12679 + }, + { + "epoch": 1.0805283340434597, + "grad_norm": 45.0, + "learning_rate": 1.0999707481842816e-06, + "loss": 0.8158, + "step": 12680 + }, + { + "epoch": 1.0806135492117597, + "grad_norm": 15.375, + "learning_rate": 1.0998039287082412e-06, + "loss": 0.5468, + "step": 12681 + }, + { + "epoch": 1.0806987643800596, + "grad_norm": 12.4375, + "learning_rate": 1.0996371119463955e-06, + "loss": 0.3753, + "step": 12682 + }, + { + "epoch": 1.0807839795483596, + "grad_norm": 15.125, + "learning_rate": 1.0994702979017584e-06, + "loss": 0.5479, + "step": 12683 + }, + { + "epoch": 1.0808691947166595, + "grad_norm": 14.25, + "learning_rate": 1.099303486577344e-06, + "loss": 0.3836, + "step": 12684 + }, + { + "epoch": 1.0809544098849595, + "grad_norm": 20.125, + "learning_rate": 1.0991366779761673e-06, + "loss": 0.8661, + "step": 12685 + }, + { + "epoch": 1.0810396250532595, + "grad_norm": 15.1875, + "learning_rate": 1.0989698721012424e-06, + "loss": 0.3521, + "step": 12686 + }, + { + "epoch": 1.0811248402215594, + "grad_norm": 27.0, + "learning_rate": 1.098803068955584e-06, + "loss": 0.7299, + "step": 12687 + }, + { + "epoch": 1.0812100553898594, + "grad_norm": 14.6875, + "learning_rate": 1.0986362685422057e-06, + "loss": 0.5865, + "step": 12688 + }, + { + "epoch": 1.0812952705581593, + "grad_norm": 14.4375, + "learning_rate": 1.0984694708641228e-06, + "loss": 0.7113, + "step": 12689 + }, + { + "epoch": 1.0813804857264593, + "grad_norm": 14.6875, + "learning_rate": 1.0983026759243484e-06, + "loss": 0.4307, + "step": 12690 + }, + { + "epoch": 1.0814657008947592, + "grad_norm": 14.75, + "learning_rate": 1.0981358837258979e-06, + "loss": 0.5553, + "step": 12691 + }, + { + "epoch": 1.0815509160630592, + "grad_norm": 16.0, + "learning_rate": 1.0979690942717843e-06, + "loss": 0.3945, + "step": 12692 + }, + { + "epoch": 1.0816361312313592, + "grad_norm": 16.5, + "learning_rate": 1.097802307565022e-06, + "loss": 0.6793, + "step": 12693 + }, + { + "epoch": 1.0817213463996591, + "grad_norm": 34.0, + "learning_rate": 1.0976355236086252e-06, + "loss": 0.7968, + "step": 12694 + }, + { + "epoch": 1.081806561567959, + "grad_norm": 20.25, + "learning_rate": 1.0974687424056076e-06, + "loss": 0.6531, + "step": 12695 + }, + { + "epoch": 1.081891776736259, + "grad_norm": 16.875, + "learning_rate": 1.097301963958983e-06, + "loss": 0.7813, + "step": 12696 + }, + { + "epoch": 1.081976991904559, + "grad_norm": 15.4375, + "learning_rate": 1.0971351882717659e-06, + "loss": 0.5741, + "step": 12697 + }, + { + "epoch": 1.082062207072859, + "grad_norm": 34.75, + "learning_rate": 1.0969684153469692e-06, + "loss": 0.7351, + "step": 12698 + }, + { + "epoch": 1.082147422241159, + "grad_norm": 15.6875, + "learning_rate": 1.0968016451876078e-06, + "loss": 0.6071, + "step": 12699 + }, + { + "epoch": 1.0822326374094589, + "grad_norm": 14.6875, + "learning_rate": 1.0966348777966942e-06, + "loss": 0.5334, + "step": 12700 + }, + { + "epoch": 1.0823178525777588, + "grad_norm": 15.0625, + "learning_rate": 1.0964681131772426e-06, + "loss": 0.4378, + "step": 12701 + }, + { + "epoch": 1.0824030677460588, + "grad_norm": 13.8125, + "learning_rate": 1.0963013513322668e-06, + "loss": 0.5015, + "step": 12702 + }, + { + "epoch": 1.0824882829143587, + "grad_norm": 15.375, + "learning_rate": 1.0961345922647796e-06, + "loss": 0.7422, + "step": 12703 + }, + { + "epoch": 1.0825734980826587, + "grad_norm": 15.1875, + "learning_rate": 1.0959678359777956e-06, + "loss": 0.8333, + "step": 12704 + }, + { + "epoch": 1.0826587132509586, + "grad_norm": 15.1875, + "learning_rate": 1.0958010824743273e-06, + "loss": 0.5575, + "step": 12705 + }, + { + "epoch": 1.0827439284192586, + "grad_norm": 13.8125, + "learning_rate": 1.095634331757389e-06, + "loss": 0.508, + "step": 12706 + }, + { + "epoch": 1.0828291435875586, + "grad_norm": 17.375, + "learning_rate": 1.0954675838299936e-06, + "loss": 0.8037, + "step": 12707 + }, + { + "epoch": 1.0829143587558585, + "grad_norm": 18.125, + "learning_rate": 1.0953008386951538e-06, + "loss": 0.7792, + "step": 12708 + }, + { + "epoch": 1.0829995739241585, + "grad_norm": 39.25, + "learning_rate": 1.0951340963558836e-06, + "loss": 1.1254, + "step": 12709 + }, + { + "epoch": 1.0830847890924584, + "grad_norm": 13.1875, + "learning_rate": 1.0949673568151962e-06, + "loss": 0.4919, + "step": 12710 + }, + { + "epoch": 1.0831700042607584, + "grad_norm": 12.625, + "learning_rate": 1.0948006200761042e-06, + "loss": 0.3435, + "step": 12711 + }, + { + "epoch": 1.0832552194290583, + "grad_norm": 18.25, + "learning_rate": 1.0946338861416213e-06, + "loss": 0.2558, + "step": 12712 + }, + { + "epoch": 1.0833404345973583, + "grad_norm": 13.4375, + "learning_rate": 1.09446715501476e-06, + "loss": 0.5498, + "step": 12713 + }, + { + "epoch": 1.0834256497656582, + "grad_norm": 13.3125, + "learning_rate": 1.094300426698534e-06, + "loss": 0.5399, + "step": 12714 + }, + { + "epoch": 1.0835108649339582, + "grad_norm": 14.375, + "learning_rate": 1.0941337011959562e-06, + "loss": 0.83, + "step": 12715 + }, + { + "epoch": 1.0835960801022582, + "grad_norm": 12.75, + "learning_rate": 1.0939669785100384e-06, + "loss": 0.2942, + "step": 12716 + }, + { + "epoch": 1.0836812952705581, + "grad_norm": 13.0, + "learning_rate": 1.0938002586437947e-06, + "loss": 0.5229, + "step": 12717 + }, + { + "epoch": 1.083766510438858, + "grad_norm": 19.875, + "learning_rate": 1.093633541600237e-06, + "loss": 0.9032, + "step": 12718 + }, + { + "epoch": 1.083851725607158, + "grad_norm": 16.875, + "learning_rate": 1.0934668273823787e-06, + "loss": 0.7821, + "step": 12719 + }, + { + "epoch": 1.083936940775458, + "grad_norm": 15.3125, + "learning_rate": 1.093300115993232e-06, + "loss": 0.6005, + "step": 12720 + }, + { + "epoch": 1.084022155943758, + "grad_norm": 15.875, + "learning_rate": 1.09313340743581e-06, + "loss": 0.1398, + "step": 12721 + }, + { + "epoch": 1.084107371112058, + "grad_norm": 19.5, + "learning_rate": 1.0929667017131248e-06, + "loss": 0.8508, + "step": 12722 + }, + { + "epoch": 1.0841925862803579, + "grad_norm": 16.75, + "learning_rate": 1.0927999988281898e-06, + "loss": 0.7227, + "step": 12723 + }, + { + "epoch": 1.0842778014486578, + "grad_norm": 20.375, + "learning_rate": 1.0926332987840167e-06, + "loss": 0.8254, + "step": 12724 + }, + { + "epoch": 1.0843630166169578, + "grad_norm": 15.8125, + "learning_rate": 1.092466601583618e-06, + "loss": 0.7405, + "step": 12725 + }, + { + "epoch": 1.0844482317852577, + "grad_norm": 14.4375, + "learning_rate": 1.092299907230006e-06, + "loss": 0.5781, + "step": 12726 + }, + { + "epoch": 1.0845334469535577, + "grad_norm": 12.4375, + "learning_rate": 1.0921332157261935e-06, + "loss": 0.54, + "step": 12727 + }, + { + "epoch": 1.0846186621218576, + "grad_norm": 13.5625, + "learning_rate": 1.0919665270751923e-06, + "loss": 0.2435, + "step": 12728 + }, + { + "epoch": 1.0847038772901576, + "grad_norm": 23.625, + "learning_rate": 1.0917998412800151e-06, + "loss": 0.7783, + "step": 12729 + }, + { + "epoch": 1.0847890924584576, + "grad_norm": 17.625, + "learning_rate": 1.0916331583436735e-06, + "loss": 0.9214, + "step": 12730 + }, + { + "epoch": 1.0848743076267575, + "grad_norm": 21.125, + "learning_rate": 1.0914664782691805e-06, + "loss": 0.5946, + "step": 12731 + }, + { + "epoch": 1.0849595227950575, + "grad_norm": 18.0, + "learning_rate": 1.0912998010595477e-06, + "loss": 0.6533, + "step": 12732 + }, + { + "epoch": 1.0850447379633574, + "grad_norm": 15.25, + "learning_rate": 1.0911331267177866e-06, + "loss": 0.6576, + "step": 12733 + }, + { + "epoch": 1.0851299531316574, + "grad_norm": 12.1875, + "learning_rate": 1.09096645524691e-06, + "loss": 0.3513, + "step": 12734 + }, + { + "epoch": 1.0852151682999573, + "grad_norm": 17.25, + "learning_rate": 1.0907997866499293e-06, + "loss": 0.7094, + "step": 12735 + }, + { + "epoch": 1.0853003834682573, + "grad_norm": 12.25, + "learning_rate": 1.0906331209298565e-06, + "loss": 0.4623, + "step": 12736 + }, + { + "epoch": 1.0853855986365573, + "grad_norm": 14.75, + "learning_rate": 1.0904664580897034e-06, + "loss": 0.6455, + "step": 12737 + }, + { + "epoch": 1.0854708138048572, + "grad_norm": 14.6875, + "learning_rate": 1.090299798132482e-06, + "loss": 0.5296, + "step": 12738 + }, + { + "epoch": 1.0855560289731572, + "grad_norm": 25.25, + "learning_rate": 1.0901331410612036e-06, + "loss": 0.8654, + "step": 12739 + }, + { + "epoch": 1.0856412441414571, + "grad_norm": 14.4375, + "learning_rate": 1.089966486878881e-06, + "loss": 0.5028, + "step": 12740 + }, + { + "epoch": 1.085726459309757, + "grad_norm": 16.0, + "learning_rate": 1.089799835588524e-06, + "loss": 0.3618, + "step": 12741 + }, + { + "epoch": 1.085811674478057, + "grad_norm": 12.0, + "learning_rate": 1.0896331871931454e-06, + "loss": 0.4043, + "step": 12742 + }, + { + "epoch": 1.085896889646357, + "grad_norm": 14.75, + "learning_rate": 1.089466541695756e-06, + "loss": 0.679, + "step": 12743 + }, + { + "epoch": 1.085982104814657, + "grad_norm": 21.0, + "learning_rate": 1.089299899099368e-06, + "loss": 0.7628, + "step": 12744 + }, + { + "epoch": 1.086067319982957, + "grad_norm": 17.625, + "learning_rate": 1.089133259406992e-06, + "loss": 0.9788, + "step": 12745 + }, + { + "epoch": 1.0861525351512569, + "grad_norm": 26.625, + "learning_rate": 1.0889666226216404e-06, + "loss": 0.7057, + "step": 12746 + }, + { + "epoch": 1.0862377503195568, + "grad_norm": 33.75, + "learning_rate": 1.0887999887463233e-06, + "loss": 0.8494, + "step": 12747 + }, + { + "epoch": 1.0863229654878568, + "grad_norm": 14.3125, + "learning_rate": 1.088633357784053e-06, + "loss": 0.6497, + "step": 12748 + }, + { + "epoch": 1.0864081806561567, + "grad_norm": 18.875, + "learning_rate": 1.08846672973784e-06, + "loss": 0.8629, + "step": 12749 + }, + { + "epoch": 1.0864933958244567, + "grad_norm": 19.125, + "learning_rate": 1.0883001046106953e-06, + "loss": 0.6444, + "step": 12750 + }, + { + "epoch": 1.0865786109927567, + "grad_norm": 22.625, + "learning_rate": 1.0881334824056307e-06, + "loss": 0.9823, + "step": 12751 + }, + { + "epoch": 1.0866638261610566, + "grad_norm": 15.0, + "learning_rate": 1.0879668631256565e-06, + "loss": 0.6912, + "step": 12752 + }, + { + "epoch": 1.0867490413293566, + "grad_norm": 15.4375, + "learning_rate": 1.0878002467737843e-06, + "loss": 0.5737, + "step": 12753 + }, + { + "epoch": 1.0868342564976565, + "grad_norm": 13.9375, + "learning_rate": 1.0876336333530244e-06, + "loss": 0.5326, + "step": 12754 + }, + { + "epoch": 1.0869194716659565, + "grad_norm": 15.5, + "learning_rate": 1.0874670228663883e-06, + "loss": 0.7217, + "step": 12755 + }, + { + "epoch": 1.0870046868342564, + "grad_norm": 16.25, + "learning_rate": 1.087300415316887e-06, + "loss": 0.6236, + "step": 12756 + }, + { + "epoch": 1.0870899020025564, + "grad_norm": 15.125, + "learning_rate": 1.0871338107075304e-06, + "loss": 0.6636, + "step": 12757 + }, + { + "epoch": 1.0871751171708564, + "grad_norm": 14.0, + "learning_rate": 1.0869672090413294e-06, + "loss": 0.6618, + "step": 12758 + }, + { + "epoch": 1.0872603323391563, + "grad_norm": 13.875, + "learning_rate": 1.086800610321295e-06, + "loss": 0.6011, + "step": 12759 + }, + { + "epoch": 1.0873455475074563, + "grad_norm": 19.75, + "learning_rate": 1.0866340145504376e-06, + "loss": 0.7407, + "step": 12760 + }, + { + "epoch": 1.0874307626757562, + "grad_norm": 14.5625, + "learning_rate": 1.0864674217317678e-06, + "loss": 0.7223, + "step": 12761 + }, + { + "epoch": 1.0875159778440562, + "grad_norm": 16.125, + "learning_rate": 1.086300831868296e-06, + "loss": 0.8548, + "step": 12762 + }, + { + "epoch": 1.0876011930123561, + "grad_norm": 13.1875, + "learning_rate": 1.0861342449630333e-06, + "loss": 0.4573, + "step": 12763 + }, + { + "epoch": 1.087686408180656, + "grad_norm": 14.875, + "learning_rate": 1.0859676610189896e-06, + "loss": 0.5312, + "step": 12764 + }, + { + "epoch": 1.087771623348956, + "grad_norm": 14.875, + "learning_rate": 1.0858010800391746e-06, + "loss": 0.6443, + "step": 12765 + }, + { + "epoch": 1.087856838517256, + "grad_norm": 20.5, + "learning_rate": 1.0856345020265996e-06, + "loss": 0.9176, + "step": 12766 + }, + { + "epoch": 1.087942053685556, + "grad_norm": 14.75, + "learning_rate": 1.0854679269842745e-06, + "loss": 0.5941, + "step": 12767 + }, + { + "epoch": 1.088027268853856, + "grad_norm": 14.0, + "learning_rate": 1.0853013549152092e-06, + "loss": 0.5012, + "step": 12768 + }, + { + "epoch": 1.0881124840221559, + "grad_norm": 12.5625, + "learning_rate": 1.0851347858224138e-06, + "loss": 0.474, + "step": 12769 + }, + { + "epoch": 1.0881976991904558, + "grad_norm": 11.4375, + "learning_rate": 1.084968219708899e-06, + "loss": 0.1327, + "step": 12770 + }, + { + "epoch": 1.0882829143587558, + "grad_norm": 13.625, + "learning_rate": 1.0848016565776742e-06, + "loss": 0.4218, + "step": 12771 + }, + { + "epoch": 1.0883681295270557, + "grad_norm": 17.0, + "learning_rate": 1.0846350964317503e-06, + "loss": 0.6918, + "step": 12772 + }, + { + "epoch": 1.0884533446953557, + "grad_norm": 15.4375, + "learning_rate": 1.0844685392741357e-06, + "loss": 0.6078, + "step": 12773 + }, + { + "epoch": 1.0885385598636557, + "grad_norm": 17.0, + "learning_rate": 1.0843019851078413e-06, + "loss": 0.9734, + "step": 12774 + }, + { + "epoch": 1.0886237750319556, + "grad_norm": 13.875, + "learning_rate": 1.0841354339358763e-06, + "loss": 0.5221, + "step": 12775 + }, + { + "epoch": 1.0887089902002556, + "grad_norm": 13.25, + "learning_rate": 1.0839688857612512e-06, + "loss": 0.5736, + "step": 12776 + }, + { + "epoch": 1.0887942053685555, + "grad_norm": 34.75, + "learning_rate": 1.0838023405869751e-06, + "loss": 0.7536, + "step": 12777 + }, + { + "epoch": 1.0888794205368555, + "grad_norm": 14.0625, + "learning_rate": 1.083635798416058e-06, + "loss": 0.5081, + "step": 12778 + }, + { + "epoch": 1.0889646357051554, + "grad_norm": 16.375, + "learning_rate": 1.0834692592515091e-06, + "loss": 0.8246, + "step": 12779 + }, + { + "epoch": 1.0890498508734554, + "grad_norm": 15.875, + "learning_rate": 1.0833027230963387e-06, + "loss": 0.7068, + "step": 12780 + }, + { + "epoch": 1.0891350660417554, + "grad_norm": 15.375, + "learning_rate": 1.0831361899535553e-06, + "loss": 0.4467, + "step": 12781 + }, + { + "epoch": 1.0892202812100553, + "grad_norm": 11.1875, + "learning_rate": 1.0829696598261686e-06, + "loss": 0.4083, + "step": 12782 + }, + { + "epoch": 1.0893054963783553, + "grad_norm": 14.9375, + "learning_rate": 1.0828031327171883e-06, + "loss": 0.416, + "step": 12783 + }, + { + "epoch": 1.0893907115466552, + "grad_norm": 21.875, + "learning_rate": 1.0826366086296233e-06, + "loss": 0.7159, + "step": 12784 + }, + { + "epoch": 1.0894759267149552, + "grad_norm": 15.0, + "learning_rate": 1.0824700875664834e-06, + "loss": 0.6925, + "step": 12785 + }, + { + "epoch": 1.0895611418832551, + "grad_norm": 18.625, + "learning_rate": 1.082303569530777e-06, + "loss": 0.5471, + "step": 12786 + }, + { + "epoch": 1.089646357051555, + "grad_norm": 12.5625, + "learning_rate": 1.0821370545255142e-06, + "loss": 0.4998, + "step": 12787 + }, + { + "epoch": 1.089731572219855, + "grad_norm": 15.8125, + "learning_rate": 1.0819705425537038e-06, + "loss": 0.8055, + "step": 12788 + }, + { + "epoch": 1.089816787388155, + "grad_norm": 15.9375, + "learning_rate": 1.0818040336183545e-06, + "loss": 0.5433, + "step": 12789 + }, + { + "epoch": 1.089902002556455, + "grad_norm": 15.0625, + "learning_rate": 1.081637527722475e-06, + "loss": 0.6502, + "step": 12790 + }, + { + "epoch": 1.089987217724755, + "grad_norm": 14.0, + "learning_rate": 1.0814710248690751e-06, + "loss": 0.555, + "step": 12791 + }, + { + "epoch": 1.0900724328930549, + "grad_norm": 17.0, + "learning_rate": 1.081304525061163e-06, + "loss": 0.6878, + "step": 12792 + }, + { + "epoch": 1.0901576480613548, + "grad_norm": 10.6875, + "learning_rate": 1.0811380283017479e-06, + "loss": 0.4999, + "step": 12793 + }, + { + "epoch": 1.0902428632296548, + "grad_norm": 24.75, + "learning_rate": 1.0809715345938383e-06, + "loss": 0.8888, + "step": 12794 + }, + { + "epoch": 1.0903280783979548, + "grad_norm": 31.0, + "learning_rate": 1.0808050439404432e-06, + "loss": 0.6933, + "step": 12795 + }, + { + "epoch": 1.0904132935662547, + "grad_norm": 13.5625, + "learning_rate": 1.0806385563445715e-06, + "loss": 0.8258, + "step": 12796 + }, + { + "epoch": 1.0904985087345547, + "grad_norm": 16.5, + "learning_rate": 1.0804720718092306e-06, + "loss": 0.8295, + "step": 12797 + }, + { + "epoch": 1.0905837239028546, + "grad_norm": 16.125, + "learning_rate": 1.0803055903374304e-06, + "loss": 0.5162, + "step": 12798 + }, + { + "epoch": 1.0906689390711546, + "grad_norm": 16.0, + "learning_rate": 1.0801391119321785e-06, + "loss": 0.6569, + "step": 12799 + }, + { + "epoch": 1.0907541542394545, + "grad_norm": 15.1875, + "learning_rate": 1.0799726365964839e-06, + "loss": 0.7901, + "step": 12800 + }, + { + "epoch": 1.0908393694077545, + "grad_norm": 14.25, + "learning_rate": 1.0798061643333544e-06, + "loss": 0.3801, + "step": 12801 + }, + { + "epoch": 1.0909245845760545, + "grad_norm": 14.875, + "learning_rate": 1.0796396951457988e-06, + "loss": 0.498, + "step": 12802 + }, + { + "epoch": 1.0910097997443544, + "grad_norm": 19.25, + "learning_rate": 1.0794732290368252e-06, + "loss": 0.7334, + "step": 12803 + }, + { + "epoch": 1.0910950149126544, + "grad_norm": 18.75, + "learning_rate": 1.079306766009442e-06, + "loss": 0.7772, + "step": 12804 + }, + { + "epoch": 1.0911802300809543, + "grad_norm": 17.75, + "learning_rate": 1.0791403060666573e-06, + "loss": 0.7102, + "step": 12805 + }, + { + "epoch": 1.0912654452492543, + "grad_norm": 14.4375, + "learning_rate": 1.078973849211479e-06, + "loss": 0.5594, + "step": 12806 + }, + { + "epoch": 1.0913506604175542, + "grad_norm": 11.6875, + "learning_rate": 1.078807395446915e-06, + "loss": 0.4549, + "step": 12807 + }, + { + "epoch": 1.0914358755858542, + "grad_norm": 15.25, + "learning_rate": 1.0786409447759737e-06, + "loss": 0.4186, + "step": 12808 + }, + { + "epoch": 1.0915210907541542, + "grad_norm": 17.625, + "learning_rate": 1.0784744972016625e-06, + "loss": 1.0159, + "step": 12809 + }, + { + "epoch": 1.0916063059224541, + "grad_norm": 11.8125, + "learning_rate": 1.0783080527269898e-06, + "loss": 0.3958, + "step": 12810 + }, + { + "epoch": 1.091691521090754, + "grad_norm": 20.875, + "learning_rate": 1.0781416113549633e-06, + "loss": 0.7264, + "step": 12811 + }, + { + "epoch": 1.091776736259054, + "grad_norm": 22.5, + "learning_rate": 1.0779751730885907e-06, + "loss": 0.8182, + "step": 12812 + }, + { + "epoch": 1.091861951427354, + "grad_norm": 20.875, + "learning_rate": 1.07780873793088e-06, + "loss": 0.8166, + "step": 12813 + }, + { + "epoch": 1.091947166595654, + "grad_norm": 12.25, + "learning_rate": 1.077642305884838e-06, + "loss": 0.4081, + "step": 12814 + }, + { + "epoch": 1.092032381763954, + "grad_norm": 16.25, + "learning_rate": 1.0774758769534734e-06, + "loss": 0.7799, + "step": 12815 + }, + { + "epoch": 1.0921175969322539, + "grad_norm": 13.5, + "learning_rate": 1.0773094511397928e-06, + "loss": 0.5278, + "step": 12816 + }, + { + "epoch": 1.0922028121005538, + "grad_norm": 10.6875, + "learning_rate": 1.0771430284468042e-06, + "loss": 0.5195, + "step": 12817 + }, + { + "epoch": 1.0922880272688538, + "grad_norm": 15.8125, + "learning_rate": 1.0769766088775146e-06, + "loss": 0.6012, + "step": 12818 + }, + { + "epoch": 1.0923732424371537, + "grad_norm": 15.6875, + "learning_rate": 1.076810192434932e-06, + "loss": 0.6268, + "step": 12819 + }, + { + "epoch": 1.0924584576054537, + "grad_norm": 17.125, + "learning_rate": 1.0766437791220635e-06, + "loss": 0.7731, + "step": 12820 + }, + { + "epoch": 1.0925436727737536, + "grad_norm": 14.5625, + "learning_rate": 1.0764773689419164e-06, + "loss": 0.4953, + "step": 12821 + }, + { + "epoch": 1.0926288879420536, + "grad_norm": 17.375, + "learning_rate": 1.0763109618974973e-06, + "loss": 0.6061, + "step": 12822 + }, + { + "epoch": 1.0927141031103536, + "grad_norm": 15.9375, + "learning_rate": 1.0761445579918139e-06, + "loss": 0.7843, + "step": 12823 + }, + { + "epoch": 1.0927993182786535, + "grad_norm": 13.3125, + "learning_rate": 1.075978157227873e-06, + "loss": 0.4908, + "step": 12824 + }, + { + "epoch": 1.0928845334469535, + "grad_norm": 15.125, + "learning_rate": 1.0758117596086818e-06, + "loss": 0.5381, + "step": 12825 + }, + { + "epoch": 1.0929697486152534, + "grad_norm": 18.5, + "learning_rate": 1.0756453651372473e-06, + "loss": 0.8341, + "step": 12826 + }, + { + "epoch": 1.0930549637835534, + "grad_norm": 12.6875, + "learning_rate": 1.0754789738165766e-06, + "loss": 0.4798, + "step": 12827 + }, + { + "epoch": 1.0931401789518533, + "grad_norm": 11.75, + "learning_rate": 1.075312585649676e-06, + "loss": 0.4415, + "step": 12828 + }, + { + "epoch": 1.0932253941201533, + "grad_norm": 12.5625, + "learning_rate": 1.075146200639553e-06, + "loss": 0.3664, + "step": 12829 + }, + { + "epoch": 1.0933106092884533, + "grad_norm": 11.4375, + "learning_rate": 1.0749798187892142e-06, + "loss": 0.4136, + "step": 12830 + }, + { + "epoch": 1.0933958244567532, + "grad_norm": 13.4375, + "learning_rate": 1.0748134401016655e-06, + "loss": 0.5164, + "step": 12831 + }, + { + "epoch": 1.0934810396250532, + "grad_norm": 24.5, + "learning_rate": 1.0746470645799142e-06, + "loss": 0.6794, + "step": 12832 + }, + { + "epoch": 1.0935662547933531, + "grad_norm": 16.625, + "learning_rate": 1.0744806922269666e-06, + "loss": 0.6106, + "step": 12833 + }, + { + "epoch": 1.093651469961653, + "grad_norm": 17.375, + "learning_rate": 1.0743143230458298e-06, + "loss": 0.6231, + "step": 12834 + }, + { + "epoch": 1.093736685129953, + "grad_norm": 19.5, + "learning_rate": 1.0741479570395099e-06, + "loss": 0.6513, + "step": 12835 + }, + { + "epoch": 1.093821900298253, + "grad_norm": 9.75, + "learning_rate": 1.0739815942110128e-06, + "loss": 0.5876, + "step": 12836 + }, + { + "epoch": 1.093907115466553, + "grad_norm": 16.25, + "learning_rate": 1.073815234563346e-06, + "loss": 0.5734, + "step": 12837 + }, + { + "epoch": 1.093992330634853, + "grad_norm": 17.875, + "learning_rate": 1.0736488780995148e-06, + "loss": 0.5672, + "step": 12838 + }, + { + "epoch": 1.0940775458031529, + "grad_norm": 15.6875, + "learning_rate": 1.0734825248225252e-06, + "loss": 0.5338, + "step": 12839 + }, + { + "epoch": 1.0941627609714528, + "grad_norm": 12.75, + "learning_rate": 1.0733161747353842e-06, + "loss": 0.3231, + "step": 12840 + }, + { + "epoch": 1.0942479761397528, + "grad_norm": 15.1875, + "learning_rate": 1.0731498278410975e-06, + "loss": 0.8591, + "step": 12841 + }, + { + "epoch": 1.0943331913080527, + "grad_norm": 19.875, + "learning_rate": 1.0729834841426714e-06, + "loss": 0.7308, + "step": 12842 + }, + { + "epoch": 1.0944184064763527, + "grad_norm": 19.0, + "learning_rate": 1.0728171436431114e-06, + "loss": 0.8915, + "step": 12843 + }, + { + "epoch": 1.0945036216446526, + "grad_norm": 15.4375, + "learning_rate": 1.0726508063454242e-06, + "loss": 0.556, + "step": 12844 + }, + { + "epoch": 1.0945888368129526, + "grad_norm": 12.6875, + "learning_rate": 1.0724844722526156e-06, + "loss": 0.5479, + "step": 12845 + }, + { + "epoch": 1.0946740519812526, + "grad_norm": 10.125, + "learning_rate": 1.0723181413676904e-06, + "loss": 0.2236, + "step": 12846 + }, + { + "epoch": 1.0947592671495525, + "grad_norm": 12.0, + "learning_rate": 1.0721518136936555e-06, + "loss": 0.3286, + "step": 12847 + }, + { + "epoch": 1.0948444823178525, + "grad_norm": 14.375, + "learning_rate": 1.0719854892335158e-06, + "loss": 0.6719, + "step": 12848 + }, + { + "epoch": 1.0949296974861524, + "grad_norm": 13.6875, + "learning_rate": 1.0718191679902777e-06, + "loss": 0.6264, + "step": 12849 + }, + { + "epoch": 1.0950149126544524, + "grad_norm": 12.5, + "learning_rate": 1.0716528499669461e-06, + "loss": 0.297, + "step": 12850 + }, + { + "epoch": 1.0951001278227523, + "grad_norm": 13.375, + "learning_rate": 1.0714865351665267e-06, + "loss": 0.6676, + "step": 12851 + }, + { + "epoch": 1.0951853429910523, + "grad_norm": 16.375, + "learning_rate": 1.0713202235920254e-06, + "loss": 0.7653, + "step": 12852 + }, + { + "epoch": 1.0952705581593523, + "grad_norm": 15.8125, + "learning_rate": 1.0711539152464476e-06, + "loss": 0.7831, + "step": 12853 + }, + { + "epoch": 1.0953557733276522, + "grad_norm": 21.875, + "learning_rate": 1.070987610132798e-06, + "loss": 0.5187, + "step": 12854 + }, + { + "epoch": 1.0954409884959522, + "grad_norm": 19.5, + "learning_rate": 1.0708213082540826e-06, + "loss": 0.7532, + "step": 12855 + }, + { + "epoch": 1.0955262036642521, + "grad_norm": 16.5, + "learning_rate": 1.0706550096133058e-06, + "loss": 0.7686, + "step": 12856 + }, + { + "epoch": 1.095611418832552, + "grad_norm": 12.25, + "learning_rate": 1.0704887142134738e-06, + "loss": 0.4692, + "step": 12857 + }, + { + "epoch": 1.095696634000852, + "grad_norm": 9.875, + "learning_rate": 1.0703224220575908e-06, + "loss": 0.1641, + "step": 12858 + }, + { + "epoch": 1.095781849169152, + "grad_norm": 16.5, + "learning_rate": 1.0701561331486627e-06, + "loss": 0.495, + "step": 12859 + }, + { + "epoch": 1.095867064337452, + "grad_norm": 16.125, + "learning_rate": 1.069989847489694e-06, + "loss": 0.4269, + "step": 12860 + }, + { + "epoch": 1.095952279505752, + "grad_norm": 19.625, + "learning_rate": 1.06982356508369e-06, + "loss": 0.8992, + "step": 12861 + }, + { + "epoch": 1.0960374946740519, + "grad_norm": 18.125, + "learning_rate": 1.0696572859336553e-06, + "loss": 0.5528, + "step": 12862 + }, + { + "epoch": 1.0961227098423518, + "grad_norm": 15.5625, + "learning_rate": 1.0694910100425944e-06, + "loss": 0.4664, + "step": 12863 + }, + { + "epoch": 1.0962079250106518, + "grad_norm": 13.8125, + "learning_rate": 1.0693247374135127e-06, + "loss": 0.5167, + "step": 12864 + }, + { + "epoch": 1.0962931401789517, + "grad_norm": 14.4375, + "learning_rate": 1.0691584680494148e-06, + "loss": 0.6103, + "step": 12865 + }, + { + "epoch": 1.0963783553472517, + "grad_norm": 12.9375, + "learning_rate": 1.068992201953305e-06, + "loss": 0.493, + "step": 12866 + }, + { + "epoch": 1.0964635705155517, + "grad_norm": 16.0, + "learning_rate": 1.0688259391281883e-06, + "loss": 0.749, + "step": 12867 + }, + { + "epoch": 1.0965487856838516, + "grad_norm": 11.25, + "learning_rate": 1.068659679577069e-06, + "loss": 0.2643, + "step": 12868 + }, + { + "epoch": 1.0966340008521516, + "grad_norm": 16.25, + "learning_rate": 1.0684934233029517e-06, + "loss": 0.6015, + "step": 12869 + }, + { + "epoch": 1.0967192160204515, + "grad_norm": 15.4375, + "learning_rate": 1.068327170308841e-06, + "loss": 0.7019, + "step": 12870 + }, + { + "epoch": 1.0968044311887515, + "grad_norm": 18.625, + "learning_rate": 1.0681609205977407e-06, + "loss": 0.9605, + "step": 12871 + }, + { + "epoch": 1.0968896463570514, + "grad_norm": 17.75, + "learning_rate": 1.0679946741726553e-06, + "loss": 0.8498, + "step": 12872 + }, + { + "epoch": 1.0969748615253514, + "grad_norm": 15.25, + "learning_rate": 1.0678284310365891e-06, + "loss": 0.4518, + "step": 12873 + }, + { + "epoch": 1.0970600766936514, + "grad_norm": 19.25, + "learning_rate": 1.0676621911925466e-06, + "loss": 0.6102, + "step": 12874 + }, + { + "epoch": 1.0971452918619513, + "grad_norm": 19.125, + "learning_rate": 1.0674959546435315e-06, + "loss": 0.9691, + "step": 12875 + }, + { + "epoch": 1.0972305070302513, + "grad_norm": 16.375, + "learning_rate": 1.0673297213925482e-06, + "loss": 0.55, + "step": 12876 + }, + { + "epoch": 1.0973157221985512, + "grad_norm": 10.3125, + "learning_rate": 1.0671634914426003e-06, + "loss": 0.2295, + "step": 12877 + }, + { + "epoch": 1.0974009373668514, + "grad_norm": 19.25, + "learning_rate": 1.0669972647966924e-06, + "loss": 0.596, + "step": 12878 + }, + { + "epoch": 1.0974861525351514, + "grad_norm": 18.0, + "learning_rate": 1.0668310414578278e-06, + "loss": 0.4683, + "step": 12879 + }, + { + "epoch": 1.0975713677034513, + "grad_norm": 11.125, + "learning_rate": 1.0666648214290104e-06, + "loss": 0.4774, + "step": 12880 + }, + { + "epoch": 1.0976565828717513, + "grad_norm": 14.3125, + "learning_rate": 1.066498604713244e-06, + "loss": 0.4846, + "step": 12881 + }, + { + "epoch": 1.0977417980400512, + "grad_norm": 15.0625, + "learning_rate": 1.0663323913135324e-06, + "loss": 0.6343, + "step": 12882 + }, + { + "epoch": 1.0978270132083512, + "grad_norm": 24.875, + "learning_rate": 1.0661661812328787e-06, + "loss": 1.0339, + "step": 12883 + }, + { + "epoch": 1.0979122283766511, + "grad_norm": 12.875, + "learning_rate": 1.0659999744742874e-06, + "loss": 0.5148, + "step": 12884 + }, + { + "epoch": 1.097997443544951, + "grad_norm": 12.6875, + "learning_rate": 1.0658337710407616e-06, + "loss": 0.5663, + "step": 12885 + }, + { + "epoch": 1.098082658713251, + "grad_norm": 12.0, + "learning_rate": 1.065667570935305e-06, + "loss": 0.2614, + "step": 12886 + }, + { + "epoch": 1.098167873881551, + "grad_norm": 17.375, + "learning_rate": 1.0655013741609205e-06, + "loss": 0.8164, + "step": 12887 + }, + { + "epoch": 1.098253089049851, + "grad_norm": 23.125, + "learning_rate": 1.0653351807206113e-06, + "loss": 0.5434, + "step": 12888 + }, + { + "epoch": 1.098338304218151, + "grad_norm": 19.75, + "learning_rate": 1.0651689906173817e-06, + "loss": 0.6547, + "step": 12889 + }, + { + "epoch": 1.0984235193864509, + "grad_norm": 18.0, + "learning_rate": 1.0650028038542337e-06, + "loss": 0.859, + "step": 12890 + }, + { + "epoch": 1.0985087345547508, + "grad_norm": 16.0, + "learning_rate": 1.0648366204341715e-06, + "loss": 0.6401, + "step": 12891 + }, + { + "epoch": 1.0985939497230508, + "grad_norm": 13.8125, + "learning_rate": 1.0646704403601975e-06, + "loss": 0.5541, + "step": 12892 + }, + { + "epoch": 1.0986791648913508, + "grad_norm": 13.625, + "learning_rate": 1.0645042636353152e-06, + "loss": 0.5803, + "step": 12893 + }, + { + "epoch": 1.0987643800596507, + "grad_norm": 17.0, + "learning_rate": 1.0643380902625277e-06, + "loss": 0.8313, + "step": 12894 + }, + { + "epoch": 1.0988495952279507, + "grad_norm": 15.3125, + "learning_rate": 1.0641719202448373e-06, + "loss": 0.5836, + "step": 12895 + }, + { + "epoch": 1.0989348103962506, + "grad_norm": 14.1875, + "learning_rate": 1.064005753585247e-06, + "loss": 0.4099, + "step": 12896 + }, + { + "epoch": 1.0990200255645506, + "grad_norm": 28.0, + "learning_rate": 1.06383959028676e-06, + "loss": 0.9255, + "step": 12897 + }, + { + "epoch": 1.0991052407328505, + "grad_norm": 37.5, + "learning_rate": 1.0636734303523785e-06, + "loss": 0.8101, + "step": 12898 + }, + { + "epoch": 1.0991904559011505, + "grad_norm": 15.4375, + "learning_rate": 1.0635072737851056e-06, + "loss": 0.5173, + "step": 12899 + }, + { + "epoch": 1.0992756710694505, + "grad_norm": 16.375, + "learning_rate": 1.0633411205879437e-06, + "loss": 0.7854, + "step": 12900 + }, + { + "epoch": 1.0993608862377504, + "grad_norm": 24.125, + "learning_rate": 1.063174970763896e-06, + "loss": 0.9447, + "step": 12901 + }, + { + "epoch": 1.0994461014060504, + "grad_norm": 22.0, + "learning_rate": 1.063008824315964e-06, + "loss": 1.0113, + "step": 12902 + }, + { + "epoch": 1.0995313165743503, + "grad_norm": 17.0, + "learning_rate": 1.0628426812471506e-06, + "loss": 0.6508, + "step": 12903 + }, + { + "epoch": 1.0996165317426503, + "grad_norm": 15.25, + "learning_rate": 1.0626765415604583e-06, + "loss": 0.5525, + "step": 12904 + }, + { + "epoch": 1.0997017469109502, + "grad_norm": 14.5, + "learning_rate": 1.062510405258889e-06, + "loss": 0.6913, + "step": 12905 + }, + { + "epoch": 1.0997869620792502, + "grad_norm": 16.875, + "learning_rate": 1.0623442723454455e-06, + "loss": 0.6473, + "step": 12906 + }, + { + "epoch": 1.0998721772475502, + "grad_norm": 15.0625, + "learning_rate": 1.0621781428231293e-06, + "loss": 0.8004, + "step": 12907 + }, + { + "epoch": 1.09995739241585, + "grad_norm": 18.375, + "learning_rate": 1.0620120166949432e-06, + "loss": 0.8201, + "step": 12908 + }, + { + "epoch": 1.10004260758415, + "grad_norm": 15.4375, + "learning_rate": 1.0618458939638888e-06, + "loss": 0.4078, + "step": 12909 + }, + { + "epoch": 1.10012782275245, + "grad_norm": 18.625, + "learning_rate": 1.0616797746329691e-06, + "loss": 0.8423, + "step": 12910 + }, + { + "epoch": 1.10021303792075, + "grad_norm": 18.375, + "learning_rate": 1.0615136587051845e-06, + "loss": 0.705, + "step": 12911 + }, + { + "epoch": 1.10029825308905, + "grad_norm": 12.375, + "learning_rate": 1.0613475461835379e-06, + "loss": 0.2477, + "step": 12912 + }, + { + "epoch": 1.10038346825735, + "grad_norm": 15.8125, + "learning_rate": 1.0611814370710305e-06, + "loss": 0.7207, + "step": 12913 + }, + { + "epoch": 1.1004686834256499, + "grad_norm": 13.5, + "learning_rate": 1.0610153313706647e-06, + "loss": 0.4509, + "step": 12914 + }, + { + "epoch": 1.1005538985939498, + "grad_norm": 16.125, + "learning_rate": 1.0608492290854417e-06, + "loss": 0.5664, + "step": 12915 + }, + { + "epoch": 1.1006391137622498, + "grad_norm": 16.375, + "learning_rate": 1.0606831302183635e-06, + "loss": 0.4848, + "step": 12916 + }, + { + "epoch": 1.1007243289305497, + "grad_norm": 15.3125, + "learning_rate": 1.0605170347724312e-06, + "loss": 0.4664, + "step": 12917 + }, + { + "epoch": 1.1008095440988497, + "grad_norm": 15.3125, + "learning_rate": 1.0603509427506473e-06, + "loss": 0.4241, + "step": 12918 + }, + { + "epoch": 1.1008947592671496, + "grad_norm": 18.625, + "learning_rate": 1.0601848541560122e-06, + "loss": 0.8123, + "step": 12919 + }, + { + "epoch": 1.1009799744354496, + "grad_norm": 12.8125, + "learning_rate": 1.0600187689915274e-06, + "loss": 0.4523, + "step": 12920 + }, + { + "epoch": 1.1010651896037496, + "grad_norm": 35.75, + "learning_rate": 1.0598526872601948e-06, + "loss": 1.0185, + "step": 12921 + }, + { + "epoch": 1.1011504047720495, + "grad_norm": 12.4375, + "learning_rate": 1.059686608965015e-06, + "loss": 0.4586, + "step": 12922 + }, + { + "epoch": 1.1012356199403495, + "grad_norm": 14.875, + "learning_rate": 1.0595205341089898e-06, + "loss": 0.4927, + "step": 12923 + }, + { + "epoch": 1.1013208351086494, + "grad_norm": 17.25, + "learning_rate": 1.0593544626951199e-06, + "loss": 0.747, + "step": 12924 + }, + { + "epoch": 1.1014060502769494, + "grad_norm": 13.9375, + "learning_rate": 1.0591883947264068e-06, + "loss": 0.7902, + "step": 12925 + }, + { + "epoch": 1.1014912654452493, + "grad_norm": 18.0, + "learning_rate": 1.0590223302058517e-06, + "loss": 0.5812, + "step": 12926 + }, + { + "epoch": 1.1015764806135493, + "grad_norm": 24.375, + "learning_rate": 1.0588562691364547e-06, + "loss": 0.8328, + "step": 12927 + }, + { + "epoch": 1.1016616957818492, + "grad_norm": 13.4375, + "learning_rate": 1.058690211521217e-06, + "loss": 0.4655, + "step": 12928 + }, + { + "epoch": 1.1017469109501492, + "grad_norm": 16.375, + "learning_rate": 1.0585241573631397e-06, + "loss": 0.6824, + "step": 12929 + }, + { + "epoch": 1.1018321261184492, + "grad_norm": 13.875, + "learning_rate": 1.0583581066652232e-06, + "loss": 0.3801, + "step": 12930 + }, + { + "epoch": 1.1019173412867491, + "grad_norm": 10.875, + "learning_rate": 1.0581920594304687e-06, + "loss": 0.1891, + "step": 12931 + }, + { + "epoch": 1.102002556455049, + "grad_norm": 14.625, + "learning_rate": 1.058026015661876e-06, + "loss": 0.6136, + "step": 12932 + }, + { + "epoch": 1.102087771623349, + "grad_norm": 14.875, + "learning_rate": 1.057859975362447e-06, + "loss": 0.6085, + "step": 12933 + }, + { + "epoch": 1.102172986791649, + "grad_norm": 17.75, + "learning_rate": 1.057693938535181e-06, + "loss": 0.6933, + "step": 12934 + }, + { + "epoch": 1.102258201959949, + "grad_norm": 17.375, + "learning_rate": 1.0575279051830793e-06, + "loss": 0.7987, + "step": 12935 + }, + { + "epoch": 1.102343417128249, + "grad_norm": 14.0, + "learning_rate": 1.0573618753091417e-06, + "loss": 0.3716, + "step": 12936 + }, + { + "epoch": 1.1024286322965489, + "grad_norm": 16.75, + "learning_rate": 1.0571958489163685e-06, + "loss": 1.1064, + "step": 12937 + }, + { + "epoch": 1.1025138474648488, + "grad_norm": 13.3125, + "learning_rate": 1.0570298260077605e-06, + "loss": 0.5074, + "step": 12938 + }, + { + "epoch": 1.1025990626331488, + "grad_norm": 14.9375, + "learning_rate": 1.0568638065863175e-06, + "loss": 0.4701, + "step": 12939 + }, + { + "epoch": 1.1026842778014487, + "grad_norm": 14.5625, + "learning_rate": 1.0566977906550396e-06, + "loss": 0.4696, + "step": 12940 + }, + { + "epoch": 1.1027694929697487, + "grad_norm": 14.125, + "learning_rate": 1.0565317782169269e-06, + "loss": 0.9267, + "step": 12941 + }, + { + "epoch": 1.1028547081380486, + "grad_norm": 13.875, + "learning_rate": 1.05636576927498e-06, + "loss": 0.5396, + "step": 12942 + }, + { + "epoch": 1.1029399233063486, + "grad_norm": 17.0, + "learning_rate": 1.0561997638321983e-06, + "loss": 0.6771, + "step": 12943 + }, + { + "epoch": 1.1030251384746486, + "grad_norm": 12.9375, + "learning_rate": 1.0560337618915815e-06, + "loss": 0.586, + "step": 12944 + }, + { + "epoch": 1.1031103536429485, + "grad_norm": 17.625, + "learning_rate": 1.0558677634561295e-06, + "loss": 0.5079, + "step": 12945 + }, + { + "epoch": 1.1031955688112485, + "grad_norm": 12.875, + "learning_rate": 1.0557017685288427e-06, + "loss": 0.2925, + "step": 12946 + }, + { + "epoch": 1.1032807839795484, + "grad_norm": 15.5625, + "learning_rate": 1.0555357771127197e-06, + "loss": 0.7335, + "step": 12947 + }, + { + "epoch": 1.1033659991478484, + "grad_norm": 16.5, + "learning_rate": 1.0553697892107614e-06, + "loss": 0.9047, + "step": 12948 + }, + { + "epoch": 1.1034512143161483, + "grad_norm": 16.0, + "learning_rate": 1.055203804825966e-06, + "loss": 0.4865, + "step": 12949 + }, + { + "epoch": 1.1035364294844483, + "grad_norm": 12.75, + "learning_rate": 1.0550378239613343e-06, + "loss": 0.4186, + "step": 12950 + }, + { + "epoch": 1.1036216446527483, + "grad_norm": 13.5, + "learning_rate": 1.0548718466198655e-06, + "loss": 0.5239, + "step": 12951 + }, + { + "epoch": 1.1037068598210482, + "grad_norm": 18.5, + "learning_rate": 1.054705872804558e-06, + "loss": 0.6996, + "step": 12952 + }, + { + "epoch": 1.1037920749893482, + "grad_norm": 13.9375, + "learning_rate": 1.0545399025184118e-06, + "loss": 0.4461, + "step": 12953 + }, + { + "epoch": 1.1038772901576481, + "grad_norm": 13.875, + "learning_rate": 1.054373935764426e-06, + "loss": 0.4231, + "step": 12954 + }, + { + "epoch": 1.103962505325948, + "grad_norm": 14.75, + "learning_rate": 1.0542079725456005e-06, + "loss": 0.6483, + "step": 12955 + }, + { + "epoch": 1.104047720494248, + "grad_norm": 18.75, + "learning_rate": 1.0540420128649332e-06, + "loss": 1.0375, + "step": 12956 + }, + { + "epoch": 1.104132935662548, + "grad_norm": 16.25, + "learning_rate": 1.0538760567254242e-06, + "loss": 0.6338, + "step": 12957 + }, + { + "epoch": 1.104218150830848, + "grad_norm": 15.625, + "learning_rate": 1.053710104130072e-06, + "loss": 0.5914, + "step": 12958 + }, + { + "epoch": 1.104303365999148, + "grad_norm": 12.125, + "learning_rate": 1.0535441550818758e-06, + "loss": 0.4574, + "step": 12959 + }, + { + "epoch": 1.1043885811674479, + "grad_norm": 21.375, + "learning_rate": 1.053378209583834e-06, + "loss": 0.7857, + "step": 12960 + }, + { + "epoch": 1.1044737963357478, + "grad_norm": 16.875, + "learning_rate": 1.0532122676389456e-06, + "loss": 0.905, + "step": 12961 + }, + { + "epoch": 1.1045590115040478, + "grad_norm": 21.0, + "learning_rate": 1.0530463292502095e-06, + "loss": 0.8267, + "step": 12962 + }, + { + "epoch": 1.1046442266723477, + "grad_norm": 11.8125, + "learning_rate": 1.0528803944206242e-06, + "loss": 0.4209, + "step": 12963 + }, + { + "epoch": 1.1047294418406477, + "grad_norm": 12.0, + "learning_rate": 1.0527144631531882e-06, + "loss": 1.0379, + "step": 12964 + }, + { + "epoch": 1.1048146570089477, + "grad_norm": 13.875, + "learning_rate": 1.0525485354509007e-06, + "loss": 0.2531, + "step": 12965 + }, + { + "epoch": 1.1048998721772476, + "grad_norm": 11.8125, + "learning_rate": 1.0523826113167592e-06, + "loss": 0.4102, + "step": 12966 + }, + { + "epoch": 1.1049850873455476, + "grad_norm": 12.875, + "learning_rate": 1.0522166907537632e-06, + "loss": 0.4423, + "step": 12967 + }, + { + "epoch": 1.1050703025138475, + "grad_norm": 13.375, + "learning_rate": 1.0520507737649105e-06, + "loss": 0.3806, + "step": 12968 + }, + { + "epoch": 1.1051555176821475, + "grad_norm": 12.6875, + "learning_rate": 1.051884860353199e-06, + "loss": 0.4255, + "step": 12969 + }, + { + "epoch": 1.1052407328504474, + "grad_norm": 13.875, + "learning_rate": 1.0517189505216272e-06, + "loss": 0.6645, + "step": 12970 + }, + { + "epoch": 1.1053259480187474, + "grad_norm": 12.0, + "learning_rate": 1.0515530442731935e-06, + "loss": 0.4232, + "step": 12971 + }, + { + "epoch": 1.1054111631870474, + "grad_norm": 36.25, + "learning_rate": 1.0513871416108959e-06, + "loss": 0.8509, + "step": 12972 + }, + { + "epoch": 1.1054963783553473, + "grad_norm": 13.4375, + "learning_rate": 1.0512212425377325e-06, + "loss": 0.5004, + "step": 12973 + }, + { + "epoch": 1.1055815935236473, + "grad_norm": 19.5, + "learning_rate": 1.0510553470567007e-06, + "loss": 0.8302, + "step": 12974 + }, + { + "epoch": 1.1056668086919472, + "grad_norm": 12.0, + "learning_rate": 1.0508894551707997e-06, + "loss": 0.3675, + "step": 12975 + }, + { + "epoch": 1.1057520238602472, + "grad_norm": 13.875, + "learning_rate": 1.0507235668830258e-06, + "loss": 0.5378, + "step": 12976 + }, + { + "epoch": 1.1058372390285471, + "grad_norm": 15.5, + "learning_rate": 1.0505576821963774e-06, + "loss": 0.7414, + "step": 12977 + }, + { + "epoch": 1.105922454196847, + "grad_norm": 9.125, + "learning_rate": 1.0503918011138524e-06, + "loss": 0.2718, + "step": 12978 + }, + { + "epoch": 1.106007669365147, + "grad_norm": 19.0, + "learning_rate": 1.0502259236384482e-06, + "loss": 1.2556, + "step": 12979 + }, + { + "epoch": 1.106092884533447, + "grad_norm": 13.375, + "learning_rate": 1.0500600497731625e-06, + "loss": 0.5361, + "step": 12980 + }, + { + "epoch": 1.106178099701747, + "grad_norm": 15.0625, + "learning_rate": 1.0498941795209924e-06, + "loss": 0.5595, + "step": 12981 + }, + { + "epoch": 1.106263314870047, + "grad_norm": 18.125, + "learning_rate": 1.049728312884936e-06, + "loss": 0.5345, + "step": 12982 + }, + { + "epoch": 1.1063485300383469, + "grad_norm": 13.25, + "learning_rate": 1.049562449867991e-06, + "loss": 0.3807, + "step": 12983 + }, + { + "epoch": 1.1064337452066468, + "grad_norm": 16.375, + "learning_rate": 1.0493965904731532e-06, + "loss": 0.4706, + "step": 12984 + }, + { + "epoch": 1.1065189603749468, + "grad_norm": 13.4375, + "learning_rate": 1.0492307347034212e-06, + "loss": 0.3667, + "step": 12985 + }, + { + "epoch": 1.1066041755432467, + "grad_norm": 14.5, + "learning_rate": 1.0490648825617913e-06, + "loss": 0.606, + "step": 12986 + }, + { + "epoch": 1.1066893907115467, + "grad_norm": 23.875, + "learning_rate": 1.0488990340512615e-06, + "loss": 1.1199, + "step": 12987 + }, + { + "epoch": 1.1067746058798467, + "grad_norm": 16.25, + "learning_rate": 1.048733189174828e-06, + "loss": 0.7381, + "step": 12988 + }, + { + "epoch": 1.1068598210481466, + "grad_norm": 13.875, + "learning_rate": 1.0485673479354882e-06, + "loss": 0.4409, + "step": 12989 + }, + { + "epoch": 1.1069450362164466, + "grad_norm": 12.75, + "learning_rate": 1.0484015103362392e-06, + "loss": 0.4177, + "step": 12990 + }, + { + "epoch": 1.1070302513847465, + "grad_norm": 16.375, + "learning_rate": 1.048235676380078e-06, + "loss": 0.683, + "step": 12991 + }, + { + "epoch": 1.1071154665530465, + "grad_norm": 20.25, + "learning_rate": 1.0480698460700003e-06, + "loss": 1.0075, + "step": 12992 + }, + { + "epoch": 1.1072006817213464, + "grad_norm": 18.625, + "learning_rate": 1.0479040194090039e-06, + "loss": 0.767, + "step": 12993 + }, + { + "epoch": 1.1072858968896464, + "grad_norm": 18.0, + "learning_rate": 1.0477381964000848e-06, + "loss": 0.5861, + "step": 12994 + }, + { + "epoch": 1.1073711120579464, + "grad_norm": 15.25, + "learning_rate": 1.0475723770462402e-06, + "loss": 0.7565, + "step": 12995 + }, + { + "epoch": 1.1074563272262463, + "grad_norm": 15.75, + "learning_rate": 1.047406561350466e-06, + "loss": 0.7408, + "step": 12996 + }, + { + "epoch": 1.1075415423945463, + "grad_norm": 15.375, + "learning_rate": 1.047240749315759e-06, + "loss": 0.534, + "step": 12997 + }, + { + "epoch": 1.1076267575628462, + "grad_norm": 14.0, + "learning_rate": 1.0470749409451156e-06, + "loss": 0.7307, + "step": 12998 + }, + { + "epoch": 1.1077119727311462, + "grad_norm": 16.125, + "learning_rate": 1.0469091362415323e-06, + "loss": 0.4105, + "step": 12999 + }, + { + "epoch": 1.1077971878994461, + "grad_norm": 16.375, + "learning_rate": 1.046743335208005e-06, + "loss": 0.6593, + "step": 13000 + }, + { + "epoch": 1.107882403067746, + "grad_norm": 16.5, + "learning_rate": 1.04657753784753e-06, + "loss": 0.5729, + "step": 13001 + }, + { + "epoch": 1.107967618236046, + "grad_norm": 14.3125, + "learning_rate": 1.0464117441631035e-06, + "loss": 0.5392, + "step": 13002 + }, + { + "epoch": 1.108052833404346, + "grad_norm": 15.6875, + "learning_rate": 1.0462459541577214e-06, + "loss": 0.8327, + "step": 13003 + }, + { + "epoch": 1.108138048572646, + "grad_norm": 29.625, + "learning_rate": 1.0460801678343797e-06, + "loss": 1.6851, + "step": 13004 + }, + { + "epoch": 1.108223263740946, + "grad_norm": 20.625, + "learning_rate": 1.0459143851960748e-06, + "loss": 0.6165, + "step": 13005 + }, + { + "epoch": 1.1083084789092459, + "grad_norm": 28.5, + "learning_rate": 1.0457486062458017e-06, + "loss": 0.9852, + "step": 13006 + }, + { + "epoch": 1.1083936940775458, + "grad_norm": 20.375, + "learning_rate": 1.045582830986557e-06, + "loss": 0.9211, + "step": 13007 + }, + { + "epoch": 1.1084789092458458, + "grad_norm": 17.125, + "learning_rate": 1.0454170594213363e-06, + "loss": 0.8565, + "step": 13008 + }, + { + "epoch": 1.1085641244141458, + "grad_norm": 15.0625, + "learning_rate": 1.0452512915531347e-06, + "loss": 0.3924, + "step": 13009 + }, + { + "epoch": 1.1086493395824457, + "grad_norm": 17.25, + "learning_rate": 1.0450855273849484e-06, + "loss": 0.6035, + "step": 13010 + }, + { + "epoch": 1.1087345547507457, + "grad_norm": 21.625, + "learning_rate": 1.0449197669197722e-06, + "loss": 0.7959, + "step": 13011 + }, + { + "epoch": 1.1088197699190456, + "grad_norm": 15.8125, + "learning_rate": 1.0447540101606025e-06, + "loss": 0.7922, + "step": 13012 + }, + { + "epoch": 1.1089049850873456, + "grad_norm": 14.0625, + "learning_rate": 1.0445882571104338e-06, + "loss": 0.6423, + "step": 13013 + }, + { + "epoch": 1.1089902002556455, + "grad_norm": 25.25, + "learning_rate": 1.0444225077722623e-06, + "loss": 0.9106, + "step": 13014 + }, + { + "epoch": 1.1090754154239455, + "grad_norm": 13.875, + "learning_rate": 1.0442567621490824e-06, + "loss": 0.6706, + "step": 13015 + }, + { + "epoch": 1.1091606305922455, + "grad_norm": 19.125, + "learning_rate": 1.0440910202438903e-06, + "loss": 0.5476, + "step": 13016 + }, + { + "epoch": 1.1092458457605454, + "grad_norm": 11.875, + "learning_rate": 1.0439252820596803e-06, + "loss": 0.3312, + "step": 13017 + }, + { + "epoch": 1.1093310609288454, + "grad_norm": 18.625, + "learning_rate": 1.0437595475994477e-06, + "loss": 0.4511, + "step": 13018 + }, + { + "epoch": 1.1094162760971453, + "grad_norm": 10.8125, + "learning_rate": 1.0435938168661872e-06, + "loss": 0.3428, + "step": 13019 + }, + { + "epoch": 1.1095014912654453, + "grad_norm": 16.0, + "learning_rate": 1.043428089862894e-06, + "loss": 0.7999, + "step": 13020 + }, + { + "epoch": 1.1095867064337452, + "grad_norm": 20.125, + "learning_rate": 1.043262366592563e-06, + "loss": 0.3464, + "step": 13021 + }, + { + "epoch": 1.1096719216020452, + "grad_norm": 12.875, + "learning_rate": 1.0430966470581889e-06, + "loss": 0.3667, + "step": 13022 + }, + { + "epoch": 1.1097571367703452, + "grad_norm": 19.625, + "learning_rate": 1.0429309312627664e-06, + "loss": 0.6985, + "step": 13023 + }, + { + "epoch": 1.109842351938645, + "grad_norm": 26.0, + "learning_rate": 1.0427652192092908e-06, + "loss": 1.0088, + "step": 13024 + }, + { + "epoch": 1.109927567106945, + "grad_norm": 16.125, + "learning_rate": 1.0425995109007555e-06, + "loss": 0.6296, + "step": 13025 + }, + { + "epoch": 1.110012782275245, + "grad_norm": 15.4375, + "learning_rate": 1.0424338063401555e-06, + "loss": 0.6085, + "step": 13026 + }, + { + "epoch": 1.110097997443545, + "grad_norm": 12.375, + "learning_rate": 1.0422681055304857e-06, + "loss": 0.4145, + "step": 13027 + }, + { + "epoch": 1.110183212611845, + "grad_norm": 15.875, + "learning_rate": 1.0421024084747397e-06, + "loss": 0.5407, + "step": 13028 + }, + { + "epoch": 1.110268427780145, + "grad_norm": 19.75, + "learning_rate": 1.0419367151759127e-06, + "loss": 0.8379, + "step": 13029 + }, + { + "epoch": 1.1103536429484449, + "grad_norm": 21.125, + "learning_rate": 1.041771025636998e-06, + "loss": 0.8537, + "step": 13030 + }, + { + "epoch": 1.1104388581167448, + "grad_norm": 12.625, + "learning_rate": 1.0416053398609907e-06, + "loss": 0.4359, + "step": 13031 + }, + { + "epoch": 1.1105240732850448, + "grad_norm": 19.5, + "learning_rate": 1.0414396578508845e-06, + "loss": 0.5, + "step": 13032 + }, + { + "epoch": 1.1106092884533447, + "grad_norm": 20.375, + "learning_rate": 1.0412739796096732e-06, + "loss": 0.5673, + "step": 13033 + }, + { + "epoch": 1.1106945036216447, + "grad_norm": 17.375, + "learning_rate": 1.0411083051403508e-06, + "loss": 0.7234, + "step": 13034 + }, + { + "epoch": 1.1107797187899446, + "grad_norm": 13.9375, + "learning_rate": 1.0409426344459117e-06, + "loss": 0.5188, + "step": 13035 + }, + { + "epoch": 1.1108649339582446, + "grad_norm": 13.9375, + "learning_rate": 1.040776967529349e-06, + "loss": 0.5039, + "step": 13036 + }, + { + "epoch": 1.1109501491265446, + "grad_norm": 19.125, + "learning_rate": 1.0406113043936573e-06, + "loss": 1.0194, + "step": 13037 + }, + { + "epoch": 1.1110353642948445, + "grad_norm": 12.8125, + "learning_rate": 1.0404456450418296e-06, + "loss": 0.4799, + "step": 13038 + }, + { + "epoch": 1.1111205794631445, + "grad_norm": 11.875, + "learning_rate": 1.04027998947686e-06, + "loss": 0.3237, + "step": 13039 + }, + { + "epoch": 1.1112057946314444, + "grad_norm": 13.8125, + "learning_rate": 1.0401143377017423e-06, + "loss": 0.1946, + "step": 13040 + }, + { + "epoch": 1.1112910097997444, + "grad_norm": 11.375, + "learning_rate": 1.0399486897194688e-06, + "loss": 0.4274, + "step": 13041 + }, + { + "epoch": 1.1113762249680443, + "grad_norm": 13.625, + "learning_rate": 1.039783045533034e-06, + "loss": 0.442, + "step": 13042 + }, + { + "epoch": 1.1114614401363443, + "grad_norm": 11.6875, + "learning_rate": 1.039617405145431e-06, + "loss": 0.3231, + "step": 13043 + }, + { + "epoch": 1.1115466553046442, + "grad_norm": 18.0, + "learning_rate": 1.039451768559653e-06, + "loss": 0.786, + "step": 13044 + }, + { + "epoch": 1.1116318704729442, + "grad_norm": 28.5, + "learning_rate": 1.0392861357786932e-06, + "loss": 0.4373, + "step": 13045 + }, + { + "epoch": 1.1117170856412442, + "grad_norm": 13.25, + "learning_rate": 1.0391205068055448e-06, + "loss": 0.6301, + "step": 13046 + }, + { + "epoch": 1.1118023008095441, + "grad_norm": 18.5, + "learning_rate": 1.0389548816432009e-06, + "loss": 0.513, + "step": 13047 + }, + { + "epoch": 1.111887515977844, + "grad_norm": 18.25, + "learning_rate": 1.0387892602946551e-06, + "loss": 0.9109, + "step": 13048 + }, + { + "epoch": 1.111972731146144, + "grad_norm": 16.25, + "learning_rate": 1.0386236427628994e-06, + "loss": 1.0801, + "step": 13049 + }, + { + "epoch": 1.112057946314444, + "grad_norm": 14.25, + "learning_rate": 1.0384580290509268e-06, + "loss": 0.461, + "step": 13050 + }, + { + "epoch": 1.112143161482744, + "grad_norm": 23.25, + "learning_rate": 1.0382924191617301e-06, + "loss": 0.5671, + "step": 13051 + }, + { + "epoch": 1.112228376651044, + "grad_norm": 10.5625, + "learning_rate": 1.0381268130983029e-06, + "loss": 0.3289, + "step": 13052 + }, + { + "epoch": 1.1123135918193439, + "grad_norm": 16.75, + "learning_rate": 1.0379612108636367e-06, + "loss": 0.6262, + "step": 13053 + }, + { + "epoch": 1.1123988069876438, + "grad_norm": 17.375, + "learning_rate": 1.037795612460725e-06, + "loss": 0.7017, + "step": 13054 + }, + { + "epoch": 1.1124840221559438, + "grad_norm": 15.1875, + "learning_rate": 1.0376300178925597e-06, + "loss": 0.4279, + "step": 13055 + }, + { + "epoch": 1.1125692373242437, + "grad_norm": 18.5, + "learning_rate": 1.037464427162134e-06, + "loss": 0.5531, + "step": 13056 + }, + { + "epoch": 1.1126544524925437, + "grad_norm": 13.0, + "learning_rate": 1.0372988402724395e-06, + "loss": 0.5188, + "step": 13057 + }, + { + "epoch": 1.1127396676608436, + "grad_norm": 11.375, + "learning_rate": 1.0371332572264687e-06, + "loss": 0.2751, + "step": 13058 + }, + { + "epoch": 1.1128248828291436, + "grad_norm": 12.375, + "learning_rate": 1.0369676780272142e-06, + "loss": 0.4019, + "step": 13059 + }, + { + "epoch": 1.1129100979974436, + "grad_norm": 13.8125, + "learning_rate": 1.0368021026776678e-06, + "loss": 0.5013, + "step": 13060 + }, + { + "epoch": 1.1129953131657435, + "grad_norm": 21.0, + "learning_rate": 1.036636531180822e-06, + "loss": 1.0437, + "step": 13061 + }, + { + "epoch": 1.1130805283340435, + "grad_norm": 22.0, + "learning_rate": 1.0364709635396684e-06, + "loss": 0.5727, + "step": 13062 + }, + { + "epoch": 1.1131657435023434, + "grad_norm": 16.125, + "learning_rate": 1.0363053997571995e-06, + "loss": 0.6494, + "step": 13063 + }, + { + "epoch": 1.1132509586706434, + "grad_norm": 14.8125, + "learning_rate": 1.036139839836407e-06, + "loss": 0.6923, + "step": 13064 + }, + { + "epoch": 1.1133361738389433, + "grad_norm": 13.375, + "learning_rate": 1.0359742837802826e-06, + "loss": 0.3933, + "step": 13065 + }, + { + "epoch": 1.1134213890072433, + "grad_norm": 15.3125, + "learning_rate": 1.0358087315918178e-06, + "loss": 0.5149, + "step": 13066 + }, + { + "epoch": 1.1135066041755433, + "grad_norm": 14.1875, + "learning_rate": 1.0356431832740047e-06, + "loss": 0.5284, + "step": 13067 + }, + { + "epoch": 1.1135918193438432, + "grad_norm": 16.125, + "learning_rate": 1.0354776388298346e-06, + "loss": 0.4593, + "step": 13068 + }, + { + "epoch": 1.1136770345121432, + "grad_norm": 11.5, + "learning_rate": 1.0353120982622998e-06, + "loss": 0.3664, + "step": 13069 + }, + { + "epoch": 1.1137622496804431, + "grad_norm": 16.125, + "learning_rate": 1.035146561574391e-06, + "loss": 0.569, + "step": 13070 + }, + { + "epoch": 1.113847464848743, + "grad_norm": 15.5, + "learning_rate": 1.0349810287690998e-06, + "loss": 0.4902, + "step": 13071 + }, + { + "epoch": 1.113932680017043, + "grad_norm": 16.25, + "learning_rate": 1.0348154998494176e-06, + "loss": 0.6578, + "step": 13072 + }, + { + "epoch": 1.114017895185343, + "grad_norm": 16.125, + "learning_rate": 1.0346499748183361e-06, + "loss": 0.4136, + "step": 13073 + }, + { + "epoch": 1.114103110353643, + "grad_norm": 12.625, + "learning_rate": 1.0344844536788461e-06, + "loss": 0.4552, + "step": 13074 + }, + { + "epoch": 1.114188325521943, + "grad_norm": 22.75, + "learning_rate": 1.0343189364339382e-06, + "loss": 0.7652, + "step": 13075 + }, + { + "epoch": 1.1142735406902429, + "grad_norm": 59.0, + "learning_rate": 1.0341534230866044e-06, + "loss": 0.6078, + "step": 13076 + }, + { + "epoch": 1.1143587558585428, + "grad_norm": 15.0, + "learning_rate": 1.033987913639835e-06, + "loss": 0.4455, + "step": 13077 + }, + { + "epoch": 1.1144439710268428, + "grad_norm": 14.625, + "learning_rate": 1.0338224080966216e-06, + "loss": 0.4727, + "step": 13078 + }, + { + "epoch": 1.1145291861951427, + "grad_norm": 15.875, + "learning_rate": 1.0336569064599542e-06, + "loss": 0.6407, + "step": 13079 + }, + { + "epoch": 1.1146144013634427, + "grad_norm": 16.125, + "learning_rate": 1.0334914087328244e-06, + "loss": 0.616, + "step": 13080 + }, + { + "epoch": 1.1146996165317427, + "grad_norm": 18.875, + "learning_rate": 1.0333259149182228e-06, + "loss": 0.887, + "step": 13081 + }, + { + "epoch": 1.1147848317000426, + "grad_norm": 17.375, + "learning_rate": 1.0331604250191396e-06, + "loss": 0.5509, + "step": 13082 + }, + { + "epoch": 1.1148700468683426, + "grad_norm": 14.0, + "learning_rate": 1.0329949390385653e-06, + "loss": 0.5918, + "step": 13083 + }, + { + "epoch": 1.1149552620366425, + "grad_norm": 24.0, + "learning_rate": 1.032829456979491e-06, + "loss": 1.0502, + "step": 13084 + }, + { + "epoch": 1.1150404772049425, + "grad_norm": 13.8125, + "learning_rate": 1.0326639788449063e-06, + "loss": 0.7153, + "step": 13085 + }, + { + "epoch": 1.1151256923732424, + "grad_norm": 13.8125, + "learning_rate": 1.0324985046378024e-06, + "loss": 0.5648, + "step": 13086 + }, + { + "epoch": 1.1152109075415424, + "grad_norm": 16.875, + "learning_rate": 1.0323330343611692e-06, + "loss": 0.3868, + "step": 13087 + }, + { + "epoch": 1.1152961227098424, + "grad_norm": 17.25, + "learning_rate": 1.032167568017997e-06, + "loss": 0.598, + "step": 13088 + }, + { + "epoch": 1.1153813378781423, + "grad_norm": 12.9375, + "learning_rate": 1.032002105611276e-06, + "loss": 0.5334, + "step": 13089 + }, + { + "epoch": 1.1154665530464423, + "grad_norm": 16.125, + "learning_rate": 1.0318366471439959e-06, + "loss": 0.7744, + "step": 13090 + }, + { + "epoch": 1.1155517682147422, + "grad_norm": 11.9375, + "learning_rate": 1.0316711926191473e-06, + "loss": 0.3883, + "step": 13091 + }, + { + "epoch": 1.1156369833830422, + "grad_norm": 14.4375, + "learning_rate": 1.0315057420397193e-06, + "loss": 0.8604, + "step": 13092 + }, + { + "epoch": 1.1157221985513421, + "grad_norm": 20.75, + "learning_rate": 1.0313402954087026e-06, + "loss": 1.0109, + "step": 13093 + }, + { + "epoch": 1.115807413719642, + "grad_norm": 13.6875, + "learning_rate": 1.0311748527290864e-06, + "loss": 0.3822, + "step": 13094 + }, + { + "epoch": 1.115892628887942, + "grad_norm": 16.0, + "learning_rate": 1.031009414003861e-06, + "loss": 0.4772, + "step": 13095 + }, + { + "epoch": 1.115977844056242, + "grad_norm": 15.9375, + "learning_rate": 1.0308439792360153e-06, + "loss": 0.9222, + "step": 13096 + }, + { + "epoch": 1.116063059224542, + "grad_norm": 21.0, + "learning_rate": 1.03067854842854e-06, + "loss": 0.353, + "step": 13097 + }, + { + "epoch": 1.116148274392842, + "grad_norm": 12.5625, + "learning_rate": 1.0305131215844233e-06, + "loss": 0.2562, + "step": 13098 + }, + { + "epoch": 1.1162334895611419, + "grad_norm": 12.875, + "learning_rate": 1.0303476987066554e-06, + "loss": 0.4775, + "step": 13099 + }, + { + "epoch": 1.1163187047294418, + "grad_norm": 15.1875, + "learning_rate": 1.030182279798225e-06, + "loss": 0.734, + "step": 13100 + }, + { + "epoch": 1.1164039198977418, + "grad_norm": 16.25, + "learning_rate": 1.0300168648621223e-06, + "loss": 0.319, + "step": 13101 + }, + { + "epoch": 1.1164891350660417, + "grad_norm": 25.125, + "learning_rate": 1.0298514539013357e-06, + "loss": 0.6406, + "step": 13102 + }, + { + "epoch": 1.1165743502343417, + "grad_norm": 14.625, + "learning_rate": 1.0296860469188553e-06, + "loss": 0.6202, + "step": 13103 + }, + { + "epoch": 1.1166595654026417, + "grad_norm": 14.875, + "learning_rate": 1.0295206439176688e-06, + "loss": 0.6371, + "step": 13104 + }, + { + "epoch": 1.1167447805709416, + "grad_norm": 13.625, + "learning_rate": 1.029355244900767e-06, + "loss": 0.3767, + "step": 13105 + }, + { + "epoch": 1.1168299957392416, + "grad_norm": 16.125, + "learning_rate": 1.0291898498711373e-06, + "loss": 0.7552, + "step": 13106 + }, + { + "epoch": 1.1169152109075415, + "grad_norm": 20.625, + "learning_rate": 1.0290244588317686e-06, + "loss": 0.5777, + "step": 13107 + }, + { + "epoch": 1.1170004260758415, + "grad_norm": 14.75, + "learning_rate": 1.0288590717856507e-06, + "loss": 0.4842, + "step": 13108 + }, + { + "epoch": 1.1170856412441414, + "grad_norm": 14.8125, + "learning_rate": 1.0286936887357713e-06, + "loss": 0.5881, + "step": 13109 + }, + { + "epoch": 1.1171708564124414, + "grad_norm": 11.8125, + "learning_rate": 1.0285283096851197e-06, + "loss": 0.3175, + "step": 13110 + }, + { + "epoch": 1.1172560715807414, + "grad_norm": 16.875, + "learning_rate": 1.0283629346366842e-06, + "loss": 0.6931, + "step": 13111 + }, + { + "epoch": 1.1173412867490413, + "grad_norm": 13.4375, + "learning_rate": 1.0281975635934534e-06, + "loss": 0.4789, + "step": 13112 + }, + { + "epoch": 1.1174265019173413, + "grad_norm": 14.625, + "learning_rate": 1.028032196558416e-06, + "loss": 0.5222, + "step": 13113 + }, + { + "epoch": 1.1175117170856412, + "grad_norm": 28.875, + "learning_rate": 1.0278668335345597e-06, + "loss": 0.8614, + "step": 13114 + }, + { + "epoch": 1.1175969322539412, + "grad_norm": 15.125, + "learning_rate": 1.027701474524873e-06, + "loss": 0.8517, + "step": 13115 + }, + { + "epoch": 1.1176821474222411, + "grad_norm": 12.0, + "learning_rate": 1.0275361195323445e-06, + "loss": 0.3669, + "step": 13116 + }, + { + "epoch": 1.117767362590541, + "grad_norm": 14.125, + "learning_rate": 1.0273707685599616e-06, + "loss": 0.4826, + "step": 13117 + }, + { + "epoch": 1.117852577758841, + "grad_norm": 16.75, + "learning_rate": 1.027205421610713e-06, + "loss": 0.833, + "step": 13118 + }, + { + "epoch": 1.117937792927141, + "grad_norm": 12.5, + "learning_rate": 1.0270400786875864e-06, + "loss": 0.3049, + "step": 13119 + }, + { + "epoch": 1.118023008095441, + "grad_norm": 17.5, + "learning_rate": 1.02687473979357e-06, + "loss": 0.6963, + "step": 13120 + }, + { + "epoch": 1.118108223263741, + "grad_norm": 12.5, + "learning_rate": 1.0267094049316514e-06, + "loss": 0.5541, + "step": 13121 + }, + { + "epoch": 1.1181934384320409, + "grad_norm": 16.5, + "learning_rate": 1.0265440741048183e-06, + "loss": 0.4425, + "step": 13122 + }, + { + "epoch": 1.1182786536003408, + "grad_norm": 15.3125, + "learning_rate": 1.0263787473160584e-06, + "loss": 0.8104, + "step": 13123 + }, + { + "epoch": 1.1183638687686408, + "grad_norm": 13.875, + "learning_rate": 1.0262134245683593e-06, + "loss": 0.476, + "step": 13124 + }, + { + "epoch": 1.1184490839369408, + "grad_norm": 16.125, + "learning_rate": 1.026048105864709e-06, + "loss": 0.6981, + "step": 13125 + }, + { + "epoch": 1.1185342991052407, + "grad_norm": 13.5625, + "learning_rate": 1.0258827912080941e-06, + "loss": 0.385, + "step": 13126 + }, + { + "epoch": 1.1186195142735407, + "grad_norm": 11.6875, + "learning_rate": 1.025717480601503e-06, + "loss": 0.4115, + "step": 13127 + }, + { + "epoch": 1.1187047294418406, + "grad_norm": 12.5, + "learning_rate": 1.0255521740479225e-06, + "loss": 0.237, + "step": 13128 + }, + { + "epoch": 1.1187899446101406, + "grad_norm": 19.875, + "learning_rate": 1.02538687155034e-06, + "loss": 1.0516, + "step": 13129 + }, + { + "epoch": 1.1188751597784405, + "grad_norm": 12.5625, + "learning_rate": 1.0252215731117424e-06, + "loss": 0.3883, + "step": 13130 + }, + { + "epoch": 1.1189603749467405, + "grad_norm": 17.5, + "learning_rate": 1.025056278735117e-06, + "loss": 0.9291, + "step": 13131 + }, + { + "epoch": 1.1190455901150405, + "grad_norm": 14.875, + "learning_rate": 1.0248909884234509e-06, + "loss": 0.5778, + "step": 13132 + }, + { + "epoch": 1.1191308052833404, + "grad_norm": 15.6875, + "learning_rate": 1.0247257021797308e-06, + "loss": 0.7954, + "step": 13133 + }, + { + "epoch": 1.1192160204516404, + "grad_norm": 15.8125, + "learning_rate": 1.0245604200069439e-06, + "loss": 0.6463, + "step": 13134 + }, + { + "epoch": 1.1193012356199403, + "grad_norm": 14.9375, + "learning_rate": 1.024395141908077e-06, + "loss": 0.6824, + "step": 13135 + }, + { + "epoch": 1.1193864507882403, + "grad_norm": 12.4375, + "learning_rate": 1.0242298678861166e-06, + "loss": 0.4984, + "step": 13136 + }, + { + "epoch": 1.1194716659565402, + "grad_norm": 13.0625, + "learning_rate": 1.0240645979440496e-06, + "loss": 0.5623, + "step": 13137 + }, + { + "epoch": 1.1195568811248402, + "grad_norm": 16.875, + "learning_rate": 1.0238993320848631e-06, + "loss": 0.73, + "step": 13138 + }, + { + "epoch": 1.1196420962931402, + "grad_norm": 15.625, + "learning_rate": 1.0237340703115423e-06, + "loss": 0.8047, + "step": 13139 + }, + { + "epoch": 1.1197273114614401, + "grad_norm": 19.375, + "learning_rate": 1.0235688126270746e-06, + "loss": 1.034, + "step": 13140 + }, + { + "epoch": 1.11981252662974, + "grad_norm": 13.6875, + "learning_rate": 1.023403559034446e-06, + "loss": 0.3668, + "step": 13141 + }, + { + "epoch": 1.11989774179804, + "grad_norm": 19.125, + "learning_rate": 1.023238309536643e-06, + "loss": 0.5025, + "step": 13142 + }, + { + "epoch": 1.11998295696634, + "grad_norm": 17.75, + "learning_rate": 1.023073064136652e-06, + "loss": 0.8449, + "step": 13143 + }, + { + "epoch": 1.12006817213464, + "grad_norm": 13.8125, + "learning_rate": 1.0229078228374585e-06, + "loss": 0.4999, + "step": 13144 + }, + { + "epoch": 1.12015338730294, + "grad_norm": 12.0, + "learning_rate": 1.0227425856420492e-06, + "loss": 0.4149, + "step": 13145 + }, + { + "epoch": 1.1202386024712399, + "grad_norm": 15.6875, + "learning_rate": 1.0225773525534104e-06, + "loss": 0.5834, + "step": 13146 + }, + { + "epoch": 1.1203238176395398, + "grad_norm": 14.9375, + "learning_rate": 1.022412123574527e-06, + "loss": 0.857, + "step": 13147 + }, + { + "epoch": 1.1204090328078398, + "grad_norm": 15.1875, + "learning_rate": 1.0222468987083855e-06, + "loss": 0.6731, + "step": 13148 + }, + { + "epoch": 1.1204942479761397, + "grad_norm": 15.0, + "learning_rate": 1.0220816779579714e-06, + "loss": 0.4319, + "step": 13149 + }, + { + "epoch": 1.1205794631444397, + "grad_norm": 12.9375, + "learning_rate": 1.0219164613262707e-06, + "loss": 0.5113, + "step": 13150 + }, + { + "epoch": 1.1206646783127396, + "grad_norm": 17.75, + "learning_rate": 1.0217512488162688e-06, + "loss": 0.6501, + "step": 13151 + }, + { + "epoch": 1.1207498934810396, + "grad_norm": 12.75, + "learning_rate": 1.0215860404309516e-06, + "loss": 0.3189, + "step": 13152 + }, + { + "epoch": 1.1208351086493396, + "grad_norm": 17.875, + "learning_rate": 1.021420836173304e-06, + "loss": 0.8867, + "step": 13153 + }, + { + "epoch": 1.1209203238176395, + "grad_norm": 18.0, + "learning_rate": 1.0212556360463122e-06, + "loss": 0.7039, + "step": 13154 + }, + { + "epoch": 1.1210055389859395, + "grad_norm": 14.0625, + "learning_rate": 1.021090440052961e-06, + "loss": 0.7019, + "step": 13155 + }, + { + "epoch": 1.1210907541542394, + "grad_norm": 15.875, + "learning_rate": 1.0209252481962352e-06, + "loss": 0.7521, + "step": 13156 + }, + { + "epoch": 1.1211759693225394, + "grad_norm": 14.9375, + "learning_rate": 1.020760060479121e-06, + "loss": 0.6964, + "step": 13157 + }, + { + "epoch": 1.1212611844908393, + "grad_norm": 19.125, + "learning_rate": 1.020594876904603e-06, + "loss": 0.7076, + "step": 13158 + }, + { + "epoch": 1.1213463996591393, + "grad_norm": 15.5, + "learning_rate": 1.020429697475666e-06, + "loss": 0.4491, + "step": 13159 + }, + { + "epoch": 1.1214316148274392, + "grad_norm": 22.25, + "learning_rate": 1.0202645221952956e-06, + "loss": 0.8533, + "step": 13160 + }, + { + "epoch": 1.1215168299957392, + "grad_norm": 16.875, + "learning_rate": 1.0200993510664758e-06, + "loss": 0.3223, + "step": 13161 + }, + { + "epoch": 1.1216020451640392, + "grad_norm": 14.0, + "learning_rate": 1.0199341840921924e-06, + "loss": 0.5168, + "step": 13162 + }, + { + "epoch": 1.1216872603323391, + "grad_norm": 21.0, + "learning_rate": 1.0197690212754296e-06, + "loss": 0.7427, + "step": 13163 + }, + { + "epoch": 1.121772475500639, + "grad_norm": 24.375, + "learning_rate": 1.0196038626191717e-06, + "loss": 0.9318, + "step": 13164 + }, + { + "epoch": 1.121857690668939, + "grad_norm": 32.25, + "learning_rate": 1.0194387081264042e-06, + "loss": 1.1205, + "step": 13165 + }, + { + "epoch": 1.121942905837239, + "grad_norm": 12.375, + "learning_rate": 1.0192735578001105e-06, + "loss": 0.4297, + "step": 13166 + }, + { + "epoch": 1.122028121005539, + "grad_norm": 19.875, + "learning_rate": 1.019108411643276e-06, + "loss": 0.5472, + "step": 13167 + }, + { + "epoch": 1.122113336173839, + "grad_norm": 13.375, + "learning_rate": 1.0189432696588843e-06, + "loss": 0.483, + "step": 13168 + }, + { + "epoch": 1.1221985513421389, + "grad_norm": 15.6875, + "learning_rate": 1.0187781318499208e-06, + "loss": 0.6847, + "step": 13169 + }, + { + "epoch": 1.1222837665104388, + "grad_norm": 22.5, + "learning_rate": 1.0186129982193688e-06, + "loss": 0.9738, + "step": 13170 + }, + { + "epoch": 1.1223689816787388, + "grad_norm": 29.0, + "learning_rate": 1.018447868770212e-06, + "loss": 0.8715, + "step": 13171 + }, + { + "epoch": 1.1224541968470387, + "grad_norm": 19.625, + "learning_rate": 1.0182827435054356e-06, + "loss": 0.5456, + "step": 13172 + }, + { + "epoch": 1.1225394120153387, + "grad_norm": 15.3125, + "learning_rate": 1.018117622428023e-06, + "loss": 0.3682, + "step": 13173 + }, + { + "epoch": 1.1226246271836386, + "grad_norm": 15.0625, + "learning_rate": 1.0179525055409576e-06, + "loss": 1.0837, + "step": 13174 + }, + { + "epoch": 1.1227098423519386, + "grad_norm": 13.4375, + "learning_rate": 1.0177873928472242e-06, + "loss": 0.2664, + "step": 13175 + }, + { + "epoch": 1.1227950575202386, + "grad_norm": 11.4375, + "learning_rate": 1.017622284349806e-06, + "loss": 0.3701, + "step": 13176 + }, + { + "epoch": 1.1228802726885385, + "grad_norm": 19.5, + "learning_rate": 1.017457180051687e-06, + "loss": 0.859, + "step": 13177 + }, + { + "epoch": 1.1229654878568385, + "grad_norm": 17.875, + "learning_rate": 1.0172920799558508e-06, + "loss": 0.4577, + "step": 13178 + }, + { + "epoch": 1.1230507030251384, + "grad_norm": 14.1875, + "learning_rate": 1.01712698406528e-06, + "loss": 0.3891, + "step": 13179 + }, + { + "epoch": 1.1231359181934384, + "grad_norm": 17.0, + "learning_rate": 1.0169618923829593e-06, + "loss": 0.7796, + "step": 13180 + }, + { + "epoch": 1.1232211333617383, + "grad_norm": 13.4375, + "learning_rate": 1.0167968049118712e-06, + "loss": 0.3676, + "step": 13181 + }, + { + "epoch": 1.1233063485300383, + "grad_norm": 12.125, + "learning_rate": 1.0166317216549995e-06, + "loss": 0.2856, + "step": 13182 + }, + { + "epoch": 1.1233915636983383, + "grad_norm": 17.5, + "learning_rate": 1.0164666426153271e-06, + "loss": 0.8112, + "step": 13183 + }, + { + "epoch": 1.1234767788666382, + "grad_norm": 19.5, + "learning_rate": 1.0163015677958376e-06, + "loss": 0.716, + "step": 13184 + }, + { + "epoch": 1.1235619940349382, + "grad_norm": 19.25, + "learning_rate": 1.0161364971995136e-06, + "loss": 0.509, + "step": 13185 + }, + { + "epoch": 1.1236472092032381, + "grad_norm": 17.875, + "learning_rate": 1.0159714308293387e-06, + "loss": 0.756, + "step": 13186 + }, + { + "epoch": 1.123732424371538, + "grad_norm": 12.8125, + "learning_rate": 1.0158063686882948e-06, + "loss": 0.4773, + "step": 13187 + }, + { + "epoch": 1.123817639539838, + "grad_norm": 13.1875, + "learning_rate": 1.0156413107793658e-06, + "loss": 0.4807, + "step": 13188 + }, + { + "epoch": 1.123902854708138, + "grad_norm": 13.0, + "learning_rate": 1.0154762571055335e-06, + "loss": 0.5748, + "step": 13189 + }, + { + "epoch": 1.123988069876438, + "grad_norm": 14.625, + "learning_rate": 1.0153112076697813e-06, + "loss": 0.399, + "step": 13190 + }, + { + "epoch": 1.124073285044738, + "grad_norm": 12.6875, + "learning_rate": 1.0151461624750914e-06, + "loss": 0.4311, + "step": 13191 + }, + { + "epoch": 1.1241585002130379, + "grad_norm": 12.625, + "learning_rate": 1.0149811215244468e-06, + "loss": 0.8434, + "step": 13192 + }, + { + "epoch": 1.1242437153813378, + "grad_norm": 17.625, + "learning_rate": 1.0148160848208293e-06, + "loss": 0.6584, + "step": 13193 + }, + { + "epoch": 1.1243289305496378, + "grad_norm": 20.75, + "learning_rate": 1.0146510523672221e-06, + "loss": 0.7541, + "step": 13194 + }, + { + "epoch": 1.1244141457179377, + "grad_norm": 13.5625, + "learning_rate": 1.0144860241666068e-06, + "loss": 0.2725, + "step": 13195 + }, + { + "epoch": 1.1244993608862377, + "grad_norm": 23.0, + "learning_rate": 1.0143210002219657e-06, + "loss": 0.8792, + "step": 13196 + }, + { + "epoch": 1.1245845760545377, + "grad_norm": 15.625, + "learning_rate": 1.0141559805362815e-06, + "loss": 0.6599, + "step": 13197 + }, + { + "epoch": 1.1246697912228376, + "grad_norm": 12.5625, + "learning_rate": 1.0139909651125354e-06, + "loss": 0.4025, + "step": 13198 + }, + { + "epoch": 1.1247550063911376, + "grad_norm": 12.8125, + "learning_rate": 1.0138259539537103e-06, + "loss": 0.4674, + "step": 13199 + }, + { + "epoch": 1.1248402215594375, + "grad_norm": 12.625, + "learning_rate": 1.0136609470627874e-06, + "loss": 0.3598, + "step": 13200 + }, + { + "epoch": 1.1249254367277375, + "grad_norm": 18.0, + "learning_rate": 1.0134959444427492e-06, + "loss": 0.6847, + "step": 13201 + }, + { + "epoch": 1.1250106518960374, + "grad_norm": 11.75, + "learning_rate": 1.0133309460965771e-06, + "loss": 0.3975, + "step": 13202 + }, + { + "epoch": 1.1250958670643374, + "grad_norm": 15.5, + "learning_rate": 1.0131659520272527e-06, + "loss": 0.3462, + "step": 13203 + }, + { + "epoch": 1.1251810822326374, + "grad_norm": 15.1875, + "learning_rate": 1.0130009622377575e-06, + "loss": 0.4185, + "step": 13204 + }, + { + "epoch": 1.1252662974009373, + "grad_norm": 16.75, + "learning_rate": 1.0128359767310734e-06, + "loss": 0.8716, + "step": 13205 + }, + { + "epoch": 1.1253515125692373, + "grad_norm": 15.25, + "learning_rate": 1.0126709955101815e-06, + "loss": 0.7207, + "step": 13206 + }, + { + "epoch": 1.1254367277375372, + "grad_norm": 16.625, + "learning_rate": 1.0125060185780636e-06, + "loss": 0.6605, + "step": 13207 + }, + { + "epoch": 1.1255219429058372, + "grad_norm": 16.5, + "learning_rate": 1.0123410459377004e-06, + "loss": 0.7494, + "step": 13208 + }, + { + "epoch": 1.1256071580741371, + "grad_norm": 18.125, + "learning_rate": 1.0121760775920739e-06, + "loss": 0.6069, + "step": 13209 + }, + { + "epoch": 1.125692373242437, + "grad_norm": 14.625, + "learning_rate": 1.0120111135441644e-06, + "loss": 0.5261, + "step": 13210 + }, + { + "epoch": 1.125777588410737, + "grad_norm": 9.625, + "learning_rate": 1.011846153796954e-06, + "loss": 0.2247, + "step": 13211 + }, + { + "epoch": 1.125862803579037, + "grad_norm": 18.0, + "learning_rate": 1.011681198353423e-06, + "loss": 0.5614, + "step": 13212 + }, + { + "epoch": 1.125948018747337, + "grad_norm": 15.4375, + "learning_rate": 1.0115162472165517e-06, + "loss": 0.6167, + "step": 13213 + }, + { + "epoch": 1.126033233915637, + "grad_norm": 16.0, + "learning_rate": 1.0113513003893223e-06, + "loss": 0.4604, + "step": 13214 + }, + { + "epoch": 1.1261184490839369, + "grad_norm": 18.0, + "learning_rate": 1.0111863578747145e-06, + "loss": 0.5409, + "step": 13215 + }, + { + "epoch": 1.1262036642522368, + "grad_norm": 15.6875, + "learning_rate": 1.0110214196757095e-06, + "loss": 0.4521, + "step": 13216 + }, + { + "epoch": 1.1262888794205368, + "grad_norm": 17.25, + "learning_rate": 1.0108564857952876e-06, + "loss": 0.4512, + "step": 13217 + }, + { + "epoch": 1.1263740945888367, + "grad_norm": 17.75, + "learning_rate": 1.0106915562364298e-06, + "loss": 0.7895, + "step": 13218 + }, + { + "epoch": 1.1264593097571367, + "grad_norm": 16.5, + "learning_rate": 1.0105266310021164e-06, + "loss": 0.6925, + "step": 13219 + }, + { + "epoch": 1.1265445249254367, + "grad_norm": 17.25, + "learning_rate": 1.0103617100953273e-06, + "loss": 0.5922, + "step": 13220 + }, + { + "epoch": 1.1266297400937366, + "grad_norm": 16.875, + "learning_rate": 1.010196793519043e-06, + "loss": 0.7036, + "step": 13221 + }, + { + "epoch": 1.1267149552620366, + "grad_norm": 24.625, + "learning_rate": 1.010031881276244e-06, + "loss": 0.8022, + "step": 13222 + }, + { + "epoch": 1.1268001704303365, + "grad_norm": 14.125, + "learning_rate": 1.0098669733699097e-06, + "loss": 0.4868, + "step": 13223 + }, + { + "epoch": 1.1268853855986365, + "grad_norm": 12.875, + "learning_rate": 1.009702069803021e-06, + "loss": 0.4106, + "step": 13224 + }, + { + "epoch": 1.1269706007669364, + "grad_norm": 18.875, + "learning_rate": 1.0095371705785574e-06, + "loss": 0.5967, + "step": 13225 + }, + { + "epoch": 1.1270558159352364, + "grad_norm": 17.75, + "learning_rate": 1.0093722756994992e-06, + "loss": 0.8156, + "step": 13226 + }, + { + "epoch": 1.1271410311035364, + "grad_norm": 13.25, + "learning_rate": 1.0092073851688262e-06, + "loss": 0.2918, + "step": 13227 + }, + { + "epoch": 1.1272262462718363, + "grad_norm": 21.25, + "learning_rate": 1.0090424989895172e-06, + "loss": 0.7232, + "step": 13228 + }, + { + "epoch": 1.1273114614401363, + "grad_norm": 17.5, + "learning_rate": 1.0088776171645529e-06, + "loss": 0.828, + "step": 13229 + }, + { + "epoch": 1.1273966766084362, + "grad_norm": 12.9375, + "learning_rate": 1.0087127396969124e-06, + "loss": 0.5333, + "step": 13230 + }, + { + "epoch": 1.1274818917767362, + "grad_norm": 20.25, + "learning_rate": 1.0085478665895753e-06, + "loss": 0.8569, + "step": 13231 + }, + { + "epoch": 1.1275671069450361, + "grad_norm": 16.75, + "learning_rate": 1.008382997845521e-06, + "loss": 0.6145, + "step": 13232 + }, + { + "epoch": 1.127652322113336, + "grad_norm": 11.0625, + "learning_rate": 1.008218133467729e-06, + "loss": 0.2534, + "step": 13233 + }, + { + "epoch": 1.127737537281636, + "grad_norm": 13.9375, + "learning_rate": 1.0080532734591784e-06, + "loss": 0.616, + "step": 13234 + }, + { + "epoch": 1.127822752449936, + "grad_norm": 20.125, + "learning_rate": 1.0078884178228492e-06, + "loss": 0.481, + "step": 13235 + }, + { + "epoch": 1.127907967618236, + "grad_norm": 14.3125, + "learning_rate": 1.0077235665617188e-06, + "loss": 0.5209, + "step": 13236 + }, + { + "epoch": 1.127993182786536, + "grad_norm": 10.375, + "learning_rate": 1.0075587196787677e-06, + "loss": 0.2546, + "step": 13237 + }, + { + "epoch": 1.1280783979548359, + "grad_norm": 16.875, + "learning_rate": 1.0073938771769738e-06, + "loss": 0.7662, + "step": 13238 + }, + { + "epoch": 1.1281636131231358, + "grad_norm": 18.5, + "learning_rate": 1.007229039059317e-06, + "loss": 0.7725, + "step": 13239 + }, + { + "epoch": 1.1282488282914358, + "grad_norm": 16.625, + "learning_rate": 1.0070642053287751e-06, + "loss": 0.7769, + "step": 13240 + }, + { + "epoch": 1.1283340434597358, + "grad_norm": 15.0625, + "learning_rate": 1.0068993759883278e-06, + "loss": 0.4144, + "step": 13241 + }, + { + "epoch": 1.1284192586280357, + "grad_norm": 9.75, + "learning_rate": 1.0067345510409527e-06, + "loss": 0.2127, + "step": 13242 + }, + { + "epoch": 1.1285044737963357, + "grad_norm": 16.625, + "learning_rate": 1.0065697304896296e-06, + "loss": 0.9401, + "step": 13243 + }, + { + "epoch": 1.1285896889646356, + "grad_norm": 13.625, + "learning_rate": 1.006404914337336e-06, + "loss": 0.4444, + "step": 13244 + }, + { + "epoch": 1.1286749041329356, + "grad_norm": 16.25, + "learning_rate": 1.0062401025870501e-06, + "loss": 0.7788, + "step": 13245 + }, + { + "epoch": 1.1287601193012355, + "grad_norm": 18.25, + "learning_rate": 1.006075295241751e-06, + "loss": 0.8345, + "step": 13246 + }, + { + "epoch": 1.1288453344695355, + "grad_norm": 29.25, + "learning_rate": 1.0059104923044166e-06, + "loss": 0.7864, + "step": 13247 + }, + { + "epoch": 1.1289305496378355, + "grad_norm": 10.625, + "learning_rate": 1.0057456937780252e-06, + "loss": 0.2497, + "step": 13248 + }, + { + "epoch": 1.1290157648061354, + "grad_norm": 13.1875, + "learning_rate": 1.0055808996655542e-06, + "loss": 0.5187, + "step": 13249 + }, + { + "epoch": 1.1291009799744354, + "grad_norm": 16.125, + "learning_rate": 1.0054161099699826e-06, + "loss": 0.745, + "step": 13250 + }, + { + "epoch": 1.1291861951427353, + "grad_norm": 15.6875, + "learning_rate": 1.0052513246942879e-06, + "loss": 0.6369, + "step": 13251 + }, + { + "epoch": 1.1292714103110353, + "grad_norm": 25.75, + "learning_rate": 1.005086543841448e-06, + "loss": 0.4194, + "step": 13252 + }, + { + "epoch": 1.1293566254793352, + "grad_norm": 10.375, + "learning_rate": 1.00492176741444e-06, + "loss": 0.37, + "step": 13253 + }, + { + "epoch": 1.1294418406476352, + "grad_norm": 17.75, + "learning_rate": 1.0047569954162424e-06, + "loss": 0.7184, + "step": 13254 + }, + { + "epoch": 1.1295270558159352, + "grad_norm": 16.875, + "learning_rate": 1.0045922278498323e-06, + "loss": 0.5522, + "step": 13255 + }, + { + "epoch": 1.1296122709842351, + "grad_norm": 10.1875, + "learning_rate": 1.0044274647181876e-06, + "loss": 0.24, + "step": 13256 + }, + { + "epoch": 1.129697486152535, + "grad_norm": 22.875, + "learning_rate": 1.0042627060242854e-06, + "loss": 0.8819, + "step": 13257 + }, + { + "epoch": 1.129782701320835, + "grad_norm": 15.875, + "learning_rate": 1.0040979517711034e-06, + "loss": 0.6125, + "step": 13258 + }, + { + "epoch": 1.129867916489135, + "grad_norm": 19.0, + "learning_rate": 1.003933201961619e-06, + "loss": 0.5613, + "step": 13259 + }, + { + "epoch": 1.129953131657435, + "grad_norm": 27.5, + "learning_rate": 1.0037684565988084e-06, + "loss": 0.707, + "step": 13260 + }, + { + "epoch": 1.130038346825735, + "grad_norm": 16.875, + "learning_rate": 1.00360371568565e-06, + "loss": 0.7815, + "step": 13261 + }, + { + "epoch": 1.1301235619940349, + "grad_norm": 18.375, + "learning_rate": 1.0034389792251196e-06, + "loss": 0.7995, + "step": 13262 + }, + { + "epoch": 1.1302087771623348, + "grad_norm": 20.625, + "learning_rate": 1.0032742472201952e-06, + "loss": 0.5258, + "step": 13263 + }, + { + "epoch": 1.1302939923306348, + "grad_norm": 16.0, + "learning_rate": 1.003109519673853e-06, + "loss": 0.5172, + "step": 13264 + }, + { + "epoch": 1.1303792074989347, + "grad_norm": 31.0, + "learning_rate": 1.0029447965890702e-06, + "loss": 0.5182, + "step": 13265 + }, + { + "epoch": 1.1304644226672347, + "grad_norm": 16.25, + "learning_rate": 1.0027800779688236e-06, + "loss": 0.498, + "step": 13266 + }, + { + "epoch": 1.1305496378355346, + "grad_norm": 13.9375, + "learning_rate": 1.0026153638160896e-06, + "loss": 0.513, + "step": 13267 + }, + { + "epoch": 1.1306348530038346, + "grad_norm": 20.625, + "learning_rate": 1.0024506541338444e-06, + "loss": 0.8987, + "step": 13268 + }, + { + "epoch": 1.1307200681721346, + "grad_norm": 14.0625, + "learning_rate": 1.002285948925065e-06, + "loss": 0.5029, + "step": 13269 + }, + { + "epoch": 1.1308052833404345, + "grad_norm": 16.25, + "learning_rate": 1.0021212481927272e-06, + "loss": 0.7434, + "step": 13270 + }, + { + "epoch": 1.1308904985087345, + "grad_norm": 12.5, + "learning_rate": 1.001956551939808e-06, + "loss": 0.5046, + "step": 13271 + }, + { + "epoch": 1.1309757136770344, + "grad_norm": 13.5, + "learning_rate": 1.0017918601692833e-06, + "loss": 0.5205, + "step": 13272 + }, + { + "epoch": 1.1310609288453344, + "grad_norm": 14.0, + "learning_rate": 1.0016271728841294e-06, + "loss": 0.6635, + "step": 13273 + }, + { + "epoch": 1.1311461440136343, + "grad_norm": 22.0, + "learning_rate": 1.001462490087322e-06, + "loss": 0.3674, + "step": 13274 + }, + { + "epoch": 1.1312313591819343, + "grad_norm": 11.875, + "learning_rate": 1.0012978117818375e-06, + "loss": 0.4243, + "step": 13275 + }, + { + "epoch": 1.1313165743502342, + "grad_norm": 11.5625, + "learning_rate": 1.0011331379706522e-06, + "loss": 0.6768, + "step": 13276 + }, + { + "epoch": 1.1314017895185342, + "grad_norm": 15.625, + "learning_rate": 1.0009684686567404e-06, + "loss": 0.9479, + "step": 13277 + }, + { + "epoch": 1.1314870046868342, + "grad_norm": 29.125, + "learning_rate": 1.0008038038430791e-06, + "loss": 0.3934, + "step": 13278 + }, + { + "epoch": 1.1315722198551341, + "grad_norm": 15.6875, + "learning_rate": 1.0006391435326438e-06, + "loss": 0.7388, + "step": 13279 + }, + { + "epoch": 1.131657435023434, + "grad_norm": 20.875, + "learning_rate": 1.0004744877284099e-06, + "loss": 1.0583, + "step": 13280 + }, + { + "epoch": 1.131742650191734, + "grad_norm": 17.25, + "learning_rate": 1.000309836433353e-06, + "loss": 0.8705, + "step": 13281 + }, + { + "epoch": 1.131827865360034, + "grad_norm": 11.8125, + "learning_rate": 1.000145189650448e-06, + "loss": 0.3678, + "step": 13282 + }, + { + "epoch": 1.131913080528334, + "grad_norm": 15.875, + "learning_rate": 9.99980547382671e-07, + "loss": 0.6978, + "step": 13283 + }, + { + "epoch": 1.131998295696634, + "grad_norm": 14.3125, + "learning_rate": 9.998159096329974e-07, + "loss": 0.6857, + "step": 13284 + }, + { + "epoch": 1.1320835108649339, + "grad_norm": 10.6875, + "learning_rate": 9.99651276404401e-07, + "loss": 0.2527, + "step": 13285 + }, + { + "epoch": 1.1321687260332338, + "grad_norm": 13.8125, + "learning_rate": 9.994866476998583e-07, + "loss": 0.5626, + "step": 13286 + }, + { + "epoch": 1.132253941201534, + "grad_norm": 15.3125, + "learning_rate": 9.993220235223437e-07, + "loss": 0.6575, + "step": 13287 + }, + { + "epoch": 1.132339156369834, + "grad_norm": 13.75, + "learning_rate": 9.991574038748322e-07, + "loss": 0.6853, + "step": 13288 + }, + { + "epoch": 1.132424371538134, + "grad_norm": 13.75, + "learning_rate": 9.989927887602984e-07, + "loss": 0.3997, + "step": 13289 + }, + { + "epoch": 1.1325095867064339, + "grad_norm": 15.625, + "learning_rate": 9.988281781817178e-07, + "loss": 0.7059, + "step": 13290 + }, + { + "epoch": 1.1325948018747338, + "grad_norm": 13.3125, + "learning_rate": 9.98663572142064e-07, + "loss": 0.4542, + "step": 13291 + }, + { + "epoch": 1.1326800170430338, + "grad_norm": 36.0, + "learning_rate": 9.984989706443129e-07, + "loss": 0.7358, + "step": 13292 + }, + { + "epoch": 1.1327652322113337, + "grad_norm": 57.5, + "learning_rate": 9.98334373691438e-07, + "loss": 0.7715, + "step": 13293 + }, + { + "epoch": 1.1328504473796337, + "grad_norm": 21.75, + "learning_rate": 9.981697812864138e-07, + "loss": 0.3737, + "step": 13294 + }, + { + "epoch": 1.1329356625479337, + "grad_norm": 22.5, + "learning_rate": 9.980051934322152e-07, + "loss": 0.5305, + "step": 13295 + }, + { + "epoch": 1.1330208777162336, + "grad_norm": 16.75, + "learning_rate": 9.97840610131816e-07, + "loss": 0.7419, + "step": 13296 + }, + { + "epoch": 1.1331060928845336, + "grad_norm": 19.25, + "learning_rate": 9.976760313881903e-07, + "loss": 0.5595, + "step": 13297 + }, + { + "epoch": 1.1331913080528335, + "grad_norm": 17.25, + "learning_rate": 9.97511457204313e-07, + "loss": 0.7737, + "step": 13298 + }, + { + "epoch": 1.1332765232211335, + "grad_norm": 19.625, + "learning_rate": 9.973468875831569e-07, + "loss": 0.7084, + "step": 13299 + }, + { + "epoch": 1.1333617383894334, + "grad_norm": 12.1875, + "learning_rate": 9.971823225276973e-07, + "loss": 0.4992, + "step": 13300 + }, + { + "epoch": 1.1334469535577334, + "grad_norm": 18.875, + "learning_rate": 9.970177620409072e-07, + "loss": 1.0615, + "step": 13301 + }, + { + "epoch": 1.1335321687260334, + "grad_norm": 14.0625, + "learning_rate": 9.968532061257603e-07, + "loss": 0.7005, + "step": 13302 + }, + { + "epoch": 1.1336173838943333, + "grad_norm": 21.875, + "learning_rate": 9.966886547852305e-07, + "loss": 0.8372, + "step": 13303 + }, + { + "epoch": 1.1337025990626333, + "grad_norm": 18.75, + "learning_rate": 9.965241080222913e-07, + "loss": 1.094, + "step": 13304 + }, + { + "epoch": 1.1337878142309332, + "grad_norm": 22.125, + "learning_rate": 9.963595658399167e-07, + "loss": 0.8261, + "step": 13305 + }, + { + "epoch": 1.1338730293992332, + "grad_norm": 18.375, + "learning_rate": 9.961950282410794e-07, + "loss": 0.4777, + "step": 13306 + }, + { + "epoch": 1.1339582445675331, + "grad_norm": 13.4375, + "learning_rate": 9.960304952287536e-07, + "loss": 0.3664, + "step": 13307 + }, + { + "epoch": 1.134043459735833, + "grad_norm": 14.75, + "learning_rate": 9.958659668059123e-07, + "loss": 0.5838, + "step": 13308 + }, + { + "epoch": 1.134128674904133, + "grad_norm": 17.875, + "learning_rate": 9.95701442975528e-07, + "loss": 0.7497, + "step": 13309 + }, + { + "epoch": 1.134213890072433, + "grad_norm": 11.0, + "learning_rate": 9.955369237405745e-07, + "loss": 0.2642, + "step": 13310 + }, + { + "epoch": 1.134299105240733, + "grad_norm": 17.125, + "learning_rate": 9.95372409104025e-07, + "loss": 0.5166, + "step": 13311 + }, + { + "epoch": 1.134384320409033, + "grad_norm": 16.5, + "learning_rate": 9.952078990688515e-07, + "loss": 0.7807, + "step": 13312 + }, + { + "epoch": 1.1344695355773329, + "grad_norm": 13.375, + "learning_rate": 9.95043393638028e-07, + "loss": 0.6093, + "step": 13313 + }, + { + "epoch": 1.1345547507456328, + "grad_norm": 18.375, + "learning_rate": 9.948788928145262e-07, + "loss": 1.1307, + "step": 13314 + }, + { + "epoch": 1.1346399659139328, + "grad_norm": 14.125, + "learning_rate": 9.9471439660132e-07, + "loss": 0.5128, + "step": 13315 + }, + { + "epoch": 1.1347251810822327, + "grad_norm": 14.6875, + "learning_rate": 9.945499050013813e-07, + "loss": 0.6648, + "step": 13316 + }, + { + "epoch": 1.1348103962505327, + "grad_norm": 15.0, + "learning_rate": 9.943854180176824e-07, + "loss": 0.5517, + "step": 13317 + }, + { + "epoch": 1.1348956114188327, + "grad_norm": 18.0, + "learning_rate": 9.942209356531962e-07, + "loss": 0.904, + "step": 13318 + }, + { + "epoch": 1.1349808265871326, + "grad_norm": 22.25, + "learning_rate": 9.940564579108946e-07, + "loss": 0.7452, + "step": 13319 + }, + { + "epoch": 1.1350660417554326, + "grad_norm": 17.5, + "learning_rate": 9.938919847937504e-07, + "loss": 0.6283, + "step": 13320 + }, + { + "epoch": 1.1351512569237325, + "grad_norm": 17.0, + "learning_rate": 9.937275163047353e-07, + "loss": 0.589, + "step": 13321 + }, + { + "epoch": 1.1352364720920325, + "grad_norm": 14.3125, + "learning_rate": 9.93563052446822e-07, + "loss": 0.5597, + "step": 13322 + }, + { + "epoch": 1.1353216872603324, + "grad_norm": 15.75, + "learning_rate": 9.93398593222982e-07, + "loss": 0.6667, + "step": 13323 + }, + { + "epoch": 1.1354069024286324, + "grad_norm": 18.75, + "learning_rate": 9.932341386361878e-07, + "loss": 0.7381, + "step": 13324 + }, + { + "epoch": 1.1354921175969324, + "grad_norm": 18.375, + "learning_rate": 9.93069688689411e-07, + "loss": 0.7696, + "step": 13325 + }, + { + "epoch": 1.1355773327652323, + "grad_norm": 15.8125, + "learning_rate": 9.929052433856227e-07, + "loss": 0.6384, + "step": 13326 + }, + { + "epoch": 1.1356625479335323, + "grad_norm": 15.125, + "learning_rate": 9.927408027277957e-07, + "loss": 0.8175, + "step": 13327 + }, + { + "epoch": 1.1357477631018322, + "grad_norm": 15.625, + "learning_rate": 9.925763667189009e-07, + "loss": 0.6364, + "step": 13328 + }, + { + "epoch": 1.1358329782701322, + "grad_norm": 24.0, + "learning_rate": 9.924119353619099e-07, + "loss": 0.9324, + "step": 13329 + }, + { + "epoch": 1.1359181934384321, + "grad_norm": 14.0625, + "learning_rate": 9.922475086597944e-07, + "loss": 0.6118, + "step": 13330 + }, + { + "epoch": 1.136003408606732, + "grad_norm": 14.875, + "learning_rate": 9.920830866155253e-07, + "loss": 0.5916, + "step": 13331 + }, + { + "epoch": 1.136088623775032, + "grad_norm": 18.875, + "learning_rate": 9.919186692320748e-07, + "loss": 0.4644, + "step": 13332 + }, + { + "epoch": 1.136173838943332, + "grad_norm": 20.75, + "learning_rate": 9.91754256512413e-07, + "loss": 0.5872, + "step": 13333 + }, + { + "epoch": 1.136259054111632, + "grad_norm": 16.875, + "learning_rate": 9.915898484595115e-07, + "loss": 1.0306, + "step": 13334 + }, + { + "epoch": 1.136344269279932, + "grad_norm": 19.375, + "learning_rate": 9.914254450763414e-07, + "loss": 0.7943, + "step": 13335 + }, + { + "epoch": 1.1364294844482319, + "grad_norm": 41.0, + "learning_rate": 9.912610463658734e-07, + "loss": 1.1428, + "step": 13336 + }, + { + "epoch": 1.1365146996165318, + "grad_norm": 11.0, + "learning_rate": 9.910966523310786e-07, + "loss": 0.3458, + "step": 13337 + }, + { + "epoch": 1.1365999147848318, + "grad_norm": 24.875, + "learning_rate": 9.909322629749273e-07, + "loss": 0.7317, + "step": 13338 + }, + { + "epoch": 1.1366851299531318, + "grad_norm": 16.375, + "learning_rate": 9.90767878300391e-07, + "loss": 0.5539, + "step": 13339 + }, + { + "epoch": 1.1367703451214317, + "grad_norm": 13.25, + "learning_rate": 9.906034983104393e-07, + "loss": 0.5132, + "step": 13340 + }, + { + "epoch": 1.1368555602897317, + "grad_norm": 17.5, + "learning_rate": 9.90439123008044e-07, + "loss": 0.7343, + "step": 13341 + }, + { + "epoch": 1.1369407754580316, + "grad_norm": 19.125, + "learning_rate": 9.902747523961744e-07, + "loss": 0.5739, + "step": 13342 + }, + { + "epoch": 1.1370259906263316, + "grad_norm": 14.5, + "learning_rate": 9.901103864778012e-07, + "loss": 0.2709, + "step": 13343 + }, + { + "epoch": 1.1371112057946315, + "grad_norm": 14.375, + "learning_rate": 9.899460252558942e-07, + "loss": 0.2251, + "step": 13344 + }, + { + "epoch": 1.1371964209629315, + "grad_norm": 18.75, + "learning_rate": 9.897816687334247e-07, + "loss": 0.932, + "step": 13345 + }, + { + "epoch": 1.1372816361312315, + "grad_norm": 14.6875, + "learning_rate": 9.896173169133615e-07, + "loss": 0.591, + "step": 13346 + }, + { + "epoch": 1.1373668512995314, + "grad_norm": 16.375, + "learning_rate": 9.894529697986758e-07, + "loss": 0.755, + "step": 13347 + }, + { + "epoch": 1.1374520664678314, + "grad_norm": 15.875, + "learning_rate": 9.892886273923363e-07, + "loss": 0.8592, + "step": 13348 + }, + { + "epoch": 1.1375372816361313, + "grad_norm": 21.75, + "learning_rate": 9.891242896973143e-07, + "loss": 0.5365, + "step": 13349 + }, + { + "epoch": 1.1376224968044313, + "grad_norm": 15.625, + "learning_rate": 9.889599567165782e-07, + "loss": 0.7757, + "step": 13350 + }, + { + "epoch": 1.1377077119727312, + "grad_norm": 18.375, + "learning_rate": 9.88795628453098e-07, + "loss": 0.7902, + "step": 13351 + }, + { + "epoch": 1.1377929271410312, + "grad_norm": 13.5, + "learning_rate": 9.88631304909844e-07, + "loss": 0.6232, + "step": 13352 + }, + { + "epoch": 1.1378781423093312, + "grad_norm": 19.625, + "learning_rate": 9.884669860897845e-07, + "loss": 0.5116, + "step": 13353 + }, + { + "epoch": 1.137963357477631, + "grad_norm": 22.25, + "learning_rate": 9.883026719958899e-07, + "loss": 1.0277, + "step": 13354 + }, + { + "epoch": 1.138048572645931, + "grad_norm": 21.75, + "learning_rate": 9.88138362631129e-07, + "loss": 0.7742, + "step": 13355 + }, + { + "epoch": 1.138133787814231, + "grad_norm": 17.125, + "learning_rate": 9.879740579984716e-07, + "loss": 0.7077, + "step": 13356 + }, + { + "epoch": 1.138219002982531, + "grad_norm": 13.6875, + "learning_rate": 9.878097581008864e-07, + "loss": 0.4034, + "step": 13357 + }, + { + "epoch": 1.138304218150831, + "grad_norm": 12.3125, + "learning_rate": 9.876454629413425e-07, + "loss": 0.3284, + "step": 13358 + }, + { + "epoch": 1.138389433319131, + "grad_norm": 16.875, + "learning_rate": 9.874811725228086e-07, + "loss": 0.7156, + "step": 13359 + }, + { + "epoch": 1.1384746484874309, + "grad_norm": 16.0, + "learning_rate": 9.87316886848254e-07, + "loss": 0.7129, + "step": 13360 + }, + { + "epoch": 1.1385598636557308, + "grad_norm": 19.75, + "learning_rate": 9.871526059206473e-07, + "loss": 0.6555, + "step": 13361 + }, + { + "epoch": 1.1386450788240308, + "grad_norm": 11.75, + "learning_rate": 9.869883297429575e-07, + "loss": 0.249, + "step": 13362 + }, + { + "epoch": 1.1387302939923307, + "grad_norm": 12.25, + "learning_rate": 9.868240583181528e-07, + "loss": 0.447, + "step": 13363 + }, + { + "epoch": 1.1388155091606307, + "grad_norm": 13.125, + "learning_rate": 9.866597916492024e-07, + "loss": 0.4346, + "step": 13364 + }, + { + "epoch": 1.1389007243289306, + "grad_norm": 15.4375, + "learning_rate": 9.864955297390746e-07, + "loss": 0.8351, + "step": 13365 + }, + { + "epoch": 1.1389859394972306, + "grad_norm": 17.5, + "learning_rate": 9.86331272590737e-07, + "loss": 0.7371, + "step": 13366 + }, + { + "epoch": 1.1390711546655305, + "grad_norm": 18.875, + "learning_rate": 9.861670202071586e-07, + "loss": 1.1333, + "step": 13367 + }, + { + "epoch": 1.1391563698338305, + "grad_norm": 18.875, + "learning_rate": 9.860027725913074e-07, + "loss": 0.8659, + "step": 13368 + }, + { + "epoch": 1.1392415850021305, + "grad_norm": 15.375, + "learning_rate": 9.858385297461518e-07, + "loss": 0.6953, + "step": 13369 + }, + { + "epoch": 1.1393268001704304, + "grad_norm": 13.9375, + "learning_rate": 9.856742916746592e-07, + "loss": 0.3334, + "step": 13370 + }, + { + "epoch": 1.1394120153387304, + "grad_norm": 23.75, + "learning_rate": 9.855100583797983e-07, + "loss": 1.0496, + "step": 13371 + }, + { + "epoch": 1.1394972305070303, + "grad_norm": 16.875, + "learning_rate": 9.853458298645363e-07, + "loss": 0.6059, + "step": 13372 + }, + { + "epoch": 1.1395824456753303, + "grad_norm": 18.375, + "learning_rate": 9.85181606131842e-07, + "loss": 0.4935, + "step": 13373 + }, + { + "epoch": 1.1396676608436302, + "grad_norm": 13.3125, + "learning_rate": 9.850173871846817e-07, + "loss": 0.5685, + "step": 13374 + }, + { + "epoch": 1.1397528760119302, + "grad_norm": 14.375, + "learning_rate": 9.848531730260238e-07, + "loss": 0.5107, + "step": 13375 + }, + { + "epoch": 1.1398380911802302, + "grad_norm": 13.3125, + "learning_rate": 9.846889636588357e-07, + "loss": 0.4051, + "step": 13376 + }, + { + "epoch": 1.1399233063485301, + "grad_norm": 12.375, + "learning_rate": 9.845247590860848e-07, + "loss": 0.4834, + "step": 13377 + }, + { + "epoch": 1.14000852151683, + "grad_norm": 13.5, + "learning_rate": 9.843605593107383e-07, + "loss": 0.5974, + "step": 13378 + }, + { + "epoch": 1.14009373668513, + "grad_norm": 18.75, + "learning_rate": 9.84196364335764e-07, + "loss": 0.7099, + "step": 13379 + }, + { + "epoch": 1.14017895185343, + "grad_norm": 15.8125, + "learning_rate": 9.840321741641281e-07, + "loss": 0.6355, + "step": 13380 + }, + { + "epoch": 1.14026416702173, + "grad_norm": 16.375, + "learning_rate": 9.83867988798799e-07, + "loss": 0.9901, + "step": 13381 + }, + { + "epoch": 1.14034938219003, + "grad_norm": 15.1875, + "learning_rate": 9.83703808242743e-07, + "loss": 0.354, + "step": 13382 + }, + { + "epoch": 1.1404345973583299, + "grad_norm": 13.9375, + "learning_rate": 9.835396324989263e-07, + "loss": 0.4641, + "step": 13383 + }, + { + "epoch": 1.1405198125266298, + "grad_norm": 12.5, + "learning_rate": 9.833754615703166e-07, + "loss": 0.3003, + "step": 13384 + }, + { + "epoch": 1.1406050276949298, + "grad_norm": 12.9375, + "learning_rate": 9.832112954598804e-07, + "loss": 0.5548, + "step": 13385 + }, + { + "epoch": 1.1406902428632297, + "grad_norm": 9.75, + "learning_rate": 9.830471341705846e-07, + "loss": 0.2483, + "step": 13386 + }, + { + "epoch": 1.1407754580315297, + "grad_norm": 14.9375, + "learning_rate": 9.828829777053952e-07, + "loss": 0.5521, + "step": 13387 + }, + { + "epoch": 1.1408606731998296, + "grad_norm": 19.75, + "learning_rate": 9.827188260672792e-07, + "loss": 0.6579, + "step": 13388 + }, + { + "epoch": 1.1409458883681296, + "grad_norm": 18.625, + "learning_rate": 9.82554679259203e-07, + "loss": 0.7812, + "step": 13389 + }, + { + "epoch": 1.1410311035364296, + "grad_norm": 13.6875, + "learning_rate": 9.823905372841325e-07, + "loss": 0.3598, + "step": 13390 + }, + { + "epoch": 1.1411163187047295, + "grad_norm": 22.5, + "learning_rate": 9.82226400145034e-07, + "loss": 0.5544, + "step": 13391 + }, + { + "epoch": 1.1412015338730295, + "grad_norm": 21.25, + "learning_rate": 9.820622678448738e-07, + "loss": 0.7871, + "step": 13392 + }, + { + "epoch": 1.1412867490413294, + "grad_norm": 17.125, + "learning_rate": 9.818981403866177e-07, + "loss": 0.6372, + "step": 13393 + }, + { + "epoch": 1.1413719642096294, + "grad_norm": 16.125, + "learning_rate": 9.81734017773232e-07, + "loss": 0.6488, + "step": 13394 + }, + { + "epoch": 1.1414571793779293, + "grad_norm": 13.75, + "learning_rate": 9.81569900007682e-07, + "loss": 0.461, + "step": 13395 + }, + { + "epoch": 1.1415423945462293, + "grad_norm": 34.0, + "learning_rate": 9.81405787092934e-07, + "loss": 0.6902, + "step": 13396 + }, + { + "epoch": 1.1416276097145293, + "grad_norm": 14.625, + "learning_rate": 9.812416790319542e-07, + "loss": 0.272, + "step": 13397 + }, + { + "epoch": 1.1417128248828292, + "grad_norm": 21.75, + "learning_rate": 9.810775758277067e-07, + "loss": 0.5177, + "step": 13398 + }, + { + "epoch": 1.1417980400511292, + "grad_norm": 14.125, + "learning_rate": 9.809134774831583e-07, + "loss": 0.5513, + "step": 13399 + }, + { + "epoch": 1.1418832552194291, + "grad_norm": 15.1875, + "learning_rate": 9.807493840012735e-07, + "loss": 0.5496, + "step": 13400 + }, + { + "epoch": 1.141968470387729, + "grad_norm": 12.6875, + "learning_rate": 9.805852953850183e-07, + "loss": 0.327, + "step": 13401 + }, + { + "epoch": 1.142053685556029, + "grad_norm": 20.25, + "learning_rate": 9.804212116373577e-07, + "loss": 0.5069, + "step": 13402 + }, + { + "epoch": 1.142138900724329, + "grad_norm": 13.5, + "learning_rate": 9.802571327612568e-07, + "loss": 0.5047, + "step": 13403 + }, + { + "epoch": 1.142224115892629, + "grad_norm": 14.3125, + "learning_rate": 9.800930587596807e-07, + "loss": 0.5939, + "step": 13404 + }, + { + "epoch": 1.142309331060929, + "grad_norm": 18.375, + "learning_rate": 9.799289896355955e-07, + "loss": 0.6073, + "step": 13405 + }, + { + "epoch": 1.1423945462292289, + "grad_norm": 17.375, + "learning_rate": 9.79764925391964e-07, + "loss": 0.8255, + "step": 13406 + }, + { + "epoch": 1.1424797613975288, + "grad_norm": 13.375, + "learning_rate": 9.796008660317522e-07, + "loss": 0.637, + "step": 13407 + }, + { + "epoch": 1.1425649765658288, + "grad_norm": 18.25, + "learning_rate": 9.794368115579244e-07, + "loss": 0.6506, + "step": 13408 + }, + { + "epoch": 1.1426501917341287, + "grad_norm": 14.3125, + "learning_rate": 9.792727619734457e-07, + "loss": 0.7287, + "step": 13409 + }, + { + "epoch": 1.1427354069024287, + "grad_norm": 10.4375, + "learning_rate": 9.791087172812804e-07, + "loss": 0.2484, + "step": 13410 + }, + { + "epoch": 1.1428206220707287, + "grad_norm": 16.25, + "learning_rate": 9.78944677484393e-07, + "loss": 0.6125, + "step": 13411 + }, + { + "epoch": 1.1429058372390286, + "grad_norm": 12.8125, + "learning_rate": 9.787806425857478e-07, + "loss": 0.2827, + "step": 13412 + }, + { + "epoch": 1.1429910524073286, + "grad_norm": 19.625, + "learning_rate": 9.786166125883092e-07, + "loss": 1.1507, + "step": 13413 + }, + { + "epoch": 1.1430762675756285, + "grad_norm": 16.25, + "learning_rate": 9.784525874950417e-07, + "loss": 0.6389, + "step": 13414 + }, + { + "epoch": 1.1431614827439285, + "grad_norm": 16.875, + "learning_rate": 9.782885673089084e-07, + "loss": 0.969, + "step": 13415 + }, + { + "epoch": 1.1432466979122284, + "grad_norm": 20.75, + "learning_rate": 9.781245520328745e-07, + "loss": 1.0094, + "step": 13416 + }, + { + "epoch": 1.1433319130805284, + "grad_norm": 42.0, + "learning_rate": 9.779605416699028e-07, + "loss": 1.3504, + "step": 13417 + }, + { + "epoch": 1.1434171282488284, + "grad_norm": 18.25, + "learning_rate": 9.777965362229578e-07, + "loss": 0.7894, + "step": 13418 + }, + { + "epoch": 1.1435023434171283, + "grad_norm": 16.375, + "learning_rate": 9.77632535695003e-07, + "loss": 0.5772, + "step": 13419 + }, + { + "epoch": 1.1435875585854283, + "grad_norm": 12.3125, + "learning_rate": 9.774685400890024e-07, + "loss": 0.4071, + "step": 13420 + }, + { + "epoch": 1.1436727737537282, + "grad_norm": 13.0, + "learning_rate": 9.773045494079192e-07, + "loss": 0.4914, + "step": 13421 + }, + { + "epoch": 1.1437579889220282, + "grad_norm": 14.0625, + "learning_rate": 9.771405636547175e-07, + "loss": 0.5116, + "step": 13422 + }, + { + "epoch": 1.1438432040903281, + "grad_norm": 15.625, + "learning_rate": 9.769765828323597e-07, + "loss": 0.535, + "step": 13423 + }, + { + "epoch": 1.143928419258628, + "grad_norm": 14.6875, + "learning_rate": 9.768126069438097e-07, + "loss": 0.7335, + "step": 13424 + }, + { + "epoch": 1.144013634426928, + "grad_norm": 18.0, + "learning_rate": 9.766486359920304e-07, + "loss": 1.0532, + "step": 13425 + }, + { + "epoch": 1.144098849595228, + "grad_norm": 14.125, + "learning_rate": 9.764846699799854e-07, + "loss": 0.7496, + "step": 13426 + }, + { + "epoch": 1.144184064763528, + "grad_norm": 15.5625, + "learning_rate": 9.76320708910637e-07, + "loss": 0.5166, + "step": 13427 + }, + { + "epoch": 1.144269279931828, + "grad_norm": 16.375, + "learning_rate": 9.76156752786949e-07, + "loss": 0.3215, + "step": 13428 + }, + { + "epoch": 1.1443544951001279, + "grad_norm": 15.6875, + "learning_rate": 9.759928016118838e-07, + "loss": 0.8494, + "step": 13429 + }, + { + "epoch": 1.1444397102684278, + "grad_norm": 16.625, + "learning_rate": 9.758288553884044e-07, + "loss": 0.6422, + "step": 13430 + }, + { + "epoch": 1.1445249254367278, + "grad_norm": 19.25, + "learning_rate": 9.756649141194732e-07, + "loss": 0.5523, + "step": 13431 + }, + { + "epoch": 1.1446101406050277, + "grad_norm": 23.875, + "learning_rate": 9.755009778080526e-07, + "loss": 0.771, + "step": 13432 + }, + { + "epoch": 1.1446953557733277, + "grad_norm": 14.25, + "learning_rate": 9.753370464571057e-07, + "loss": 0.5322, + "step": 13433 + }, + { + "epoch": 1.1447805709416277, + "grad_norm": 13.625, + "learning_rate": 9.75173120069594e-07, + "loss": 0.652, + "step": 13434 + }, + { + "epoch": 1.1448657861099276, + "grad_norm": 18.25, + "learning_rate": 9.75009198648481e-07, + "loss": 0.4819, + "step": 13435 + }, + { + "epoch": 1.1449510012782276, + "grad_norm": 16.625, + "learning_rate": 9.74845282196728e-07, + "loss": 0.5884, + "step": 13436 + }, + { + "epoch": 1.1450362164465275, + "grad_norm": 13.25, + "learning_rate": 9.746813707172973e-07, + "loss": 0.5908, + "step": 13437 + }, + { + "epoch": 1.1451214316148275, + "grad_norm": 20.125, + "learning_rate": 9.745174642131518e-07, + "loss": 0.5963, + "step": 13438 + }, + { + "epoch": 1.1452066467831274, + "grad_norm": 19.0, + "learning_rate": 9.74353562687252e-07, + "loss": 0.6282, + "step": 13439 + }, + { + "epoch": 1.1452918619514274, + "grad_norm": 21.875, + "learning_rate": 9.741896661425609e-07, + "loss": 0.7913, + "step": 13440 + }, + { + "epoch": 1.1453770771197274, + "grad_norm": 23.375, + "learning_rate": 9.740257745820398e-07, + "loss": 1.0568, + "step": 13441 + }, + { + "epoch": 1.1454622922880273, + "grad_norm": 18.125, + "learning_rate": 9.7386188800865e-07, + "loss": 0.8974, + "step": 13442 + }, + { + "epoch": 1.1455475074563273, + "grad_norm": 18.5, + "learning_rate": 9.736980064253542e-07, + "loss": 0.8557, + "step": 13443 + }, + { + "epoch": 1.1456327226246272, + "grad_norm": 17.25, + "learning_rate": 9.735341298351128e-07, + "loss": 0.5335, + "step": 13444 + }, + { + "epoch": 1.1457179377929272, + "grad_norm": 14.9375, + "learning_rate": 9.733702582408879e-07, + "loss": 0.692, + "step": 13445 + }, + { + "epoch": 1.1458031529612271, + "grad_norm": 19.5, + "learning_rate": 9.73206391645641e-07, + "loss": 0.3934, + "step": 13446 + }, + { + "epoch": 1.145888368129527, + "grad_norm": 16.375, + "learning_rate": 9.730425300523324e-07, + "loss": 0.8496, + "step": 13447 + }, + { + "epoch": 1.145973583297827, + "grad_norm": 16.875, + "learning_rate": 9.72878673463924e-07, + "loss": 0.6115, + "step": 13448 + }, + { + "epoch": 1.146058798466127, + "grad_norm": 17.125, + "learning_rate": 9.727148218833763e-07, + "loss": 0.6767, + "step": 13449 + }, + { + "epoch": 1.146144013634427, + "grad_norm": 14.6875, + "learning_rate": 9.72550975313651e-07, + "loss": 0.9049, + "step": 13450 + }, + { + "epoch": 1.146229228802727, + "grad_norm": 12.0, + "learning_rate": 9.723871337577085e-07, + "loss": 0.4095, + "step": 13451 + }, + { + "epoch": 1.1463144439710269, + "grad_norm": 18.25, + "learning_rate": 9.722232972185094e-07, + "loss": 0.6787, + "step": 13452 + }, + { + "epoch": 1.1463996591393268, + "grad_norm": 20.625, + "learning_rate": 9.720594656990148e-07, + "loss": 0.6588, + "step": 13453 + }, + { + "epoch": 1.1464848743076268, + "grad_norm": 21.125, + "learning_rate": 9.718956392021856e-07, + "loss": 0.8133, + "step": 13454 + }, + { + "epoch": 1.1465700894759268, + "grad_norm": 13.9375, + "learning_rate": 9.717318177309812e-07, + "loss": 0.7094, + "step": 13455 + }, + { + "epoch": 1.1466553046442267, + "grad_norm": 15.75, + "learning_rate": 9.71568001288363e-07, + "loss": 0.7029, + "step": 13456 + }, + { + "epoch": 1.1467405198125267, + "grad_norm": 19.625, + "learning_rate": 9.714041898772908e-07, + "loss": 0.396, + "step": 13457 + }, + { + "epoch": 1.1468257349808266, + "grad_norm": 14.25, + "learning_rate": 9.712403835007252e-07, + "loss": 0.3976, + "step": 13458 + }, + { + "epoch": 1.1469109501491266, + "grad_norm": 24.0, + "learning_rate": 9.71076582161626e-07, + "loss": 0.598, + "step": 13459 + }, + { + "epoch": 1.1469961653174265, + "grad_norm": 16.75, + "learning_rate": 9.709127858629536e-07, + "loss": 0.4045, + "step": 13460 + }, + { + "epoch": 1.1470813804857265, + "grad_norm": 17.75, + "learning_rate": 9.707489946076675e-07, + "loss": 0.4125, + "step": 13461 + }, + { + "epoch": 1.1471665956540265, + "grad_norm": 13.6875, + "learning_rate": 9.705852083987285e-07, + "loss": 0.5206, + "step": 13462 + }, + { + "epoch": 1.1472518108223264, + "grad_norm": 16.125, + "learning_rate": 9.704214272390956e-07, + "loss": 0.5586, + "step": 13463 + }, + { + "epoch": 1.1473370259906264, + "grad_norm": 17.5, + "learning_rate": 9.702576511317283e-07, + "loss": 0.682, + "step": 13464 + }, + { + "epoch": 1.1474222411589263, + "grad_norm": 17.625, + "learning_rate": 9.700938800795867e-07, + "loss": 0.8917, + "step": 13465 + }, + { + "epoch": 1.1475074563272263, + "grad_norm": 19.875, + "learning_rate": 9.699301140856304e-07, + "loss": 0.5282, + "step": 13466 + }, + { + "epoch": 1.1475926714955262, + "grad_norm": 12.0625, + "learning_rate": 9.69766353152818e-07, + "loss": 0.489, + "step": 13467 + }, + { + "epoch": 1.1476778866638262, + "grad_norm": 12.875, + "learning_rate": 9.696025972841099e-07, + "loss": 0.338, + "step": 13468 + }, + { + "epoch": 1.1477631018321262, + "grad_norm": 19.0, + "learning_rate": 9.694388464824645e-07, + "loss": 0.7455, + "step": 13469 + }, + { + "epoch": 1.147848317000426, + "grad_norm": 12.375, + "learning_rate": 9.692751007508416e-07, + "loss": 0.4396, + "step": 13470 + }, + { + "epoch": 1.147933532168726, + "grad_norm": 16.25, + "learning_rate": 9.691113600921998e-07, + "loss": 0.8635, + "step": 13471 + }, + { + "epoch": 1.148018747337026, + "grad_norm": 18.125, + "learning_rate": 9.68947624509498e-07, + "loss": 0.4474, + "step": 13472 + }, + { + "epoch": 1.148103962505326, + "grad_norm": 18.75, + "learning_rate": 9.687838940056956e-07, + "loss": 0.7303, + "step": 13473 + }, + { + "epoch": 1.148189177673626, + "grad_norm": 21.0, + "learning_rate": 9.686201685837504e-07, + "loss": 0.8052, + "step": 13474 + }, + { + "epoch": 1.148274392841926, + "grad_norm": 17.0, + "learning_rate": 9.684564482466222e-07, + "loss": 0.7654, + "step": 13475 + }, + { + "epoch": 1.1483596080102259, + "grad_norm": 12.9375, + "learning_rate": 9.682927329972689e-07, + "loss": 0.4802, + "step": 13476 + }, + { + "epoch": 1.1484448231785258, + "grad_norm": 17.125, + "learning_rate": 9.681290228386494e-07, + "loss": 0.7754, + "step": 13477 + }, + { + "epoch": 1.1485300383468258, + "grad_norm": 12.75, + "learning_rate": 9.679653177737216e-07, + "loss": 0.362, + "step": 13478 + }, + { + "epoch": 1.1486152535151257, + "grad_norm": 14.625, + "learning_rate": 9.678016178054446e-07, + "loss": 0.7624, + "step": 13479 + }, + { + "epoch": 1.1487004686834257, + "grad_norm": 13.4375, + "learning_rate": 9.676379229367759e-07, + "loss": 0.3636, + "step": 13480 + }, + { + "epoch": 1.1487856838517256, + "grad_norm": 16.25, + "learning_rate": 9.674742331706738e-07, + "loss": 0.5063, + "step": 13481 + }, + { + "epoch": 1.1488708990200256, + "grad_norm": 18.5, + "learning_rate": 9.673105485100963e-07, + "loss": 0.442, + "step": 13482 + }, + { + "epoch": 1.1489561141883256, + "grad_norm": 12.625, + "learning_rate": 9.671468689580017e-07, + "loss": 0.3088, + "step": 13483 + }, + { + "epoch": 1.1490413293566255, + "grad_norm": 14.125, + "learning_rate": 9.669831945173472e-07, + "loss": 0.5611, + "step": 13484 + }, + { + "epoch": 1.1491265445249255, + "grad_norm": 14.5625, + "learning_rate": 9.668195251910915e-07, + "loss": 0.5355, + "step": 13485 + }, + { + "epoch": 1.1492117596932254, + "grad_norm": 13.0, + "learning_rate": 9.66655860982191e-07, + "loss": 0.6047, + "step": 13486 + }, + { + "epoch": 1.1492969748615254, + "grad_norm": 17.875, + "learning_rate": 9.664922018936051e-07, + "loss": 0.7379, + "step": 13487 + }, + { + "epoch": 1.1493821900298253, + "grad_norm": 14.875, + "learning_rate": 9.663285479282895e-07, + "loss": 0.7082, + "step": 13488 + }, + { + "epoch": 1.1494674051981253, + "grad_norm": 15.8125, + "learning_rate": 9.661648990892023e-07, + "loss": 0.514, + "step": 13489 + }, + { + "epoch": 1.1495526203664252, + "grad_norm": 12.6875, + "learning_rate": 9.660012553793009e-07, + "loss": 0.5388, + "step": 13490 + }, + { + "epoch": 1.1496378355347252, + "grad_norm": 16.75, + "learning_rate": 9.658376168015422e-07, + "loss": 0.6193, + "step": 13491 + }, + { + "epoch": 1.1497230507030252, + "grad_norm": 16.125, + "learning_rate": 9.656739833588837e-07, + "loss": 0.8265, + "step": 13492 + }, + { + "epoch": 1.1498082658713251, + "grad_norm": 11.4375, + "learning_rate": 9.655103550542822e-07, + "loss": 0.3289, + "step": 13493 + }, + { + "epoch": 1.149893481039625, + "grad_norm": 20.25, + "learning_rate": 9.653467318906948e-07, + "loss": 0.8454, + "step": 13494 + }, + { + "epoch": 1.149978696207925, + "grad_norm": 14.75, + "learning_rate": 9.651831138710784e-07, + "loss": 0.3903, + "step": 13495 + }, + { + "epoch": 1.150063911376225, + "grad_norm": 25.25, + "learning_rate": 9.650195009983894e-07, + "loss": 0.6945, + "step": 13496 + }, + { + "epoch": 1.150149126544525, + "grad_norm": 24.625, + "learning_rate": 9.648558932755841e-07, + "loss": 0.9499, + "step": 13497 + }, + { + "epoch": 1.150234341712825, + "grad_norm": 24.625, + "learning_rate": 9.6469229070562e-07, + "loss": 0.6027, + "step": 13498 + }, + { + "epoch": 1.1503195568811249, + "grad_norm": 37.0, + "learning_rate": 9.645286932914526e-07, + "loss": 0.5165, + "step": 13499 + }, + { + "epoch": 1.1504047720494248, + "grad_norm": 24.25, + "learning_rate": 9.643651010360393e-07, + "loss": 0.8691, + "step": 13500 + }, + { + "epoch": 1.1504899872177248, + "grad_norm": 13.1875, + "learning_rate": 9.642015139423355e-07, + "loss": 0.6577, + "step": 13501 + }, + { + "epoch": 1.1505752023860247, + "grad_norm": 14.875, + "learning_rate": 9.640379320132982e-07, + "loss": 0.3295, + "step": 13502 + }, + { + "epoch": 1.1506604175543247, + "grad_norm": 15.0, + "learning_rate": 9.63874355251883e-07, + "loss": 0.6709, + "step": 13503 + }, + { + "epoch": 1.1507456327226246, + "grad_norm": 21.125, + "learning_rate": 9.637107836610455e-07, + "loss": 0.7441, + "step": 13504 + }, + { + "epoch": 1.1508308478909246, + "grad_norm": 20.0, + "learning_rate": 9.63547217243742e-07, + "loss": 0.7832, + "step": 13505 + }, + { + "epoch": 1.1509160630592246, + "grad_norm": 17.25, + "learning_rate": 9.633836560029286e-07, + "loss": 0.7449, + "step": 13506 + }, + { + "epoch": 1.1510012782275245, + "grad_norm": 32.5, + "learning_rate": 9.632200999415607e-07, + "loss": 0.544, + "step": 13507 + }, + { + "epoch": 1.1510864933958245, + "grad_norm": 23.25, + "learning_rate": 9.630565490625938e-07, + "loss": 0.9638, + "step": 13508 + }, + { + "epoch": 1.1511717085641244, + "grad_norm": 14.5625, + "learning_rate": 9.628930033689837e-07, + "loss": 0.3941, + "step": 13509 + }, + { + "epoch": 1.1512569237324244, + "grad_norm": 17.625, + "learning_rate": 9.627294628636857e-07, + "loss": 0.7794, + "step": 13510 + }, + { + "epoch": 1.1513421389007243, + "grad_norm": 20.125, + "learning_rate": 9.625659275496558e-07, + "loss": 0.8737, + "step": 13511 + }, + { + "epoch": 1.1514273540690243, + "grad_norm": 10.1875, + "learning_rate": 9.62402397429848e-07, + "loss": 0.2886, + "step": 13512 + }, + { + "epoch": 1.1515125692373243, + "grad_norm": 17.125, + "learning_rate": 9.622388725072182e-07, + "loss": 0.6164, + "step": 13513 + }, + { + "epoch": 1.1515977844056242, + "grad_norm": 14.625, + "learning_rate": 9.620753527847212e-07, + "loss": 0.5109, + "step": 13514 + }, + { + "epoch": 1.1516829995739242, + "grad_norm": 26.125, + "learning_rate": 9.619118382653123e-07, + "loss": 0.93, + "step": 13515 + }, + { + "epoch": 1.1517682147422241, + "grad_norm": 15.5625, + "learning_rate": 9.61748328951946e-07, + "loss": 0.633, + "step": 13516 + }, + { + "epoch": 1.151853429910524, + "grad_norm": 15.0625, + "learning_rate": 9.615848248475774e-07, + "loss": 0.7057, + "step": 13517 + }, + { + "epoch": 1.151938645078824, + "grad_norm": 19.75, + "learning_rate": 9.614213259551607e-07, + "loss": 0.7244, + "step": 13518 + }, + { + "epoch": 1.152023860247124, + "grad_norm": 11.5625, + "learning_rate": 9.612578322776516e-07, + "loss": 0.2486, + "step": 13519 + }, + { + "epoch": 1.152109075415424, + "grad_norm": 15.375, + "learning_rate": 9.610943438180033e-07, + "loss": 0.956, + "step": 13520 + }, + { + "epoch": 1.152194290583724, + "grad_norm": 17.0, + "learning_rate": 9.609308605791705e-07, + "loss": 0.6803, + "step": 13521 + }, + { + "epoch": 1.1522795057520239, + "grad_norm": 13.4375, + "learning_rate": 9.607673825641083e-07, + "loss": 0.4474, + "step": 13522 + }, + { + "epoch": 1.1523647209203238, + "grad_norm": 12.4375, + "learning_rate": 9.606039097757698e-07, + "loss": 0.3392, + "step": 13523 + }, + { + "epoch": 1.1524499360886238, + "grad_norm": 16.875, + "learning_rate": 9.6044044221711e-07, + "loss": 0.8876, + "step": 13524 + }, + { + "epoch": 1.1525351512569237, + "grad_norm": 14.3125, + "learning_rate": 9.602769798910824e-07, + "loss": 0.7512, + "step": 13525 + }, + { + "epoch": 1.1526203664252237, + "grad_norm": 18.5, + "learning_rate": 9.601135228006412e-07, + "loss": 0.3645, + "step": 13526 + }, + { + "epoch": 1.1527055815935237, + "grad_norm": 15.5, + "learning_rate": 9.599500709487405e-07, + "loss": 0.7833, + "step": 13527 + }, + { + "epoch": 1.1527907967618236, + "grad_norm": 20.625, + "learning_rate": 9.597866243383334e-07, + "loss": 0.6381, + "step": 13528 + }, + { + "epoch": 1.1528760119301236, + "grad_norm": 12.875, + "learning_rate": 9.596231829723736e-07, + "loss": 0.7029, + "step": 13529 + }, + { + "epoch": 1.1529612270984235, + "grad_norm": 22.75, + "learning_rate": 9.594597468538151e-07, + "loss": 0.9073, + "step": 13530 + }, + { + "epoch": 1.1530464422667235, + "grad_norm": 14.6875, + "learning_rate": 9.59296315985611e-07, + "loss": 0.5385, + "step": 13531 + }, + { + "epoch": 1.1531316574350234, + "grad_norm": 15.5625, + "learning_rate": 9.59132890370715e-07, + "loss": 0.505, + "step": 13532 + }, + { + "epoch": 1.1532168726033234, + "grad_norm": 14.625, + "learning_rate": 9.589694700120798e-07, + "loss": 0.7409, + "step": 13533 + }, + { + "epoch": 1.1533020877716234, + "grad_norm": 14.625, + "learning_rate": 9.588060549126591e-07, + "loss": 0.5704, + "step": 13534 + }, + { + "epoch": 1.1533873029399233, + "grad_norm": 10.25, + "learning_rate": 9.586426450754063e-07, + "loss": 0.2538, + "step": 13535 + }, + { + "epoch": 1.1534725181082233, + "grad_norm": 13.0, + "learning_rate": 9.584792405032734e-07, + "loss": 0.3588, + "step": 13536 + }, + { + "epoch": 1.1535577332765232, + "grad_norm": 17.5, + "learning_rate": 9.583158411992139e-07, + "loss": 0.8585, + "step": 13537 + }, + { + "epoch": 1.1536429484448232, + "grad_norm": 18.125, + "learning_rate": 9.581524471661802e-07, + "loss": 0.6913, + "step": 13538 + }, + { + "epoch": 1.1537281636131231, + "grad_norm": 16.875, + "learning_rate": 9.579890584071254e-07, + "loss": 0.6071, + "step": 13539 + }, + { + "epoch": 1.153813378781423, + "grad_norm": 14.4375, + "learning_rate": 9.578256749250018e-07, + "loss": 0.7357, + "step": 13540 + }, + { + "epoch": 1.153898593949723, + "grad_norm": 16.625, + "learning_rate": 9.576622967227626e-07, + "loss": 0.7746, + "step": 13541 + }, + { + "epoch": 1.153983809118023, + "grad_norm": 18.0, + "learning_rate": 9.57498923803359e-07, + "loss": 0.7199, + "step": 13542 + }, + { + "epoch": 1.154069024286323, + "grad_norm": 13.375, + "learning_rate": 9.573355561697443e-07, + "loss": 0.4421, + "step": 13543 + }, + { + "epoch": 1.154154239454623, + "grad_norm": 13.3125, + "learning_rate": 9.57172193824871e-07, + "loss": 0.6116, + "step": 13544 + }, + { + "epoch": 1.1542394546229229, + "grad_norm": 10.875, + "learning_rate": 9.5700883677169e-07, + "loss": 0.3635, + "step": 13545 + }, + { + "epoch": 1.1543246697912228, + "grad_norm": 20.625, + "learning_rate": 9.568454850131536e-07, + "loss": 0.7502, + "step": 13546 + }, + { + "epoch": 1.1544098849595228, + "grad_norm": 9.6875, + "learning_rate": 9.566821385522145e-07, + "loss": 0.2851, + "step": 13547 + }, + { + "epoch": 1.1544951001278227, + "grad_norm": 18.5, + "learning_rate": 9.56518797391824e-07, + "loss": 0.8178, + "step": 13548 + }, + { + "epoch": 1.1545803152961227, + "grad_norm": 20.75, + "learning_rate": 9.56355461534934e-07, + "loss": 0.7996, + "step": 13549 + }, + { + "epoch": 1.1546655304644227, + "grad_norm": 22.75, + "learning_rate": 9.56192130984496e-07, + "loss": 0.5011, + "step": 13550 + }, + { + "epoch": 1.1547507456327226, + "grad_norm": 16.25, + "learning_rate": 9.560288057434618e-07, + "loss": 0.7961, + "step": 13551 + }, + { + "epoch": 1.1548359608010226, + "grad_norm": 14.1875, + "learning_rate": 9.558654858147829e-07, + "loss": 0.5697, + "step": 13552 + }, + { + "epoch": 1.1549211759693225, + "grad_norm": 18.875, + "learning_rate": 9.5570217120141e-07, + "loss": 0.5314, + "step": 13553 + }, + { + "epoch": 1.1550063911376225, + "grad_norm": 14.0625, + "learning_rate": 9.55538861906295e-07, + "loss": 0.6523, + "step": 13554 + }, + { + "epoch": 1.1550916063059224, + "grad_norm": 19.5, + "learning_rate": 9.553755579323888e-07, + "loss": 0.799, + "step": 13555 + }, + { + "epoch": 1.1551768214742224, + "grad_norm": 10.0, + "learning_rate": 9.552122592826427e-07, + "loss": 0.2523, + "step": 13556 + }, + { + "epoch": 1.1552620366425224, + "grad_norm": 15.6875, + "learning_rate": 9.55048965960007e-07, + "loss": 0.5398, + "step": 13557 + }, + { + "epoch": 1.1553472518108223, + "grad_norm": 14.8125, + "learning_rate": 9.548856779674338e-07, + "loss": 0.4667, + "step": 13558 + }, + { + "epoch": 1.1554324669791223, + "grad_norm": 15.125, + "learning_rate": 9.54722395307873e-07, + "loss": 0.5802, + "step": 13559 + }, + { + "epoch": 1.1555176821474222, + "grad_norm": 20.125, + "learning_rate": 9.545591179842755e-07, + "loss": 0.6534, + "step": 13560 + }, + { + "epoch": 1.1556028973157222, + "grad_norm": 13.8125, + "learning_rate": 9.543958459995914e-07, + "loss": 0.3055, + "step": 13561 + }, + { + "epoch": 1.1556881124840221, + "grad_norm": 14.9375, + "learning_rate": 9.542325793567718e-07, + "loss": 0.6445, + "step": 13562 + }, + { + "epoch": 1.155773327652322, + "grad_norm": 13.1875, + "learning_rate": 9.540693180587667e-07, + "loss": 0.361, + "step": 13563 + }, + { + "epoch": 1.155858542820622, + "grad_norm": 11.1875, + "learning_rate": 9.53906062108527e-07, + "loss": 0.3159, + "step": 13564 + }, + { + "epoch": 1.155943757988922, + "grad_norm": 23.5, + "learning_rate": 9.53742811509002e-07, + "loss": 0.8819, + "step": 13565 + }, + { + "epoch": 1.156028973157222, + "grad_norm": 13.25, + "learning_rate": 9.535795662631428e-07, + "loss": 0.5447, + "step": 13566 + }, + { + "epoch": 1.156114188325522, + "grad_norm": 12.4375, + "learning_rate": 9.534163263738985e-07, + "loss": 0.3405, + "step": 13567 + }, + { + "epoch": 1.1561994034938219, + "grad_norm": 15.375, + "learning_rate": 9.532530918442201e-07, + "loss": 0.6456, + "step": 13568 + }, + { + "epoch": 1.1562846186621218, + "grad_norm": 13.1875, + "learning_rate": 9.530898626770564e-07, + "loss": 0.5433, + "step": 13569 + }, + { + "epoch": 1.1563698338304218, + "grad_norm": 15.125, + "learning_rate": 9.529266388753572e-07, + "loss": 0.6616, + "step": 13570 + }, + { + "epoch": 1.1564550489987218, + "grad_norm": 49.0, + "learning_rate": 9.527634204420726e-07, + "loss": 0.9666, + "step": 13571 + }, + { + "epoch": 1.1565402641670217, + "grad_norm": 11.6875, + "learning_rate": 9.526002073801516e-07, + "loss": 0.4294, + "step": 13572 + }, + { + "epoch": 1.1566254793353217, + "grad_norm": 13.3125, + "learning_rate": 9.524369996925442e-07, + "loss": 0.5046, + "step": 13573 + }, + { + "epoch": 1.1567106945036216, + "grad_norm": 12.6875, + "learning_rate": 9.522737973821994e-07, + "loss": 0.3816, + "step": 13574 + }, + { + "epoch": 1.1567959096719216, + "grad_norm": 15.0625, + "learning_rate": 9.521106004520662e-07, + "loss": 0.7074, + "step": 13575 + }, + { + "epoch": 1.1568811248402215, + "grad_norm": 14.8125, + "learning_rate": 9.519474089050945e-07, + "loss": 0.5165, + "step": 13576 + }, + { + "epoch": 1.1569663400085215, + "grad_norm": 15.5625, + "learning_rate": 9.517842227442327e-07, + "loss": 0.8808, + "step": 13577 + }, + { + "epoch": 1.1570515551768215, + "grad_norm": 16.125, + "learning_rate": 9.516210419724295e-07, + "loss": 0.5519, + "step": 13578 + }, + { + "epoch": 1.1571367703451214, + "grad_norm": 16.625, + "learning_rate": 9.514578665926345e-07, + "loss": 0.7716, + "step": 13579 + }, + { + "epoch": 1.1572219855134214, + "grad_norm": 20.25, + "learning_rate": 9.512946966077957e-07, + "loss": 0.6582, + "step": 13580 + }, + { + "epoch": 1.1573072006817213, + "grad_norm": 14.0625, + "learning_rate": 9.511315320208623e-07, + "loss": 0.5491, + "step": 13581 + }, + { + "epoch": 1.1573924158500213, + "grad_norm": 13.6875, + "learning_rate": 9.509683728347825e-07, + "loss": 0.326, + "step": 13582 + }, + { + "epoch": 1.1574776310183212, + "grad_norm": 15.3125, + "learning_rate": 9.508052190525052e-07, + "loss": 0.5705, + "step": 13583 + }, + { + "epoch": 1.1575628461866212, + "grad_norm": 13.25, + "learning_rate": 9.506420706769787e-07, + "loss": 0.627, + "step": 13584 + }, + { + "epoch": 1.1576480613549212, + "grad_norm": 13.875, + "learning_rate": 9.504789277111504e-07, + "loss": 0.4842, + "step": 13585 + }, + { + "epoch": 1.157733276523221, + "grad_norm": 15.5625, + "learning_rate": 9.503157901579693e-07, + "loss": 0.7308, + "step": 13586 + }, + { + "epoch": 1.157818491691521, + "grad_norm": 13.25, + "learning_rate": 9.501526580203832e-07, + "loss": 0.5224, + "step": 13587 + }, + { + "epoch": 1.157903706859821, + "grad_norm": 13.875, + "learning_rate": 9.499895313013401e-07, + "loss": 0.6051, + "step": 13588 + }, + { + "epoch": 1.157988922028121, + "grad_norm": 13.0625, + "learning_rate": 9.498264100037879e-07, + "loss": 0.552, + "step": 13589 + }, + { + "epoch": 1.158074137196421, + "grad_norm": 22.125, + "learning_rate": 9.496632941306741e-07, + "loss": 0.5993, + "step": 13590 + }, + { + "epoch": 1.158159352364721, + "grad_norm": 12.5625, + "learning_rate": 9.495001836849468e-07, + "loss": 0.4013, + "step": 13591 + }, + { + "epoch": 1.1582445675330209, + "grad_norm": 15.4375, + "learning_rate": 9.493370786695538e-07, + "loss": 0.5326, + "step": 13592 + }, + { + "epoch": 1.1583297827013208, + "grad_norm": 15.0, + "learning_rate": 9.491739790874414e-07, + "loss": 0.598, + "step": 13593 + }, + { + "epoch": 1.1584149978696208, + "grad_norm": 16.5, + "learning_rate": 9.49010884941558e-07, + "loss": 0.6714, + "step": 13594 + }, + { + "epoch": 1.1585002130379207, + "grad_norm": 11.0, + "learning_rate": 9.488477962348506e-07, + "loss": 0.3743, + "step": 13595 + }, + { + "epoch": 1.1585854282062207, + "grad_norm": 20.5, + "learning_rate": 9.486847129702661e-07, + "loss": 0.8852, + "step": 13596 + }, + { + "epoch": 1.1586706433745206, + "grad_norm": 10.875, + "learning_rate": 9.48521635150752e-07, + "loss": 0.3323, + "step": 13597 + }, + { + "epoch": 1.1587558585428206, + "grad_norm": 17.25, + "learning_rate": 9.483585627792552e-07, + "loss": 0.9642, + "step": 13598 + }, + { + "epoch": 1.1588410737111206, + "grad_norm": 29.0, + "learning_rate": 9.481954958587223e-07, + "loss": 0.4036, + "step": 13599 + }, + { + "epoch": 1.1589262888794205, + "grad_norm": 16.0, + "learning_rate": 9.480324343921006e-07, + "loss": 0.614, + "step": 13600 + }, + { + "epoch": 1.1590115040477205, + "grad_norm": 21.125, + "learning_rate": 9.478693783823364e-07, + "loss": 0.7071, + "step": 13601 + }, + { + "epoch": 1.1590967192160204, + "grad_norm": 15.75, + "learning_rate": 9.47706327832376e-07, + "loss": 0.5266, + "step": 13602 + }, + { + "epoch": 1.1591819343843204, + "grad_norm": 14.4375, + "learning_rate": 9.475432827451666e-07, + "loss": 0.5305, + "step": 13603 + }, + { + "epoch": 1.1592671495526203, + "grad_norm": 12.9375, + "learning_rate": 9.473802431236538e-07, + "loss": 0.3812, + "step": 13604 + }, + { + "epoch": 1.1593523647209203, + "grad_norm": 13.875, + "learning_rate": 9.472172089707847e-07, + "loss": 0.4425, + "step": 13605 + }, + { + "epoch": 1.1594375798892202, + "grad_norm": 17.5, + "learning_rate": 9.47054180289505e-07, + "loss": 0.4946, + "step": 13606 + }, + { + "epoch": 1.1595227950575202, + "grad_norm": 17.0, + "learning_rate": 9.468911570827607e-07, + "loss": 0.7057, + "step": 13607 + }, + { + "epoch": 1.1596080102258202, + "grad_norm": 22.25, + "learning_rate": 9.467281393534986e-07, + "loss": 1.0526, + "step": 13608 + }, + { + "epoch": 1.1596932253941201, + "grad_norm": 15.0625, + "learning_rate": 9.465651271046637e-07, + "loss": 0.6051, + "step": 13609 + }, + { + "epoch": 1.15977844056242, + "grad_norm": 14.1875, + "learning_rate": 9.464021203392018e-07, + "loss": 0.5853, + "step": 13610 + }, + { + "epoch": 1.15986365573072, + "grad_norm": 17.5, + "learning_rate": 9.46239119060059e-07, + "loss": 0.7617, + "step": 13611 + }, + { + "epoch": 1.15994887089902, + "grad_norm": 15.4375, + "learning_rate": 9.460761232701807e-07, + "loss": 0.2305, + "step": 13612 + }, + { + "epoch": 1.16003408606732, + "grad_norm": 15.125, + "learning_rate": 9.459131329725127e-07, + "loss": 0.7347, + "step": 13613 + }, + { + "epoch": 1.16011930123562, + "grad_norm": 17.625, + "learning_rate": 9.457501481699997e-07, + "loss": 0.5301, + "step": 13614 + }, + { + "epoch": 1.1602045164039199, + "grad_norm": 14.125, + "learning_rate": 9.45587168865588e-07, + "loss": 0.6893, + "step": 13615 + }, + { + "epoch": 1.1602897315722198, + "grad_norm": 17.625, + "learning_rate": 9.454241950622219e-07, + "loss": 0.7127, + "step": 13616 + }, + { + "epoch": 1.1603749467405198, + "grad_norm": 20.25, + "learning_rate": 9.452612267628473e-07, + "loss": 0.9187, + "step": 13617 + }, + { + "epoch": 1.1604601619088197, + "grad_norm": 14.75, + "learning_rate": 9.450982639704087e-07, + "loss": 0.6128, + "step": 13618 + }, + { + "epoch": 1.1605453770771197, + "grad_norm": 12.125, + "learning_rate": 9.449353066878507e-07, + "loss": 0.4894, + "step": 13619 + }, + { + "epoch": 1.1606305922454196, + "grad_norm": 17.75, + "learning_rate": 9.447723549181187e-07, + "loss": 0.9162, + "step": 13620 + }, + { + "epoch": 1.1607158074137196, + "grad_norm": 15.5625, + "learning_rate": 9.446094086641573e-07, + "loss": 0.989, + "step": 13621 + }, + { + "epoch": 1.1608010225820196, + "grad_norm": 17.0, + "learning_rate": 9.444464679289106e-07, + "loss": 0.6169, + "step": 13622 + }, + { + "epoch": 1.1608862377503195, + "grad_norm": 15.9375, + "learning_rate": 9.442835327153238e-07, + "loss": 0.4222, + "step": 13623 + }, + { + "epoch": 1.1609714529186195, + "grad_norm": 16.5, + "learning_rate": 9.441206030263407e-07, + "loss": 0.6476, + "step": 13624 + }, + { + "epoch": 1.1610566680869194, + "grad_norm": 14.9375, + "learning_rate": 9.439576788649065e-07, + "loss": 0.4112, + "step": 13625 + }, + { + "epoch": 1.1611418832552194, + "grad_norm": 14.0625, + "learning_rate": 9.437947602339644e-07, + "loss": 0.6339, + "step": 13626 + }, + { + "epoch": 1.1612270984235193, + "grad_norm": 16.5, + "learning_rate": 9.436318471364586e-07, + "loss": 0.6932, + "step": 13627 + }, + { + "epoch": 1.1613123135918193, + "grad_norm": 18.75, + "learning_rate": 9.434689395753338e-07, + "loss": 0.9164, + "step": 13628 + }, + { + "epoch": 1.1613975287601193, + "grad_norm": 23.125, + "learning_rate": 9.433060375535333e-07, + "loss": 0.5634, + "step": 13629 + }, + { + "epoch": 1.1614827439284192, + "grad_norm": 18.75, + "learning_rate": 9.431431410740012e-07, + "loss": 0.7379, + "step": 13630 + }, + { + "epoch": 1.1615679590967192, + "grad_norm": 16.5, + "learning_rate": 9.42980250139681e-07, + "loss": 0.6562, + "step": 13631 + }, + { + "epoch": 1.1616531742650191, + "grad_norm": 12.25, + "learning_rate": 9.428173647535165e-07, + "loss": 0.3048, + "step": 13632 + }, + { + "epoch": 1.161738389433319, + "grad_norm": 16.625, + "learning_rate": 9.426544849184514e-07, + "loss": 0.6302, + "step": 13633 + }, + { + "epoch": 1.161823604601619, + "grad_norm": 17.875, + "learning_rate": 9.424916106374283e-07, + "loss": 0.4418, + "step": 13634 + }, + { + "epoch": 1.161908819769919, + "grad_norm": 13.8125, + "learning_rate": 9.423287419133914e-07, + "loss": 0.3419, + "step": 13635 + }, + { + "epoch": 1.161994034938219, + "grad_norm": 15.875, + "learning_rate": 9.421658787492835e-07, + "loss": 0.4799, + "step": 13636 + }, + { + "epoch": 1.162079250106519, + "grad_norm": 28.0, + "learning_rate": 9.420030211480474e-07, + "loss": 0.6386, + "step": 13637 + }, + { + "epoch": 1.1621644652748189, + "grad_norm": 11.125, + "learning_rate": 9.418401691126266e-07, + "loss": 0.1991, + "step": 13638 + }, + { + "epoch": 1.1622496804431188, + "grad_norm": 18.375, + "learning_rate": 9.416773226459637e-07, + "loss": 0.7864, + "step": 13639 + }, + { + "epoch": 1.1623348956114188, + "grad_norm": 18.375, + "learning_rate": 9.415144817510019e-07, + "loss": 0.9134, + "step": 13640 + }, + { + "epoch": 1.1624201107797187, + "grad_norm": 18.75, + "learning_rate": 9.413516464306838e-07, + "loss": 0.3923, + "step": 13641 + }, + { + "epoch": 1.1625053259480187, + "grad_norm": 15.1875, + "learning_rate": 9.411888166879512e-07, + "loss": 0.3891, + "step": 13642 + }, + { + "epoch": 1.1625905411163187, + "grad_norm": 15.8125, + "learning_rate": 9.410259925257477e-07, + "loss": 0.7715, + "step": 13643 + }, + { + "epoch": 1.1626757562846186, + "grad_norm": 15.0625, + "learning_rate": 9.408631739470149e-07, + "loss": 0.5817, + "step": 13644 + }, + { + "epoch": 1.1627609714529186, + "grad_norm": 12.9375, + "learning_rate": 9.407003609546955e-07, + "loss": 0.2861, + "step": 13645 + }, + { + "epoch": 1.1628461866212185, + "grad_norm": 13.8125, + "learning_rate": 9.405375535517315e-07, + "loss": 0.3416, + "step": 13646 + }, + { + "epoch": 1.1629314017895185, + "grad_norm": 13.375, + "learning_rate": 9.403747517410653e-07, + "loss": 0.3693, + "step": 13647 + }, + { + "epoch": 1.1630166169578184, + "grad_norm": 16.0, + "learning_rate": 9.402119555256385e-07, + "loss": 0.6877, + "step": 13648 + }, + { + "epoch": 1.1631018321261184, + "grad_norm": 13.3125, + "learning_rate": 9.400491649083937e-07, + "loss": 0.4036, + "step": 13649 + }, + { + "epoch": 1.1631870472944184, + "grad_norm": 18.375, + "learning_rate": 9.398863798922718e-07, + "loss": 0.5185, + "step": 13650 + }, + { + "epoch": 1.1632722624627183, + "grad_norm": 10.9375, + "learning_rate": 9.397236004802149e-07, + "loss": 0.2909, + "step": 13651 + }, + { + "epoch": 1.1633574776310183, + "grad_norm": 14.375, + "learning_rate": 9.395608266751643e-07, + "loss": 0.6689, + "step": 13652 + }, + { + "epoch": 1.1634426927993182, + "grad_norm": 16.125, + "learning_rate": 9.393980584800621e-07, + "loss": 0.7906, + "step": 13653 + }, + { + "epoch": 1.1635279079676182, + "grad_norm": 16.125, + "learning_rate": 9.392352958978489e-07, + "loss": 0.9041, + "step": 13654 + }, + { + "epoch": 1.1636131231359181, + "grad_norm": 16.0, + "learning_rate": 9.390725389314667e-07, + "loss": 0.6811, + "step": 13655 + }, + { + "epoch": 1.163698338304218, + "grad_norm": 17.625, + "learning_rate": 9.389097875838562e-07, + "loss": 0.8108, + "step": 13656 + }, + { + "epoch": 1.163783553472518, + "grad_norm": 15.1875, + "learning_rate": 9.38747041857959e-07, + "loss": 0.5247, + "step": 13657 + }, + { + "epoch": 1.163868768640818, + "grad_norm": 14.6875, + "learning_rate": 9.385843017567156e-07, + "loss": 0.6277, + "step": 13658 + }, + { + "epoch": 1.163953983809118, + "grad_norm": 13.875, + "learning_rate": 9.384215672830668e-07, + "loss": 0.5695, + "step": 13659 + }, + { + "epoch": 1.164039198977418, + "grad_norm": 13.3125, + "learning_rate": 9.382588384399538e-07, + "loss": 0.8171, + "step": 13660 + }, + { + "epoch": 1.1641244141457179, + "grad_norm": 21.5, + "learning_rate": 9.380961152303168e-07, + "loss": 0.9488, + "step": 13661 + }, + { + "epoch": 1.1642096293140178, + "grad_norm": 12.625, + "learning_rate": 9.37933397657097e-07, + "loss": 0.4288, + "step": 13662 + }, + { + "epoch": 1.1642948444823178, + "grad_norm": 17.75, + "learning_rate": 9.377706857232341e-07, + "loss": 0.4675, + "step": 13663 + }, + { + "epoch": 1.1643800596506177, + "grad_norm": 11.8125, + "learning_rate": 9.376079794316692e-07, + "loss": 0.4697, + "step": 13664 + }, + { + "epoch": 1.1644652748189177, + "grad_norm": 13.5, + "learning_rate": 9.374452787853426e-07, + "loss": 0.602, + "step": 13665 + }, + { + "epoch": 1.1645504899872177, + "grad_norm": 13.375, + "learning_rate": 9.372825837871938e-07, + "loss": 0.5384, + "step": 13666 + }, + { + "epoch": 1.1646357051555176, + "grad_norm": 8.5, + "learning_rate": 9.371198944401628e-07, + "loss": 0.1984, + "step": 13667 + }, + { + "epoch": 1.1647209203238176, + "grad_norm": 13.4375, + "learning_rate": 9.369572107471903e-07, + "loss": 0.6151, + "step": 13668 + }, + { + "epoch": 1.1648061354921175, + "grad_norm": 17.25, + "learning_rate": 9.367945327112155e-07, + "loss": 0.5947, + "step": 13669 + }, + { + "epoch": 1.1648913506604175, + "grad_norm": 13.1875, + "learning_rate": 9.366318603351787e-07, + "loss": 0.5175, + "step": 13670 + }, + { + "epoch": 1.1649765658287174, + "grad_norm": 14.5625, + "learning_rate": 9.364691936220191e-07, + "loss": 0.6977, + "step": 13671 + }, + { + "epoch": 1.1650617809970174, + "grad_norm": 20.375, + "learning_rate": 9.363065325746767e-07, + "loss": 0.5001, + "step": 13672 + }, + { + "epoch": 1.1651469961653174, + "grad_norm": 22.125, + "learning_rate": 9.361438771960909e-07, + "loss": 0.8938, + "step": 13673 + }, + { + "epoch": 1.1652322113336173, + "grad_norm": 20.5, + "learning_rate": 9.359812274892004e-07, + "loss": 1.0476, + "step": 13674 + }, + { + "epoch": 1.1653174265019173, + "grad_norm": 14.0, + "learning_rate": 9.35818583456945e-07, + "loss": 0.3981, + "step": 13675 + }, + { + "epoch": 1.1654026416702172, + "grad_norm": 16.0, + "learning_rate": 9.356559451022637e-07, + "loss": 0.6305, + "step": 13676 + }, + { + "epoch": 1.1654878568385172, + "grad_norm": 15.25, + "learning_rate": 9.354933124280957e-07, + "loss": 0.7307, + "step": 13677 + }, + { + "epoch": 1.1655730720068171, + "grad_norm": 12.8125, + "learning_rate": 9.353306854373794e-07, + "loss": 0.3523, + "step": 13678 + }, + { + "epoch": 1.165658287175117, + "grad_norm": 14.25, + "learning_rate": 9.351680641330546e-07, + "loss": 0.4692, + "step": 13679 + }, + { + "epoch": 1.165743502343417, + "grad_norm": 16.5, + "learning_rate": 9.35005448518059e-07, + "loss": 0.7912, + "step": 13680 + }, + { + "epoch": 1.165828717511717, + "grad_norm": 14.125, + "learning_rate": 9.34842838595332e-07, + "loss": 0.4384, + "step": 13681 + }, + { + "epoch": 1.165913932680017, + "grad_norm": 15.0, + "learning_rate": 9.346802343678121e-07, + "loss": 0.7061, + "step": 13682 + }, + { + "epoch": 1.165999147848317, + "grad_norm": 16.875, + "learning_rate": 9.345176358384373e-07, + "loss": 1.0105, + "step": 13683 + }, + { + "epoch": 1.1660843630166169, + "grad_norm": 11.4375, + "learning_rate": 9.343550430101455e-07, + "loss": 0.2593, + "step": 13684 + }, + { + "epoch": 1.1661695781849168, + "grad_norm": 13.1875, + "learning_rate": 9.341924558858759e-07, + "loss": 0.4889, + "step": 13685 + }, + { + "epoch": 1.1662547933532168, + "grad_norm": 17.875, + "learning_rate": 9.34029874468566e-07, + "loss": 0.6194, + "step": 13686 + }, + { + "epoch": 1.1663400085215168, + "grad_norm": 13.5, + "learning_rate": 9.338672987611542e-07, + "loss": 0.6946, + "step": 13687 + }, + { + "epoch": 1.1664252236898167, + "grad_norm": 15.6875, + "learning_rate": 9.337047287665779e-07, + "loss": 0.592, + "step": 13688 + }, + { + "epoch": 1.1665104388581167, + "grad_norm": 15.6875, + "learning_rate": 9.335421644877757e-07, + "loss": 0.6409, + "step": 13689 + }, + { + "epoch": 1.1665956540264166, + "grad_norm": 11.0625, + "learning_rate": 9.333796059276847e-07, + "loss": 0.3282, + "step": 13690 + }, + { + "epoch": 1.1666808691947166, + "grad_norm": 15.5, + "learning_rate": 9.332170530892423e-07, + "loss": 0.6501, + "step": 13691 + }, + { + "epoch": 1.1667660843630165, + "grad_norm": 25.375, + "learning_rate": 9.330545059753866e-07, + "loss": 0.5105, + "step": 13692 + }, + { + "epoch": 1.1668512995313165, + "grad_norm": 11.0625, + "learning_rate": 9.328919645890544e-07, + "loss": 0.7121, + "step": 13693 + }, + { + "epoch": 1.1669365146996165, + "grad_norm": 22.125, + "learning_rate": 9.327294289331835e-07, + "loss": 0.8792, + "step": 13694 + }, + { + "epoch": 1.1670217298679164, + "grad_norm": 17.25, + "learning_rate": 9.325668990107106e-07, + "loss": 0.9586, + "step": 13695 + }, + { + "epoch": 1.1671069450362164, + "grad_norm": 14.875, + "learning_rate": 9.324043748245734e-07, + "loss": 0.5544, + "step": 13696 + }, + { + "epoch": 1.1671921602045163, + "grad_norm": 15.375, + "learning_rate": 9.32241856377708e-07, + "loss": 0.5766, + "step": 13697 + }, + { + "epoch": 1.1672773753728163, + "grad_norm": 14.8125, + "learning_rate": 9.32079343673053e-07, + "loss": 0.4521, + "step": 13698 + }, + { + "epoch": 1.1673625905411162, + "grad_norm": 13.5625, + "learning_rate": 9.319168367135428e-07, + "loss": 0.5398, + "step": 13699 + }, + { + "epoch": 1.1674478057094162, + "grad_norm": 14.5, + "learning_rate": 9.317543355021155e-07, + "loss": 0.6367, + "step": 13700 + }, + { + "epoch": 1.1675330208777162, + "grad_norm": 14.75, + "learning_rate": 9.315918400417073e-07, + "loss": 0.5078, + "step": 13701 + }, + { + "epoch": 1.1676182360460161, + "grad_norm": 18.0, + "learning_rate": 9.314293503352547e-07, + "loss": 0.7722, + "step": 13702 + }, + { + "epoch": 1.167703451214316, + "grad_norm": 11.1875, + "learning_rate": 9.312668663856941e-07, + "loss": 0.3216, + "step": 13703 + }, + { + "epoch": 1.167788666382616, + "grad_norm": 18.5, + "learning_rate": 9.311043881959619e-07, + "loss": 0.7287, + "step": 13704 + }, + { + "epoch": 1.167873881550916, + "grad_norm": 18.625, + "learning_rate": 9.309419157689939e-07, + "loss": 0.8367, + "step": 13705 + }, + { + "epoch": 1.167959096719216, + "grad_norm": 12.625, + "learning_rate": 9.307794491077269e-07, + "loss": 0.4761, + "step": 13706 + }, + { + "epoch": 1.168044311887516, + "grad_norm": 25.875, + "learning_rate": 9.306169882150959e-07, + "loss": 0.9498, + "step": 13707 + }, + { + "epoch": 1.1681295270558159, + "grad_norm": 15.125, + "learning_rate": 9.304545330940368e-07, + "loss": 0.7091, + "step": 13708 + }, + { + "epoch": 1.1682147422241158, + "grad_norm": 14.6875, + "learning_rate": 9.302920837474857e-07, + "loss": 0.7179, + "step": 13709 + }, + { + "epoch": 1.1682999573924158, + "grad_norm": 15.5625, + "learning_rate": 9.301296401783781e-07, + "loss": 0.7572, + "step": 13710 + }, + { + "epoch": 1.1683851725607157, + "grad_norm": 17.375, + "learning_rate": 9.299672023896497e-07, + "loss": 0.7923, + "step": 13711 + }, + { + "epoch": 1.1684703877290157, + "grad_norm": 12.1875, + "learning_rate": 9.298047703842356e-07, + "loss": 0.3043, + "step": 13712 + }, + { + "epoch": 1.1685556028973156, + "grad_norm": 14.9375, + "learning_rate": 9.296423441650715e-07, + "loss": 0.643, + "step": 13713 + }, + { + "epoch": 1.1686408180656156, + "grad_norm": 13.625, + "learning_rate": 9.294799237350926e-07, + "loss": 0.4616, + "step": 13714 + }, + { + "epoch": 1.1687260332339156, + "grad_norm": 13.0625, + "learning_rate": 9.293175090972337e-07, + "loss": 0.3855, + "step": 13715 + }, + { + "epoch": 1.1688112484022155, + "grad_norm": 16.25, + "learning_rate": 9.291551002544295e-07, + "loss": 0.6121, + "step": 13716 + }, + { + "epoch": 1.1688964635705155, + "grad_norm": 20.125, + "learning_rate": 9.289926972096156e-07, + "loss": 0.6761, + "step": 13717 + }, + { + "epoch": 1.1689816787388154, + "grad_norm": 13.125, + "learning_rate": 9.288302999657262e-07, + "loss": 0.5429, + "step": 13718 + }, + { + "epoch": 1.1690668939071154, + "grad_norm": 21.0, + "learning_rate": 9.286679085256963e-07, + "loss": 0.6478, + "step": 13719 + }, + { + "epoch": 1.1691521090754153, + "grad_norm": 16.5, + "learning_rate": 9.285055228924604e-07, + "loss": 0.639, + "step": 13720 + }, + { + "epoch": 1.1692373242437153, + "grad_norm": 18.875, + "learning_rate": 9.283431430689532e-07, + "loss": 0.4565, + "step": 13721 + }, + { + "epoch": 1.1693225394120152, + "grad_norm": 16.25, + "learning_rate": 9.281807690581091e-07, + "loss": 0.5764, + "step": 13722 + }, + { + "epoch": 1.1694077545803152, + "grad_norm": 14.6875, + "learning_rate": 9.280184008628614e-07, + "loss": 0.5861, + "step": 13723 + }, + { + "epoch": 1.1694929697486152, + "grad_norm": 18.75, + "learning_rate": 9.278560384861454e-07, + "loss": 1.1764, + "step": 13724 + }, + { + "epoch": 1.1695781849169151, + "grad_norm": 17.0, + "learning_rate": 9.276936819308946e-07, + "loss": 0.4114, + "step": 13725 + }, + { + "epoch": 1.169663400085215, + "grad_norm": 16.25, + "learning_rate": 9.275313312000431e-07, + "loss": 0.5187, + "step": 13726 + }, + { + "epoch": 1.169748615253515, + "grad_norm": 15.0, + "learning_rate": 9.273689862965246e-07, + "loss": 0.6566, + "step": 13727 + }, + { + "epoch": 1.169833830421815, + "grad_norm": 17.0, + "learning_rate": 9.272066472232733e-07, + "loss": 0.5981, + "step": 13728 + }, + { + "epoch": 1.169919045590115, + "grad_norm": 14.125, + "learning_rate": 9.270443139832222e-07, + "loss": 0.5107, + "step": 13729 + }, + { + "epoch": 1.170004260758415, + "grad_norm": 16.375, + "learning_rate": 9.268819865793055e-07, + "loss": 0.484, + "step": 13730 + }, + { + "epoch": 1.1700894759267149, + "grad_norm": 20.875, + "learning_rate": 9.267196650144558e-07, + "loss": 0.5193, + "step": 13731 + }, + { + "epoch": 1.1701746910950148, + "grad_norm": 11.5625, + "learning_rate": 9.265573492916069e-07, + "loss": 0.4721, + "step": 13732 + }, + { + "epoch": 1.1702599062633148, + "grad_norm": 13.9375, + "learning_rate": 9.263950394136918e-07, + "loss": 0.5427, + "step": 13733 + }, + { + "epoch": 1.1703451214316147, + "grad_norm": 14.8125, + "learning_rate": 9.26232735383644e-07, + "loss": 0.4442, + "step": 13734 + }, + { + "epoch": 1.1704303365999147, + "grad_norm": 19.875, + "learning_rate": 9.260704372043961e-07, + "loss": 0.6256, + "step": 13735 + }, + { + "epoch": 1.1705155517682146, + "grad_norm": 26.875, + "learning_rate": 9.259081448788812e-07, + "loss": 0.7938, + "step": 13736 + }, + { + "epoch": 1.1706007669365146, + "grad_norm": 13.5625, + "learning_rate": 9.257458584100318e-07, + "loss": 0.4748, + "step": 13737 + }, + { + "epoch": 1.1706859821048146, + "grad_norm": 18.375, + "learning_rate": 9.255835778007813e-07, + "loss": 0.4346, + "step": 13738 + }, + { + "epoch": 1.1707711972731145, + "grad_norm": 13.6875, + "learning_rate": 9.254213030540615e-07, + "loss": 0.5094, + "step": 13739 + }, + { + "epoch": 1.1708564124414145, + "grad_norm": 17.625, + "learning_rate": 9.252590341728052e-07, + "loss": 0.7029, + "step": 13740 + }, + { + "epoch": 1.1709416276097144, + "grad_norm": 17.0, + "learning_rate": 9.250967711599446e-07, + "loss": 0.8056, + "step": 13741 + }, + { + "epoch": 1.1710268427780144, + "grad_norm": 18.5, + "learning_rate": 9.249345140184118e-07, + "loss": 0.8193, + "step": 13742 + }, + { + "epoch": 1.1711120579463143, + "grad_norm": 15.875, + "learning_rate": 9.247722627511396e-07, + "loss": 0.7369, + "step": 13743 + }, + { + "epoch": 1.1711972731146143, + "grad_norm": 16.125, + "learning_rate": 9.246100173610598e-07, + "loss": 0.3672, + "step": 13744 + }, + { + "epoch": 1.1712824882829143, + "grad_norm": 17.875, + "learning_rate": 9.244477778511038e-07, + "loss": 0.2977, + "step": 13745 + }, + { + "epoch": 1.1713677034512142, + "grad_norm": 15.1875, + "learning_rate": 9.24285544224204e-07, + "loss": 0.9983, + "step": 13746 + }, + { + "epoch": 1.1714529186195142, + "grad_norm": 17.0, + "learning_rate": 9.241233164832922e-07, + "loss": 0.8154, + "step": 13747 + }, + { + "epoch": 1.1715381337878141, + "grad_norm": 15.8125, + "learning_rate": 9.239610946312993e-07, + "loss": 0.4281, + "step": 13748 + }, + { + "epoch": 1.171623348956114, + "grad_norm": 12.4375, + "learning_rate": 9.237988786711577e-07, + "loss": 0.3146, + "step": 13749 + }, + { + "epoch": 1.171708564124414, + "grad_norm": 16.25, + "learning_rate": 9.236366686057979e-07, + "loss": 0.7633, + "step": 13750 + }, + { + "epoch": 1.171793779292714, + "grad_norm": 13.75, + "learning_rate": 9.23474464438152e-07, + "loss": 0.4556, + "step": 13751 + }, + { + "epoch": 1.171878994461014, + "grad_norm": 19.5, + "learning_rate": 9.233122661711507e-07, + "loss": 0.9847, + "step": 13752 + }, + { + "epoch": 1.171964209629314, + "grad_norm": 16.5, + "learning_rate": 9.231500738077254e-07, + "loss": 0.484, + "step": 13753 + }, + { + "epoch": 1.1720494247976139, + "grad_norm": 19.0, + "learning_rate": 9.229878873508066e-07, + "loss": 0.804, + "step": 13754 + }, + { + "epoch": 1.1721346399659138, + "grad_norm": 11.0625, + "learning_rate": 9.228257068033262e-07, + "loss": 0.351, + "step": 13755 + }, + { + "epoch": 1.1722198551342138, + "grad_norm": 12.4375, + "learning_rate": 9.22663532168214e-07, + "loss": 0.3316, + "step": 13756 + }, + { + "epoch": 1.1723050703025137, + "grad_norm": 13.5625, + "learning_rate": 9.225013634484006e-07, + "loss": 0.505, + "step": 13757 + }, + { + "epoch": 1.1723902854708137, + "grad_norm": 20.75, + "learning_rate": 9.223392006468174e-07, + "loss": 0.9544, + "step": 13758 + }, + { + "epoch": 1.1724755006391137, + "grad_norm": 13.375, + "learning_rate": 9.22177043766394e-07, + "loss": 0.4804, + "step": 13759 + }, + { + "epoch": 1.1725607158074136, + "grad_norm": 10.6875, + "learning_rate": 9.220148928100609e-07, + "loss": 0.3642, + "step": 13760 + }, + { + "epoch": 1.1726459309757136, + "grad_norm": 16.125, + "learning_rate": 9.218527477807488e-07, + "loss": 1.0494, + "step": 13761 + }, + { + "epoch": 1.1727311461440135, + "grad_norm": 9.6875, + "learning_rate": 9.216906086813873e-07, + "loss": 0.2008, + "step": 13762 + }, + { + "epoch": 1.1728163613123135, + "grad_norm": 15.375, + "learning_rate": 9.215284755149071e-07, + "loss": 0.5116, + "step": 13763 + }, + { + "epoch": 1.1729015764806134, + "grad_norm": 13.625, + "learning_rate": 9.213663482842375e-07, + "loss": 0.4188, + "step": 13764 + }, + { + "epoch": 1.1729867916489134, + "grad_norm": 18.75, + "learning_rate": 9.21204226992308e-07, + "loss": 0.6175, + "step": 13765 + }, + { + "epoch": 1.1730720068172134, + "grad_norm": 22.375, + "learning_rate": 9.210421116420493e-07, + "loss": 0.7288, + "step": 13766 + }, + { + "epoch": 1.1731572219855133, + "grad_norm": 12.5625, + "learning_rate": 9.208800022363899e-07, + "loss": 0.53, + "step": 13767 + }, + { + "epoch": 1.1732424371538133, + "grad_norm": 12.875, + "learning_rate": 9.207178987782602e-07, + "loss": 0.4425, + "step": 13768 + }, + { + "epoch": 1.1733276523221132, + "grad_norm": 15.5625, + "learning_rate": 9.20555801270589e-07, + "loss": 0.6485, + "step": 13769 + }, + { + "epoch": 1.1734128674904132, + "grad_norm": 11.9375, + "learning_rate": 9.203937097163058e-07, + "loss": 0.4844, + "step": 13770 + }, + { + "epoch": 1.1734980826587131, + "grad_norm": 14.5, + "learning_rate": 9.202316241183401e-07, + "loss": 0.4934, + "step": 13771 + }, + { + "epoch": 1.173583297827013, + "grad_norm": 14.5625, + "learning_rate": 9.200695444796201e-07, + "loss": 0.6111, + "step": 13772 + }, + { + "epoch": 1.173668512995313, + "grad_norm": 11.5, + "learning_rate": 9.199074708030753e-07, + "loss": 0.406, + "step": 13773 + }, + { + "epoch": 1.173753728163613, + "grad_norm": 34.5, + "learning_rate": 9.197454030916345e-07, + "loss": 0.8554, + "step": 13774 + }, + { + "epoch": 1.173838943331913, + "grad_norm": 18.875, + "learning_rate": 9.195833413482261e-07, + "loss": 0.5937, + "step": 13775 + }, + { + "epoch": 1.173924158500213, + "grad_norm": 15.3125, + "learning_rate": 9.194212855757793e-07, + "loss": 0.4867, + "step": 13776 + }, + { + "epoch": 1.1740093736685129, + "grad_norm": 23.0, + "learning_rate": 9.19259235777222e-07, + "loss": 0.93, + "step": 13777 + }, + { + "epoch": 1.1740945888368128, + "grad_norm": 17.125, + "learning_rate": 9.190971919554831e-07, + "loss": 0.5663, + "step": 13778 + }, + { + "epoch": 1.1741798040051128, + "grad_norm": 19.875, + "learning_rate": 9.189351541134911e-07, + "loss": 0.9228, + "step": 13779 + }, + { + "epoch": 1.1742650191734127, + "grad_norm": 15.875, + "learning_rate": 9.187731222541732e-07, + "loss": 0.918, + "step": 13780 + }, + { + "epoch": 1.1743502343417127, + "grad_norm": 13.4375, + "learning_rate": 9.186110963804582e-07, + "loss": 0.646, + "step": 13781 + }, + { + "epoch": 1.1744354495100127, + "grad_norm": 12.6875, + "learning_rate": 9.184490764952737e-07, + "loss": 0.4019, + "step": 13782 + }, + { + "epoch": 1.1745206646783126, + "grad_norm": 11.4375, + "learning_rate": 9.18287062601548e-07, + "loss": 0.3562, + "step": 13783 + }, + { + "epoch": 1.1746058798466126, + "grad_norm": 22.375, + "learning_rate": 9.181250547022086e-07, + "loss": 0.6302, + "step": 13784 + }, + { + "epoch": 1.1746910950149125, + "grad_norm": 16.875, + "learning_rate": 9.179630528001833e-07, + "loss": 0.485, + "step": 13785 + }, + { + "epoch": 1.1747763101832125, + "grad_norm": 22.5, + "learning_rate": 9.178010568983993e-07, + "loss": 0.9022, + "step": 13786 + }, + { + "epoch": 1.1748615253515124, + "grad_norm": 13.5625, + "learning_rate": 9.176390669997848e-07, + "loss": 0.4142, + "step": 13787 + }, + { + "epoch": 1.1749467405198124, + "grad_norm": 19.875, + "learning_rate": 9.174770831072663e-07, + "loss": 0.6718, + "step": 13788 + }, + { + "epoch": 1.1750319556881124, + "grad_norm": 27.125, + "learning_rate": 9.173151052237712e-07, + "loss": 0.6082, + "step": 13789 + }, + { + "epoch": 1.1751171708564123, + "grad_norm": 19.75, + "learning_rate": 9.171531333522266e-07, + "loss": 0.7176, + "step": 13790 + }, + { + "epoch": 1.1752023860247123, + "grad_norm": 21.5, + "learning_rate": 9.169911674955596e-07, + "loss": 0.4735, + "step": 13791 + }, + { + "epoch": 1.1752876011930122, + "grad_norm": 16.25, + "learning_rate": 9.16829207656697e-07, + "loss": 0.7935, + "step": 13792 + }, + { + "epoch": 1.1753728163613122, + "grad_norm": 13.625, + "learning_rate": 9.166672538385659e-07, + "loss": 0.6668, + "step": 13793 + }, + { + "epoch": 1.1754580315296121, + "grad_norm": 15.4375, + "learning_rate": 9.165053060440923e-07, + "loss": 0.6673, + "step": 13794 + }, + { + "epoch": 1.1755432466979123, + "grad_norm": 21.75, + "learning_rate": 9.163433642762038e-07, + "loss": 0.7526, + "step": 13795 + }, + { + "epoch": 1.1756284618662123, + "grad_norm": 14.0625, + "learning_rate": 9.16181428537826e-07, + "loss": 0.5234, + "step": 13796 + }, + { + "epoch": 1.1757136770345122, + "grad_norm": 20.625, + "learning_rate": 9.160194988318851e-07, + "loss": 0.3877, + "step": 13797 + }, + { + "epoch": 1.1757988922028122, + "grad_norm": 14.1875, + "learning_rate": 9.15857575161308e-07, + "loss": 0.6685, + "step": 13798 + }, + { + "epoch": 1.1758841073711122, + "grad_norm": 14.75, + "learning_rate": 9.156956575290205e-07, + "loss": 0.4307, + "step": 13799 + }, + { + "epoch": 1.175969322539412, + "grad_norm": 18.875, + "learning_rate": 9.155337459379485e-07, + "loss": 0.8261, + "step": 13800 + }, + { + "epoch": 1.176054537707712, + "grad_norm": 12.25, + "learning_rate": 9.15371840391018e-07, + "loss": 0.4362, + "step": 13801 + }, + { + "epoch": 1.176139752876012, + "grad_norm": 16.625, + "learning_rate": 9.152099408911551e-07, + "loss": 0.7719, + "step": 13802 + }, + { + "epoch": 1.176224968044312, + "grad_norm": 12.875, + "learning_rate": 9.150480474412854e-07, + "loss": 0.4996, + "step": 13803 + }, + { + "epoch": 1.176310183212612, + "grad_norm": 17.375, + "learning_rate": 9.148861600443341e-07, + "loss": 0.9614, + "step": 13804 + }, + { + "epoch": 1.176395398380912, + "grad_norm": 20.25, + "learning_rate": 9.147242787032266e-07, + "loss": 0.5946, + "step": 13805 + }, + { + "epoch": 1.1764806135492119, + "grad_norm": 16.25, + "learning_rate": 9.145624034208889e-07, + "loss": 0.7093, + "step": 13806 + }, + { + "epoch": 1.1765658287175118, + "grad_norm": 14.125, + "learning_rate": 9.144005342002455e-07, + "loss": 0.3679, + "step": 13807 + }, + { + "epoch": 1.1766510438858118, + "grad_norm": 14.1875, + "learning_rate": 9.142386710442222e-07, + "loss": 0.6171, + "step": 13808 + }, + { + "epoch": 1.1767362590541117, + "grad_norm": 17.375, + "learning_rate": 9.140768139557435e-07, + "loss": 0.5618, + "step": 13809 + }, + { + "epoch": 1.1768214742224117, + "grad_norm": 18.0, + "learning_rate": 9.139149629377347e-07, + "loss": 0.7244, + "step": 13810 + }, + { + "epoch": 1.1769066893907116, + "grad_norm": 13.0625, + "learning_rate": 9.137531179931208e-07, + "loss": 0.6206, + "step": 13811 + }, + { + "epoch": 1.1769919045590116, + "grad_norm": 14.75, + "learning_rate": 9.135912791248258e-07, + "loss": 0.4439, + "step": 13812 + }, + { + "epoch": 1.1770771197273115, + "grad_norm": 15.875, + "learning_rate": 9.134294463357749e-07, + "loss": 0.4244, + "step": 13813 + }, + { + "epoch": 1.1771623348956115, + "grad_norm": 13.125, + "learning_rate": 9.13267619628892e-07, + "loss": 0.4025, + "step": 13814 + }, + { + "epoch": 1.1772475500639115, + "grad_norm": 12.4375, + "learning_rate": 9.131057990071021e-07, + "loss": 0.3839, + "step": 13815 + }, + { + "epoch": 1.1773327652322114, + "grad_norm": 15.5625, + "learning_rate": 9.12943984473329e-07, + "loss": 0.6203, + "step": 13816 + }, + { + "epoch": 1.1774179804005114, + "grad_norm": 14.5, + "learning_rate": 9.127821760304972e-07, + "loss": 0.4257, + "step": 13817 + }, + { + "epoch": 1.1775031955688113, + "grad_norm": 11.5, + "learning_rate": 9.126203736815303e-07, + "loss": 0.4317, + "step": 13818 + }, + { + "epoch": 1.1775884107371113, + "grad_norm": 17.625, + "learning_rate": 9.124585774293529e-07, + "loss": 0.5322, + "step": 13819 + }, + { + "epoch": 1.1776736259054112, + "grad_norm": 40.0, + "learning_rate": 9.122967872768887e-07, + "loss": 0.4118, + "step": 13820 + }, + { + "epoch": 1.1777588410737112, + "grad_norm": 15.625, + "learning_rate": 9.121350032270607e-07, + "loss": 0.7193, + "step": 13821 + }, + { + "epoch": 1.1778440562420112, + "grad_norm": 12.8125, + "learning_rate": 9.119732252827928e-07, + "loss": 0.4431, + "step": 13822 + }, + { + "epoch": 1.1779292714103111, + "grad_norm": 18.625, + "learning_rate": 9.118114534470089e-07, + "loss": 0.5925, + "step": 13823 + }, + { + "epoch": 1.178014486578611, + "grad_norm": 14.875, + "learning_rate": 9.116496877226319e-07, + "loss": 0.7285, + "step": 13824 + }, + { + "epoch": 1.178099701746911, + "grad_norm": 17.25, + "learning_rate": 9.114879281125854e-07, + "loss": 0.4615, + "step": 13825 + }, + { + "epoch": 1.178184916915211, + "grad_norm": 11.8125, + "learning_rate": 9.113261746197923e-07, + "loss": 0.3311, + "step": 13826 + }, + { + "epoch": 1.178270132083511, + "grad_norm": 14.375, + "learning_rate": 9.111644272471759e-07, + "loss": 0.5541, + "step": 13827 + }, + { + "epoch": 1.178355347251811, + "grad_norm": 20.875, + "learning_rate": 9.110026859976595e-07, + "loss": 0.8298, + "step": 13828 + }, + { + "epoch": 1.1784405624201109, + "grad_norm": 16.375, + "learning_rate": 9.108409508741647e-07, + "loss": 0.6907, + "step": 13829 + }, + { + "epoch": 1.1785257775884108, + "grad_norm": 13.8125, + "learning_rate": 9.106792218796154e-07, + "loss": 0.5155, + "step": 13830 + }, + { + "epoch": 1.1786109927567108, + "grad_norm": 18.25, + "learning_rate": 9.105174990169334e-07, + "loss": 0.56, + "step": 13831 + }, + { + "epoch": 1.1786962079250107, + "grad_norm": 15.75, + "learning_rate": 9.103557822890418e-07, + "loss": 0.641, + "step": 13832 + }, + { + "epoch": 1.1787814230933107, + "grad_norm": 12.75, + "learning_rate": 9.101940716988626e-07, + "loss": 0.5763, + "step": 13833 + }, + { + "epoch": 1.1788666382616106, + "grad_norm": 12.0625, + "learning_rate": 9.100323672493186e-07, + "loss": 0.2797, + "step": 13834 + }, + { + "epoch": 1.1789518534299106, + "grad_norm": 17.0, + "learning_rate": 9.09870668943331e-07, + "loss": 0.6225, + "step": 13835 + }, + { + "epoch": 1.1790370685982106, + "grad_norm": 39.5, + "learning_rate": 9.097089767838235e-07, + "loss": 1.3705, + "step": 13836 + }, + { + "epoch": 1.1791222837665105, + "grad_norm": 17.375, + "learning_rate": 9.095472907737162e-07, + "loss": 0.6613, + "step": 13837 + }, + { + "epoch": 1.1792074989348105, + "grad_norm": 15.1875, + "learning_rate": 9.093856109159319e-07, + "loss": 0.7647, + "step": 13838 + }, + { + "epoch": 1.1792927141031104, + "grad_norm": 16.0, + "learning_rate": 9.092239372133918e-07, + "loss": 0.5336, + "step": 13839 + }, + { + "epoch": 1.1793779292714104, + "grad_norm": 12.0625, + "learning_rate": 9.090622696690179e-07, + "loss": 0.4046, + "step": 13840 + }, + { + "epoch": 1.1794631444397103, + "grad_norm": 37.75, + "learning_rate": 9.089006082857317e-07, + "loss": 1.3158, + "step": 13841 + }, + { + "epoch": 1.1795483596080103, + "grad_norm": 14.75, + "learning_rate": 9.087389530664547e-07, + "loss": 1.0166, + "step": 13842 + }, + { + "epoch": 1.1796335747763103, + "grad_norm": 17.625, + "learning_rate": 9.085773040141077e-07, + "loss": 0.5921, + "step": 13843 + }, + { + "epoch": 1.1797187899446102, + "grad_norm": 13.8125, + "learning_rate": 9.084156611316126e-07, + "loss": 0.4077, + "step": 13844 + }, + { + "epoch": 1.1798040051129102, + "grad_norm": 18.75, + "learning_rate": 9.082540244218899e-07, + "loss": 0.734, + "step": 13845 + }, + { + "epoch": 1.1798892202812101, + "grad_norm": 23.75, + "learning_rate": 9.080923938878602e-07, + "loss": 1.0257, + "step": 13846 + }, + { + "epoch": 1.17997443544951, + "grad_norm": 16.0, + "learning_rate": 9.079307695324451e-07, + "loss": 0.7709, + "step": 13847 + }, + { + "epoch": 1.18005965061781, + "grad_norm": 16.875, + "learning_rate": 9.077691513585645e-07, + "loss": 0.632, + "step": 13848 + }, + { + "epoch": 1.18014486578611, + "grad_norm": 17.875, + "learning_rate": 9.0760753936914e-07, + "loss": 0.814, + "step": 13849 + }, + { + "epoch": 1.18023008095441, + "grad_norm": 13.75, + "learning_rate": 9.074459335670913e-07, + "loss": 0.3345, + "step": 13850 + }, + { + "epoch": 1.18031529612271, + "grad_norm": 31.5, + "learning_rate": 9.072843339553391e-07, + "loss": 0.5699, + "step": 13851 + }, + { + "epoch": 1.1804005112910099, + "grad_norm": 15.125, + "learning_rate": 9.071227405368041e-07, + "loss": 0.5589, + "step": 13852 + }, + { + "epoch": 1.1804857264593098, + "grad_norm": 29.875, + "learning_rate": 9.069611533144057e-07, + "loss": 0.3087, + "step": 13853 + }, + { + "epoch": 1.1805709416276098, + "grad_norm": 16.125, + "learning_rate": 9.067995722910639e-07, + "loss": 0.5764, + "step": 13854 + }, + { + "epoch": 1.1806561567959097, + "grad_norm": 18.75, + "learning_rate": 9.066379974696989e-07, + "loss": 0.6613, + "step": 13855 + }, + { + "epoch": 1.1807413719642097, + "grad_norm": 16.0, + "learning_rate": 9.064764288532308e-07, + "loss": 0.7291, + "step": 13856 + }, + { + "epoch": 1.1808265871325097, + "grad_norm": 19.25, + "learning_rate": 9.063148664445789e-07, + "loss": 0.6896, + "step": 13857 + }, + { + "epoch": 1.1809118023008096, + "grad_norm": 13.4375, + "learning_rate": 9.06153310246663e-07, + "loss": 0.4896, + "step": 13858 + }, + { + "epoch": 1.1809970174691096, + "grad_norm": 13.625, + "learning_rate": 9.059917602624027e-07, + "loss": 0.525, + "step": 13859 + }, + { + "epoch": 1.1810822326374095, + "grad_norm": 19.25, + "learning_rate": 9.058302164947174e-07, + "loss": 0.7842, + "step": 13860 + }, + { + "epoch": 1.1811674478057095, + "grad_norm": 19.125, + "learning_rate": 9.056686789465258e-07, + "loss": 0.3967, + "step": 13861 + }, + { + "epoch": 1.1812526629740094, + "grad_norm": 15.0, + "learning_rate": 9.055071476207477e-07, + "loss": 0.7483, + "step": 13862 + }, + { + "epoch": 1.1813378781423094, + "grad_norm": 15.4375, + "learning_rate": 9.053456225203016e-07, + "loss": 0.6002, + "step": 13863 + }, + { + "epoch": 1.1814230933106094, + "grad_norm": 17.5, + "learning_rate": 9.051841036481068e-07, + "loss": 0.7455, + "step": 13864 + }, + { + "epoch": 1.1815083084789093, + "grad_norm": 13.125, + "learning_rate": 9.050225910070817e-07, + "loss": 0.4873, + "step": 13865 + }, + { + "epoch": 1.1815935236472093, + "grad_norm": 15.5625, + "learning_rate": 9.048610846001457e-07, + "loss": 0.5632, + "step": 13866 + }, + { + "epoch": 1.1816787388155092, + "grad_norm": 14.5, + "learning_rate": 9.04699584430217e-07, + "loss": 0.6127, + "step": 13867 + }, + { + "epoch": 1.1817639539838092, + "grad_norm": 15.375, + "learning_rate": 9.045380905002141e-07, + "loss": 0.7485, + "step": 13868 + }, + { + "epoch": 1.1818491691521091, + "grad_norm": 14.4375, + "learning_rate": 9.043766028130549e-07, + "loss": 0.8294, + "step": 13869 + }, + { + "epoch": 1.181934384320409, + "grad_norm": 17.5, + "learning_rate": 9.042151213716582e-07, + "loss": 0.548, + "step": 13870 + }, + { + "epoch": 1.182019599488709, + "grad_norm": 17.625, + "learning_rate": 9.040536461789418e-07, + "loss": 0.7102, + "step": 13871 + }, + { + "epoch": 1.182104814657009, + "grad_norm": 17.0, + "learning_rate": 9.038921772378242e-07, + "loss": 0.4423, + "step": 13872 + }, + { + "epoch": 1.182190029825309, + "grad_norm": 17.125, + "learning_rate": 9.037307145512225e-07, + "loss": 0.8903, + "step": 13873 + }, + { + "epoch": 1.182275244993609, + "grad_norm": 17.5, + "learning_rate": 9.035692581220555e-07, + "loss": 0.5376, + "step": 13874 + }, + { + "epoch": 1.1823604601619089, + "grad_norm": 12.4375, + "learning_rate": 9.0340780795324e-07, + "loss": 0.4616, + "step": 13875 + }, + { + "epoch": 1.1824456753302088, + "grad_norm": 15.375, + "learning_rate": 9.032463640476947e-07, + "loss": 0.7267, + "step": 13876 + }, + { + "epoch": 1.1825308904985088, + "grad_norm": 15.6875, + "learning_rate": 9.030849264083357e-07, + "loss": 0.6071, + "step": 13877 + }, + { + "epoch": 1.1826161056668087, + "grad_norm": 14.4375, + "learning_rate": 9.029234950380808e-07, + "loss": 0.4414, + "step": 13878 + }, + { + "epoch": 1.1827013208351087, + "grad_norm": 19.0, + "learning_rate": 9.027620699398474e-07, + "loss": 0.7998, + "step": 13879 + }, + { + "epoch": 1.1827865360034087, + "grad_norm": 13.6875, + "learning_rate": 9.026006511165527e-07, + "loss": 0.4086, + "step": 13880 + }, + { + "epoch": 1.1828717511717086, + "grad_norm": 13.3125, + "learning_rate": 9.024392385711135e-07, + "loss": 0.5875, + "step": 13881 + }, + { + "epoch": 1.1829569663400086, + "grad_norm": 19.0, + "learning_rate": 9.022778323064466e-07, + "loss": 0.8234, + "step": 13882 + }, + { + "epoch": 1.1830421815083085, + "grad_norm": 15.5, + "learning_rate": 9.021164323254692e-07, + "loss": 0.5214, + "step": 13883 + }, + { + "epoch": 1.1831273966766085, + "grad_norm": 15.75, + "learning_rate": 9.019550386310978e-07, + "loss": 0.7611, + "step": 13884 + }, + { + "epoch": 1.1832126118449084, + "grad_norm": 19.75, + "learning_rate": 9.01793651226249e-07, + "loss": 0.7538, + "step": 13885 + }, + { + "epoch": 1.1832978270132084, + "grad_norm": 13.6875, + "learning_rate": 9.016322701138386e-07, + "loss": 0.6235, + "step": 13886 + }, + { + "epoch": 1.1833830421815084, + "grad_norm": 13.9375, + "learning_rate": 9.014708952967837e-07, + "loss": 0.5189, + "step": 13887 + }, + { + "epoch": 1.1834682573498083, + "grad_norm": 18.0, + "learning_rate": 9.013095267779998e-07, + "loss": 1.1996, + "step": 13888 + }, + { + "epoch": 1.1835534725181083, + "grad_norm": 13.4375, + "learning_rate": 9.011481645604036e-07, + "loss": 0.8322, + "step": 13889 + }, + { + "epoch": 1.1836386876864082, + "grad_norm": 15.4375, + "learning_rate": 9.009868086469108e-07, + "loss": 0.5539, + "step": 13890 + }, + { + "epoch": 1.1837239028547082, + "grad_norm": 14.375, + "learning_rate": 9.008254590404375e-07, + "loss": 0.6563, + "step": 13891 + }, + { + "epoch": 1.1838091180230081, + "grad_norm": 21.125, + "learning_rate": 9.006641157438989e-07, + "loss": 0.6192, + "step": 13892 + }, + { + "epoch": 1.183894333191308, + "grad_norm": 16.625, + "learning_rate": 9.005027787602117e-07, + "loss": 0.585, + "step": 13893 + }, + { + "epoch": 1.183979548359608, + "grad_norm": 14.0, + "learning_rate": 9.003414480922904e-07, + "loss": 0.544, + "step": 13894 + }, + { + "epoch": 1.184064763527908, + "grad_norm": 17.125, + "learning_rate": 9.001801237430505e-07, + "loss": 0.7996, + "step": 13895 + }, + { + "epoch": 1.184149978696208, + "grad_norm": 17.875, + "learning_rate": 9.000188057154077e-07, + "loss": 0.7155, + "step": 13896 + }, + { + "epoch": 1.184235193864508, + "grad_norm": 10.1875, + "learning_rate": 8.99857494012277e-07, + "loss": 0.368, + "step": 13897 + }, + { + "epoch": 1.1843204090328079, + "grad_norm": 14.0, + "learning_rate": 8.996961886365735e-07, + "loss": 0.5922, + "step": 13898 + }, + { + "epoch": 1.1844056242011078, + "grad_norm": 20.0, + "learning_rate": 8.995348895912123e-07, + "loss": 0.6272, + "step": 13899 + }, + { + "epoch": 1.1844908393694078, + "grad_norm": 20.625, + "learning_rate": 8.993735968791076e-07, + "loss": 0.8241, + "step": 13900 + }, + { + "epoch": 1.1845760545377078, + "grad_norm": 20.0, + "learning_rate": 8.992123105031752e-07, + "loss": 0.796, + "step": 13901 + }, + { + "epoch": 1.1846612697060077, + "grad_norm": 14.8125, + "learning_rate": 8.990510304663288e-07, + "loss": 0.6819, + "step": 13902 + }, + { + "epoch": 1.1847464848743077, + "grad_norm": 17.375, + "learning_rate": 8.988897567714831e-07, + "loss": 0.6611, + "step": 13903 + }, + { + "epoch": 1.1848317000426076, + "grad_norm": 16.375, + "learning_rate": 8.987284894215528e-07, + "loss": 0.4065, + "step": 13904 + }, + { + "epoch": 1.1849169152109076, + "grad_norm": 39.5, + "learning_rate": 8.985672284194517e-07, + "loss": 0.6687, + "step": 13905 + }, + { + "epoch": 1.1850021303792075, + "grad_norm": 10.0, + "learning_rate": 8.984059737680941e-07, + "loss": 0.2108, + "step": 13906 + }, + { + "epoch": 1.1850873455475075, + "grad_norm": 15.4375, + "learning_rate": 8.982447254703941e-07, + "loss": 0.6192, + "step": 13907 + }, + { + "epoch": 1.1851725607158075, + "grad_norm": 16.125, + "learning_rate": 8.980834835292658e-07, + "loss": 0.7532, + "step": 13908 + }, + { + "epoch": 1.1852577758841074, + "grad_norm": 16.5, + "learning_rate": 8.979222479476232e-07, + "loss": 0.7998, + "step": 13909 + }, + { + "epoch": 1.1853429910524074, + "grad_norm": 19.25, + "learning_rate": 8.97761018728379e-07, + "loss": 0.9819, + "step": 13910 + }, + { + "epoch": 1.1854282062207073, + "grad_norm": 14.8125, + "learning_rate": 8.975997958744477e-07, + "loss": 0.6064, + "step": 13911 + }, + { + "epoch": 1.1855134213890073, + "grad_norm": 18.5, + "learning_rate": 8.974385793887422e-07, + "loss": 0.6795, + "step": 13912 + }, + { + "epoch": 1.1855986365573072, + "grad_norm": 14.0, + "learning_rate": 8.972773692741762e-07, + "loss": 0.5897, + "step": 13913 + }, + { + "epoch": 1.1856838517256072, + "grad_norm": 15.125, + "learning_rate": 8.971161655336631e-07, + "loss": 0.4291, + "step": 13914 + }, + { + "epoch": 1.1857690668939072, + "grad_norm": 14.0, + "learning_rate": 8.969549681701153e-07, + "loss": 0.6136, + "step": 13915 + }, + { + "epoch": 1.185854282062207, + "grad_norm": 13.8125, + "learning_rate": 8.967937771864466e-07, + "loss": 0.5899, + "step": 13916 + }, + { + "epoch": 1.185939497230507, + "grad_norm": 15.625, + "learning_rate": 8.966325925855698e-07, + "loss": 0.6131, + "step": 13917 + }, + { + "epoch": 1.186024712398807, + "grad_norm": 16.5, + "learning_rate": 8.964714143703968e-07, + "loss": 0.3838, + "step": 13918 + }, + { + "epoch": 1.186109927567107, + "grad_norm": 12.375, + "learning_rate": 8.963102425438411e-07, + "loss": 0.3785, + "step": 13919 + }, + { + "epoch": 1.186195142735407, + "grad_norm": 15.625, + "learning_rate": 8.961490771088146e-07, + "loss": 0.7403, + "step": 13920 + }, + { + "epoch": 1.186280357903707, + "grad_norm": 13.4375, + "learning_rate": 8.959879180682307e-07, + "loss": 0.3227, + "step": 13921 + }, + { + "epoch": 1.1863655730720069, + "grad_norm": 19.0, + "learning_rate": 8.958267654250005e-07, + "loss": 0.942, + "step": 13922 + }, + { + "epoch": 1.1864507882403068, + "grad_norm": 20.625, + "learning_rate": 8.956656191820373e-07, + "loss": 0.6239, + "step": 13923 + }, + { + "epoch": 1.1865360034086068, + "grad_norm": 10.375, + "learning_rate": 8.955044793422522e-07, + "loss": 0.2948, + "step": 13924 + }, + { + "epoch": 1.1866212185769067, + "grad_norm": 13.4375, + "learning_rate": 8.953433459085584e-07, + "loss": 0.5747, + "step": 13925 + }, + { + "epoch": 1.1867064337452067, + "grad_norm": 10.6875, + "learning_rate": 8.951822188838666e-07, + "loss": 0.235, + "step": 13926 + }, + { + "epoch": 1.1867916489135066, + "grad_norm": 16.125, + "learning_rate": 8.950210982710885e-07, + "loss": 0.7468, + "step": 13927 + }, + { + "epoch": 1.1868768640818066, + "grad_norm": 19.0, + "learning_rate": 8.948599840731364e-07, + "loss": 0.8483, + "step": 13928 + }, + { + "epoch": 1.1869620792501065, + "grad_norm": 17.375, + "learning_rate": 8.946988762929218e-07, + "loss": 0.7305, + "step": 13929 + }, + { + "epoch": 1.1870472944184065, + "grad_norm": 25.75, + "learning_rate": 8.945377749333553e-07, + "loss": 0.6006, + "step": 13930 + }, + { + "epoch": 1.1871325095867065, + "grad_norm": 17.75, + "learning_rate": 8.943766799973489e-07, + "loss": 0.5573, + "step": 13931 + }, + { + "epoch": 1.1872177247550064, + "grad_norm": 19.125, + "learning_rate": 8.942155914878132e-07, + "loss": 0.7629, + "step": 13932 + }, + { + "epoch": 1.1873029399233064, + "grad_norm": 21.0, + "learning_rate": 8.940545094076602e-07, + "loss": 0.6928, + "step": 13933 + }, + { + "epoch": 1.1873881550916063, + "grad_norm": 18.0, + "learning_rate": 8.938934337597999e-07, + "loss": 0.5207, + "step": 13934 + }, + { + "epoch": 1.1874733702599063, + "grad_norm": 10.75, + "learning_rate": 8.937323645471429e-07, + "loss": 0.2774, + "step": 13935 + }, + { + "epoch": 1.1875585854282062, + "grad_norm": 12.5625, + "learning_rate": 8.935713017726006e-07, + "loss": 0.3826, + "step": 13936 + }, + { + "epoch": 1.1876438005965062, + "grad_norm": 11.3125, + "learning_rate": 8.934102454390832e-07, + "loss": 0.5029, + "step": 13937 + }, + { + "epoch": 1.1877290157648062, + "grad_norm": 12.3125, + "learning_rate": 8.932491955495015e-07, + "loss": 0.3429, + "step": 13938 + }, + { + "epoch": 1.1878142309331061, + "grad_norm": 18.0, + "learning_rate": 8.930881521067652e-07, + "loss": 0.5205, + "step": 13939 + }, + { + "epoch": 1.187899446101406, + "grad_norm": 16.375, + "learning_rate": 8.929271151137852e-07, + "loss": 0.6538, + "step": 13940 + }, + { + "epoch": 1.187984661269706, + "grad_norm": 20.25, + "learning_rate": 8.927660845734714e-07, + "loss": 0.6821, + "step": 13941 + }, + { + "epoch": 1.188069876438006, + "grad_norm": 17.75, + "learning_rate": 8.926050604887334e-07, + "loss": 0.7899, + "step": 13942 + }, + { + "epoch": 1.188155091606306, + "grad_norm": 12.3125, + "learning_rate": 8.924440428624815e-07, + "loss": 0.491, + "step": 13943 + }, + { + "epoch": 1.188240306774606, + "grad_norm": 11.4375, + "learning_rate": 8.922830316976252e-07, + "loss": 0.3275, + "step": 13944 + }, + { + "epoch": 1.1883255219429059, + "grad_norm": 16.75, + "learning_rate": 8.921220269970743e-07, + "loss": 0.6647, + "step": 13945 + }, + { + "epoch": 1.1884107371112058, + "grad_norm": 16.375, + "learning_rate": 8.919610287637381e-07, + "loss": 0.7892, + "step": 13946 + }, + { + "epoch": 1.1884959522795058, + "grad_norm": 16.75, + "learning_rate": 8.918000370005261e-07, + "loss": 0.8509, + "step": 13947 + }, + { + "epoch": 1.1885811674478057, + "grad_norm": 13.1875, + "learning_rate": 8.916390517103479e-07, + "loss": 0.3842, + "step": 13948 + }, + { + "epoch": 1.1886663826161057, + "grad_norm": 12.625, + "learning_rate": 8.914780728961121e-07, + "loss": 0.3759, + "step": 13949 + }, + { + "epoch": 1.1887515977844056, + "grad_norm": 15.8125, + "learning_rate": 8.913171005607285e-07, + "loss": 0.6726, + "step": 13950 + }, + { + "epoch": 1.1888368129527056, + "grad_norm": 20.375, + "learning_rate": 8.911561347071055e-07, + "loss": 0.6583, + "step": 13951 + }, + { + "epoch": 1.1889220281210056, + "grad_norm": 16.125, + "learning_rate": 8.909951753381515e-07, + "loss": 0.7319, + "step": 13952 + }, + { + "epoch": 1.1890072432893055, + "grad_norm": 10.6875, + "learning_rate": 8.908342224567762e-07, + "loss": 0.2693, + "step": 13953 + }, + { + "epoch": 1.1890924584576055, + "grad_norm": 18.75, + "learning_rate": 8.906732760658871e-07, + "loss": 0.4593, + "step": 13954 + }, + { + "epoch": 1.1891776736259054, + "grad_norm": 12.875, + "learning_rate": 8.905123361683937e-07, + "loss": 0.3662, + "step": 13955 + }, + { + "epoch": 1.1892628887942054, + "grad_norm": 13.3125, + "learning_rate": 8.903514027672036e-07, + "loss": 0.6233, + "step": 13956 + }, + { + "epoch": 1.1893481039625053, + "grad_norm": 15.0625, + "learning_rate": 8.901904758652254e-07, + "loss": 0.7191, + "step": 13957 + }, + { + "epoch": 1.1894333191308053, + "grad_norm": 13.375, + "learning_rate": 8.900295554653676e-07, + "loss": 0.4753, + "step": 13958 + }, + { + "epoch": 1.1895185342991053, + "grad_norm": 21.5, + "learning_rate": 8.898686415705374e-07, + "loss": 0.5589, + "step": 13959 + }, + { + "epoch": 1.1896037494674052, + "grad_norm": 17.75, + "learning_rate": 8.897077341836427e-07, + "loss": 0.9451, + "step": 13960 + }, + { + "epoch": 1.1896889646357052, + "grad_norm": 14.5625, + "learning_rate": 8.895468333075918e-07, + "loss": 0.7401, + "step": 13961 + }, + { + "epoch": 1.1897741798040051, + "grad_norm": 14.5625, + "learning_rate": 8.893859389452919e-07, + "loss": 0.2874, + "step": 13962 + }, + { + "epoch": 1.189859394972305, + "grad_norm": 16.625, + "learning_rate": 8.892250510996508e-07, + "loss": 0.7872, + "step": 13963 + }, + { + "epoch": 1.189944610140605, + "grad_norm": 12.8125, + "learning_rate": 8.890641697735757e-07, + "loss": 0.3925, + "step": 13964 + }, + { + "epoch": 1.190029825308905, + "grad_norm": 12.375, + "learning_rate": 8.889032949699742e-07, + "loss": 0.4832, + "step": 13965 + }, + { + "epoch": 1.190115040477205, + "grad_norm": 17.75, + "learning_rate": 8.887424266917535e-07, + "loss": 0.5644, + "step": 13966 + }, + { + "epoch": 1.190200255645505, + "grad_norm": 11.0, + "learning_rate": 8.8858156494182e-07, + "loss": 0.3026, + "step": 13967 + }, + { + "epoch": 1.1902854708138049, + "grad_norm": 12.8125, + "learning_rate": 8.884207097230812e-07, + "loss": 0.3956, + "step": 13968 + }, + { + "epoch": 1.1903706859821048, + "grad_norm": 14.9375, + "learning_rate": 8.882598610384437e-07, + "loss": 0.5994, + "step": 13969 + }, + { + "epoch": 1.1904559011504048, + "grad_norm": 14.1875, + "learning_rate": 8.880990188908143e-07, + "loss": 0.6867, + "step": 13970 + }, + { + "epoch": 1.1905411163187047, + "grad_norm": 12.25, + "learning_rate": 8.879381832830994e-07, + "loss": 0.3718, + "step": 13971 + }, + { + "epoch": 1.1906263314870047, + "grad_norm": 17.125, + "learning_rate": 8.87777354218206e-07, + "loss": 0.4537, + "step": 13972 + }, + { + "epoch": 1.1907115466553047, + "grad_norm": 14.6875, + "learning_rate": 8.876165316990396e-07, + "loss": 0.7113, + "step": 13973 + }, + { + "epoch": 1.1907967618236046, + "grad_norm": 14.0, + "learning_rate": 8.874557157285077e-07, + "loss": 0.418, + "step": 13974 + }, + { + "epoch": 1.1908819769919046, + "grad_norm": 14.5625, + "learning_rate": 8.872949063095148e-07, + "loss": 0.6324, + "step": 13975 + }, + { + "epoch": 1.1909671921602045, + "grad_norm": 13.1875, + "learning_rate": 8.871341034449679e-07, + "loss": 0.5445, + "step": 13976 + }, + { + "epoch": 1.1910524073285045, + "grad_norm": 12.75, + "learning_rate": 8.869733071377725e-07, + "loss": 0.3347, + "step": 13977 + }, + { + "epoch": 1.1911376224968044, + "grad_norm": 18.375, + "learning_rate": 8.868125173908346e-07, + "loss": 0.9493, + "step": 13978 + }, + { + "epoch": 1.1912228376651044, + "grad_norm": 17.875, + "learning_rate": 8.866517342070596e-07, + "loss": 0.7775, + "step": 13979 + }, + { + "epoch": 1.1913080528334044, + "grad_norm": 11.0625, + "learning_rate": 8.864909575893534e-07, + "loss": 0.4234, + "step": 13980 + }, + { + "epoch": 1.1913932680017043, + "grad_norm": 15.25, + "learning_rate": 8.863301875406208e-07, + "loss": 0.5289, + "step": 13981 + }, + { + "epoch": 1.1914784831700043, + "grad_norm": 17.625, + "learning_rate": 8.86169424063768e-07, + "loss": 0.7465, + "step": 13982 + }, + { + "epoch": 1.1915636983383042, + "grad_norm": 12.875, + "learning_rate": 8.860086671616991e-07, + "loss": 0.3823, + "step": 13983 + }, + { + "epoch": 1.1916489135066042, + "grad_norm": 23.0, + "learning_rate": 8.858479168373196e-07, + "loss": 0.7136, + "step": 13984 + }, + { + "epoch": 1.1917341286749041, + "grad_norm": 14.125, + "learning_rate": 8.856871730935346e-07, + "loss": 0.605, + "step": 13985 + }, + { + "epoch": 1.191819343843204, + "grad_norm": 9.6875, + "learning_rate": 8.855264359332483e-07, + "loss": 0.1831, + "step": 13986 + }, + { + "epoch": 1.191904559011504, + "grad_norm": 12.75, + "learning_rate": 8.853657053593664e-07, + "loss": 0.2646, + "step": 13987 + }, + { + "epoch": 1.191989774179804, + "grad_norm": 15.75, + "learning_rate": 8.852049813747924e-07, + "loss": 0.3613, + "step": 13988 + }, + { + "epoch": 1.192074989348104, + "grad_norm": 10.375, + "learning_rate": 8.850442639824314e-07, + "loss": 0.3577, + "step": 13989 + }, + { + "epoch": 1.192160204516404, + "grad_norm": 13.9375, + "learning_rate": 8.848835531851879e-07, + "loss": 0.5365, + "step": 13990 + }, + { + "epoch": 1.1922454196847039, + "grad_norm": 13.125, + "learning_rate": 8.847228489859655e-07, + "loss": 0.5703, + "step": 13991 + }, + { + "epoch": 1.1923306348530038, + "grad_norm": 14.875, + "learning_rate": 8.845621513876683e-07, + "loss": 0.4117, + "step": 13992 + }, + { + "epoch": 1.1924158500213038, + "grad_norm": 16.75, + "learning_rate": 8.844014603932008e-07, + "loss": 0.8187, + "step": 13993 + }, + { + "epoch": 1.1925010651896037, + "grad_norm": 13.375, + "learning_rate": 8.842407760054661e-07, + "loss": 0.414, + "step": 13994 + }, + { + "epoch": 1.1925862803579037, + "grad_norm": 16.75, + "learning_rate": 8.840800982273687e-07, + "loss": 0.7227, + "step": 13995 + }, + { + "epoch": 1.1926714955262037, + "grad_norm": 12.5, + "learning_rate": 8.839194270618118e-07, + "loss": 0.2721, + "step": 13996 + }, + { + "epoch": 1.1927567106945036, + "grad_norm": 20.625, + "learning_rate": 8.83758762511699e-07, + "loss": 0.6246, + "step": 13997 + }, + { + "epoch": 1.1928419258628036, + "grad_norm": 16.5, + "learning_rate": 8.835981045799341e-07, + "loss": 0.7627, + "step": 13998 + }, + { + "epoch": 1.1929271410311035, + "grad_norm": 13.1875, + "learning_rate": 8.834374532694193e-07, + "loss": 0.4493, + "step": 13999 + }, + { + "epoch": 1.1930123561994035, + "grad_norm": 15.6875, + "learning_rate": 8.832768085830585e-07, + "loss": 0.6372, + "step": 14000 + }, + { + "epoch": 1.1930975713677034, + "grad_norm": 25.5, + "learning_rate": 8.831161705237544e-07, + "loss": 0.6108, + "step": 14001 + }, + { + "epoch": 1.1931827865360034, + "grad_norm": 14.5625, + "learning_rate": 8.829555390944103e-07, + "loss": 0.6974, + "step": 14002 + }, + { + "epoch": 1.1932680017043034, + "grad_norm": 21.0, + "learning_rate": 8.827949142979284e-07, + "loss": 1.0006, + "step": 14003 + }, + { + "epoch": 1.1933532168726033, + "grad_norm": 16.375, + "learning_rate": 8.826342961372118e-07, + "loss": 0.6715, + "step": 14004 + }, + { + "epoch": 1.1934384320409033, + "grad_norm": 15.0, + "learning_rate": 8.824736846151627e-07, + "loss": 0.5677, + "step": 14005 + }, + { + "epoch": 1.1935236472092032, + "grad_norm": 15.9375, + "learning_rate": 8.823130797346846e-07, + "loss": 0.5664, + "step": 14006 + }, + { + "epoch": 1.1936088623775032, + "grad_norm": 18.25, + "learning_rate": 8.821524814986778e-07, + "loss": 0.973, + "step": 14007 + }, + { + "epoch": 1.1936940775458031, + "grad_norm": 15.4375, + "learning_rate": 8.819918899100462e-07, + "loss": 0.6489, + "step": 14008 + }, + { + "epoch": 1.193779292714103, + "grad_norm": 17.625, + "learning_rate": 8.818313049716908e-07, + "loss": 0.7236, + "step": 14009 + }, + { + "epoch": 1.193864507882403, + "grad_norm": 17.75, + "learning_rate": 8.81670726686514e-07, + "loss": 0.7463, + "step": 14010 + }, + { + "epoch": 1.193949723050703, + "grad_norm": 23.25, + "learning_rate": 8.815101550574177e-07, + "loss": 1.1033, + "step": 14011 + }, + { + "epoch": 1.194034938219003, + "grad_norm": 11.125, + "learning_rate": 8.813495900873034e-07, + "loss": 0.3491, + "step": 14012 + }, + { + "epoch": 1.194120153387303, + "grad_norm": 11.1875, + "learning_rate": 8.811890317790725e-07, + "loss": 0.306, + "step": 14013 + }, + { + "epoch": 1.1942053685556029, + "grad_norm": 15.0625, + "learning_rate": 8.81028480135627e-07, + "loss": 0.4726, + "step": 14014 + }, + { + "epoch": 1.1942905837239028, + "grad_norm": 20.25, + "learning_rate": 8.808679351598681e-07, + "loss": 0.5591, + "step": 14015 + }, + { + "epoch": 1.1943757988922028, + "grad_norm": 17.0, + "learning_rate": 8.807073968546963e-07, + "loss": 0.6262, + "step": 14016 + }, + { + "epoch": 1.1944610140605028, + "grad_norm": 16.25, + "learning_rate": 8.805468652230134e-07, + "loss": 0.5463, + "step": 14017 + }, + { + "epoch": 1.1945462292288027, + "grad_norm": 18.375, + "learning_rate": 8.803863402677199e-07, + "loss": 0.9624, + "step": 14018 + }, + { + "epoch": 1.1946314443971027, + "grad_norm": 19.0, + "learning_rate": 8.802258219917173e-07, + "loss": 0.4853, + "step": 14019 + }, + { + "epoch": 1.1947166595654026, + "grad_norm": 29.625, + "learning_rate": 8.800653103979055e-07, + "loss": 0.911, + "step": 14020 + }, + { + "epoch": 1.1948018747337026, + "grad_norm": 17.25, + "learning_rate": 8.79904805489186e-07, + "loss": 0.6465, + "step": 14021 + }, + { + "epoch": 1.1948870899020025, + "grad_norm": 16.875, + "learning_rate": 8.797443072684586e-07, + "loss": 0.6253, + "step": 14022 + }, + { + "epoch": 1.1949723050703025, + "grad_norm": 33.75, + "learning_rate": 8.795838157386242e-07, + "loss": 0.8825, + "step": 14023 + }, + { + "epoch": 1.1950575202386025, + "grad_norm": 12.5625, + "learning_rate": 8.794233309025824e-07, + "loss": 0.5183, + "step": 14024 + }, + { + "epoch": 1.1951427354069024, + "grad_norm": 22.625, + "learning_rate": 8.792628527632337e-07, + "loss": 0.8591, + "step": 14025 + }, + { + "epoch": 1.1952279505752024, + "grad_norm": 16.5, + "learning_rate": 8.791023813234778e-07, + "loss": 0.4932, + "step": 14026 + }, + { + "epoch": 1.1953131657435023, + "grad_norm": 13.9375, + "learning_rate": 8.789419165862153e-07, + "loss": 0.4498, + "step": 14027 + }, + { + "epoch": 1.1953983809118023, + "grad_norm": 16.0, + "learning_rate": 8.78781458554345e-07, + "loss": 0.7703, + "step": 14028 + }, + { + "epoch": 1.1954835960801022, + "grad_norm": 24.0, + "learning_rate": 8.786210072307675e-07, + "loss": 0.7145, + "step": 14029 + }, + { + "epoch": 1.1955688112484022, + "grad_norm": 20.625, + "learning_rate": 8.784605626183813e-07, + "loss": 0.6528, + "step": 14030 + }, + { + "epoch": 1.1956540264167022, + "grad_norm": 15.3125, + "learning_rate": 8.78300124720087e-07, + "loss": 0.5606, + "step": 14031 + }, + { + "epoch": 1.195739241585002, + "grad_norm": 14.625, + "learning_rate": 8.781396935387832e-07, + "loss": 0.5712, + "step": 14032 + }, + { + "epoch": 1.195824456753302, + "grad_norm": 11.25, + "learning_rate": 8.779792690773686e-07, + "loss": 0.3576, + "step": 14033 + }, + { + "epoch": 1.195909671921602, + "grad_norm": 17.875, + "learning_rate": 8.778188513387428e-07, + "loss": 0.8496, + "step": 14034 + }, + { + "epoch": 1.195994887089902, + "grad_norm": 15.5, + "learning_rate": 8.776584403258046e-07, + "loss": 0.8143, + "step": 14035 + }, + { + "epoch": 1.196080102258202, + "grad_norm": 15.3125, + "learning_rate": 8.774980360414531e-07, + "loss": 0.3802, + "step": 14036 + }, + { + "epoch": 1.196165317426502, + "grad_norm": 13.4375, + "learning_rate": 8.773376384885865e-07, + "loss": 0.3343, + "step": 14037 + }, + { + "epoch": 1.1962505325948019, + "grad_norm": 13.6875, + "learning_rate": 8.771772476701033e-07, + "loss": 0.5611, + "step": 14038 + }, + { + "epoch": 1.1963357477631018, + "grad_norm": 14.875, + "learning_rate": 8.770168635889027e-07, + "loss": 0.7927, + "step": 14039 + }, + { + "epoch": 1.1964209629314018, + "grad_norm": 11.6875, + "learning_rate": 8.768564862478822e-07, + "loss": 0.5856, + "step": 14040 + }, + { + "epoch": 1.1965061780997017, + "grad_norm": 17.5, + "learning_rate": 8.766961156499398e-07, + "loss": 0.5329, + "step": 14041 + }, + { + "epoch": 1.1965913932680017, + "grad_norm": 12.125, + "learning_rate": 8.765357517979745e-07, + "loss": 0.5797, + "step": 14042 + }, + { + "epoch": 1.1966766084363016, + "grad_norm": 13.4375, + "learning_rate": 8.763753946948833e-07, + "loss": 0.3884, + "step": 14043 + }, + { + "epoch": 1.1967618236046016, + "grad_norm": 14.5625, + "learning_rate": 8.762150443435646e-07, + "loss": 0.4425, + "step": 14044 + }, + { + "epoch": 1.1968470387729016, + "grad_norm": 16.25, + "learning_rate": 8.760547007469158e-07, + "loss": 0.4221, + "step": 14045 + }, + { + "epoch": 1.1969322539412015, + "grad_norm": 12.5, + "learning_rate": 8.758943639078347e-07, + "loss": 0.4596, + "step": 14046 + }, + { + "epoch": 1.1970174691095015, + "grad_norm": 17.5, + "learning_rate": 8.757340338292191e-07, + "loss": 0.5352, + "step": 14047 + }, + { + "epoch": 1.1971026842778014, + "grad_norm": 14.6875, + "learning_rate": 8.755737105139653e-07, + "loss": 0.5569, + "step": 14048 + }, + { + "epoch": 1.1971878994461014, + "grad_norm": 16.375, + "learning_rate": 8.754133939649712e-07, + "loss": 0.6973, + "step": 14049 + }, + { + "epoch": 1.1972731146144013, + "grad_norm": 19.875, + "learning_rate": 8.752530841851335e-07, + "loss": 0.7481, + "step": 14050 + }, + { + "epoch": 1.1973583297827013, + "grad_norm": 13.0, + "learning_rate": 8.750927811773496e-07, + "loss": 0.467, + "step": 14051 + }, + { + "epoch": 1.1974435449510012, + "grad_norm": 18.75, + "learning_rate": 8.749324849445165e-07, + "loss": 0.5897, + "step": 14052 + }, + { + "epoch": 1.1975287601193012, + "grad_norm": 15.1875, + "learning_rate": 8.7477219548953e-07, + "loss": 0.8792, + "step": 14053 + }, + { + "epoch": 1.1976139752876012, + "grad_norm": 10.25, + "learning_rate": 8.746119128152875e-07, + "loss": 0.3574, + "step": 14054 + }, + { + "epoch": 1.1976991904559011, + "grad_norm": 14.3125, + "learning_rate": 8.744516369246855e-07, + "loss": 0.4556, + "step": 14055 + }, + { + "epoch": 1.197784405624201, + "grad_norm": 12.875, + "learning_rate": 8.742913678206197e-07, + "loss": 0.5211, + "step": 14056 + }, + { + "epoch": 1.197869620792501, + "grad_norm": 18.375, + "learning_rate": 8.741311055059868e-07, + "loss": 0.75, + "step": 14057 + }, + { + "epoch": 1.197954835960801, + "grad_norm": 14.0, + "learning_rate": 8.739708499836824e-07, + "loss": 0.4655, + "step": 14058 + }, + { + "epoch": 1.198040051129101, + "grad_norm": 17.25, + "learning_rate": 8.738106012566034e-07, + "loss": 1.029, + "step": 14059 + }, + { + "epoch": 1.198125266297401, + "grad_norm": 16.375, + "learning_rate": 8.736503593276448e-07, + "loss": 0.8588, + "step": 14060 + }, + { + "epoch": 1.1982104814657009, + "grad_norm": 14.125, + "learning_rate": 8.734901241997029e-07, + "loss": 0.5043, + "step": 14061 + }, + { + "epoch": 1.1982956966340008, + "grad_norm": 16.25, + "learning_rate": 8.733298958756727e-07, + "loss": 0.7785, + "step": 14062 + }, + { + "epoch": 1.1983809118023008, + "grad_norm": 18.625, + "learning_rate": 8.731696743584505e-07, + "loss": 0.6924, + "step": 14063 + }, + { + "epoch": 1.1984661269706007, + "grad_norm": 19.375, + "learning_rate": 8.730094596509311e-07, + "loss": 0.5844, + "step": 14064 + }, + { + "epoch": 1.1985513421389007, + "grad_norm": 14.0, + "learning_rate": 8.728492517560097e-07, + "loss": 0.5937, + "step": 14065 + }, + { + "epoch": 1.1986365573072006, + "grad_norm": 13.0625, + "learning_rate": 8.726890506765818e-07, + "loss": 0.4447, + "step": 14066 + }, + { + "epoch": 1.1987217724755006, + "grad_norm": 13.5, + "learning_rate": 8.725288564155421e-07, + "loss": 0.6119, + "step": 14067 + }, + { + "epoch": 1.1988069876438006, + "grad_norm": 26.75, + "learning_rate": 8.723686689757852e-07, + "loss": 0.8273, + "step": 14068 + }, + { + "epoch": 1.1988922028121005, + "grad_norm": 13.625, + "learning_rate": 8.722084883602067e-07, + "loss": 0.5634, + "step": 14069 + }, + { + "epoch": 1.1989774179804005, + "grad_norm": 15.0625, + "learning_rate": 8.720483145717001e-07, + "loss": 0.7647, + "step": 14070 + }, + { + "epoch": 1.1990626331487004, + "grad_norm": 15.4375, + "learning_rate": 8.718881476131613e-07, + "loss": 0.7546, + "step": 14071 + }, + { + "epoch": 1.1991478483170004, + "grad_norm": 15.4375, + "learning_rate": 8.717279874874836e-07, + "loss": 0.72, + "step": 14072 + }, + { + "epoch": 1.1992330634853003, + "grad_norm": 12.9375, + "learning_rate": 8.715678341975612e-07, + "loss": 0.4732, + "step": 14073 + }, + { + "epoch": 1.1993182786536003, + "grad_norm": 19.375, + "learning_rate": 8.714076877462889e-07, + "loss": 0.6179, + "step": 14074 + }, + { + "epoch": 1.1994034938219003, + "grad_norm": 30.125, + "learning_rate": 8.7124754813656e-07, + "loss": 1.0587, + "step": 14075 + }, + { + "epoch": 1.1994887089902002, + "grad_norm": 19.25, + "learning_rate": 8.710874153712693e-07, + "loss": 0.9087, + "step": 14076 + }, + { + "epoch": 1.1995739241585002, + "grad_norm": 13.9375, + "learning_rate": 8.709272894533096e-07, + "loss": 0.5339, + "step": 14077 + }, + { + "epoch": 1.1996591393268001, + "grad_norm": 13.9375, + "learning_rate": 8.707671703855752e-07, + "loss": 0.492, + "step": 14078 + }, + { + "epoch": 1.1997443544951, + "grad_norm": 13.9375, + "learning_rate": 8.706070581709597e-07, + "loss": 0.8418, + "step": 14079 + }, + { + "epoch": 1.1998295696634, + "grad_norm": 17.5, + "learning_rate": 8.704469528123554e-07, + "loss": 0.6859, + "step": 14080 + }, + { + "epoch": 1.1999147848317, + "grad_norm": 17.375, + "learning_rate": 8.702868543126569e-07, + "loss": 0.5632, + "step": 14081 + }, + { + "epoch": 1.2, + "grad_norm": 13.375, + "learning_rate": 8.701267626747567e-07, + "loss": 0.5121, + "step": 14082 + }, + { + "epoch": 1.2000852151683, + "grad_norm": 17.75, + "learning_rate": 8.699666779015476e-07, + "loss": 0.3943, + "step": 14083 + }, + { + "epoch": 1.2001704303365999, + "grad_norm": 12.6875, + "learning_rate": 8.698065999959232e-07, + "loss": 0.431, + "step": 14084 + }, + { + "epoch": 1.2002556455048998, + "grad_norm": 33.0, + "learning_rate": 8.696465289607754e-07, + "loss": 0.5714, + "step": 14085 + }, + { + "epoch": 1.2003408606731998, + "grad_norm": 15.0625, + "learning_rate": 8.694864647989975e-07, + "loss": 0.4568, + "step": 14086 + }, + { + "epoch": 1.2004260758414997, + "grad_norm": 23.125, + "learning_rate": 8.693264075134818e-07, + "loss": 0.3078, + "step": 14087 + }, + { + "epoch": 1.2005112910097997, + "grad_norm": 31.375, + "learning_rate": 8.691663571071211e-07, + "loss": 0.6994, + "step": 14088 + }, + { + "epoch": 1.2005965061780997, + "grad_norm": 14.75, + "learning_rate": 8.690063135828069e-07, + "loss": 0.7324, + "step": 14089 + }, + { + "epoch": 1.2006817213463996, + "grad_norm": 16.25, + "learning_rate": 8.688462769434316e-07, + "loss": 0.7098, + "step": 14090 + }, + { + "epoch": 1.2007669365146996, + "grad_norm": 17.125, + "learning_rate": 8.686862471918877e-07, + "loss": 0.4589, + "step": 14091 + }, + { + "epoch": 1.2008521516829995, + "grad_norm": 38.0, + "learning_rate": 8.685262243310665e-07, + "loss": 0.7908, + "step": 14092 + }, + { + "epoch": 1.2009373668512995, + "grad_norm": 23.75, + "learning_rate": 8.683662083638601e-07, + "loss": 1.1645, + "step": 14093 + }, + { + "epoch": 1.2010225820195994, + "grad_norm": 18.75, + "learning_rate": 8.6820619929316e-07, + "loss": 0.7679, + "step": 14094 + }, + { + "epoch": 1.2011077971878994, + "grad_norm": 13.3125, + "learning_rate": 8.680461971218579e-07, + "loss": 0.4885, + "step": 14095 + }, + { + "epoch": 1.2011930123561994, + "grad_norm": 15.1875, + "learning_rate": 8.678862018528455e-07, + "loss": 0.4537, + "step": 14096 + }, + { + "epoch": 1.2012782275244993, + "grad_norm": 21.625, + "learning_rate": 8.677262134890133e-07, + "loss": 0.8193, + "step": 14097 + }, + { + "epoch": 1.2013634426927993, + "grad_norm": 20.125, + "learning_rate": 8.675662320332526e-07, + "loss": 0.514, + "step": 14098 + }, + { + "epoch": 1.2014486578610992, + "grad_norm": 18.5, + "learning_rate": 8.674062574884551e-07, + "loss": 0.8436, + "step": 14099 + }, + { + "epoch": 1.2015338730293992, + "grad_norm": 16.125, + "learning_rate": 8.672462898575107e-07, + "loss": 0.9233, + "step": 14100 + }, + { + "epoch": 1.2016190881976991, + "grad_norm": 18.625, + "learning_rate": 8.670863291433113e-07, + "loss": 1.1457, + "step": 14101 + }, + { + "epoch": 1.201704303365999, + "grad_norm": 16.875, + "learning_rate": 8.669263753487464e-07, + "loss": 0.7729, + "step": 14102 + }, + { + "epoch": 1.201789518534299, + "grad_norm": 17.5, + "learning_rate": 8.667664284767075e-07, + "loss": 0.9133, + "step": 14103 + }, + { + "epoch": 1.201874733702599, + "grad_norm": 32.5, + "learning_rate": 8.666064885300848e-07, + "loss": 0.6118, + "step": 14104 + }, + { + "epoch": 1.201959948870899, + "grad_norm": 10.6875, + "learning_rate": 8.664465555117681e-07, + "loss": 0.3658, + "step": 14105 + }, + { + "epoch": 1.202045164039199, + "grad_norm": 14.4375, + "learning_rate": 8.662866294246479e-07, + "loss": 0.6481, + "step": 14106 + }, + { + "epoch": 1.2021303792074989, + "grad_norm": 17.625, + "learning_rate": 8.661267102716139e-07, + "loss": 0.4409, + "step": 14107 + }, + { + "epoch": 1.2022155943757988, + "grad_norm": 11.4375, + "learning_rate": 8.659667980555567e-07, + "loss": 0.314, + "step": 14108 + }, + { + "epoch": 1.2023008095440988, + "grad_norm": 19.625, + "learning_rate": 8.658068927793651e-07, + "loss": 0.6957, + "step": 14109 + }, + { + "epoch": 1.2023860247123987, + "grad_norm": 14.625, + "learning_rate": 8.656469944459297e-07, + "loss": 0.4927, + "step": 14110 + }, + { + "epoch": 1.2024712398806987, + "grad_norm": 13.0, + "learning_rate": 8.654871030581394e-07, + "loss": 0.3295, + "step": 14111 + }, + { + "epoch": 1.2025564550489987, + "grad_norm": 13.0, + "learning_rate": 8.653272186188844e-07, + "loss": 0.5615, + "step": 14112 + }, + { + "epoch": 1.2026416702172986, + "grad_norm": 17.375, + "learning_rate": 8.651673411310526e-07, + "loss": 0.7884, + "step": 14113 + }, + { + "epoch": 1.2027268853855986, + "grad_norm": 13.8125, + "learning_rate": 8.650074705975345e-07, + "loss": 0.5495, + "step": 14114 + }, + { + "epoch": 1.2028121005538985, + "grad_norm": 18.25, + "learning_rate": 8.64847607021218e-07, + "loss": 0.7949, + "step": 14115 + }, + { + "epoch": 1.2028973157221985, + "grad_norm": 13.875, + "learning_rate": 8.646877504049929e-07, + "loss": 0.6935, + "step": 14116 + }, + { + "epoch": 1.2029825308904984, + "grad_norm": 20.125, + "learning_rate": 8.645279007517474e-07, + "loss": 0.632, + "step": 14117 + }, + { + "epoch": 1.2030677460587984, + "grad_norm": 28.25, + "learning_rate": 8.643680580643704e-07, + "loss": 1.6578, + "step": 14118 + }, + { + "epoch": 1.2031529612270984, + "grad_norm": 15.75, + "learning_rate": 8.642082223457504e-07, + "loss": 0.7916, + "step": 14119 + }, + { + "epoch": 1.2032381763953983, + "grad_norm": 14.1875, + "learning_rate": 8.640483935987762e-07, + "loss": 0.5402, + "step": 14120 + }, + { + "epoch": 1.2033233915636983, + "grad_norm": 17.25, + "learning_rate": 8.638885718263352e-07, + "loss": 0.8929, + "step": 14121 + }, + { + "epoch": 1.2034086067319982, + "grad_norm": 13.375, + "learning_rate": 8.637287570313159e-07, + "loss": 0.2864, + "step": 14122 + }, + { + "epoch": 1.2034938219002982, + "grad_norm": 17.625, + "learning_rate": 8.635689492166066e-07, + "loss": 0.5802, + "step": 14123 + }, + { + "epoch": 1.2035790370685981, + "grad_norm": 14.3125, + "learning_rate": 8.634091483850946e-07, + "loss": 0.5603, + "step": 14124 + }, + { + "epoch": 1.203664252236898, + "grad_norm": 12.125, + "learning_rate": 8.632493545396683e-07, + "loss": 0.3803, + "step": 14125 + }, + { + "epoch": 1.203749467405198, + "grad_norm": 14.75, + "learning_rate": 8.630895676832151e-07, + "loss": 0.4944, + "step": 14126 + }, + { + "epoch": 1.203834682573498, + "grad_norm": 15.3125, + "learning_rate": 8.629297878186224e-07, + "loss": 0.6526, + "step": 14127 + }, + { + "epoch": 1.203919897741798, + "grad_norm": 23.5, + "learning_rate": 8.627700149487782e-07, + "loss": 0.8024, + "step": 14128 + }, + { + "epoch": 1.204005112910098, + "grad_norm": 19.875, + "learning_rate": 8.626102490765688e-07, + "loss": 0.563, + "step": 14129 + }, + { + "epoch": 1.2040903280783979, + "grad_norm": 14.5625, + "learning_rate": 8.624504902048815e-07, + "loss": 0.7062, + "step": 14130 + }, + { + "epoch": 1.2041755432466978, + "grad_norm": 19.125, + "learning_rate": 8.622907383366039e-07, + "loss": 0.8156, + "step": 14131 + }, + { + "epoch": 1.2042607584149978, + "grad_norm": 9.8125, + "learning_rate": 8.621309934746222e-07, + "loss": 0.2526, + "step": 14132 + }, + { + "epoch": 1.2043459735832978, + "grad_norm": 17.125, + "learning_rate": 8.619712556218236e-07, + "loss": 0.5359, + "step": 14133 + }, + { + "epoch": 1.2044311887515977, + "grad_norm": 17.0, + "learning_rate": 8.618115247810945e-07, + "loss": 1.0013, + "step": 14134 + }, + { + "epoch": 1.2045164039198977, + "grad_norm": 16.375, + "learning_rate": 8.616518009553218e-07, + "loss": 0.8414, + "step": 14135 + }, + { + "epoch": 1.2046016190881976, + "grad_norm": 22.25, + "learning_rate": 8.614920841473917e-07, + "loss": 0.729, + "step": 14136 + }, + { + "epoch": 1.2046868342564976, + "grad_norm": 13.1875, + "learning_rate": 8.613323743601899e-07, + "loss": 0.5106, + "step": 14137 + }, + { + "epoch": 1.2047720494247975, + "grad_norm": 24.0, + "learning_rate": 8.611726715966032e-07, + "loss": 0.9207, + "step": 14138 + }, + { + "epoch": 1.2048572645930975, + "grad_norm": 23.5, + "learning_rate": 8.61012975859517e-07, + "loss": 0.7182, + "step": 14139 + }, + { + "epoch": 1.2049424797613975, + "grad_norm": 17.5, + "learning_rate": 8.608532871518177e-07, + "loss": 0.6125, + "step": 14140 + }, + { + "epoch": 1.2050276949296974, + "grad_norm": 15.9375, + "learning_rate": 8.606936054763907e-07, + "loss": 0.7645, + "step": 14141 + }, + { + "epoch": 1.2051129100979974, + "grad_norm": 21.75, + "learning_rate": 8.605339308361219e-07, + "loss": 0.6365, + "step": 14142 + }, + { + "epoch": 1.2051981252662973, + "grad_norm": 12.25, + "learning_rate": 8.603742632338963e-07, + "loss": 0.3891, + "step": 14143 + }, + { + "epoch": 1.2052833404345973, + "grad_norm": 17.25, + "learning_rate": 8.602146026726005e-07, + "loss": 0.8285, + "step": 14144 + }, + { + "epoch": 1.2053685556028972, + "grad_norm": 12.5, + "learning_rate": 8.600549491551179e-07, + "loss": 0.3931, + "step": 14145 + }, + { + "epoch": 1.2054537707711972, + "grad_norm": 16.125, + "learning_rate": 8.598953026843347e-07, + "loss": 0.7789, + "step": 14146 + }, + { + "epoch": 1.2055389859394972, + "grad_norm": 17.25, + "learning_rate": 8.597356632631354e-07, + "loss": 0.6328, + "step": 14147 + }, + { + "epoch": 1.205624201107797, + "grad_norm": 12.375, + "learning_rate": 8.595760308944056e-07, + "loss": 0.1813, + "step": 14148 + }, + { + "epoch": 1.205709416276097, + "grad_norm": 21.375, + "learning_rate": 8.594164055810289e-07, + "loss": 0.8747, + "step": 14149 + }, + { + "epoch": 1.205794631444397, + "grad_norm": 20.375, + "learning_rate": 8.592567873258911e-07, + "loss": 0.6306, + "step": 14150 + }, + { + "epoch": 1.205879846612697, + "grad_norm": 20.75, + "learning_rate": 8.590971761318756e-07, + "loss": 0.7399, + "step": 14151 + }, + { + "epoch": 1.205965061780997, + "grad_norm": 17.5, + "learning_rate": 8.589375720018675e-07, + "loss": 0.6904, + "step": 14152 + }, + { + "epoch": 1.206050276949297, + "grad_norm": 17.0, + "learning_rate": 8.58777974938751e-07, + "loss": 0.5246, + "step": 14153 + }, + { + "epoch": 1.2061354921175969, + "grad_norm": 18.75, + "learning_rate": 8.586183849454094e-07, + "loss": 0.6141, + "step": 14154 + }, + { + "epoch": 1.2062207072858968, + "grad_norm": 15.0, + "learning_rate": 8.584588020247275e-07, + "loss": 0.5386, + "step": 14155 + }, + { + "epoch": 1.2063059224541968, + "grad_norm": 32.25, + "learning_rate": 8.582992261795883e-07, + "loss": 0.7592, + "step": 14156 + }, + { + "epoch": 1.2063911376224967, + "grad_norm": 16.375, + "learning_rate": 8.581396574128764e-07, + "loss": 0.7365, + "step": 14157 + }, + { + "epoch": 1.2064763527907967, + "grad_norm": 16.25, + "learning_rate": 8.579800957274748e-07, + "loss": 0.5283, + "step": 14158 + }, + { + "epoch": 1.2065615679590966, + "grad_norm": 14.6875, + "learning_rate": 8.57820541126267e-07, + "loss": 0.4268, + "step": 14159 + }, + { + "epoch": 1.2066467831273966, + "grad_norm": 19.0, + "learning_rate": 8.576609936121363e-07, + "loss": 0.9223, + "step": 14160 + }, + { + "epoch": 1.2067319982956966, + "grad_norm": 14.0625, + "learning_rate": 8.575014531879669e-07, + "loss": 0.7523, + "step": 14161 + }, + { + "epoch": 1.2068172134639965, + "grad_norm": 30.5, + "learning_rate": 8.573419198566399e-07, + "loss": 0.7288, + "step": 14162 + }, + { + "epoch": 1.2069024286322965, + "grad_norm": 15.4375, + "learning_rate": 8.571823936210397e-07, + "loss": 0.7984, + "step": 14163 + }, + { + "epoch": 1.2069876438005964, + "grad_norm": 11.75, + "learning_rate": 8.570228744840482e-07, + "loss": 0.3422, + "step": 14164 + }, + { + "epoch": 1.2070728589688964, + "grad_norm": 21.25, + "learning_rate": 8.568633624485489e-07, + "loss": 0.9414, + "step": 14165 + }, + { + "epoch": 1.2071580741371963, + "grad_norm": 17.125, + "learning_rate": 8.567038575174237e-07, + "loss": 0.488, + "step": 14166 + }, + { + "epoch": 1.2072432893054963, + "grad_norm": 15.6875, + "learning_rate": 8.565443596935555e-07, + "loss": 0.8764, + "step": 14167 + }, + { + "epoch": 1.2073285044737962, + "grad_norm": 19.75, + "learning_rate": 8.563848689798262e-07, + "loss": 0.7281, + "step": 14168 + }, + { + "epoch": 1.2074137196420962, + "grad_norm": 18.875, + "learning_rate": 8.562253853791186e-07, + "loss": 0.8583, + "step": 14169 + }, + { + "epoch": 1.2074989348103962, + "grad_norm": 16.375, + "learning_rate": 8.560659088943141e-07, + "loss": 0.5999, + "step": 14170 + }, + { + "epoch": 1.2075841499786961, + "grad_norm": 24.875, + "learning_rate": 8.559064395282942e-07, + "loss": 0.6782, + "step": 14171 + }, + { + "epoch": 1.207669365146996, + "grad_norm": 18.5, + "learning_rate": 8.557469772839419e-07, + "loss": 0.6869, + "step": 14172 + }, + { + "epoch": 1.207754580315296, + "grad_norm": 17.0, + "learning_rate": 8.555875221641377e-07, + "loss": 0.394, + "step": 14173 + }, + { + "epoch": 1.207839795483596, + "grad_norm": 13.125, + "learning_rate": 8.554280741717639e-07, + "loss": 0.4108, + "step": 14174 + }, + { + "epoch": 1.207925010651896, + "grad_norm": 12.75, + "learning_rate": 8.552686333097011e-07, + "loss": 0.4517, + "step": 14175 + }, + { + "epoch": 1.208010225820196, + "grad_norm": 11.8125, + "learning_rate": 8.551091995808316e-07, + "loss": 0.4063, + "step": 14176 + }, + { + "epoch": 1.2080954409884959, + "grad_norm": 13.5, + "learning_rate": 8.54949772988036e-07, + "loss": 0.3967, + "step": 14177 + }, + { + "epoch": 1.2081806561567958, + "grad_norm": 21.375, + "learning_rate": 8.54790353534195e-07, + "loss": 0.8404, + "step": 14178 + }, + { + "epoch": 1.2082658713250958, + "grad_norm": 23.0, + "learning_rate": 8.546309412221895e-07, + "loss": 0.9861, + "step": 14179 + }, + { + "epoch": 1.2083510864933957, + "grad_norm": 21.625, + "learning_rate": 8.544715360549008e-07, + "loss": 0.3546, + "step": 14180 + }, + { + "epoch": 1.2084363016616957, + "grad_norm": 18.75, + "learning_rate": 8.543121380352088e-07, + "loss": 0.5103, + "step": 14181 + }, + { + "epoch": 1.2085215168299956, + "grad_norm": 16.125, + "learning_rate": 8.541527471659948e-07, + "loss": 0.7188, + "step": 14182 + }, + { + "epoch": 1.2086067319982956, + "grad_norm": 15.25, + "learning_rate": 8.539933634501384e-07, + "loss": 0.5575, + "step": 14183 + }, + { + "epoch": 1.2086919471665956, + "grad_norm": 13.9375, + "learning_rate": 8.538339868905203e-07, + "loss": 0.4004, + "step": 14184 + }, + { + "epoch": 1.2087771623348955, + "grad_norm": 11.625, + "learning_rate": 8.536746174900206e-07, + "loss": 0.2434, + "step": 14185 + }, + { + "epoch": 1.2088623775031955, + "grad_norm": 12.9375, + "learning_rate": 8.535152552515188e-07, + "loss": 0.48, + "step": 14186 + }, + { + "epoch": 1.2089475926714954, + "grad_norm": 13.25, + "learning_rate": 8.533559001778952e-07, + "loss": 0.5442, + "step": 14187 + }, + { + "epoch": 1.2090328078397954, + "grad_norm": 14.0625, + "learning_rate": 8.531965522720292e-07, + "loss": 0.2867, + "step": 14188 + }, + { + "epoch": 1.2091180230080953, + "grad_norm": 12.5625, + "learning_rate": 8.530372115368006e-07, + "loss": 0.4273, + "step": 14189 + }, + { + "epoch": 1.2092032381763955, + "grad_norm": 14.0625, + "learning_rate": 8.528778779750886e-07, + "loss": 0.5339, + "step": 14190 + }, + { + "epoch": 1.2092884533446955, + "grad_norm": 15.25, + "learning_rate": 8.527185515897729e-07, + "loss": 0.683, + "step": 14191 + }, + { + "epoch": 1.2093736685129954, + "grad_norm": 17.125, + "learning_rate": 8.525592323837325e-07, + "loss": 0.4142, + "step": 14192 + }, + { + "epoch": 1.2094588836812954, + "grad_norm": 11.5, + "learning_rate": 8.523999203598467e-07, + "loss": 0.3446, + "step": 14193 + }, + { + "epoch": 1.2095440988495954, + "grad_norm": 10.9375, + "learning_rate": 8.522406155209937e-07, + "loss": 0.4636, + "step": 14194 + }, + { + "epoch": 1.2096293140178953, + "grad_norm": 10.6875, + "learning_rate": 8.520813178700528e-07, + "loss": 0.2821, + "step": 14195 + }, + { + "epoch": 1.2097145291861953, + "grad_norm": 14.3125, + "learning_rate": 8.519220274099024e-07, + "loss": 0.5829, + "step": 14196 + }, + { + "epoch": 1.2097997443544952, + "grad_norm": 19.0, + "learning_rate": 8.517627441434217e-07, + "loss": 0.5798, + "step": 14197 + }, + { + "epoch": 1.2098849595227952, + "grad_norm": 18.75, + "learning_rate": 8.516034680734883e-07, + "loss": 0.8952, + "step": 14198 + }, + { + "epoch": 1.2099701746910951, + "grad_norm": 12.375, + "learning_rate": 8.514441992029812e-07, + "loss": 0.5965, + "step": 14199 + }, + { + "epoch": 1.210055389859395, + "grad_norm": 13.0, + "learning_rate": 8.512849375347779e-07, + "loss": 0.4492, + "step": 14200 + }, + { + "epoch": 1.210140605027695, + "grad_norm": 14.75, + "learning_rate": 8.511256830717571e-07, + "loss": 0.3748, + "step": 14201 + }, + { + "epoch": 1.210225820195995, + "grad_norm": 12.5625, + "learning_rate": 8.509664358167963e-07, + "loss": 0.3002, + "step": 14202 + }, + { + "epoch": 1.210311035364295, + "grad_norm": 15.875, + "learning_rate": 8.508071957727728e-07, + "loss": 0.856, + "step": 14203 + }, + { + "epoch": 1.210396250532595, + "grad_norm": 14.25, + "learning_rate": 8.50647962942565e-07, + "loss": 0.4694, + "step": 14204 + }, + { + "epoch": 1.2104814657008949, + "grad_norm": 16.75, + "learning_rate": 8.504887373290499e-07, + "loss": 0.6661, + "step": 14205 + }, + { + "epoch": 1.2105666808691948, + "grad_norm": 14.3125, + "learning_rate": 8.503295189351052e-07, + "loss": 0.614, + "step": 14206 + }, + { + "epoch": 1.2106518960374948, + "grad_norm": 17.0, + "learning_rate": 8.501703077636081e-07, + "loss": 0.4606, + "step": 14207 + }, + { + "epoch": 1.2107371112057947, + "grad_norm": 15.5, + "learning_rate": 8.500111038174352e-07, + "loss": 0.6667, + "step": 14208 + }, + { + "epoch": 1.2108223263740947, + "grad_norm": 15.0625, + "learning_rate": 8.498519070994646e-07, + "loss": 0.4457, + "step": 14209 + }, + { + "epoch": 1.2109075415423947, + "grad_norm": 14.375, + "learning_rate": 8.49692717612572e-07, + "loss": 0.5797, + "step": 14210 + }, + { + "epoch": 1.2109927567106946, + "grad_norm": 15.125, + "learning_rate": 8.495335353596343e-07, + "loss": 0.7663, + "step": 14211 + }, + { + "epoch": 1.2110779718789946, + "grad_norm": 24.0, + "learning_rate": 8.493743603435286e-07, + "loss": 1.0169, + "step": 14212 + }, + { + "epoch": 1.2111631870472945, + "grad_norm": 19.875, + "learning_rate": 8.492151925671307e-07, + "loss": 0.8912, + "step": 14213 + }, + { + "epoch": 1.2112484022155945, + "grad_norm": 14.125, + "learning_rate": 8.490560320333177e-07, + "loss": 0.4532, + "step": 14214 + }, + { + "epoch": 1.2113336173838944, + "grad_norm": 12.75, + "learning_rate": 8.488968787449652e-07, + "loss": 0.4928, + "step": 14215 + }, + { + "epoch": 1.2114188325521944, + "grad_norm": 16.75, + "learning_rate": 8.487377327049495e-07, + "loss": 0.6004, + "step": 14216 + }, + { + "epoch": 1.2115040477204944, + "grad_norm": 19.5, + "learning_rate": 8.485785939161463e-07, + "loss": 0.929, + "step": 14217 + }, + { + "epoch": 1.2115892628887943, + "grad_norm": 14.9375, + "learning_rate": 8.48419462381432e-07, + "loss": 0.7083, + "step": 14218 + }, + { + "epoch": 1.2116744780570943, + "grad_norm": 18.25, + "learning_rate": 8.482603381036816e-07, + "loss": 0.7162, + "step": 14219 + }, + { + "epoch": 1.2117596932253942, + "grad_norm": 12.4375, + "learning_rate": 8.481012210857708e-07, + "loss": 0.5165, + "step": 14220 + }, + { + "epoch": 1.2118449083936942, + "grad_norm": 18.25, + "learning_rate": 8.479421113305753e-07, + "loss": 0.6443, + "step": 14221 + }, + { + "epoch": 1.2119301235619941, + "grad_norm": 21.375, + "learning_rate": 8.477830088409699e-07, + "loss": 0.6465, + "step": 14222 + }, + { + "epoch": 1.212015338730294, + "grad_norm": 17.75, + "learning_rate": 8.4762391361983e-07, + "loss": 0.678, + "step": 14223 + }, + { + "epoch": 1.212100553898594, + "grad_norm": 12.125, + "learning_rate": 8.474648256700307e-07, + "loss": 0.2822, + "step": 14224 + }, + { + "epoch": 1.212185769066894, + "grad_norm": 13.9375, + "learning_rate": 8.473057449944464e-07, + "loss": 0.5982, + "step": 14225 + }, + { + "epoch": 1.212270984235194, + "grad_norm": 16.5, + "learning_rate": 8.471466715959531e-07, + "loss": 0.5527, + "step": 14226 + }, + { + "epoch": 1.212356199403494, + "grad_norm": 16.75, + "learning_rate": 8.46987605477424e-07, + "loss": 0.6016, + "step": 14227 + }, + { + "epoch": 1.2124414145717939, + "grad_norm": 14.875, + "learning_rate": 8.46828546641734e-07, + "loss": 0.3789, + "step": 14228 + }, + { + "epoch": 1.2125266297400938, + "grad_norm": 19.625, + "learning_rate": 8.466694950917579e-07, + "loss": 0.669, + "step": 14229 + }, + { + "epoch": 1.2126118449083938, + "grad_norm": 12.625, + "learning_rate": 8.465104508303691e-07, + "loss": 0.4945, + "step": 14230 + }, + { + "epoch": 1.2126970600766938, + "grad_norm": 81.0, + "learning_rate": 8.463514138604425e-07, + "loss": 0.3029, + "step": 14231 + }, + { + "epoch": 1.2127822752449937, + "grad_norm": 11.875, + "learning_rate": 8.461923841848517e-07, + "loss": 0.4501, + "step": 14232 + }, + { + "epoch": 1.2128674904132937, + "grad_norm": 12.375, + "learning_rate": 8.460333618064708e-07, + "loss": 0.4638, + "step": 14233 + }, + { + "epoch": 1.2129527055815936, + "grad_norm": 19.0, + "learning_rate": 8.458743467281734e-07, + "loss": 0.7756, + "step": 14234 + }, + { + "epoch": 1.2130379207498936, + "grad_norm": 20.875, + "learning_rate": 8.457153389528325e-07, + "loss": 0.7872, + "step": 14235 + }, + { + "epoch": 1.2131231359181935, + "grad_norm": 14.0, + "learning_rate": 8.455563384833221e-07, + "loss": 0.845, + "step": 14236 + }, + { + "epoch": 1.2132083510864935, + "grad_norm": 23.0, + "learning_rate": 8.453973453225156e-07, + "loss": 0.755, + "step": 14237 + }, + { + "epoch": 1.2132935662547935, + "grad_norm": 15.0, + "learning_rate": 8.452383594732855e-07, + "loss": 0.6139, + "step": 14238 + }, + { + "epoch": 1.2133787814230934, + "grad_norm": 15.5, + "learning_rate": 8.450793809385055e-07, + "loss": 0.5804, + "step": 14239 + }, + { + "epoch": 1.2134639965913934, + "grad_norm": 18.0, + "learning_rate": 8.449204097210481e-07, + "loss": 0.4057, + "step": 14240 + }, + { + "epoch": 1.2135492117596933, + "grad_norm": 18.75, + "learning_rate": 8.447614458237865e-07, + "loss": 0.7369, + "step": 14241 + }, + { + "epoch": 1.2136344269279933, + "grad_norm": 16.875, + "learning_rate": 8.446024892495932e-07, + "loss": 0.6297, + "step": 14242 + }, + { + "epoch": 1.2137196420962932, + "grad_norm": 14.25, + "learning_rate": 8.444435400013401e-07, + "loss": 0.5253, + "step": 14243 + }, + { + "epoch": 1.2138048572645932, + "grad_norm": 22.375, + "learning_rate": 8.442845980819005e-07, + "loss": 0.5096, + "step": 14244 + }, + { + "epoch": 1.2138900724328932, + "grad_norm": 16.25, + "learning_rate": 8.441256634941458e-07, + "loss": 0.7037, + "step": 14245 + }, + { + "epoch": 1.213975287601193, + "grad_norm": 15.25, + "learning_rate": 8.439667362409487e-07, + "loss": 0.3959, + "step": 14246 + }, + { + "epoch": 1.214060502769493, + "grad_norm": 18.875, + "learning_rate": 8.438078163251808e-07, + "loss": 0.8084, + "step": 14247 + }, + { + "epoch": 1.214145717937793, + "grad_norm": 16.125, + "learning_rate": 8.436489037497142e-07, + "loss": 0.4822, + "step": 14248 + }, + { + "epoch": 1.214230933106093, + "grad_norm": 13.625, + "learning_rate": 8.434899985174205e-07, + "loss": 0.7781, + "step": 14249 + }, + { + "epoch": 1.214316148274393, + "grad_norm": 13.5, + "learning_rate": 8.433311006311718e-07, + "loss": 0.6064, + "step": 14250 + }, + { + "epoch": 1.214401363442693, + "grad_norm": 13.125, + "learning_rate": 8.431722100938386e-07, + "loss": 0.5313, + "step": 14251 + }, + { + "epoch": 1.2144865786109929, + "grad_norm": 14.5625, + "learning_rate": 8.430133269082928e-07, + "loss": 0.3832, + "step": 14252 + }, + { + "epoch": 1.2145717937792928, + "grad_norm": 16.375, + "learning_rate": 8.428544510774052e-07, + "loss": 1.0254, + "step": 14253 + }, + { + "epoch": 1.2146570089475928, + "grad_norm": 13.3125, + "learning_rate": 8.426955826040473e-07, + "loss": 0.4287, + "step": 14254 + }, + { + "epoch": 1.2147422241158927, + "grad_norm": 16.25, + "learning_rate": 8.425367214910896e-07, + "loss": 0.7628, + "step": 14255 + }, + { + "epoch": 1.2148274392841927, + "grad_norm": 13.8125, + "learning_rate": 8.423778677414032e-07, + "loss": 0.5263, + "step": 14256 + }, + { + "epoch": 1.2149126544524926, + "grad_norm": 16.875, + "learning_rate": 8.422190213578585e-07, + "loss": 0.8243, + "step": 14257 + }, + { + "epoch": 1.2149978696207926, + "grad_norm": 15.125, + "learning_rate": 8.420601823433266e-07, + "loss": 0.4406, + "step": 14258 + }, + { + "epoch": 1.2150830847890925, + "grad_norm": 13.6875, + "learning_rate": 8.41901350700677e-07, + "loss": 0.6284, + "step": 14259 + }, + { + "epoch": 1.2151682999573925, + "grad_norm": 13.0, + "learning_rate": 8.417425264327804e-07, + "loss": 0.6977, + "step": 14260 + }, + { + "epoch": 1.2152535151256925, + "grad_norm": 15.9375, + "learning_rate": 8.415837095425069e-07, + "loss": 0.7631, + "step": 14261 + }, + { + "epoch": 1.2153387302939924, + "grad_norm": 15.1875, + "learning_rate": 8.414249000327265e-07, + "loss": 0.5302, + "step": 14262 + }, + { + "epoch": 1.2154239454622924, + "grad_norm": 14.3125, + "learning_rate": 8.412660979063091e-07, + "loss": 0.6646, + "step": 14263 + }, + { + "epoch": 1.2155091606305923, + "grad_norm": 16.875, + "learning_rate": 8.41107303166124e-07, + "loss": 0.8917, + "step": 14264 + }, + { + "epoch": 1.2155943757988923, + "grad_norm": 13.375, + "learning_rate": 8.409485158150414e-07, + "loss": 0.4421, + "step": 14265 + }, + { + "epoch": 1.2156795909671922, + "grad_norm": 19.5, + "learning_rate": 8.407897358559308e-07, + "loss": 0.8879, + "step": 14266 + }, + { + "epoch": 1.2157648061354922, + "grad_norm": 13.125, + "learning_rate": 8.406309632916608e-07, + "loss": 0.5283, + "step": 14267 + }, + { + "epoch": 1.2158500213037922, + "grad_norm": 18.25, + "learning_rate": 8.404721981251007e-07, + "loss": 0.8967, + "step": 14268 + }, + { + "epoch": 1.2159352364720921, + "grad_norm": 17.375, + "learning_rate": 8.403134403591201e-07, + "loss": 0.7255, + "step": 14269 + }, + { + "epoch": 1.216020451640392, + "grad_norm": 17.625, + "learning_rate": 8.401546899965873e-07, + "loss": 0.6983, + "step": 14270 + }, + { + "epoch": 1.216105666808692, + "grad_norm": 20.75, + "learning_rate": 8.399959470403718e-07, + "loss": 0.7689, + "step": 14271 + }, + { + "epoch": 1.216190881976992, + "grad_norm": 13.9375, + "learning_rate": 8.398372114933413e-07, + "loss": 0.5063, + "step": 14272 + }, + { + "epoch": 1.216276097145292, + "grad_norm": 17.625, + "learning_rate": 8.396784833583653e-07, + "loss": 0.8025, + "step": 14273 + }, + { + "epoch": 1.216361312313592, + "grad_norm": 19.375, + "learning_rate": 8.395197626383119e-07, + "loss": 0.9521, + "step": 14274 + }, + { + "epoch": 1.2164465274818919, + "grad_norm": 14.5, + "learning_rate": 8.393610493360487e-07, + "loss": 0.4882, + "step": 14275 + }, + { + "epoch": 1.2165317426501918, + "grad_norm": 15.3125, + "learning_rate": 8.392023434544444e-07, + "loss": 0.5282, + "step": 14276 + }, + { + "epoch": 1.2166169578184918, + "grad_norm": 18.375, + "learning_rate": 8.390436449963667e-07, + "loss": 0.8183, + "step": 14277 + }, + { + "epoch": 1.2167021729867917, + "grad_norm": 18.625, + "learning_rate": 8.388849539646837e-07, + "loss": 0.8769, + "step": 14278 + }, + { + "epoch": 1.2167873881550917, + "grad_norm": 17.375, + "learning_rate": 8.387262703622629e-07, + "loss": 0.9322, + "step": 14279 + }, + { + "epoch": 1.2168726033233916, + "grad_norm": 19.125, + "learning_rate": 8.385675941919722e-07, + "loss": 0.7885, + "step": 14280 + }, + { + "epoch": 1.2169578184916916, + "grad_norm": 17.125, + "learning_rate": 8.384089254566785e-07, + "loss": 0.6721, + "step": 14281 + }, + { + "epoch": 1.2170430336599916, + "grad_norm": 13.6875, + "learning_rate": 8.382502641592502e-07, + "loss": 0.6553, + "step": 14282 + }, + { + "epoch": 1.2171282488282915, + "grad_norm": 20.25, + "learning_rate": 8.380916103025527e-07, + "loss": 1.0162, + "step": 14283 + }, + { + "epoch": 1.2172134639965915, + "grad_norm": 13.875, + "learning_rate": 8.379329638894546e-07, + "loss": 0.3203, + "step": 14284 + }, + { + "epoch": 1.2172986791648914, + "grad_norm": 17.125, + "learning_rate": 8.377743249228218e-07, + "loss": 0.3779, + "step": 14285 + }, + { + "epoch": 1.2173838943331914, + "grad_norm": 18.625, + "learning_rate": 8.376156934055216e-07, + "loss": 0.6119, + "step": 14286 + }, + { + "epoch": 1.2174691095014913, + "grad_norm": 18.25, + "learning_rate": 8.374570693404205e-07, + "loss": 0.8944, + "step": 14287 + }, + { + "epoch": 1.2175543246697913, + "grad_norm": 16.375, + "learning_rate": 8.37298452730385e-07, + "loss": 0.4667, + "step": 14288 + }, + { + "epoch": 1.2176395398380913, + "grad_norm": 18.375, + "learning_rate": 8.371398435782813e-07, + "loss": 0.4449, + "step": 14289 + }, + { + "epoch": 1.2177247550063912, + "grad_norm": 20.625, + "learning_rate": 8.369812418869761e-07, + "loss": 0.998, + "step": 14290 + }, + { + "epoch": 1.2178099701746912, + "grad_norm": 14.5, + "learning_rate": 8.368226476593355e-07, + "loss": 0.731, + "step": 14291 + }, + { + "epoch": 1.2178951853429911, + "grad_norm": 18.75, + "learning_rate": 8.366640608982245e-07, + "loss": 0.4268, + "step": 14292 + }, + { + "epoch": 1.217980400511291, + "grad_norm": 16.375, + "learning_rate": 8.365054816065099e-07, + "loss": 0.6161, + "step": 14293 + }, + { + "epoch": 1.218065615679591, + "grad_norm": 19.625, + "learning_rate": 8.363469097870566e-07, + "loss": 0.5126, + "step": 14294 + }, + { + "epoch": 1.218150830847891, + "grad_norm": 16.5, + "learning_rate": 8.361883454427311e-07, + "loss": 0.397, + "step": 14295 + }, + { + "epoch": 1.218236046016191, + "grad_norm": 15.5, + "learning_rate": 8.36029788576398e-07, + "loss": 0.6022, + "step": 14296 + }, + { + "epoch": 1.218321261184491, + "grad_norm": 14.0625, + "learning_rate": 8.358712391909231e-07, + "loss": 0.4326, + "step": 14297 + }, + { + "epoch": 1.2184064763527909, + "grad_norm": 19.625, + "learning_rate": 8.35712697289171e-07, + "loss": 0.6293, + "step": 14298 + }, + { + "epoch": 1.2184916915210908, + "grad_norm": 15.25, + "learning_rate": 8.355541628740079e-07, + "loss": 0.5858, + "step": 14299 + }, + { + "epoch": 1.2185769066893908, + "grad_norm": 18.5, + "learning_rate": 8.353956359482973e-07, + "loss": 0.642, + "step": 14300 + }, + { + "epoch": 1.2186621218576907, + "grad_norm": 14.875, + "learning_rate": 8.352371165149045e-07, + "loss": 0.4365, + "step": 14301 + }, + { + "epoch": 1.2187473370259907, + "grad_norm": 12.6875, + "learning_rate": 8.350786045766939e-07, + "loss": 0.5058, + "step": 14302 + }, + { + "epoch": 1.2188325521942907, + "grad_norm": 16.25, + "learning_rate": 8.349201001365304e-07, + "loss": 0.7715, + "step": 14303 + }, + { + "epoch": 1.2189177673625906, + "grad_norm": 41.75, + "learning_rate": 8.347616031972778e-07, + "loss": 0.7642, + "step": 14304 + }, + { + "epoch": 1.2190029825308906, + "grad_norm": 11.625, + "learning_rate": 8.346031137618008e-07, + "loss": 0.372, + "step": 14305 + }, + { + "epoch": 1.2190881976991905, + "grad_norm": 11.6875, + "learning_rate": 8.34444631832963e-07, + "loss": 0.4241, + "step": 14306 + }, + { + "epoch": 1.2191734128674905, + "grad_norm": 10.875, + "learning_rate": 8.34286157413629e-07, + "loss": 0.3683, + "step": 14307 + }, + { + "epoch": 1.2192586280357904, + "grad_norm": 18.375, + "learning_rate": 8.341276905066621e-07, + "loss": 0.8122, + "step": 14308 + }, + { + "epoch": 1.2193438432040904, + "grad_norm": 10.875, + "learning_rate": 8.339692311149257e-07, + "loss": 0.2567, + "step": 14309 + }, + { + "epoch": 1.2194290583723904, + "grad_norm": 11.0625, + "learning_rate": 8.338107792412839e-07, + "loss": 0.3534, + "step": 14310 + }, + { + "epoch": 1.2195142735406903, + "grad_norm": 25.125, + "learning_rate": 8.336523348885997e-07, + "loss": 0.8438, + "step": 14311 + }, + { + "epoch": 1.2195994887089903, + "grad_norm": 10.1875, + "learning_rate": 8.334938980597365e-07, + "loss": 0.3333, + "step": 14312 + }, + { + "epoch": 1.2196847038772902, + "grad_norm": 25.125, + "learning_rate": 8.333354687575573e-07, + "loss": 0.9861, + "step": 14313 + }, + { + "epoch": 1.2197699190455902, + "grad_norm": 16.25, + "learning_rate": 8.331770469849254e-07, + "loss": 0.5639, + "step": 14314 + }, + { + "epoch": 1.2198551342138901, + "grad_norm": 12.9375, + "learning_rate": 8.330186327447036e-07, + "loss": 0.4092, + "step": 14315 + }, + { + "epoch": 1.21994034938219, + "grad_norm": 24.375, + "learning_rate": 8.328602260397541e-07, + "loss": 0.7765, + "step": 14316 + }, + { + "epoch": 1.22002556455049, + "grad_norm": 10.125, + "learning_rate": 8.327018268729398e-07, + "loss": 0.3123, + "step": 14317 + }, + { + "epoch": 1.22011077971879, + "grad_norm": 13.4375, + "learning_rate": 8.325434352471231e-07, + "loss": 0.4499, + "step": 14318 + }, + { + "epoch": 1.22019599488709, + "grad_norm": 12.875, + "learning_rate": 8.323850511651662e-07, + "loss": 0.4656, + "step": 14319 + }, + { + "epoch": 1.22028121005539, + "grad_norm": 16.25, + "learning_rate": 8.322266746299316e-07, + "loss": 0.6067, + "step": 14320 + }, + { + "epoch": 1.2203664252236899, + "grad_norm": 20.875, + "learning_rate": 8.320683056442807e-07, + "loss": 0.5858, + "step": 14321 + }, + { + "epoch": 1.2204516403919898, + "grad_norm": 15.5, + "learning_rate": 8.319099442110762e-07, + "loss": 0.4778, + "step": 14322 + }, + { + "epoch": 1.2205368555602898, + "grad_norm": 13.625, + "learning_rate": 8.317515903331796e-07, + "loss": 0.5522, + "step": 14323 + }, + { + "epoch": 1.2206220707285897, + "grad_norm": 11.4375, + "learning_rate": 8.315932440134518e-07, + "loss": 0.2924, + "step": 14324 + }, + { + "epoch": 1.2207072858968897, + "grad_norm": 14.1875, + "learning_rate": 8.314349052547552e-07, + "loss": 0.6461, + "step": 14325 + }, + { + "epoch": 1.2207925010651897, + "grad_norm": 15.5625, + "learning_rate": 8.312765740599505e-07, + "loss": 0.5876, + "step": 14326 + }, + { + "epoch": 1.2208777162334896, + "grad_norm": 16.375, + "learning_rate": 8.311182504318993e-07, + "loss": 0.6497, + "step": 14327 + }, + { + "epoch": 1.2209629314017896, + "grad_norm": 17.625, + "learning_rate": 8.309599343734623e-07, + "loss": 0.7302, + "step": 14328 + }, + { + "epoch": 1.2210481465700895, + "grad_norm": 15.4375, + "learning_rate": 8.308016258875009e-07, + "loss": 0.5807, + "step": 14329 + }, + { + "epoch": 1.2211333617383895, + "grad_norm": 16.25, + "learning_rate": 8.306433249768759e-07, + "loss": 0.5513, + "step": 14330 + }, + { + "epoch": 1.2212185769066894, + "grad_norm": 19.25, + "learning_rate": 8.304850316444477e-07, + "loss": 0.8578, + "step": 14331 + }, + { + "epoch": 1.2213037920749894, + "grad_norm": 14.4375, + "learning_rate": 8.303267458930763e-07, + "loss": 0.4727, + "step": 14332 + }, + { + "epoch": 1.2213890072432894, + "grad_norm": 14.875, + "learning_rate": 8.30168467725623e-07, + "loss": 0.3806, + "step": 14333 + }, + { + "epoch": 1.2214742224115893, + "grad_norm": 14.75, + "learning_rate": 8.300101971449474e-07, + "loss": 0.6695, + "step": 14334 + }, + { + "epoch": 1.2215594375798893, + "grad_norm": 13.875, + "learning_rate": 8.298519341539102e-07, + "loss": 0.6018, + "step": 14335 + }, + { + "epoch": 1.2216446527481892, + "grad_norm": 17.75, + "learning_rate": 8.296936787553706e-07, + "loss": 0.5728, + "step": 14336 + }, + { + "epoch": 1.2217298679164892, + "grad_norm": 12.4375, + "learning_rate": 8.295354309521893e-07, + "loss": 0.4016, + "step": 14337 + }, + { + "epoch": 1.2218150830847891, + "grad_norm": 11.0, + "learning_rate": 8.293771907472252e-07, + "loss": 0.3203, + "step": 14338 + }, + { + "epoch": 1.221900298253089, + "grad_norm": 20.875, + "learning_rate": 8.292189581433388e-07, + "loss": 1.08, + "step": 14339 + }, + { + "epoch": 1.221985513421389, + "grad_norm": 13.5, + "learning_rate": 8.290607331433886e-07, + "loss": 0.4541, + "step": 14340 + }, + { + "epoch": 1.222070728589689, + "grad_norm": 15.9375, + "learning_rate": 8.289025157502343e-07, + "loss": 0.7774, + "step": 14341 + }, + { + "epoch": 1.222155943757989, + "grad_norm": 17.75, + "learning_rate": 8.287443059667349e-07, + "loss": 0.6292, + "step": 14342 + }, + { + "epoch": 1.222241158926289, + "grad_norm": 11.4375, + "learning_rate": 8.285861037957493e-07, + "loss": 0.231, + "step": 14343 + }, + { + "epoch": 1.2223263740945889, + "grad_norm": 19.0, + "learning_rate": 8.284279092401368e-07, + "loss": 0.6922, + "step": 14344 + }, + { + "epoch": 1.2224115892628888, + "grad_norm": 22.875, + "learning_rate": 8.282697223027561e-07, + "loss": 1.0454, + "step": 14345 + }, + { + "epoch": 1.2224968044311888, + "grad_norm": 20.375, + "learning_rate": 8.281115429864652e-07, + "loss": 0.7021, + "step": 14346 + }, + { + "epoch": 1.2225820195994888, + "grad_norm": 17.25, + "learning_rate": 8.279533712941234e-07, + "loss": 0.6982, + "step": 14347 + }, + { + "epoch": 1.2226672347677887, + "grad_norm": 14.8125, + "learning_rate": 8.277952072285886e-07, + "loss": 0.568, + "step": 14348 + }, + { + "epoch": 1.2227524499360887, + "grad_norm": 14.75, + "learning_rate": 8.276370507927184e-07, + "loss": 0.2369, + "step": 14349 + }, + { + "epoch": 1.2228376651043886, + "grad_norm": 12.8125, + "learning_rate": 8.274789019893718e-07, + "loss": 0.5487, + "step": 14350 + }, + { + "epoch": 1.2229228802726886, + "grad_norm": 15.1875, + "learning_rate": 8.273207608214061e-07, + "loss": 0.6964, + "step": 14351 + }, + { + "epoch": 1.2230080954409885, + "grad_norm": 23.75, + "learning_rate": 8.271626272916793e-07, + "loss": 0.7589, + "step": 14352 + }, + { + "epoch": 1.2230933106092885, + "grad_norm": 15.5, + "learning_rate": 8.270045014030492e-07, + "loss": 0.4066, + "step": 14353 + }, + { + "epoch": 1.2231785257775885, + "grad_norm": 16.25, + "learning_rate": 8.268463831583731e-07, + "loss": 0.5047, + "step": 14354 + }, + { + "epoch": 1.2232637409458884, + "grad_norm": 12.5625, + "learning_rate": 8.266882725605083e-07, + "loss": 0.3694, + "step": 14355 + }, + { + "epoch": 1.2233489561141884, + "grad_norm": 11.1875, + "learning_rate": 8.265301696123124e-07, + "loss": 0.377, + "step": 14356 + }, + { + "epoch": 1.2234341712824883, + "grad_norm": 22.875, + "learning_rate": 8.263720743166418e-07, + "loss": 0.7289, + "step": 14357 + }, + { + "epoch": 1.2235193864507883, + "grad_norm": 13.0, + "learning_rate": 8.262139866763538e-07, + "loss": 0.5662, + "step": 14358 + }, + { + "epoch": 1.2236046016190882, + "grad_norm": 27.0, + "learning_rate": 8.260559066943053e-07, + "loss": 1.1001, + "step": 14359 + }, + { + "epoch": 1.2236898167873882, + "grad_norm": 14.375, + "learning_rate": 8.258978343733532e-07, + "loss": 0.5106, + "step": 14360 + }, + { + "epoch": 1.2237750319556882, + "grad_norm": 11.6875, + "learning_rate": 8.25739769716353e-07, + "loss": 0.3155, + "step": 14361 + }, + { + "epoch": 1.223860247123988, + "grad_norm": 18.0, + "learning_rate": 8.255817127261623e-07, + "loss": 0.4861, + "step": 14362 + }, + { + "epoch": 1.223945462292288, + "grad_norm": 13.3125, + "learning_rate": 8.254236634056365e-07, + "loss": 0.5578, + "step": 14363 + }, + { + "epoch": 1.224030677460588, + "grad_norm": 14.375, + "learning_rate": 8.252656217576327e-07, + "loss": 0.7464, + "step": 14364 + }, + { + "epoch": 1.224115892628888, + "grad_norm": 12.6875, + "learning_rate": 8.251075877850058e-07, + "loss": 0.3164, + "step": 14365 + }, + { + "epoch": 1.224201107797188, + "grad_norm": 13.1875, + "learning_rate": 8.249495614906118e-07, + "loss": 0.4481, + "step": 14366 + }, + { + "epoch": 1.224286322965488, + "grad_norm": 14.0625, + "learning_rate": 8.24791542877307e-07, + "loss": 0.6494, + "step": 14367 + }, + { + "epoch": 1.2243715381337879, + "grad_norm": 16.5, + "learning_rate": 8.246335319479462e-07, + "loss": 0.8217, + "step": 14368 + }, + { + "epoch": 1.2244567533020878, + "grad_norm": 16.25, + "learning_rate": 8.244755287053857e-07, + "loss": 0.732, + "step": 14369 + }, + { + "epoch": 1.2245419684703878, + "grad_norm": 17.125, + "learning_rate": 8.243175331524799e-07, + "loss": 0.5777, + "step": 14370 + }, + { + "epoch": 1.2246271836386877, + "grad_norm": 23.25, + "learning_rate": 8.241595452920845e-07, + "loss": 0.4402, + "step": 14371 + }, + { + "epoch": 1.2247123988069877, + "grad_norm": 28.25, + "learning_rate": 8.240015651270547e-07, + "loss": 1.1677, + "step": 14372 + }, + { + "epoch": 1.2247976139752876, + "grad_norm": 20.125, + "learning_rate": 8.238435926602445e-07, + "loss": 0.7971, + "step": 14373 + }, + { + "epoch": 1.2248828291435876, + "grad_norm": 20.5, + "learning_rate": 8.236856278945094e-07, + "loss": 0.8393, + "step": 14374 + }, + { + "epoch": 1.2249680443118875, + "grad_norm": 12.3125, + "learning_rate": 8.235276708327035e-07, + "loss": 0.4389, + "step": 14375 + }, + { + "epoch": 1.2250532594801875, + "grad_norm": 15.125, + "learning_rate": 8.233697214776814e-07, + "loss": 0.5732, + "step": 14376 + }, + { + "epoch": 1.2251384746484875, + "grad_norm": 18.375, + "learning_rate": 8.232117798322975e-07, + "loss": 0.8083, + "step": 14377 + }, + { + "epoch": 1.2252236898167874, + "grad_norm": 13.875, + "learning_rate": 8.230538458994058e-07, + "loss": 0.4711, + "step": 14378 + }, + { + "epoch": 1.2253089049850874, + "grad_norm": 14.375, + "learning_rate": 8.228959196818606e-07, + "loss": 0.5149, + "step": 14379 + }, + { + "epoch": 1.2253941201533873, + "grad_norm": 15.9375, + "learning_rate": 8.227380011825159e-07, + "loss": 0.6433, + "step": 14380 + }, + { + "epoch": 1.2254793353216873, + "grad_norm": 56.25, + "learning_rate": 8.225800904042247e-07, + "loss": 0.7149, + "step": 14381 + }, + { + "epoch": 1.2255645504899872, + "grad_norm": 21.125, + "learning_rate": 8.224221873498414e-07, + "loss": 0.8643, + "step": 14382 + }, + { + "epoch": 1.2256497656582872, + "grad_norm": 24.125, + "learning_rate": 8.222642920222187e-07, + "loss": 0.8243, + "step": 14383 + }, + { + "epoch": 1.2257349808265872, + "grad_norm": 18.25, + "learning_rate": 8.221064044242107e-07, + "loss": 0.6944, + "step": 14384 + }, + { + "epoch": 1.2258201959948871, + "grad_norm": 24.625, + "learning_rate": 8.219485245586701e-07, + "loss": 0.9646, + "step": 14385 + }, + { + "epoch": 1.225905411163187, + "grad_norm": 14.4375, + "learning_rate": 8.217906524284502e-07, + "loss": 0.7437, + "step": 14386 + }, + { + "epoch": 1.225990626331487, + "grad_norm": 18.5, + "learning_rate": 8.216327880364036e-07, + "loss": 0.5139, + "step": 14387 + }, + { + "epoch": 1.226075841499787, + "grad_norm": 21.75, + "learning_rate": 8.214749313853837e-07, + "loss": 1.1702, + "step": 14388 + }, + { + "epoch": 1.226161056668087, + "grad_norm": 14.8125, + "learning_rate": 8.213170824782426e-07, + "loss": 0.7668, + "step": 14389 + }, + { + "epoch": 1.226246271836387, + "grad_norm": 16.625, + "learning_rate": 8.211592413178329e-07, + "loss": 0.703, + "step": 14390 + }, + { + "epoch": 1.2263314870046869, + "grad_norm": 17.125, + "learning_rate": 8.210014079070066e-07, + "loss": 0.8535, + "step": 14391 + }, + { + "epoch": 1.2264167021729868, + "grad_norm": 24.375, + "learning_rate": 8.208435822486165e-07, + "loss": 0.6355, + "step": 14392 + }, + { + "epoch": 1.2265019173412868, + "grad_norm": 23.0, + "learning_rate": 8.206857643455142e-07, + "loss": 0.7368, + "step": 14393 + }, + { + "epoch": 1.2265871325095867, + "grad_norm": 23.25, + "learning_rate": 8.20527954200552e-07, + "loss": 0.8731, + "step": 14394 + }, + { + "epoch": 1.2266723476778867, + "grad_norm": 17.125, + "learning_rate": 8.203701518165814e-07, + "loss": 0.9526, + "step": 14395 + }, + { + "epoch": 1.2267575628461866, + "grad_norm": 16.0, + "learning_rate": 8.202123571964547e-07, + "loss": 0.7149, + "step": 14396 + }, + { + "epoch": 1.2268427780144866, + "grad_norm": 18.75, + "learning_rate": 8.200545703430227e-07, + "loss": 0.6252, + "step": 14397 + }, + { + "epoch": 1.2269279931827866, + "grad_norm": 17.0, + "learning_rate": 8.198967912591366e-07, + "loss": 0.7001, + "step": 14398 + }, + { + "epoch": 1.2270132083510865, + "grad_norm": 13.5, + "learning_rate": 8.197390199476482e-07, + "loss": 0.6932, + "step": 14399 + }, + { + "epoch": 1.2270984235193865, + "grad_norm": 13.0, + "learning_rate": 8.195812564114081e-07, + "loss": 0.4058, + "step": 14400 + }, + { + "epoch": 1.2271836386876864, + "grad_norm": 10.0, + "learning_rate": 8.194235006532678e-07, + "loss": 0.2526, + "step": 14401 + }, + { + "epoch": 1.2272688538559864, + "grad_norm": 14.875, + "learning_rate": 8.192657526760777e-07, + "loss": 0.6565, + "step": 14402 + }, + { + "epoch": 1.2273540690242863, + "grad_norm": 13.9375, + "learning_rate": 8.191080124826886e-07, + "loss": 0.4241, + "step": 14403 + }, + { + "epoch": 1.2274392841925863, + "grad_norm": 16.125, + "learning_rate": 8.189502800759515e-07, + "loss": 0.5987, + "step": 14404 + }, + { + "epoch": 1.2275244993608863, + "grad_norm": 10.1875, + "learning_rate": 8.187925554587159e-07, + "loss": 0.2314, + "step": 14405 + }, + { + "epoch": 1.2276097145291862, + "grad_norm": 13.75, + "learning_rate": 8.186348386338322e-07, + "loss": 0.3858, + "step": 14406 + }, + { + "epoch": 1.2276949296974862, + "grad_norm": 16.625, + "learning_rate": 8.18477129604151e-07, + "loss": 0.3744, + "step": 14407 + }, + { + "epoch": 1.2277801448657861, + "grad_norm": 21.125, + "learning_rate": 8.183194283725218e-07, + "loss": 0.5223, + "step": 14408 + }, + { + "epoch": 1.227865360034086, + "grad_norm": 18.625, + "learning_rate": 8.181617349417949e-07, + "loss": 0.7838, + "step": 14409 + }, + { + "epoch": 1.227950575202386, + "grad_norm": 18.25, + "learning_rate": 8.180040493148193e-07, + "loss": 0.6126, + "step": 14410 + }, + { + "epoch": 1.228035790370686, + "grad_norm": 18.25, + "learning_rate": 8.178463714944452e-07, + "loss": 0.7033, + "step": 14411 + }, + { + "epoch": 1.228121005538986, + "grad_norm": 17.125, + "learning_rate": 8.176887014835222e-07, + "loss": 0.4473, + "step": 14412 + }, + { + "epoch": 1.228206220707286, + "grad_norm": 21.125, + "learning_rate": 8.175310392848984e-07, + "loss": 0.7426, + "step": 14413 + }, + { + "epoch": 1.2282914358755859, + "grad_norm": 14.5, + "learning_rate": 8.173733849014238e-07, + "loss": 0.7649, + "step": 14414 + }, + { + "epoch": 1.2283766510438858, + "grad_norm": 15.875, + "learning_rate": 8.17215738335947e-07, + "loss": 0.8257, + "step": 14415 + }, + { + "epoch": 1.2284618662121858, + "grad_norm": 15.125, + "learning_rate": 8.170580995913173e-07, + "loss": 0.5833, + "step": 14416 + }, + { + "epoch": 1.2285470813804857, + "grad_norm": 15.9375, + "learning_rate": 8.169004686703828e-07, + "loss": 0.7669, + "step": 14417 + }, + { + "epoch": 1.2286322965487857, + "grad_norm": 19.625, + "learning_rate": 8.167428455759926e-07, + "loss": 0.6515, + "step": 14418 + }, + { + "epoch": 1.2287175117170857, + "grad_norm": 13.875, + "learning_rate": 8.165852303109946e-07, + "loss": 0.4103, + "step": 14419 + }, + { + "epoch": 1.2288027268853856, + "grad_norm": 12.3125, + "learning_rate": 8.164276228782375e-07, + "loss": 0.2408, + "step": 14420 + }, + { + "epoch": 1.2288879420536856, + "grad_norm": 15.5, + "learning_rate": 8.162700232805696e-07, + "loss": 0.73, + "step": 14421 + }, + { + "epoch": 1.2289731572219855, + "grad_norm": 31.25, + "learning_rate": 8.161124315208382e-07, + "loss": 0.8404, + "step": 14422 + }, + { + "epoch": 1.2290583723902855, + "grad_norm": 12.3125, + "learning_rate": 8.159548476018913e-07, + "loss": 0.387, + "step": 14423 + }, + { + "epoch": 1.2291435875585854, + "grad_norm": 15.1875, + "learning_rate": 8.157972715265771e-07, + "loss": 0.6301, + "step": 14424 + }, + { + "epoch": 1.2292288027268854, + "grad_norm": 11.375, + "learning_rate": 8.156397032977426e-07, + "loss": 0.2258, + "step": 14425 + }, + { + "epoch": 1.2293140178951854, + "grad_norm": 18.75, + "learning_rate": 8.154821429182355e-07, + "loss": 1.0408, + "step": 14426 + }, + { + "epoch": 1.2293992330634853, + "grad_norm": 13.875, + "learning_rate": 8.153245903909026e-07, + "loss": 0.607, + "step": 14427 + }, + { + "epoch": 1.2294844482317853, + "grad_norm": 10.625, + "learning_rate": 8.15167045718592e-07, + "loss": 0.2481, + "step": 14428 + }, + { + "epoch": 1.2295696634000852, + "grad_norm": 26.625, + "learning_rate": 8.150095089041504e-07, + "loss": 1.0483, + "step": 14429 + }, + { + "epoch": 1.2296548785683852, + "grad_norm": 17.125, + "learning_rate": 8.148519799504238e-07, + "loss": 0.5979, + "step": 14430 + }, + { + "epoch": 1.2297400937366851, + "grad_norm": 14.4375, + "learning_rate": 8.146944588602598e-07, + "loss": 0.7043, + "step": 14431 + }, + { + "epoch": 1.229825308904985, + "grad_norm": 14.5625, + "learning_rate": 8.145369456365044e-07, + "loss": 0.4015, + "step": 14432 + }, + { + "epoch": 1.229910524073285, + "grad_norm": 13.0, + "learning_rate": 8.143794402820044e-07, + "loss": 0.4216, + "step": 14433 + }, + { + "epoch": 1.229995739241585, + "grad_norm": 13.5625, + "learning_rate": 8.142219427996059e-07, + "loss": 0.4798, + "step": 14434 + }, + { + "epoch": 1.230080954409885, + "grad_norm": 15.875, + "learning_rate": 8.140644531921553e-07, + "loss": 0.6938, + "step": 14435 + }, + { + "epoch": 1.230166169578185, + "grad_norm": 15.1875, + "learning_rate": 8.139069714624981e-07, + "loss": 0.5764, + "step": 14436 + }, + { + "epoch": 1.2302513847464849, + "grad_norm": 11.125, + "learning_rate": 8.137494976134813e-07, + "loss": 0.3643, + "step": 14437 + }, + { + "epoch": 1.2303365999147848, + "grad_norm": 14.9375, + "learning_rate": 8.13592031647949e-07, + "loss": 0.5891, + "step": 14438 + }, + { + "epoch": 1.2304218150830848, + "grad_norm": 15.5625, + "learning_rate": 8.134345735687476e-07, + "loss": 0.6498, + "step": 14439 + }, + { + "epoch": 1.2305070302513847, + "grad_norm": 17.0, + "learning_rate": 8.132771233787223e-07, + "loss": 0.7822, + "step": 14440 + }, + { + "epoch": 1.2305922454196847, + "grad_norm": 18.125, + "learning_rate": 8.131196810807189e-07, + "loss": 0.494, + "step": 14441 + }, + { + "epoch": 1.2306774605879847, + "grad_norm": 17.25, + "learning_rate": 8.129622466775819e-07, + "loss": 0.7106, + "step": 14442 + }, + { + "epoch": 1.2307626757562846, + "grad_norm": 17.25, + "learning_rate": 8.128048201721567e-07, + "loss": 0.7046, + "step": 14443 + }, + { + "epoch": 1.2308478909245846, + "grad_norm": 15.5, + "learning_rate": 8.126474015672878e-07, + "loss": 0.7518, + "step": 14444 + }, + { + "epoch": 1.2309331060928845, + "grad_norm": 15.4375, + "learning_rate": 8.124899908658207e-07, + "loss": 0.6698, + "step": 14445 + }, + { + "epoch": 1.2310183212611845, + "grad_norm": 15.4375, + "learning_rate": 8.123325880705992e-07, + "loss": 0.7122, + "step": 14446 + }, + { + "epoch": 1.2311035364294844, + "grad_norm": 17.625, + "learning_rate": 8.121751931844676e-07, + "loss": 0.6943, + "step": 14447 + }, + { + "epoch": 1.2311887515977844, + "grad_norm": 20.5, + "learning_rate": 8.12017806210271e-07, + "loss": 0.8931, + "step": 14448 + }, + { + "epoch": 1.2312739667660844, + "grad_norm": 22.625, + "learning_rate": 8.118604271508526e-07, + "loss": 1.0771, + "step": 14449 + }, + { + "epoch": 1.2313591819343843, + "grad_norm": 15.5, + "learning_rate": 8.117030560090572e-07, + "loss": 0.7453, + "step": 14450 + }, + { + "epoch": 1.2314443971026843, + "grad_norm": 16.375, + "learning_rate": 8.11545692787728e-07, + "loss": 0.4901, + "step": 14451 + }, + { + "epoch": 1.2315296122709842, + "grad_norm": 19.625, + "learning_rate": 8.113883374897093e-07, + "loss": 0.8172, + "step": 14452 + }, + { + "epoch": 1.2316148274392842, + "grad_norm": 13.875, + "learning_rate": 8.112309901178448e-07, + "loss": 0.5826, + "step": 14453 + }, + { + "epoch": 1.2317000426075841, + "grad_norm": 27.0, + "learning_rate": 8.110736506749769e-07, + "loss": 0.7465, + "step": 14454 + }, + { + "epoch": 1.231785257775884, + "grad_norm": 20.25, + "learning_rate": 8.109163191639496e-07, + "loss": 0.5639, + "step": 14455 + }, + { + "epoch": 1.231870472944184, + "grad_norm": 14.8125, + "learning_rate": 8.107589955876061e-07, + "loss": 0.5941, + "step": 14456 + }, + { + "epoch": 1.231955688112484, + "grad_norm": 18.625, + "learning_rate": 8.10601679948789e-07, + "loss": 0.8677, + "step": 14457 + }, + { + "epoch": 1.232040903280784, + "grad_norm": 10.375, + "learning_rate": 8.104443722503416e-07, + "loss": 0.3252, + "step": 14458 + }, + { + "epoch": 1.232126118449084, + "grad_norm": 13.9375, + "learning_rate": 8.102870724951062e-07, + "loss": 0.4186, + "step": 14459 + }, + { + "epoch": 1.2322113336173839, + "grad_norm": 21.625, + "learning_rate": 8.101297806859257e-07, + "loss": 0.6037, + "step": 14460 + }, + { + "epoch": 1.2322965487856838, + "grad_norm": 14.875, + "learning_rate": 8.099724968256429e-07, + "loss": 0.4335, + "step": 14461 + }, + { + "epoch": 1.2323817639539838, + "grad_norm": 19.0, + "learning_rate": 8.098152209170988e-07, + "loss": 1.0173, + "step": 14462 + }, + { + "epoch": 1.2324669791222838, + "grad_norm": 10.6875, + "learning_rate": 8.096579529631367e-07, + "loss": 0.3258, + "step": 14463 + }, + { + "epoch": 1.2325521942905837, + "grad_norm": 15.875, + "learning_rate": 8.095006929665979e-07, + "loss": 0.7742, + "step": 14464 + }, + { + "epoch": 1.2326374094588837, + "grad_norm": 12.75, + "learning_rate": 8.093434409303247e-07, + "loss": 0.437, + "step": 14465 + }, + { + "epoch": 1.2327226246271836, + "grad_norm": 13.75, + "learning_rate": 8.091861968571585e-07, + "loss": 0.5079, + "step": 14466 + }, + { + "epoch": 1.2328078397954836, + "grad_norm": 13.0, + "learning_rate": 8.090289607499412e-07, + "loss": 0.4581, + "step": 14467 + }, + { + "epoch": 1.2328930549637835, + "grad_norm": 16.625, + "learning_rate": 8.088717326115137e-07, + "loss": 0.7145, + "step": 14468 + }, + { + "epoch": 1.2329782701320835, + "grad_norm": 19.625, + "learning_rate": 8.087145124447185e-07, + "loss": 0.713, + "step": 14469 + }, + { + "epoch": 1.2330634853003835, + "grad_norm": 11.4375, + "learning_rate": 8.085573002523948e-07, + "loss": 0.8115, + "step": 14470 + }, + { + "epoch": 1.2331487004686834, + "grad_norm": 15.8125, + "learning_rate": 8.084000960373851e-07, + "loss": 0.6105, + "step": 14471 + }, + { + "epoch": 1.2332339156369834, + "grad_norm": 11.1875, + "learning_rate": 8.082428998025293e-07, + "loss": 0.3658, + "step": 14472 + }, + { + "epoch": 1.2333191308052833, + "grad_norm": 23.5, + "learning_rate": 8.080857115506688e-07, + "loss": 0.8776, + "step": 14473 + }, + { + "epoch": 1.2334043459735833, + "grad_norm": 12.9375, + "learning_rate": 8.079285312846437e-07, + "loss": 0.3254, + "step": 14474 + }, + { + "epoch": 1.2334895611418832, + "grad_norm": 14.1875, + "learning_rate": 8.077713590072947e-07, + "loss": 0.5601, + "step": 14475 + }, + { + "epoch": 1.2335747763101832, + "grad_norm": 10.9375, + "learning_rate": 8.076141947214619e-07, + "loss": 0.5177, + "step": 14476 + }, + { + "epoch": 1.2336599914784832, + "grad_norm": 11.9375, + "learning_rate": 8.074570384299857e-07, + "loss": 0.3548, + "step": 14477 + }, + { + "epoch": 1.233745206646783, + "grad_norm": 20.125, + "learning_rate": 8.072998901357057e-07, + "loss": 0.7749, + "step": 14478 + }, + { + "epoch": 1.233830421815083, + "grad_norm": 16.375, + "learning_rate": 8.071427498414616e-07, + "loss": 0.5837, + "step": 14479 + }, + { + "epoch": 1.233915636983383, + "grad_norm": 19.5, + "learning_rate": 8.069856175500937e-07, + "loss": 0.9381, + "step": 14480 + }, + { + "epoch": 1.234000852151683, + "grad_norm": 11.8125, + "learning_rate": 8.068284932644407e-07, + "loss": 0.529, + "step": 14481 + }, + { + "epoch": 1.234086067319983, + "grad_norm": 16.625, + "learning_rate": 8.06671376987343e-07, + "loss": 0.7009, + "step": 14482 + }, + { + "epoch": 1.234171282488283, + "grad_norm": 17.375, + "learning_rate": 8.065142687216387e-07, + "loss": 0.7, + "step": 14483 + }, + { + "epoch": 1.2342564976565829, + "grad_norm": 16.75, + "learning_rate": 8.063571684701682e-07, + "loss": 0.6999, + "step": 14484 + }, + { + "epoch": 1.2343417128248828, + "grad_norm": 15.4375, + "learning_rate": 8.062000762357697e-07, + "loss": 0.5865, + "step": 14485 + }, + { + "epoch": 1.2344269279931828, + "grad_norm": 11.25, + "learning_rate": 8.06042992021282e-07, + "loss": 0.4193, + "step": 14486 + }, + { + "epoch": 1.2345121431614827, + "grad_norm": 13.625, + "learning_rate": 8.058859158295435e-07, + "loss": 0.4003, + "step": 14487 + }, + { + "epoch": 1.2345973583297827, + "grad_norm": 16.875, + "learning_rate": 8.057288476633935e-07, + "loss": 0.6689, + "step": 14488 + }, + { + "epoch": 1.2346825734980826, + "grad_norm": 22.125, + "learning_rate": 8.055717875256697e-07, + "loss": 0.7927, + "step": 14489 + }, + { + "epoch": 1.2347677886663826, + "grad_norm": 15.25, + "learning_rate": 8.054147354192107e-07, + "loss": 0.4423, + "step": 14490 + }, + { + "epoch": 1.2348530038346825, + "grad_norm": 15.5, + "learning_rate": 8.052576913468544e-07, + "loss": 0.5075, + "step": 14491 + }, + { + "epoch": 1.2349382190029825, + "grad_norm": 16.625, + "learning_rate": 8.051006553114391e-07, + "loss": 0.5374, + "step": 14492 + }, + { + "epoch": 1.2350234341712825, + "grad_norm": 15.625, + "learning_rate": 8.04943627315802e-07, + "loss": 0.7394, + "step": 14493 + }, + { + "epoch": 1.2351086493395824, + "grad_norm": 15.25, + "learning_rate": 8.047866073627816e-07, + "loss": 0.4514, + "step": 14494 + }, + { + "epoch": 1.2351938645078824, + "grad_norm": 17.25, + "learning_rate": 8.046295954552147e-07, + "loss": 0.7461, + "step": 14495 + }, + { + "epoch": 1.2352790796761823, + "grad_norm": 17.25, + "learning_rate": 8.044725915959387e-07, + "loss": 0.8528, + "step": 14496 + }, + { + "epoch": 1.2353642948444823, + "grad_norm": 13.1875, + "learning_rate": 8.04315595787791e-07, + "loss": 0.2919, + "step": 14497 + }, + { + "epoch": 1.2354495100127822, + "grad_norm": 15.6875, + "learning_rate": 8.041586080336085e-07, + "loss": 0.5813, + "step": 14498 + }, + { + "epoch": 1.2355347251810822, + "grad_norm": 14.4375, + "learning_rate": 8.040016283362286e-07, + "loss": 0.6932, + "step": 14499 + }, + { + "epoch": 1.2356199403493822, + "grad_norm": 13.8125, + "learning_rate": 8.038446566984876e-07, + "loss": 0.3093, + "step": 14500 + }, + { + "epoch": 1.2357051555176821, + "grad_norm": 22.375, + "learning_rate": 8.036876931232222e-07, + "loss": 0.9861, + "step": 14501 + }, + { + "epoch": 1.235790370685982, + "grad_norm": 14.4375, + "learning_rate": 8.035307376132695e-07, + "loss": 0.5617, + "step": 14502 + }, + { + "epoch": 1.235875585854282, + "grad_norm": 15.5, + "learning_rate": 8.03373790171465e-07, + "loss": 0.8113, + "step": 14503 + }, + { + "epoch": 1.235960801022582, + "grad_norm": 16.625, + "learning_rate": 8.032168508006451e-07, + "loss": 0.743, + "step": 14504 + }, + { + "epoch": 1.236046016190882, + "grad_norm": 19.875, + "learning_rate": 8.030599195036462e-07, + "loss": 0.6202, + "step": 14505 + }, + { + "epoch": 1.236131231359182, + "grad_norm": 21.125, + "learning_rate": 8.029029962833037e-07, + "loss": 0.9888, + "step": 14506 + }, + { + "epoch": 1.2362164465274819, + "grad_norm": 11.5625, + "learning_rate": 8.027460811424537e-07, + "loss": 0.3226, + "step": 14507 + }, + { + "epoch": 1.2363016616957818, + "grad_norm": 15.75, + "learning_rate": 8.025891740839317e-07, + "loss": 0.7289, + "step": 14508 + }, + { + "epoch": 1.2363868768640818, + "grad_norm": 15.5625, + "learning_rate": 8.024322751105736e-07, + "loss": 0.9043, + "step": 14509 + }, + { + "epoch": 1.2364720920323817, + "grad_norm": 20.625, + "learning_rate": 8.022753842252144e-07, + "loss": 0.8845, + "step": 14510 + }, + { + "epoch": 1.2365573072006817, + "grad_norm": 16.875, + "learning_rate": 8.021185014306888e-07, + "loss": 0.5755, + "step": 14511 + }, + { + "epoch": 1.2366425223689816, + "grad_norm": 16.5, + "learning_rate": 8.019616267298324e-07, + "loss": 0.4918, + "step": 14512 + }, + { + "epoch": 1.2367277375372816, + "grad_norm": 25.375, + "learning_rate": 8.0180476012548e-07, + "loss": 0.615, + "step": 14513 + }, + { + "epoch": 1.2368129527055816, + "grad_norm": 14.5, + "learning_rate": 8.016479016204662e-07, + "loss": 0.7477, + "step": 14514 + }, + { + "epoch": 1.2368981678738815, + "grad_norm": 16.875, + "learning_rate": 8.014910512176258e-07, + "loss": 0.7153, + "step": 14515 + }, + { + "epoch": 1.2369833830421815, + "grad_norm": 15.5, + "learning_rate": 8.013342089197928e-07, + "loss": 0.5963, + "step": 14516 + }, + { + "epoch": 1.2370685982104814, + "grad_norm": 16.75, + "learning_rate": 8.011773747298022e-07, + "loss": 0.4796, + "step": 14517 + }, + { + "epoch": 1.2371538133787814, + "grad_norm": 25.875, + "learning_rate": 8.01020548650488e-07, + "loss": 0.8417, + "step": 14518 + }, + { + "epoch": 1.2372390285470813, + "grad_norm": 15.1875, + "learning_rate": 8.008637306846834e-07, + "loss": 0.7155, + "step": 14519 + }, + { + "epoch": 1.2373242437153813, + "grad_norm": 13.25, + "learning_rate": 8.00706920835223e-07, + "loss": 0.7163, + "step": 14520 + }, + { + "epoch": 1.2374094588836813, + "grad_norm": 18.25, + "learning_rate": 8.005501191049403e-07, + "loss": 0.5048, + "step": 14521 + }, + { + "epoch": 1.2374946740519812, + "grad_norm": 15.0625, + "learning_rate": 8.00393325496669e-07, + "loss": 0.5039, + "step": 14522 + }, + { + "epoch": 1.2375798892202812, + "grad_norm": 13.3125, + "learning_rate": 8.002365400132423e-07, + "loss": 0.5394, + "step": 14523 + }, + { + "epoch": 1.2376651043885811, + "grad_norm": 12.1875, + "learning_rate": 8.000797626574938e-07, + "loss": 0.2966, + "step": 14524 + }, + { + "epoch": 1.237750319556881, + "grad_norm": 17.75, + "learning_rate": 7.999229934322561e-07, + "loss": 0.6193, + "step": 14525 + }, + { + "epoch": 1.237835534725181, + "grad_norm": 17.375, + "learning_rate": 7.997662323403629e-07, + "loss": 0.813, + "step": 14526 + }, + { + "epoch": 1.237920749893481, + "grad_norm": 25.25, + "learning_rate": 7.996094793846465e-07, + "loss": 0.6272, + "step": 14527 + }, + { + "epoch": 1.238005965061781, + "grad_norm": 16.5, + "learning_rate": 7.994527345679395e-07, + "loss": 0.6968, + "step": 14528 + }, + { + "epoch": 1.238091180230081, + "grad_norm": 14.0625, + "learning_rate": 7.992959978930749e-07, + "loss": 0.4154, + "step": 14529 + }, + { + "epoch": 1.2381763953983809, + "grad_norm": 13.375, + "learning_rate": 7.991392693628848e-07, + "loss": 0.34, + "step": 14530 + }, + { + "epoch": 1.2382616105666808, + "grad_norm": 12.625, + "learning_rate": 7.989825489802011e-07, + "loss": 0.3067, + "step": 14531 + }, + { + "epoch": 1.2383468257349808, + "grad_norm": 13.125, + "learning_rate": 7.988258367478568e-07, + "loss": 0.5455, + "step": 14532 + }, + { + "epoch": 1.2384320409032807, + "grad_norm": 17.0, + "learning_rate": 7.98669132668683e-07, + "loss": 0.6555, + "step": 14533 + }, + { + "epoch": 1.2385172560715807, + "grad_norm": 15.5625, + "learning_rate": 7.985124367455122e-07, + "loss": 0.7298, + "step": 14534 + }, + { + "epoch": 1.2386024712398807, + "grad_norm": 23.125, + "learning_rate": 7.983557489811754e-07, + "loss": 1.0131, + "step": 14535 + }, + { + "epoch": 1.2386876864081806, + "grad_norm": 24.75, + "learning_rate": 7.981990693785041e-07, + "loss": 0.7111, + "step": 14536 + }, + { + "epoch": 1.2387729015764806, + "grad_norm": 10.5, + "learning_rate": 7.980423979403304e-07, + "loss": 0.2446, + "step": 14537 + }, + { + "epoch": 1.2388581167447805, + "grad_norm": 24.125, + "learning_rate": 7.978857346694846e-07, + "loss": 1.0147, + "step": 14538 + }, + { + "epoch": 1.2389433319130805, + "grad_norm": 14.625, + "learning_rate": 7.977290795687985e-07, + "loss": 0.739, + "step": 14539 + }, + { + "epoch": 1.2390285470813804, + "grad_norm": 15.0, + "learning_rate": 7.975724326411025e-07, + "loss": 0.3423, + "step": 14540 + }, + { + "epoch": 1.2391137622496804, + "grad_norm": 8.9375, + "learning_rate": 7.97415793889228e-07, + "loss": 0.2164, + "step": 14541 + }, + { + "epoch": 1.2391989774179804, + "grad_norm": 20.125, + "learning_rate": 7.972591633160051e-07, + "loss": 0.4979, + "step": 14542 + }, + { + "epoch": 1.2392841925862803, + "grad_norm": 9.9375, + "learning_rate": 7.971025409242641e-07, + "loss": 0.9076, + "step": 14543 + }, + { + "epoch": 1.2393694077545803, + "grad_norm": 17.25, + "learning_rate": 7.96945926716836e-07, + "loss": 0.4437, + "step": 14544 + }, + { + "epoch": 1.2394546229228802, + "grad_norm": 13.0625, + "learning_rate": 7.967893206965505e-07, + "loss": 0.2974, + "step": 14545 + }, + { + "epoch": 1.2395398380911802, + "grad_norm": 37.25, + "learning_rate": 7.966327228662373e-07, + "loss": 1.3901, + "step": 14546 + }, + { + "epoch": 1.2396250532594801, + "grad_norm": 14.375, + "learning_rate": 7.964761332287271e-07, + "loss": 0.6615, + "step": 14547 + }, + { + "epoch": 1.23971026842778, + "grad_norm": 18.875, + "learning_rate": 7.963195517868488e-07, + "loss": 0.7583, + "step": 14548 + }, + { + "epoch": 1.23979548359608, + "grad_norm": 12.5, + "learning_rate": 7.961629785434329e-07, + "loss": 0.3806, + "step": 14549 + }, + { + "epoch": 1.23988069876438, + "grad_norm": 12.375, + "learning_rate": 7.960064135013085e-07, + "loss": 0.3351, + "step": 14550 + }, + { + "epoch": 1.23996591393268, + "grad_norm": 12.0625, + "learning_rate": 7.958498566633041e-07, + "loss": 0.4748, + "step": 14551 + }, + { + "epoch": 1.24005112910098, + "grad_norm": 13.1875, + "learning_rate": 7.956933080322499e-07, + "loss": 0.4207, + "step": 14552 + }, + { + "epoch": 1.2401363442692799, + "grad_norm": 13.6875, + "learning_rate": 7.955367676109742e-07, + "loss": 0.4808, + "step": 14553 + }, + { + "epoch": 1.2402215594375798, + "grad_norm": 14.125, + "learning_rate": 7.953802354023063e-07, + "loss": 0.3672, + "step": 14554 + }, + { + "epoch": 1.2403067746058798, + "grad_norm": 14.0625, + "learning_rate": 7.952237114090744e-07, + "loss": 0.5251, + "step": 14555 + }, + { + "epoch": 1.2403919897741797, + "grad_norm": 12.9375, + "learning_rate": 7.950671956341077e-07, + "loss": 0.5081, + "step": 14556 + }, + { + "epoch": 1.2404772049424797, + "grad_norm": 24.25, + "learning_rate": 7.949106880802341e-07, + "loss": 0.9129, + "step": 14557 + }, + { + "epoch": 1.2405624201107797, + "grad_norm": 14.8125, + "learning_rate": 7.94754188750282e-07, + "loss": 0.5264, + "step": 14558 + }, + { + "epoch": 1.2406476352790796, + "grad_norm": 15.1875, + "learning_rate": 7.945976976470798e-07, + "loss": 0.519, + "step": 14559 + }, + { + "epoch": 1.2407328504473796, + "grad_norm": 14.8125, + "learning_rate": 7.94441214773455e-07, + "loss": 0.4413, + "step": 14560 + }, + { + "epoch": 1.2408180656156795, + "grad_norm": 15.3125, + "learning_rate": 7.942847401322353e-07, + "loss": 0.4321, + "step": 14561 + }, + { + "epoch": 1.2409032807839795, + "grad_norm": 15.8125, + "learning_rate": 7.941282737262487e-07, + "loss": 0.6936, + "step": 14562 + }, + { + "epoch": 1.2409884959522794, + "grad_norm": 17.5, + "learning_rate": 7.939718155583225e-07, + "loss": 0.6167, + "step": 14563 + }, + { + "epoch": 1.2410737111205794, + "grad_norm": 17.0, + "learning_rate": 7.938153656312844e-07, + "loss": 0.7481, + "step": 14564 + }, + { + "epoch": 1.2411589262888794, + "grad_norm": 20.25, + "learning_rate": 7.936589239479611e-07, + "loss": 1.0365, + "step": 14565 + }, + { + "epoch": 1.2412441414571793, + "grad_norm": 18.875, + "learning_rate": 7.935024905111802e-07, + "loss": 0.7845, + "step": 14566 + }, + { + "epoch": 1.2413293566254793, + "grad_norm": 29.75, + "learning_rate": 7.933460653237684e-07, + "loss": 0.9666, + "step": 14567 + }, + { + "epoch": 1.2414145717937792, + "grad_norm": 15.9375, + "learning_rate": 7.93189648388552e-07, + "loss": 0.7358, + "step": 14568 + }, + { + "epoch": 1.2414997869620792, + "grad_norm": 19.0, + "learning_rate": 7.930332397083585e-07, + "loss": 0.4884, + "step": 14569 + }, + { + "epoch": 1.2415850021303791, + "grad_norm": 14.5, + "learning_rate": 7.928768392860134e-07, + "loss": 0.3494, + "step": 14570 + }, + { + "epoch": 1.241670217298679, + "grad_norm": 18.625, + "learning_rate": 7.927204471243438e-07, + "loss": 0.6889, + "step": 14571 + }, + { + "epoch": 1.241755432466979, + "grad_norm": 14.1875, + "learning_rate": 7.925640632261753e-07, + "loss": 0.3671, + "step": 14572 + }, + { + "epoch": 1.241840647635279, + "grad_norm": 20.875, + "learning_rate": 7.924076875943343e-07, + "loss": 0.6434, + "step": 14573 + }, + { + "epoch": 1.241925862803579, + "grad_norm": 20.625, + "learning_rate": 7.922513202316465e-07, + "loss": 1.2731, + "step": 14574 + }, + { + "epoch": 1.242011077971879, + "grad_norm": 31.25, + "learning_rate": 7.920949611409382e-07, + "loss": 0.8002, + "step": 14575 + }, + { + "epoch": 1.2420962931401789, + "grad_norm": 14.9375, + "learning_rate": 7.919386103250337e-07, + "loss": 0.8563, + "step": 14576 + }, + { + "epoch": 1.2421815083084788, + "grad_norm": 9.25, + "learning_rate": 7.917822677867595e-07, + "loss": 0.2108, + "step": 14577 + }, + { + "epoch": 1.2422667234767788, + "grad_norm": 16.125, + "learning_rate": 7.916259335289401e-07, + "loss": 0.7071, + "step": 14578 + }, + { + "epoch": 1.2423519386450788, + "grad_norm": 30.625, + "learning_rate": 7.914696075544013e-07, + "loss": 0.6715, + "step": 14579 + }, + { + "epoch": 1.2424371538133787, + "grad_norm": 14.1875, + "learning_rate": 7.913132898659676e-07, + "loss": 0.6789, + "step": 14580 + }, + { + "epoch": 1.2425223689816787, + "grad_norm": 13.375, + "learning_rate": 7.911569804664643e-07, + "loss": 0.2534, + "step": 14581 + }, + { + "epoch": 1.2426075841499786, + "grad_norm": 15.4375, + "learning_rate": 7.910006793587152e-07, + "loss": 0.359, + "step": 14582 + }, + { + "epoch": 1.2426927993182786, + "grad_norm": 15.8125, + "learning_rate": 7.908443865455462e-07, + "loss": 0.7314, + "step": 14583 + }, + { + "epoch": 1.2427780144865785, + "grad_norm": 13.9375, + "learning_rate": 7.906881020297805e-07, + "loss": 0.6248, + "step": 14584 + }, + { + "epoch": 1.2428632296548785, + "grad_norm": 17.125, + "learning_rate": 7.905318258142425e-07, + "loss": 0.6161, + "step": 14585 + }, + { + "epoch": 1.2429484448231785, + "grad_norm": 16.875, + "learning_rate": 7.903755579017565e-07, + "loss": 0.5821, + "step": 14586 + }, + { + "epoch": 1.2430336599914784, + "grad_norm": 12.75, + "learning_rate": 7.902192982951462e-07, + "loss": 0.3288, + "step": 14587 + }, + { + "epoch": 1.2431188751597784, + "grad_norm": 19.25, + "learning_rate": 7.900630469972359e-07, + "loss": 1.1231, + "step": 14588 + }, + { + "epoch": 1.2432040903280783, + "grad_norm": 14.75, + "learning_rate": 7.899068040108484e-07, + "loss": 0.6788, + "step": 14589 + }, + { + "epoch": 1.2432893054963783, + "grad_norm": 13.125, + "learning_rate": 7.89750569338808e-07, + "loss": 0.4566, + "step": 14590 + }, + { + "epoch": 1.2433745206646782, + "grad_norm": 11.6875, + "learning_rate": 7.895943429839379e-07, + "loss": 0.3042, + "step": 14591 + }, + { + "epoch": 1.2434597358329782, + "grad_norm": 16.875, + "learning_rate": 7.894381249490605e-07, + "loss": 0.5243, + "step": 14592 + }, + { + "epoch": 1.2435449510012782, + "grad_norm": 14.75, + "learning_rate": 7.892819152369993e-07, + "loss": 0.4836, + "step": 14593 + }, + { + "epoch": 1.243630166169578, + "grad_norm": 12.8125, + "learning_rate": 7.891257138505773e-07, + "loss": 0.5922, + "step": 14594 + }, + { + "epoch": 1.243715381337878, + "grad_norm": 15.5625, + "learning_rate": 7.88969520792617e-07, + "loss": 0.4425, + "step": 14595 + }, + { + "epoch": 1.243800596506178, + "grad_norm": 13.625, + "learning_rate": 7.888133360659412e-07, + "loss": 0.4358, + "step": 14596 + }, + { + "epoch": 1.243885811674478, + "grad_norm": 18.0, + "learning_rate": 7.886571596733719e-07, + "loss": 0.5741, + "step": 14597 + }, + { + "epoch": 1.243971026842778, + "grad_norm": 10.6875, + "learning_rate": 7.88500991617732e-07, + "loss": 0.3579, + "step": 14598 + }, + { + "epoch": 1.244056242011078, + "grad_norm": 15.0625, + "learning_rate": 7.883448319018433e-07, + "loss": 0.6104, + "step": 14599 + }, + { + "epoch": 1.2441414571793779, + "grad_norm": 16.5, + "learning_rate": 7.881886805285274e-07, + "loss": 0.5722, + "step": 14600 + }, + { + "epoch": 1.2442266723476778, + "grad_norm": 18.0, + "learning_rate": 7.880325375006066e-07, + "loss": 0.8445, + "step": 14601 + }, + { + "epoch": 1.2443118875159778, + "grad_norm": 20.375, + "learning_rate": 7.878764028209023e-07, + "loss": 0.8089, + "step": 14602 + }, + { + "epoch": 1.2443971026842777, + "grad_norm": 13.8125, + "learning_rate": 7.877202764922362e-07, + "loss": 0.5218, + "step": 14603 + }, + { + "epoch": 1.2444823178525777, + "grad_norm": 26.0, + "learning_rate": 7.875641585174293e-07, + "loss": 1.1188, + "step": 14604 + }, + { + "epoch": 1.2445675330208776, + "grad_norm": 24.625, + "learning_rate": 7.874080488993033e-07, + "loss": 0.9768, + "step": 14605 + }, + { + "epoch": 1.2446527481891776, + "grad_norm": 15.9375, + "learning_rate": 7.872519476406789e-07, + "loss": 1.0397, + "step": 14606 + }, + { + "epoch": 1.2447379633574776, + "grad_norm": 15.8125, + "learning_rate": 7.870958547443779e-07, + "loss": 0.6703, + "step": 14607 + }, + { + "epoch": 1.2448231785257775, + "grad_norm": 22.5, + "learning_rate": 7.869397702132195e-07, + "loss": 0.6299, + "step": 14608 + }, + { + "epoch": 1.2449083936940775, + "grad_norm": 15.875, + "learning_rate": 7.867836940500252e-07, + "loss": 0.7937, + "step": 14609 + }, + { + "epoch": 1.2449936088623774, + "grad_norm": 12.5, + "learning_rate": 7.866276262576153e-07, + "loss": 0.2633, + "step": 14610 + }, + { + "epoch": 1.2450788240306774, + "grad_norm": 12.8125, + "learning_rate": 7.864715668388103e-07, + "loss": 0.4869, + "step": 14611 + }, + { + "epoch": 1.2451640391989773, + "grad_norm": 15.3125, + "learning_rate": 7.863155157964299e-07, + "loss": 0.6115, + "step": 14612 + }, + { + "epoch": 1.2452492543672773, + "grad_norm": 18.375, + "learning_rate": 7.861594731332948e-07, + "loss": 0.513, + "step": 14613 + }, + { + "epoch": 1.2453344695355772, + "grad_norm": 21.875, + "learning_rate": 7.860034388522239e-07, + "loss": 0.5601, + "step": 14614 + }, + { + "epoch": 1.2454196847038772, + "grad_norm": 19.25, + "learning_rate": 7.858474129560382e-07, + "loss": 0.7062, + "step": 14615 + }, + { + "epoch": 1.2455048998721772, + "grad_norm": 14.8125, + "learning_rate": 7.856913954475562e-07, + "loss": 0.4848, + "step": 14616 + }, + { + "epoch": 1.2455901150404771, + "grad_norm": 35.25, + "learning_rate": 7.855353863295974e-07, + "loss": 0.843, + "step": 14617 + }, + { + "epoch": 1.245675330208777, + "grad_norm": 21.375, + "learning_rate": 7.853793856049814e-07, + "loss": 0.7631, + "step": 14618 + }, + { + "epoch": 1.245760545377077, + "grad_norm": 15.5, + "learning_rate": 7.852233932765269e-07, + "loss": 0.7349, + "step": 14619 + }, + { + "epoch": 1.245845760545377, + "grad_norm": 13.25, + "learning_rate": 7.850674093470533e-07, + "loss": 0.5253, + "step": 14620 + }, + { + "epoch": 1.245930975713677, + "grad_norm": 15.75, + "learning_rate": 7.849114338193788e-07, + "loss": 0.5871, + "step": 14621 + }, + { + "epoch": 1.246016190881977, + "grad_norm": 14.0, + "learning_rate": 7.847554666963226e-07, + "loss": 0.8007, + "step": 14622 + }, + { + "epoch": 1.2461014060502769, + "grad_norm": 18.0, + "learning_rate": 7.845995079807031e-07, + "loss": 0.8773, + "step": 14623 + }, + { + "epoch": 1.2461866212185768, + "grad_norm": 12.625, + "learning_rate": 7.844435576753387e-07, + "loss": 0.4596, + "step": 14624 + }, + { + "epoch": 1.2462718363868768, + "grad_norm": 30.625, + "learning_rate": 7.842876157830467e-07, + "loss": 0.5978, + "step": 14625 + }, + { + "epoch": 1.2463570515551767, + "grad_norm": 14.3125, + "learning_rate": 7.841316823066463e-07, + "loss": 0.5941, + "step": 14626 + }, + { + "epoch": 1.2464422667234767, + "grad_norm": 16.75, + "learning_rate": 7.839757572489545e-07, + "loss": 0.6234, + "step": 14627 + }, + { + "epoch": 1.2465274818917766, + "grad_norm": 15.0, + "learning_rate": 7.838198406127896e-07, + "loss": 0.6071, + "step": 14628 + }, + { + "epoch": 1.2466126970600766, + "grad_norm": 12.0, + "learning_rate": 7.836639324009687e-07, + "loss": 0.2458, + "step": 14629 + }, + { + "epoch": 1.2466979122283766, + "grad_norm": 16.125, + "learning_rate": 7.8350803261631e-07, + "loss": 0.5096, + "step": 14630 + }, + { + "epoch": 1.2467831273966765, + "grad_norm": 12.375, + "learning_rate": 7.833521412616297e-07, + "loss": 0.3598, + "step": 14631 + }, + { + "epoch": 1.2468683425649765, + "grad_norm": 15.5625, + "learning_rate": 7.831962583397459e-07, + "loss": 0.6971, + "step": 14632 + }, + { + "epoch": 1.2469535577332764, + "grad_norm": 16.0, + "learning_rate": 7.830403838534751e-07, + "loss": 0.6217, + "step": 14633 + }, + { + "epoch": 1.2470387729015764, + "grad_norm": 14.625, + "learning_rate": 7.828845178056338e-07, + "loss": 0.502, + "step": 14634 + }, + { + "epoch": 1.2471239880698763, + "grad_norm": 13.0625, + "learning_rate": 7.827286601990391e-07, + "loss": 0.4965, + "step": 14635 + }, + { + "epoch": 1.2472092032381763, + "grad_norm": 15.125, + "learning_rate": 7.825728110365073e-07, + "loss": 0.7139, + "step": 14636 + }, + { + "epoch": 1.2472944184064763, + "grad_norm": 14.9375, + "learning_rate": 7.824169703208551e-07, + "loss": 0.7267, + "step": 14637 + }, + { + "epoch": 1.2473796335747762, + "grad_norm": 13.5, + "learning_rate": 7.822611380548983e-07, + "loss": 0.4705, + "step": 14638 + }, + { + "epoch": 1.2474648487430762, + "grad_norm": 13.6875, + "learning_rate": 7.82105314241453e-07, + "loss": 0.439, + "step": 14639 + }, + { + "epoch": 1.2475500639113761, + "grad_norm": 13.3125, + "learning_rate": 7.819494988833357e-07, + "loss": 0.3843, + "step": 14640 + }, + { + "epoch": 1.247635279079676, + "grad_norm": 16.125, + "learning_rate": 7.817936919833611e-07, + "loss": 0.7717, + "step": 14641 + }, + { + "epoch": 1.247720494247976, + "grad_norm": 16.125, + "learning_rate": 7.816378935443453e-07, + "loss": 0.7799, + "step": 14642 + }, + { + "epoch": 1.247805709416276, + "grad_norm": 18.75, + "learning_rate": 7.814821035691039e-07, + "loss": 0.5608, + "step": 14643 + }, + { + "epoch": 1.247890924584576, + "grad_norm": 16.875, + "learning_rate": 7.813263220604517e-07, + "loss": 0.5795, + "step": 14644 + }, + { + "epoch": 1.247976139752876, + "grad_norm": 9.75, + "learning_rate": 7.811705490212044e-07, + "loss": 0.2612, + "step": 14645 + }, + { + "epoch": 1.2480613549211759, + "grad_norm": 13.8125, + "learning_rate": 7.810147844541766e-07, + "loss": 0.6701, + "step": 14646 + }, + { + "epoch": 1.2481465700894758, + "grad_norm": 22.5, + "learning_rate": 7.808590283621834e-07, + "loss": 1.0441, + "step": 14647 + }, + { + "epoch": 1.2482317852577758, + "grad_norm": 18.75, + "learning_rate": 7.807032807480397e-07, + "loss": 0.8453, + "step": 14648 + }, + { + "epoch": 1.2483170004260757, + "grad_norm": 13.75, + "learning_rate": 7.805475416145589e-07, + "loss": 0.5625, + "step": 14649 + }, + { + "epoch": 1.2484022155943757, + "grad_norm": 14.8125, + "learning_rate": 7.803918109645566e-07, + "loss": 0.5269, + "step": 14650 + }, + { + "epoch": 1.2484874307626757, + "grad_norm": 17.375, + "learning_rate": 7.80236088800846e-07, + "loss": 0.5127, + "step": 14651 + }, + { + "epoch": 1.2485726459309756, + "grad_norm": 14.0, + "learning_rate": 7.80080375126242e-07, + "loss": 0.4743, + "step": 14652 + }, + { + "epoch": 1.2486578610992756, + "grad_norm": 17.25, + "learning_rate": 7.799246699435583e-07, + "loss": 0.5051, + "step": 14653 + }, + { + "epoch": 1.2487430762675755, + "grad_norm": 13.6875, + "learning_rate": 7.797689732556082e-07, + "loss": 0.2845, + "step": 14654 + }, + { + "epoch": 1.2488282914358755, + "grad_norm": 14.5, + "learning_rate": 7.796132850652058e-07, + "loss": 0.4436, + "step": 14655 + }, + { + "epoch": 1.2489135066041754, + "grad_norm": 17.125, + "learning_rate": 7.794576053751646e-07, + "loss": 0.6844, + "step": 14656 + }, + { + "epoch": 1.2489987217724754, + "grad_norm": 16.875, + "learning_rate": 7.79301934188297e-07, + "loss": 0.6172, + "step": 14657 + }, + { + "epoch": 1.2490839369407754, + "grad_norm": 16.5, + "learning_rate": 7.791462715074173e-07, + "loss": 0.919, + "step": 14658 + }, + { + "epoch": 1.2491691521090753, + "grad_norm": 17.875, + "learning_rate": 7.789906173353376e-07, + "loss": 0.6099, + "step": 14659 + }, + { + "epoch": 1.2492543672773753, + "grad_norm": 24.375, + "learning_rate": 7.788349716748714e-07, + "loss": 0.6801, + "step": 14660 + }, + { + "epoch": 1.2493395824456752, + "grad_norm": 14.875, + "learning_rate": 7.786793345288307e-07, + "loss": 0.628, + "step": 14661 + }, + { + "epoch": 1.2494247976139752, + "grad_norm": 12.25, + "learning_rate": 7.785237059000286e-07, + "loss": 0.7181, + "step": 14662 + }, + { + "epoch": 1.2495100127822751, + "grad_norm": 23.25, + "learning_rate": 7.78368085791277e-07, + "loss": 0.7345, + "step": 14663 + }, + { + "epoch": 1.249595227950575, + "grad_norm": 18.125, + "learning_rate": 7.78212474205389e-07, + "loss": 0.8377, + "step": 14664 + }, + { + "epoch": 1.249680443118875, + "grad_norm": 12.75, + "learning_rate": 7.780568711451756e-07, + "loss": 0.4615, + "step": 14665 + }, + { + "epoch": 1.249765658287175, + "grad_norm": 19.375, + "learning_rate": 7.779012766134489e-07, + "loss": 0.4634, + "step": 14666 + }, + { + "epoch": 1.249850873455475, + "grad_norm": 21.875, + "learning_rate": 7.777456906130211e-07, + "loss": 0.8786, + "step": 14667 + }, + { + "epoch": 1.249936088623775, + "grad_norm": 22.75, + "learning_rate": 7.775901131467036e-07, + "loss": 0.5075, + "step": 14668 + }, + { + "epoch": 1.2500213037920749, + "grad_norm": 26.25, + "learning_rate": 7.774345442173076e-07, + "loss": 0.8304, + "step": 14669 + }, + { + "epoch": 1.2501065189603748, + "grad_norm": 16.875, + "learning_rate": 7.772789838276445e-07, + "loss": 0.7385, + "step": 14670 + }, + { + "epoch": 1.2501917341286748, + "grad_norm": 16.5, + "learning_rate": 7.771234319805257e-07, + "loss": 0.7161, + "step": 14671 + }, + { + "epoch": 1.2502769492969747, + "grad_norm": 15.8125, + "learning_rate": 7.769678886787624e-07, + "loss": 0.7023, + "step": 14672 + }, + { + "epoch": 1.2503621644652747, + "grad_norm": 25.5, + "learning_rate": 7.768123539251646e-07, + "loss": 0.7773, + "step": 14673 + }, + { + "epoch": 1.2504473796335747, + "grad_norm": 15.1875, + "learning_rate": 7.766568277225433e-07, + "loss": 0.5063, + "step": 14674 + }, + { + "epoch": 1.2505325948018746, + "grad_norm": 16.5, + "learning_rate": 7.765013100737092e-07, + "loss": 0.6495, + "step": 14675 + }, + { + "epoch": 1.2506178099701746, + "grad_norm": 12.5625, + "learning_rate": 7.763458009814725e-07, + "loss": 0.481, + "step": 14676 + }, + { + "epoch": 1.2507030251384745, + "grad_norm": 20.125, + "learning_rate": 7.761903004486438e-07, + "loss": 0.7506, + "step": 14677 + }, + { + "epoch": 1.2507882403067745, + "grad_norm": 20.875, + "learning_rate": 7.760348084780323e-07, + "loss": 0.6135, + "step": 14678 + }, + { + "epoch": 1.2508734554750744, + "grad_norm": 32.25, + "learning_rate": 7.758793250724488e-07, + "loss": 0.7601, + "step": 14679 + }, + { + "epoch": 1.2509586706433744, + "grad_norm": 12.5625, + "learning_rate": 7.757238502347028e-07, + "loss": 0.4277, + "step": 14680 + }, + { + "epoch": 1.2510438858116744, + "grad_norm": 13.3125, + "learning_rate": 7.755683839676034e-07, + "loss": 0.4344, + "step": 14681 + }, + { + "epoch": 1.2511291009799743, + "grad_norm": 15.125, + "learning_rate": 7.754129262739607e-07, + "loss": 0.4443, + "step": 14682 + }, + { + "epoch": 1.2512143161482743, + "grad_norm": 14.8125, + "learning_rate": 7.752574771565835e-07, + "loss": 0.838, + "step": 14683 + }, + { + "epoch": 1.2512995313165742, + "grad_norm": 17.0, + "learning_rate": 7.751020366182807e-07, + "loss": 0.712, + "step": 14684 + }, + { + "epoch": 1.2513847464848742, + "grad_norm": 16.25, + "learning_rate": 7.749466046618622e-07, + "loss": 0.6198, + "step": 14685 + }, + { + "epoch": 1.2514699616531741, + "grad_norm": 16.625, + "learning_rate": 7.747911812901359e-07, + "loss": 0.6036, + "step": 14686 + }, + { + "epoch": 1.251555176821474, + "grad_norm": 27.125, + "learning_rate": 7.74635766505911e-07, + "loss": 0.6246, + "step": 14687 + }, + { + "epoch": 1.251640391989774, + "grad_norm": 17.375, + "learning_rate": 7.74480360311996e-07, + "loss": 0.8158, + "step": 14688 + }, + { + "epoch": 1.251725607158074, + "grad_norm": 21.75, + "learning_rate": 7.743249627111987e-07, + "loss": 0.7725, + "step": 14689 + }, + { + "epoch": 1.251810822326374, + "grad_norm": 20.875, + "learning_rate": 7.741695737063278e-07, + "loss": 0.8485, + "step": 14690 + }, + { + "epoch": 1.251896037494674, + "grad_norm": 15.4375, + "learning_rate": 7.74014193300191e-07, + "loss": 0.4862, + "step": 14691 + }, + { + "epoch": 1.2519812526629739, + "grad_norm": 19.125, + "learning_rate": 7.738588214955967e-07, + "loss": 0.6126, + "step": 14692 + }, + { + "epoch": 1.2520664678312738, + "grad_norm": 16.25, + "learning_rate": 7.737034582953519e-07, + "loss": 0.3129, + "step": 14693 + }, + { + "epoch": 1.2521516829995738, + "grad_norm": 15.4375, + "learning_rate": 7.73548103702265e-07, + "loss": 0.6435, + "step": 14694 + }, + { + "epoch": 1.2522368981678738, + "grad_norm": 17.5, + "learning_rate": 7.733927577191425e-07, + "loss": 0.8859, + "step": 14695 + }, + { + "epoch": 1.2523221133361737, + "grad_norm": 12.625, + "learning_rate": 7.732374203487925e-07, + "loss": 0.455, + "step": 14696 + }, + { + "epoch": 1.2524073285044737, + "grad_norm": 22.125, + "learning_rate": 7.73082091594022e-07, + "loss": 0.8317, + "step": 14697 + }, + { + "epoch": 1.2524925436727736, + "grad_norm": 15.9375, + "learning_rate": 7.729267714576374e-07, + "loss": 0.7693, + "step": 14698 + }, + { + "epoch": 1.2525777588410736, + "grad_norm": 17.75, + "learning_rate": 7.727714599424454e-07, + "loss": 0.7912, + "step": 14699 + }, + { + "epoch": 1.2526629740093735, + "grad_norm": 12.375, + "learning_rate": 7.726161570512534e-07, + "loss": 0.6755, + "step": 14700 + }, + { + "epoch": 1.2527481891776735, + "grad_norm": 17.75, + "learning_rate": 7.724608627868673e-07, + "loss": 0.6475, + "step": 14701 + }, + { + "epoch": 1.2528334043459735, + "grad_norm": 13.375, + "learning_rate": 7.723055771520938e-07, + "loss": 0.6995, + "step": 14702 + }, + { + "epoch": 1.2529186195142734, + "grad_norm": 18.25, + "learning_rate": 7.721503001497385e-07, + "loss": 0.6928, + "step": 14703 + }, + { + "epoch": 1.2530038346825734, + "grad_norm": 29.25, + "learning_rate": 7.719950317826082e-07, + "loss": 0.4541, + "step": 14704 + }, + { + "epoch": 1.2530890498508733, + "grad_norm": 14.875, + "learning_rate": 7.718397720535086e-07, + "loss": 0.6209, + "step": 14705 + }, + { + "epoch": 1.2531742650191733, + "grad_norm": 15.0625, + "learning_rate": 7.716845209652446e-07, + "loss": 0.7165, + "step": 14706 + }, + { + "epoch": 1.2532594801874732, + "grad_norm": 15.5, + "learning_rate": 7.715292785206224e-07, + "loss": 0.5183, + "step": 14707 + }, + { + "epoch": 1.2533446953557732, + "grad_norm": 22.0, + "learning_rate": 7.713740447224471e-07, + "loss": 0.7164, + "step": 14708 + }, + { + "epoch": 1.2534299105240732, + "grad_norm": 15.3125, + "learning_rate": 7.712188195735245e-07, + "loss": 0.4587, + "step": 14709 + }, + { + "epoch": 1.2535151256923731, + "grad_norm": 21.625, + "learning_rate": 7.710636030766589e-07, + "loss": 1.2628, + "step": 14710 + }, + { + "epoch": 1.253600340860673, + "grad_norm": 11.8125, + "learning_rate": 7.70908395234656e-07, + "loss": 0.3048, + "step": 14711 + }, + { + "epoch": 1.253685556028973, + "grad_norm": 18.5, + "learning_rate": 7.707531960503198e-07, + "loss": 0.4974, + "step": 14712 + }, + { + "epoch": 1.253770771197273, + "grad_norm": 18.125, + "learning_rate": 7.705980055264559e-07, + "loss": 0.7327, + "step": 14713 + }, + { + "epoch": 1.253855986365573, + "grad_norm": 14.3125, + "learning_rate": 7.704428236658676e-07, + "loss": 0.5049, + "step": 14714 + }, + { + "epoch": 1.253941201533873, + "grad_norm": 15.3125, + "learning_rate": 7.702876504713599e-07, + "loss": 0.6619, + "step": 14715 + }, + { + "epoch": 1.2540264167021729, + "grad_norm": 12.3125, + "learning_rate": 7.701324859457364e-07, + "loss": 0.5473, + "step": 14716 + }, + { + "epoch": 1.2541116318704728, + "grad_norm": 24.25, + "learning_rate": 7.699773300918017e-07, + "loss": 0.8938, + "step": 14717 + }, + { + "epoch": 1.2541968470387728, + "grad_norm": 14.8125, + "learning_rate": 7.698221829123592e-07, + "loss": 0.6725, + "step": 14718 + }, + { + "epoch": 1.2542820622070727, + "grad_norm": 13.4375, + "learning_rate": 7.696670444102128e-07, + "loss": 0.6757, + "step": 14719 + }, + { + "epoch": 1.2543672773753727, + "grad_norm": 19.625, + "learning_rate": 7.695119145881657e-07, + "loss": 0.4822, + "step": 14720 + }, + { + "epoch": 1.2544524925436726, + "grad_norm": 13.75, + "learning_rate": 7.693567934490221e-07, + "loss": 0.4454, + "step": 14721 + }, + { + "epoch": 1.2545377077119726, + "grad_norm": 18.75, + "learning_rate": 7.692016809955842e-07, + "loss": 0.682, + "step": 14722 + }, + { + "epoch": 1.2546229228802726, + "grad_norm": 17.125, + "learning_rate": 7.690465772306552e-07, + "loss": 1.0014, + "step": 14723 + }, + { + "epoch": 1.2547081380485725, + "grad_norm": 14.375, + "learning_rate": 7.688914821570384e-07, + "loss": 0.6769, + "step": 14724 + }, + { + "epoch": 1.2547933532168725, + "grad_norm": 19.0, + "learning_rate": 7.68736395777536e-07, + "loss": 0.594, + "step": 14725 + }, + { + "epoch": 1.2548785683851726, + "grad_norm": 20.125, + "learning_rate": 7.685813180949514e-07, + "loss": 0.8626, + "step": 14726 + }, + { + "epoch": 1.2549637835534726, + "grad_norm": 21.875, + "learning_rate": 7.684262491120858e-07, + "loss": 0.669, + "step": 14727 + }, + { + "epoch": 1.2550489987217726, + "grad_norm": 15.5, + "learning_rate": 7.682711888317427e-07, + "loss": 0.6762, + "step": 14728 + }, + { + "epoch": 1.2551342138900725, + "grad_norm": 20.125, + "learning_rate": 7.681161372567239e-07, + "loss": 0.4479, + "step": 14729 + }, + { + "epoch": 1.2552194290583725, + "grad_norm": 14.0625, + "learning_rate": 7.679610943898308e-07, + "loss": 0.5165, + "step": 14730 + }, + { + "epoch": 1.2553046442266724, + "grad_norm": 15.1875, + "learning_rate": 7.678060602338651e-07, + "loss": 0.7642, + "step": 14731 + }, + { + "epoch": 1.2553898593949724, + "grad_norm": 15.125, + "learning_rate": 7.676510347916291e-07, + "loss": 0.3226, + "step": 14732 + }, + { + "epoch": 1.2554750745632723, + "grad_norm": 16.25, + "learning_rate": 7.674960180659235e-07, + "loss": 0.3955, + "step": 14733 + }, + { + "epoch": 1.2555602897315723, + "grad_norm": 15.4375, + "learning_rate": 7.673410100595505e-07, + "loss": 0.7236, + "step": 14734 + }, + { + "epoch": 1.2556455048998723, + "grad_norm": 15.75, + "learning_rate": 7.671860107753105e-07, + "loss": 0.6195, + "step": 14735 + }, + { + "epoch": 1.2557307200681722, + "grad_norm": 18.0, + "learning_rate": 7.670310202160052e-07, + "loss": 0.6392, + "step": 14736 + }, + { + "epoch": 1.2558159352364722, + "grad_norm": 12.0, + "learning_rate": 7.66876038384435e-07, + "loss": 0.2498, + "step": 14737 + }, + { + "epoch": 1.2559011504047721, + "grad_norm": 22.25, + "learning_rate": 7.667210652834004e-07, + "loss": 1.0402, + "step": 14738 + }, + { + "epoch": 1.255986365573072, + "grad_norm": 13.5, + "learning_rate": 7.665661009157022e-07, + "loss": 0.5617, + "step": 14739 + }, + { + "epoch": 1.256071580741372, + "grad_norm": 17.75, + "learning_rate": 7.664111452841402e-07, + "loss": 0.5832, + "step": 14740 + }, + { + "epoch": 1.256156795909672, + "grad_norm": 18.625, + "learning_rate": 7.662561983915157e-07, + "loss": 0.9422, + "step": 14741 + }, + { + "epoch": 1.256242011077972, + "grad_norm": 14.5625, + "learning_rate": 7.661012602406278e-07, + "loss": 0.6766, + "step": 14742 + }, + { + "epoch": 1.256327226246272, + "grad_norm": 11.8125, + "learning_rate": 7.659463308342768e-07, + "loss": 0.2537, + "step": 14743 + }, + { + "epoch": 1.2564124414145719, + "grad_norm": 15.75, + "learning_rate": 7.657914101752622e-07, + "loss": 0.5953, + "step": 14744 + }, + { + "epoch": 1.2564976565828718, + "grad_norm": 16.75, + "learning_rate": 7.656364982663847e-07, + "loss": 0.5801, + "step": 14745 + }, + { + "epoch": 1.2565828717511718, + "grad_norm": 15.6875, + "learning_rate": 7.654815951104419e-07, + "loss": 0.4965, + "step": 14746 + }, + { + "epoch": 1.2566680869194717, + "grad_norm": 15.125, + "learning_rate": 7.65326700710234e-07, + "loss": 0.3834, + "step": 14747 + }, + { + "epoch": 1.2567533020877717, + "grad_norm": 12.75, + "learning_rate": 7.651718150685597e-07, + "loss": 0.3145, + "step": 14748 + }, + { + "epoch": 1.2568385172560717, + "grad_norm": 15.5625, + "learning_rate": 7.650169381882186e-07, + "loss": 0.4912, + "step": 14749 + }, + { + "epoch": 1.2569237324243716, + "grad_norm": 16.625, + "learning_rate": 7.648620700720088e-07, + "loss": 0.534, + "step": 14750 + }, + { + "epoch": 1.2570089475926716, + "grad_norm": 16.625, + "learning_rate": 7.647072107227297e-07, + "loss": 0.3808, + "step": 14751 + }, + { + "epoch": 1.2570941627609715, + "grad_norm": 10.5625, + "learning_rate": 7.64552360143179e-07, + "loss": 0.358, + "step": 14752 + }, + { + "epoch": 1.2571793779292715, + "grad_norm": 14.1875, + "learning_rate": 7.643975183361555e-07, + "loss": 0.5785, + "step": 14753 + }, + { + "epoch": 1.2572645930975714, + "grad_norm": 15.8125, + "learning_rate": 7.642426853044573e-07, + "loss": 0.3668, + "step": 14754 + }, + { + "epoch": 1.2573498082658714, + "grad_norm": 11.75, + "learning_rate": 7.640878610508817e-07, + "loss": 0.3408, + "step": 14755 + }, + { + "epoch": 1.2574350234341714, + "grad_norm": 14.6875, + "learning_rate": 7.639330455782275e-07, + "loss": 0.4472, + "step": 14756 + }, + { + "epoch": 1.2575202386024713, + "grad_norm": 14.375, + "learning_rate": 7.637782388892916e-07, + "loss": 0.7079, + "step": 14757 + }, + { + "epoch": 1.2576054537707713, + "grad_norm": 17.625, + "learning_rate": 7.63623440986872e-07, + "loss": 0.6583, + "step": 14758 + }, + { + "epoch": 1.2576906689390712, + "grad_norm": 15.4375, + "learning_rate": 7.634686518737659e-07, + "loss": 0.5979, + "step": 14759 + }, + { + "epoch": 1.2577758841073712, + "grad_norm": 17.25, + "learning_rate": 7.633138715527706e-07, + "loss": 0.8534, + "step": 14760 + }, + { + "epoch": 1.2578610992756711, + "grad_norm": 15.5, + "learning_rate": 7.631591000266829e-07, + "loss": 0.5489, + "step": 14761 + }, + { + "epoch": 1.257946314443971, + "grad_norm": 13.875, + "learning_rate": 7.630043372983005e-07, + "loss": 0.491, + "step": 14762 + }, + { + "epoch": 1.258031529612271, + "grad_norm": 17.375, + "learning_rate": 7.628495833704186e-07, + "loss": 0.8262, + "step": 14763 + }, + { + "epoch": 1.258116744780571, + "grad_norm": 14.375, + "learning_rate": 7.62694838245835e-07, + "loss": 0.5647, + "step": 14764 + }, + { + "epoch": 1.258201959948871, + "grad_norm": 20.75, + "learning_rate": 7.625401019273454e-07, + "loss": 0.5822, + "step": 14765 + }, + { + "epoch": 1.258287175117171, + "grad_norm": 17.5, + "learning_rate": 7.623853744177466e-07, + "loss": 0.6623, + "step": 14766 + }, + { + "epoch": 1.2583723902854709, + "grad_norm": 13.1875, + "learning_rate": 7.622306557198341e-07, + "loss": 0.4083, + "step": 14767 + }, + { + "epoch": 1.2584576054537708, + "grad_norm": 15.9375, + "learning_rate": 7.620759458364043e-07, + "loss": 0.4822, + "step": 14768 + }, + { + "epoch": 1.2585428206220708, + "grad_norm": 20.875, + "learning_rate": 7.619212447702528e-07, + "loss": 0.6139, + "step": 14769 + }, + { + "epoch": 1.2586280357903707, + "grad_norm": 18.75, + "learning_rate": 7.617665525241757e-07, + "loss": 0.7264, + "step": 14770 + }, + { + "epoch": 1.2587132509586707, + "grad_norm": 14.8125, + "learning_rate": 7.616118691009675e-07, + "loss": 0.6282, + "step": 14771 + }, + { + "epoch": 1.2587984661269707, + "grad_norm": 12.375, + "learning_rate": 7.614571945034238e-07, + "loss": 0.4528, + "step": 14772 + }, + { + "epoch": 1.2588836812952706, + "grad_norm": 16.875, + "learning_rate": 7.613025287343401e-07, + "loss": 0.4314, + "step": 14773 + }, + { + "epoch": 1.2589688964635706, + "grad_norm": 16.75, + "learning_rate": 7.611478717965109e-07, + "loss": 0.596, + "step": 14774 + }, + { + "epoch": 1.2590541116318705, + "grad_norm": 13.0, + "learning_rate": 7.609932236927315e-07, + "loss": 0.3991, + "step": 14775 + }, + { + "epoch": 1.2591393268001705, + "grad_norm": 16.875, + "learning_rate": 7.608385844257959e-07, + "loss": 0.4711, + "step": 14776 + }, + { + "epoch": 1.2592245419684704, + "grad_norm": 13.625, + "learning_rate": 7.606839539984995e-07, + "loss": 0.3357, + "step": 14777 + }, + { + "epoch": 1.2593097571367704, + "grad_norm": 16.5, + "learning_rate": 7.60529332413636e-07, + "loss": 0.7104, + "step": 14778 + }, + { + "epoch": 1.2593949723050704, + "grad_norm": 21.375, + "learning_rate": 7.603747196739997e-07, + "loss": 0.998, + "step": 14779 + }, + { + "epoch": 1.2594801874733703, + "grad_norm": 14.1875, + "learning_rate": 7.602201157823841e-07, + "loss": 0.729, + "step": 14780 + }, + { + "epoch": 1.2595654026416703, + "grad_norm": 16.375, + "learning_rate": 7.600655207415838e-07, + "loss": 0.6283, + "step": 14781 + }, + { + "epoch": 1.2596506178099702, + "grad_norm": 13.4375, + "learning_rate": 7.59910934554392e-07, + "loss": 0.2892, + "step": 14782 + }, + { + "epoch": 1.2597358329782702, + "grad_norm": 16.75, + "learning_rate": 7.597563572236027e-07, + "loss": 0.8029, + "step": 14783 + }, + { + "epoch": 1.2598210481465701, + "grad_norm": 18.5, + "learning_rate": 7.596017887520088e-07, + "loss": 0.737, + "step": 14784 + }, + { + "epoch": 1.25990626331487, + "grad_norm": 16.125, + "learning_rate": 7.59447229142404e-07, + "loss": 0.6369, + "step": 14785 + }, + { + "epoch": 1.25999147848317, + "grad_norm": 18.25, + "learning_rate": 7.592926783975813e-07, + "loss": 0.3987, + "step": 14786 + }, + { + "epoch": 1.26007669365147, + "grad_norm": 38.25, + "learning_rate": 7.591381365203327e-07, + "loss": 0.6789, + "step": 14787 + }, + { + "epoch": 1.26016190881977, + "grad_norm": 17.75, + "learning_rate": 7.58983603513452e-07, + "loss": 0.5868, + "step": 14788 + }, + { + "epoch": 1.26024712398807, + "grad_norm": 19.0, + "learning_rate": 7.58829079379731e-07, + "loss": 0.9919, + "step": 14789 + }, + { + "epoch": 1.2603323391563699, + "grad_norm": 15.0625, + "learning_rate": 7.586745641219626e-07, + "loss": 0.5793, + "step": 14790 + }, + { + "epoch": 1.2604175543246698, + "grad_norm": 19.125, + "learning_rate": 7.585200577429387e-07, + "loss": 0.6396, + "step": 14791 + }, + { + "epoch": 1.2605027694929698, + "grad_norm": 11.5, + "learning_rate": 7.58365560245452e-07, + "loss": 0.3634, + "step": 14792 + }, + { + "epoch": 1.2605879846612698, + "grad_norm": 17.125, + "learning_rate": 7.582110716322938e-07, + "loss": 0.5278, + "step": 14793 + }, + { + "epoch": 1.2606731998295697, + "grad_norm": 15.75, + "learning_rate": 7.580565919062563e-07, + "loss": 0.4552, + "step": 14794 + }, + { + "epoch": 1.2607584149978697, + "grad_norm": 20.75, + "learning_rate": 7.579021210701304e-07, + "loss": 0.7806, + "step": 14795 + }, + { + "epoch": 1.2608436301661696, + "grad_norm": 18.5, + "learning_rate": 7.577476591267084e-07, + "loss": 0.7404, + "step": 14796 + }, + { + "epoch": 1.2609288453344696, + "grad_norm": 27.125, + "learning_rate": 7.575932060787808e-07, + "loss": 1.1459, + "step": 14797 + }, + { + "epoch": 1.2610140605027695, + "grad_norm": 14.625, + "learning_rate": 7.574387619291396e-07, + "loss": 0.6001, + "step": 14798 + }, + { + "epoch": 1.2610992756710695, + "grad_norm": 12.375, + "learning_rate": 7.572843266805747e-07, + "loss": 0.5163, + "step": 14799 + }, + { + "epoch": 1.2611844908393695, + "grad_norm": 15.6875, + "learning_rate": 7.571299003358779e-07, + "loss": 0.8591, + "step": 14800 + }, + { + "epoch": 1.2612697060076694, + "grad_norm": 18.625, + "learning_rate": 7.569754828978393e-07, + "loss": 0.7369, + "step": 14801 + }, + { + "epoch": 1.2613549211759694, + "grad_norm": 17.375, + "learning_rate": 7.568210743692497e-07, + "loss": 0.5494, + "step": 14802 + }, + { + "epoch": 1.2614401363442693, + "grad_norm": 16.0, + "learning_rate": 7.566666747528992e-07, + "loss": 0.6382, + "step": 14803 + }, + { + "epoch": 1.2615253515125693, + "grad_norm": 12.875, + "learning_rate": 7.565122840515776e-07, + "loss": 0.4514, + "step": 14804 + }, + { + "epoch": 1.2616105666808692, + "grad_norm": 18.25, + "learning_rate": 7.563579022680756e-07, + "loss": 0.5869, + "step": 14805 + }, + { + "epoch": 1.2616957818491692, + "grad_norm": 14.375, + "learning_rate": 7.562035294051824e-07, + "loss": 0.5325, + "step": 14806 + }, + { + "epoch": 1.2617809970174692, + "grad_norm": 17.0, + "learning_rate": 7.560491654656882e-07, + "loss": 0.4941, + "step": 14807 + }, + { + "epoch": 1.261866212185769, + "grad_norm": 11.875, + "learning_rate": 7.558948104523825e-07, + "loss": 0.415, + "step": 14808 + }, + { + "epoch": 1.261951427354069, + "grad_norm": 17.125, + "learning_rate": 7.557404643680538e-07, + "loss": 0.6768, + "step": 14809 + }, + { + "epoch": 1.262036642522369, + "grad_norm": 12.5625, + "learning_rate": 7.555861272154927e-07, + "loss": 0.3011, + "step": 14810 + }, + { + "epoch": 1.262121857690669, + "grad_norm": 13.6875, + "learning_rate": 7.554317989974871e-07, + "loss": 0.5505, + "step": 14811 + }, + { + "epoch": 1.262207072858969, + "grad_norm": 19.0, + "learning_rate": 7.552774797168261e-07, + "loss": 0.4301, + "step": 14812 + }, + { + "epoch": 1.262292288027269, + "grad_norm": 17.875, + "learning_rate": 7.551231693762987e-07, + "loss": 0.5846, + "step": 14813 + }, + { + "epoch": 1.2623775031955689, + "grad_norm": 10.9375, + "learning_rate": 7.549688679786929e-07, + "loss": 0.3565, + "step": 14814 + }, + { + "epoch": 1.2624627183638688, + "grad_norm": 12.8125, + "learning_rate": 7.548145755267979e-07, + "loss": 0.3375, + "step": 14815 + }, + { + "epoch": 1.2625479335321688, + "grad_norm": 18.375, + "learning_rate": 7.546602920234011e-07, + "loss": 0.5924, + "step": 14816 + }, + { + "epoch": 1.2626331487004687, + "grad_norm": 24.5, + "learning_rate": 7.545060174712912e-07, + "loss": 1.164, + "step": 14817 + }, + { + "epoch": 1.2627183638687687, + "grad_norm": 13.4375, + "learning_rate": 7.54351751873256e-07, + "loss": 0.5373, + "step": 14818 + }, + { + "epoch": 1.2628035790370686, + "grad_norm": 14.5625, + "learning_rate": 7.541974952320827e-07, + "loss": 0.3873, + "step": 14819 + }, + { + "epoch": 1.2628887942053686, + "grad_norm": 16.125, + "learning_rate": 7.540432475505593e-07, + "loss": 0.6975, + "step": 14820 + }, + { + "epoch": 1.2629740093736685, + "grad_norm": 22.375, + "learning_rate": 7.538890088314731e-07, + "loss": 0.7877, + "step": 14821 + }, + { + "epoch": 1.2630592245419685, + "grad_norm": 17.25, + "learning_rate": 7.537347790776114e-07, + "loss": 0.7288, + "step": 14822 + }, + { + "epoch": 1.2631444397102685, + "grad_norm": 12.375, + "learning_rate": 7.535805582917616e-07, + "loss": 0.2239, + "step": 14823 + }, + { + "epoch": 1.2632296548785684, + "grad_norm": 17.375, + "learning_rate": 7.534263464767097e-07, + "loss": 0.6621, + "step": 14824 + }, + { + "epoch": 1.2633148700468684, + "grad_norm": 16.625, + "learning_rate": 7.532721436352436e-07, + "loss": 0.5718, + "step": 14825 + }, + { + "epoch": 1.2634000852151683, + "grad_norm": 13.875, + "learning_rate": 7.53117949770149e-07, + "loss": 0.48, + "step": 14826 + }, + { + "epoch": 1.2634853003834683, + "grad_norm": 15.875, + "learning_rate": 7.529637648842131e-07, + "loss": 0.6214, + "step": 14827 + }, + { + "epoch": 1.2635705155517682, + "grad_norm": 22.0, + "learning_rate": 7.528095889802217e-07, + "loss": 0.7493, + "step": 14828 + }, + { + "epoch": 1.2636557307200682, + "grad_norm": 10.8125, + "learning_rate": 7.526554220609605e-07, + "loss": 0.2894, + "step": 14829 + }, + { + "epoch": 1.2637409458883682, + "grad_norm": 23.0, + "learning_rate": 7.525012641292165e-07, + "loss": 0.6788, + "step": 14830 + }, + { + "epoch": 1.2638261610566681, + "grad_norm": 12.75, + "learning_rate": 7.523471151877744e-07, + "loss": 0.5408, + "step": 14831 + }, + { + "epoch": 1.263911376224968, + "grad_norm": 12.3125, + "learning_rate": 7.521929752394209e-07, + "loss": 0.5056, + "step": 14832 + }, + { + "epoch": 1.263996591393268, + "grad_norm": 15.1875, + "learning_rate": 7.520388442869403e-07, + "loss": 0.629, + "step": 14833 + }, + { + "epoch": 1.264081806561568, + "grad_norm": 16.5, + "learning_rate": 7.518847223331191e-07, + "loss": 0.6272, + "step": 14834 + }, + { + "epoch": 1.264167021729868, + "grad_norm": 18.75, + "learning_rate": 7.51730609380742e-07, + "loss": 0.6811, + "step": 14835 + }, + { + "epoch": 1.264252236898168, + "grad_norm": 15.9375, + "learning_rate": 7.515765054325932e-07, + "loss": 0.5443, + "step": 14836 + }, + { + "epoch": 1.2643374520664679, + "grad_norm": 16.75, + "learning_rate": 7.514224104914586e-07, + "loss": 0.5802, + "step": 14837 + }, + { + "epoch": 1.2644226672347678, + "grad_norm": 17.875, + "learning_rate": 7.512683245601224e-07, + "loss": 0.9035, + "step": 14838 + }, + { + "epoch": 1.2645078824030678, + "grad_norm": 13.125, + "learning_rate": 7.511142476413687e-07, + "loss": 0.6229, + "step": 14839 + }, + { + "epoch": 1.2645930975713677, + "grad_norm": 12.5625, + "learning_rate": 7.509601797379824e-07, + "loss": 0.484, + "step": 14840 + }, + { + "epoch": 1.2646783127396677, + "grad_norm": 23.125, + "learning_rate": 7.508061208527474e-07, + "loss": 0.7213, + "step": 14841 + }, + { + "epoch": 1.2647635279079676, + "grad_norm": 15.9375, + "learning_rate": 7.506520709884479e-07, + "loss": 0.6746, + "step": 14842 + }, + { + "epoch": 1.2648487430762676, + "grad_norm": 18.5, + "learning_rate": 7.50498030147868e-07, + "loss": 0.9253, + "step": 14843 + }, + { + "epoch": 1.2649339582445676, + "grad_norm": 17.375, + "learning_rate": 7.503439983337904e-07, + "loss": 0.5983, + "step": 14844 + }, + { + "epoch": 1.2650191734128675, + "grad_norm": 13.5, + "learning_rate": 7.501899755489994e-07, + "loss": 0.446, + "step": 14845 + }, + { + "epoch": 1.2651043885811675, + "grad_norm": 15.0, + "learning_rate": 7.50035961796278e-07, + "loss": 0.5689, + "step": 14846 + }, + { + "epoch": 1.2651896037494674, + "grad_norm": 14.5, + "learning_rate": 7.498819570784098e-07, + "loss": 0.448, + "step": 14847 + }, + { + "epoch": 1.2652748189177674, + "grad_norm": 12.25, + "learning_rate": 7.497279613981773e-07, + "loss": 0.3957, + "step": 14848 + }, + { + "epoch": 1.2653600340860673, + "grad_norm": 20.5, + "learning_rate": 7.495739747583636e-07, + "loss": 0.6283, + "step": 14849 + }, + { + "epoch": 1.2654452492543673, + "grad_norm": 17.0, + "learning_rate": 7.494199971617513e-07, + "loss": 0.9095, + "step": 14850 + }, + { + "epoch": 1.2655304644226673, + "grad_norm": 11.875, + "learning_rate": 7.492660286111237e-07, + "loss": 0.503, + "step": 14851 + }, + { + "epoch": 1.2656156795909672, + "grad_norm": 21.0, + "learning_rate": 7.49112069109262e-07, + "loss": 0.8372, + "step": 14852 + }, + { + "epoch": 1.2657008947592672, + "grad_norm": 10.9375, + "learning_rate": 7.489581186589489e-07, + "loss": 0.3218, + "step": 14853 + }, + { + "epoch": 1.2657861099275671, + "grad_norm": 36.25, + "learning_rate": 7.488041772629663e-07, + "loss": 0.7869, + "step": 14854 + }, + { + "epoch": 1.265871325095867, + "grad_norm": 16.0, + "learning_rate": 7.486502449240965e-07, + "loss": 0.723, + "step": 14855 + }, + { + "epoch": 1.265956540264167, + "grad_norm": 16.5, + "learning_rate": 7.484963216451205e-07, + "loss": 0.5091, + "step": 14856 + }, + { + "epoch": 1.266041755432467, + "grad_norm": 15.6875, + "learning_rate": 7.483424074288207e-07, + "loss": 0.5811, + "step": 14857 + }, + { + "epoch": 1.266126970600767, + "grad_norm": 11.625, + "learning_rate": 7.481885022779775e-07, + "loss": 0.2233, + "step": 14858 + }, + { + "epoch": 1.266212185769067, + "grad_norm": 15.375, + "learning_rate": 7.480346061953735e-07, + "loss": 0.7656, + "step": 14859 + }, + { + "epoch": 1.2662974009373669, + "grad_norm": 14.25, + "learning_rate": 7.478807191837885e-07, + "loss": 0.3737, + "step": 14860 + }, + { + "epoch": 1.2663826161056668, + "grad_norm": 18.875, + "learning_rate": 7.477268412460037e-07, + "loss": 0.8529, + "step": 14861 + }, + { + "epoch": 1.2664678312739668, + "grad_norm": 14.3125, + "learning_rate": 7.475729723848e-07, + "loss": 0.6417, + "step": 14862 + }, + { + "epoch": 1.2665530464422667, + "grad_norm": 11.8125, + "learning_rate": 7.474191126029579e-07, + "loss": 0.3944, + "step": 14863 + }, + { + "epoch": 1.2666382616105667, + "grad_norm": 11.625, + "learning_rate": 7.47265261903258e-07, + "loss": 0.5039, + "step": 14864 + }, + { + "epoch": 1.2667234767788667, + "grad_norm": 19.375, + "learning_rate": 7.4711142028848e-07, + "loss": 0.695, + "step": 14865 + }, + { + "epoch": 1.2668086919471666, + "grad_norm": 13.875, + "learning_rate": 7.469575877614047e-07, + "loss": 0.6105, + "step": 14866 + }, + { + "epoch": 1.2668939071154666, + "grad_norm": 14.5625, + "learning_rate": 7.468037643248119e-07, + "loss": 0.4175, + "step": 14867 + }, + { + "epoch": 1.2669791222837665, + "grad_norm": 14.6875, + "learning_rate": 7.466499499814807e-07, + "loss": 0.7527, + "step": 14868 + }, + { + "epoch": 1.2670643374520665, + "grad_norm": 22.125, + "learning_rate": 7.464961447341909e-07, + "loss": 0.5648, + "step": 14869 + }, + { + "epoch": 1.2671495526203664, + "grad_norm": 18.5, + "learning_rate": 7.463423485857224e-07, + "loss": 0.7705, + "step": 14870 + }, + { + "epoch": 1.2672347677886664, + "grad_norm": 13.375, + "learning_rate": 7.461885615388536e-07, + "loss": 0.5381, + "step": 14871 + }, + { + "epoch": 1.2673199829569664, + "grad_norm": 13.0, + "learning_rate": 7.460347835963647e-07, + "loss": 0.4936, + "step": 14872 + }, + { + "epoch": 1.2674051981252663, + "grad_norm": 20.875, + "learning_rate": 7.458810147610336e-07, + "loss": 0.8413, + "step": 14873 + }, + { + "epoch": 1.2674904132935663, + "grad_norm": 14.8125, + "learning_rate": 7.457272550356399e-07, + "loss": 0.7646, + "step": 14874 + }, + { + "epoch": 1.2675756284618662, + "grad_norm": 16.375, + "learning_rate": 7.455735044229617e-07, + "loss": 0.7063, + "step": 14875 + }, + { + "epoch": 1.2676608436301662, + "grad_norm": 15.5, + "learning_rate": 7.454197629257773e-07, + "loss": 0.6136, + "step": 14876 + }, + { + "epoch": 1.2677460587984661, + "grad_norm": 13.5, + "learning_rate": 7.452660305468653e-07, + "loss": 0.3719, + "step": 14877 + }, + { + "epoch": 1.267831273966766, + "grad_norm": 17.875, + "learning_rate": 7.451123072890037e-07, + "loss": 0.8116, + "step": 14878 + }, + { + "epoch": 1.267916489135066, + "grad_norm": 16.125, + "learning_rate": 7.449585931549704e-07, + "loss": 0.5316, + "step": 14879 + }, + { + "epoch": 1.268001704303366, + "grad_norm": 16.125, + "learning_rate": 7.44804888147543e-07, + "loss": 0.4795, + "step": 14880 + }, + { + "epoch": 1.268086919471666, + "grad_norm": 20.5, + "learning_rate": 7.446511922694994e-07, + "loss": 0.5703, + "step": 14881 + }, + { + "epoch": 1.268172134639966, + "grad_norm": 11.9375, + "learning_rate": 7.444975055236169e-07, + "loss": 0.35, + "step": 14882 + }, + { + "epoch": 1.2682573498082659, + "grad_norm": 14.0, + "learning_rate": 7.443438279126734e-07, + "loss": 0.5845, + "step": 14883 + }, + { + "epoch": 1.2683425649765658, + "grad_norm": 21.875, + "learning_rate": 7.441901594394446e-07, + "loss": 0.7929, + "step": 14884 + }, + { + "epoch": 1.2684277801448658, + "grad_norm": 12.25, + "learning_rate": 7.440365001067087e-07, + "loss": 0.3127, + "step": 14885 + }, + { + "epoch": 1.2685129953131657, + "grad_norm": 14.25, + "learning_rate": 7.438828499172416e-07, + "loss": 0.3749, + "step": 14886 + }, + { + "epoch": 1.2685982104814657, + "grad_norm": 11.6875, + "learning_rate": 7.437292088738207e-07, + "loss": 0.5223, + "step": 14887 + }, + { + "epoch": 1.2686834256497657, + "grad_norm": 15.5, + "learning_rate": 7.435755769792218e-07, + "loss": 0.5193, + "step": 14888 + }, + { + "epoch": 1.2687686408180656, + "grad_norm": 11.25, + "learning_rate": 7.434219542362217e-07, + "loss": 0.3711, + "step": 14889 + }, + { + "epoch": 1.2688538559863656, + "grad_norm": 12.0625, + "learning_rate": 7.432683406475962e-07, + "loss": 0.3422, + "step": 14890 + }, + { + "epoch": 1.2689390711546655, + "grad_norm": 18.5, + "learning_rate": 7.431147362161217e-07, + "loss": 0.9403, + "step": 14891 + }, + { + "epoch": 1.2690242863229655, + "grad_norm": 14.3125, + "learning_rate": 7.429611409445733e-07, + "loss": 0.3545, + "step": 14892 + }, + { + "epoch": 1.2691095014912654, + "grad_norm": 16.25, + "learning_rate": 7.42807554835727e-07, + "loss": 0.5702, + "step": 14893 + }, + { + "epoch": 1.2691947166595654, + "grad_norm": 22.875, + "learning_rate": 7.42653977892358e-07, + "loss": 0.5754, + "step": 14894 + }, + { + "epoch": 1.2692799318278654, + "grad_norm": 17.75, + "learning_rate": 7.42500410117242e-07, + "loss": 0.7299, + "step": 14895 + }, + { + "epoch": 1.2693651469961653, + "grad_norm": 15.125, + "learning_rate": 7.423468515131538e-07, + "loss": 0.6058, + "step": 14896 + }, + { + "epoch": 1.2694503621644653, + "grad_norm": 17.875, + "learning_rate": 7.421933020828684e-07, + "loss": 0.8696, + "step": 14897 + }, + { + "epoch": 1.2695355773327652, + "grad_norm": 15.75, + "learning_rate": 7.42039761829161e-07, + "loss": 0.6012, + "step": 14898 + }, + { + "epoch": 1.2696207925010652, + "grad_norm": 19.125, + "learning_rate": 7.418862307548055e-07, + "loss": 0.8615, + "step": 14899 + }, + { + "epoch": 1.2697060076693651, + "grad_norm": 14.625, + "learning_rate": 7.417327088625776e-07, + "loss": 0.624, + "step": 14900 + }, + { + "epoch": 1.269791222837665, + "grad_norm": 11.75, + "learning_rate": 7.4157919615525e-07, + "loss": 0.2909, + "step": 14901 + }, + { + "epoch": 1.269876438005965, + "grad_norm": 16.625, + "learning_rate": 7.414256926355978e-07, + "loss": 0.8275, + "step": 14902 + }, + { + "epoch": 1.269961653174265, + "grad_norm": 20.0, + "learning_rate": 7.412721983063944e-07, + "loss": 0.3769, + "step": 14903 + }, + { + "epoch": 1.270046868342565, + "grad_norm": 19.25, + "learning_rate": 7.411187131704143e-07, + "loss": 0.7938, + "step": 14904 + }, + { + "epoch": 1.270132083510865, + "grad_norm": 23.75, + "learning_rate": 7.409652372304306e-07, + "loss": 0.9351, + "step": 14905 + }, + { + "epoch": 1.2702172986791649, + "grad_norm": 16.875, + "learning_rate": 7.40811770489217e-07, + "loss": 0.5219, + "step": 14906 + }, + { + "epoch": 1.2703025138474648, + "grad_norm": 14.5625, + "learning_rate": 7.406583129495466e-07, + "loss": 0.6657, + "step": 14907 + }, + { + "epoch": 1.2703877290157648, + "grad_norm": 15.25, + "learning_rate": 7.405048646141932e-07, + "loss": 0.4659, + "step": 14908 + }, + { + "epoch": 1.2704729441840648, + "grad_norm": 16.125, + "learning_rate": 7.40351425485929e-07, + "loss": 0.5535, + "step": 14909 + }, + { + "epoch": 1.2705581593523647, + "grad_norm": 21.625, + "learning_rate": 7.401979955675267e-07, + "loss": 0.8114, + "step": 14910 + }, + { + "epoch": 1.2706433745206647, + "grad_norm": 13.25, + "learning_rate": 7.400445748617595e-07, + "loss": 0.6602, + "step": 14911 + }, + { + "epoch": 1.2707285896889646, + "grad_norm": 16.125, + "learning_rate": 7.398911633713994e-07, + "loss": 0.6098, + "step": 14912 + }, + { + "epoch": 1.2708138048572646, + "grad_norm": 20.0, + "learning_rate": 7.397377610992193e-07, + "loss": 1.0105, + "step": 14913 + }, + { + "epoch": 1.2708990200255645, + "grad_norm": 16.5, + "learning_rate": 7.395843680479906e-07, + "loss": 0.5999, + "step": 14914 + }, + { + "epoch": 1.2709842351938645, + "grad_norm": 24.0, + "learning_rate": 7.394309842204858e-07, + "loss": 0.8848, + "step": 14915 + }, + { + "epoch": 1.2710694503621645, + "grad_norm": 13.125, + "learning_rate": 7.392776096194769e-07, + "loss": 0.4414, + "step": 14916 + }, + { + "epoch": 1.2711546655304644, + "grad_norm": 25.0, + "learning_rate": 7.391242442477349e-07, + "loss": 0.6606, + "step": 14917 + }, + { + "epoch": 1.2712398806987644, + "grad_norm": 18.125, + "learning_rate": 7.389708881080312e-07, + "loss": 0.5772, + "step": 14918 + }, + { + "epoch": 1.2713250958670643, + "grad_norm": 14.375, + "learning_rate": 7.388175412031379e-07, + "loss": 0.4497, + "step": 14919 + }, + { + "epoch": 1.2714103110353643, + "grad_norm": 19.125, + "learning_rate": 7.386642035358252e-07, + "loss": 0.4895, + "step": 14920 + }, + { + "epoch": 1.2714955262036642, + "grad_norm": 13.8125, + "learning_rate": 7.385108751088649e-07, + "loss": 0.5445, + "step": 14921 + }, + { + "epoch": 1.2715807413719642, + "grad_norm": 14.0, + "learning_rate": 7.383575559250271e-07, + "loss": 0.4578, + "step": 14922 + }, + { + "epoch": 1.2716659565402642, + "grad_norm": 14.0625, + "learning_rate": 7.382042459870831e-07, + "loss": 0.3584, + "step": 14923 + }, + { + "epoch": 1.271751171708564, + "grad_norm": 13.25, + "learning_rate": 7.380509452978032e-07, + "loss": 0.5988, + "step": 14924 + }, + { + "epoch": 1.271836386876864, + "grad_norm": 17.25, + "learning_rate": 7.378976538599569e-07, + "loss": 0.6437, + "step": 14925 + }, + { + "epoch": 1.271921602045164, + "grad_norm": 14.25, + "learning_rate": 7.377443716763155e-07, + "loss": 0.4798, + "step": 14926 + }, + { + "epoch": 1.272006817213464, + "grad_norm": 6.5, + "learning_rate": 7.375910987496479e-07, + "loss": 0.107, + "step": 14927 + }, + { + "epoch": 1.272092032381764, + "grad_norm": 19.875, + "learning_rate": 7.374378350827248e-07, + "loss": 0.9183, + "step": 14928 + }, + { + "epoch": 1.272177247550064, + "grad_norm": 19.75, + "learning_rate": 7.372845806783152e-07, + "loss": 0.8643, + "step": 14929 + }, + { + "epoch": 1.2722624627183639, + "grad_norm": 13.625, + "learning_rate": 7.37131335539189e-07, + "loss": 0.5338, + "step": 14930 + }, + { + "epoch": 1.2723476778866638, + "grad_norm": 14.6875, + "learning_rate": 7.369780996681153e-07, + "loss": 0.3288, + "step": 14931 + }, + { + "epoch": 1.2724328930549638, + "grad_norm": 10.75, + "learning_rate": 7.368248730678634e-07, + "loss": 0.3136, + "step": 14932 + }, + { + "epoch": 1.2725181082232637, + "grad_norm": 19.75, + "learning_rate": 7.366716557412019e-07, + "loss": 0.7263, + "step": 14933 + }, + { + "epoch": 1.2726033233915637, + "grad_norm": 17.5, + "learning_rate": 7.365184476908995e-07, + "loss": 0.9171, + "step": 14934 + }, + { + "epoch": 1.2726885385598636, + "grad_norm": 15.5625, + "learning_rate": 7.363652489197252e-07, + "loss": 0.8063, + "step": 14935 + }, + { + "epoch": 1.2727737537281636, + "grad_norm": 21.25, + "learning_rate": 7.362120594304476e-07, + "loss": 0.5757, + "step": 14936 + }, + { + "epoch": 1.2728589688964635, + "grad_norm": 19.875, + "learning_rate": 7.360588792258344e-07, + "loss": 0.9526, + "step": 14937 + }, + { + "epoch": 1.2729441840647635, + "grad_norm": 27.875, + "learning_rate": 7.359057083086543e-07, + "loss": 1.1087, + "step": 14938 + }, + { + "epoch": 1.2730293992330635, + "grad_norm": 9.875, + "learning_rate": 7.357525466816748e-07, + "loss": 0.2731, + "step": 14939 + }, + { + "epoch": 1.2731146144013634, + "grad_norm": 38.75, + "learning_rate": 7.355993943476644e-07, + "loss": 0.7923, + "step": 14940 + }, + { + "epoch": 1.2731998295696634, + "grad_norm": 26.125, + "learning_rate": 7.3544625130939e-07, + "loss": 0.7759, + "step": 14941 + }, + { + "epoch": 1.2732850447379633, + "grad_norm": 17.0, + "learning_rate": 7.352931175696191e-07, + "loss": 0.5691, + "step": 14942 + }, + { + "epoch": 1.2733702599062633, + "grad_norm": 17.125, + "learning_rate": 7.351399931311191e-07, + "loss": 0.7022, + "step": 14943 + }, + { + "epoch": 1.2734554750745632, + "grad_norm": 15.125, + "learning_rate": 7.349868779966571e-07, + "loss": 0.6164, + "step": 14944 + }, + { + "epoch": 1.2735406902428632, + "grad_norm": 10.875, + "learning_rate": 7.348337721690004e-07, + "loss": 0.2313, + "step": 14945 + }, + { + "epoch": 1.2736259054111632, + "grad_norm": 13.3125, + "learning_rate": 7.346806756509153e-07, + "loss": 0.5515, + "step": 14946 + }, + { + "epoch": 1.2737111205794631, + "grad_norm": 13.6875, + "learning_rate": 7.345275884451686e-07, + "loss": 0.466, + "step": 14947 + }, + { + "epoch": 1.273796335747763, + "grad_norm": 15.5625, + "learning_rate": 7.343745105545269e-07, + "loss": 0.7773, + "step": 14948 + }, + { + "epoch": 1.273881550916063, + "grad_norm": 16.375, + "learning_rate": 7.342214419817561e-07, + "loss": 0.7301, + "step": 14949 + }, + { + "epoch": 1.273966766084363, + "grad_norm": 10.6875, + "learning_rate": 7.340683827296224e-07, + "loss": 0.2507, + "step": 14950 + }, + { + "epoch": 1.274051981252663, + "grad_norm": 18.375, + "learning_rate": 7.339153328008922e-07, + "loss": 0.4547, + "step": 14951 + }, + { + "epoch": 1.274137196420963, + "grad_norm": 15.125, + "learning_rate": 7.337622921983303e-07, + "loss": 0.5418, + "step": 14952 + }, + { + "epoch": 1.2742224115892629, + "grad_norm": 16.375, + "learning_rate": 7.336092609247034e-07, + "loss": 0.3821, + "step": 14953 + }, + { + "epoch": 1.2743076267575628, + "grad_norm": 15.125, + "learning_rate": 7.334562389827762e-07, + "loss": 0.5002, + "step": 14954 + }, + { + "epoch": 1.2743928419258628, + "grad_norm": 18.125, + "learning_rate": 7.333032263753142e-07, + "loss": 0.7862, + "step": 14955 + }, + { + "epoch": 1.2744780570941627, + "grad_norm": 15.125, + "learning_rate": 7.33150223105083e-07, + "loss": 0.7382, + "step": 14956 + }, + { + "epoch": 1.2745632722624627, + "grad_norm": 16.25, + "learning_rate": 7.329972291748464e-07, + "loss": 0.3691, + "step": 14957 + }, + { + "epoch": 1.2746484874307626, + "grad_norm": 18.125, + "learning_rate": 7.328442445873699e-07, + "loss": 0.555, + "step": 14958 + }, + { + "epoch": 1.2747337025990626, + "grad_norm": 22.875, + "learning_rate": 7.326912693454177e-07, + "loss": 0.9039, + "step": 14959 + }, + { + "epoch": 1.2748189177673626, + "grad_norm": 16.375, + "learning_rate": 7.325383034517549e-07, + "loss": 0.4336, + "step": 14960 + }, + { + "epoch": 1.2749041329356625, + "grad_norm": 15.0, + "learning_rate": 7.323853469091452e-07, + "loss": 0.6374, + "step": 14961 + }, + { + "epoch": 1.2749893481039625, + "grad_norm": 18.0, + "learning_rate": 7.322323997203525e-07, + "loss": 0.8151, + "step": 14962 + }, + { + "epoch": 1.2750745632722624, + "grad_norm": 18.75, + "learning_rate": 7.320794618881412e-07, + "loss": 0.4047, + "step": 14963 + }, + { + "epoch": 1.2751597784405624, + "grad_norm": 13.4375, + "learning_rate": 7.319265334152748e-07, + "loss": 0.4835, + "step": 14964 + }, + { + "epoch": 1.2752449936088623, + "grad_norm": 12.0625, + "learning_rate": 7.317736143045171e-07, + "loss": 0.5499, + "step": 14965 + }, + { + "epoch": 1.2753302087771623, + "grad_norm": 14.5, + "learning_rate": 7.316207045586313e-07, + "loss": 0.7171, + "step": 14966 + }, + { + "epoch": 1.2754154239454623, + "grad_norm": 12.8125, + "learning_rate": 7.314678041803802e-07, + "loss": 0.4786, + "step": 14967 + }, + { + "epoch": 1.2755006391137622, + "grad_norm": 15.875, + "learning_rate": 7.313149131725277e-07, + "loss": 0.5722, + "step": 14968 + }, + { + "epoch": 1.2755858542820622, + "grad_norm": 16.125, + "learning_rate": 7.31162031537836e-07, + "loss": 0.6202, + "step": 14969 + }, + { + "epoch": 1.2756710694503621, + "grad_norm": 16.625, + "learning_rate": 7.310091592790683e-07, + "loss": 0.5454, + "step": 14970 + }, + { + "epoch": 1.275756284618662, + "grad_norm": 12.0625, + "learning_rate": 7.308562963989869e-07, + "loss": 0.4022, + "step": 14971 + }, + { + "epoch": 1.275841499786962, + "grad_norm": 16.375, + "learning_rate": 7.307034429003542e-07, + "loss": 0.4146, + "step": 14972 + }, + { + "epoch": 1.275926714955262, + "grad_norm": 15.625, + "learning_rate": 7.30550598785933e-07, + "loss": 0.4525, + "step": 14973 + }, + { + "epoch": 1.276011930123562, + "grad_norm": 14.0625, + "learning_rate": 7.303977640584842e-07, + "loss": 0.5165, + "step": 14974 + }, + { + "epoch": 1.276097145291862, + "grad_norm": 15.0625, + "learning_rate": 7.302449387207707e-07, + "loss": 0.8541, + "step": 14975 + }, + { + "epoch": 1.2761823604601619, + "grad_norm": 19.75, + "learning_rate": 7.300921227755537e-07, + "loss": 0.2242, + "step": 14976 + }, + { + "epoch": 1.2762675756284618, + "grad_norm": 13.125, + "learning_rate": 7.299393162255946e-07, + "loss": 0.5531, + "step": 14977 + }, + { + "epoch": 1.2763527907967618, + "grad_norm": 15.75, + "learning_rate": 7.297865190736552e-07, + "loss": 0.4541, + "step": 14978 + }, + { + "epoch": 1.2764380059650617, + "grad_norm": 21.625, + "learning_rate": 7.296337313224965e-07, + "loss": 0.7248, + "step": 14979 + }, + { + "epoch": 1.2765232211333617, + "grad_norm": 16.125, + "learning_rate": 7.294809529748794e-07, + "loss": 0.5078, + "step": 14980 + }, + { + "epoch": 1.2766084363016617, + "grad_norm": 22.125, + "learning_rate": 7.293281840335653e-07, + "loss": 0.5162, + "step": 14981 + }, + { + "epoch": 1.2766936514699616, + "grad_norm": 12.0, + "learning_rate": 7.291754245013147e-07, + "loss": 0.3598, + "step": 14982 + }, + { + "epoch": 1.2767788666382616, + "grad_norm": 22.125, + "learning_rate": 7.290226743808873e-07, + "loss": 1.011, + "step": 14983 + }, + { + "epoch": 1.2768640818065615, + "grad_norm": 15.875, + "learning_rate": 7.288699336750441e-07, + "loss": 0.7183, + "step": 14984 + }, + { + "epoch": 1.2769492969748615, + "grad_norm": 33.25, + "learning_rate": 7.287172023865452e-07, + "loss": 0.7252, + "step": 14985 + }, + { + "epoch": 1.2770345121431614, + "grad_norm": 14.875, + "learning_rate": 7.28564480518151e-07, + "loss": 0.7955, + "step": 14986 + }, + { + "epoch": 1.2771197273114614, + "grad_norm": 15.1875, + "learning_rate": 7.284117680726208e-07, + "loss": 0.7062, + "step": 14987 + }, + { + "epoch": 1.2772049424797614, + "grad_norm": 10.375, + "learning_rate": 7.282590650527141e-07, + "loss": 0.1982, + "step": 14988 + }, + { + "epoch": 1.2772901576480613, + "grad_norm": 12.8125, + "learning_rate": 7.281063714611914e-07, + "loss": 0.5037, + "step": 14989 + }, + { + "epoch": 1.2773753728163613, + "grad_norm": 21.375, + "learning_rate": 7.27953687300811e-07, + "loss": 0.5884, + "step": 14990 + }, + { + "epoch": 1.2774605879846612, + "grad_norm": 14.25, + "learning_rate": 7.278010125743327e-07, + "loss": 0.55, + "step": 14991 + }, + { + "epoch": 1.2775458031529612, + "grad_norm": 13.4375, + "learning_rate": 7.276483472845148e-07, + "loss": 0.6246, + "step": 14992 + }, + { + "epoch": 1.2776310183212611, + "grad_norm": 15.25, + "learning_rate": 7.274956914341166e-07, + "loss": 0.8062, + "step": 14993 + }, + { + "epoch": 1.277716233489561, + "grad_norm": 14.875, + "learning_rate": 7.273430450258973e-07, + "loss": 0.6742, + "step": 14994 + }, + { + "epoch": 1.277801448657861, + "grad_norm": 19.25, + "learning_rate": 7.27190408062614e-07, + "loss": 0.7489, + "step": 14995 + }, + { + "epoch": 1.277886663826161, + "grad_norm": 18.125, + "learning_rate": 7.270377805470258e-07, + "loss": 0.7734, + "step": 14996 + }, + { + "epoch": 1.277971878994461, + "grad_norm": 15.625, + "learning_rate": 7.268851624818913e-07, + "loss": 0.6351, + "step": 14997 + }, + { + "epoch": 1.278057094162761, + "grad_norm": 11.6875, + "learning_rate": 7.267325538699674e-07, + "loss": 0.3188, + "step": 14998 + }, + { + "epoch": 1.2781423093310609, + "grad_norm": 13.25, + "learning_rate": 7.265799547140132e-07, + "loss": 0.5073, + "step": 14999 + }, + { + "epoch": 1.2782275244993608, + "grad_norm": 20.25, + "learning_rate": 7.264273650167847e-07, + "loss": 0.6926, + "step": 15000 + }, + { + "epoch": 1.2783127396676608, + "grad_norm": 14.1875, + "learning_rate": 7.262747847810404e-07, + "loss": 0.5763, + "step": 15001 + }, + { + "epoch": 1.2783979548359607, + "grad_norm": 18.375, + "learning_rate": 7.261222140095374e-07, + "loss": 0.6478, + "step": 15002 + }, + { + "epoch": 1.2784831700042607, + "grad_norm": 19.375, + "learning_rate": 7.259696527050332e-07, + "loss": 0.7442, + "step": 15003 + }, + { + "epoch": 1.2785683851725607, + "grad_norm": 15.875, + "learning_rate": 7.258171008702839e-07, + "loss": 0.9312, + "step": 15004 + }, + { + "epoch": 1.2786536003408606, + "grad_norm": 19.75, + "learning_rate": 7.256645585080472e-07, + "loss": 0.966, + "step": 15005 + }, + { + "epoch": 1.2787388155091606, + "grad_norm": 16.125, + "learning_rate": 7.255120256210786e-07, + "loss": 0.5236, + "step": 15006 + }, + { + "epoch": 1.2788240306774605, + "grad_norm": 14.5625, + "learning_rate": 7.253595022121357e-07, + "loss": 0.5449, + "step": 15007 + }, + { + "epoch": 1.2789092458457605, + "grad_norm": 10.5625, + "learning_rate": 7.252069882839735e-07, + "loss": 0.3431, + "step": 15008 + }, + { + "epoch": 1.2789944610140604, + "grad_norm": 15.625, + "learning_rate": 7.250544838393489e-07, + "loss": 0.5911, + "step": 15009 + }, + { + "epoch": 1.2790796761823604, + "grad_norm": 15.4375, + "learning_rate": 7.249019888810177e-07, + "loss": 0.357, + "step": 15010 + }, + { + "epoch": 1.2791648913506604, + "grad_norm": 16.875, + "learning_rate": 7.247495034117361e-07, + "loss": 0.7202, + "step": 15011 + }, + { + "epoch": 1.2792501065189603, + "grad_norm": 18.875, + "learning_rate": 7.245970274342584e-07, + "loss": 0.4696, + "step": 15012 + }, + { + "epoch": 1.2793353216872603, + "grad_norm": 10.75, + "learning_rate": 7.244445609513414e-07, + "loss": 0.2505, + "step": 15013 + }, + { + "epoch": 1.2794205368555602, + "grad_norm": 15.6875, + "learning_rate": 7.242921039657393e-07, + "loss": 0.4527, + "step": 15014 + }, + { + "epoch": 1.2795057520238602, + "grad_norm": 19.625, + "learning_rate": 7.241396564802074e-07, + "loss": 0.6797, + "step": 15015 + }, + { + "epoch": 1.2795909671921601, + "grad_norm": 38.5, + "learning_rate": 7.239872184975011e-07, + "loss": 0.7694, + "step": 15016 + }, + { + "epoch": 1.27967618236046, + "grad_norm": 13.8125, + "learning_rate": 7.238347900203744e-07, + "loss": 0.625, + "step": 15017 + }, + { + "epoch": 1.27976139752876, + "grad_norm": 18.0, + "learning_rate": 7.236823710515819e-07, + "loss": 0.5112, + "step": 15018 + }, + { + "epoch": 1.27984661269706, + "grad_norm": 20.25, + "learning_rate": 7.235299615938785e-07, + "loss": 0.6654, + "step": 15019 + }, + { + "epoch": 1.27993182786536, + "grad_norm": 12.5625, + "learning_rate": 7.233775616500185e-07, + "loss": 0.4417, + "step": 15020 + }, + { + "epoch": 1.28001704303366, + "grad_norm": 12.5625, + "learning_rate": 7.232251712227554e-07, + "loss": 0.5317, + "step": 15021 + }, + { + "epoch": 1.2801022582019599, + "grad_norm": 17.75, + "learning_rate": 7.230727903148428e-07, + "loss": 0.8129, + "step": 15022 + }, + { + "epoch": 1.2801874733702598, + "grad_norm": 19.375, + "learning_rate": 7.229204189290347e-07, + "loss": 0.5367, + "step": 15023 + }, + { + "epoch": 1.2802726885385598, + "grad_norm": 12.375, + "learning_rate": 7.227680570680851e-07, + "loss": 0.3759, + "step": 15024 + }, + { + "epoch": 1.2803579037068598, + "grad_norm": 20.5, + "learning_rate": 7.226157047347465e-07, + "loss": 0.6855, + "step": 15025 + }, + { + "epoch": 1.2804431188751597, + "grad_norm": 13.125, + "learning_rate": 7.224633619317723e-07, + "loss": 0.7443, + "step": 15026 + }, + { + "epoch": 1.2805283340434597, + "grad_norm": 15.0, + "learning_rate": 7.223110286619157e-07, + "loss": 0.4937, + "step": 15027 + }, + { + "epoch": 1.2806135492117596, + "grad_norm": 13.375, + "learning_rate": 7.2215870492793e-07, + "loss": 0.5423, + "step": 15028 + }, + { + "epoch": 1.2806987643800596, + "grad_norm": 18.5, + "learning_rate": 7.220063907325666e-07, + "loss": 0.8018, + "step": 15029 + }, + { + "epoch": 1.2807839795483595, + "grad_norm": 15.8125, + "learning_rate": 7.218540860785795e-07, + "loss": 0.6262, + "step": 15030 + }, + { + "epoch": 1.2808691947166595, + "grad_norm": 24.375, + "learning_rate": 7.217017909687193e-07, + "loss": 0.5389, + "step": 15031 + }, + { + "epoch": 1.2809544098849595, + "grad_norm": 18.375, + "learning_rate": 7.215495054057392e-07, + "loss": 0.6472, + "step": 15032 + }, + { + "epoch": 1.2810396250532594, + "grad_norm": 22.625, + "learning_rate": 7.213972293923913e-07, + "loss": 0.8111, + "step": 15033 + }, + { + "epoch": 1.2811248402215594, + "grad_norm": 21.5, + "learning_rate": 7.212449629314267e-07, + "loss": 0.8046, + "step": 15034 + }, + { + "epoch": 1.2812100553898593, + "grad_norm": 14.0, + "learning_rate": 7.210927060255971e-07, + "loss": 0.3659, + "step": 15035 + }, + { + "epoch": 1.2812952705581593, + "grad_norm": 9.375, + "learning_rate": 7.209404586776543e-07, + "loss": 0.2725, + "step": 15036 + }, + { + "epoch": 1.2813804857264592, + "grad_norm": 24.5, + "learning_rate": 7.2078822089035e-07, + "loss": 0.6344, + "step": 15037 + }, + { + "epoch": 1.2814657008947592, + "grad_norm": 23.5, + "learning_rate": 7.206359926664345e-07, + "loss": 1.3018, + "step": 15038 + }, + { + "epoch": 1.2815509160630592, + "grad_norm": 14.875, + "learning_rate": 7.204837740086584e-07, + "loss": 0.6955, + "step": 15039 + }, + { + "epoch": 1.281636131231359, + "grad_norm": 16.375, + "learning_rate": 7.203315649197731e-07, + "loss": 0.919, + "step": 15040 + }, + { + "epoch": 1.281721346399659, + "grad_norm": 21.75, + "learning_rate": 7.201793654025297e-07, + "loss": 0.8063, + "step": 15041 + }, + { + "epoch": 1.281806561567959, + "grad_norm": 12.9375, + "learning_rate": 7.200271754596771e-07, + "loss": 0.5939, + "step": 15042 + }, + { + "epoch": 1.281891776736259, + "grad_norm": 14.125, + "learning_rate": 7.198749950939667e-07, + "loss": 0.6959, + "step": 15043 + }, + { + "epoch": 1.281976991904559, + "grad_norm": 16.625, + "learning_rate": 7.19722824308148e-07, + "loss": 0.4886, + "step": 15044 + }, + { + "epoch": 1.282062207072859, + "grad_norm": 16.125, + "learning_rate": 7.195706631049718e-07, + "loss": 0.6447, + "step": 15045 + }, + { + "epoch": 1.2821474222411589, + "grad_norm": 29.25, + "learning_rate": 7.194185114871868e-07, + "loss": 0.6003, + "step": 15046 + }, + { + "epoch": 1.2822326374094588, + "grad_norm": 14.625, + "learning_rate": 7.192663694575428e-07, + "loss": 0.7729, + "step": 15047 + }, + { + "epoch": 1.2823178525777588, + "grad_norm": 19.0, + "learning_rate": 7.191142370187887e-07, + "loss": 0.5621, + "step": 15048 + }, + { + "epoch": 1.2824030677460587, + "grad_norm": 20.375, + "learning_rate": 7.189621141736744e-07, + "loss": 0.6954, + "step": 15049 + }, + { + "epoch": 1.2824882829143587, + "grad_norm": 20.0, + "learning_rate": 7.188100009249494e-07, + "loss": 0.5112, + "step": 15050 + }, + { + "epoch": 1.2825734980826586, + "grad_norm": 10.75, + "learning_rate": 7.18657897275361e-07, + "loss": 0.2816, + "step": 15051 + }, + { + "epoch": 1.2826587132509586, + "grad_norm": 17.375, + "learning_rate": 7.185058032276589e-07, + "loss": 0.7607, + "step": 15052 + }, + { + "epoch": 1.2827439284192586, + "grad_norm": 11.3125, + "learning_rate": 7.183537187845913e-07, + "loss": 0.3999, + "step": 15053 + }, + { + "epoch": 1.2828291435875585, + "grad_norm": 15.75, + "learning_rate": 7.182016439489076e-07, + "loss": 0.4299, + "step": 15054 + }, + { + "epoch": 1.2829143587558585, + "grad_norm": 12.25, + "learning_rate": 7.180495787233541e-07, + "loss": 0.3483, + "step": 15055 + }, + { + "epoch": 1.2829995739241584, + "grad_norm": 30.75, + "learning_rate": 7.178975231106794e-07, + "loss": 0.7989, + "step": 15056 + }, + { + "epoch": 1.2830847890924584, + "grad_norm": 19.875, + "learning_rate": 7.177454771136316e-07, + "loss": 0.6729, + "step": 15057 + }, + { + "epoch": 1.2831700042607583, + "grad_norm": 15.1875, + "learning_rate": 7.175934407349589e-07, + "loss": 0.5917, + "step": 15058 + }, + { + "epoch": 1.2832552194290583, + "grad_norm": 14.625, + "learning_rate": 7.174414139774075e-07, + "loss": 0.615, + "step": 15059 + }, + { + "epoch": 1.2833404345973582, + "grad_norm": 21.0, + "learning_rate": 7.172893968437252e-07, + "loss": 0.8324, + "step": 15060 + }, + { + "epoch": 1.2834256497656582, + "grad_norm": 15.0625, + "learning_rate": 7.171373893366594e-07, + "loss": 0.7575, + "step": 15061 + }, + { + "epoch": 1.2835108649339582, + "grad_norm": 23.5, + "learning_rate": 7.169853914589572e-07, + "loss": 0.9775, + "step": 15062 + }, + { + "epoch": 1.2835960801022581, + "grad_norm": 16.375, + "learning_rate": 7.168334032133651e-07, + "loss": 0.7289, + "step": 15063 + }, + { + "epoch": 1.2836812952705583, + "grad_norm": 14.3125, + "learning_rate": 7.166814246026288e-07, + "loss": 0.6016, + "step": 15064 + }, + { + "epoch": 1.2837665104388583, + "grad_norm": 15.3125, + "learning_rate": 7.165294556294955e-07, + "loss": 0.6615, + "step": 15065 + }, + { + "epoch": 1.2838517256071582, + "grad_norm": 14.8125, + "learning_rate": 7.163774962967114e-07, + "loss": 0.4791, + "step": 15066 + }, + { + "epoch": 1.2839369407754582, + "grad_norm": 20.0, + "learning_rate": 7.162255466070231e-07, + "loss": 1.0902, + "step": 15067 + }, + { + "epoch": 1.2840221559437581, + "grad_norm": 18.25, + "learning_rate": 7.160736065631755e-07, + "loss": 0.8088, + "step": 15068 + }, + { + "epoch": 1.284107371112058, + "grad_norm": 10.375, + "learning_rate": 7.159216761679144e-07, + "loss": 0.4919, + "step": 15069 + }, + { + "epoch": 1.284192586280358, + "grad_norm": 13.875, + "learning_rate": 7.157697554239864e-07, + "loss": 0.5626, + "step": 15070 + }, + { + "epoch": 1.284277801448658, + "grad_norm": 14.9375, + "learning_rate": 7.156178443341362e-07, + "loss": 0.5965, + "step": 15071 + }, + { + "epoch": 1.284363016616958, + "grad_norm": 13.375, + "learning_rate": 7.154659429011084e-07, + "loss": 0.4038, + "step": 15072 + }, + { + "epoch": 1.284448231785258, + "grad_norm": 14.5, + "learning_rate": 7.153140511276485e-07, + "loss": 0.6984, + "step": 15073 + }, + { + "epoch": 1.2845334469535579, + "grad_norm": 16.375, + "learning_rate": 7.151621690165013e-07, + "loss": 0.5937, + "step": 15074 + }, + { + "epoch": 1.2846186621218578, + "grad_norm": 13.1875, + "learning_rate": 7.15010296570412e-07, + "loss": 0.6077, + "step": 15075 + }, + { + "epoch": 1.2847038772901578, + "grad_norm": 18.5, + "learning_rate": 7.148584337921242e-07, + "loss": 0.6863, + "step": 15076 + }, + { + "epoch": 1.2847890924584577, + "grad_norm": 17.0, + "learning_rate": 7.147065806843826e-07, + "loss": 0.7406, + "step": 15077 + }, + { + "epoch": 1.2848743076267577, + "grad_norm": 12.0625, + "learning_rate": 7.145547372499319e-07, + "loss": 0.5194, + "step": 15078 + }, + { + "epoch": 1.2849595227950577, + "grad_norm": 18.75, + "learning_rate": 7.144029034915149e-07, + "loss": 0.5087, + "step": 15079 + }, + { + "epoch": 1.2850447379633576, + "grad_norm": 12.1875, + "learning_rate": 7.142510794118767e-07, + "loss": 0.507, + "step": 15080 + }, + { + "epoch": 1.2851299531316576, + "grad_norm": 14.3125, + "learning_rate": 7.140992650137596e-07, + "loss": 0.6032, + "step": 15081 + }, + { + "epoch": 1.2852151682999575, + "grad_norm": 13.75, + "learning_rate": 7.139474602999077e-07, + "loss": 0.4407, + "step": 15082 + }, + { + "epoch": 1.2853003834682575, + "grad_norm": 17.375, + "learning_rate": 7.137956652730644e-07, + "loss": 0.5077, + "step": 15083 + }, + { + "epoch": 1.2853855986365574, + "grad_norm": 14.3125, + "learning_rate": 7.13643879935973e-07, + "loss": 0.5777, + "step": 15084 + }, + { + "epoch": 1.2854708138048574, + "grad_norm": 14.3125, + "learning_rate": 7.134921042913753e-07, + "loss": 0.4061, + "step": 15085 + }, + { + "epoch": 1.2855560289731573, + "grad_norm": 14.625, + "learning_rate": 7.133403383420156e-07, + "loss": 0.5027, + "step": 15086 + }, + { + "epoch": 1.2856412441414573, + "grad_norm": 13.1875, + "learning_rate": 7.131885820906351e-07, + "loss": 0.3994, + "step": 15087 + }, + { + "epoch": 1.2857264593097573, + "grad_norm": 15.25, + "learning_rate": 7.13036835539977e-07, + "loss": 0.6309, + "step": 15088 + }, + { + "epoch": 1.2858116744780572, + "grad_norm": 14.75, + "learning_rate": 7.12885098692783e-07, + "loss": 0.3851, + "step": 15089 + }, + { + "epoch": 1.2858968896463572, + "grad_norm": 11.9375, + "learning_rate": 7.127333715517954e-07, + "loss": 0.3319, + "step": 15090 + }, + { + "epoch": 1.2859821048146571, + "grad_norm": 22.875, + "learning_rate": 7.125816541197559e-07, + "loss": 0.6869, + "step": 15091 + }, + { + "epoch": 1.286067319982957, + "grad_norm": 16.625, + "learning_rate": 7.124299463994067e-07, + "loss": 0.4867, + "step": 15092 + }, + { + "epoch": 1.286152535151257, + "grad_norm": 17.0, + "learning_rate": 7.122782483934885e-07, + "loss": 0.9977, + "step": 15093 + }, + { + "epoch": 1.286237750319557, + "grad_norm": 14.3125, + "learning_rate": 7.121265601047436e-07, + "loss": 0.4308, + "step": 15094 + }, + { + "epoch": 1.286322965487857, + "grad_norm": 17.0, + "learning_rate": 7.119748815359123e-07, + "loss": 0.3849, + "step": 15095 + }, + { + "epoch": 1.286408180656157, + "grad_norm": 18.25, + "learning_rate": 7.118232126897358e-07, + "loss": 0.5573, + "step": 15096 + }, + { + "epoch": 1.2864933958244569, + "grad_norm": 10.9375, + "learning_rate": 7.116715535689556e-07, + "loss": 0.3496, + "step": 15097 + }, + { + "epoch": 1.2865786109927568, + "grad_norm": 12.9375, + "learning_rate": 7.11519904176311e-07, + "loss": 0.5815, + "step": 15098 + }, + { + "epoch": 1.2866638261610568, + "grad_norm": 17.375, + "learning_rate": 7.113682645145436e-07, + "loss": 0.7165, + "step": 15099 + }, + { + "epoch": 1.2867490413293567, + "grad_norm": 13.875, + "learning_rate": 7.112166345863936e-07, + "loss": 0.5607, + "step": 15100 + }, + { + "epoch": 1.2868342564976567, + "grad_norm": 15.8125, + "learning_rate": 7.110650143946001e-07, + "loss": 0.7267, + "step": 15101 + }, + { + "epoch": 1.2869194716659567, + "grad_norm": 16.625, + "learning_rate": 7.109134039419041e-07, + "loss": 0.7336, + "step": 15102 + }, + { + "epoch": 1.2870046868342566, + "grad_norm": 16.25, + "learning_rate": 7.107618032310453e-07, + "loss": 0.8601, + "step": 15103 + }, + { + "epoch": 1.2870899020025566, + "grad_norm": 16.375, + "learning_rate": 7.106102122647624e-07, + "loss": 0.5396, + "step": 15104 + }, + { + "epoch": 1.2871751171708565, + "grad_norm": 14.0625, + "learning_rate": 7.104586310457959e-07, + "loss": 0.4023, + "step": 15105 + }, + { + "epoch": 1.2872603323391565, + "grad_norm": 18.125, + "learning_rate": 7.103070595768841e-07, + "loss": 0.6298, + "step": 15106 + }, + { + "epoch": 1.2873455475074564, + "grad_norm": 10.1875, + "learning_rate": 7.101554978607663e-07, + "loss": 0.2957, + "step": 15107 + }, + { + "epoch": 1.2874307626757564, + "grad_norm": 15.625, + "learning_rate": 7.100039459001816e-07, + "loss": 0.6626, + "step": 15108 + }, + { + "epoch": 1.2875159778440564, + "grad_norm": 15.8125, + "learning_rate": 7.098524036978691e-07, + "loss": 0.5311, + "step": 15109 + }, + { + "epoch": 1.2876011930123563, + "grad_norm": 13.9375, + "learning_rate": 7.097008712565662e-07, + "loss": 0.5253, + "step": 15110 + }, + { + "epoch": 1.2876864081806563, + "grad_norm": 13.25, + "learning_rate": 7.095493485790126e-07, + "loss": 0.338, + "step": 15111 + }, + { + "epoch": 1.2877716233489562, + "grad_norm": 18.0, + "learning_rate": 7.093978356679451e-07, + "loss": 0.7971, + "step": 15112 + }, + { + "epoch": 1.2878568385172562, + "grad_norm": 14.1875, + "learning_rate": 7.092463325261023e-07, + "loss": 0.5011, + "step": 15113 + }, + { + "epoch": 1.2879420536855561, + "grad_norm": 15.8125, + "learning_rate": 7.090948391562225e-07, + "loss": 0.6159, + "step": 15114 + }, + { + "epoch": 1.288027268853856, + "grad_norm": 23.375, + "learning_rate": 7.089433555610426e-07, + "loss": 0.6046, + "step": 15115 + }, + { + "epoch": 1.288112484022156, + "grad_norm": 12.3125, + "learning_rate": 7.087918817433001e-07, + "loss": 0.5112, + "step": 15116 + }, + { + "epoch": 1.288197699190456, + "grad_norm": 18.125, + "learning_rate": 7.086404177057331e-07, + "loss": 0.287, + "step": 15117 + }, + { + "epoch": 1.288282914358756, + "grad_norm": 14.375, + "learning_rate": 7.084889634510776e-07, + "loss": 0.5479, + "step": 15118 + }, + { + "epoch": 1.288368129527056, + "grad_norm": 12.375, + "learning_rate": 7.083375189820717e-07, + "loss": 0.3994, + "step": 15119 + }, + { + "epoch": 1.2884533446953559, + "grad_norm": 17.625, + "learning_rate": 7.08186084301451e-07, + "loss": 0.5072, + "step": 15120 + }, + { + "epoch": 1.2885385598636558, + "grad_norm": 16.5, + "learning_rate": 7.080346594119524e-07, + "loss": 0.676, + "step": 15121 + }, + { + "epoch": 1.2886237750319558, + "grad_norm": 12.0, + "learning_rate": 7.07883244316313e-07, + "loss": 0.3134, + "step": 15122 + }, + { + "epoch": 1.2887089902002558, + "grad_norm": 15.5625, + "learning_rate": 7.077318390172682e-07, + "loss": 0.5933, + "step": 15123 + }, + { + "epoch": 1.2887942053685557, + "grad_norm": 31.25, + "learning_rate": 7.075804435175543e-07, + "loss": 0.7032, + "step": 15124 + }, + { + "epoch": 1.2888794205368557, + "grad_norm": 20.0, + "learning_rate": 7.074290578199072e-07, + "loss": 0.6434, + "step": 15125 + }, + { + "epoch": 1.2889646357051556, + "grad_norm": 12.5625, + "learning_rate": 7.072776819270633e-07, + "loss": 0.4171, + "step": 15126 + }, + { + "epoch": 1.2890498508734556, + "grad_norm": 14.4375, + "learning_rate": 7.071263158417571e-07, + "loss": 0.3842, + "step": 15127 + }, + { + "epoch": 1.2891350660417555, + "grad_norm": 23.125, + "learning_rate": 7.06974959566724e-07, + "loss": 0.769, + "step": 15128 + }, + { + "epoch": 1.2892202812100555, + "grad_norm": 21.125, + "learning_rate": 7.068236131046994e-07, + "loss": 0.3633, + "step": 15129 + }, + { + "epoch": 1.2893054963783555, + "grad_norm": 11.5, + "learning_rate": 7.066722764584183e-07, + "loss": 0.2246, + "step": 15130 + }, + { + "epoch": 1.2893907115466554, + "grad_norm": 14.4375, + "learning_rate": 7.06520949630616e-07, + "loss": 0.4088, + "step": 15131 + }, + { + "epoch": 1.2894759267149554, + "grad_norm": 18.125, + "learning_rate": 7.063696326240262e-07, + "loss": 0.5391, + "step": 15132 + }, + { + "epoch": 1.2895611418832553, + "grad_norm": 13.5625, + "learning_rate": 7.062183254413838e-07, + "loss": 0.477, + "step": 15133 + }, + { + "epoch": 1.2896463570515553, + "grad_norm": 23.125, + "learning_rate": 7.060670280854234e-07, + "loss": 0.681, + "step": 15134 + }, + { + "epoch": 1.2897315722198552, + "grad_norm": 12.5, + "learning_rate": 7.05915740558879e-07, + "loss": 0.3855, + "step": 15135 + }, + { + "epoch": 1.2898167873881552, + "grad_norm": 18.125, + "learning_rate": 7.057644628644838e-07, + "loss": 0.8117, + "step": 15136 + }, + { + "epoch": 1.2899020025564552, + "grad_norm": 13.0, + "learning_rate": 7.056131950049718e-07, + "loss": 0.5801, + "step": 15137 + }, + { + "epoch": 1.289987217724755, + "grad_norm": 17.875, + "learning_rate": 7.05461936983077e-07, + "loss": 0.6438, + "step": 15138 + }, + { + "epoch": 1.290072432893055, + "grad_norm": 16.875, + "learning_rate": 7.053106888015329e-07, + "loss": 1.2043, + "step": 15139 + }, + { + "epoch": 1.290157648061355, + "grad_norm": 16.875, + "learning_rate": 7.051594504630721e-07, + "loss": 0.8353, + "step": 15140 + }, + { + "epoch": 1.290242863229655, + "grad_norm": 20.625, + "learning_rate": 7.050082219704276e-07, + "loss": 0.5865, + "step": 15141 + }, + { + "epoch": 1.290328078397955, + "grad_norm": 21.625, + "learning_rate": 7.04857003326333e-07, + "loss": 0.7834, + "step": 15142 + }, + { + "epoch": 1.290413293566255, + "grad_norm": 19.875, + "learning_rate": 7.047057945335206e-07, + "loss": 0.6894, + "step": 15143 + }, + { + "epoch": 1.2904985087345549, + "grad_norm": 13.5, + "learning_rate": 7.04554595594723e-07, + "loss": 0.2951, + "step": 15144 + }, + { + "epoch": 1.2905837239028548, + "grad_norm": 31.25, + "learning_rate": 7.04403406512672e-07, + "loss": 0.3598, + "step": 15145 + }, + { + "epoch": 1.2906689390711548, + "grad_norm": 21.625, + "learning_rate": 7.042522272900999e-07, + "loss": 0.6158, + "step": 15146 + }, + { + "epoch": 1.2907541542394547, + "grad_norm": 14.5, + "learning_rate": 7.041010579297394e-07, + "loss": 0.5333, + "step": 15147 + }, + { + "epoch": 1.2908393694077547, + "grad_norm": 20.875, + "learning_rate": 7.039498984343212e-07, + "loss": 0.694, + "step": 15148 + }, + { + "epoch": 1.2909245845760546, + "grad_norm": 18.75, + "learning_rate": 7.037987488065775e-07, + "loss": 0.7569, + "step": 15149 + }, + { + "epoch": 1.2910097997443546, + "grad_norm": 23.125, + "learning_rate": 7.036476090492395e-07, + "loss": 1.1767, + "step": 15150 + }, + { + "epoch": 1.2910950149126545, + "grad_norm": 17.0, + "learning_rate": 7.034964791650392e-07, + "loss": 0.5989, + "step": 15151 + }, + { + "epoch": 1.2911802300809545, + "grad_norm": 15.625, + "learning_rate": 7.033453591567071e-07, + "loss": 0.7496, + "step": 15152 + }, + { + "epoch": 1.2912654452492545, + "grad_norm": 13.0, + "learning_rate": 7.031942490269735e-07, + "loss": 0.57, + "step": 15153 + }, + { + "epoch": 1.2913506604175544, + "grad_norm": 14.3125, + "learning_rate": 7.030431487785698e-07, + "loss": 0.719, + "step": 15154 + }, + { + "epoch": 1.2914358755858544, + "grad_norm": 14.875, + "learning_rate": 7.028920584142263e-07, + "loss": 0.502, + "step": 15155 + }, + { + "epoch": 1.2915210907541543, + "grad_norm": 20.75, + "learning_rate": 7.02740977936674e-07, + "loss": 0.6913, + "step": 15156 + }, + { + "epoch": 1.2916063059224543, + "grad_norm": 22.875, + "learning_rate": 7.025899073486419e-07, + "loss": 0.5518, + "step": 15157 + }, + { + "epoch": 1.2916915210907542, + "grad_norm": 14.375, + "learning_rate": 7.024388466528608e-07, + "loss": 0.6213, + "step": 15158 + }, + { + "epoch": 1.2917767362590542, + "grad_norm": 16.375, + "learning_rate": 7.022877958520611e-07, + "loss": 0.6652, + "step": 15159 + }, + { + "epoch": 1.2918619514273542, + "grad_norm": 10.5, + "learning_rate": 7.021367549489708e-07, + "loss": 0.2571, + "step": 15160 + }, + { + "epoch": 1.2919471665956541, + "grad_norm": 18.5, + "learning_rate": 7.019857239463209e-07, + "loss": 0.9149, + "step": 15161 + }, + { + "epoch": 1.292032381763954, + "grad_norm": 19.625, + "learning_rate": 7.018347028468395e-07, + "loss": 0.5043, + "step": 15162 + }, + { + "epoch": 1.292117596932254, + "grad_norm": 12.75, + "learning_rate": 7.016836916532563e-07, + "loss": 0.5463, + "step": 15163 + }, + { + "epoch": 1.292202812100554, + "grad_norm": 14.6875, + "learning_rate": 7.015326903683006e-07, + "loss": 0.5039, + "step": 15164 + }, + { + "epoch": 1.292288027268854, + "grad_norm": 15.5625, + "learning_rate": 7.013816989947002e-07, + "loss": 0.5184, + "step": 15165 + }, + { + "epoch": 1.292373242437154, + "grad_norm": 18.75, + "learning_rate": 7.012307175351843e-07, + "loss": 0.6132, + "step": 15166 + }, + { + "epoch": 1.2924584576054539, + "grad_norm": 13.625, + "learning_rate": 7.01079745992481e-07, + "loss": 0.3651, + "step": 15167 + }, + { + "epoch": 1.2925436727737538, + "grad_norm": 16.75, + "learning_rate": 7.009287843693192e-07, + "loss": 0.3371, + "step": 15168 + }, + { + "epoch": 1.2926288879420538, + "grad_norm": 17.375, + "learning_rate": 7.007778326684264e-07, + "loss": 0.7953, + "step": 15169 + }, + { + "epoch": 1.2927141031103537, + "grad_norm": 16.0, + "learning_rate": 7.006268908925301e-07, + "loss": 0.8093, + "step": 15170 + }, + { + "epoch": 1.2927993182786537, + "grad_norm": 14.125, + "learning_rate": 7.004759590443583e-07, + "loss": 0.4036, + "step": 15171 + }, + { + "epoch": 1.2928845334469536, + "grad_norm": 16.75, + "learning_rate": 7.003250371266383e-07, + "loss": 0.5849, + "step": 15172 + }, + { + "epoch": 1.2929697486152536, + "grad_norm": 15.25, + "learning_rate": 7.001741251420982e-07, + "loss": 0.6896, + "step": 15173 + }, + { + "epoch": 1.2930549637835536, + "grad_norm": 15.0625, + "learning_rate": 7.000232230934642e-07, + "loss": 0.4887, + "step": 15174 + }, + { + "epoch": 1.2931401789518535, + "grad_norm": 20.5, + "learning_rate": 6.998723309834634e-07, + "loss": 0.7003, + "step": 15175 + }, + { + "epoch": 1.2932253941201535, + "grad_norm": 13.8125, + "learning_rate": 6.997214488148235e-07, + "loss": 0.4471, + "step": 15176 + }, + { + "epoch": 1.2933106092884534, + "grad_norm": 19.0, + "learning_rate": 6.995705765902702e-07, + "loss": 0.6771, + "step": 15177 + }, + { + "epoch": 1.2933958244567534, + "grad_norm": 15.625, + "learning_rate": 6.994197143125297e-07, + "loss": 0.5664, + "step": 15178 + }, + { + "epoch": 1.2934810396250533, + "grad_norm": 16.75, + "learning_rate": 6.992688619843286e-07, + "loss": 0.5801, + "step": 15179 + }, + { + "epoch": 1.2935662547933533, + "grad_norm": 13.0, + "learning_rate": 6.991180196083929e-07, + "loss": 0.5631, + "step": 15180 + }, + { + "epoch": 1.2936514699616533, + "grad_norm": 15.0, + "learning_rate": 6.98967187187449e-07, + "loss": 0.7203, + "step": 15181 + }, + { + "epoch": 1.2937366851299532, + "grad_norm": 21.625, + "learning_rate": 6.988163647242216e-07, + "loss": 0.5454, + "step": 15182 + }, + { + "epoch": 1.2938219002982532, + "grad_norm": 16.625, + "learning_rate": 6.986655522214369e-07, + "loss": 0.8698, + "step": 15183 + }, + { + "epoch": 1.2939071154665531, + "grad_norm": 15.1875, + "learning_rate": 6.985147496818205e-07, + "loss": 0.7245, + "step": 15184 + }, + { + "epoch": 1.293992330634853, + "grad_norm": 19.5, + "learning_rate": 6.983639571080966e-07, + "loss": 0.6196, + "step": 15185 + }, + { + "epoch": 1.294077545803153, + "grad_norm": 15.875, + "learning_rate": 6.982131745029913e-07, + "loss": 0.5569, + "step": 15186 + }, + { + "epoch": 1.294162760971453, + "grad_norm": 12.0, + "learning_rate": 6.980624018692282e-07, + "loss": 0.4614, + "step": 15187 + }, + { + "epoch": 1.294247976139753, + "grad_norm": 14.3125, + "learning_rate": 6.979116392095325e-07, + "loss": 0.3476, + "step": 15188 + }, + { + "epoch": 1.294333191308053, + "grad_norm": 16.0, + "learning_rate": 6.977608865266289e-07, + "loss": 0.4849, + "step": 15189 + }, + { + "epoch": 1.2944184064763529, + "grad_norm": 17.75, + "learning_rate": 6.976101438232416e-07, + "loss": 0.8338, + "step": 15190 + }, + { + "epoch": 1.2945036216446528, + "grad_norm": 13.4375, + "learning_rate": 6.974594111020941e-07, + "loss": 0.5814, + "step": 15191 + }, + { + "epoch": 1.2945888368129528, + "grad_norm": 13.5625, + "learning_rate": 6.973086883659111e-07, + "loss": 0.489, + "step": 15192 + }, + { + "epoch": 1.2946740519812527, + "grad_norm": 23.125, + "learning_rate": 6.971579756174154e-07, + "loss": 0.696, + "step": 15193 + }, + { + "epoch": 1.2947592671495527, + "grad_norm": 13.4375, + "learning_rate": 6.970072728593316e-07, + "loss": 0.6349, + "step": 15194 + }, + { + "epoch": 1.2948444823178527, + "grad_norm": 15.375, + "learning_rate": 6.968565800943819e-07, + "loss": 0.5272, + "step": 15195 + }, + { + "epoch": 1.2949296974861526, + "grad_norm": 20.625, + "learning_rate": 6.9670589732529e-07, + "loss": 0.6985, + "step": 15196 + }, + { + "epoch": 1.2950149126544526, + "grad_norm": 20.5, + "learning_rate": 6.965552245547791e-07, + "loss": 0.5732, + "step": 15197 + }, + { + "epoch": 1.2951001278227525, + "grad_norm": 25.0, + "learning_rate": 6.964045617855721e-07, + "loss": 0.8428, + "step": 15198 + }, + { + "epoch": 1.2951853429910525, + "grad_norm": 17.25, + "learning_rate": 6.962539090203908e-07, + "loss": 0.7555, + "step": 15199 + }, + { + "epoch": 1.2952705581593524, + "grad_norm": 11.5625, + "learning_rate": 6.961032662619588e-07, + "loss": 0.3381, + "step": 15200 + }, + { + "epoch": 1.2953557733276524, + "grad_norm": 11.3125, + "learning_rate": 6.959526335129974e-07, + "loss": 0.4085, + "step": 15201 + }, + { + "epoch": 1.2954409884959524, + "grad_norm": 14.75, + "learning_rate": 6.958020107762289e-07, + "loss": 0.6015, + "step": 15202 + }, + { + "epoch": 1.2955262036642523, + "grad_norm": 12.625, + "learning_rate": 6.956513980543759e-07, + "loss": 0.5905, + "step": 15203 + }, + { + "epoch": 1.2956114188325523, + "grad_norm": 17.375, + "learning_rate": 6.955007953501589e-07, + "loss": 0.6573, + "step": 15204 + }, + { + "epoch": 1.2956966340008522, + "grad_norm": 18.625, + "learning_rate": 6.953502026663003e-07, + "loss": 0.6264, + "step": 15205 + }, + { + "epoch": 1.2957818491691522, + "grad_norm": 11.9375, + "learning_rate": 6.951996200055214e-07, + "loss": 0.478, + "step": 15206 + }, + { + "epoch": 1.2958670643374521, + "grad_norm": 13.75, + "learning_rate": 6.950490473705435e-07, + "loss": 0.5811, + "step": 15207 + }, + { + "epoch": 1.295952279505752, + "grad_norm": 12.5, + "learning_rate": 6.948984847640875e-07, + "loss": 0.499, + "step": 15208 + }, + { + "epoch": 1.296037494674052, + "grad_norm": 16.25, + "learning_rate": 6.947479321888736e-07, + "loss": 0.8175, + "step": 15209 + }, + { + "epoch": 1.296122709842352, + "grad_norm": 19.125, + "learning_rate": 6.945973896476228e-07, + "loss": 0.7917, + "step": 15210 + }, + { + "epoch": 1.296207925010652, + "grad_norm": 25.875, + "learning_rate": 6.944468571430563e-07, + "loss": 0.9344, + "step": 15211 + }, + { + "epoch": 1.296293140178952, + "grad_norm": 17.25, + "learning_rate": 6.942963346778932e-07, + "loss": 0.5976, + "step": 15212 + }, + { + "epoch": 1.2963783553472519, + "grad_norm": 32.5, + "learning_rate": 6.941458222548542e-07, + "loss": 0.7255, + "step": 15213 + }, + { + "epoch": 1.2964635705155518, + "grad_norm": 14.0, + "learning_rate": 6.939953198766589e-07, + "loss": 0.5307, + "step": 15214 + }, + { + "epoch": 1.2965487856838518, + "grad_norm": 16.625, + "learning_rate": 6.93844827546028e-07, + "loss": 0.4613, + "step": 15215 + }, + { + "epoch": 1.2966340008521517, + "grad_norm": 16.375, + "learning_rate": 6.936943452656802e-07, + "loss": 0.5695, + "step": 15216 + }, + { + "epoch": 1.2967192160204517, + "grad_norm": 18.25, + "learning_rate": 6.935438730383345e-07, + "loss": 0.6086, + "step": 15217 + }, + { + "epoch": 1.2968044311887517, + "grad_norm": 13.375, + "learning_rate": 6.933934108667108e-07, + "loss": 0.4479, + "step": 15218 + }, + { + "epoch": 1.2968896463570516, + "grad_norm": 19.875, + "learning_rate": 6.932429587535277e-07, + "loss": 0.6651, + "step": 15219 + }, + { + "epoch": 1.2969748615253516, + "grad_norm": 17.125, + "learning_rate": 6.930925167015046e-07, + "loss": 0.7064, + "step": 15220 + }, + { + "epoch": 1.2970600766936515, + "grad_norm": 17.875, + "learning_rate": 6.929420847133594e-07, + "loss": 1.0303, + "step": 15221 + }, + { + "epoch": 1.2971452918619515, + "grad_norm": 12.0625, + "learning_rate": 6.927916627918107e-07, + "loss": 0.3923, + "step": 15222 + }, + { + "epoch": 1.2972305070302514, + "grad_norm": 15.5625, + "learning_rate": 6.926412509395772e-07, + "loss": 0.484, + "step": 15223 + }, + { + "epoch": 1.2973157221985514, + "grad_norm": 14.25, + "learning_rate": 6.924908491593776e-07, + "loss": 0.5768, + "step": 15224 + }, + { + "epoch": 1.2974009373668514, + "grad_norm": 21.625, + "learning_rate": 6.923404574539282e-07, + "loss": 1.0559, + "step": 15225 + }, + { + "epoch": 1.2974861525351513, + "grad_norm": 13.4375, + "learning_rate": 6.921900758259472e-07, + "loss": 0.524, + "step": 15226 + }, + { + "epoch": 1.2975713677034513, + "grad_norm": 15.4375, + "learning_rate": 6.920397042781526e-07, + "loss": 0.7534, + "step": 15227 + }, + { + "epoch": 1.2976565828717512, + "grad_norm": 14.1875, + "learning_rate": 6.918893428132623e-07, + "loss": 0.4391, + "step": 15228 + }, + { + "epoch": 1.2977417980400512, + "grad_norm": 12.4375, + "learning_rate": 6.91738991433992e-07, + "loss": 0.4048, + "step": 15229 + }, + { + "epoch": 1.2978270132083511, + "grad_norm": 13.8125, + "learning_rate": 6.915886501430596e-07, + "loss": 0.4819, + "step": 15230 + }, + { + "epoch": 1.297912228376651, + "grad_norm": 22.25, + "learning_rate": 6.914383189431818e-07, + "loss": 0.7331, + "step": 15231 + }, + { + "epoch": 1.297997443544951, + "grad_norm": 22.25, + "learning_rate": 6.912879978370758e-07, + "loss": 0.9725, + "step": 15232 + }, + { + "epoch": 1.298082658713251, + "grad_norm": 14.6875, + "learning_rate": 6.911376868274574e-07, + "loss": 0.7844, + "step": 15233 + }, + { + "epoch": 1.298167873881551, + "grad_norm": 19.5, + "learning_rate": 6.909873859170427e-07, + "loss": 0.844, + "step": 15234 + }, + { + "epoch": 1.298253089049851, + "grad_norm": 14.4375, + "learning_rate": 6.90837095108548e-07, + "loss": 0.4212, + "step": 15235 + }, + { + "epoch": 1.2983383042181509, + "grad_norm": 19.125, + "learning_rate": 6.906868144046891e-07, + "loss": 0.607, + "step": 15236 + }, + { + "epoch": 1.2984235193864508, + "grad_norm": 28.875, + "learning_rate": 6.905365438081826e-07, + "loss": 0.6844, + "step": 15237 + }, + { + "epoch": 1.2985087345547508, + "grad_norm": 18.625, + "learning_rate": 6.903862833217427e-07, + "loss": 0.5836, + "step": 15238 + }, + { + "epoch": 1.2985939497230508, + "grad_norm": 16.125, + "learning_rate": 6.902360329480854e-07, + "loss": 0.5235, + "step": 15239 + }, + { + "epoch": 1.2986791648913507, + "grad_norm": 12.1875, + "learning_rate": 6.900857926899264e-07, + "loss": 0.3698, + "step": 15240 + }, + { + "epoch": 1.2987643800596507, + "grad_norm": 16.75, + "learning_rate": 6.899355625499801e-07, + "loss": 0.6695, + "step": 15241 + }, + { + "epoch": 1.2988495952279506, + "grad_norm": 14.5625, + "learning_rate": 6.897853425309611e-07, + "loss": 0.7095, + "step": 15242 + }, + { + "epoch": 1.2989348103962506, + "grad_norm": 14.25, + "learning_rate": 6.896351326355841e-07, + "loss": 0.5926, + "step": 15243 + }, + { + "epoch": 1.2990200255645505, + "grad_norm": 14.3125, + "learning_rate": 6.894849328665637e-07, + "loss": 0.5323, + "step": 15244 + }, + { + "epoch": 1.2991052407328505, + "grad_norm": 16.125, + "learning_rate": 6.893347432266148e-07, + "loss": 0.6236, + "step": 15245 + }, + { + "epoch": 1.2991904559011505, + "grad_norm": 20.25, + "learning_rate": 6.891845637184505e-07, + "loss": 0.4984, + "step": 15246 + }, + { + "epoch": 1.2992756710694504, + "grad_norm": 25.5, + "learning_rate": 6.89034394344785e-07, + "loss": 1.1318, + "step": 15247 + }, + { + "epoch": 1.2993608862377504, + "grad_norm": 17.25, + "learning_rate": 6.888842351083319e-07, + "loss": 0.8634, + "step": 15248 + }, + { + "epoch": 1.2994461014060503, + "grad_norm": 14.75, + "learning_rate": 6.887340860118056e-07, + "loss": 0.5517, + "step": 15249 + }, + { + "epoch": 1.2995313165743503, + "grad_norm": 25.375, + "learning_rate": 6.885839470579187e-07, + "loss": 0.7941, + "step": 15250 + }, + { + "epoch": 1.2996165317426502, + "grad_norm": 14.8125, + "learning_rate": 6.88433818249384e-07, + "loss": 0.5883, + "step": 15251 + }, + { + "epoch": 1.2997017469109502, + "grad_norm": 13.0, + "learning_rate": 6.882836995889147e-07, + "loss": 0.466, + "step": 15252 + }, + { + "epoch": 1.2997869620792502, + "grad_norm": 16.5, + "learning_rate": 6.881335910792242e-07, + "loss": 0.3988, + "step": 15253 + }, + { + "epoch": 1.29987217724755, + "grad_norm": 14.0625, + "learning_rate": 6.879834927230247e-07, + "loss": 0.7602, + "step": 15254 + }, + { + "epoch": 1.29995739241585, + "grad_norm": 15.0, + "learning_rate": 6.878334045230285e-07, + "loss": 0.6885, + "step": 15255 + }, + { + "epoch": 1.30004260758415, + "grad_norm": 17.625, + "learning_rate": 6.87683326481948e-07, + "loss": 0.9881, + "step": 15256 + }, + { + "epoch": 1.30012782275245, + "grad_norm": 17.375, + "learning_rate": 6.875332586024955e-07, + "loss": 0.5717, + "step": 15257 + }, + { + "epoch": 1.30021303792075, + "grad_norm": 24.625, + "learning_rate": 6.873832008873828e-07, + "loss": 0.5091, + "step": 15258 + }, + { + "epoch": 1.30029825308905, + "grad_norm": 21.375, + "learning_rate": 6.872331533393209e-07, + "loss": 0.5149, + "step": 15259 + }, + { + "epoch": 1.3003834682573499, + "grad_norm": 21.375, + "learning_rate": 6.870831159610217e-07, + "loss": 0.9007, + "step": 15260 + }, + { + "epoch": 1.3004686834256498, + "grad_norm": 13.75, + "learning_rate": 6.869330887551967e-07, + "loss": 0.5163, + "step": 15261 + }, + { + "epoch": 1.3005538985939498, + "grad_norm": 15.8125, + "learning_rate": 6.867830717245575e-07, + "loss": 0.5828, + "step": 15262 + }, + { + "epoch": 1.3006391137622497, + "grad_norm": 15.4375, + "learning_rate": 6.866330648718142e-07, + "loss": 0.3296, + "step": 15263 + }, + { + "epoch": 1.3007243289305497, + "grad_norm": 13.5625, + "learning_rate": 6.864830681996777e-07, + "loss": 0.4653, + "step": 15264 + }, + { + "epoch": 1.3008095440988496, + "grad_norm": 12.625, + "learning_rate": 6.863330817108594e-07, + "loss": 0.3393, + "step": 15265 + }, + { + "epoch": 1.3008947592671496, + "grad_norm": 15.3125, + "learning_rate": 6.861831054080686e-07, + "loss": 0.7411, + "step": 15266 + }, + { + "epoch": 1.3009799744354495, + "grad_norm": 16.75, + "learning_rate": 6.860331392940167e-07, + "loss": 0.4362, + "step": 15267 + }, + { + "epoch": 1.3010651896037495, + "grad_norm": 19.375, + "learning_rate": 6.858831833714125e-07, + "loss": 0.6598, + "step": 15268 + }, + { + "epoch": 1.3011504047720495, + "grad_norm": 18.0, + "learning_rate": 6.857332376429663e-07, + "loss": 0.4877, + "step": 15269 + }, + { + "epoch": 1.3012356199403494, + "grad_norm": 15.4375, + "learning_rate": 6.855833021113886e-07, + "loss": 0.5588, + "step": 15270 + }, + { + "epoch": 1.3013208351086494, + "grad_norm": 19.0, + "learning_rate": 6.854333767793877e-07, + "loss": 0.5448, + "step": 15271 + }, + { + "epoch": 1.3014060502769493, + "grad_norm": 12.5, + "learning_rate": 6.852834616496733e-07, + "loss": 0.536, + "step": 15272 + }, + { + "epoch": 1.3014912654452493, + "grad_norm": 14.9375, + "learning_rate": 6.85133556724955e-07, + "loss": 0.5267, + "step": 15273 + }, + { + "epoch": 1.3015764806135492, + "grad_norm": 17.125, + "learning_rate": 6.849836620079412e-07, + "loss": 0.6593, + "step": 15274 + }, + { + "epoch": 1.3016616957818492, + "grad_norm": 15.5625, + "learning_rate": 6.848337775013409e-07, + "loss": 0.4954, + "step": 15275 + }, + { + "epoch": 1.3017469109501492, + "grad_norm": 17.875, + "learning_rate": 6.846839032078622e-07, + "loss": 1.0108, + "step": 15276 + }, + { + "epoch": 1.3018321261184491, + "grad_norm": 12.9375, + "learning_rate": 6.845340391302139e-07, + "loss": 0.2836, + "step": 15277 + }, + { + "epoch": 1.301917341286749, + "grad_norm": 13.75, + "learning_rate": 6.843841852711042e-07, + "loss": 0.6242, + "step": 15278 + }, + { + "epoch": 1.302002556455049, + "grad_norm": 19.5, + "learning_rate": 6.842343416332414e-07, + "loss": 0.9831, + "step": 15279 + }, + { + "epoch": 1.302087771623349, + "grad_norm": 13.0, + "learning_rate": 6.840845082193326e-07, + "loss": 0.3596, + "step": 15280 + }, + { + "epoch": 1.302172986791649, + "grad_norm": 13.5625, + "learning_rate": 6.839346850320862e-07, + "loss": 0.6221, + "step": 15281 + }, + { + "epoch": 1.302258201959949, + "grad_norm": 19.625, + "learning_rate": 6.837848720742088e-07, + "loss": 0.9982, + "step": 15282 + }, + { + "epoch": 1.3023434171282489, + "grad_norm": 15.0625, + "learning_rate": 6.836350693484081e-07, + "loss": 0.5154, + "step": 15283 + }, + { + "epoch": 1.3024286322965488, + "grad_norm": 13.5625, + "learning_rate": 6.834852768573919e-07, + "loss": 0.674, + "step": 15284 + }, + { + "epoch": 1.3025138474648488, + "grad_norm": 13.25, + "learning_rate": 6.833354946038657e-07, + "loss": 0.3937, + "step": 15285 + }, + { + "epoch": 1.3025990626331487, + "grad_norm": 15.4375, + "learning_rate": 6.831857225905369e-07, + "loss": 0.6341, + "step": 15286 + }, + { + "epoch": 1.3026842778014487, + "grad_norm": 19.625, + "learning_rate": 6.830359608201127e-07, + "loss": 0.4433, + "step": 15287 + }, + { + "epoch": 1.3027694929697486, + "grad_norm": 17.875, + "learning_rate": 6.828862092952985e-07, + "loss": 0.7904, + "step": 15288 + }, + { + "epoch": 1.3028547081380486, + "grad_norm": 16.25, + "learning_rate": 6.827364680188009e-07, + "loss": 0.574, + "step": 15289 + }, + { + "epoch": 1.3029399233063486, + "grad_norm": 12.625, + "learning_rate": 6.825867369933254e-07, + "loss": 0.3381, + "step": 15290 + }, + { + "epoch": 1.3030251384746485, + "grad_norm": 11.9375, + "learning_rate": 6.824370162215783e-07, + "loss": 0.3572, + "step": 15291 + }, + { + "epoch": 1.3031103536429485, + "grad_norm": 16.75, + "learning_rate": 6.822873057062653e-07, + "loss": 0.6405, + "step": 15292 + }, + { + "epoch": 1.3031955688112484, + "grad_norm": 16.125, + "learning_rate": 6.821376054500911e-07, + "loss": 0.4735, + "step": 15293 + }, + { + "epoch": 1.3032807839795484, + "grad_norm": 13.5, + "learning_rate": 6.819879154557615e-07, + "loss": 0.6004, + "step": 15294 + }, + { + "epoch": 1.3033659991478483, + "grad_norm": 13.125, + "learning_rate": 6.818382357259814e-07, + "loss": 0.606, + "step": 15295 + }, + { + "epoch": 1.3034512143161483, + "grad_norm": 14.375, + "learning_rate": 6.816885662634561e-07, + "loss": 0.6028, + "step": 15296 + }, + { + "epoch": 1.3035364294844483, + "grad_norm": 11.875, + "learning_rate": 6.815389070708903e-07, + "loss": 0.3505, + "step": 15297 + }, + { + "epoch": 1.3036216446527482, + "grad_norm": 16.125, + "learning_rate": 6.813892581509872e-07, + "loss": 0.6732, + "step": 15298 + }, + { + "epoch": 1.3037068598210482, + "grad_norm": 14.0625, + "learning_rate": 6.812396195064521e-07, + "loss": 0.5071, + "step": 15299 + }, + { + "epoch": 1.3037920749893481, + "grad_norm": 15.375, + "learning_rate": 6.810899911399895e-07, + "loss": 0.8301, + "step": 15300 + }, + { + "epoch": 1.303877290157648, + "grad_norm": 17.875, + "learning_rate": 6.809403730543023e-07, + "loss": 0.8184, + "step": 15301 + }, + { + "epoch": 1.303962505325948, + "grad_norm": 15.1875, + "learning_rate": 6.807907652520948e-07, + "loss": 0.5431, + "step": 15302 + }, + { + "epoch": 1.304047720494248, + "grad_norm": 13.5, + "learning_rate": 6.806411677360705e-07, + "loss": 0.4563, + "step": 15303 + }, + { + "epoch": 1.304132935662548, + "grad_norm": 13.75, + "learning_rate": 6.804915805089335e-07, + "loss": 0.5589, + "step": 15304 + }, + { + "epoch": 1.304218150830848, + "grad_norm": 22.5, + "learning_rate": 6.803420035733857e-07, + "loss": 0.5234, + "step": 15305 + }, + { + "epoch": 1.3043033659991479, + "grad_norm": 14.5, + "learning_rate": 6.801924369321314e-07, + "loss": 0.8253, + "step": 15306 + }, + { + "epoch": 1.3043885811674478, + "grad_norm": 19.0, + "learning_rate": 6.80042880587872e-07, + "loss": 0.9325, + "step": 15307 + }, + { + "epoch": 1.3044737963357478, + "grad_norm": 17.5, + "learning_rate": 6.798933345433111e-07, + "loss": 0.7194, + "step": 15308 + }, + { + "epoch": 1.3045590115040477, + "grad_norm": 15.9375, + "learning_rate": 6.797437988011513e-07, + "loss": 0.7592, + "step": 15309 + }, + { + "epoch": 1.3046442266723477, + "grad_norm": 22.875, + "learning_rate": 6.795942733640939e-07, + "loss": 0.7976, + "step": 15310 + }, + { + "epoch": 1.3047294418406477, + "grad_norm": 17.25, + "learning_rate": 6.794447582348418e-07, + "loss": 0.7394, + "step": 15311 + }, + { + "epoch": 1.3048146570089476, + "grad_norm": 15.0625, + "learning_rate": 6.792952534160965e-07, + "loss": 0.71, + "step": 15312 + }, + { + "epoch": 1.3048998721772476, + "grad_norm": 14.5625, + "learning_rate": 6.791457589105605e-07, + "loss": 0.4231, + "step": 15313 + }, + { + "epoch": 1.3049850873455475, + "grad_norm": 19.5, + "learning_rate": 6.789962747209345e-07, + "loss": 0.6185, + "step": 15314 + }, + { + "epoch": 1.3050703025138475, + "grad_norm": 11.6875, + "learning_rate": 6.788468008499196e-07, + "loss": 0.4752, + "step": 15315 + }, + { + "epoch": 1.3051555176821474, + "grad_norm": 15.875, + "learning_rate": 6.786973373002174e-07, + "loss": 0.7148, + "step": 15316 + }, + { + "epoch": 1.3052407328504474, + "grad_norm": 19.125, + "learning_rate": 6.785478840745294e-07, + "loss": 0.6581, + "step": 15317 + }, + { + "epoch": 1.3053259480187474, + "grad_norm": 18.875, + "learning_rate": 6.783984411755551e-07, + "loss": 0.6178, + "step": 15318 + }, + { + "epoch": 1.3054111631870473, + "grad_norm": 14.3125, + "learning_rate": 6.782490086059957e-07, + "loss": 0.7363, + "step": 15319 + }, + { + "epoch": 1.3054963783553473, + "grad_norm": 19.75, + "learning_rate": 6.780995863685518e-07, + "loss": 0.633, + "step": 15320 + }, + { + "epoch": 1.3055815935236472, + "grad_norm": 13.375, + "learning_rate": 6.779501744659238e-07, + "loss": 0.3973, + "step": 15321 + }, + { + "epoch": 1.3056668086919472, + "grad_norm": 13.125, + "learning_rate": 6.778007729008115e-07, + "loss": 0.4486, + "step": 15322 + }, + { + "epoch": 1.3057520238602471, + "grad_norm": 14.5, + "learning_rate": 6.776513816759142e-07, + "loss": 0.5381, + "step": 15323 + }, + { + "epoch": 1.305837239028547, + "grad_norm": 18.875, + "learning_rate": 6.775020007939317e-07, + "loss": 0.9599, + "step": 15324 + }, + { + "epoch": 1.305922454196847, + "grad_norm": 15.625, + "learning_rate": 6.77352630257564e-07, + "loss": 0.6201, + "step": 15325 + }, + { + "epoch": 1.306007669365147, + "grad_norm": 16.125, + "learning_rate": 6.772032700695105e-07, + "loss": 0.4252, + "step": 15326 + }, + { + "epoch": 1.306092884533447, + "grad_norm": 17.375, + "learning_rate": 6.770539202324694e-07, + "loss": 0.6245, + "step": 15327 + }, + { + "epoch": 1.306178099701747, + "grad_norm": 15.25, + "learning_rate": 6.769045807491402e-07, + "loss": 0.4163, + "step": 15328 + }, + { + "epoch": 1.3062633148700469, + "grad_norm": 20.25, + "learning_rate": 6.767552516222214e-07, + "loss": 0.7515, + "step": 15329 + }, + { + "epoch": 1.3063485300383468, + "grad_norm": 24.875, + "learning_rate": 6.766059328544128e-07, + "loss": 0.557, + "step": 15330 + }, + { + "epoch": 1.3064337452066468, + "grad_norm": 14.875, + "learning_rate": 6.764566244484106e-07, + "loss": 0.5739, + "step": 15331 + }, + { + "epoch": 1.3065189603749467, + "grad_norm": 15.8125, + "learning_rate": 6.763073264069137e-07, + "loss": 0.6625, + "step": 15332 + }, + { + "epoch": 1.3066041755432467, + "grad_norm": 17.625, + "learning_rate": 6.761580387326205e-07, + "loss": 0.9779, + "step": 15333 + }, + { + "epoch": 1.3066893907115467, + "grad_norm": 18.375, + "learning_rate": 6.76008761428229e-07, + "loss": 0.689, + "step": 15334 + }, + { + "epoch": 1.3067746058798466, + "grad_norm": 17.0, + "learning_rate": 6.758594944964358e-07, + "loss": 0.7297, + "step": 15335 + }, + { + "epoch": 1.3068598210481466, + "grad_norm": 15.5, + "learning_rate": 6.757102379399389e-07, + "loss": 0.3316, + "step": 15336 + }, + { + "epoch": 1.3069450362164465, + "grad_norm": 16.625, + "learning_rate": 6.755609917614356e-07, + "loss": 0.4785, + "step": 15337 + }, + { + "epoch": 1.3070302513847465, + "grad_norm": 10.8125, + "learning_rate": 6.754117559636233e-07, + "loss": 0.7587, + "step": 15338 + }, + { + "epoch": 1.3071154665530464, + "grad_norm": 11.3125, + "learning_rate": 6.752625305491983e-07, + "loss": 0.4599, + "step": 15339 + }, + { + "epoch": 1.3072006817213464, + "grad_norm": 13.875, + "learning_rate": 6.75113315520857e-07, + "loss": 0.6435, + "step": 15340 + }, + { + "epoch": 1.3072858968896464, + "grad_norm": 13.3125, + "learning_rate": 6.74964110881296e-07, + "loss": 0.3003, + "step": 15341 + }, + { + "epoch": 1.3073711120579463, + "grad_norm": 12.875, + "learning_rate": 6.748149166332119e-07, + "loss": 0.4318, + "step": 15342 + }, + { + "epoch": 1.3074563272262463, + "grad_norm": 15.0, + "learning_rate": 6.74665732779301e-07, + "loss": 0.5902, + "step": 15343 + }, + { + "epoch": 1.3075415423945462, + "grad_norm": 18.375, + "learning_rate": 6.745165593222583e-07, + "loss": 0.894, + "step": 15344 + }, + { + "epoch": 1.3076267575628462, + "grad_norm": 15.625, + "learning_rate": 6.743673962647802e-07, + "loss": 0.5885, + "step": 15345 + }, + { + "epoch": 1.3077119727311461, + "grad_norm": 14.75, + "learning_rate": 6.742182436095626e-07, + "loss": 0.4738, + "step": 15346 + }, + { + "epoch": 1.307797187899446, + "grad_norm": 20.5, + "learning_rate": 6.740691013593002e-07, + "loss": 0.5926, + "step": 15347 + }, + { + "epoch": 1.307882403067746, + "grad_norm": 13.75, + "learning_rate": 6.739199695166879e-07, + "loss": 0.3915, + "step": 15348 + }, + { + "epoch": 1.307967618236046, + "grad_norm": 11.5, + "learning_rate": 6.73770848084421e-07, + "loss": 0.3932, + "step": 15349 + }, + { + "epoch": 1.308052833404346, + "grad_norm": 19.875, + "learning_rate": 6.736217370651944e-07, + "loss": 0.9585, + "step": 15350 + }, + { + "epoch": 1.308138048572646, + "grad_norm": 13.3125, + "learning_rate": 6.734726364617031e-07, + "loss": 0.4649, + "step": 15351 + }, + { + "epoch": 1.3082232637409459, + "grad_norm": 11.75, + "learning_rate": 6.733235462766405e-07, + "loss": 0.3234, + "step": 15352 + }, + { + "epoch": 1.3083084789092458, + "grad_norm": 11.5625, + "learning_rate": 6.731744665127012e-07, + "loss": 0.3865, + "step": 15353 + }, + { + "epoch": 1.3083936940775458, + "grad_norm": 18.5, + "learning_rate": 6.730253971725798e-07, + "loss": 0.4972, + "step": 15354 + }, + { + "epoch": 1.3084789092458458, + "grad_norm": 17.125, + "learning_rate": 6.728763382589693e-07, + "loss": 1.1915, + "step": 15355 + }, + { + "epoch": 1.3085641244141457, + "grad_norm": 18.625, + "learning_rate": 6.727272897745641e-07, + "loss": 0.8562, + "step": 15356 + }, + { + "epoch": 1.3086493395824457, + "grad_norm": 15.1875, + "learning_rate": 6.72578251722057e-07, + "loss": 0.4997, + "step": 15357 + }, + { + "epoch": 1.3087345547507456, + "grad_norm": 41.75, + "learning_rate": 6.724292241041413e-07, + "loss": 1.1714, + "step": 15358 + }, + { + "epoch": 1.3088197699190456, + "grad_norm": 18.25, + "learning_rate": 6.722802069235104e-07, + "loss": 0.3617, + "step": 15359 + }, + { + "epoch": 1.3089049850873455, + "grad_norm": 43.75, + "learning_rate": 6.721312001828576e-07, + "loss": 0.9277, + "step": 15360 + }, + { + "epoch": 1.3089902002556455, + "grad_norm": 9.5625, + "learning_rate": 6.719822038848745e-07, + "loss": 0.1717, + "step": 15361 + }, + { + "epoch": 1.3090754154239455, + "grad_norm": 17.375, + "learning_rate": 6.718332180322548e-07, + "loss": 0.6324, + "step": 15362 + }, + { + "epoch": 1.3091606305922454, + "grad_norm": 14.0, + "learning_rate": 6.716842426276898e-07, + "loss": 0.6823, + "step": 15363 + }, + { + "epoch": 1.3092458457605454, + "grad_norm": 15.4375, + "learning_rate": 6.715352776738724e-07, + "loss": 0.7052, + "step": 15364 + }, + { + "epoch": 1.3093310609288453, + "grad_norm": 16.375, + "learning_rate": 6.713863231734939e-07, + "loss": 0.669, + "step": 15365 + }, + { + "epoch": 1.3094162760971453, + "grad_norm": 14.5625, + "learning_rate": 6.712373791292463e-07, + "loss": 0.7044, + "step": 15366 + }, + { + "epoch": 1.3095014912654452, + "grad_norm": 14.5, + "learning_rate": 6.710884455438213e-07, + "loss": 0.5283, + "step": 15367 + }, + { + "epoch": 1.3095867064337452, + "grad_norm": 12.25, + "learning_rate": 6.709395224199107e-07, + "loss": 0.591, + "step": 15368 + }, + { + "epoch": 1.3096719216020452, + "grad_norm": 16.625, + "learning_rate": 6.707906097602048e-07, + "loss": 0.8888, + "step": 15369 + }, + { + "epoch": 1.309757136770345, + "grad_norm": 14.8125, + "learning_rate": 6.706417075673948e-07, + "loss": 0.5586, + "step": 15370 + }, + { + "epoch": 1.309842351938645, + "grad_norm": 14.25, + "learning_rate": 6.704928158441725e-07, + "loss": 0.5235, + "step": 15371 + }, + { + "epoch": 1.309927567106945, + "grad_norm": 14.375, + "learning_rate": 6.703439345932272e-07, + "loss": 0.5794, + "step": 15372 + }, + { + "epoch": 1.310012782275245, + "grad_norm": 16.5, + "learning_rate": 6.701950638172503e-07, + "loss": 0.412, + "step": 15373 + }, + { + "epoch": 1.310097997443545, + "grad_norm": 10.25, + "learning_rate": 6.700462035189312e-07, + "loss": 0.3629, + "step": 15374 + }, + { + "epoch": 1.310183212611845, + "grad_norm": 16.0, + "learning_rate": 6.698973537009604e-07, + "loss": 0.6084, + "step": 15375 + }, + { + "epoch": 1.3102684277801449, + "grad_norm": 20.0, + "learning_rate": 6.697485143660276e-07, + "loss": 0.6572, + "step": 15376 + }, + { + "epoch": 1.3103536429484448, + "grad_norm": 16.875, + "learning_rate": 6.695996855168232e-07, + "loss": 0.85, + "step": 15377 + }, + { + "epoch": 1.3104388581167448, + "grad_norm": 36.5, + "learning_rate": 6.694508671560357e-07, + "loss": 0.8257, + "step": 15378 + }, + { + "epoch": 1.3105240732850447, + "grad_norm": 13.1875, + "learning_rate": 6.693020592863553e-07, + "loss": 0.4859, + "step": 15379 + }, + { + "epoch": 1.3106092884533447, + "grad_norm": 14.3125, + "learning_rate": 6.691532619104702e-07, + "loss": 0.5618, + "step": 15380 + }, + { + "epoch": 1.3106945036216446, + "grad_norm": 17.0, + "learning_rate": 6.690044750310701e-07, + "loss": 0.759, + "step": 15381 + }, + { + "epoch": 1.3107797187899446, + "grad_norm": 15.0, + "learning_rate": 6.688556986508429e-07, + "loss": 0.6163, + "step": 15382 + }, + { + "epoch": 1.3108649339582445, + "grad_norm": 23.375, + "learning_rate": 6.687069327724776e-07, + "loss": 1.1852, + "step": 15383 + }, + { + "epoch": 1.3109501491265445, + "grad_norm": 19.0, + "learning_rate": 6.685581773986627e-07, + "loss": 0.49, + "step": 15384 + }, + { + "epoch": 1.3110353642948445, + "grad_norm": 13.875, + "learning_rate": 6.684094325320866e-07, + "loss": 0.478, + "step": 15385 + }, + { + "epoch": 1.3111205794631444, + "grad_norm": 16.75, + "learning_rate": 6.682606981754364e-07, + "loss": 0.4598, + "step": 15386 + }, + { + "epoch": 1.3112057946314444, + "grad_norm": 19.375, + "learning_rate": 6.681119743314007e-07, + "loss": 0.9594, + "step": 15387 + }, + { + "epoch": 1.3112910097997443, + "grad_norm": 17.0, + "learning_rate": 6.679632610026665e-07, + "loss": 0.6137, + "step": 15388 + }, + { + "epoch": 1.3113762249680443, + "grad_norm": 13.6875, + "learning_rate": 6.678145581919215e-07, + "loss": 0.4175, + "step": 15389 + }, + { + "epoch": 1.3114614401363442, + "grad_norm": 13.125, + "learning_rate": 6.676658659018532e-07, + "loss": 0.3571, + "step": 15390 + }, + { + "epoch": 1.3115466553046442, + "grad_norm": 12.875, + "learning_rate": 6.675171841351479e-07, + "loss": 0.299, + "step": 15391 + }, + { + "epoch": 1.3116318704729442, + "grad_norm": 15.0625, + "learning_rate": 6.673685128944926e-07, + "loss": 0.5589, + "step": 15392 + }, + { + "epoch": 1.3117170856412441, + "grad_norm": 19.0, + "learning_rate": 6.672198521825748e-07, + "loss": 0.7522, + "step": 15393 + }, + { + "epoch": 1.311802300809544, + "grad_norm": 14.5625, + "learning_rate": 6.670712020020801e-07, + "loss": 0.5869, + "step": 15394 + }, + { + "epoch": 1.311887515977844, + "grad_norm": 13.4375, + "learning_rate": 6.669225623556949e-07, + "loss": 0.2488, + "step": 15395 + }, + { + "epoch": 1.311972731146144, + "grad_norm": 26.0, + "learning_rate": 6.667739332461051e-07, + "loss": 0.8275, + "step": 15396 + }, + { + "epoch": 1.312057946314444, + "grad_norm": 13.875, + "learning_rate": 6.666253146759968e-07, + "loss": 0.402, + "step": 15397 + }, + { + "epoch": 1.312143161482744, + "grad_norm": 14.1875, + "learning_rate": 6.664767066480562e-07, + "loss": 0.6103, + "step": 15398 + }, + { + "epoch": 1.3122283766510439, + "grad_norm": 14.5, + "learning_rate": 6.663281091649679e-07, + "loss": 0.5477, + "step": 15399 + }, + { + "epoch": 1.3123135918193438, + "grad_norm": 15.75, + "learning_rate": 6.661795222294175e-07, + "loss": 0.7335, + "step": 15400 + }, + { + "epoch": 1.3123988069876438, + "grad_norm": 16.375, + "learning_rate": 6.660309458440901e-07, + "loss": 0.625, + "step": 15401 + }, + { + "epoch": 1.3124840221559437, + "grad_norm": 11.8125, + "learning_rate": 6.658823800116714e-07, + "loss": 0.2921, + "step": 15402 + }, + { + "epoch": 1.3125692373242437, + "grad_norm": 18.75, + "learning_rate": 6.657338247348454e-07, + "loss": 0.6292, + "step": 15403 + }, + { + "epoch": 1.3126544524925436, + "grad_norm": 15.9375, + "learning_rate": 6.655852800162963e-07, + "loss": 0.6587, + "step": 15404 + }, + { + "epoch": 1.3127396676608436, + "grad_norm": 13.5, + "learning_rate": 6.654367458587087e-07, + "loss": 0.3783, + "step": 15405 + }, + { + "epoch": 1.3128248828291436, + "grad_norm": 13.6875, + "learning_rate": 6.652882222647669e-07, + "loss": 0.4794, + "step": 15406 + }, + { + "epoch": 1.3129100979974435, + "grad_norm": 16.5, + "learning_rate": 6.651397092371556e-07, + "loss": 0.5813, + "step": 15407 + }, + { + "epoch": 1.3129953131657435, + "grad_norm": 12.25, + "learning_rate": 6.649912067785572e-07, + "loss": 0.4377, + "step": 15408 + }, + { + "epoch": 1.3130805283340434, + "grad_norm": 14.875, + "learning_rate": 6.648427148916561e-07, + "loss": 0.555, + "step": 15409 + }, + { + "epoch": 1.3131657435023434, + "grad_norm": 14.625, + "learning_rate": 6.646942335791359e-07, + "loss": 0.5567, + "step": 15410 + }, + { + "epoch": 1.3132509586706433, + "grad_norm": 33.75, + "learning_rate": 6.645457628436797e-07, + "loss": 0.7929, + "step": 15411 + }, + { + "epoch": 1.3133361738389433, + "grad_norm": 12.25, + "learning_rate": 6.643973026879697e-07, + "loss": 0.9013, + "step": 15412 + }, + { + "epoch": 1.3134213890072433, + "grad_norm": 17.0, + "learning_rate": 6.642488531146891e-07, + "loss": 0.8858, + "step": 15413 + }, + { + "epoch": 1.3135066041755432, + "grad_norm": 14.0, + "learning_rate": 6.641004141265208e-07, + "loss": 0.7271, + "step": 15414 + }, + { + "epoch": 1.3135918193438432, + "grad_norm": 23.25, + "learning_rate": 6.63951985726148e-07, + "loss": 1.1887, + "step": 15415 + }, + { + "epoch": 1.3136770345121431, + "grad_norm": 17.75, + "learning_rate": 6.638035679162513e-07, + "loss": 0.6856, + "step": 15416 + }, + { + "epoch": 1.313762249680443, + "grad_norm": 15.75, + "learning_rate": 6.636551606995137e-07, + "loss": 0.8047, + "step": 15417 + }, + { + "epoch": 1.313847464848743, + "grad_norm": 17.375, + "learning_rate": 6.635067640786171e-07, + "loss": 0.7164, + "step": 15418 + }, + { + "epoch": 1.313932680017043, + "grad_norm": 24.0, + "learning_rate": 6.633583780562433e-07, + "loss": 0.7173, + "step": 15419 + }, + { + "epoch": 1.314017895185343, + "grad_norm": 18.0, + "learning_rate": 6.632100026350738e-07, + "loss": 0.811, + "step": 15420 + }, + { + "epoch": 1.314103110353643, + "grad_norm": 13.5, + "learning_rate": 6.630616378177889e-07, + "loss": 0.3032, + "step": 15421 + }, + { + "epoch": 1.3141883255219429, + "grad_norm": 19.0, + "learning_rate": 6.629132836070704e-07, + "loss": 0.9835, + "step": 15422 + }, + { + "epoch": 1.3142735406902428, + "grad_norm": 13.5, + "learning_rate": 6.627649400055994e-07, + "loss": 0.5907, + "step": 15423 + }, + { + "epoch": 1.3143587558585428, + "grad_norm": 16.75, + "learning_rate": 6.626166070160568e-07, + "loss": 0.5778, + "step": 15424 + }, + { + "epoch": 1.3144439710268427, + "grad_norm": 16.25, + "learning_rate": 6.624682846411224e-07, + "loss": 0.8411, + "step": 15425 + }, + { + "epoch": 1.3145291861951427, + "grad_norm": 10.0625, + "learning_rate": 6.623199728834767e-07, + "loss": 0.3694, + "step": 15426 + }, + { + "epoch": 1.3146144013634427, + "grad_norm": 14.5, + "learning_rate": 6.621716717458007e-07, + "loss": 0.5971, + "step": 15427 + }, + { + "epoch": 1.3146996165317426, + "grad_norm": 12.6875, + "learning_rate": 6.620233812307736e-07, + "loss": 0.3018, + "step": 15428 + }, + { + "epoch": 1.3147848317000426, + "grad_norm": 14.625, + "learning_rate": 6.618751013410748e-07, + "loss": 0.6595, + "step": 15429 + }, + { + "epoch": 1.3148700468683425, + "grad_norm": 22.25, + "learning_rate": 6.617268320793844e-07, + "loss": 0.5228, + "step": 15430 + }, + { + "epoch": 1.3149552620366425, + "grad_norm": 21.25, + "learning_rate": 6.615785734483818e-07, + "loss": 0.6278, + "step": 15431 + }, + { + "epoch": 1.3150404772049424, + "grad_norm": 20.5, + "learning_rate": 6.614303254507464e-07, + "loss": 0.7388, + "step": 15432 + }, + { + "epoch": 1.3151256923732424, + "grad_norm": 17.625, + "learning_rate": 6.612820880891563e-07, + "loss": 0.5935, + "step": 15433 + }, + { + "epoch": 1.3152109075415424, + "grad_norm": 16.0, + "learning_rate": 6.61133861366291e-07, + "loss": 0.505, + "step": 15434 + }, + { + "epoch": 1.3152961227098423, + "grad_norm": 15.6875, + "learning_rate": 6.60985645284829e-07, + "loss": 0.5201, + "step": 15435 + }, + { + "epoch": 1.3153813378781423, + "grad_norm": 16.5, + "learning_rate": 6.60837439847449e-07, + "loss": 0.4904, + "step": 15436 + }, + { + "epoch": 1.3154665530464422, + "grad_norm": 18.625, + "learning_rate": 6.606892450568292e-07, + "loss": 0.6562, + "step": 15437 + }, + { + "epoch": 1.3155517682147422, + "grad_norm": 24.0, + "learning_rate": 6.605410609156465e-07, + "loss": 0.7016, + "step": 15438 + }, + { + "epoch": 1.3156369833830421, + "grad_norm": 16.5, + "learning_rate": 6.603928874265799e-07, + "loss": 0.4223, + "step": 15439 + }, + { + "epoch": 1.315722198551342, + "grad_norm": 25.5, + "learning_rate": 6.60244724592307e-07, + "loss": 0.5433, + "step": 15440 + }, + { + "epoch": 1.315807413719642, + "grad_norm": 12.8125, + "learning_rate": 6.600965724155045e-07, + "loss": 0.3768, + "step": 15441 + }, + { + "epoch": 1.315892628887942, + "grad_norm": 14.5, + "learning_rate": 6.599484308988502e-07, + "loss": 0.5223, + "step": 15442 + }, + { + "epoch": 1.315977844056242, + "grad_norm": 18.0, + "learning_rate": 6.598003000450212e-07, + "loss": 0.7828, + "step": 15443 + }, + { + "epoch": 1.316063059224542, + "grad_norm": 15.5625, + "learning_rate": 6.596521798566946e-07, + "loss": 0.5904, + "step": 15444 + }, + { + "epoch": 1.3161482743928419, + "grad_norm": 15.1875, + "learning_rate": 6.595040703365469e-07, + "loss": 0.6069, + "step": 15445 + }, + { + "epoch": 1.3162334895611418, + "grad_norm": 24.375, + "learning_rate": 6.593559714872538e-07, + "loss": 0.8387, + "step": 15446 + }, + { + "epoch": 1.3163187047294418, + "grad_norm": 21.25, + "learning_rate": 6.592078833114922e-07, + "loss": 0.5262, + "step": 15447 + }, + { + "epoch": 1.3164039198977417, + "grad_norm": 11.1875, + "learning_rate": 6.590598058119382e-07, + "loss": 0.3016, + "step": 15448 + }, + { + "epoch": 1.3164891350660417, + "grad_norm": 14.5, + "learning_rate": 6.589117389912684e-07, + "loss": 0.5915, + "step": 15449 + }, + { + "epoch": 1.3165743502343417, + "grad_norm": 19.125, + "learning_rate": 6.587636828521574e-07, + "loss": 0.5429, + "step": 15450 + }, + { + "epoch": 1.3166595654026416, + "grad_norm": 14.0625, + "learning_rate": 6.58615637397281e-07, + "loss": 0.3568, + "step": 15451 + }, + { + "epoch": 1.3167447805709416, + "grad_norm": 22.25, + "learning_rate": 6.584676026293154e-07, + "loss": 0.505, + "step": 15452 + }, + { + "epoch": 1.3168299957392415, + "grad_norm": 15.9375, + "learning_rate": 6.583195785509345e-07, + "loss": 0.6568, + "step": 15453 + }, + { + "epoch": 1.3169152109075415, + "grad_norm": 16.0, + "learning_rate": 6.581715651648142e-07, + "loss": 0.7261, + "step": 15454 + }, + { + "epoch": 1.3170004260758414, + "grad_norm": 11.0625, + "learning_rate": 6.580235624736283e-07, + "loss": 0.4172, + "step": 15455 + }, + { + "epoch": 1.3170856412441414, + "grad_norm": 16.125, + "learning_rate": 6.578755704800521e-07, + "loss": 0.6075, + "step": 15456 + }, + { + "epoch": 1.3171708564124414, + "grad_norm": 15.9375, + "learning_rate": 6.577275891867602e-07, + "loss": 0.5733, + "step": 15457 + }, + { + "epoch": 1.3172560715807413, + "grad_norm": 15.875, + "learning_rate": 6.575796185964259e-07, + "loss": 0.6118, + "step": 15458 + }, + { + "epoch": 1.3173412867490413, + "grad_norm": 15.0, + "learning_rate": 6.574316587117236e-07, + "loss": 0.652, + "step": 15459 + }, + { + "epoch": 1.3174265019173412, + "grad_norm": 16.5, + "learning_rate": 6.572837095353274e-07, + "loss": 0.7427, + "step": 15460 + }, + { + "epoch": 1.3175117170856412, + "grad_norm": 15.0625, + "learning_rate": 6.571357710699103e-07, + "loss": 0.6524, + "step": 15461 + }, + { + "epoch": 1.3175969322539411, + "grad_norm": 17.5, + "learning_rate": 6.569878433181465e-07, + "loss": 0.6554, + "step": 15462 + }, + { + "epoch": 1.317682147422241, + "grad_norm": 16.375, + "learning_rate": 6.568399262827079e-07, + "loss": 0.7124, + "step": 15463 + }, + { + "epoch": 1.317767362590541, + "grad_norm": 13.6875, + "learning_rate": 6.566920199662686e-07, + "loss": 0.3508, + "step": 15464 + }, + { + "epoch": 1.317852577758841, + "grad_norm": 24.75, + "learning_rate": 6.565441243715011e-07, + "loss": 0.9498, + "step": 15465 + }, + { + "epoch": 1.317937792927141, + "grad_norm": 29.875, + "learning_rate": 6.563962395010784e-07, + "loss": 1.5344, + "step": 15466 + }, + { + "epoch": 1.318023008095441, + "grad_norm": 15.5625, + "learning_rate": 6.56248365357672e-07, + "loss": 0.6284, + "step": 15467 + }, + { + "epoch": 1.3181082232637409, + "grad_norm": 15.9375, + "learning_rate": 6.561005019439555e-07, + "loss": 0.5028, + "step": 15468 + }, + { + "epoch": 1.3181934384320408, + "grad_norm": 31.75, + "learning_rate": 6.559526492625995e-07, + "loss": 0.9941, + "step": 15469 + }, + { + "epoch": 1.3182786536003408, + "grad_norm": 12.5, + "learning_rate": 6.558048073162769e-07, + "loss": 0.4775, + "step": 15470 + }, + { + "epoch": 1.3183638687686408, + "grad_norm": 12.5625, + "learning_rate": 6.556569761076585e-07, + "loss": 0.3441, + "step": 15471 + }, + { + "epoch": 1.3184490839369407, + "grad_norm": 15.625, + "learning_rate": 6.555091556394163e-07, + "loss": 0.7119, + "step": 15472 + }, + { + "epoch": 1.3185342991052407, + "grad_norm": 16.25, + "learning_rate": 6.553613459142214e-07, + "loss": 0.6743, + "step": 15473 + }, + { + "epoch": 1.3186195142735406, + "grad_norm": 17.125, + "learning_rate": 6.552135469347453e-07, + "loss": 0.8856, + "step": 15474 + }, + { + "epoch": 1.3187047294418406, + "grad_norm": 20.375, + "learning_rate": 6.550657587036581e-07, + "loss": 0.7738, + "step": 15475 + }, + { + "epoch": 1.3187899446101405, + "grad_norm": 21.5, + "learning_rate": 6.549179812236314e-07, + "loss": 0.7166, + "step": 15476 + }, + { + "epoch": 1.3188751597784405, + "grad_norm": 15.5625, + "learning_rate": 6.547702144973348e-07, + "loss": 0.6175, + "step": 15477 + }, + { + "epoch": 1.3189603749467405, + "grad_norm": 13.4375, + "learning_rate": 6.546224585274388e-07, + "loss": 0.3014, + "step": 15478 + }, + { + "epoch": 1.3190455901150404, + "grad_norm": 13.125, + "learning_rate": 6.544747133166144e-07, + "loss": 0.5052, + "step": 15479 + }, + { + "epoch": 1.3191308052833404, + "grad_norm": 10.8125, + "learning_rate": 6.543269788675299e-07, + "loss": 0.5391, + "step": 15480 + }, + { + "epoch": 1.3192160204516403, + "grad_norm": 14.4375, + "learning_rate": 6.541792551828561e-07, + "loss": 0.516, + "step": 15481 + }, + { + "epoch": 1.3193012356199403, + "grad_norm": 15.0625, + "learning_rate": 6.540315422652624e-07, + "loss": 0.6095, + "step": 15482 + }, + { + "epoch": 1.3193864507882402, + "grad_norm": 16.25, + "learning_rate": 6.538838401174183e-07, + "loss": 0.4185, + "step": 15483 + }, + { + "epoch": 1.3194716659565402, + "grad_norm": 14.125, + "learning_rate": 6.537361487419928e-07, + "loss": 0.5646, + "step": 15484 + }, + { + "epoch": 1.3195568811248402, + "grad_norm": 17.5, + "learning_rate": 6.535884681416542e-07, + "loss": 0.6418, + "step": 15485 + }, + { + "epoch": 1.31964209629314, + "grad_norm": 17.125, + "learning_rate": 6.534407983190717e-07, + "loss": 0.8113, + "step": 15486 + }, + { + "epoch": 1.31972731146144, + "grad_norm": 21.75, + "learning_rate": 6.532931392769143e-07, + "loss": 0.8686, + "step": 15487 + }, + { + "epoch": 1.31981252662974, + "grad_norm": 15.3125, + "learning_rate": 6.531454910178494e-07, + "loss": 0.6186, + "step": 15488 + }, + { + "epoch": 1.31989774179804, + "grad_norm": 13.6875, + "learning_rate": 6.529978535445458e-07, + "loss": 0.4389, + "step": 15489 + }, + { + "epoch": 1.31998295696634, + "grad_norm": 16.5, + "learning_rate": 6.528502268596712e-07, + "loss": 0.6286, + "step": 15490 + }, + { + "epoch": 1.32006817213464, + "grad_norm": 16.0, + "learning_rate": 6.527026109658941e-07, + "loss": 0.4586, + "step": 15491 + }, + { + "epoch": 1.3201533873029399, + "grad_norm": 10.75, + "learning_rate": 6.525550058658813e-07, + "loss": 0.35, + "step": 15492 + }, + { + "epoch": 1.3202386024712398, + "grad_norm": 15.4375, + "learning_rate": 6.524074115622998e-07, + "loss": 0.5575, + "step": 15493 + }, + { + "epoch": 1.3203238176395398, + "grad_norm": 11.6875, + "learning_rate": 6.522598280578174e-07, + "loss": 0.3634, + "step": 15494 + }, + { + "epoch": 1.3204090328078397, + "grad_norm": 13.8125, + "learning_rate": 6.521122553551009e-07, + "loss": 0.3248, + "step": 15495 + }, + { + "epoch": 1.3204942479761397, + "grad_norm": 20.5, + "learning_rate": 6.519646934568177e-07, + "loss": 0.8944, + "step": 15496 + }, + { + "epoch": 1.3205794631444396, + "grad_norm": 17.0, + "learning_rate": 6.518171423656336e-07, + "loss": 0.6229, + "step": 15497 + }, + { + "epoch": 1.3206646783127396, + "grad_norm": 13.625, + "learning_rate": 6.516696020842151e-07, + "loss": 0.5192, + "step": 15498 + }, + { + "epoch": 1.3207498934810395, + "grad_norm": 20.5, + "learning_rate": 6.515220726152284e-07, + "loss": 1.0354, + "step": 15499 + }, + { + "epoch": 1.3208351086493395, + "grad_norm": 25.625, + "learning_rate": 6.513745539613403e-07, + "loss": 0.7525, + "step": 15500 + }, + { + "epoch": 1.3209203238176395, + "grad_norm": 14.8125, + "learning_rate": 6.512270461252158e-07, + "loss": 0.5107, + "step": 15501 + }, + { + "epoch": 1.3210055389859394, + "grad_norm": 17.625, + "learning_rate": 6.510795491095203e-07, + "loss": 0.8479, + "step": 15502 + }, + { + "epoch": 1.3210907541542394, + "grad_norm": 12.375, + "learning_rate": 6.509320629169197e-07, + "loss": 0.4749, + "step": 15503 + }, + { + "epoch": 1.3211759693225393, + "grad_norm": 13.6875, + "learning_rate": 6.507845875500796e-07, + "loss": 0.3673, + "step": 15504 + }, + { + "epoch": 1.3212611844908393, + "grad_norm": 14.8125, + "learning_rate": 6.50637123011664e-07, + "loss": 0.4624, + "step": 15505 + }, + { + "epoch": 1.3213463996591392, + "grad_norm": 15.8125, + "learning_rate": 6.504896693043384e-07, + "loss": 0.6702, + "step": 15506 + }, + { + "epoch": 1.3214316148274392, + "grad_norm": 19.25, + "learning_rate": 6.503422264307673e-07, + "loss": 0.8024, + "step": 15507 + }, + { + "epoch": 1.3215168299957392, + "grad_norm": 15.375, + "learning_rate": 6.501947943936156e-07, + "loss": 0.6587, + "step": 15508 + }, + { + "epoch": 1.3216020451640391, + "grad_norm": 12.4375, + "learning_rate": 6.50047373195547e-07, + "loss": 0.4783, + "step": 15509 + }, + { + "epoch": 1.321687260332339, + "grad_norm": 14.5, + "learning_rate": 6.498999628392254e-07, + "loss": 0.7538, + "step": 15510 + }, + { + "epoch": 1.321772475500639, + "grad_norm": 12.5625, + "learning_rate": 6.497525633273148e-07, + "loss": 0.4592, + "step": 15511 + }, + { + "epoch": 1.321857690668939, + "grad_norm": 9.5625, + "learning_rate": 6.49605174662479e-07, + "loss": 0.17, + "step": 15512 + }, + { + "epoch": 1.321942905837239, + "grad_norm": 18.75, + "learning_rate": 6.49457796847382e-07, + "loss": 0.5805, + "step": 15513 + }, + { + "epoch": 1.322028121005539, + "grad_norm": 13.6875, + "learning_rate": 6.493104298846859e-07, + "loss": 0.688, + "step": 15514 + }, + { + "epoch": 1.3221133361738389, + "grad_norm": 13.0625, + "learning_rate": 6.491630737770543e-07, + "loss": 0.4344, + "step": 15515 + }, + { + "epoch": 1.3221985513421388, + "grad_norm": 14.25, + "learning_rate": 6.490157285271504e-07, + "loss": 0.4381, + "step": 15516 + }, + { + "epoch": 1.3222837665104388, + "grad_norm": 19.125, + "learning_rate": 6.488683941376375e-07, + "loss": 0.8797, + "step": 15517 + }, + { + "epoch": 1.3223689816787387, + "grad_norm": 12.1875, + "learning_rate": 6.487210706111761e-07, + "loss": 0.377, + "step": 15518 + }, + { + "epoch": 1.3224541968470387, + "grad_norm": 15.0625, + "learning_rate": 6.485737579504297e-07, + "loss": 0.6436, + "step": 15519 + }, + { + "epoch": 1.3225394120153386, + "grad_norm": 17.375, + "learning_rate": 6.484264561580603e-07, + "loss": 0.7755, + "step": 15520 + }, + { + "epoch": 1.3226246271836386, + "grad_norm": 14.875, + "learning_rate": 6.482791652367302e-07, + "loss": 0.7225, + "step": 15521 + }, + { + "epoch": 1.3227098423519386, + "grad_norm": 17.75, + "learning_rate": 6.481318851891002e-07, + "loss": 0.6543, + "step": 15522 + }, + { + "epoch": 1.3227950575202385, + "grad_norm": 14.3125, + "learning_rate": 6.479846160178321e-07, + "loss": 0.6054, + "step": 15523 + }, + { + "epoch": 1.3228802726885385, + "grad_norm": 12.5, + "learning_rate": 6.478373577255875e-07, + "loss": 0.374, + "step": 15524 + }, + { + "epoch": 1.3229654878568384, + "grad_norm": 14.25, + "learning_rate": 6.476901103150279e-07, + "loss": 0.6184, + "step": 15525 + }, + { + "epoch": 1.3230507030251384, + "grad_norm": 15.4375, + "learning_rate": 6.475428737888137e-07, + "loss": 0.7745, + "step": 15526 + }, + { + "epoch": 1.3231359181934383, + "grad_norm": 25.5, + "learning_rate": 6.473956481496049e-07, + "loss": 0.6662, + "step": 15527 + }, + { + "epoch": 1.3232211333617383, + "grad_norm": 12.875, + "learning_rate": 6.472484334000629e-07, + "loss": 0.3165, + "step": 15528 + }, + { + "epoch": 1.3233063485300383, + "grad_norm": 16.125, + "learning_rate": 6.471012295428478e-07, + "loss": 0.5944, + "step": 15529 + }, + { + "epoch": 1.3233915636983382, + "grad_norm": 14.375, + "learning_rate": 6.469540365806202e-07, + "loss": 0.3362, + "step": 15530 + }, + { + "epoch": 1.3234767788666382, + "grad_norm": 11.9375, + "learning_rate": 6.468068545160391e-07, + "loss": 0.4027, + "step": 15531 + }, + { + "epoch": 1.3235619940349381, + "grad_norm": 14.4375, + "learning_rate": 6.466596833517647e-07, + "loss": 0.6648, + "step": 15532 + }, + { + "epoch": 1.323647209203238, + "grad_norm": 18.5, + "learning_rate": 6.465125230904568e-07, + "loss": 0.4131, + "step": 15533 + }, + { + "epoch": 1.323732424371538, + "grad_norm": 15.75, + "learning_rate": 6.463653737347748e-07, + "loss": 0.6232, + "step": 15534 + }, + { + "epoch": 1.323817639539838, + "grad_norm": 16.5, + "learning_rate": 6.462182352873768e-07, + "loss": 0.2563, + "step": 15535 + }, + { + "epoch": 1.323902854708138, + "grad_norm": 15.8125, + "learning_rate": 6.460711077509224e-07, + "loss": 1.2542, + "step": 15536 + }, + { + "epoch": 1.323988069876438, + "grad_norm": 16.5, + "learning_rate": 6.459239911280706e-07, + "loss": 0.424, + "step": 15537 + }, + { + "epoch": 1.3240732850447379, + "grad_norm": 10.8125, + "learning_rate": 6.4577688542148e-07, + "loss": 0.3378, + "step": 15538 + }, + { + "epoch": 1.3241585002130378, + "grad_norm": 11.9375, + "learning_rate": 6.456297906338083e-07, + "loss": 0.387, + "step": 15539 + }, + { + "epoch": 1.3242437153813378, + "grad_norm": 12.1875, + "learning_rate": 6.45482706767714e-07, + "loss": 0.4798, + "step": 15540 + }, + { + "epoch": 1.3243289305496377, + "grad_norm": 14.9375, + "learning_rate": 6.453356338258555e-07, + "loss": 0.4521, + "step": 15541 + }, + { + "epoch": 1.3244141457179377, + "grad_norm": 11.9375, + "learning_rate": 6.451885718108897e-07, + "loss": 0.3642, + "step": 15542 + }, + { + "epoch": 1.3244993608862377, + "grad_norm": 19.0, + "learning_rate": 6.450415207254751e-07, + "loss": 0.4517, + "step": 15543 + }, + { + "epoch": 1.3245845760545376, + "grad_norm": 15.6875, + "learning_rate": 6.448944805722682e-07, + "loss": 0.3924, + "step": 15544 + }, + { + "epoch": 1.3246697912228376, + "grad_norm": 22.75, + "learning_rate": 6.447474513539262e-07, + "loss": 0.7529, + "step": 15545 + }, + { + "epoch": 1.3247550063911375, + "grad_norm": 23.0, + "learning_rate": 6.446004330731066e-07, + "loss": 0.9091, + "step": 15546 + }, + { + "epoch": 1.3248402215594375, + "grad_norm": 20.5, + "learning_rate": 6.444534257324664e-07, + "loss": 0.8651, + "step": 15547 + }, + { + "epoch": 1.3249254367277374, + "grad_norm": 11.75, + "learning_rate": 6.443064293346613e-07, + "loss": 0.4996, + "step": 15548 + }, + { + "epoch": 1.3250106518960374, + "grad_norm": 19.0, + "learning_rate": 6.441594438823485e-07, + "loss": 0.6837, + "step": 15549 + }, + { + "epoch": 1.3250958670643374, + "grad_norm": 22.25, + "learning_rate": 6.440124693781833e-07, + "loss": 0.6429, + "step": 15550 + }, + { + "epoch": 1.3251810822326373, + "grad_norm": 17.375, + "learning_rate": 6.438655058248225e-07, + "loss": 0.8109, + "step": 15551 + }, + { + "epoch": 1.3252662974009373, + "grad_norm": 15.125, + "learning_rate": 6.437185532249213e-07, + "loss": 0.6294, + "step": 15552 + }, + { + "epoch": 1.3253515125692372, + "grad_norm": 16.625, + "learning_rate": 6.435716115811352e-07, + "loss": 0.9326, + "step": 15553 + }, + { + "epoch": 1.3254367277375372, + "grad_norm": 17.125, + "learning_rate": 6.4342468089612e-07, + "loss": 0.6685, + "step": 15554 + }, + { + "epoch": 1.3255219429058371, + "grad_norm": 14.125, + "learning_rate": 6.432777611725315e-07, + "loss": 0.4013, + "step": 15555 + }, + { + "epoch": 1.325607158074137, + "grad_norm": 15.25, + "learning_rate": 6.431308524130233e-07, + "loss": 0.4833, + "step": 15556 + }, + { + "epoch": 1.325692373242437, + "grad_norm": 15.125, + "learning_rate": 6.429839546202513e-07, + "loss": 0.5507, + "step": 15557 + }, + { + "epoch": 1.325777588410737, + "grad_norm": 15.625, + "learning_rate": 6.428370677968691e-07, + "loss": 0.725, + "step": 15558 + }, + { + "epoch": 1.325862803579037, + "grad_norm": 15.5, + "learning_rate": 6.426901919455317e-07, + "loss": 0.3917, + "step": 15559 + }, + { + "epoch": 1.325948018747337, + "grad_norm": 13.4375, + "learning_rate": 6.425433270688939e-07, + "loss": 0.5027, + "step": 15560 + }, + { + "epoch": 1.3260332339156369, + "grad_norm": 25.75, + "learning_rate": 6.423964731696084e-07, + "loss": 0.6568, + "step": 15561 + }, + { + "epoch": 1.3261184490839368, + "grad_norm": 14.5, + "learning_rate": 6.422496302503295e-07, + "loss": 0.3881, + "step": 15562 + }, + { + "epoch": 1.3262036642522368, + "grad_norm": 16.875, + "learning_rate": 6.421027983137115e-07, + "loss": 0.631, + "step": 15563 + }, + { + "epoch": 1.3262888794205367, + "grad_norm": 21.375, + "learning_rate": 6.419559773624068e-07, + "loss": 0.8545, + "step": 15564 + }, + { + "epoch": 1.3263740945888367, + "grad_norm": 35.5, + "learning_rate": 6.418091673990692e-07, + "loss": 1.1535, + "step": 15565 + }, + { + "epoch": 1.3264593097571367, + "grad_norm": 19.375, + "learning_rate": 6.416623684263515e-07, + "loss": 0.7635, + "step": 15566 + }, + { + "epoch": 1.3265445249254366, + "grad_norm": 18.75, + "learning_rate": 6.415155804469061e-07, + "loss": 0.6744, + "step": 15567 + }, + { + "epoch": 1.3266297400937366, + "grad_norm": 13.1875, + "learning_rate": 6.413688034633869e-07, + "loss": 0.5515, + "step": 15568 + }, + { + "epoch": 1.3267149552620365, + "grad_norm": 31.5, + "learning_rate": 6.412220374784448e-07, + "loss": 0.7286, + "step": 15569 + }, + { + "epoch": 1.3268001704303365, + "grad_norm": 73.5, + "learning_rate": 6.410752824947327e-07, + "loss": 0.3932, + "step": 15570 + }, + { + "epoch": 1.3268853855986364, + "grad_norm": 14.875, + "learning_rate": 6.409285385149025e-07, + "loss": 0.3961, + "step": 15571 + }, + { + "epoch": 1.3269706007669364, + "grad_norm": 18.625, + "learning_rate": 6.407818055416066e-07, + "loss": 0.6192, + "step": 15572 + }, + { + "epoch": 1.3270558159352364, + "grad_norm": 18.0, + "learning_rate": 6.406350835774956e-07, + "loss": 0.8179, + "step": 15573 + }, + { + "epoch": 1.3271410311035363, + "grad_norm": 15.125, + "learning_rate": 6.404883726252219e-07, + "loss": 0.5299, + "step": 15574 + }, + { + "epoch": 1.3272262462718363, + "grad_norm": 14.9375, + "learning_rate": 6.403416726874356e-07, + "loss": 0.6936, + "step": 15575 + }, + { + "epoch": 1.3273114614401362, + "grad_norm": 16.25, + "learning_rate": 6.401949837667886e-07, + "loss": 0.8529, + "step": 15576 + }, + { + "epoch": 1.3273966766084362, + "grad_norm": 15.375, + "learning_rate": 6.400483058659319e-07, + "loss": 0.7732, + "step": 15577 + }, + { + "epoch": 1.3274818917767361, + "grad_norm": 15.5, + "learning_rate": 6.399016389875149e-07, + "loss": 0.5452, + "step": 15578 + }, + { + "epoch": 1.327567106945036, + "grad_norm": 13.5, + "learning_rate": 6.39754983134189e-07, + "loss": 0.3446, + "step": 15579 + }, + { + "epoch": 1.327652322113336, + "grad_norm": 13.3125, + "learning_rate": 6.396083383086046e-07, + "loss": 0.31, + "step": 15580 + }, + { + "epoch": 1.327737537281636, + "grad_norm": 15.25, + "learning_rate": 6.394617045134109e-07, + "loss": 0.5258, + "step": 15581 + }, + { + "epoch": 1.327822752449936, + "grad_norm": 16.125, + "learning_rate": 6.393150817512586e-07, + "loss": 0.6396, + "step": 15582 + }, + { + "epoch": 1.327907967618236, + "grad_norm": 16.25, + "learning_rate": 6.391684700247963e-07, + "loss": 0.7739, + "step": 15583 + }, + { + "epoch": 1.3279931827865359, + "grad_norm": 15.0625, + "learning_rate": 6.390218693366743e-07, + "loss": 0.6652, + "step": 15584 + }, + { + "epoch": 1.3280783979548358, + "grad_norm": 15.625, + "learning_rate": 6.388752796895417e-07, + "loss": 0.7387, + "step": 15585 + }, + { + "epoch": 1.3281636131231358, + "grad_norm": 16.125, + "learning_rate": 6.387287010860468e-07, + "loss": 0.7027, + "step": 15586 + }, + { + "epoch": 1.3282488282914358, + "grad_norm": 14.1875, + "learning_rate": 6.385821335288392e-07, + "loss": 0.4567, + "step": 15587 + }, + { + "epoch": 1.3283340434597357, + "grad_norm": 15.0625, + "learning_rate": 6.38435577020567e-07, + "loss": 0.6818, + "step": 15588 + }, + { + "epoch": 1.3284192586280357, + "grad_norm": 12.9375, + "learning_rate": 6.382890315638796e-07, + "loss": 0.4347, + "step": 15589 + }, + { + "epoch": 1.3285044737963356, + "grad_norm": 18.625, + "learning_rate": 6.381424971614244e-07, + "loss": 1.0646, + "step": 15590 + }, + { + "epoch": 1.3285896889646356, + "grad_norm": 15.4375, + "learning_rate": 6.379959738158492e-07, + "loss": 0.728, + "step": 15591 + }, + { + "epoch": 1.3286749041329355, + "grad_norm": 15.3125, + "learning_rate": 6.37849461529802e-07, + "loss": 0.3933, + "step": 15592 + }, + { + "epoch": 1.3287601193012355, + "grad_norm": 12.25, + "learning_rate": 6.377029603059312e-07, + "loss": 0.3987, + "step": 15593 + }, + { + "epoch": 1.3288453344695355, + "grad_norm": 15.5625, + "learning_rate": 6.375564701468831e-07, + "loss": 0.4494, + "step": 15594 + }, + { + "epoch": 1.3289305496378354, + "grad_norm": 18.125, + "learning_rate": 6.374099910553054e-07, + "loss": 0.6786, + "step": 15595 + }, + { + "epoch": 1.3290157648061354, + "grad_norm": 18.5, + "learning_rate": 6.372635230338453e-07, + "loss": 0.6173, + "step": 15596 + }, + { + "epoch": 1.3291009799744353, + "grad_norm": 22.25, + "learning_rate": 6.371170660851498e-07, + "loss": 0.7306, + "step": 15597 + }, + { + "epoch": 1.3291861951427353, + "grad_norm": 13.4375, + "learning_rate": 6.369706202118653e-07, + "loss": 0.4737, + "step": 15598 + }, + { + "epoch": 1.3292714103110352, + "grad_norm": 13.6875, + "learning_rate": 6.368241854166377e-07, + "loss": 0.4495, + "step": 15599 + }, + { + "epoch": 1.3293566254793352, + "grad_norm": 19.75, + "learning_rate": 6.366777617021136e-07, + "loss": 0.7747, + "step": 15600 + }, + { + "epoch": 1.3294418406476352, + "grad_norm": 14.1875, + "learning_rate": 6.365313490709391e-07, + "loss": 0.7337, + "step": 15601 + }, + { + "epoch": 1.329527055815935, + "grad_norm": 12.75, + "learning_rate": 6.363849475257605e-07, + "loss": 0.4389, + "step": 15602 + }, + { + "epoch": 1.329612270984235, + "grad_norm": 17.125, + "learning_rate": 6.362385570692225e-07, + "loss": 0.6661, + "step": 15603 + }, + { + "epoch": 1.329697486152535, + "grad_norm": 22.25, + "learning_rate": 6.360921777039709e-07, + "loss": 0.7444, + "step": 15604 + }, + { + "epoch": 1.329782701320835, + "grad_norm": 20.625, + "learning_rate": 6.35945809432651e-07, + "loss": 0.8543, + "step": 15605 + }, + { + "epoch": 1.329867916489135, + "grad_norm": 10.8125, + "learning_rate": 6.357994522579083e-07, + "loss": 0.3088, + "step": 15606 + }, + { + "epoch": 1.329953131657435, + "grad_norm": 13.8125, + "learning_rate": 6.35653106182387e-07, + "loss": 0.5507, + "step": 15607 + }, + { + "epoch": 1.3300383468257349, + "grad_norm": 14.0, + "learning_rate": 6.355067712087315e-07, + "loss": 0.404, + "step": 15608 + }, + { + "epoch": 1.3301235619940348, + "grad_norm": 12.9375, + "learning_rate": 6.353604473395866e-07, + "loss": 0.3057, + "step": 15609 + }, + { + "epoch": 1.3302087771623348, + "grad_norm": 13.1875, + "learning_rate": 6.35214134577597e-07, + "loss": 0.4884, + "step": 15610 + }, + { + "epoch": 1.3302939923306347, + "grad_norm": 13.75, + "learning_rate": 6.350678329254057e-07, + "loss": 0.4596, + "step": 15611 + }, + { + "epoch": 1.3303792074989347, + "grad_norm": 19.375, + "learning_rate": 6.349215423856568e-07, + "loss": 0.7621, + "step": 15612 + }, + { + "epoch": 1.3304644226672346, + "grad_norm": 33.0, + "learning_rate": 6.347752629609943e-07, + "loss": 1.1956, + "step": 15613 + }, + { + "epoch": 1.3305496378355346, + "grad_norm": 16.5, + "learning_rate": 6.34628994654062e-07, + "loss": 0.9324, + "step": 15614 + }, + { + "epoch": 1.3306348530038346, + "grad_norm": 13.25, + "learning_rate": 6.344827374675024e-07, + "loss": 0.3379, + "step": 15615 + }, + { + "epoch": 1.3307200681721345, + "grad_norm": 22.125, + "learning_rate": 6.343364914039583e-07, + "loss": 0.7926, + "step": 15616 + }, + { + "epoch": 1.3308052833404345, + "grad_norm": 20.5, + "learning_rate": 6.341902564660729e-07, + "loss": 0.6884, + "step": 15617 + }, + { + "epoch": 1.3308904985087344, + "grad_norm": 15.0, + "learning_rate": 6.340440326564888e-07, + "loss": 0.7279, + "step": 15618 + }, + { + "epoch": 1.3309757136770344, + "grad_norm": 15.5625, + "learning_rate": 6.338978199778489e-07, + "loss": 0.5522, + "step": 15619 + }, + { + "epoch": 1.3310609288453343, + "grad_norm": 13.9375, + "learning_rate": 6.337516184327944e-07, + "loss": 0.4233, + "step": 15620 + }, + { + "epoch": 1.3311461440136343, + "grad_norm": 11.8125, + "learning_rate": 6.33605428023968e-07, + "loss": 0.5019, + "step": 15621 + }, + { + "epoch": 1.3312313591819342, + "grad_norm": 18.75, + "learning_rate": 6.334592487540117e-07, + "loss": 0.8607, + "step": 15622 + }, + { + "epoch": 1.3313165743502342, + "grad_norm": 18.625, + "learning_rate": 6.333130806255668e-07, + "loss": 0.5629, + "step": 15623 + }, + { + "epoch": 1.3314017895185342, + "grad_norm": 17.0, + "learning_rate": 6.33166923641274e-07, + "loss": 0.6582, + "step": 15624 + }, + { + "epoch": 1.3314870046868341, + "grad_norm": 11.0, + "learning_rate": 6.330207778037754e-07, + "loss": 0.354, + "step": 15625 + }, + { + "epoch": 1.331572219855134, + "grad_norm": 14.9375, + "learning_rate": 6.328746431157118e-07, + "loss": 0.6407, + "step": 15626 + }, + { + "epoch": 1.331657435023434, + "grad_norm": 18.625, + "learning_rate": 6.327285195797242e-07, + "loss": 0.6945, + "step": 15627 + }, + { + "epoch": 1.3317426501917342, + "grad_norm": 16.25, + "learning_rate": 6.325824071984526e-07, + "loss": 0.5635, + "step": 15628 + }, + { + "epoch": 1.3318278653600342, + "grad_norm": 12.875, + "learning_rate": 6.324363059745378e-07, + "loss": 0.4307, + "step": 15629 + }, + { + "epoch": 1.3319130805283341, + "grad_norm": 16.0, + "learning_rate": 6.322902159106205e-07, + "loss": 0.8672, + "step": 15630 + }, + { + "epoch": 1.331998295696634, + "grad_norm": 15.5, + "learning_rate": 6.321441370093396e-07, + "loss": 0.7528, + "step": 15631 + }, + { + "epoch": 1.332083510864934, + "grad_norm": 17.0, + "learning_rate": 6.31998069273336e-07, + "loss": 0.6148, + "step": 15632 + }, + { + "epoch": 1.332168726033234, + "grad_norm": 12.8125, + "learning_rate": 6.318520127052481e-07, + "loss": 0.4522, + "step": 15633 + }, + { + "epoch": 1.332253941201534, + "grad_norm": 19.0, + "learning_rate": 6.317059673077161e-07, + "loss": 0.8924, + "step": 15634 + }, + { + "epoch": 1.332339156369834, + "grad_norm": 16.5, + "learning_rate": 6.315599330833788e-07, + "loss": 0.615, + "step": 15635 + }, + { + "epoch": 1.3324243715381339, + "grad_norm": 14.125, + "learning_rate": 6.314139100348759e-07, + "loss": 0.4373, + "step": 15636 + }, + { + "epoch": 1.3325095867064338, + "grad_norm": 12.625, + "learning_rate": 6.312678981648454e-07, + "loss": 0.4595, + "step": 15637 + }, + { + "epoch": 1.3325948018747338, + "grad_norm": 12.5, + "learning_rate": 6.311218974759262e-07, + "loss": 0.3226, + "step": 15638 + }, + { + "epoch": 1.3326800170430337, + "grad_norm": 17.5, + "learning_rate": 6.30975907970757e-07, + "loss": 0.7782, + "step": 15639 + }, + { + "epoch": 1.3327652322113337, + "grad_norm": 16.375, + "learning_rate": 6.308299296519754e-07, + "loss": 0.5207, + "step": 15640 + }, + { + "epoch": 1.3328504473796337, + "grad_norm": 15.375, + "learning_rate": 6.306839625222194e-07, + "loss": 0.4249, + "step": 15641 + }, + { + "epoch": 1.3329356625479336, + "grad_norm": 13.75, + "learning_rate": 6.305380065841269e-07, + "loss": 0.6355, + "step": 15642 + }, + { + "epoch": 1.3330208777162336, + "grad_norm": 12.625, + "learning_rate": 6.303920618403356e-07, + "loss": 0.4853, + "step": 15643 + }, + { + "epoch": 1.3331060928845335, + "grad_norm": 26.0, + "learning_rate": 6.302461282934832e-07, + "loss": 0.9384, + "step": 15644 + }, + { + "epoch": 1.3331913080528335, + "grad_norm": 26.125, + "learning_rate": 6.301002059462059e-07, + "loss": 0.8621, + "step": 15645 + }, + { + "epoch": 1.3332765232211334, + "grad_norm": 13.625, + "learning_rate": 6.299542948011413e-07, + "loss": 0.5483, + "step": 15646 + }, + { + "epoch": 1.3333617383894334, + "grad_norm": 13.3125, + "learning_rate": 6.298083948609268e-07, + "loss": 0.5552, + "step": 15647 + }, + { + "epoch": 1.3334469535577333, + "grad_norm": 16.5, + "learning_rate": 6.296625061281975e-07, + "loss": 0.8563, + "step": 15648 + }, + { + "epoch": 1.3335321687260333, + "grad_norm": 10.375, + "learning_rate": 6.295166286055912e-07, + "loss": 0.2181, + "step": 15649 + }, + { + "epoch": 1.3336173838943333, + "grad_norm": 18.5, + "learning_rate": 6.293707622957426e-07, + "loss": 0.4174, + "step": 15650 + }, + { + "epoch": 1.3337025990626332, + "grad_norm": 17.75, + "learning_rate": 6.292249072012888e-07, + "loss": 0.6716, + "step": 15651 + }, + { + "epoch": 1.3337878142309332, + "grad_norm": 13.75, + "learning_rate": 6.290790633248648e-07, + "loss": 0.4388, + "step": 15652 + }, + { + "epoch": 1.3338730293992331, + "grad_norm": 17.25, + "learning_rate": 6.289332306691072e-07, + "loss": 0.4727, + "step": 15653 + }, + { + "epoch": 1.333958244567533, + "grad_norm": 15.5625, + "learning_rate": 6.2878740923665e-07, + "loss": 0.4555, + "step": 15654 + }, + { + "epoch": 1.334043459735833, + "grad_norm": 13.875, + "learning_rate": 6.286415990301297e-07, + "loss": 0.5666, + "step": 15655 + }, + { + "epoch": 1.334128674904133, + "grad_norm": 23.75, + "learning_rate": 6.284958000521799e-07, + "loss": 1.01, + "step": 15656 + }, + { + "epoch": 1.334213890072433, + "grad_norm": 16.875, + "learning_rate": 6.283500123054365e-07, + "loss": 0.7886, + "step": 15657 + }, + { + "epoch": 1.334299105240733, + "grad_norm": 10.9375, + "learning_rate": 6.28204235792533e-07, + "loss": 0.4274, + "step": 15658 + }, + { + "epoch": 1.3343843204090329, + "grad_norm": 15.8125, + "learning_rate": 6.280584705161041e-07, + "loss": 0.5967, + "step": 15659 + }, + { + "epoch": 1.3344695355773328, + "grad_norm": 15.125, + "learning_rate": 6.279127164787843e-07, + "loss": 0.4639, + "step": 15660 + }, + { + "epoch": 1.3345547507456328, + "grad_norm": 33.25, + "learning_rate": 6.277669736832076e-07, + "loss": 0.9024, + "step": 15661 + }, + { + "epoch": 1.3346399659139327, + "grad_norm": 17.75, + "learning_rate": 6.276212421320069e-07, + "loss": 0.6282, + "step": 15662 + }, + { + "epoch": 1.3347251810822327, + "grad_norm": 12.3125, + "learning_rate": 6.274755218278167e-07, + "loss": 0.3909, + "step": 15663 + }, + { + "epoch": 1.3348103962505327, + "grad_norm": 15.3125, + "learning_rate": 6.273298127732693e-07, + "loss": 0.3907, + "step": 15664 + }, + { + "epoch": 1.3348956114188326, + "grad_norm": 18.75, + "learning_rate": 6.271841149709984e-07, + "loss": 0.4667, + "step": 15665 + }, + { + "epoch": 1.3349808265871326, + "grad_norm": 21.75, + "learning_rate": 6.270384284236373e-07, + "loss": 0.4605, + "step": 15666 + }, + { + "epoch": 1.3350660417554325, + "grad_norm": 15.0, + "learning_rate": 6.268927531338178e-07, + "loss": 0.5187, + "step": 15667 + }, + { + "epoch": 1.3351512569237325, + "grad_norm": 16.875, + "learning_rate": 6.267470891041727e-07, + "loss": 0.9368, + "step": 15668 + }, + { + "epoch": 1.3352364720920324, + "grad_norm": 15.25, + "learning_rate": 6.266014363373347e-07, + "loss": 0.4576, + "step": 15669 + }, + { + "epoch": 1.3353216872603324, + "grad_norm": 14.6875, + "learning_rate": 6.264557948359358e-07, + "loss": 0.4103, + "step": 15670 + }, + { + "epoch": 1.3354069024286324, + "grad_norm": 18.0, + "learning_rate": 6.263101646026079e-07, + "loss": 0.6336, + "step": 15671 + }, + { + "epoch": 1.3354921175969323, + "grad_norm": 16.875, + "learning_rate": 6.261645456399818e-07, + "loss": 0.8661, + "step": 15672 + }, + { + "epoch": 1.3355773327652323, + "grad_norm": 19.875, + "learning_rate": 6.260189379506898e-07, + "loss": 0.576, + "step": 15673 + }, + { + "epoch": 1.3356625479335322, + "grad_norm": 19.875, + "learning_rate": 6.258733415373636e-07, + "loss": 0.969, + "step": 15674 + }, + { + "epoch": 1.3357477631018322, + "grad_norm": 10.8125, + "learning_rate": 6.257277564026332e-07, + "loss": 0.3217, + "step": 15675 + }, + { + "epoch": 1.3358329782701321, + "grad_norm": 25.875, + "learning_rate": 6.2558218254913e-07, + "loss": 1.048, + "step": 15676 + }, + { + "epoch": 1.335918193438432, + "grad_norm": 11.0625, + "learning_rate": 6.254366199794847e-07, + "loss": 0.3052, + "step": 15677 + }, + { + "epoch": 1.336003408606732, + "grad_norm": 18.375, + "learning_rate": 6.252910686963282e-07, + "loss": 0.6116, + "step": 15678 + }, + { + "epoch": 1.336088623775032, + "grad_norm": 16.25, + "learning_rate": 6.251455287022901e-07, + "loss": 0.5934, + "step": 15679 + }, + { + "epoch": 1.336173838943332, + "grad_norm": 12.4375, + "learning_rate": 6.250000000000004e-07, + "loss": 0.5959, + "step": 15680 + }, + { + "epoch": 1.336259054111632, + "grad_norm": 14.0625, + "learning_rate": 6.248544825920889e-07, + "loss": 0.493, + "step": 15681 + }, + { + "epoch": 1.3363442692799319, + "grad_norm": 13.375, + "learning_rate": 6.247089764811857e-07, + "loss": 0.5888, + "step": 15682 + }, + { + "epoch": 1.3364294844482318, + "grad_norm": 16.625, + "learning_rate": 6.245634816699206e-07, + "loss": 0.6976, + "step": 15683 + }, + { + "epoch": 1.3365146996165318, + "grad_norm": 15.25, + "learning_rate": 6.244179981609217e-07, + "loss": 0.9093, + "step": 15684 + }, + { + "epoch": 1.3365999147848318, + "grad_norm": 12.625, + "learning_rate": 6.242725259568189e-07, + "loss": 0.3541, + "step": 15685 + }, + { + "epoch": 1.3366851299531317, + "grad_norm": 13.375, + "learning_rate": 6.241270650602406e-07, + "loss": 0.3877, + "step": 15686 + }, + { + "epoch": 1.3367703451214317, + "grad_norm": 14.5625, + "learning_rate": 6.239816154738166e-07, + "loss": 0.6845, + "step": 15687 + }, + { + "epoch": 1.3368555602897316, + "grad_norm": 13.75, + "learning_rate": 6.238361772001733e-07, + "loss": 0.5618, + "step": 15688 + }, + { + "epoch": 1.3369407754580316, + "grad_norm": 15.25, + "learning_rate": 6.2369075024194e-07, + "loss": 0.6768, + "step": 15689 + }, + { + "epoch": 1.3370259906263315, + "grad_norm": 16.5, + "learning_rate": 6.235453346017444e-07, + "loss": 0.7249, + "step": 15690 + }, + { + "epoch": 1.3371112057946315, + "grad_norm": 18.0, + "learning_rate": 6.233999302822154e-07, + "loss": 0.5653, + "step": 15691 + }, + { + "epoch": 1.3371964209629315, + "grad_norm": 15.0625, + "learning_rate": 6.232545372859791e-07, + "loss": 0.3707, + "step": 15692 + }, + { + "epoch": 1.3372816361312314, + "grad_norm": 13.9375, + "learning_rate": 6.231091556156634e-07, + "loss": 0.5516, + "step": 15693 + }, + { + "epoch": 1.3373668512995314, + "grad_norm": 17.375, + "learning_rate": 6.22963785273896e-07, + "loss": 0.7924, + "step": 15694 + }, + { + "epoch": 1.3374520664678313, + "grad_norm": 14.9375, + "learning_rate": 6.228184262633037e-07, + "loss": 0.5161, + "step": 15695 + }, + { + "epoch": 1.3375372816361313, + "grad_norm": 19.125, + "learning_rate": 6.226730785865133e-07, + "loss": 0.7232, + "step": 15696 + }, + { + "epoch": 1.3376224968044312, + "grad_norm": 17.0, + "learning_rate": 6.225277422461507e-07, + "loss": 0.8308, + "step": 15697 + }, + { + "epoch": 1.3377077119727312, + "grad_norm": 11.75, + "learning_rate": 6.223824172448427e-07, + "loss": 0.4218, + "step": 15698 + }, + { + "epoch": 1.3377929271410312, + "grad_norm": 21.125, + "learning_rate": 6.222371035852156e-07, + "loss": 0.7345, + "step": 15699 + }, + { + "epoch": 1.337878142309331, + "grad_norm": 13.125, + "learning_rate": 6.220918012698957e-07, + "loss": 0.25, + "step": 15700 + }, + { + "epoch": 1.337963357477631, + "grad_norm": 15.9375, + "learning_rate": 6.219465103015079e-07, + "loss": 0.6474, + "step": 15701 + }, + { + "epoch": 1.338048572645931, + "grad_norm": 21.125, + "learning_rate": 6.218012306826783e-07, + "loss": 0.7271, + "step": 15702 + }, + { + "epoch": 1.338133787814231, + "grad_norm": 11.9375, + "learning_rate": 6.216559624160326e-07, + "loss": 0.4292, + "step": 15703 + }, + { + "epoch": 1.338219002982531, + "grad_norm": 14.6875, + "learning_rate": 6.215107055041955e-07, + "loss": 0.4744, + "step": 15704 + }, + { + "epoch": 1.338304218150831, + "grad_norm": 21.5, + "learning_rate": 6.213654599497914e-07, + "loss": 0.7093, + "step": 15705 + }, + { + "epoch": 1.3383894333191309, + "grad_norm": 15.3125, + "learning_rate": 6.212202257554458e-07, + "loss": 0.5499, + "step": 15706 + }, + { + "epoch": 1.3384746484874308, + "grad_norm": 19.875, + "learning_rate": 6.210750029237827e-07, + "loss": 0.9656, + "step": 15707 + }, + { + "epoch": 1.3385598636557308, + "grad_norm": 18.5, + "learning_rate": 6.209297914574274e-07, + "loss": 0.9574, + "step": 15708 + }, + { + "epoch": 1.3386450788240307, + "grad_norm": 14.625, + "learning_rate": 6.207845913590028e-07, + "loss": 0.71, + "step": 15709 + }, + { + "epoch": 1.3387302939923307, + "grad_norm": 13.4375, + "learning_rate": 6.206394026311334e-07, + "loss": 0.5585, + "step": 15710 + }, + { + "epoch": 1.3388155091606306, + "grad_norm": 16.25, + "learning_rate": 6.204942252764429e-07, + "loss": 0.5981, + "step": 15711 + }, + { + "epoch": 1.3389007243289306, + "grad_norm": 15.6875, + "learning_rate": 6.203490592975552e-07, + "loss": 0.7555, + "step": 15712 + }, + { + "epoch": 1.3389859394972305, + "grad_norm": 20.625, + "learning_rate": 6.202039046970932e-07, + "loss": 0.6557, + "step": 15713 + }, + { + "epoch": 1.3390711546655305, + "grad_norm": 14.3125, + "learning_rate": 6.200587614776795e-07, + "loss": 0.5673, + "step": 15714 + }, + { + "epoch": 1.3391563698338305, + "grad_norm": 12.4375, + "learning_rate": 6.199136296419372e-07, + "loss": 0.3766, + "step": 15715 + }, + { + "epoch": 1.3392415850021304, + "grad_norm": 16.75, + "learning_rate": 6.197685091924895e-07, + "loss": 0.6367, + "step": 15716 + }, + { + "epoch": 1.3393268001704304, + "grad_norm": 17.125, + "learning_rate": 6.196234001319589e-07, + "loss": 0.7573, + "step": 15717 + }, + { + "epoch": 1.3394120153387303, + "grad_norm": 18.75, + "learning_rate": 6.19478302462967e-07, + "loss": 0.8693, + "step": 15718 + }, + { + "epoch": 1.3394972305070303, + "grad_norm": 22.75, + "learning_rate": 6.19333216188136e-07, + "loss": 1.0281, + "step": 15719 + }, + { + "epoch": 1.3395824456753302, + "grad_norm": 16.0, + "learning_rate": 6.191881413100886e-07, + "loss": 0.5265, + "step": 15720 + }, + { + "epoch": 1.3396676608436302, + "grad_norm": 13.6875, + "learning_rate": 6.190430778314457e-07, + "loss": 0.5743, + "step": 15721 + }, + { + "epoch": 1.3397528760119302, + "grad_norm": 12.4375, + "learning_rate": 6.188980257548286e-07, + "loss": 0.3437, + "step": 15722 + }, + { + "epoch": 1.3398380911802301, + "grad_norm": 16.625, + "learning_rate": 6.187529850828587e-07, + "loss": 0.5188, + "step": 15723 + }, + { + "epoch": 1.33992330634853, + "grad_norm": 12.3125, + "learning_rate": 6.186079558181571e-07, + "loss": 0.516, + "step": 15724 + }, + { + "epoch": 1.34000852151683, + "grad_norm": 16.5, + "learning_rate": 6.184629379633452e-07, + "loss": 0.6354, + "step": 15725 + }, + { + "epoch": 1.34009373668513, + "grad_norm": 12.75, + "learning_rate": 6.183179315210425e-07, + "loss": 0.3984, + "step": 15726 + }, + { + "epoch": 1.34017895185343, + "grad_norm": 25.375, + "learning_rate": 6.181729364938699e-07, + "loss": 1.2726, + "step": 15727 + }, + { + "epoch": 1.34026416702173, + "grad_norm": 18.0, + "learning_rate": 6.180279528844482e-07, + "loss": 0.767, + "step": 15728 + }, + { + "epoch": 1.3403493821900299, + "grad_norm": 16.375, + "learning_rate": 6.178829806953966e-07, + "loss": 0.7979, + "step": 15729 + }, + { + "epoch": 1.3404345973583298, + "grad_norm": 20.5, + "learning_rate": 6.177380199293355e-07, + "loss": 0.707, + "step": 15730 + }, + { + "epoch": 1.3405198125266298, + "grad_norm": 20.375, + "learning_rate": 6.175930705888838e-07, + "loss": 0.6921, + "step": 15731 + }, + { + "epoch": 1.3406050276949297, + "grad_norm": 17.5, + "learning_rate": 6.174481326766611e-07, + "loss": 0.813, + "step": 15732 + }, + { + "epoch": 1.3406902428632297, + "grad_norm": 16.75, + "learning_rate": 6.173032061952873e-07, + "loss": 0.5818, + "step": 15733 + }, + { + "epoch": 1.3407754580315296, + "grad_norm": 17.125, + "learning_rate": 6.171582911473803e-07, + "loss": 0.7163, + "step": 15734 + }, + { + "epoch": 1.3408606731998296, + "grad_norm": 17.875, + "learning_rate": 6.170133875355594e-07, + "loss": 0.7903, + "step": 15735 + }, + { + "epoch": 1.3409458883681296, + "grad_norm": 13.1875, + "learning_rate": 6.168684953624434e-07, + "loss": 0.3583, + "step": 15736 + }, + { + "epoch": 1.3410311035364295, + "grad_norm": 16.25, + "learning_rate": 6.167236146306501e-07, + "loss": 0.493, + "step": 15737 + }, + { + "epoch": 1.3411163187047295, + "grad_norm": 17.0, + "learning_rate": 6.165787453427982e-07, + "loss": 0.9997, + "step": 15738 + }, + { + "epoch": 1.3412015338730294, + "grad_norm": 16.0, + "learning_rate": 6.16433887501505e-07, + "loss": 0.439, + "step": 15739 + }, + { + "epoch": 1.3412867490413294, + "grad_norm": 15.125, + "learning_rate": 6.162890411093885e-07, + "loss": 0.4338, + "step": 15740 + }, + { + "epoch": 1.3413719642096293, + "grad_norm": 14.9375, + "learning_rate": 6.161442061690662e-07, + "loss": 0.5005, + "step": 15741 + }, + { + "epoch": 1.3414571793779293, + "grad_norm": 16.625, + "learning_rate": 6.15999382683156e-07, + "loss": 0.4024, + "step": 15742 + }, + { + "epoch": 1.3415423945462293, + "grad_norm": 16.75, + "learning_rate": 6.15854570654274e-07, + "loss": 0.3623, + "step": 15743 + }, + { + "epoch": 1.3416276097145292, + "grad_norm": 12.5, + "learning_rate": 6.157097700850381e-07, + "loss": 0.4578, + "step": 15744 + }, + { + "epoch": 1.3417128248828292, + "grad_norm": 12.1875, + "learning_rate": 6.155649809780641e-07, + "loss": 0.4782, + "step": 15745 + }, + { + "epoch": 1.3417980400511291, + "grad_norm": 18.5, + "learning_rate": 6.154202033359686e-07, + "loss": 0.5755, + "step": 15746 + }, + { + "epoch": 1.341883255219429, + "grad_norm": 18.875, + "learning_rate": 6.15275437161369e-07, + "loss": 0.4891, + "step": 15747 + }, + { + "epoch": 1.341968470387729, + "grad_norm": 18.25, + "learning_rate": 6.151306824568797e-07, + "loss": 0.4314, + "step": 15748 + }, + { + "epoch": 1.342053685556029, + "grad_norm": 24.125, + "learning_rate": 6.149859392251176e-07, + "loss": 0.8643, + "step": 15749 + }, + { + "epoch": 1.342138900724329, + "grad_norm": 36.25, + "learning_rate": 6.148412074686984e-07, + "loss": 0.674, + "step": 15750 + }, + { + "epoch": 1.342224115892629, + "grad_norm": 23.0, + "learning_rate": 6.146964871902369e-07, + "loss": 0.6592, + "step": 15751 + }, + { + "epoch": 1.3423093310609289, + "grad_norm": 16.125, + "learning_rate": 6.145517783923493e-07, + "loss": 0.6739, + "step": 15752 + }, + { + "epoch": 1.3423945462292288, + "grad_norm": 18.0, + "learning_rate": 6.144070810776495e-07, + "loss": 0.5365, + "step": 15753 + }, + { + "epoch": 1.3424797613975288, + "grad_norm": 20.75, + "learning_rate": 6.142623952487528e-07, + "loss": 0.5746, + "step": 15754 + }, + { + "epoch": 1.3425649765658287, + "grad_norm": 9.0625, + "learning_rate": 6.141177209082745e-07, + "loss": 0.2225, + "step": 15755 + }, + { + "epoch": 1.3426501917341287, + "grad_norm": 17.0, + "learning_rate": 6.13973058058828e-07, + "loss": 0.8738, + "step": 15756 + }, + { + "epoch": 1.3427354069024287, + "grad_norm": 17.875, + "learning_rate": 6.138284067030277e-07, + "loss": 0.5033, + "step": 15757 + }, + { + "epoch": 1.3428206220707286, + "grad_norm": 14.375, + "learning_rate": 6.136837668434879e-07, + "loss": 0.5787, + "step": 15758 + }, + { + "epoch": 1.3429058372390286, + "grad_norm": 11.8125, + "learning_rate": 6.135391384828229e-07, + "loss": 0.2792, + "step": 15759 + }, + { + "epoch": 1.3429910524073285, + "grad_norm": 13.5, + "learning_rate": 6.133945216236456e-07, + "loss": 0.4844, + "step": 15760 + }, + { + "epoch": 1.3430762675756285, + "grad_norm": 25.875, + "learning_rate": 6.132499162685689e-07, + "loss": 0.4957, + "step": 15761 + }, + { + "epoch": 1.3431614827439284, + "grad_norm": 16.75, + "learning_rate": 6.131053224202066e-07, + "loss": 0.6553, + "step": 15762 + }, + { + "epoch": 1.3432466979122284, + "grad_norm": 14.75, + "learning_rate": 6.12960740081172e-07, + "loss": 0.7764, + "step": 15763 + }, + { + "epoch": 1.3433319130805284, + "grad_norm": 14.75, + "learning_rate": 6.128161692540769e-07, + "loss": 0.6774, + "step": 15764 + }, + { + "epoch": 1.3434171282488283, + "grad_norm": 17.75, + "learning_rate": 6.126716099415345e-07, + "loss": 0.6884, + "step": 15765 + }, + { + "epoch": 1.3435023434171283, + "grad_norm": 20.125, + "learning_rate": 6.125270621461568e-07, + "loss": 0.7481, + "step": 15766 + }, + { + "epoch": 1.3435875585854282, + "grad_norm": 13.875, + "learning_rate": 6.123825258705568e-07, + "loss": 0.4403, + "step": 15767 + }, + { + "epoch": 1.3436727737537282, + "grad_norm": 15.9375, + "learning_rate": 6.122380011173457e-07, + "loss": 0.3084, + "step": 15768 + }, + { + "epoch": 1.3437579889220281, + "grad_norm": 23.5, + "learning_rate": 6.120934878891346e-07, + "loss": 0.5549, + "step": 15769 + }, + { + "epoch": 1.343843204090328, + "grad_norm": 18.25, + "learning_rate": 6.119489861885357e-07, + "loss": 0.7841, + "step": 15770 + }, + { + "epoch": 1.343928419258628, + "grad_norm": 18.0, + "learning_rate": 6.118044960181603e-07, + "loss": 0.5578, + "step": 15771 + }, + { + "epoch": 1.344013634426928, + "grad_norm": 13.25, + "learning_rate": 6.116600173806199e-07, + "loss": 0.5436, + "step": 15772 + }, + { + "epoch": 1.344098849595228, + "grad_norm": 11.8125, + "learning_rate": 6.115155502785243e-07, + "loss": 0.3165, + "step": 15773 + }, + { + "epoch": 1.344184064763528, + "grad_norm": 16.875, + "learning_rate": 6.113710947144847e-07, + "loss": 0.5749, + "step": 15774 + }, + { + "epoch": 1.3442692799318279, + "grad_norm": 16.875, + "learning_rate": 6.112266506911118e-07, + "loss": 0.6834, + "step": 15775 + }, + { + "epoch": 1.3443544951001278, + "grad_norm": 17.0, + "learning_rate": 6.110822182110162e-07, + "loss": 0.6688, + "step": 15776 + }, + { + "epoch": 1.3444397102684278, + "grad_norm": 15.6875, + "learning_rate": 6.109377972768071e-07, + "loss": 0.5449, + "step": 15777 + }, + { + "epoch": 1.3445249254367277, + "grad_norm": 15.1875, + "learning_rate": 6.107933878910943e-07, + "loss": 0.4237, + "step": 15778 + }, + { + "epoch": 1.3446101406050277, + "grad_norm": 14.375, + "learning_rate": 6.106489900564879e-07, + "loss": 0.7421, + "step": 15779 + }, + { + "epoch": 1.3446953557733277, + "grad_norm": 22.125, + "learning_rate": 6.105046037755974e-07, + "loss": 0.9309, + "step": 15780 + }, + { + "epoch": 1.3447805709416276, + "grad_norm": 16.75, + "learning_rate": 6.103602290510315e-07, + "loss": 0.5672, + "step": 15781 + }, + { + "epoch": 1.3448657861099276, + "grad_norm": 13.375, + "learning_rate": 6.102158658853994e-07, + "loss": 0.4657, + "step": 15782 + }, + { + "epoch": 1.3449510012782275, + "grad_norm": 20.375, + "learning_rate": 6.100715142813098e-07, + "loss": 0.5265, + "step": 15783 + }, + { + "epoch": 1.3450362164465275, + "grad_norm": 33.75, + "learning_rate": 6.099271742413721e-07, + "loss": 0.7699, + "step": 15784 + }, + { + "epoch": 1.3451214316148274, + "grad_norm": 15.125, + "learning_rate": 6.09782845768194e-07, + "loss": 0.4127, + "step": 15785 + }, + { + "epoch": 1.3452066467831274, + "grad_norm": 22.375, + "learning_rate": 6.09638528864383e-07, + "loss": 0.9485, + "step": 15786 + }, + { + "epoch": 1.3452918619514274, + "grad_norm": 16.25, + "learning_rate": 6.09494223532548e-07, + "loss": 0.7417, + "step": 15787 + }, + { + "epoch": 1.3453770771197273, + "grad_norm": 9.4375, + "learning_rate": 6.093499297752962e-07, + "loss": 0.2085, + "step": 15788 + }, + { + "epoch": 1.3454622922880273, + "grad_norm": 12.375, + "learning_rate": 6.09205647595236e-07, + "loss": 0.4626, + "step": 15789 + }, + { + "epoch": 1.3455475074563272, + "grad_norm": 19.75, + "learning_rate": 6.090613769949735e-07, + "loss": 0.6636, + "step": 15790 + }, + { + "epoch": 1.3456327226246272, + "grad_norm": 18.75, + "learning_rate": 6.089171179771167e-07, + "loss": 0.4606, + "step": 15791 + }, + { + "epoch": 1.3457179377929271, + "grad_norm": 12.8125, + "learning_rate": 6.087728705442721e-07, + "loss": 0.4842, + "step": 15792 + }, + { + "epoch": 1.345803152961227, + "grad_norm": 16.625, + "learning_rate": 6.086286346990474e-07, + "loss": 0.4564, + "step": 15793 + }, + { + "epoch": 1.345888368129527, + "grad_norm": 18.5, + "learning_rate": 6.084844104440472e-07, + "loss": 0.9401, + "step": 15794 + }, + { + "epoch": 1.345973583297827, + "grad_norm": 13.6875, + "learning_rate": 6.083401977818792e-07, + "loss": 0.5254, + "step": 15795 + }, + { + "epoch": 1.346058798466127, + "grad_norm": 17.875, + "learning_rate": 6.081959967151487e-07, + "loss": 0.5924, + "step": 15796 + }, + { + "epoch": 1.346144013634427, + "grad_norm": 13.3125, + "learning_rate": 6.080518072464624e-07, + "loss": 0.5881, + "step": 15797 + }, + { + "epoch": 1.3462292288027269, + "grad_norm": 17.875, + "learning_rate": 6.079076293784252e-07, + "loss": 0.6528, + "step": 15798 + }, + { + "epoch": 1.3463144439710268, + "grad_norm": 12.9375, + "learning_rate": 6.077634631136427e-07, + "loss": 0.5699, + "step": 15799 + }, + { + "epoch": 1.3463996591393268, + "grad_norm": 15.375, + "learning_rate": 6.076193084547202e-07, + "loss": 0.4941, + "step": 15800 + }, + { + "epoch": 1.3464848743076268, + "grad_norm": 11.625, + "learning_rate": 6.074751654042632e-07, + "loss": 0.3549, + "step": 15801 + }, + { + "epoch": 1.3465700894759267, + "grad_norm": 25.25, + "learning_rate": 6.073310339648762e-07, + "loss": 0.7699, + "step": 15802 + }, + { + "epoch": 1.3466553046442267, + "grad_norm": 16.125, + "learning_rate": 6.071869141391631e-07, + "loss": 0.736, + "step": 15803 + }, + { + "epoch": 1.3467405198125266, + "grad_norm": 15.5625, + "learning_rate": 6.07042805929729e-07, + "loss": 0.5478, + "step": 15804 + }, + { + "epoch": 1.3468257349808266, + "grad_norm": 20.625, + "learning_rate": 6.068987093391778e-07, + "loss": 1.0594, + "step": 15805 + }, + { + "epoch": 1.3469109501491265, + "grad_norm": 28.5, + "learning_rate": 6.067546243701142e-07, + "loss": 0.7303, + "step": 15806 + }, + { + "epoch": 1.3469961653174265, + "grad_norm": 15.6875, + "learning_rate": 6.066105510251408e-07, + "loss": 0.6478, + "step": 15807 + }, + { + "epoch": 1.3470813804857265, + "grad_norm": 15.5625, + "learning_rate": 6.064664893068617e-07, + "loss": 0.5998, + "step": 15808 + }, + { + "epoch": 1.3471665956540264, + "grad_norm": 14.5625, + "learning_rate": 6.063224392178808e-07, + "loss": 0.8845, + "step": 15809 + }, + { + "epoch": 1.3472518108223264, + "grad_norm": 15.3125, + "learning_rate": 6.061784007608007e-07, + "loss": 0.4496, + "step": 15810 + }, + { + "epoch": 1.3473370259906263, + "grad_norm": 10.3125, + "learning_rate": 6.060343739382237e-07, + "loss": 0.3702, + "step": 15811 + }, + { + "epoch": 1.3474222411589263, + "grad_norm": 15.5625, + "learning_rate": 6.058903587527533e-07, + "loss": 0.4573, + "step": 15812 + }, + { + "epoch": 1.3475074563272262, + "grad_norm": 15.375, + "learning_rate": 6.057463552069917e-07, + "loss": 0.6626, + "step": 15813 + }, + { + "epoch": 1.3475926714955262, + "grad_norm": 12.9375, + "learning_rate": 6.056023633035418e-07, + "loss": 0.3749, + "step": 15814 + }, + { + "epoch": 1.3476778866638262, + "grad_norm": 28.0, + "learning_rate": 6.054583830450049e-07, + "loss": 0.9343, + "step": 15815 + }, + { + "epoch": 1.347763101832126, + "grad_norm": 11.625, + "learning_rate": 6.05314414433983e-07, + "loss": 0.3671, + "step": 15816 + }, + { + "epoch": 1.347848317000426, + "grad_norm": 17.625, + "learning_rate": 6.051704574730784e-07, + "loss": 0.4965, + "step": 15817 + }, + { + "epoch": 1.347933532168726, + "grad_norm": 18.0, + "learning_rate": 6.050265121648915e-07, + "loss": 0.6225, + "step": 15818 + }, + { + "epoch": 1.348018747337026, + "grad_norm": 14.25, + "learning_rate": 6.048825785120247e-07, + "loss": 0.3876, + "step": 15819 + }, + { + "epoch": 1.348103962505326, + "grad_norm": 16.75, + "learning_rate": 6.047386565170778e-07, + "loss": 0.7574, + "step": 15820 + }, + { + "epoch": 1.348189177673626, + "grad_norm": 17.0, + "learning_rate": 6.045947461826524e-07, + "loss": 0.6694, + "step": 15821 + }, + { + "epoch": 1.3482743928419259, + "grad_norm": 13.75, + "learning_rate": 6.044508475113488e-07, + "loss": 0.4245, + "step": 15822 + }, + { + "epoch": 1.3483596080102258, + "grad_norm": 17.125, + "learning_rate": 6.043069605057682e-07, + "loss": 0.7726, + "step": 15823 + }, + { + "epoch": 1.3484448231785258, + "grad_norm": 16.625, + "learning_rate": 6.041630851685095e-07, + "loss": 0.7617, + "step": 15824 + }, + { + "epoch": 1.3485300383468257, + "grad_norm": 16.125, + "learning_rate": 6.040192215021737e-07, + "loss": 0.4532, + "step": 15825 + }, + { + "epoch": 1.3486152535151257, + "grad_norm": 12.5, + "learning_rate": 6.038753695093598e-07, + "loss": 0.5056, + "step": 15826 + }, + { + "epoch": 1.3487004686834256, + "grad_norm": 13.625, + "learning_rate": 6.037315291926682e-07, + "loss": 0.4505, + "step": 15827 + }, + { + "epoch": 1.3487856838517256, + "grad_norm": 16.25, + "learning_rate": 6.035877005546971e-07, + "loss": 1.1094, + "step": 15828 + }, + { + "epoch": 1.3488708990200255, + "grad_norm": 14.125, + "learning_rate": 6.034438835980465e-07, + "loss": 0.6765, + "step": 15829 + }, + { + "epoch": 1.3489561141883255, + "grad_norm": 11.8125, + "learning_rate": 6.033000783253148e-07, + "loss": 0.3139, + "step": 15830 + }, + { + "epoch": 1.3490413293566255, + "grad_norm": 26.625, + "learning_rate": 6.031562847391017e-07, + "loss": 0.8411, + "step": 15831 + }, + { + "epoch": 1.3491265445249254, + "grad_norm": 27.875, + "learning_rate": 6.030125028420045e-07, + "loss": 1.1612, + "step": 15832 + }, + { + "epoch": 1.3492117596932254, + "grad_norm": 13.875, + "learning_rate": 6.028687326366224e-07, + "loss": 0.6284, + "step": 15833 + }, + { + "epoch": 1.3492969748615253, + "grad_norm": 21.625, + "learning_rate": 6.027249741255525e-07, + "loss": 0.8207, + "step": 15834 + }, + { + "epoch": 1.3493821900298253, + "grad_norm": 15.8125, + "learning_rate": 6.025812273113933e-07, + "loss": 0.8499, + "step": 15835 + }, + { + "epoch": 1.3494674051981252, + "grad_norm": 21.25, + "learning_rate": 6.024374921967427e-07, + "loss": 0.7364, + "step": 15836 + }, + { + "epoch": 1.3495526203664252, + "grad_norm": 16.375, + "learning_rate": 6.022937687841975e-07, + "loss": 0.6, + "step": 15837 + }, + { + "epoch": 1.3496378355347252, + "grad_norm": 17.375, + "learning_rate": 6.02150057076355e-07, + "loss": 0.6208, + "step": 15838 + }, + { + "epoch": 1.3497230507030251, + "grad_norm": 18.5, + "learning_rate": 6.020063570758126e-07, + "loss": 0.3062, + "step": 15839 + }, + { + "epoch": 1.349808265871325, + "grad_norm": 19.875, + "learning_rate": 6.018626687851674e-07, + "loss": 0.8002, + "step": 15840 + }, + { + "epoch": 1.349893481039625, + "grad_norm": 16.25, + "learning_rate": 6.017189922070149e-07, + "loss": 0.6893, + "step": 15841 + }, + { + "epoch": 1.349978696207925, + "grad_norm": 19.375, + "learning_rate": 6.015753273439527e-07, + "loss": 0.7421, + "step": 15842 + }, + { + "epoch": 1.350063911376225, + "grad_norm": 14.9375, + "learning_rate": 6.01431674198576e-07, + "loss": 0.5787, + "step": 15843 + }, + { + "epoch": 1.350149126544525, + "grad_norm": 18.25, + "learning_rate": 6.012880327734814e-07, + "loss": 0.7974, + "step": 15844 + }, + { + "epoch": 1.3502343417128249, + "grad_norm": 17.375, + "learning_rate": 6.01144403071264e-07, + "loss": 1.1023, + "step": 15845 + }, + { + "epoch": 1.3503195568811248, + "grad_norm": 18.0, + "learning_rate": 6.010007850945197e-07, + "loss": 0.6663, + "step": 15846 + }, + { + "epoch": 1.3504047720494248, + "grad_norm": 14.125, + "learning_rate": 6.00857178845844e-07, + "loss": 0.5281, + "step": 15847 + }, + { + "epoch": 1.3504899872177247, + "grad_norm": 17.625, + "learning_rate": 6.007135843278321e-07, + "loss": 0.785, + "step": 15848 + }, + { + "epoch": 1.3505752023860247, + "grad_norm": 19.875, + "learning_rate": 6.005700015430782e-07, + "loss": 0.9119, + "step": 15849 + }, + { + "epoch": 1.3506604175543246, + "grad_norm": 16.25, + "learning_rate": 6.00426430494178e-07, + "loss": 0.712, + "step": 15850 + }, + { + "epoch": 1.3507456327226246, + "grad_norm": 17.125, + "learning_rate": 6.002828711837248e-07, + "loss": 0.8806, + "step": 15851 + }, + { + "epoch": 1.3508308478909246, + "grad_norm": 12.8125, + "learning_rate": 6.001393236143135e-07, + "loss": 0.3643, + "step": 15852 + }, + { + "epoch": 1.3509160630592245, + "grad_norm": 12.9375, + "learning_rate": 5.999957877885386e-07, + "loss": 0.2465, + "step": 15853 + }, + { + "epoch": 1.3510012782275245, + "grad_norm": 16.625, + "learning_rate": 5.998522637089931e-07, + "loss": 0.5329, + "step": 15854 + }, + { + "epoch": 1.3510864933958244, + "grad_norm": 14.3125, + "learning_rate": 5.997087513782707e-07, + "loss": 0.6052, + "step": 15855 + }, + { + "epoch": 1.3511717085641244, + "grad_norm": 14.25, + "learning_rate": 5.995652507989657e-07, + "loss": 0.4683, + "step": 15856 + }, + { + "epoch": 1.3512569237324243, + "grad_norm": 14.8125, + "learning_rate": 5.994217619736702e-07, + "loss": 0.2995, + "step": 15857 + }, + { + "epoch": 1.3513421389007243, + "grad_norm": 18.625, + "learning_rate": 5.992782849049782e-07, + "loss": 0.9606, + "step": 15858 + }, + { + "epoch": 1.3514273540690243, + "grad_norm": 19.25, + "learning_rate": 5.991348195954813e-07, + "loss": 0.7654, + "step": 15859 + }, + { + "epoch": 1.3515125692373242, + "grad_norm": 15.3125, + "learning_rate": 5.989913660477728e-07, + "loss": 0.7139, + "step": 15860 + }, + { + "epoch": 1.3515977844056242, + "grad_norm": 15.4375, + "learning_rate": 5.988479242644455e-07, + "loss": 0.6454, + "step": 15861 + }, + { + "epoch": 1.3516829995739241, + "grad_norm": 13.6875, + "learning_rate": 5.987044942480904e-07, + "loss": 0.4046, + "step": 15862 + }, + { + "epoch": 1.351768214742224, + "grad_norm": 15.0625, + "learning_rate": 5.985610760013001e-07, + "loss": 0.7737, + "step": 15863 + }, + { + "epoch": 1.351853429910524, + "grad_norm": 12.0625, + "learning_rate": 5.984176695266662e-07, + "loss": 0.3874, + "step": 15864 + }, + { + "epoch": 1.351938645078824, + "grad_norm": 13.5625, + "learning_rate": 5.982742748267807e-07, + "loss": 0.3902, + "step": 15865 + }, + { + "epoch": 1.352023860247124, + "grad_norm": 13.6875, + "learning_rate": 5.981308919042344e-07, + "loss": 0.5255, + "step": 15866 + }, + { + "epoch": 1.352109075415424, + "grad_norm": 18.75, + "learning_rate": 5.97987520761618e-07, + "loss": 0.8196, + "step": 15867 + }, + { + "epoch": 1.3521942905837239, + "grad_norm": 11.875, + "learning_rate": 5.978441614015228e-07, + "loss": 0.3243, + "step": 15868 + }, + { + "epoch": 1.3522795057520238, + "grad_norm": 15.125, + "learning_rate": 5.977008138265392e-07, + "loss": 0.537, + "step": 15869 + }, + { + "epoch": 1.3523647209203238, + "grad_norm": 17.75, + "learning_rate": 5.975574780392583e-07, + "loss": 0.832, + "step": 15870 + }, + { + "epoch": 1.3524499360886237, + "grad_norm": 17.75, + "learning_rate": 5.974141540422696e-07, + "loss": 0.8173, + "step": 15871 + }, + { + "epoch": 1.3525351512569237, + "grad_norm": 11.8125, + "learning_rate": 5.972708418381631e-07, + "loss": 0.3605, + "step": 15872 + }, + { + "epoch": 1.3526203664252237, + "grad_norm": 13.6875, + "learning_rate": 5.971275414295295e-07, + "loss": 0.5674, + "step": 15873 + }, + { + "epoch": 1.3527055815935236, + "grad_norm": 26.75, + "learning_rate": 5.969842528189577e-07, + "loss": 0.6826, + "step": 15874 + }, + { + "epoch": 1.3527907967618236, + "grad_norm": 14.6875, + "learning_rate": 5.968409760090367e-07, + "loss": 0.583, + "step": 15875 + }, + { + "epoch": 1.3528760119301235, + "grad_norm": 17.125, + "learning_rate": 5.96697711002356e-07, + "loss": 0.7757, + "step": 15876 + }, + { + "epoch": 1.3529612270984235, + "grad_norm": 13.25, + "learning_rate": 5.965544578015046e-07, + "loss": 0.5738, + "step": 15877 + }, + { + "epoch": 1.3530464422667234, + "grad_norm": 13.3125, + "learning_rate": 5.964112164090718e-07, + "loss": 0.6923, + "step": 15878 + }, + { + "epoch": 1.3531316574350234, + "grad_norm": 15.5, + "learning_rate": 5.962679868276449e-07, + "loss": 0.7008, + "step": 15879 + }, + { + "epoch": 1.3532168726033234, + "grad_norm": 11.9375, + "learning_rate": 5.96124769059813e-07, + "loss": 0.334, + "step": 15880 + }, + { + "epoch": 1.3533020877716233, + "grad_norm": 13.1875, + "learning_rate": 5.95981563108164e-07, + "loss": 0.4529, + "step": 15881 + }, + { + "epoch": 1.3533873029399233, + "grad_norm": 15.875, + "learning_rate": 5.958383689752863e-07, + "loss": 0.7905, + "step": 15882 + }, + { + "epoch": 1.3534725181082232, + "grad_norm": 22.75, + "learning_rate": 5.956951866637671e-07, + "loss": 0.7251, + "step": 15883 + }, + { + "epoch": 1.3535577332765232, + "grad_norm": 16.25, + "learning_rate": 5.955520161761934e-07, + "loss": 0.4628, + "step": 15884 + }, + { + "epoch": 1.3536429484448231, + "grad_norm": 17.125, + "learning_rate": 5.954088575151527e-07, + "loss": 0.9171, + "step": 15885 + }, + { + "epoch": 1.353728163613123, + "grad_norm": 20.125, + "learning_rate": 5.952657106832328e-07, + "loss": 0.8455, + "step": 15886 + }, + { + "epoch": 1.353813378781423, + "grad_norm": 14.6875, + "learning_rate": 5.951225756830194e-07, + "loss": 0.7517, + "step": 15887 + }, + { + "epoch": 1.353898593949723, + "grad_norm": 12.625, + "learning_rate": 5.949794525170996e-07, + "loss": 0.5419, + "step": 15888 + }, + { + "epoch": 1.353983809118023, + "grad_norm": 10.8125, + "learning_rate": 5.948363411880598e-07, + "loss": 0.2975, + "step": 15889 + }, + { + "epoch": 1.354069024286323, + "grad_norm": 9.875, + "learning_rate": 5.946932416984864e-07, + "loss": 0.2283, + "step": 15890 + }, + { + "epoch": 1.3541542394546229, + "grad_norm": 13.5, + "learning_rate": 5.945501540509654e-07, + "loss": 0.3838, + "step": 15891 + }, + { + "epoch": 1.3542394546229228, + "grad_norm": 18.75, + "learning_rate": 5.944070782480815e-07, + "loss": 0.8425, + "step": 15892 + }, + { + "epoch": 1.3543246697912228, + "grad_norm": 17.125, + "learning_rate": 5.94264014292421e-07, + "loss": 0.8381, + "step": 15893 + }, + { + "epoch": 1.3544098849595227, + "grad_norm": 14.4375, + "learning_rate": 5.941209621865691e-07, + "loss": 0.4299, + "step": 15894 + }, + { + "epoch": 1.3544951001278227, + "grad_norm": 17.625, + "learning_rate": 5.939779219331115e-07, + "loss": 0.731, + "step": 15895 + }, + { + "epoch": 1.3545803152961227, + "grad_norm": 19.375, + "learning_rate": 5.938348935346322e-07, + "loss": 0.8002, + "step": 15896 + }, + { + "epoch": 1.3546655304644226, + "grad_norm": 17.875, + "learning_rate": 5.93691876993716e-07, + "loss": 0.7089, + "step": 15897 + }, + { + "epoch": 1.3547507456327226, + "grad_norm": 15.5625, + "learning_rate": 5.935488723129481e-07, + "loss": 0.4617, + "step": 15898 + }, + { + "epoch": 1.3548359608010225, + "grad_norm": 16.0, + "learning_rate": 5.934058794949117e-07, + "loss": 0.5794, + "step": 15899 + }, + { + "epoch": 1.3549211759693225, + "grad_norm": 17.625, + "learning_rate": 5.932628985421918e-07, + "loss": 0.6229, + "step": 15900 + }, + { + "epoch": 1.3550063911376224, + "grad_norm": 17.75, + "learning_rate": 5.931199294573713e-07, + "loss": 0.7742, + "step": 15901 + }, + { + "epoch": 1.3550916063059224, + "grad_norm": 23.0, + "learning_rate": 5.929769722430341e-07, + "loss": 0.9518, + "step": 15902 + }, + { + "epoch": 1.3551768214742224, + "grad_norm": 17.625, + "learning_rate": 5.928340269017642e-07, + "loss": 0.5512, + "step": 15903 + }, + { + "epoch": 1.3552620366425223, + "grad_norm": 16.25, + "learning_rate": 5.926910934361439e-07, + "loss": 0.6373, + "step": 15904 + }, + { + "epoch": 1.3553472518108223, + "grad_norm": 15.6875, + "learning_rate": 5.925481718487565e-07, + "loss": 0.9097, + "step": 15905 + }, + { + "epoch": 1.3554324669791222, + "grad_norm": 12.3125, + "learning_rate": 5.924052621421847e-07, + "loss": 0.2453, + "step": 15906 + }, + { + "epoch": 1.3555176821474222, + "grad_norm": 15.75, + "learning_rate": 5.922623643190114e-07, + "loss": 0.6213, + "step": 15907 + }, + { + "epoch": 1.3556028973157221, + "grad_norm": 15.125, + "learning_rate": 5.921194783818188e-07, + "loss": 0.5509, + "step": 15908 + }, + { + "epoch": 1.355688112484022, + "grad_norm": 13.5, + "learning_rate": 5.919766043331883e-07, + "loss": 0.28, + "step": 15909 + }, + { + "epoch": 1.355773327652322, + "grad_norm": 14.1875, + "learning_rate": 5.918337421757022e-07, + "loss": 0.4226, + "step": 15910 + }, + { + "epoch": 1.355858542820622, + "grad_norm": 19.0, + "learning_rate": 5.916908919119421e-07, + "loss": 0.5925, + "step": 15911 + }, + { + "epoch": 1.355943757988922, + "grad_norm": 22.375, + "learning_rate": 5.915480535444903e-07, + "loss": 1.2797, + "step": 15912 + }, + { + "epoch": 1.356028973157222, + "grad_norm": 15.1875, + "learning_rate": 5.914052270759265e-07, + "loss": 0.6694, + "step": 15913 + }, + { + "epoch": 1.3561141883255219, + "grad_norm": 14.875, + "learning_rate": 5.912624125088328e-07, + "loss": 0.8276, + "step": 15914 + }, + { + "epoch": 1.3561994034938218, + "grad_norm": 28.125, + "learning_rate": 5.9111960984579e-07, + "loss": 0.8987, + "step": 15915 + }, + { + "epoch": 1.3562846186621218, + "grad_norm": 15.375, + "learning_rate": 5.909768190893784e-07, + "loss": 0.5611, + "step": 15916 + }, + { + "epoch": 1.3563698338304218, + "grad_norm": 20.375, + "learning_rate": 5.908340402421781e-07, + "loss": 0.2776, + "step": 15917 + }, + { + "epoch": 1.3564550489987217, + "grad_norm": 12.25, + "learning_rate": 5.906912733067693e-07, + "loss": 0.5888, + "step": 15918 + }, + { + "epoch": 1.3565402641670217, + "grad_norm": 15.0, + "learning_rate": 5.905485182857323e-07, + "loss": 0.6105, + "step": 15919 + }, + { + "epoch": 1.3566254793353216, + "grad_norm": 13.5, + "learning_rate": 5.904057751816473e-07, + "loss": 0.6149, + "step": 15920 + }, + { + "epoch": 1.3567106945036216, + "grad_norm": 15.0, + "learning_rate": 5.902630439970926e-07, + "loss": 0.3949, + "step": 15921 + }, + { + "epoch": 1.3567959096719215, + "grad_norm": 24.875, + "learning_rate": 5.901203247346481e-07, + "loss": 1.1055, + "step": 15922 + }, + { + "epoch": 1.3568811248402215, + "grad_norm": 16.125, + "learning_rate": 5.899776173968934e-07, + "loss": 0.8457, + "step": 15923 + }, + { + "epoch": 1.3569663400085215, + "grad_norm": 14.5, + "learning_rate": 5.898349219864064e-07, + "loss": 0.6871, + "step": 15924 + }, + { + "epoch": 1.3570515551768214, + "grad_norm": 14.75, + "learning_rate": 5.896922385057666e-07, + "loss": 0.6489, + "step": 15925 + }, + { + "epoch": 1.3571367703451214, + "grad_norm": 11.9375, + "learning_rate": 5.895495669575515e-07, + "loss": 0.2933, + "step": 15926 + }, + { + "epoch": 1.3572219855134213, + "grad_norm": 15.3125, + "learning_rate": 5.894069073443401e-07, + "loss": 0.5319, + "step": 15927 + }, + { + "epoch": 1.3573072006817213, + "grad_norm": 11.5, + "learning_rate": 5.892642596687101e-07, + "loss": 0.3115, + "step": 15928 + }, + { + "epoch": 1.3573924158500212, + "grad_norm": 14.8125, + "learning_rate": 5.891216239332397e-07, + "loss": 0.6167, + "step": 15929 + }, + { + "epoch": 1.3574776310183212, + "grad_norm": 14.875, + "learning_rate": 5.889790001405058e-07, + "loss": 0.481, + "step": 15930 + }, + { + "epoch": 1.3575628461866212, + "grad_norm": 15.875, + "learning_rate": 5.888363882930865e-07, + "loss": 0.8069, + "step": 15931 + }, + { + "epoch": 1.357648061354921, + "grad_norm": 16.375, + "learning_rate": 5.886937883935579e-07, + "loss": 0.8347, + "step": 15932 + }, + { + "epoch": 1.357733276523221, + "grad_norm": 15.0, + "learning_rate": 5.885512004444981e-07, + "loss": 0.6962, + "step": 15933 + }, + { + "epoch": 1.357818491691521, + "grad_norm": 15.75, + "learning_rate": 5.884086244484828e-07, + "loss": 0.5843, + "step": 15934 + }, + { + "epoch": 1.357903706859821, + "grad_norm": 12.25, + "learning_rate": 5.882660604080889e-07, + "loss": 0.4408, + "step": 15935 + }, + { + "epoch": 1.357988922028121, + "grad_norm": 18.625, + "learning_rate": 5.881235083258928e-07, + "loss": 0.4395, + "step": 15936 + }, + { + "epoch": 1.358074137196421, + "grad_norm": 12.75, + "learning_rate": 5.879809682044708e-07, + "loss": 0.7589, + "step": 15937 + }, + { + "epoch": 1.3581593523647209, + "grad_norm": 18.125, + "learning_rate": 5.878384400463981e-07, + "loss": 0.5083, + "step": 15938 + }, + { + "epoch": 1.3582445675330208, + "grad_norm": 14.3125, + "learning_rate": 5.87695923854251e-07, + "loss": 0.5216, + "step": 15939 + }, + { + "epoch": 1.3583297827013208, + "grad_norm": 15.25, + "learning_rate": 5.875534196306042e-07, + "loss": 0.4046, + "step": 15940 + }, + { + "epoch": 1.3584149978696207, + "grad_norm": 16.125, + "learning_rate": 5.87410927378033e-07, + "loss": 0.6059, + "step": 15941 + }, + { + "epoch": 1.3585002130379207, + "grad_norm": 15.375, + "learning_rate": 5.872684470991134e-07, + "loss": 0.4912, + "step": 15942 + }, + { + "epoch": 1.3585854282062206, + "grad_norm": 13.0625, + "learning_rate": 5.871259787964189e-07, + "loss": 0.5477, + "step": 15943 + }, + { + "epoch": 1.3586706433745206, + "grad_norm": 12.6875, + "learning_rate": 5.869835224725243e-07, + "loss": 0.2879, + "step": 15944 + }, + { + "epoch": 1.3587558585428205, + "grad_norm": 11.6875, + "learning_rate": 5.868410781300044e-07, + "loss": 0.2783, + "step": 15945 + }, + { + "epoch": 1.3588410737111205, + "grad_norm": 17.75, + "learning_rate": 5.866986457714332e-07, + "loss": 0.5897, + "step": 15946 + }, + { + "epoch": 1.3589262888794205, + "grad_norm": 15.75, + "learning_rate": 5.865562253993848e-07, + "loss": 0.618, + "step": 15947 + }, + { + "epoch": 1.3590115040477204, + "grad_norm": 23.0, + "learning_rate": 5.86413817016432e-07, + "loss": 0.6742, + "step": 15948 + }, + { + "epoch": 1.3590967192160204, + "grad_norm": 14.125, + "learning_rate": 5.862714206251488e-07, + "loss": 0.5644, + "step": 15949 + }, + { + "epoch": 1.3591819343843203, + "grad_norm": 30.125, + "learning_rate": 5.86129036228109e-07, + "loss": 0.6692, + "step": 15950 + }, + { + "epoch": 1.3592671495526203, + "grad_norm": 26.0, + "learning_rate": 5.859866638278845e-07, + "loss": 0.3863, + "step": 15951 + }, + { + "epoch": 1.3593523647209202, + "grad_norm": 11.0, + "learning_rate": 5.858443034270488e-07, + "loss": 0.3295, + "step": 15952 + }, + { + "epoch": 1.3594375798892202, + "grad_norm": 15.125, + "learning_rate": 5.857019550281744e-07, + "loss": 0.4746, + "step": 15953 + }, + { + "epoch": 1.3595227950575202, + "grad_norm": 17.125, + "learning_rate": 5.85559618633834e-07, + "loss": 0.4506, + "step": 15954 + }, + { + "epoch": 1.3596080102258201, + "grad_norm": 14.0625, + "learning_rate": 5.854172942465995e-07, + "loss": 0.3714, + "step": 15955 + }, + { + "epoch": 1.35969322539412, + "grad_norm": 15.75, + "learning_rate": 5.852749818690425e-07, + "loss": 0.6052, + "step": 15956 + }, + { + "epoch": 1.35977844056242, + "grad_norm": 14.5, + "learning_rate": 5.851326815037347e-07, + "loss": 0.69, + "step": 15957 + }, + { + "epoch": 1.35986365573072, + "grad_norm": 18.625, + "learning_rate": 5.849903931532482e-07, + "loss": 0.5228, + "step": 15958 + }, + { + "epoch": 1.35994887089902, + "grad_norm": 14.6875, + "learning_rate": 5.848481168201543e-07, + "loss": 0.424, + "step": 15959 + }, + { + "epoch": 1.36003408606732, + "grad_norm": 13.75, + "learning_rate": 5.847058525070235e-07, + "loss": 0.6474, + "step": 15960 + }, + { + "epoch": 1.3601193012356199, + "grad_norm": 16.75, + "learning_rate": 5.845636002164267e-07, + "loss": 0.6883, + "step": 15961 + }, + { + "epoch": 1.3602045164039198, + "grad_norm": 15.5625, + "learning_rate": 5.844213599509349e-07, + "loss": 0.5883, + "step": 15962 + }, + { + "epoch": 1.3602897315722198, + "grad_norm": 18.125, + "learning_rate": 5.842791317131194e-07, + "loss": 0.5319, + "step": 15963 + }, + { + "epoch": 1.3603749467405197, + "grad_norm": 18.375, + "learning_rate": 5.841369155055485e-07, + "loss": 0.7101, + "step": 15964 + }, + { + "epoch": 1.3604601619088197, + "grad_norm": 17.125, + "learning_rate": 5.839947113307931e-07, + "loss": 0.6981, + "step": 15965 + }, + { + "epoch": 1.3605453770771196, + "grad_norm": 12.8125, + "learning_rate": 5.838525191914229e-07, + "loss": 0.4287, + "step": 15966 + }, + { + "epoch": 1.3606305922454198, + "grad_norm": 21.75, + "learning_rate": 5.837103390900081e-07, + "loss": 0.6906, + "step": 15967 + }, + { + "epoch": 1.3607158074137198, + "grad_norm": 17.0, + "learning_rate": 5.83568171029117e-07, + "loss": 0.7192, + "step": 15968 + }, + { + "epoch": 1.3608010225820197, + "grad_norm": 15.25, + "learning_rate": 5.834260150113192e-07, + "loss": 0.4449, + "step": 15969 + }, + { + "epoch": 1.3608862377503197, + "grad_norm": 13.3125, + "learning_rate": 5.832838710391836e-07, + "loss": 0.4807, + "step": 15970 + }, + { + "epoch": 1.3609714529186197, + "grad_norm": 17.25, + "learning_rate": 5.831417391152794e-07, + "loss": 0.7733, + "step": 15971 + }, + { + "epoch": 1.3610566680869196, + "grad_norm": 12.375, + "learning_rate": 5.829996192421744e-07, + "loss": 0.4813, + "step": 15972 + }, + { + "epoch": 1.3611418832552196, + "grad_norm": 15.125, + "learning_rate": 5.828575114224366e-07, + "loss": 0.8302, + "step": 15973 + }, + { + "epoch": 1.3612270984235195, + "grad_norm": 13.3125, + "learning_rate": 5.827154156586346e-07, + "loss": 0.5226, + "step": 15974 + }, + { + "epoch": 1.3613123135918195, + "grad_norm": 14.875, + "learning_rate": 5.825733319533359e-07, + "loss": 0.6715, + "step": 15975 + }, + { + "epoch": 1.3613975287601194, + "grad_norm": 19.25, + "learning_rate": 5.824312603091087e-07, + "loss": 0.779, + "step": 15976 + }, + { + "epoch": 1.3614827439284194, + "grad_norm": 23.625, + "learning_rate": 5.822892007285194e-07, + "loss": 0.6616, + "step": 15977 + }, + { + "epoch": 1.3615679590967193, + "grad_norm": 11.5625, + "learning_rate": 5.821471532141357e-07, + "loss": 0.2809, + "step": 15978 + }, + { + "epoch": 1.3616531742650193, + "grad_norm": 10.125, + "learning_rate": 5.820051177685246e-07, + "loss": 0.3028, + "step": 15979 + }, + { + "epoch": 1.3617383894333193, + "grad_norm": 19.375, + "learning_rate": 5.818630943942537e-07, + "loss": 0.678, + "step": 15980 + }, + { + "epoch": 1.3618236046016192, + "grad_norm": 17.875, + "learning_rate": 5.817210830938876e-07, + "loss": 0.5908, + "step": 15981 + }, + { + "epoch": 1.3619088197699192, + "grad_norm": 14.5625, + "learning_rate": 5.815790838699934e-07, + "loss": 0.5574, + "step": 15982 + }, + { + "epoch": 1.3619940349382191, + "grad_norm": 15.0625, + "learning_rate": 5.814370967251373e-07, + "loss": 0.6033, + "step": 15983 + }, + { + "epoch": 1.362079250106519, + "grad_norm": 17.875, + "learning_rate": 5.812951216618858e-07, + "loss": 0.7202, + "step": 15984 + }, + { + "epoch": 1.362164465274819, + "grad_norm": 20.375, + "learning_rate": 5.811531586828033e-07, + "loss": 0.9735, + "step": 15985 + }, + { + "epoch": 1.362249680443119, + "grad_norm": 26.375, + "learning_rate": 5.810112077904559e-07, + "loss": 0.5277, + "step": 15986 + }, + { + "epoch": 1.362334895611419, + "grad_norm": 19.875, + "learning_rate": 5.808692689874086e-07, + "loss": 1.0465, + "step": 15987 + }, + { + "epoch": 1.362420110779719, + "grad_norm": 14.8125, + "learning_rate": 5.80727342276227e-07, + "loss": 0.6369, + "step": 15988 + }, + { + "epoch": 1.3625053259480189, + "grad_norm": 19.375, + "learning_rate": 5.805854276594753e-07, + "loss": 0.8363, + "step": 15989 + }, + { + "epoch": 1.3625905411163188, + "grad_norm": 14.8125, + "learning_rate": 5.804435251397176e-07, + "loss": 0.6467, + "step": 15990 + }, + { + "epoch": 1.3626757562846188, + "grad_norm": 15.25, + "learning_rate": 5.803016347195187e-07, + "loss": 0.8417, + "step": 15991 + }, + { + "epoch": 1.3627609714529187, + "grad_norm": 11.1875, + "learning_rate": 5.801597564014426e-07, + "loss": 0.315, + "step": 15992 + }, + { + "epoch": 1.3628461866212187, + "grad_norm": 13.875, + "learning_rate": 5.800178901880539e-07, + "loss": 0.3547, + "step": 15993 + }, + { + "epoch": 1.3629314017895187, + "grad_norm": 11.1875, + "learning_rate": 5.79876036081915e-07, + "loss": 0.9094, + "step": 15994 + }, + { + "epoch": 1.3630166169578186, + "grad_norm": 14.625, + "learning_rate": 5.797341940855901e-07, + "loss": 0.4358, + "step": 15995 + }, + { + "epoch": 1.3631018321261186, + "grad_norm": 14.4375, + "learning_rate": 5.795923642016428e-07, + "loss": 0.9059, + "step": 15996 + }, + { + "epoch": 1.3631870472944185, + "grad_norm": 16.5, + "learning_rate": 5.794505464326357e-07, + "loss": 0.7799, + "step": 15997 + }, + { + "epoch": 1.3632722624627185, + "grad_norm": 13.375, + "learning_rate": 5.793087407811309e-07, + "loss": 0.4759, + "step": 15998 + }, + { + "epoch": 1.3633574776310184, + "grad_norm": 13.25, + "learning_rate": 5.791669472496916e-07, + "loss": 0.5153, + "step": 15999 + }, + { + "epoch": 1.3634426927993184, + "grad_norm": 19.5, + "learning_rate": 5.790251658408802e-07, + "loss": 0.7079, + "step": 16000 + }, + { + "epoch": 1.3635279079676184, + "grad_norm": 16.5, + "learning_rate": 5.788833965572592e-07, + "loss": 0.4141, + "step": 16001 + }, + { + "epoch": 1.3636131231359183, + "grad_norm": 16.25, + "learning_rate": 5.787416394013897e-07, + "loss": 0.5057, + "step": 16002 + }, + { + "epoch": 1.3636983383042183, + "grad_norm": 11.625, + "learning_rate": 5.785998943758336e-07, + "loss": 0.344, + "step": 16003 + }, + { + "epoch": 1.3637835534725182, + "grad_norm": 16.625, + "learning_rate": 5.784581614831532e-07, + "loss": 0.5798, + "step": 16004 + }, + { + "epoch": 1.3638687686408182, + "grad_norm": 18.625, + "learning_rate": 5.783164407259086e-07, + "loss": 0.5144, + "step": 16005 + }, + { + "epoch": 1.3639539838091181, + "grad_norm": 15.1875, + "learning_rate": 5.781747321066617e-07, + "loss": 0.6022, + "step": 16006 + }, + { + "epoch": 1.364039198977418, + "grad_norm": 13.875, + "learning_rate": 5.780330356279728e-07, + "loss": 0.5603, + "step": 16007 + }, + { + "epoch": 1.364124414145718, + "grad_norm": 16.125, + "learning_rate": 5.778913512924024e-07, + "loss": 0.5789, + "step": 16008 + }, + { + "epoch": 1.364209629314018, + "grad_norm": 16.0, + "learning_rate": 5.777496791025111e-07, + "loss": 0.6167, + "step": 16009 + }, + { + "epoch": 1.364294844482318, + "grad_norm": 13.125, + "learning_rate": 5.776080190608596e-07, + "loss": 0.2654, + "step": 16010 + }, + { + "epoch": 1.364380059650618, + "grad_norm": 18.375, + "learning_rate": 5.77466371170007e-07, + "loss": 0.582, + "step": 16011 + }, + { + "epoch": 1.3644652748189179, + "grad_norm": 19.125, + "learning_rate": 5.773247354325137e-07, + "loss": 0.7123, + "step": 16012 + }, + { + "epoch": 1.3645504899872178, + "grad_norm": 17.25, + "learning_rate": 5.771831118509384e-07, + "loss": 0.4806, + "step": 16013 + }, + { + "epoch": 1.3646357051555178, + "grad_norm": 13.375, + "learning_rate": 5.770415004278413e-07, + "loss": 0.6413, + "step": 16014 + }, + { + "epoch": 1.3647209203238178, + "grad_norm": 15.875, + "learning_rate": 5.768999011657806e-07, + "loss": 0.4478, + "step": 16015 + }, + { + "epoch": 1.3648061354921177, + "grad_norm": 19.5, + "learning_rate": 5.767583140673156e-07, + "loss": 1.0446, + "step": 16016 + }, + { + "epoch": 1.3648913506604177, + "grad_norm": 16.625, + "learning_rate": 5.766167391350048e-07, + "loss": 0.6855, + "step": 16017 + }, + { + "epoch": 1.3649765658287176, + "grad_norm": 16.75, + "learning_rate": 5.764751763714072e-07, + "loss": 0.6523, + "step": 16018 + }, + { + "epoch": 1.3650617809970176, + "grad_norm": 16.875, + "learning_rate": 5.763336257790798e-07, + "loss": 0.7014, + "step": 16019 + }, + { + "epoch": 1.3651469961653175, + "grad_norm": 12.25, + "learning_rate": 5.761920873605818e-07, + "loss": 0.469, + "step": 16020 + }, + { + "epoch": 1.3652322113336175, + "grad_norm": 19.0, + "learning_rate": 5.760505611184698e-07, + "loss": 0.9983, + "step": 16021 + }, + { + "epoch": 1.3653174265019175, + "grad_norm": 12.5625, + "learning_rate": 5.759090470553019e-07, + "loss": 0.4177, + "step": 16022 + }, + { + "epoch": 1.3654026416702174, + "grad_norm": 15.0, + "learning_rate": 5.757675451736358e-07, + "loss": 0.6135, + "step": 16023 + }, + { + "epoch": 1.3654878568385174, + "grad_norm": 16.625, + "learning_rate": 5.756260554760278e-07, + "loss": 0.6252, + "step": 16024 + }, + { + "epoch": 1.3655730720068173, + "grad_norm": 17.5, + "learning_rate": 5.75484577965035e-07, + "loss": 0.8448, + "step": 16025 + }, + { + "epoch": 1.3656582871751173, + "grad_norm": 16.25, + "learning_rate": 5.753431126432147e-07, + "loss": 0.6765, + "step": 16026 + }, + { + "epoch": 1.3657435023434172, + "grad_norm": 21.875, + "learning_rate": 5.752016595131223e-07, + "loss": 0.6313, + "step": 16027 + }, + { + "epoch": 1.3658287175117172, + "grad_norm": 16.5, + "learning_rate": 5.750602185773149e-07, + "loss": 0.5846, + "step": 16028 + }, + { + "epoch": 1.3659139326800172, + "grad_norm": 22.25, + "learning_rate": 5.749187898383476e-07, + "loss": 0.7186, + "step": 16029 + }, + { + "epoch": 1.365999147848317, + "grad_norm": 18.25, + "learning_rate": 5.747773732987765e-07, + "loss": 0.9485, + "step": 16030 + }, + { + "epoch": 1.366084363016617, + "grad_norm": 17.25, + "learning_rate": 5.74635968961158e-07, + "loss": 0.58, + "step": 16031 + }, + { + "epoch": 1.366169578184917, + "grad_norm": 17.25, + "learning_rate": 5.74494576828046e-07, + "loss": 0.6079, + "step": 16032 + }, + { + "epoch": 1.366254793353217, + "grad_norm": 15.5, + "learning_rate": 5.743531969019964e-07, + "loss": 0.6933, + "step": 16033 + }, + { + "epoch": 1.366340008521517, + "grad_norm": 14.9375, + "learning_rate": 5.742118291855637e-07, + "loss": 0.7072, + "step": 16034 + }, + { + "epoch": 1.366425223689817, + "grad_norm": 18.25, + "learning_rate": 5.740704736813035e-07, + "loss": 0.6852, + "step": 16035 + }, + { + "epoch": 1.3665104388581168, + "grad_norm": 15.5, + "learning_rate": 5.739291303917695e-07, + "loss": 0.5727, + "step": 16036 + }, + { + "epoch": 1.3665956540264168, + "grad_norm": 17.875, + "learning_rate": 5.737877993195153e-07, + "loss": 1.0102, + "step": 16037 + }, + { + "epoch": 1.3666808691947168, + "grad_norm": 12.0625, + "learning_rate": 5.736464804670959e-07, + "loss": 0.3029, + "step": 16038 + }, + { + "epoch": 1.3667660843630167, + "grad_norm": 23.625, + "learning_rate": 5.735051738370646e-07, + "loss": 0.7543, + "step": 16039 + }, + { + "epoch": 1.3668512995313167, + "grad_norm": 10.0625, + "learning_rate": 5.733638794319754e-07, + "loss": 0.2312, + "step": 16040 + }, + { + "epoch": 1.3669365146996166, + "grad_norm": 13.4375, + "learning_rate": 5.732225972543809e-07, + "loss": 0.6068, + "step": 16041 + }, + { + "epoch": 1.3670217298679166, + "grad_norm": 18.5, + "learning_rate": 5.730813273068346e-07, + "loss": 0.8657, + "step": 16042 + }, + { + "epoch": 1.3671069450362165, + "grad_norm": 13.0, + "learning_rate": 5.729400695918899e-07, + "loss": 0.4295, + "step": 16043 + }, + { + "epoch": 1.3671921602045165, + "grad_norm": 12.375, + "learning_rate": 5.727988241120986e-07, + "loss": 0.4244, + "step": 16044 + }, + { + "epoch": 1.3672773753728165, + "grad_norm": 17.25, + "learning_rate": 5.726575908700139e-07, + "loss": 0.5043, + "step": 16045 + }, + { + "epoch": 1.3673625905411164, + "grad_norm": 21.75, + "learning_rate": 5.725163698681873e-07, + "loss": 0.7034, + "step": 16046 + }, + { + "epoch": 1.3674478057094164, + "grad_norm": 16.25, + "learning_rate": 5.72375161109171e-07, + "loss": 0.5909, + "step": 16047 + }, + { + "epoch": 1.3675330208777163, + "grad_norm": 11.25, + "learning_rate": 5.722339645955175e-07, + "loss": 0.5964, + "step": 16048 + }, + { + "epoch": 1.3676182360460163, + "grad_norm": 11.9375, + "learning_rate": 5.720927803297772e-07, + "loss": 0.3673, + "step": 16049 + }, + { + "epoch": 1.3677034512143162, + "grad_norm": 22.375, + "learning_rate": 5.719516083145023e-07, + "loss": 0.777, + "step": 16050 + }, + { + "epoch": 1.3677886663826162, + "grad_norm": 14.5625, + "learning_rate": 5.718104485522435e-07, + "loss": 0.3528, + "step": 16051 + }, + { + "epoch": 1.3678738815509162, + "grad_norm": 14.625, + "learning_rate": 5.716693010455524e-07, + "loss": 0.5524, + "step": 16052 + }, + { + "epoch": 1.3679590967192161, + "grad_norm": 22.75, + "learning_rate": 5.71528165796979e-07, + "loss": 1.1691, + "step": 16053 + }, + { + "epoch": 1.368044311887516, + "grad_norm": 13.0, + "learning_rate": 5.713870428090735e-07, + "loss": 0.3076, + "step": 16054 + }, + { + "epoch": 1.368129527055816, + "grad_norm": 21.375, + "learning_rate": 5.712459320843867e-07, + "loss": 0.7869, + "step": 16055 + }, + { + "epoch": 1.368214742224116, + "grad_norm": 18.25, + "learning_rate": 5.711048336254688e-07, + "loss": 0.8944, + "step": 16056 + }, + { + "epoch": 1.368299957392416, + "grad_norm": 14.3125, + "learning_rate": 5.709637474348688e-07, + "loss": 0.4312, + "step": 16057 + }, + { + "epoch": 1.368385172560716, + "grad_norm": 13.3125, + "learning_rate": 5.708226735151366e-07, + "loss": 0.3191, + "step": 16058 + }, + { + "epoch": 1.3684703877290159, + "grad_norm": 18.5, + "learning_rate": 5.706816118688217e-07, + "loss": 0.7212, + "step": 16059 + }, + { + "epoch": 1.3685556028973158, + "grad_norm": 32.75, + "learning_rate": 5.705405624984738e-07, + "loss": 0.6661, + "step": 16060 + }, + { + "epoch": 1.3686408180656158, + "grad_norm": 19.0, + "learning_rate": 5.703995254066409e-07, + "loss": 0.6797, + "step": 16061 + }, + { + "epoch": 1.3687260332339157, + "grad_norm": 33.5, + "learning_rate": 5.702585005958718e-07, + "loss": 1.162, + "step": 16062 + }, + { + "epoch": 1.3688112484022157, + "grad_norm": 13.625, + "learning_rate": 5.701174880687149e-07, + "loss": 0.3124, + "step": 16063 + }, + { + "epoch": 1.3688964635705156, + "grad_norm": 15.3125, + "learning_rate": 5.699764878277188e-07, + "loss": 0.7422, + "step": 16064 + }, + { + "epoch": 1.3689816787388156, + "grad_norm": 17.75, + "learning_rate": 5.698354998754319e-07, + "loss": 0.7739, + "step": 16065 + }, + { + "epoch": 1.3690668939071156, + "grad_norm": 11.75, + "learning_rate": 5.696945242144008e-07, + "loss": 0.4205, + "step": 16066 + }, + { + "epoch": 1.3691521090754155, + "grad_norm": 16.0, + "learning_rate": 5.695535608471739e-07, + "loss": 0.3947, + "step": 16067 + }, + { + "epoch": 1.3692373242437155, + "grad_norm": 9.625, + "learning_rate": 5.694126097762984e-07, + "loss": 0.4074, + "step": 16068 + }, + { + "epoch": 1.3693225394120154, + "grad_norm": 16.0, + "learning_rate": 5.692716710043218e-07, + "loss": 0.5328, + "step": 16069 + }, + { + "epoch": 1.3694077545803154, + "grad_norm": 18.25, + "learning_rate": 5.691307445337906e-07, + "loss": 0.8244, + "step": 16070 + }, + { + "epoch": 1.3694929697486153, + "grad_norm": 16.375, + "learning_rate": 5.689898303672512e-07, + "loss": 0.6688, + "step": 16071 + }, + { + "epoch": 1.3695781849169153, + "grad_norm": 19.625, + "learning_rate": 5.688489285072503e-07, + "loss": 1.0184, + "step": 16072 + }, + { + "epoch": 1.3696634000852153, + "grad_norm": 15.8125, + "learning_rate": 5.687080389563346e-07, + "loss": 0.6495, + "step": 16073 + }, + { + "epoch": 1.3697486152535152, + "grad_norm": 14.3125, + "learning_rate": 5.685671617170493e-07, + "loss": 0.4662, + "step": 16074 + }, + { + "epoch": 1.3698338304218152, + "grad_norm": 19.375, + "learning_rate": 5.684262967919407e-07, + "loss": 0.9025, + "step": 16075 + }, + { + "epoch": 1.3699190455901151, + "grad_norm": 19.25, + "learning_rate": 5.682854441835543e-07, + "loss": 0.9998, + "step": 16076 + }, + { + "epoch": 1.370004260758415, + "grad_norm": 15.0625, + "learning_rate": 5.681446038944359e-07, + "loss": 0.6481, + "step": 16077 + }, + { + "epoch": 1.370089475926715, + "grad_norm": 15.125, + "learning_rate": 5.680037759271301e-07, + "loss": 0.2297, + "step": 16078 + }, + { + "epoch": 1.370174691095015, + "grad_norm": 19.75, + "learning_rate": 5.678629602841815e-07, + "loss": 0.5858, + "step": 16079 + }, + { + "epoch": 1.370259906263315, + "grad_norm": 22.25, + "learning_rate": 5.67722156968135e-07, + "loss": 0.8521, + "step": 16080 + }, + { + "epoch": 1.370345121431615, + "grad_norm": 33.5, + "learning_rate": 5.675813659815354e-07, + "loss": 0.9786, + "step": 16081 + }, + { + "epoch": 1.3704303365999149, + "grad_norm": 14.5625, + "learning_rate": 5.67440587326927e-07, + "loss": 0.6121, + "step": 16082 + }, + { + "epoch": 1.3705155517682148, + "grad_norm": 16.75, + "learning_rate": 5.672998210068531e-07, + "loss": 0.6145, + "step": 16083 + }, + { + "epoch": 1.3706007669365148, + "grad_norm": 12.25, + "learning_rate": 5.671590670238581e-07, + "loss": 0.2772, + "step": 16084 + }, + { + "epoch": 1.3706859821048147, + "grad_norm": 14.6875, + "learning_rate": 5.670183253804857e-07, + "loss": 0.5794, + "step": 16085 + }, + { + "epoch": 1.3707711972731147, + "grad_norm": 17.375, + "learning_rate": 5.668775960792791e-07, + "loss": 0.7444, + "step": 16086 + }, + { + "epoch": 1.3708564124414147, + "grad_norm": 18.125, + "learning_rate": 5.667368791227808e-07, + "loss": 0.8679, + "step": 16087 + }, + { + "epoch": 1.3709416276097146, + "grad_norm": 15.0625, + "learning_rate": 5.665961745135342e-07, + "loss": 0.6098, + "step": 16088 + }, + { + "epoch": 1.3710268427780146, + "grad_norm": 16.375, + "learning_rate": 5.664554822540817e-07, + "loss": 0.6882, + "step": 16089 + }, + { + "epoch": 1.3711120579463145, + "grad_norm": 14.0, + "learning_rate": 5.663148023469667e-07, + "loss": 0.3834, + "step": 16090 + }, + { + "epoch": 1.3711972731146145, + "grad_norm": 18.125, + "learning_rate": 5.661741347947301e-07, + "loss": 0.7943, + "step": 16091 + }, + { + "epoch": 1.3712824882829144, + "grad_norm": 14.4375, + "learning_rate": 5.660334795999146e-07, + "loss": 0.8472, + "step": 16092 + }, + { + "epoch": 1.3713677034512144, + "grad_norm": 13.0, + "learning_rate": 5.658928367650625e-07, + "loss": 0.4627, + "step": 16093 + }, + { + "epoch": 1.3714529186195143, + "grad_norm": 14.75, + "learning_rate": 5.657522062927142e-07, + "loss": 0.5775, + "step": 16094 + }, + { + "epoch": 1.3715381337878143, + "grad_norm": 15.8125, + "learning_rate": 5.65611588185412e-07, + "loss": 0.8127, + "step": 16095 + }, + { + "epoch": 1.3716233489561143, + "grad_norm": 14.75, + "learning_rate": 5.65470982445696e-07, + "loss": 0.4273, + "step": 16096 + }, + { + "epoch": 1.3717085641244142, + "grad_norm": 18.25, + "learning_rate": 5.653303890761081e-07, + "loss": 0.7115, + "step": 16097 + }, + { + "epoch": 1.3717937792927142, + "grad_norm": 17.25, + "learning_rate": 5.651898080791881e-07, + "loss": 0.4268, + "step": 16098 + }, + { + "epoch": 1.3718789944610141, + "grad_norm": 12.75, + "learning_rate": 5.650492394574777e-07, + "loss": 0.4857, + "step": 16099 + }, + { + "epoch": 1.371964209629314, + "grad_norm": 16.75, + "learning_rate": 5.649086832135158e-07, + "loss": 0.6239, + "step": 16100 + }, + { + "epoch": 1.372049424797614, + "grad_norm": 12.0625, + "learning_rate": 5.647681393498433e-07, + "loss": 0.4281, + "step": 16101 + }, + { + "epoch": 1.372134639965914, + "grad_norm": 15.25, + "learning_rate": 5.646276078689991e-07, + "loss": 0.6254, + "step": 16102 + }, + { + "epoch": 1.372219855134214, + "grad_norm": 15.5, + "learning_rate": 5.644870887735236e-07, + "loss": 0.8104, + "step": 16103 + }, + { + "epoch": 1.372305070302514, + "grad_norm": 13.3125, + "learning_rate": 5.643465820659555e-07, + "loss": 0.4898, + "step": 16104 + }, + { + "epoch": 1.3723902854708139, + "grad_norm": 16.0, + "learning_rate": 5.64206087748834e-07, + "loss": 0.7035, + "step": 16105 + }, + { + "epoch": 1.3724755006391138, + "grad_norm": 18.25, + "learning_rate": 5.64065605824698e-07, + "loss": 0.7884, + "step": 16106 + }, + { + "epoch": 1.3725607158074138, + "grad_norm": 15.625, + "learning_rate": 5.63925136296087e-07, + "loss": 0.471, + "step": 16107 + }, + { + "epoch": 1.3726459309757137, + "grad_norm": 15.8125, + "learning_rate": 5.637846791655381e-07, + "loss": 0.7667, + "step": 16108 + }, + { + "epoch": 1.3727311461440137, + "grad_norm": 14.25, + "learning_rate": 5.636442344355901e-07, + "loss": 0.6779, + "step": 16109 + }, + { + "epoch": 1.3728163613123137, + "grad_norm": 13.0, + "learning_rate": 5.635038021087815e-07, + "loss": 0.3917, + "step": 16110 + }, + { + "epoch": 1.3729015764806136, + "grad_norm": 12.375, + "learning_rate": 5.633633821876488e-07, + "loss": 0.4201, + "step": 16111 + }, + { + "epoch": 1.3729867916489136, + "grad_norm": 12.4375, + "learning_rate": 5.63222974674731e-07, + "loss": 0.5646, + "step": 16112 + }, + { + "epoch": 1.3730720068172135, + "grad_norm": 13.5625, + "learning_rate": 5.630825795725641e-07, + "loss": 0.6594, + "step": 16113 + }, + { + "epoch": 1.3731572219855135, + "grad_norm": 17.125, + "learning_rate": 5.629421968836857e-07, + "loss": 0.6535, + "step": 16114 + }, + { + "epoch": 1.3732424371538134, + "grad_norm": 15.0625, + "learning_rate": 5.628018266106326e-07, + "loss": 0.5179, + "step": 16115 + }, + { + "epoch": 1.3733276523221134, + "grad_norm": 12.25, + "learning_rate": 5.626614687559422e-07, + "loss": 0.4165, + "step": 16116 + }, + { + "epoch": 1.3734128674904134, + "grad_norm": 14.1875, + "learning_rate": 5.625211233221496e-07, + "loss": 0.4319, + "step": 16117 + }, + { + "epoch": 1.3734980826587133, + "grad_norm": 16.625, + "learning_rate": 5.623807903117921e-07, + "loss": 0.7363, + "step": 16118 + }, + { + "epoch": 1.3735832978270133, + "grad_norm": 17.0, + "learning_rate": 5.622404697274047e-07, + "loss": 0.4946, + "step": 16119 + }, + { + "epoch": 1.3736685129953132, + "grad_norm": 12.1875, + "learning_rate": 5.621001615715242e-07, + "loss": 0.3595, + "step": 16120 + }, + { + "epoch": 1.3737537281636132, + "grad_norm": 15.875, + "learning_rate": 5.61959865846685e-07, + "loss": 0.7856, + "step": 16121 + }, + { + "epoch": 1.3738389433319131, + "grad_norm": 15.4375, + "learning_rate": 5.618195825554229e-07, + "loss": 0.5068, + "step": 16122 + }, + { + "epoch": 1.373924158500213, + "grad_norm": 11.125, + "learning_rate": 5.616793117002729e-07, + "loss": 0.2673, + "step": 16123 + }, + { + "epoch": 1.374009373668513, + "grad_norm": 12.125, + "learning_rate": 5.615390532837704e-07, + "loss": 0.5927, + "step": 16124 + }, + { + "epoch": 1.374094588836813, + "grad_norm": 20.75, + "learning_rate": 5.613988073084491e-07, + "loss": 0.8598, + "step": 16125 + }, + { + "epoch": 1.374179804005113, + "grad_norm": 22.75, + "learning_rate": 5.612585737768442e-07, + "loss": 0.9524, + "step": 16126 + }, + { + "epoch": 1.374265019173413, + "grad_norm": 23.75, + "learning_rate": 5.611183526914889e-07, + "loss": 1.0647, + "step": 16127 + }, + { + "epoch": 1.3743502343417129, + "grad_norm": 12.4375, + "learning_rate": 5.609781440549178e-07, + "loss": 0.5867, + "step": 16128 + }, + { + "epoch": 1.3744354495100128, + "grad_norm": 15.0, + "learning_rate": 5.608379478696649e-07, + "loss": 0.5602, + "step": 16129 + }, + { + "epoch": 1.3745206646783128, + "grad_norm": 13.8125, + "learning_rate": 5.60697764138263e-07, + "loss": 0.5048, + "step": 16130 + }, + { + "epoch": 1.3746058798466128, + "grad_norm": 18.625, + "learning_rate": 5.605575928632455e-07, + "loss": 0.6622, + "step": 16131 + }, + { + "epoch": 1.3746910950149127, + "grad_norm": 16.75, + "learning_rate": 5.604174340471455e-07, + "loss": 0.6081, + "step": 16132 + }, + { + "epoch": 1.3747763101832127, + "grad_norm": 15.3125, + "learning_rate": 5.602772876924965e-07, + "loss": 0.7895, + "step": 16133 + }, + { + "epoch": 1.3748615253515126, + "grad_norm": 12.875, + "learning_rate": 5.601371538018304e-07, + "loss": 0.3368, + "step": 16134 + }, + { + "epoch": 1.3749467405198126, + "grad_norm": 18.75, + "learning_rate": 5.599970323776794e-07, + "loss": 0.7931, + "step": 16135 + }, + { + "epoch": 1.3750319556881125, + "grad_norm": 15.4375, + "learning_rate": 5.598569234225757e-07, + "loss": 0.8102, + "step": 16136 + }, + { + "epoch": 1.3751171708564125, + "grad_norm": 10.5, + "learning_rate": 5.597168269390517e-07, + "loss": 0.2217, + "step": 16137 + }, + { + "epoch": 1.3752023860247125, + "grad_norm": 9.375, + "learning_rate": 5.595767429296385e-07, + "loss": 0.3478, + "step": 16138 + }, + { + "epoch": 1.3752876011930124, + "grad_norm": 15.75, + "learning_rate": 5.594366713968678e-07, + "loss": 0.3109, + "step": 16139 + }, + { + "epoch": 1.3753728163613124, + "grad_norm": 19.5, + "learning_rate": 5.592966123432708e-07, + "loss": 0.8687, + "step": 16140 + }, + { + "epoch": 1.3754580315296123, + "grad_norm": 17.125, + "learning_rate": 5.591565657713792e-07, + "loss": 0.7723, + "step": 16141 + }, + { + "epoch": 1.3755432466979123, + "grad_norm": 16.25, + "learning_rate": 5.59016531683723e-07, + "loss": 0.8457, + "step": 16142 + }, + { + "epoch": 1.3756284618662122, + "grad_norm": 20.0, + "learning_rate": 5.588765100828325e-07, + "loss": 0.8141, + "step": 16143 + }, + { + "epoch": 1.3757136770345122, + "grad_norm": 14.5625, + "learning_rate": 5.587365009712384e-07, + "loss": 0.6031, + "step": 16144 + }, + { + "epoch": 1.3757988922028122, + "grad_norm": 11.5625, + "learning_rate": 5.585965043514709e-07, + "loss": 0.2805, + "step": 16145 + }, + { + "epoch": 1.375884107371112, + "grad_norm": 14.625, + "learning_rate": 5.584565202260602e-07, + "loss": 0.8765, + "step": 16146 + }, + { + "epoch": 1.375969322539412, + "grad_norm": 18.25, + "learning_rate": 5.583165485975352e-07, + "loss": 0.7144, + "step": 16147 + }, + { + "epoch": 1.376054537707712, + "grad_norm": 9.8125, + "learning_rate": 5.581765894684257e-07, + "loss": 0.2466, + "step": 16148 + }, + { + "epoch": 1.376139752876012, + "grad_norm": 16.5, + "learning_rate": 5.580366428412613e-07, + "loss": 0.7653, + "step": 16149 + }, + { + "epoch": 1.376224968044312, + "grad_norm": 17.5, + "learning_rate": 5.578967087185708e-07, + "loss": 0.8808, + "step": 16150 + }, + { + "epoch": 1.376310183212612, + "grad_norm": 16.875, + "learning_rate": 5.57756787102882e-07, + "loss": 0.6259, + "step": 16151 + }, + { + "epoch": 1.3763953983809118, + "grad_norm": 11.4375, + "learning_rate": 5.576168779967243e-07, + "loss": 0.3255, + "step": 16152 + }, + { + "epoch": 1.3764806135492118, + "grad_norm": 13.375, + "learning_rate": 5.574769814026258e-07, + "loss": 0.5066, + "step": 16153 + }, + { + "epoch": 1.3765658287175118, + "grad_norm": 15.875, + "learning_rate": 5.573370973231153e-07, + "loss": 0.5777, + "step": 16154 + }, + { + "epoch": 1.3766510438858117, + "grad_norm": 14.25, + "learning_rate": 5.571972257607193e-07, + "loss": 0.7687, + "step": 16155 + }, + { + "epoch": 1.3767362590541117, + "grad_norm": 12.0, + "learning_rate": 5.570573667179662e-07, + "loss": 0.5043, + "step": 16156 + }, + { + "epoch": 1.3768214742224116, + "grad_norm": 17.25, + "learning_rate": 5.569175201973831e-07, + "loss": 0.6726, + "step": 16157 + }, + { + "epoch": 1.3769066893907116, + "grad_norm": 17.0, + "learning_rate": 5.56777686201498e-07, + "loss": 0.7493, + "step": 16158 + }, + { + "epoch": 1.3769919045590115, + "grad_norm": 13.625, + "learning_rate": 5.566378647328372e-07, + "loss": 0.4753, + "step": 16159 + }, + { + "epoch": 1.3770771197273115, + "grad_norm": 14.75, + "learning_rate": 5.564980557939269e-07, + "loss": 0.5677, + "step": 16160 + }, + { + "epoch": 1.3771623348956115, + "grad_norm": 17.125, + "learning_rate": 5.563582593872941e-07, + "loss": 0.4901, + "step": 16161 + }, + { + "epoch": 1.3772475500639114, + "grad_norm": 23.375, + "learning_rate": 5.562184755154651e-07, + "loss": 0.7359, + "step": 16162 + }, + { + "epoch": 1.3773327652322114, + "grad_norm": 22.25, + "learning_rate": 5.560787041809664e-07, + "loss": 0.5983, + "step": 16163 + }, + { + "epoch": 1.3774179804005113, + "grad_norm": 12.9375, + "learning_rate": 5.559389453863227e-07, + "loss": 0.3806, + "step": 16164 + }, + { + "epoch": 1.3775031955688113, + "grad_norm": 15.375, + "learning_rate": 5.557991991340603e-07, + "loss": 0.6937, + "step": 16165 + }, + { + "epoch": 1.3775884107371112, + "grad_norm": 17.625, + "learning_rate": 5.556594654267049e-07, + "loss": 0.7008, + "step": 16166 + }, + { + "epoch": 1.3776736259054112, + "grad_norm": 15.625, + "learning_rate": 5.555197442667811e-07, + "loss": 0.6989, + "step": 16167 + }, + { + "epoch": 1.3777588410737112, + "grad_norm": 14.75, + "learning_rate": 5.553800356568136e-07, + "loss": 0.5194, + "step": 16168 + }, + { + "epoch": 1.3778440562420111, + "grad_norm": 31.5, + "learning_rate": 5.552403395993271e-07, + "loss": 1.1434, + "step": 16169 + }, + { + "epoch": 1.377929271410311, + "grad_norm": 17.125, + "learning_rate": 5.551006560968464e-07, + "loss": 1.2713, + "step": 16170 + }, + { + "epoch": 1.378014486578611, + "grad_norm": 17.625, + "learning_rate": 5.549609851518961e-07, + "loss": 0.5922, + "step": 16171 + }, + { + "epoch": 1.378099701746911, + "grad_norm": 13.8125, + "learning_rate": 5.548213267669993e-07, + "loss": 0.2707, + "step": 16172 + }, + { + "epoch": 1.378184916915211, + "grad_norm": 25.25, + "learning_rate": 5.546816809446802e-07, + "loss": 1.1922, + "step": 16173 + }, + { + "epoch": 1.378270132083511, + "grad_norm": 12.3125, + "learning_rate": 5.545420476874626e-07, + "loss": 0.3915, + "step": 16174 + }, + { + "epoch": 1.3783553472518109, + "grad_norm": 21.75, + "learning_rate": 5.544024269978692e-07, + "loss": 0.5879, + "step": 16175 + }, + { + "epoch": 1.3784405624201108, + "grad_norm": 17.25, + "learning_rate": 5.542628188784239e-07, + "loss": 0.6716, + "step": 16176 + }, + { + "epoch": 1.3785257775884108, + "grad_norm": 15.6875, + "learning_rate": 5.541232233316486e-07, + "loss": 0.4344, + "step": 16177 + }, + { + "epoch": 1.3786109927567107, + "grad_norm": 15.625, + "learning_rate": 5.539836403600664e-07, + "loss": 0.5461, + "step": 16178 + }, + { + "epoch": 1.3786962079250107, + "grad_norm": 15.0625, + "learning_rate": 5.538440699662002e-07, + "loss": 0.6723, + "step": 16179 + }, + { + "epoch": 1.3787814230933106, + "grad_norm": 15.375, + "learning_rate": 5.537045121525712e-07, + "loss": 0.6039, + "step": 16180 + }, + { + "epoch": 1.3788666382616106, + "grad_norm": 17.25, + "learning_rate": 5.535649669217019e-07, + "loss": 0.8407, + "step": 16181 + }, + { + "epoch": 1.3789518534299106, + "grad_norm": 27.375, + "learning_rate": 5.534254342761139e-07, + "loss": 0.9365, + "step": 16182 + }, + { + "epoch": 1.3790370685982105, + "grad_norm": 12.0625, + "learning_rate": 5.532859142183293e-07, + "loss": 0.4117, + "step": 16183 + }, + { + "epoch": 1.3791222837665105, + "grad_norm": 17.5, + "learning_rate": 5.531464067508688e-07, + "loss": 0.5665, + "step": 16184 + }, + { + "epoch": 1.3792074989348104, + "grad_norm": 16.25, + "learning_rate": 5.53006911876253e-07, + "loss": 0.8722, + "step": 16185 + }, + { + "epoch": 1.3792927141031104, + "grad_norm": 15.625, + "learning_rate": 5.528674295970031e-07, + "loss": 0.5691, + "step": 16186 + }, + { + "epoch": 1.3793779292714103, + "grad_norm": 17.375, + "learning_rate": 5.527279599156399e-07, + "loss": 0.8627, + "step": 16187 + }, + { + "epoch": 1.3794631444397103, + "grad_norm": 16.75, + "learning_rate": 5.525885028346839e-07, + "loss": 0.624, + "step": 16188 + }, + { + "epoch": 1.3795483596080103, + "grad_norm": 11.1875, + "learning_rate": 5.524490583566547e-07, + "loss": 0.2264, + "step": 16189 + }, + { + "epoch": 1.3796335747763102, + "grad_norm": 16.875, + "learning_rate": 5.523096264840724e-07, + "loss": 0.5036, + "step": 16190 + }, + { + "epoch": 1.3797187899446102, + "grad_norm": 15.375, + "learning_rate": 5.52170207219457e-07, + "loss": 0.5264, + "step": 16191 + }, + { + "epoch": 1.3798040051129101, + "grad_norm": 12.3125, + "learning_rate": 5.520308005653276e-07, + "loss": 0.379, + "step": 16192 + }, + { + "epoch": 1.37988922028121, + "grad_norm": 15.6875, + "learning_rate": 5.518914065242037e-07, + "loss": 0.5936, + "step": 16193 + }, + { + "epoch": 1.37997443544951, + "grad_norm": 15.3125, + "learning_rate": 5.517520250986036e-07, + "loss": 0.7415, + "step": 16194 + }, + { + "epoch": 1.38005965061781, + "grad_norm": 11.5, + "learning_rate": 5.516126562910467e-07, + "loss": 0.267, + "step": 16195 + }, + { + "epoch": 1.38014486578611, + "grad_norm": 21.375, + "learning_rate": 5.514733001040517e-07, + "loss": 0.6606, + "step": 16196 + }, + { + "epoch": 1.38023008095441, + "grad_norm": 18.0, + "learning_rate": 5.513339565401362e-07, + "loss": 0.7756, + "step": 16197 + }, + { + "epoch": 1.3803152961227099, + "grad_norm": 14.3125, + "learning_rate": 5.511946256018187e-07, + "loss": 0.6675, + "step": 16198 + }, + { + "epoch": 1.3804005112910098, + "grad_norm": 16.75, + "learning_rate": 5.510553072916175e-07, + "loss": 0.6884, + "step": 16199 + }, + { + "epoch": 1.3804857264593098, + "grad_norm": 17.5, + "learning_rate": 5.509160016120492e-07, + "loss": 0.5537, + "step": 16200 + }, + { + "epoch": 1.3805709416276097, + "grad_norm": 18.25, + "learning_rate": 5.507767085656323e-07, + "loss": 0.8159, + "step": 16201 + }, + { + "epoch": 1.3806561567959097, + "grad_norm": 9.6875, + "learning_rate": 5.506374281548827e-07, + "loss": 0.2083, + "step": 16202 + }, + { + "epoch": 1.3807413719642097, + "grad_norm": 22.0, + "learning_rate": 5.504981603823182e-07, + "loss": 0.9635, + "step": 16203 + }, + { + "epoch": 1.3808265871325096, + "grad_norm": 11.0625, + "learning_rate": 5.503589052504554e-07, + "loss": 0.3829, + "step": 16204 + }, + { + "epoch": 1.3809118023008096, + "grad_norm": 14.5625, + "learning_rate": 5.502196627618112e-07, + "loss": 0.4774, + "step": 16205 + }, + { + "epoch": 1.3809970174691095, + "grad_norm": 14.3125, + "learning_rate": 5.500804329189008e-07, + "loss": 0.6848, + "step": 16206 + }, + { + "epoch": 1.3810822326374095, + "grad_norm": 16.5, + "learning_rate": 5.499412157242411e-07, + "loss": 0.6415, + "step": 16207 + }, + { + "epoch": 1.3811674478057094, + "grad_norm": 31.0, + "learning_rate": 5.498020111803475e-07, + "loss": 0.8314, + "step": 16208 + }, + { + "epoch": 1.3812526629740094, + "grad_norm": 12.375, + "learning_rate": 5.496628192897359e-07, + "loss": 0.4089, + "step": 16209 + }, + { + "epoch": 1.3813378781423093, + "grad_norm": 16.25, + "learning_rate": 5.49523640054921e-07, + "loss": 0.7087, + "step": 16210 + }, + { + "epoch": 1.3814230933106093, + "grad_norm": 18.125, + "learning_rate": 5.493844734784183e-07, + "loss": 0.6688, + "step": 16211 + }, + { + "epoch": 1.3815083084789093, + "grad_norm": 11.9375, + "learning_rate": 5.492453195627427e-07, + "loss": 0.4449, + "step": 16212 + }, + { + "epoch": 1.3815935236472092, + "grad_norm": 13.0, + "learning_rate": 5.491061783104093e-07, + "loss": 0.6191, + "step": 16213 + }, + { + "epoch": 1.3816787388155092, + "grad_norm": 16.375, + "learning_rate": 5.489670497239316e-07, + "loss": 0.7047, + "step": 16214 + }, + { + "epoch": 1.3817639539838091, + "grad_norm": 9.6875, + "learning_rate": 5.488279338058247e-07, + "loss": 0.2489, + "step": 16215 + }, + { + "epoch": 1.381849169152109, + "grad_norm": 12.25, + "learning_rate": 5.486888305586018e-07, + "loss": 0.4164, + "step": 16216 + }, + { + "epoch": 1.381934384320409, + "grad_norm": 17.875, + "learning_rate": 5.485497399847766e-07, + "loss": 0.7807, + "step": 16217 + }, + { + "epoch": 1.382019599488709, + "grad_norm": 26.125, + "learning_rate": 5.484106620868637e-07, + "loss": 0.3848, + "step": 16218 + }, + { + "epoch": 1.382104814657009, + "grad_norm": 12.3125, + "learning_rate": 5.482715968673752e-07, + "loss": 0.4489, + "step": 16219 + }, + { + "epoch": 1.382190029825309, + "grad_norm": 21.125, + "learning_rate": 5.481325443288245e-07, + "loss": 0.5143, + "step": 16220 + }, + { + "epoch": 1.3822752449936089, + "grad_norm": 16.375, + "learning_rate": 5.479935044737245e-07, + "loss": 0.8863, + "step": 16221 + }, + { + "epoch": 1.3823604601619088, + "grad_norm": 12.9375, + "learning_rate": 5.478544773045882e-07, + "loss": 0.4683, + "step": 16222 + }, + { + "epoch": 1.3824456753302088, + "grad_norm": 16.625, + "learning_rate": 5.477154628239277e-07, + "loss": 0.4664, + "step": 16223 + }, + { + "epoch": 1.3825308904985087, + "grad_norm": 13.0, + "learning_rate": 5.475764610342544e-07, + "loss": 0.3818, + "step": 16224 + }, + { + "epoch": 1.3826161056668087, + "grad_norm": 15.3125, + "learning_rate": 5.47437471938081e-07, + "loss": 0.6835, + "step": 16225 + }, + { + "epoch": 1.3827013208351087, + "grad_norm": 14.875, + "learning_rate": 5.472984955379192e-07, + "loss": 0.5074, + "step": 16226 + }, + { + "epoch": 1.3827865360034086, + "grad_norm": 13.4375, + "learning_rate": 5.471595318362797e-07, + "loss": 0.4608, + "step": 16227 + }, + { + "epoch": 1.3828717511717086, + "grad_norm": 16.75, + "learning_rate": 5.470205808356745e-07, + "loss": 0.5839, + "step": 16228 + }, + { + "epoch": 1.3829569663400085, + "grad_norm": 11.875, + "learning_rate": 5.468816425386141e-07, + "loss": 0.4228, + "step": 16229 + }, + { + "epoch": 1.3830421815083085, + "grad_norm": 13.0, + "learning_rate": 5.4674271694761e-07, + "loss": 0.4436, + "step": 16230 + }, + { + "epoch": 1.3831273966766084, + "grad_norm": 13.25, + "learning_rate": 5.466038040651721e-07, + "loss": 0.4287, + "step": 16231 + }, + { + "epoch": 1.3832126118449084, + "grad_norm": 14.0, + "learning_rate": 5.464649038938104e-07, + "loss": 0.6583, + "step": 16232 + }, + { + "epoch": 1.3832978270132084, + "grad_norm": 19.0, + "learning_rate": 5.463260164360353e-07, + "loss": 0.7777, + "step": 16233 + }, + { + "epoch": 1.3833830421815083, + "grad_norm": 29.5, + "learning_rate": 5.461871416943566e-07, + "loss": 1.0111, + "step": 16234 + }, + { + "epoch": 1.3834682573498083, + "grad_norm": 20.25, + "learning_rate": 5.460482796712846e-07, + "loss": 0.9076, + "step": 16235 + }, + { + "epoch": 1.3835534725181082, + "grad_norm": 12.1875, + "learning_rate": 5.459094303693276e-07, + "loss": 0.311, + "step": 16236 + }, + { + "epoch": 1.3836386876864082, + "grad_norm": 21.75, + "learning_rate": 5.457705937909951e-07, + "loss": 0.5241, + "step": 16237 + }, + { + "epoch": 1.3837239028547081, + "grad_norm": 19.625, + "learning_rate": 5.456317699387962e-07, + "loss": 0.7599, + "step": 16238 + }, + { + "epoch": 1.383809118023008, + "grad_norm": 13.625, + "learning_rate": 5.4549295881524e-07, + "loss": 0.465, + "step": 16239 + }, + { + "epoch": 1.383894333191308, + "grad_norm": 21.375, + "learning_rate": 5.453541604228343e-07, + "loss": 0.7384, + "step": 16240 + }, + { + "epoch": 1.383979548359608, + "grad_norm": 20.875, + "learning_rate": 5.452153747640873e-07, + "loss": 0.7375, + "step": 16241 + }, + { + "epoch": 1.384064763527908, + "grad_norm": 18.625, + "learning_rate": 5.45076601841507e-07, + "loss": 0.7463, + "step": 16242 + }, + { + "epoch": 1.384149978696208, + "grad_norm": 18.875, + "learning_rate": 5.449378416576019e-07, + "loss": 0.9104, + "step": 16243 + }, + { + "epoch": 1.3842351938645079, + "grad_norm": 15.75, + "learning_rate": 5.447990942148786e-07, + "loss": 0.685, + "step": 16244 + }, + { + "epoch": 1.3843204090328078, + "grad_norm": 14.5, + "learning_rate": 5.446603595158447e-07, + "loss": 0.5088, + "step": 16245 + }, + { + "epoch": 1.3844056242011078, + "grad_norm": 14.3125, + "learning_rate": 5.445216375630073e-07, + "loss": 0.5055, + "step": 16246 + }, + { + "epoch": 1.3844908393694078, + "grad_norm": 17.125, + "learning_rate": 5.443829283588739e-07, + "loss": 0.6934, + "step": 16247 + }, + { + "epoch": 1.3845760545377077, + "grad_norm": 21.125, + "learning_rate": 5.442442319059506e-07, + "loss": 0.8992, + "step": 16248 + }, + { + "epoch": 1.3846612697060077, + "grad_norm": 21.0, + "learning_rate": 5.441055482067431e-07, + "loss": 0.6825, + "step": 16249 + }, + { + "epoch": 1.3847464848743076, + "grad_norm": 12.375, + "learning_rate": 5.439668772637582e-07, + "loss": 0.392, + "step": 16250 + }, + { + "epoch": 1.3848317000426076, + "grad_norm": 15.1875, + "learning_rate": 5.43828219079502e-07, + "loss": 0.4674, + "step": 16251 + }, + { + "epoch": 1.3849169152109075, + "grad_norm": 20.625, + "learning_rate": 5.436895736564802e-07, + "loss": 0.384, + "step": 16252 + }, + { + "epoch": 1.3850021303792075, + "grad_norm": 14.6875, + "learning_rate": 5.435509409971977e-07, + "loss": 0.8089, + "step": 16253 + }, + { + "epoch": 1.3850873455475075, + "grad_norm": 12.875, + "learning_rate": 5.434123211041601e-07, + "loss": 0.4132, + "step": 16254 + }, + { + "epoch": 1.3851725607158074, + "grad_norm": 17.25, + "learning_rate": 5.432737139798725e-07, + "loss": 0.6034, + "step": 16255 + }, + { + "epoch": 1.3852577758841074, + "grad_norm": 17.125, + "learning_rate": 5.431351196268403e-07, + "loss": 0.4918, + "step": 16256 + }, + { + "epoch": 1.3853429910524073, + "grad_norm": 12.5, + "learning_rate": 5.429965380475665e-07, + "loss": 0.3789, + "step": 16257 + }, + { + "epoch": 1.3854282062207073, + "grad_norm": 9.1875, + "learning_rate": 5.42857969244556e-07, + "loss": 0.1938, + "step": 16258 + }, + { + "epoch": 1.3855134213890072, + "grad_norm": 18.375, + "learning_rate": 5.427194132203133e-07, + "loss": 0.8812, + "step": 16259 + }, + { + "epoch": 1.3855986365573072, + "grad_norm": 21.75, + "learning_rate": 5.425808699773423e-07, + "loss": 0.684, + "step": 16260 + }, + { + "epoch": 1.3856838517256072, + "grad_norm": 16.25, + "learning_rate": 5.424423395181461e-07, + "loss": 0.6042, + "step": 16261 + }, + { + "epoch": 1.385769066893907, + "grad_norm": 16.5, + "learning_rate": 5.423038218452281e-07, + "loss": 0.7886, + "step": 16262 + }, + { + "epoch": 1.385854282062207, + "grad_norm": 17.625, + "learning_rate": 5.421653169610919e-07, + "loss": 0.6448, + "step": 16263 + }, + { + "epoch": 1.385939497230507, + "grad_norm": 20.75, + "learning_rate": 5.420268248682405e-07, + "loss": 0.7995, + "step": 16264 + }, + { + "epoch": 1.386024712398807, + "grad_norm": 15.9375, + "learning_rate": 5.418883455691763e-07, + "loss": 0.7082, + "step": 16265 + }, + { + "epoch": 1.386109927567107, + "grad_norm": 13.9375, + "learning_rate": 5.417498790664014e-07, + "loss": 0.4509, + "step": 16266 + }, + { + "epoch": 1.386195142735407, + "grad_norm": 19.25, + "learning_rate": 5.416114253624183e-07, + "loss": 0.7585, + "step": 16267 + }, + { + "epoch": 1.3862803579037069, + "grad_norm": 20.5, + "learning_rate": 5.414729844597291e-07, + "loss": 0.7591, + "step": 16268 + }, + { + "epoch": 1.3863655730720068, + "grad_norm": 19.625, + "learning_rate": 5.41334556360836e-07, + "loss": 0.9383, + "step": 16269 + }, + { + "epoch": 1.3864507882403068, + "grad_norm": 16.625, + "learning_rate": 5.411961410682398e-07, + "loss": 0.7127, + "step": 16270 + }, + { + "epoch": 1.3865360034086067, + "grad_norm": 14.875, + "learning_rate": 5.410577385844419e-07, + "loss": 0.7662, + "step": 16271 + }, + { + "epoch": 1.3866212185769067, + "grad_norm": 14.5625, + "learning_rate": 5.409193489119441e-07, + "loss": 0.7139, + "step": 16272 + }, + { + "epoch": 1.3867064337452066, + "grad_norm": 19.0, + "learning_rate": 5.407809720532466e-07, + "loss": 0.7459, + "step": 16273 + }, + { + "epoch": 1.3867916489135066, + "grad_norm": 12.625, + "learning_rate": 5.406426080108497e-07, + "loss": 0.4128, + "step": 16274 + }, + { + "epoch": 1.3868768640818065, + "grad_norm": 16.5, + "learning_rate": 5.405042567872543e-07, + "loss": 0.7258, + "step": 16275 + }, + { + "epoch": 1.3869620792501065, + "grad_norm": 20.75, + "learning_rate": 5.403659183849604e-07, + "loss": 0.5366, + "step": 16276 + }, + { + "epoch": 1.3870472944184065, + "grad_norm": 9.5, + "learning_rate": 5.402275928064682e-07, + "loss": 0.3374, + "step": 16277 + }, + { + "epoch": 1.3871325095867064, + "grad_norm": 28.5, + "learning_rate": 5.400892800542769e-07, + "loss": 0.7199, + "step": 16278 + }, + { + "epoch": 1.3872177247550064, + "grad_norm": 12.5625, + "learning_rate": 5.39950980130886e-07, + "loss": 0.5541, + "step": 16279 + }, + { + "epoch": 1.3873029399233063, + "grad_norm": 16.875, + "learning_rate": 5.398126930387955e-07, + "loss": 0.3705, + "step": 16280 + }, + { + "epoch": 1.3873881550916063, + "grad_norm": 17.625, + "learning_rate": 5.396744187805032e-07, + "loss": 0.564, + "step": 16281 + }, + { + "epoch": 1.3874733702599062, + "grad_norm": 20.625, + "learning_rate": 5.395361573585091e-07, + "loss": 0.8389, + "step": 16282 + }, + { + "epoch": 1.3875585854282062, + "grad_norm": 14.8125, + "learning_rate": 5.393979087753104e-07, + "loss": 0.5423, + "step": 16283 + }, + { + "epoch": 1.3876438005965062, + "grad_norm": 18.125, + "learning_rate": 5.392596730334061e-07, + "loss": 0.7961, + "step": 16284 + }, + { + "epoch": 1.3877290157648061, + "grad_norm": 27.125, + "learning_rate": 5.391214501352943e-07, + "loss": 0.3258, + "step": 16285 + }, + { + "epoch": 1.387814230933106, + "grad_norm": 12.25, + "learning_rate": 5.389832400834732e-07, + "loss": 0.3986, + "step": 16286 + }, + { + "epoch": 1.387899446101406, + "grad_norm": 11.125, + "learning_rate": 5.388450428804394e-07, + "loss": 0.3016, + "step": 16287 + }, + { + "epoch": 1.387984661269706, + "grad_norm": 14.0, + "learning_rate": 5.387068585286913e-07, + "loss": 0.6687, + "step": 16288 + }, + { + "epoch": 1.388069876438006, + "grad_norm": 19.875, + "learning_rate": 5.385686870307251e-07, + "loss": 0.691, + "step": 16289 + }, + { + "epoch": 1.388155091606306, + "grad_norm": 16.875, + "learning_rate": 5.384305283890384e-07, + "loss": 0.4801, + "step": 16290 + }, + { + "epoch": 1.3882403067746059, + "grad_norm": 14.5625, + "learning_rate": 5.382923826061274e-07, + "loss": 0.7809, + "step": 16291 + }, + { + "epoch": 1.3883255219429058, + "grad_norm": 21.25, + "learning_rate": 5.381542496844885e-07, + "loss": 0.8043, + "step": 16292 + }, + { + "epoch": 1.3884107371112058, + "grad_norm": 16.0, + "learning_rate": 5.380161296266183e-07, + "loss": 0.609, + "step": 16293 + }, + { + "epoch": 1.3884959522795057, + "grad_norm": 21.125, + "learning_rate": 5.378780224350129e-07, + "loss": 1.0666, + "step": 16294 + }, + { + "epoch": 1.3885811674478057, + "grad_norm": 15.8125, + "learning_rate": 5.377399281121673e-07, + "loss": 0.7965, + "step": 16295 + }, + { + "epoch": 1.3886663826161056, + "grad_norm": 11.9375, + "learning_rate": 5.37601846660578e-07, + "loss": 0.4348, + "step": 16296 + }, + { + "epoch": 1.3887515977844056, + "grad_norm": 11.6875, + "learning_rate": 5.374637780827391e-07, + "loss": 0.2844, + "step": 16297 + }, + { + "epoch": 1.3888368129527056, + "grad_norm": 25.625, + "learning_rate": 5.373257223811464e-07, + "loss": 0.3522, + "step": 16298 + }, + { + "epoch": 1.3889220281210055, + "grad_norm": 11.4375, + "learning_rate": 5.371876795582948e-07, + "loss": 0.3121, + "step": 16299 + }, + { + "epoch": 1.3890072432893055, + "grad_norm": 14.5, + "learning_rate": 5.370496496166783e-07, + "loss": 0.5959, + "step": 16300 + }, + { + "epoch": 1.3890924584576054, + "grad_norm": 23.125, + "learning_rate": 5.369116325587916e-07, + "loss": 0.9148, + "step": 16301 + }, + { + "epoch": 1.3891776736259054, + "grad_norm": 17.25, + "learning_rate": 5.367736283871286e-07, + "loss": 0.596, + "step": 16302 + }, + { + "epoch": 1.3892628887942053, + "grad_norm": 17.375, + "learning_rate": 5.36635637104184e-07, + "loss": 0.7201, + "step": 16303 + }, + { + "epoch": 1.3893481039625053, + "grad_norm": 21.5, + "learning_rate": 5.364976587124506e-07, + "loss": 0.5346, + "step": 16304 + }, + { + "epoch": 1.3894333191308053, + "grad_norm": 12.25, + "learning_rate": 5.363596932144216e-07, + "loss": 0.2148, + "step": 16305 + }, + { + "epoch": 1.3895185342991052, + "grad_norm": 15.75, + "learning_rate": 5.362217406125906e-07, + "loss": 0.6497, + "step": 16306 + }, + { + "epoch": 1.3896037494674052, + "grad_norm": 15.5, + "learning_rate": 5.360838009094508e-07, + "loss": 0.4809, + "step": 16307 + }, + { + "epoch": 1.3896889646357051, + "grad_norm": 18.625, + "learning_rate": 5.359458741074944e-07, + "loss": 0.7103, + "step": 16308 + }, + { + "epoch": 1.389774179804005, + "grad_norm": 32.0, + "learning_rate": 5.35807960209214e-07, + "loss": 0.7895, + "step": 16309 + }, + { + "epoch": 1.389859394972305, + "grad_norm": 26.0, + "learning_rate": 5.356700592171019e-07, + "loss": 1.0872, + "step": 16310 + }, + { + "epoch": 1.389944610140605, + "grad_norm": 12.125, + "learning_rate": 5.355321711336506e-07, + "loss": 0.4133, + "step": 16311 + }, + { + "epoch": 1.390029825308905, + "grad_norm": 15.625, + "learning_rate": 5.35394295961351e-07, + "loss": 0.7255, + "step": 16312 + }, + { + "epoch": 1.390115040477205, + "grad_norm": 24.625, + "learning_rate": 5.352564337026955e-07, + "loss": 1.0113, + "step": 16313 + }, + { + "epoch": 1.3902002556455049, + "grad_norm": 14.625, + "learning_rate": 5.351185843601745e-07, + "loss": 0.5893, + "step": 16314 + }, + { + "epoch": 1.3902854708138048, + "grad_norm": 11.4375, + "learning_rate": 5.349807479362796e-07, + "loss": 0.3435, + "step": 16315 + }, + { + "epoch": 1.3903706859821048, + "grad_norm": 14.125, + "learning_rate": 5.348429244335019e-07, + "loss": 0.6399, + "step": 16316 + }, + { + "epoch": 1.3904559011504047, + "grad_norm": 12.25, + "learning_rate": 5.347051138543315e-07, + "loss": 0.3775, + "step": 16317 + }, + { + "epoch": 1.3905411163187047, + "grad_norm": 17.0, + "learning_rate": 5.345673162012588e-07, + "loss": 0.8836, + "step": 16318 + }, + { + "epoch": 1.3906263314870047, + "grad_norm": 15.4375, + "learning_rate": 5.344295314767746e-07, + "loss": 0.491, + "step": 16319 + }, + { + "epoch": 1.3907115466553046, + "grad_norm": 13.3125, + "learning_rate": 5.34291759683368e-07, + "loss": 0.5502, + "step": 16320 + }, + { + "epoch": 1.3907967618236046, + "grad_norm": 15.0, + "learning_rate": 5.341540008235293e-07, + "loss": 0.7206, + "step": 16321 + }, + { + "epoch": 1.3908819769919045, + "grad_norm": 20.375, + "learning_rate": 5.340162548997473e-07, + "loss": 0.8401, + "step": 16322 + }, + { + "epoch": 1.3909671921602045, + "grad_norm": 17.25, + "learning_rate": 5.338785219145114e-07, + "loss": 0.7415, + "step": 16323 + }, + { + "epoch": 1.3910524073285044, + "grad_norm": 14.6875, + "learning_rate": 5.337408018703112e-07, + "loss": 0.6126, + "step": 16324 + }, + { + "epoch": 1.3911376224968044, + "grad_norm": 16.25, + "learning_rate": 5.336030947696346e-07, + "loss": 0.4307, + "step": 16325 + }, + { + "epoch": 1.3912228376651044, + "grad_norm": 11.3125, + "learning_rate": 5.334654006149705e-07, + "loss": 0.8282, + "step": 16326 + }, + { + "epoch": 1.3913080528334043, + "grad_norm": 14.0625, + "learning_rate": 5.333277194088071e-07, + "loss": 0.5582, + "step": 16327 + }, + { + "epoch": 1.3913932680017043, + "grad_norm": 15.75, + "learning_rate": 5.33190051153633e-07, + "loss": 0.5498, + "step": 16328 + }, + { + "epoch": 1.3914784831700042, + "grad_norm": 18.625, + "learning_rate": 5.330523958519353e-07, + "loss": 0.5722, + "step": 16329 + }, + { + "epoch": 1.3915636983383042, + "grad_norm": 22.75, + "learning_rate": 5.329147535062014e-07, + "loss": 0.7093, + "step": 16330 + }, + { + "epoch": 1.3916489135066041, + "grad_norm": 12.5625, + "learning_rate": 5.32777124118919e-07, + "loss": 0.4678, + "step": 16331 + }, + { + "epoch": 1.391734128674904, + "grad_norm": 16.375, + "learning_rate": 5.326395076925753e-07, + "loss": 0.7793, + "step": 16332 + }, + { + "epoch": 1.391819343843204, + "grad_norm": 14.75, + "learning_rate": 5.325019042296572e-07, + "loss": 0.4395, + "step": 16333 + }, + { + "epoch": 1.391904559011504, + "grad_norm": 17.5, + "learning_rate": 5.32364313732651e-07, + "loss": 0.785, + "step": 16334 + }, + { + "epoch": 1.391989774179804, + "grad_norm": 16.0, + "learning_rate": 5.322267362040433e-07, + "loss": 0.7196, + "step": 16335 + }, + { + "epoch": 1.392074989348104, + "grad_norm": 9.875, + "learning_rate": 5.320891716463206e-07, + "loss": 0.5907, + "step": 16336 + }, + { + "epoch": 1.3921602045164039, + "grad_norm": 17.75, + "learning_rate": 5.319516200619686e-07, + "loss": 0.688, + "step": 16337 + }, + { + "epoch": 1.3922454196847038, + "grad_norm": 13.3125, + "learning_rate": 5.318140814534722e-07, + "loss": 0.3144, + "step": 16338 + }, + { + "epoch": 1.3923306348530038, + "grad_norm": 16.25, + "learning_rate": 5.316765558233178e-07, + "loss": 0.4942, + "step": 16339 + }, + { + "epoch": 1.3924158500213037, + "grad_norm": 19.375, + "learning_rate": 5.315390431739902e-07, + "loss": 0.6378, + "step": 16340 + }, + { + "epoch": 1.3925010651896037, + "grad_norm": 17.375, + "learning_rate": 5.31401543507975e-07, + "loss": 0.7395, + "step": 16341 + }, + { + "epoch": 1.3925862803579037, + "grad_norm": 16.0, + "learning_rate": 5.312640568277562e-07, + "loss": 0.7385, + "step": 16342 + }, + { + "epoch": 1.3926714955262036, + "grad_norm": 18.0, + "learning_rate": 5.311265831358185e-07, + "loss": 0.5627, + "step": 16343 + }, + { + "epoch": 1.3927567106945036, + "grad_norm": 11.5625, + "learning_rate": 5.309891224346464e-07, + "loss": 0.2779, + "step": 16344 + }, + { + "epoch": 1.3928419258628035, + "grad_norm": 16.375, + "learning_rate": 5.308516747267243e-07, + "loss": 0.6792, + "step": 16345 + }, + { + "epoch": 1.3929271410311035, + "grad_norm": 25.375, + "learning_rate": 5.307142400145357e-07, + "loss": 0.7731, + "step": 16346 + }, + { + "epoch": 1.3930123561994034, + "grad_norm": 13.375, + "learning_rate": 5.305768183005634e-07, + "loss": 0.6184, + "step": 16347 + }, + { + "epoch": 1.3930975713677034, + "grad_norm": 16.75, + "learning_rate": 5.304394095872917e-07, + "loss": 0.5288, + "step": 16348 + }, + { + "epoch": 1.3931827865360034, + "grad_norm": 14.875, + "learning_rate": 5.303020138772038e-07, + "loss": 0.6564, + "step": 16349 + }, + { + "epoch": 1.3932680017043033, + "grad_norm": 13.625, + "learning_rate": 5.301646311727816e-07, + "loss": 0.5452, + "step": 16350 + }, + { + "epoch": 1.3933532168726033, + "grad_norm": 15.3125, + "learning_rate": 5.300272614765086e-07, + "loss": 0.3012, + "step": 16351 + }, + { + "epoch": 1.3934384320409032, + "grad_norm": 15.5625, + "learning_rate": 5.298899047908667e-07, + "loss": 0.5805, + "step": 16352 + }, + { + "epoch": 1.3935236472092032, + "grad_norm": 15.375, + "learning_rate": 5.29752561118339e-07, + "loss": 0.5796, + "step": 16353 + }, + { + "epoch": 1.3936088623775031, + "grad_norm": 18.625, + "learning_rate": 5.296152304614069e-07, + "loss": 0.8866, + "step": 16354 + }, + { + "epoch": 1.393694077545803, + "grad_norm": 17.75, + "learning_rate": 5.294779128225514e-07, + "loss": 0.7419, + "step": 16355 + }, + { + "epoch": 1.393779292714103, + "grad_norm": 19.25, + "learning_rate": 5.293406082042545e-07, + "loss": 0.9076, + "step": 16356 + }, + { + "epoch": 1.393864507882403, + "grad_norm": 17.0, + "learning_rate": 5.292033166089975e-07, + "loss": 0.8882, + "step": 16357 + }, + { + "epoch": 1.393949723050703, + "grad_norm": 18.375, + "learning_rate": 5.290660380392619e-07, + "loss": 0.843, + "step": 16358 + }, + { + "epoch": 1.394034938219003, + "grad_norm": 13.4375, + "learning_rate": 5.289287724975276e-07, + "loss": 0.5389, + "step": 16359 + }, + { + "epoch": 1.3941201533873029, + "grad_norm": 15.875, + "learning_rate": 5.287915199862753e-07, + "loss": 0.7784, + "step": 16360 + }, + { + "epoch": 1.3942053685556028, + "grad_norm": 17.375, + "learning_rate": 5.28654280507986e-07, + "loss": 0.8619, + "step": 16361 + }, + { + "epoch": 1.3942905837239028, + "grad_norm": 13.0625, + "learning_rate": 5.285170540651387e-07, + "loss": 0.5326, + "step": 16362 + }, + { + "epoch": 1.3943757988922028, + "grad_norm": 16.125, + "learning_rate": 5.283798406602142e-07, + "loss": 0.8422, + "step": 16363 + }, + { + "epoch": 1.3944610140605027, + "grad_norm": 18.375, + "learning_rate": 5.282426402956911e-07, + "loss": 0.6701, + "step": 16364 + }, + { + "epoch": 1.3945462292288027, + "grad_norm": 17.0, + "learning_rate": 5.281054529740494e-07, + "loss": 0.9546, + "step": 16365 + }, + { + "epoch": 1.3946314443971026, + "grad_norm": 15.5, + "learning_rate": 5.279682786977686e-07, + "loss": 0.549, + "step": 16366 + }, + { + "epoch": 1.3947166595654026, + "grad_norm": 18.875, + "learning_rate": 5.278311174693265e-07, + "loss": 0.808, + "step": 16367 + }, + { + "epoch": 1.3948018747337025, + "grad_norm": 18.875, + "learning_rate": 5.276939692912023e-07, + "loss": 0.6242, + "step": 16368 + }, + { + "epoch": 1.3948870899020025, + "grad_norm": 15.625, + "learning_rate": 5.275568341658749e-07, + "loss": 0.4896, + "step": 16369 + }, + { + "epoch": 1.3949723050703025, + "grad_norm": 15.125, + "learning_rate": 5.274197120958215e-07, + "loss": 0.7505, + "step": 16370 + }, + { + "epoch": 1.3950575202386024, + "grad_norm": 12.8125, + "learning_rate": 5.27282603083521e-07, + "loss": 0.4665, + "step": 16371 + }, + { + "epoch": 1.3951427354069024, + "grad_norm": 16.125, + "learning_rate": 5.271455071314502e-07, + "loss": 0.8909, + "step": 16372 + }, + { + "epoch": 1.3952279505752023, + "grad_norm": 13.375, + "learning_rate": 5.270084242420868e-07, + "loss": 0.438, + "step": 16373 + }, + { + "epoch": 1.3953131657435023, + "grad_norm": 13.6875, + "learning_rate": 5.268713544179084e-07, + "loss": 0.4716, + "step": 16374 + }, + { + "epoch": 1.3953983809118022, + "grad_norm": 17.0, + "learning_rate": 5.267342976613921e-07, + "loss": 0.8214, + "step": 16375 + }, + { + "epoch": 1.3954835960801022, + "grad_norm": 14.3125, + "learning_rate": 5.265972539750139e-07, + "loss": 0.5241, + "step": 16376 + }, + { + "epoch": 1.3955688112484022, + "grad_norm": 16.375, + "learning_rate": 5.264602233612514e-07, + "loss": 0.8926, + "step": 16377 + }, + { + "epoch": 1.395654026416702, + "grad_norm": 14.375, + "learning_rate": 5.263232058225795e-07, + "loss": 0.6483, + "step": 16378 + }, + { + "epoch": 1.395739241585002, + "grad_norm": 15.8125, + "learning_rate": 5.261862013614756e-07, + "loss": 0.4542, + "step": 16379 + }, + { + "epoch": 1.395824456753302, + "grad_norm": 18.0, + "learning_rate": 5.260492099804144e-07, + "loss": 0.6637, + "step": 16380 + }, + { + "epoch": 1.395909671921602, + "grad_norm": 17.5, + "learning_rate": 5.259122316818719e-07, + "loss": 0.5686, + "step": 16381 + }, + { + "epoch": 1.395994887089902, + "grad_norm": 15.3125, + "learning_rate": 5.257752664683237e-07, + "loss": 0.5135, + "step": 16382 + }, + { + "epoch": 1.396080102258202, + "grad_norm": 17.625, + "learning_rate": 5.25638314342245e-07, + "loss": 1.0912, + "step": 16383 + }, + { + "epoch": 1.3961653174265019, + "grad_norm": 17.875, + "learning_rate": 5.255013753061099e-07, + "loss": 0.6753, + "step": 16384 + }, + { + "epoch": 1.3962505325948018, + "grad_norm": 13.5625, + "learning_rate": 5.253644493623936e-07, + "loss": 0.4023, + "step": 16385 + }, + { + "epoch": 1.3963357477631018, + "grad_norm": 12.0, + "learning_rate": 5.252275365135708e-07, + "loss": 0.3412, + "step": 16386 + }, + { + "epoch": 1.3964209629314017, + "grad_norm": 18.625, + "learning_rate": 5.250906367621146e-07, + "loss": 0.9212, + "step": 16387 + }, + { + "epoch": 1.3965061780997017, + "grad_norm": 13.9375, + "learning_rate": 5.249537501105001e-07, + "loss": 0.6369, + "step": 16388 + }, + { + "epoch": 1.3965913932680016, + "grad_norm": 23.875, + "learning_rate": 5.248168765612e-07, + "loss": 1.1704, + "step": 16389 + }, + { + "epoch": 1.3966766084363016, + "grad_norm": 13.625, + "learning_rate": 5.246800161166882e-07, + "loss": 0.3576, + "step": 16390 + }, + { + "epoch": 1.3967618236046015, + "grad_norm": 19.25, + "learning_rate": 5.245431687794379e-07, + "loss": 0.692, + "step": 16391 + }, + { + "epoch": 1.3968470387729015, + "grad_norm": 13.1875, + "learning_rate": 5.244063345519223e-07, + "loss": 0.3199, + "step": 16392 + }, + { + "epoch": 1.3969322539412015, + "grad_norm": 16.25, + "learning_rate": 5.242695134366137e-07, + "loss": 0.3489, + "step": 16393 + }, + { + "epoch": 1.3970174691095014, + "grad_norm": 12.875, + "learning_rate": 5.24132705435985e-07, + "loss": 0.5171, + "step": 16394 + }, + { + "epoch": 1.3971026842778014, + "grad_norm": 14.1875, + "learning_rate": 5.239959105525079e-07, + "loss": 0.6713, + "step": 16395 + }, + { + "epoch": 1.3971878994461013, + "grad_norm": 13.25, + "learning_rate": 5.238591287886552e-07, + "loss": 0.4684, + "step": 16396 + }, + { + "epoch": 1.3972731146144013, + "grad_norm": 13.3125, + "learning_rate": 5.237223601468977e-07, + "loss": 0.2055, + "step": 16397 + }, + { + "epoch": 1.3973583297827012, + "grad_norm": 13.75, + "learning_rate": 5.235856046297076e-07, + "loss": 0.399, + "step": 16398 + }, + { + "epoch": 1.3974435449510012, + "grad_norm": 14.75, + "learning_rate": 5.23448862239556e-07, + "loss": 0.5487, + "step": 16399 + }, + { + "epoch": 1.3975287601193012, + "grad_norm": 27.125, + "learning_rate": 5.233121329789145e-07, + "loss": 0.4427, + "step": 16400 + }, + { + "epoch": 1.3976139752876011, + "grad_norm": 25.125, + "learning_rate": 5.231754168502532e-07, + "loss": 0.8255, + "step": 16401 + }, + { + "epoch": 1.397699190455901, + "grad_norm": 20.75, + "learning_rate": 5.230387138560433e-07, + "loss": 1.1072, + "step": 16402 + }, + { + "epoch": 1.397784405624201, + "grad_norm": 19.625, + "learning_rate": 5.229020239987544e-07, + "loss": 0.8599, + "step": 16403 + }, + { + "epoch": 1.397869620792501, + "grad_norm": 14.75, + "learning_rate": 5.22765347280857e-07, + "loss": 0.5858, + "step": 16404 + }, + { + "epoch": 1.397954835960801, + "grad_norm": 17.625, + "learning_rate": 5.226286837048218e-07, + "loss": 0.8556, + "step": 16405 + }, + { + "epoch": 1.398040051129101, + "grad_norm": 33.25, + "learning_rate": 5.224920332731171e-07, + "loss": 1.0935, + "step": 16406 + }, + { + "epoch": 1.3981252662974009, + "grad_norm": 13.0625, + "learning_rate": 5.223553959882129e-07, + "loss": 0.4323, + "step": 16407 + }, + { + "epoch": 1.3982104814657008, + "grad_norm": 13.25, + "learning_rate": 5.222187718525784e-07, + "loss": 0.5871, + "step": 16408 + }, + { + "epoch": 1.3982956966340008, + "grad_norm": 14.125, + "learning_rate": 5.220821608686829e-07, + "loss": 0.5311, + "step": 16409 + }, + { + "epoch": 1.3983809118023007, + "grad_norm": 26.625, + "learning_rate": 5.21945563038995e-07, + "loss": 0.7321, + "step": 16410 + }, + { + "epoch": 1.3984661269706007, + "grad_norm": 12.875, + "learning_rate": 5.218089783659822e-07, + "loss": 0.2922, + "step": 16411 + }, + { + "epoch": 1.3985513421389006, + "grad_norm": 17.25, + "learning_rate": 5.216724068521135e-07, + "loss": 0.7021, + "step": 16412 + }, + { + "epoch": 1.3986365573072006, + "grad_norm": 14.1875, + "learning_rate": 5.215358484998572e-07, + "loss": 0.6954, + "step": 16413 + }, + { + "epoch": 1.3987217724755006, + "grad_norm": 14.3125, + "learning_rate": 5.213993033116803e-07, + "loss": 0.4491, + "step": 16414 + }, + { + "epoch": 1.3988069876438005, + "grad_norm": 18.625, + "learning_rate": 5.212627712900505e-07, + "loss": 0.5117, + "step": 16415 + }, + { + "epoch": 1.3988922028121005, + "grad_norm": 17.25, + "learning_rate": 5.211262524374354e-07, + "loss": 0.8865, + "step": 16416 + }, + { + "epoch": 1.3989774179804004, + "grad_norm": 12.75, + "learning_rate": 5.209897467563021e-07, + "loss": 0.4027, + "step": 16417 + }, + { + "epoch": 1.3990626331487004, + "grad_norm": 13.625, + "learning_rate": 5.208532542491174e-07, + "loss": 0.5065, + "step": 16418 + }, + { + "epoch": 1.3991478483170003, + "grad_norm": 15.5, + "learning_rate": 5.20716774918347e-07, + "loss": 0.5038, + "step": 16419 + }, + { + "epoch": 1.3992330634853003, + "grad_norm": 12.6875, + "learning_rate": 5.205803087664577e-07, + "loss": 0.4637, + "step": 16420 + }, + { + "epoch": 1.3993182786536003, + "grad_norm": 16.125, + "learning_rate": 5.20443855795916e-07, + "loss": 0.5285, + "step": 16421 + }, + { + "epoch": 1.3994034938219002, + "grad_norm": 22.125, + "learning_rate": 5.203074160091877e-07, + "loss": 0.6923, + "step": 16422 + }, + { + "epoch": 1.3994887089902002, + "grad_norm": 18.125, + "learning_rate": 5.201709894087379e-07, + "loss": 0.5117, + "step": 16423 + }, + { + "epoch": 1.3995739241585001, + "grad_norm": 17.125, + "learning_rate": 5.20034575997032e-07, + "loss": 0.681, + "step": 16424 + }, + { + "epoch": 1.3996591393268, + "grad_norm": 10.3125, + "learning_rate": 5.198981757765355e-07, + "loss": 0.3168, + "step": 16425 + }, + { + "epoch": 1.3997443544951, + "grad_norm": 17.125, + "learning_rate": 5.197617887497139e-07, + "loss": 0.7524, + "step": 16426 + }, + { + "epoch": 1.3998295696634, + "grad_norm": 13.25, + "learning_rate": 5.196254149190303e-07, + "loss": 0.4455, + "step": 16427 + }, + { + "epoch": 1.3999147848317, + "grad_norm": 13.375, + "learning_rate": 5.194890542869499e-07, + "loss": 0.6791, + "step": 16428 + }, + { + "epoch": 1.4, + "grad_norm": 12.75, + "learning_rate": 5.193527068559368e-07, + "loss": 0.4361, + "step": 16429 + }, + { + "epoch": 1.4000852151682999, + "grad_norm": 12.375, + "learning_rate": 5.192163726284555e-07, + "loss": 0.3726, + "step": 16430 + }, + { + "epoch": 1.4001704303365998, + "grad_norm": 20.375, + "learning_rate": 5.190800516069687e-07, + "loss": 0.8526, + "step": 16431 + }, + { + "epoch": 1.4002556455048998, + "grad_norm": 16.625, + "learning_rate": 5.189437437939406e-07, + "loss": 0.664, + "step": 16432 + }, + { + "epoch": 1.4003408606731997, + "grad_norm": 21.375, + "learning_rate": 5.188074491918339e-07, + "loss": 1.0777, + "step": 16433 + }, + { + "epoch": 1.4004260758414997, + "grad_norm": 19.25, + "learning_rate": 5.186711678031124e-07, + "loss": 0.7175, + "step": 16434 + }, + { + "epoch": 1.4005112910097997, + "grad_norm": 18.875, + "learning_rate": 5.185348996302383e-07, + "loss": 1.1691, + "step": 16435 + }, + { + "epoch": 1.4005965061780996, + "grad_norm": 15.875, + "learning_rate": 5.183986446756738e-07, + "loss": 0.4921, + "step": 16436 + }, + { + "epoch": 1.4006817213463996, + "grad_norm": 13.3125, + "learning_rate": 5.182624029418813e-07, + "loss": 0.58, + "step": 16437 + }, + { + "epoch": 1.4007669365146995, + "grad_norm": 11.0625, + "learning_rate": 5.181261744313231e-07, + "loss": 0.3165, + "step": 16438 + }, + { + "epoch": 1.4008521516829995, + "grad_norm": 20.125, + "learning_rate": 5.179899591464613e-07, + "loss": 0.7726, + "step": 16439 + }, + { + "epoch": 1.4009373668512994, + "grad_norm": 11.9375, + "learning_rate": 5.178537570897565e-07, + "loss": 0.4885, + "step": 16440 + }, + { + "epoch": 1.4010225820195994, + "grad_norm": 13.6875, + "learning_rate": 5.177175682636707e-07, + "loss": 0.6742, + "step": 16441 + }, + { + "epoch": 1.4011077971878994, + "grad_norm": 13.9375, + "learning_rate": 5.175813926706651e-07, + "loss": 0.4466, + "step": 16442 + }, + { + "epoch": 1.4011930123561993, + "grad_norm": 16.5, + "learning_rate": 5.174452303132004e-07, + "loss": 0.5636, + "step": 16443 + }, + { + "epoch": 1.4012782275244993, + "grad_norm": 14.4375, + "learning_rate": 5.173090811937364e-07, + "loss": 0.7403, + "step": 16444 + }, + { + "epoch": 1.4013634426927992, + "grad_norm": 17.25, + "learning_rate": 5.171729453147343e-07, + "loss": 0.6367, + "step": 16445 + }, + { + "epoch": 1.4014486578610992, + "grad_norm": 14.75, + "learning_rate": 5.170368226786537e-07, + "loss": 0.5159, + "step": 16446 + }, + { + "epoch": 1.4015338730293991, + "grad_norm": 12.875, + "learning_rate": 5.169007132879554e-07, + "loss": 0.6055, + "step": 16447 + }, + { + "epoch": 1.401619088197699, + "grad_norm": 15.5, + "learning_rate": 5.16764617145098e-07, + "loss": 0.5841, + "step": 16448 + }, + { + "epoch": 1.401704303365999, + "grad_norm": 14.5, + "learning_rate": 5.166285342525411e-07, + "loss": 0.572, + "step": 16449 + }, + { + "epoch": 1.401789518534299, + "grad_norm": 12.125, + "learning_rate": 5.164924646127442e-07, + "loss": 0.5027, + "step": 16450 + }, + { + "epoch": 1.401874733702599, + "grad_norm": 19.0, + "learning_rate": 5.163564082281665e-07, + "loss": 0.8262, + "step": 16451 + }, + { + "epoch": 1.401959948870899, + "grad_norm": 10.875, + "learning_rate": 5.162203651012662e-07, + "loss": 0.2286, + "step": 16452 + }, + { + "epoch": 1.4020451640391989, + "grad_norm": 14.125, + "learning_rate": 5.160843352345013e-07, + "loss": 0.4557, + "step": 16453 + }, + { + "epoch": 1.4021303792074988, + "grad_norm": 17.5, + "learning_rate": 5.159483186303306e-07, + "loss": 0.698, + "step": 16454 + }, + { + "epoch": 1.4022155943757988, + "grad_norm": 17.25, + "learning_rate": 5.158123152912118e-07, + "loss": 0.5511, + "step": 16455 + }, + { + "epoch": 1.4023008095440987, + "grad_norm": 16.125, + "learning_rate": 5.156763252196033e-07, + "loss": 0.5065, + "step": 16456 + }, + { + "epoch": 1.4023860247123987, + "grad_norm": 23.0, + "learning_rate": 5.155403484179615e-07, + "loss": 0.5213, + "step": 16457 + }, + { + "epoch": 1.4024712398806987, + "grad_norm": 14.3125, + "learning_rate": 5.154043848887441e-07, + "loss": 0.4562, + "step": 16458 + }, + { + "epoch": 1.4025564550489986, + "grad_norm": 11.6875, + "learning_rate": 5.152684346344087e-07, + "loss": 0.3337, + "step": 16459 + }, + { + "epoch": 1.4026416702172986, + "grad_norm": 15.25, + "learning_rate": 5.151324976574115e-07, + "loss": 0.6697, + "step": 16460 + }, + { + "epoch": 1.4027268853855985, + "grad_norm": 14.625, + "learning_rate": 5.149965739602086e-07, + "loss": 0.5686, + "step": 16461 + }, + { + "epoch": 1.4028121005538985, + "grad_norm": 11.5625, + "learning_rate": 5.148606635452568e-07, + "loss": 0.453, + "step": 16462 + }, + { + "epoch": 1.4028973157221984, + "grad_norm": 29.125, + "learning_rate": 5.147247664150118e-07, + "loss": 0.6343, + "step": 16463 + }, + { + "epoch": 1.4029825308904984, + "grad_norm": 14.625, + "learning_rate": 5.145888825719303e-07, + "loss": 0.5851, + "step": 16464 + }, + { + "epoch": 1.4030677460587984, + "grad_norm": 17.125, + "learning_rate": 5.144530120184667e-07, + "loss": 0.461, + "step": 16465 + }, + { + "epoch": 1.4031529612270983, + "grad_norm": 13.8125, + "learning_rate": 5.143171547570767e-07, + "loss": 0.3404, + "step": 16466 + }, + { + "epoch": 1.4032381763953983, + "grad_norm": 19.375, + "learning_rate": 5.141813107902162e-07, + "loss": 1.0303, + "step": 16467 + }, + { + "epoch": 1.4033233915636982, + "grad_norm": 17.75, + "learning_rate": 5.140454801203387e-07, + "loss": 0.6771, + "step": 16468 + }, + { + "epoch": 1.4034086067319982, + "grad_norm": 14.625, + "learning_rate": 5.139096627498999e-07, + "loss": 0.4882, + "step": 16469 + }, + { + "epoch": 1.4034938219002981, + "grad_norm": 17.375, + "learning_rate": 5.137738586813533e-07, + "loss": 0.7031, + "step": 16470 + }, + { + "epoch": 1.403579037068598, + "grad_norm": 40.5, + "learning_rate": 5.136380679171534e-07, + "loss": 1.2443, + "step": 16471 + }, + { + "epoch": 1.403664252236898, + "grad_norm": 13.625, + "learning_rate": 5.135022904597544e-07, + "loss": 0.4532, + "step": 16472 + }, + { + "epoch": 1.403749467405198, + "grad_norm": 19.875, + "learning_rate": 5.133665263116094e-07, + "loss": 0.4926, + "step": 16473 + }, + { + "epoch": 1.403834682573498, + "grad_norm": 25.25, + "learning_rate": 5.132307754751719e-07, + "loss": 0.9242, + "step": 16474 + }, + { + "epoch": 1.403919897741798, + "grad_norm": 16.125, + "learning_rate": 5.130950379528956e-07, + "loss": 0.8087, + "step": 16475 + }, + { + "epoch": 1.4040051129100979, + "grad_norm": 19.125, + "learning_rate": 5.129593137472326e-07, + "loss": 1.1997, + "step": 16476 + }, + { + "epoch": 1.4040903280783978, + "grad_norm": 17.75, + "learning_rate": 5.128236028606363e-07, + "loss": 0.9101, + "step": 16477 + }, + { + "epoch": 1.4041755432466978, + "grad_norm": 15.1875, + "learning_rate": 5.126879052955584e-07, + "loss": 0.6631, + "step": 16478 + }, + { + "epoch": 1.4042607584149978, + "grad_norm": 15.0625, + "learning_rate": 5.125522210544515e-07, + "loss": 0.5251, + "step": 16479 + }, + { + "epoch": 1.4043459735832977, + "grad_norm": 16.5, + "learning_rate": 5.124165501397674e-07, + "loss": 0.5776, + "step": 16480 + }, + { + "epoch": 1.4044311887515977, + "grad_norm": 21.75, + "learning_rate": 5.122808925539585e-07, + "loss": 0.772, + "step": 16481 + }, + { + "epoch": 1.4045164039198976, + "grad_norm": 14.1875, + "learning_rate": 5.121452482994751e-07, + "loss": 0.5172, + "step": 16482 + }, + { + "epoch": 1.4046016190881976, + "grad_norm": 12.1875, + "learning_rate": 5.120096173787696e-07, + "loss": 0.7046, + "step": 16483 + }, + { + "epoch": 1.4046868342564975, + "grad_norm": 19.375, + "learning_rate": 5.118739997942921e-07, + "loss": 1.0487, + "step": 16484 + }, + { + "epoch": 1.4047720494247975, + "grad_norm": 15.5, + "learning_rate": 5.117383955484935e-07, + "loss": 0.7468, + "step": 16485 + }, + { + "epoch": 1.4048572645930975, + "grad_norm": 13.625, + "learning_rate": 5.11602804643825e-07, + "loss": 0.6507, + "step": 16486 + }, + { + "epoch": 1.4049424797613974, + "grad_norm": 12.4375, + "learning_rate": 5.114672270827358e-07, + "loss": 0.4388, + "step": 16487 + }, + { + "epoch": 1.4050276949296974, + "grad_norm": 16.25, + "learning_rate": 5.113316628676765e-07, + "loss": 0.5757, + "step": 16488 + }, + { + "epoch": 1.4051129100979973, + "grad_norm": 19.25, + "learning_rate": 5.111961120010974e-07, + "loss": 0.7725, + "step": 16489 + }, + { + "epoch": 1.4051981252662973, + "grad_norm": 18.875, + "learning_rate": 5.110605744854469e-07, + "loss": 0.7043, + "step": 16490 + }, + { + "epoch": 1.4052833404345972, + "grad_norm": 17.125, + "learning_rate": 5.109250503231752e-07, + "loss": 0.8088, + "step": 16491 + }, + { + "epoch": 1.4053685556028972, + "grad_norm": 18.125, + "learning_rate": 5.107895395167309e-07, + "loss": 0.7808, + "step": 16492 + }, + { + "epoch": 1.4054537707711972, + "grad_norm": 18.75, + "learning_rate": 5.106540420685629e-07, + "loss": 0.8131, + "step": 16493 + }, + { + "epoch": 1.405538985939497, + "grad_norm": 15.75, + "learning_rate": 5.105185579811201e-07, + "loss": 0.7294, + "step": 16494 + }, + { + "epoch": 1.405624201107797, + "grad_norm": 15.0, + "learning_rate": 5.103830872568501e-07, + "loss": 0.7818, + "step": 16495 + }, + { + "epoch": 1.405709416276097, + "grad_norm": 12.9375, + "learning_rate": 5.102476298982017e-07, + "loss": 0.4041, + "step": 16496 + }, + { + "epoch": 1.405794631444397, + "grad_norm": 15.9375, + "learning_rate": 5.101121859076222e-07, + "loss": 0.7404, + "step": 16497 + }, + { + "epoch": 1.405879846612697, + "grad_norm": 17.5, + "learning_rate": 5.099767552875602e-07, + "loss": 0.6422, + "step": 16498 + }, + { + "epoch": 1.405965061780997, + "grad_norm": 15.625, + "learning_rate": 5.098413380404622e-07, + "loss": 0.5735, + "step": 16499 + }, + { + "epoch": 1.4060502769492969, + "grad_norm": 17.125, + "learning_rate": 5.097059341687751e-07, + "loss": 0.5003, + "step": 16500 + }, + { + "epoch": 1.4061354921175968, + "grad_norm": 14.4375, + "learning_rate": 5.095705436749464e-07, + "loss": 0.4601, + "step": 16501 + }, + { + "epoch": 1.4062207072858968, + "grad_norm": 16.625, + "learning_rate": 5.094351665614228e-07, + "loss": 0.8413, + "step": 16502 + }, + { + "epoch": 1.4063059224541967, + "grad_norm": 16.125, + "learning_rate": 5.0929980283065e-07, + "loss": 0.8693, + "step": 16503 + }, + { + "epoch": 1.4063911376224967, + "grad_norm": 11.9375, + "learning_rate": 5.091644524850747e-07, + "loss": 0.2907, + "step": 16504 + }, + { + "epoch": 1.4064763527907966, + "grad_norm": 14.8125, + "learning_rate": 5.090291155271427e-07, + "loss": 0.4796, + "step": 16505 + }, + { + "epoch": 1.4065615679590966, + "grad_norm": 17.875, + "learning_rate": 5.088937919593e-07, + "loss": 0.7473, + "step": 16506 + }, + { + "epoch": 1.4066467831273965, + "grad_norm": 33.5, + "learning_rate": 5.087584817839917e-07, + "loss": 0.7585, + "step": 16507 + }, + { + "epoch": 1.4067319982956965, + "grad_norm": 15.0, + "learning_rate": 5.086231850036627e-07, + "loss": 0.6283, + "step": 16508 + }, + { + "epoch": 1.4068172134639965, + "grad_norm": 16.75, + "learning_rate": 5.084879016207582e-07, + "loss": 0.5981, + "step": 16509 + }, + { + "epoch": 1.4069024286322964, + "grad_norm": 10.5, + "learning_rate": 5.083526316377229e-07, + "loss": 0.8238, + "step": 16510 + }, + { + "epoch": 1.4069876438005964, + "grad_norm": 18.0, + "learning_rate": 5.08217375057002e-07, + "loss": 0.6106, + "step": 16511 + }, + { + "epoch": 1.4070728589688963, + "grad_norm": 16.5, + "learning_rate": 5.080821318810385e-07, + "loss": 0.8085, + "step": 16512 + }, + { + "epoch": 1.4071580741371963, + "grad_norm": 16.5, + "learning_rate": 5.079469021122768e-07, + "loss": 0.7634, + "step": 16513 + }, + { + "epoch": 1.4072432893054962, + "grad_norm": 23.375, + "learning_rate": 5.078116857531609e-07, + "loss": 0.711, + "step": 16514 + }, + { + "epoch": 1.4073285044737962, + "grad_norm": 14.125, + "learning_rate": 5.076764828061345e-07, + "loss": 0.5395, + "step": 16515 + }, + { + "epoch": 1.4074137196420962, + "grad_norm": 17.125, + "learning_rate": 5.075412932736406e-07, + "loss": 0.4234, + "step": 16516 + }, + { + "epoch": 1.4074989348103961, + "grad_norm": 12.375, + "learning_rate": 5.074061171581216e-07, + "loss": 0.4247, + "step": 16517 + }, + { + "epoch": 1.407584149978696, + "grad_norm": 11.9375, + "learning_rate": 5.072709544620211e-07, + "loss": 0.2638, + "step": 16518 + }, + { + "epoch": 1.407669365146996, + "grad_norm": 21.875, + "learning_rate": 5.071358051877814e-07, + "loss": 1.008, + "step": 16519 + }, + { + "epoch": 1.407754580315296, + "grad_norm": 35.5, + "learning_rate": 5.070006693378445e-07, + "loss": 1.0408, + "step": 16520 + }, + { + "epoch": 1.407839795483596, + "grad_norm": 14.4375, + "learning_rate": 5.068655469146526e-07, + "loss": 0.5524, + "step": 16521 + }, + { + "epoch": 1.407925010651896, + "grad_norm": 15.5625, + "learning_rate": 5.067304379206473e-07, + "loss": 0.3218, + "step": 16522 + }, + { + "epoch": 1.4080102258201959, + "grad_norm": 17.875, + "learning_rate": 5.065953423582712e-07, + "loss": 0.8556, + "step": 16523 + }, + { + "epoch": 1.4080954409884958, + "grad_norm": 22.75, + "learning_rate": 5.064602602299647e-07, + "loss": 0.9158, + "step": 16524 + }, + { + "epoch": 1.4081806561567958, + "grad_norm": 15.375, + "learning_rate": 5.063251915381683e-07, + "loss": 0.4169, + "step": 16525 + }, + { + "epoch": 1.4082658713250957, + "grad_norm": 14.75, + "learning_rate": 5.061901362853239e-07, + "loss": 0.988, + "step": 16526 + }, + { + "epoch": 1.4083510864933957, + "grad_norm": 18.625, + "learning_rate": 5.060550944738715e-07, + "loss": 0.7488, + "step": 16527 + }, + { + "epoch": 1.4084363016616956, + "grad_norm": 11.125, + "learning_rate": 5.059200661062522e-07, + "loss": 0.3206, + "step": 16528 + }, + { + "epoch": 1.4085215168299956, + "grad_norm": 19.125, + "learning_rate": 5.057850511849049e-07, + "loss": 0.882, + "step": 16529 + }, + { + "epoch": 1.4086067319982956, + "grad_norm": 16.75, + "learning_rate": 5.056500497122702e-07, + "loss": 0.6264, + "step": 16530 + }, + { + "epoch": 1.4086919471665957, + "grad_norm": 28.375, + "learning_rate": 5.055150616907875e-07, + "loss": 0.8639, + "step": 16531 + }, + { + "epoch": 1.4087771623348957, + "grad_norm": 10.8125, + "learning_rate": 5.053800871228969e-07, + "loss": 0.3065, + "step": 16532 + }, + { + "epoch": 1.4088623775031957, + "grad_norm": 14.6875, + "learning_rate": 5.052451260110368e-07, + "loss": 0.3775, + "step": 16533 + }, + { + "epoch": 1.4089475926714956, + "grad_norm": 21.0, + "learning_rate": 5.051101783576456e-07, + "loss": 0.4998, + "step": 16534 + }, + { + "epoch": 1.4090328078397956, + "grad_norm": 12.3125, + "learning_rate": 5.049752441651628e-07, + "loss": 0.4299, + "step": 16535 + }, + { + "epoch": 1.4091180230080955, + "grad_norm": 15.8125, + "learning_rate": 5.048403234360268e-07, + "loss": 0.4584, + "step": 16536 + }, + { + "epoch": 1.4092032381763955, + "grad_norm": 13.25, + "learning_rate": 5.047054161726749e-07, + "loss": 0.4749, + "step": 16537 + }, + { + "epoch": 1.4092884533446954, + "grad_norm": 22.375, + "learning_rate": 5.045705223775459e-07, + "loss": 0.7882, + "step": 16538 + }, + { + "epoch": 1.4093736685129954, + "grad_norm": 19.875, + "learning_rate": 5.044356420530769e-07, + "loss": 0.787, + "step": 16539 + }, + { + "epoch": 1.4094588836812953, + "grad_norm": 10.8125, + "learning_rate": 5.043007752017061e-07, + "loss": 0.2653, + "step": 16540 + }, + { + "epoch": 1.4095440988495953, + "grad_norm": 17.875, + "learning_rate": 5.041659218258702e-07, + "loss": 0.6112, + "step": 16541 + }, + { + "epoch": 1.4096293140178953, + "grad_norm": 13.25, + "learning_rate": 5.040310819280056e-07, + "loss": 0.6303, + "step": 16542 + }, + { + "epoch": 1.4097145291861952, + "grad_norm": 12.875, + "learning_rate": 5.038962555105493e-07, + "loss": 0.3604, + "step": 16543 + }, + { + "epoch": 1.4097997443544952, + "grad_norm": 12.25, + "learning_rate": 5.037614425759382e-07, + "loss": 0.3688, + "step": 16544 + }, + { + "epoch": 1.4098849595227951, + "grad_norm": 13.8125, + "learning_rate": 5.036266431266086e-07, + "loss": 0.3841, + "step": 16545 + }, + { + "epoch": 1.409970174691095, + "grad_norm": 15.6875, + "learning_rate": 5.034918571649956e-07, + "loss": 0.7882, + "step": 16546 + }, + { + "epoch": 1.410055389859395, + "grad_norm": 66.0, + "learning_rate": 5.033570846935354e-07, + "loss": 0.6905, + "step": 16547 + }, + { + "epoch": 1.410140605027695, + "grad_norm": 12.625, + "learning_rate": 5.03222325714664e-07, + "loss": 0.4332, + "step": 16548 + }, + { + "epoch": 1.410225820195995, + "grad_norm": 10.625, + "learning_rate": 5.030875802308161e-07, + "loss": 0.3606, + "step": 16549 + }, + { + "epoch": 1.410311035364295, + "grad_norm": 19.375, + "learning_rate": 5.029528482444262e-07, + "loss": 0.6726, + "step": 16550 + }, + { + "epoch": 1.4103962505325949, + "grad_norm": 29.25, + "learning_rate": 5.028181297579296e-07, + "loss": 0.946, + "step": 16551 + }, + { + "epoch": 1.4104814657008948, + "grad_norm": 17.0, + "learning_rate": 5.026834247737608e-07, + "loss": 0.7243, + "step": 16552 + }, + { + "epoch": 1.4105666808691948, + "grad_norm": 22.875, + "learning_rate": 5.025487332943543e-07, + "loss": 0.6982, + "step": 16553 + }, + { + "epoch": 1.4106518960374947, + "grad_norm": 15.125, + "learning_rate": 5.024140553221435e-07, + "loss": 0.7655, + "step": 16554 + }, + { + "epoch": 1.4107371112057947, + "grad_norm": 15.75, + "learning_rate": 5.022793908595625e-07, + "loss": 0.6237, + "step": 16555 + }, + { + "epoch": 1.4108223263740947, + "grad_norm": 13.25, + "learning_rate": 5.021447399090451e-07, + "loss": 0.5613, + "step": 16556 + }, + { + "epoch": 1.4109075415423946, + "grad_norm": 19.375, + "learning_rate": 5.020101024730238e-07, + "loss": 1.0481, + "step": 16557 + }, + { + "epoch": 1.4109927567106946, + "grad_norm": 13.8125, + "learning_rate": 5.018754785539328e-07, + "loss": 0.4068, + "step": 16558 + }, + { + "epoch": 1.4110779718789945, + "grad_norm": 14.9375, + "learning_rate": 5.017408681542035e-07, + "loss": 0.5206, + "step": 16559 + }, + { + "epoch": 1.4111631870472945, + "grad_norm": 21.625, + "learning_rate": 5.016062712762693e-07, + "loss": 0.8158, + "step": 16560 + }, + { + "epoch": 1.4112484022155944, + "grad_norm": 24.125, + "learning_rate": 5.014716879225623e-07, + "loss": 0.6583, + "step": 16561 + }, + { + "epoch": 1.4113336173838944, + "grad_norm": 18.75, + "learning_rate": 5.01337118095515e-07, + "loss": 0.6697, + "step": 16562 + }, + { + "epoch": 1.4114188325521944, + "grad_norm": 16.5, + "learning_rate": 5.012025617975584e-07, + "loss": 0.4441, + "step": 16563 + }, + { + "epoch": 1.4115040477204943, + "grad_norm": 11.5, + "learning_rate": 5.01068019031125e-07, + "loss": 1.2402, + "step": 16564 + }, + { + "epoch": 1.4115892628887943, + "grad_norm": 16.0, + "learning_rate": 5.009334897986451e-07, + "loss": 0.5753, + "step": 16565 + }, + { + "epoch": 1.4116744780570942, + "grad_norm": 16.375, + "learning_rate": 5.007989741025507e-07, + "loss": 0.9699, + "step": 16566 + }, + { + "epoch": 1.4117596932253942, + "grad_norm": 13.5, + "learning_rate": 5.006644719452717e-07, + "loss": 0.5311, + "step": 16567 + }, + { + "epoch": 1.4118449083936941, + "grad_norm": 20.5, + "learning_rate": 5.005299833292394e-07, + "loss": 0.8108, + "step": 16568 + }, + { + "epoch": 1.411930123561994, + "grad_norm": 16.25, + "learning_rate": 5.003955082568838e-07, + "loss": 0.7582, + "step": 16569 + }, + { + "epoch": 1.412015338730294, + "grad_norm": 15.9375, + "learning_rate": 5.002610467306357e-07, + "loss": 0.5584, + "step": 16570 + }, + { + "epoch": 1.412100553898594, + "grad_norm": 23.125, + "learning_rate": 5.00126598752924e-07, + "loss": 0.489, + "step": 16571 + }, + { + "epoch": 1.412185769066894, + "grad_norm": 18.25, + "learning_rate": 4.999921643261791e-07, + "loss": 0.8342, + "step": 16572 + }, + { + "epoch": 1.412270984235194, + "grad_norm": 13.25, + "learning_rate": 4.998577434528294e-07, + "loss": 0.2673, + "step": 16573 + }, + { + "epoch": 1.4123561994034939, + "grad_norm": 15.1875, + "learning_rate": 4.997233361353047e-07, + "loss": 0.7502, + "step": 16574 + }, + { + "epoch": 1.4124414145717938, + "grad_norm": 19.5, + "learning_rate": 4.995889423760343e-07, + "loss": 0.5668, + "step": 16575 + }, + { + "epoch": 1.4125266297400938, + "grad_norm": 18.625, + "learning_rate": 4.994545621774457e-07, + "loss": 0.6787, + "step": 16576 + }, + { + "epoch": 1.4126118449083938, + "grad_norm": 20.625, + "learning_rate": 4.993201955419681e-07, + "loss": 1.0089, + "step": 16577 + }, + { + "epoch": 1.4126970600766937, + "grad_norm": 12.5, + "learning_rate": 4.991858424720293e-07, + "loss": 0.4901, + "step": 16578 + }, + { + "epoch": 1.4127822752449937, + "grad_norm": 16.875, + "learning_rate": 4.990515029700578e-07, + "loss": 0.5484, + "step": 16579 + }, + { + "epoch": 1.4128674904132936, + "grad_norm": 15.375, + "learning_rate": 4.989171770384809e-07, + "loss": 0.4544, + "step": 16580 + }, + { + "epoch": 1.4129527055815936, + "grad_norm": 11.3125, + "learning_rate": 4.987828646797252e-07, + "loss": 0.2709, + "step": 16581 + }, + { + "epoch": 1.4130379207498935, + "grad_norm": 15.5, + "learning_rate": 4.986485658962186e-07, + "loss": 0.6004, + "step": 16582 + }, + { + "epoch": 1.4131231359181935, + "grad_norm": 15.9375, + "learning_rate": 4.985142806903885e-07, + "loss": 0.3834, + "step": 16583 + }, + { + "epoch": 1.4132083510864935, + "grad_norm": 17.375, + "learning_rate": 4.983800090646607e-07, + "loss": 0.5642, + "step": 16584 + }, + { + "epoch": 1.4132935662547934, + "grad_norm": 10.875, + "learning_rate": 4.982457510214617e-07, + "loss": 0.4225, + "step": 16585 + }, + { + "epoch": 1.4133787814230934, + "grad_norm": 19.5, + "learning_rate": 4.98111506563218e-07, + "loss": 0.7194, + "step": 16586 + }, + { + "epoch": 1.4134639965913933, + "grad_norm": 14.1875, + "learning_rate": 4.979772756923559e-07, + "loss": 0.6912, + "step": 16587 + }, + { + "epoch": 1.4135492117596933, + "grad_norm": 16.5, + "learning_rate": 4.978430584113003e-07, + "loss": 0.963, + "step": 16588 + }, + { + "epoch": 1.4136344269279932, + "grad_norm": 18.625, + "learning_rate": 4.977088547224773e-07, + "loss": 0.6714, + "step": 16589 + }, + { + "epoch": 1.4137196420962932, + "grad_norm": 13.25, + "learning_rate": 4.975746646283114e-07, + "loss": 0.5975, + "step": 16590 + }, + { + "epoch": 1.4138048572645932, + "grad_norm": 14.625, + "learning_rate": 4.97440488131228e-07, + "loss": 0.3642, + "step": 16591 + }, + { + "epoch": 1.413890072432893, + "grad_norm": 25.0, + "learning_rate": 4.973063252336521e-07, + "loss": 1.146, + "step": 16592 + }, + { + "epoch": 1.413975287601193, + "grad_norm": 10.75, + "learning_rate": 4.971721759380073e-07, + "loss": 0.416, + "step": 16593 + }, + { + "epoch": 1.414060502769493, + "grad_norm": 21.25, + "learning_rate": 4.970380402467185e-07, + "loss": 1.0392, + "step": 16594 + }, + { + "epoch": 1.414145717937793, + "grad_norm": 23.375, + "learning_rate": 4.969039181622093e-07, + "loss": 0.6082, + "step": 16595 + }, + { + "epoch": 1.414230933106093, + "grad_norm": 15.0625, + "learning_rate": 4.96769809686904e-07, + "loss": 0.9213, + "step": 16596 + }, + { + "epoch": 1.414316148274393, + "grad_norm": 10.5, + "learning_rate": 4.966357148232258e-07, + "loss": 0.241, + "step": 16597 + }, + { + "epoch": 1.4144013634426928, + "grad_norm": 24.5, + "learning_rate": 4.965016335735973e-07, + "loss": 0.4489, + "step": 16598 + }, + { + "epoch": 1.4144865786109928, + "grad_norm": 10.9375, + "learning_rate": 4.963675659404419e-07, + "loss": 0.2269, + "step": 16599 + }, + { + "epoch": 1.4145717937792928, + "grad_norm": 18.125, + "learning_rate": 4.962335119261831e-07, + "loss": 0.7874, + "step": 16600 + }, + { + "epoch": 1.4146570089475927, + "grad_norm": 17.625, + "learning_rate": 4.960994715332421e-07, + "loss": 0.6893, + "step": 16601 + }, + { + "epoch": 1.4147422241158927, + "grad_norm": 22.125, + "learning_rate": 4.959654447640417e-07, + "loss": 0.4085, + "step": 16602 + }, + { + "epoch": 1.4148274392841926, + "grad_norm": 15.9375, + "learning_rate": 4.958314316210041e-07, + "loss": 0.4995, + "step": 16603 + }, + { + "epoch": 1.4149126544524926, + "grad_norm": 18.625, + "learning_rate": 4.956974321065512e-07, + "loss": 0.9185, + "step": 16604 + }, + { + "epoch": 1.4149978696207925, + "grad_norm": 13.3125, + "learning_rate": 4.955634462231043e-07, + "loss": 0.4122, + "step": 16605 + }, + { + "epoch": 1.4150830847890925, + "grad_norm": 18.0, + "learning_rate": 4.954294739730842e-07, + "loss": 0.8476, + "step": 16606 + }, + { + "epoch": 1.4151682999573925, + "grad_norm": 23.0, + "learning_rate": 4.952955153589122e-07, + "loss": 0.7513, + "step": 16607 + }, + { + "epoch": 1.4152535151256924, + "grad_norm": 14.375, + "learning_rate": 4.951615703830091e-07, + "loss": 0.4624, + "step": 16608 + }, + { + "epoch": 1.4153387302939924, + "grad_norm": 15.3125, + "learning_rate": 4.95027639047796e-07, + "loss": 0.4218, + "step": 16609 + }, + { + "epoch": 1.4154239454622923, + "grad_norm": 17.125, + "learning_rate": 4.948937213556921e-07, + "loss": 0.8306, + "step": 16610 + }, + { + "epoch": 1.4155091606305923, + "grad_norm": 23.25, + "learning_rate": 4.94759817309118e-07, + "loss": 0.7987, + "step": 16611 + }, + { + "epoch": 1.4155943757988922, + "grad_norm": 11.375, + "learning_rate": 4.946259269104941e-07, + "loss": 0.2166, + "step": 16612 + }, + { + "epoch": 1.4156795909671922, + "grad_norm": 17.75, + "learning_rate": 4.944920501622392e-07, + "loss": 0.7787, + "step": 16613 + }, + { + "epoch": 1.4157648061354922, + "grad_norm": 12.3125, + "learning_rate": 4.94358187066772e-07, + "loss": 0.3062, + "step": 16614 + }, + { + "epoch": 1.4158500213037921, + "grad_norm": 11.5625, + "learning_rate": 4.942243376265124e-07, + "loss": 0.2991, + "step": 16615 + }, + { + "epoch": 1.415935236472092, + "grad_norm": 34.75, + "learning_rate": 4.940905018438789e-07, + "loss": 0.7591, + "step": 16616 + }, + { + "epoch": 1.416020451640392, + "grad_norm": 14.0625, + "learning_rate": 4.939566797212908e-07, + "loss": 0.3228, + "step": 16617 + }, + { + "epoch": 1.416105666808692, + "grad_norm": 12.9375, + "learning_rate": 4.938228712611652e-07, + "loss": 0.5307, + "step": 16618 + }, + { + "epoch": 1.416190881976992, + "grad_norm": 12.0, + "learning_rate": 4.936890764659206e-07, + "loss": 0.3378, + "step": 16619 + }, + { + "epoch": 1.416276097145292, + "grad_norm": 12.625, + "learning_rate": 4.935552953379751e-07, + "loss": 0.3132, + "step": 16620 + }, + { + "epoch": 1.4163613123135919, + "grad_norm": 14.875, + "learning_rate": 4.934215278797465e-07, + "loss": 0.5478, + "step": 16621 + }, + { + "epoch": 1.4164465274818918, + "grad_norm": 13.25, + "learning_rate": 4.932877740936516e-07, + "loss": 0.2866, + "step": 16622 + }, + { + "epoch": 1.4165317426501918, + "grad_norm": 18.625, + "learning_rate": 4.931540339821071e-07, + "loss": 0.7415, + "step": 16623 + }, + { + "epoch": 1.4166169578184917, + "grad_norm": 15.9375, + "learning_rate": 4.930203075475304e-07, + "loss": 0.7254, + "step": 16624 + }, + { + "epoch": 1.4167021729867917, + "grad_norm": 18.25, + "learning_rate": 4.928865947923379e-07, + "loss": 0.9204, + "step": 16625 + }, + { + "epoch": 1.4167873881550916, + "grad_norm": 20.625, + "learning_rate": 4.927528957189463e-07, + "loss": 1.0934, + "step": 16626 + }, + { + "epoch": 1.4168726033233916, + "grad_norm": 16.5, + "learning_rate": 4.92619210329771e-07, + "loss": 0.6441, + "step": 16627 + }, + { + "epoch": 1.4169578184916916, + "grad_norm": 15.5625, + "learning_rate": 4.924855386272281e-07, + "loss": 0.4072, + "step": 16628 + }, + { + "epoch": 1.4170430336599915, + "grad_norm": 15.9375, + "learning_rate": 4.923518806137338e-07, + "loss": 0.6813, + "step": 16629 + }, + { + "epoch": 1.4171282488282915, + "grad_norm": 16.75, + "learning_rate": 4.922182362917029e-07, + "loss": 0.6055, + "step": 16630 + }, + { + "epoch": 1.4172134639965914, + "grad_norm": 16.125, + "learning_rate": 4.9208460566355e-07, + "loss": 0.6539, + "step": 16631 + }, + { + "epoch": 1.4172986791648914, + "grad_norm": 22.25, + "learning_rate": 4.919509887316903e-07, + "loss": 0.866, + "step": 16632 + }, + { + "epoch": 1.4173838943331913, + "grad_norm": 16.5, + "learning_rate": 4.918173854985388e-07, + "loss": 0.5245, + "step": 16633 + }, + { + "epoch": 1.4174691095014913, + "grad_norm": 15.25, + "learning_rate": 4.916837959665097e-07, + "loss": 0.6661, + "step": 16634 + }, + { + "epoch": 1.4175543246697913, + "grad_norm": 17.0, + "learning_rate": 4.915502201380166e-07, + "loss": 0.4548, + "step": 16635 + }, + { + "epoch": 1.4176395398380912, + "grad_norm": 10.75, + "learning_rate": 4.914166580154737e-07, + "loss": 0.7864, + "step": 16636 + }, + { + "epoch": 1.4177247550063912, + "grad_norm": 11.75, + "learning_rate": 4.912831096012951e-07, + "loss": 0.2432, + "step": 16637 + }, + { + "epoch": 1.4178099701746911, + "grad_norm": 19.0, + "learning_rate": 4.911495748978932e-07, + "loss": 0.8257, + "step": 16638 + }, + { + "epoch": 1.417895185342991, + "grad_norm": 14.6875, + "learning_rate": 4.91016053907682e-07, + "loss": 0.7638, + "step": 16639 + }, + { + "epoch": 1.417980400511291, + "grad_norm": 12.375, + "learning_rate": 4.908825466330735e-07, + "loss": 0.3615, + "step": 16640 + }, + { + "epoch": 1.418065615679591, + "grad_norm": 14.25, + "learning_rate": 4.907490530764808e-07, + "loss": 0.3604, + "step": 16641 + }, + { + "epoch": 1.418150830847891, + "grad_norm": 17.5, + "learning_rate": 4.906155732403164e-07, + "loss": 0.5031, + "step": 16642 + }, + { + "epoch": 1.418236046016191, + "grad_norm": 11.125, + "learning_rate": 4.90482107126992e-07, + "loss": 0.2802, + "step": 16643 + }, + { + "epoch": 1.4183212611844909, + "grad_norm": 21.25, + "learning_rate": 4.903486547389195e-07, + "loss": 0.5741, + "step": 16644 + }, + { + "epoch": 1.4184064763527908, + "grad_norm": 15.3125, + "learning_rate": 4.902152160785113e-07, + "loss": 0.513, + "step": 16645 + }, + { + "epoch": 1.4184916915210908, + "grad_norm": 17.625, + "learning_rate": 4.900817911481777e-07, + "loss": 0.6645, + "step": 16646 + }, + { + "epoch": 1.4185769066893907, + "grad_norm": 40.25, + "learning_rate": 4.899483799503307e-07, + "loss": 0.9492, + "step": 16647 + }, + { + "epoch": 1.4186621218576907, + "grad_norm": 14.1875, + "learning_rate": 4.898149824873803e-07, + "loss": 0.4754, + "step": 16648 + }, + { + "epoch": 1.4187473370259907, + "grad_norm": 17.875, + "learning_rate": 4.896815987617376e-07, + "loss": 0.8589, + "step": 16649 + }, + { + "epoch": 1.4188325521942906, + "grad_norm": 19.125, + "learning_rate": 4.89548228775813e-07, + "loss": 0.8617, + "step": 16650 + }, + { + "epoch": 1.4189177673625906, + "grad_norm": 14.125, + "learning_rate": 4.89414872532017e-07, + "loss": 0.4204, + "step": 16651 + }, + { + "epoch": 1.4190029825308905, + "grad_norm": 14.875, + "learning_rate": 4.892815300327587e-07, + "loss": 0.4971, + "step": 16652 + }, + { + "epoch": 1.4190881976991905, + "grad_norm": 14.875, + "learning_rate": 4.891482012804481e-07, + "loss": 0.4805, + "step": 16653 + }, + { + "epoch": 1.4191734128674904, + "grad_norm": 15.3125, + "learning_rate": 4.89014886277495e-07, + "loss": 0.6677, + "step": 16654 + }, + { + "epoch": 1.4192586280357904, + "grad_norm": 12.375, + "learning_rate": 4.888815850263078e-07, + "loss": 0.2897, + "step": 16655 + }, + { + "epoch": 1.4193438432040903, + "grad_norm": 11.6875, + "learning_rate": 4.887482975292959e-07, + "loss": 0.2136, + "step": 16656 + }, + { + "epoch": 1.4194290583723903, + "grad_norm": 18.25, + "learning_rate": 4.886150237888676e-07, + "loss": 0.6075, + "step": 16657 + }, + { + "epoch": 1.4195142735406903, + "grad_norm": 24.875, + "learning_rate": 4.884817638074313e-07, + "loss": 1.0217, + "step": 16658 + }, + { + "epoch": 1.4195994887089902, + "grad_norm": 24.375, + "learning_rate": 4.883485175873957e-07, + "loss": 0.6955, + "step": 16659 + }, + { + "epoch": 1.4196847038772902, + "grad_norm": 21.0, + "learning_rate": 4.88215285131168e-07, + "loss": 0.6843, + "step": 16660 + }, + { + "epoch": 1.4197699190455901, + "grad_norm": 19.75, + "learning_rate": 4.88082066441156e-07, + "loss": 1.1276, + "step": 16661 + }, + { + "epoch": 1.41985513421389, + "grad_norm": 11.0, + "learning_rate": 4.879488615197678e-07, + "loss": 0.4128, + "step": 16662 + }, + { + "epoch": 1.41994034938219, + "grad_norm": 13.4375, + "learning_rate": 4.878156703694095e-07, + "loss": 0.4682, + "step": 16663 + }, + { + "epoch": 1.42002556455049, + "grad_norm": 16.75, + "learning_rate": 4.876824929924888e-07, + "loss": 0.7233, + "step": 16664 + }, + { + "epoch": 1.42011077971879, + "grad_norm": 15.125, + "learning_rate": 4.875493293914117e-07, + "loss": 0.9003, + "step": 16665 + }, + { + "epoch": 1.42019599488709, + "grad_norm": 16.75, + "learning_rate": 4.874161795685849e-07, + "loss": 0.6798, + "step": 16666 + }, + { + "epoch": 1.4202812100553899, + "grad_norm": 15.9375, + "learning_rate": 4.872830435264144e-07, + "loss": 0.7864, + "step": 16667 + }, + { + "epoch": 1.4203664252236898, + "grad_norm": 15.375, + "learning_rate": 4.87149921267307e-07, + "loss": 0.7661, + "step": 16668 + }, + { + "epoch": 1.4204516403919898, + "grad_norm": 13.3125, + "learning_rate": 4.87016812793667e-07, + "loss": 0.4782, + "step": 16669 + }, + { + "epoch": 1.4205368555602897, + "grad_norm": 13.5, + "learning_rate": 4.86883718107901e-07, + "loss": 0.6022, + "step": 16670 + }, + { + "epoch": 1.4206220707285897, + "grad_norm": 15.3125, + "learning_rate": 4.86750637212413e-07, + "loss": 0.9416, + "step": 16671 + }, + { + "epoch": 1.4207072858968897, + "grad_norm": 17.75, + "learning_rate": 4.866175701096091e-07, + "loss": 0.7532, + "step": 16672 + }, + { + "epoch": 1.4207925010651896, + "grad_norm": 27.125, + "learning_rate": 4.864845168018926e-07, + "loss": 0.8238, + "step": 16673 + }, + { + "epoch": 1.4208777162334896, + "grad_norm": 13.3125, + "learning_rate": 4.86351477291669e-07, + "loss": 0.4847, + "step": 16674 + }, + { + "epoch": 1.4209629314017895, + "grad_norm": 11.0625, + "learning_rate": 4.86218451581342e-07, + "loss": 0.3577, + "step": 16675 + }, + { + "epoch": 1.4210481465700895, + "grad_norm": 16.875, + "learning_rate": 4.86085439673316e-07, + "loss": 0.8356, + "step": 16676 + }, + { + "epoch": 1.4211333617383894, + "grad_norm": 16.0, + "learning_rate": 4.859524415699938e-07, + "loss": 0.5045, + "step": 16677 + }, + { + "epoch": 1.4212185769066894, + "grad_norm": 15.8125, + "learning_rate": 4.858194572737798e-07, + "loss": 0.7373, + "step": 16678 + }, + { + "epoch": 1.4213037920749894, + "grad_norm": 15.0, + "learning_rate": 4.856864867870763e-07, + "loss": 0.6136, + "step": 16679 + }, + { + "epoch": 1.4213890072432893, + "grad_norm": 15.5625, + "learning_rate": 4.855535301122864e-07, + "loss": 0.2688, + "step": 16680 + }, + { + "epoch": 1.4214742224115893, + "grad_norm": 13.0, + "learning_rate": 4.854205872518134e-07, + "loss": 0.402, + "step": 16681 + }, + { + "epoch": 1.4215594375798892, + "grad_norm": 15.125, + "learning_rate": 4.85287658208059e-07, + "loss": 0.6026, + "step": 16682 + }, + { + "epoch": 1.4216446527481892, + "grad_norm": 13.125, + "learning_rate": 4.851547429834253e-07, + "loss": 0.4469, + "step": 16683 + }, + { + "epoch": 1.4217298679164891, + "grad_norm": 12.5, + "learning_rate": 4.850218415803147e-07, + "loss": 0.4459, + "step": 16684 + }, + { + "epoch": 1.421815083084789, + "grad_norm": 17.625, + "learning_rate": 4.848889540011291e-07, + "loss": 0.9097, + "step": 16685 + }, + { + "epoch": 1.421900298253089, + "grad_norm": 13.8125, + "learning_rate": 4.847560802482694e-07, + "loss": 0.4992, + "step": 16686 + }, + { + "epoch": 1.421985513421389, + "grad_norm": 22.25, + "learning_rate": 4.846232203241364e-07, + "loss": 0.8185, + "step": 16687 + }, + { + "epoch": 1.422070728589689, + "grad_norm": 13.3125, + "learning_rate": 4.844903742311315e-07, + "loss": 0.4158, + "step": 16688 + }, + { + "epoch": 1.422155943757989, + "grad_norm": 17.75, + "learning_rate": 4.843575419716556e-07, + "loss": 0.4013, + "step": 16689 + }, + { + "epoch": 1.4222411589262889, + "grad_norm": 14.1875, + "learning_rate": 4.842247235481086e-07, + "loss": 0.7037, + "step": 16690 + }, + { + "epoch": 1.4223263740945888, + "grad_norm": 17.375, + "learning_rate": 4.840919189628908e-07, + "loss": 0.4933, + "step": 16691 + }, + { + "epoch": 1.4224115892628888, + "grad_norm": 14.0625, + "learning_rate": 4.839591282184023e-07, + "loss": 0.5539, + "step": 16692 + }, + { + "epoch": 1.4224968044311888, + "grad_norm": 14.75, + "learning_rate": 4.83826351317043e-07, + "loss": 0.4446, + "step": 16693 + }, + { + "epoch": 1.4225820195994887, + "grad_norm": 12.9375, + "learning_rate": 4.83693588261212e-07, + "loss": 0.4061, + "step": 16694 + }, + { + "epoch": 1.4226672347677887, + "grad_norm": 16.375, + "learning_rate": 4.83560839053308e-07, + "loss": 0.556, + "step": 16695 + }, + { + "epoch": 1.4227524499360886, + "grad_norm": 13.5625, + "learning_rate": 4.834281036957303e-07, + "loss": 0.532, + "step": 16696 + }, + { + "epoch": 1.4228376651043886, + "grad_norm": 17.25, + "learning_rate": 4.832953821908777e-07, + "loss": 0.5811, + "step": 16697 + }, + { + "epoch": 1.4229228802726885, + "grad_norm": 18.5, + "learning_rate": 4.83162674541149e-07, + "loss": 0.634, + "step": 16698 + }, + { + "epoch": 1.4230080954409885, + "grad_norm": 16.375, + "learning_rate": 4.830299807489413e-07, + "loss": 0.4792, + "step": 16699 + }, + { + "epoch": 1.4230933106092885, + "grad_norm": 12.75, + "learning_rate": 4.828973008166532e-07, + "loss": 0.3463, + "step": 16700 + }, + { + "epoch": 1.4231785257775884, + "grad_norm": 42.5, + "learning_rate": 4.827646347466823e-07, + "loss": 0.6968, + "step": 16701 + }, + { + "epoch": 1.4232637409458884, + "grad_norm": 15.0625, + "learning_rate": 4.826319825414268e-07, + "loss": 0.6929, + "step": 16702 + }, + { + "epoch": 1.4233489561141883, + "grad_norm": 10.4375, + "learning_rate": 4.82499344203282e-07, + "loss": 0.3771, + "step": 16703 + }, + { + "epoch": 1.4234341712824883, + "grad_norm": 15.125, + "learning_rate": 4.82366719734646e-07, + "loss": 0.5272, + "step": 16704 + }, + { + "epoch": 1.4235193864507882, + "grad_norm": 18.5, + "learning_rate": 4.822341091379153e-07, + "loss": 0.9869, + "step": 16705 + }, + { + "epoch": 1.4236046016190882, + "grad_norm": 26.125, + "learning_rate": 4.821015124154867e-07, + "loss": 0.7637, + "step": 16706 + }, + { + "epoch": 1.4236898167873882, + "grad_norm": 17.0, + "learning_rate": 4.819689295697555e-07, + "loss": 0.4694, + "step": 16707 + }, + { + "epoch": 1.423775031955688, + "grad_norm": 15.4375, + "learning_rate": 4.81836360603118e-07, + "loss": 0.4784, + "step": 16708 + }, + { + "epoch": 1.423860247123988, + "grad_norm": 15.6875, + "learning_rate": 4.8170380551797e-07, + "loss": 0.6485, + "step": 16709 + }, + { + "epoch": 1.423945462292288, + "grad_norm": 18.25, + "learning_rate": 4.815712643167071e-07, + "loss": 0.5694, + "step": 16710 + }, + { + "epoch": 1.424030677460588, + "grad_norm": 19.375, + "learning_rate": 4.814387370017244e-07, + "loss": 0.8724, + "step": 16711 + }, + { + "epoch": 1.424115892628888, + "grad_norm": 17.0, + "learning_rate": 4.813062235754159e-07, + "loss": 0.4765, + "step": 16712 + }, + { + "epoch": 1.424201107797188, + "grad_norm": 14.125, + "learning_rate": 4.811737240401771e-07, + "loss": 0.8526, + "step": 16713 + }, + { + "epoch": 1.4242863229654878, + "grad_norm": 20.625, + "learning_rate": 4.810412383984022e-07, + "loss": 0.91, + "step": 16714 + }, + { + "epoch": 1.4243715381337878, + "grad_norm": 21.75, + "learning_rate": 4.809087666524857e-07, + "loss": 0.4954, + "step": 16715 + }, + { + "epoch": 1.4244567533020878, + "grad_norm": 19.0, + "learning_rate": 4.807763088048209e-07, + "loss": 0.5744, + "step": 16716 + }, + { + "epoch": 1.4245419684703877, + "grad_norm": 24.625, + "learning_rate": 4.806438648578016e-07, + "loss": 0.8589, + "step": 16717 + }, + { + "epoch": 1.4246271836386877, + "grad_norm": 11.6875, + "learning_rate": 4.805114348138213e-07, + "loss": 0.4083, + "step": 16718 + }, + { + "epoch": 1.4247123988069876, + "grad_norm": 18.0, + "learning_rate": 4.80379018675274e-07, + "loss": 0.886, + "step": 16719 + }, + { + "epoch": 1.4247976139752876, + "grad_norm": 14.3125, + "learning_rate": 4.80246616444551e-07, + "loss": 0.6228, + "step": 16720 + }, + { + "epoch": 1.4248828291435875, + "grad_norm": 11.0, + "learning_rate": 4.801142281240454e-07, + "loss": 0.2413, + "step": 16721 + }, + { + "epoch": 1.4249680443118875, + "grad_norm": 17.875, + "learning_rate": 4.799818537161501e-07, + "loss": 0.9458, + "step": 16722 + }, + { + "epoch": 1.4250532594801875, + "grad_norm": 15.6875, + "learning_rate": 4.798494932232576e-07, + "loss": 0.5363, + "step": 16723 + }, + { + "epoch": 1.4251384746484874, + "grad_norm": 20.0, + "learning_rate": 4.797171466477586e-07, + "loss": 1.0386, + "step": 16724 + }, + { + "epoch": 1.4252236898167874, + "grad_norm": 15.875, + "learning_rate": 4.795848139920453e-07, + "loss": 0.5526, + "step": 16725 + }, + { + "epoch": 1.4253089049850873, + "grad_norm": 15.3125, + "learning_rate": 4.794524952585091e-07, + "loss": 0.8001, + "step": 16726 + }, + { + "epoch": 1.4253941201533873, + "grad_norm": 19.625, + "learning_rate": 4.793201904495418e-07, + "loss": 0.6321, + "step": 16727 + }, + { + "epoch": 1.4254793353216872, + "grad_norm": 14.1875, + "learning_rate": 4.791878995675335e-07, + "loss": 0.7873, + "step": 16728 + }, + { + "epoch": 1.4255645504899872, + "grad_norm": 13.1875, + "learning_rate": 4.790556226148744e-07, + "loss": 0.4344, + "step": 16729 + }, + { + "epoch": 1.4256497656582872, + "grad_norm": 16.125, + "learning_rate": 4.789233595939556e-07, + "loss": 0.5117, + "step": 16730 + }, + { + "epoch": 1.4257349808265871, + "grad_norm": 15.875, + "learning_rate": 4.787911105071671e-07, + "loss": 0.5396, + "step": 16731 + }, + { + "epoch": 1.425820195994887, + "grad_norm": 15.8125, + "learning_rate": 4.786588753568991e-07, + "loss": 0.6739, + "step": 16732 + }, + { + "epoch": 1.425905411163187, + "grad_norm": 15.3125, + "learning_rate": 4.785266541455405e-07, + "loss": 0.6102, + "step": 16733 + }, + { + "epoch": 1.425990626331487, + "grad_norm": 25.375, + "learning_rate": 4.78394446875481e-07, + "loss": 0.7135, + "step": 16734 + }, + { + "epoch": 1.426075841499787, + "grad_norm": 10.6875, + "learning_rate": 4.782622535491102e-07, + "loss": 0.3138, + "step": 16735 + }, + { + "epoch": 1.426161056668087, + "grad_norm": 22.25, + "learning_rate": 4.781300741688165e-07, + "loss": 0.9088, + "step": 16736 + }, + { + "epoch": 1.4262462718363869, + "grad_norm": 14.875, + "learning_rate": 4.779979087369881e-07, + "loss": 0.5635, + "step": 16737 + }, + { + "epoch": 1.4263314870046868, + "grad_norm": 21.625, + "learning_rate": 4.778657572560137e-07, + "loss": 0.7276, + "step": 16738 + }, + { + "epoch": 1.4264167021729868, + "grad_norm": 11.875, + "learning_rate": 4.777336197282817e-07, + "loss": 0.3042, + "step": 16739 + }, + { + "epoch": 1.4265019173412867, + "grad_norm": 16.375, + "learning_rate": 4.776014961561801e-07, + "loss": 0.6561, + "step": 16740 + }, + { + "epoch": 1.4265871325095867, + "grad_norm": 14.625, + "learning_rate": 4.774693865420957e-07, + "loss": 0.3912, + "step": 16741 + }, + { + "epoch": 1.4266723476778866, + "grad_norm": 14.375, + "learning_rate": 4.773372908884164e-07, + "loss": 0.756, + "step": 16742 + }, + { + "epoch": 1.4267575628461866, + "grad_norm": 20.125, + "learning_rate": 4.772052091975296e-07, + "loss": 0.7665, + "step": 16743 + }, + { + "epoch": 1.4268427780144866, + "grad_norm": 20.25, + "learning_rate": 4.770731414718213e-07, + "loss": 0.5692, + "step": 16744 + }, + { + "epoch": 1.4269279931827865, + "grad_norm": 13.0, + "learning_rate": 4.769410877136792e-07, + "loss": 0.453, + "step": 16745 + }, + { + "epoch": 1.4270132083510865, + "grad_norm": 15.625, + "learning_rate": 4.7680904792548845e-07, + "loss": 0.7236, + "step": 16746 + }, + { + "epoch": 1.4270984235193864, + "grad_norm": 15.625, + "learning_rate": 4.766770221096356e-07, + "loss": 0.8154, + "step": 16747 + }, + { + "epoch": 1.4271836386876864, + "grad_norm": 17.5, + "learning_rate": 4.765450102685068e-07, + "loss": 0.5076, + "step": 16748 + }, + { + "epoch": 1.4272688538559863, + "grad_norm": 12.9375, + "learning_rate": 4.764130124044877e-07, + "loss": 0.3557, + "step": 16749 + }, + { + "epoch": 1.4273540690242863, + "grad_norm": 15.125, + "learning_rate": 4.7628102851996303e-07, + "loss": 0.4774, + "step": 16750 + }, + { + "epoch": 1.4274392841925863, + "grad_norm": 17.875, + "learning_rate": 4.761490586173187e-07, + "loss": 0.7747, + "step": 16751 + }, + { + "epoch": 1.4275244993608862, + "grad_norm": 22.75, + "learning_rate": 4.760171026989385e-07, + "loss": 1.089, + "step": 16752 + }, + { + "epoch": 1.4276097145291862, + "grad_norm": 15.6875, + "learning_rate": 4.7588516076720795e-07, + "loss": 0.8067, + "step": 16753 + }, + { + "epoch": 1.4276949296974861, + "grad_norm": 14.0, + "learning_rate": 4.757532328245107e-07, + "loss": 0.6167, + "step": 16754 + }, + { + "epoch": 1.427780144865786, + "grad_norm": 17.0, + "learning_rate": 4.75621318873231e-07, + "loss": 0.682, + "step": 16755 + }, + { + "epoch": 1.427865360034086, + "grad_norm": 15.6875, + "learning_rate": 4.754894189157527e-07, + "loss": 0.6443, + "step": 16756 + }, + { + "epoch": 1.427950575202386, + "grad_norm": 19.0, + "learning_rate": 4.7535753295446006e-07, + "loss": 0.6721, + "step": 16757 + }, + { + "epoch": 1.428035790370686, + "grad_norm": 17.0, + "learning_rate": 4.752256609917352e-07, + "loss": 1.028, + "step": 16758 + }, + { + "epoch": 1.428121005538986, + "grad_norm": 15.75, + "learning_rate": 4.750938030299622e-07, + "loss": 0.7033, + "step": 16759 + }, + { + "epoch": 1.4282062207072859, + "grad_norm": 14.8125, + "learning_rate": 4.749619590715229e-07, + "loss": 0.4882, + "step": 16760 + }, + { + "epoch": 1.4282914358755858, + "grad_norm": 15.25, + "learning_rate": 4.748301291188005e-07, + "loss": 0.6189, + "step": 16761 + }, + { + "epoch": 1.4283766510438858, + "grad_norm": 14.625, + "learning_rate": 4.746983131741775e-07, + "loss": 0.5514, + "step": 16762 + }, + { + "epoch": 1.4284618662121857, + "grad_norm": 13.375, + "learning_rate": 4.745665112400352e-07, + "loss": 0.6223, + "step": 16763 + }, + { + "epoch": 1.4285470813804857, + "grad_norm": 19.625, + "learning_rate": 4.7443472331875573e-07, + "loss": 0.6374, + "step": 16764 + }, + { + "epoch": 1.4286322965487857, + "grad_norm": 14.625, + "learning_rate": 4.743029494127213e-07, + "loss": 0.7396, + "step": 16765 + }, + { + "epoch": 1.4287175117170856, + "grad_norm": 15.8125, + "learning_rate": 4.74171189524312e-07, + "loss": 0.4851, + "step": 16766 + }, + { + "epoch": 1.4288027268853856, + "grad_norm": 20.875, + "learning_rate": 4.7403944365590993e-07, + "loss": 0.6922, + "step": 16767 + }, + { + "epoch": 1.4288879420536855, + "grad_norm": 15.875, + "learning_rate": 4.7390771180989505e-07, + "loss": 0.5071, + "step": 16768 + }, + { + "epoch": 1.4289731572219855, + "grad_norm": 14.75, + "learning_rate": 4.7377599398864814e-07, + "loss": 0.5815, + "step": 16769 + }, + { + "epoch": 1.4290583723902854, + "grad_norm": 15.8125, + "learning_rate": 4.736442901945501e-07, + "loss": 0.64, + "step": 16770 + }, + { + "epoch": 1.4291435875585854, + "grad_norm": 12.4375, + "learning_rate": 4.735126004299799e-07, + "loss": 0.4494, + "step": 16771 + }, + { + "epoch": 1.4292288027268854, + "grad_norm": 15.8125, + "learning_rate": 4.7338092469731795e-07, + "loss": 0.5254, + "step": 16772 + }, + { + "epoch": 1.4293140178951853, + "grad_norm": 13.4375, + "learning_rate": 4.7324926299894354e-07, + "loss": 0.5048, + "step": 16773 + }, + { + "epoch": 1.4293992330634853, + "grad_norm": 12.5, + "learning_rate": 4.7311761533723645e-07, + "loss": 0.374, + "step": 16774 + }, + { + "epoch": 1.4294844482317852, + "grad_norm": 15.875, + "learning_rate": 4.729859817145753e-07, + "loss": 0.2672, + "step": 16775 + }, + { + "epoch": 1.4295696634000852, + "grad_norm": 13.1875, + "learning_rate": 4.728543621333384e-07, + "loss": 0.4818, + "step": 16776 + }, + { + "epoch": 1.4296548785683851, + "grad_norm": 21.625, + "learning_rate": 4.7272275659590457e-07, + "loss": 1.0221, + "step": 16777 + }, + { + "epoch": 1.429740093736685, + "grad_norm": 14.9375, + "learning_rate": 4.725911651046522e-07, + "loss": 0.6897, + "step": 16778 + }, + { + "epoch": 1.429825308904985, + "grad_norm": 17.75, + "learning_rate": 4.7245958766195976e-07, + "loss": 0.6093, + "step": 16779 + }, + { + "epoch": 1.429910524073285, + "grad_norm": 16.0, + "learning_rate": 4.7232802427020385e-07, + "loss": 0.2735, + "step": 16780 + }, + { + "epoch": 1.429995739241585, + "grad_norm": 14.5, + "learning_rate": 4.721964749317626e-07, + "loss": 0.8889, + "step": 16781 + }, + { + "epoch": 1.430080954409885, + "grad_norm": 13.5, + "learning_rate": 4.720649396490137e-07, + "loss": 0.5683, + "step": 16782 + }, + { + "epoch": 1.4301661695781849, + "grad_norm": 16.75, + "learning_rate": 4.7193341842433347e-07, + "loss": 0.6365, + "step": 16783 + }, + { + "epoch": 1.4302513847464848, + "grad_norm": 11.8125, + "learning_rate": 4.7180191126009843e-07, + "loss": 0.4008, + "step": 16784 + }, + { + "epoch": 1.4303365999147848, + "grad_norm": 13.25, + "learning_rate": 4.716704181586855e-07, + "loss": 0.4825, + "step": 16785 + }, + { + "epoch": 1.4304218150830847, + "grad_norm": 17.5, + "learning_rate": 4.715389391224706e-07, + "loss": 0.4019, + "step": 16786 + }, + { + "epoch": 1.4305070302513847, + "grad_norm": 13.625, + "learning_rate": 4.7140747415383035e-07, + "loss": 0.2482, + "step": 16787 + }, + { + "epoch": 1.4305922454196847, + "grad_norm": 16.75, + "learning_rate": 4.7127602325513955e-07, + "loss": 0.6586, + "step": 16788 + }, + { + "epoch": 1.4306774605879846, + "grad_norm": 12.4375, + "learning_rate": 4.711445864287742e-07, + "loss": 0.5883, + "step": 16789 + }, + { + "epoch": 1.4307626757562846, + "grad_norm": 17.25, + "learning_rate": 4.7101316367710926e-07, + "loss": 0.9237, + "step": 16790 + }, + { + "epoch": 1.4308478909245845, + "grad_norm": 22.125, + "learning_rate": 4.708817550025202e-07, + "loss": 0.8415, + "step": 16791 + }, + { + "epoch": 1.4309331060928845, + "grad_norm": 13.3125, + "learning_rate": 4.707503604073814e-07, + "loss": 0.4571, + "step": 16792 + }, + { + "epoch": 1.4310183212611844, + "grad_norm": 12.8125, + "learning_rate": 4.7061897989406664e-07, + "loss": 0.4654, + "step": 16793 + }, + { + "epoch": 1.4311035364294844, + "grad_norm": 18.375, + "learning_rate": 4.704876134649507e-07, + "loss": 0.917, + "step": 16794 + }, + { + "epoch": 1.4311887515977844, + "grad_norm": 11.375, + "learning_rate": 4.703562611224073e-07, + "loss": 0.2416, + "step": 16795 + }, + { + "epoch": 1.4312739667660843, + "grad_norm": 16.75, + "learning_rate": 4.702249228688109e-07, + "loss": 0.6096, + "step": 16796 + }, + { + "epoch": 1.4313591819343843, + "grad_norm": 19.0, + "learning_rate": 4.700935987065337e-07, + "loss": 1.0781, + "step": 16797 + }, + { + "epoch": 1.4314443971026842, + "grad_norm": 11.4375, + "learning_rate": 4.699622886379494e-07, + "loss": 0.4531, + "step": 16798 + }, + { + "epoch": 1.4315296122709842, + "grad_norm": 12.9375, + "learning_rate": 4.6983099266543154e-07, + "loss": 0.3972, + "step": 16799 + }, + { + "epoch": 1.4316148274392841, + "grad_norm": 12.625, + "learning_rate": 4.69699710791352e-07, + "loss": 0.6589, + "step": 16800 + }, + { + "epoch": 1.431700042607584, + "grad_norm": 18.375, + "learning_rate": 4.69568443018083e-07, + "loss": 0.6045, + "step": 16801 + }, + { + "epoch": 1.431785257775884, + "grad_norm": 14.9375, + "learning_rate": 4.6943718934799694e-07, + "loss": 0.4343, + "step": 16802 + }, + { + "epoch": 1.431870472944184, + "grad_norm": 13.9375, + "learning_rate": 4.693059497834658e-07, + "loss": 0.9234, + "step": 16803 + }, + { + "epoch": 1.431955688112484, + "grad_norm": 19.125, + "learning_rate": 4.691747243268617e-07, + "loss": 0.7311, + "step": 16804 + }, + { + "epoch": 1.432040903280784, + "grad_norm": 12.6875, + "learning_rate": 4.69043512980555e-07, + "loss": 0.4026, + "step": 16805 + }, + { + "epoch": 1.4321261184490839, + "grad_norm": 20.125, + "learning_rate": 4.689123157469175e-07, + "loss": 0.509, + "step": 16806 + }, + { + "epoch": 1.4322113336173838, + "grad_norm": 15.5, + "learning_rate": 4.687811326283197e-07, + "loss": 0.6349, + "step": 16807 + }, + { + "epoch": 1.4322965487856838, + "grad_norm": 16.25, + "learning_rate": 4.6864996362713285e-07, + "loss": 0.358, + "step": 16808 + }, + { + "epoch": 1.4323817639539838, + "grad_norm": 17.625, + "learning_rate": 4.6851880874572694e-07, + "loss": 0.7008, + "step": 16809 + }, + { + "epoch": 1.4324669791222837, + "grad_norm": 19.375, + "learning_rate": 4.6838766798647155e-07, + "loss": 0.7913, + "step": 16810 + }, + { + "epoch": 1.4325521942905837, + "grad_norm": 14.1875, + "learning_rate": 4.6825654135173703e-07, + "loss": 0.5885, + "step": 16811 + }, + { + "epoch": 1.4326374094588836, + "grad_norm": 13.0625, + "learning_rate": 4.681254288438933e-07, + "loss": 0.4588, + "step": 16812 + }, + { + "epoch": 1.4327226246271836, + "grad_norm": 11.75, + "learning_rate": 4.679943304653088e-07, + "loss": 0.4323, + "step": 16813 + }, + { + "epoch": 1.4328078397954835, + "grad_norm": 17.0, + "learning_rate": 4.6786324621835316e-07, + "loss": 0.7268, + "step": 16814 + }, + { + "epoch": 1.4328930549637835, + "grad_norm": 15.0625, + "learning_rate": 4.6773217610539517e-07, + "loss": 0.4369, + "step": 16815 + }, + { + "epoch": 1.4329782701320835, + "grad_norm": 11.8125, + "learning_rate": 4.6760112012880387e-07, + "loss": 0.3471, + "step": 16816 + }, + { + "epoch": 1.4330634853003834, + "grad_norm": 14.9375, + "learning_rate": 4.674700782909471e-07, + "loss": 0.5582, + "step": 16817 + }, + { + "epoch": 1.4331487004686834, + "grad_norm": 14.5625, + "learning_rate": 4.673390505941923e-07, + "loss": 0.5054, + "step": 16818 + }, + { + "epoch": 1.4332339156369833, + "grad_norm": 15.375, + "learning_rate": 4.672080370409081e-07, + "loss": 0.4612, + "step": 16819 + }, + { + "epoch": 1.4333191308052833, + "grad_norm": 13.5, + "learning_rate": 4.670770376334617e-07, + "loss": 0.5002, + "step": 16820 + }, + { + "epoch": 1.4334043459735832, + "grad_norm": 16.875, + "learning_rate": 4.66946052374221e-07, + "loss": 0.4621, + "step": 16821 + }, + { + "epoch": 1.4334895611418832, + "grad_norm": 13.0, + "learning_rate": 4.668150812655521e-07, + "loss": 0.327, + "step": 16822 + }, + { + "epoch": 1.4335747763101832, + "grad_norm": 14.125, + "learning_rate": 4.666841243098222e-07, + "loss": 0.5494, + "step": 16823 + }, + { + "epoch": 1.433659991478483, + "grad_norm": 20.0, + "learning_rate": 4.6655318150939837e-07, + "loss": 0.833, + "step": 16824 + }, + { + "epoch": 1.433745206646783, + "grad_norm": 18.625, + "learning_rate": 4.664222528666459e-07, + "loss": 0.7251, + "step": 16825 + }, + { + "epoch": 1.433830421815083, + "grad_norm": 17.0, + "learning_rate": 4.6629133838393167e-07, + "loss": 0.6115, + "step": 16826 + }, + { + "epoch": 1.433915636983383, + "grad_norm": 11.875, + "learning_rate": 4.6616043806362074e-07, + "loss": 0.3541, + "step": 16827 + }, + { + "epoch": 1.434000852151683, + "grad_norm": 11.8125, + "learning_rate": 4.6602955190807873e-07, + "loss": 0.3607, + "step": 16828 + }, + { + "epoch": 1.434086067319983, + "grad_norm": 15.0, + "learning_rate": 4.6589867991967156e-07, + "loss": 0.6337, + "step": 16829 + }, + { + "epoch": 1.4341712824882829, + "grad_norm": 22.5, + "learning_rate": 4.657678221007633e-07, + "loss": 0.6134, + "step": 16830 + }, + { + "epoch": 1.4342564976565828, + "grad_norm": 17.5, + "learning_rate": 4.6563697845371916e-07, + "loss": 0.753, + "step": 16831 + }, + { + "epoch": 1.4343417128248828, + "grad_norm": 16.25, + "learning_rate": 4.6550614898090385e-07, + "loss": 0.5584, + "step": 16832 + }, + { + "epoch": 1.4344269279931827, + "grad_norm": 24.5, + "learning_rate": 4.65375333684681e-07, + "loss": 0.6933, + "step": 16833 + }, + { + "epoch": 1.4345121431614827, + "grad_norm": 11.6875, + "learning_rate": 4.652445325674151e-07, + "loss": 0.2988, + "step": 16834 + }, + { + "epoch": 1.4345973583297826, + "grad_norm": 15.875, + "learning_rate": 4.6511374563146946e-07, + "loss": 0.8899, + "step": 16835 + }, + { + "epoch": 1.4346825734980826, + "grad_norm": 15.0625, + "learning_rate": 4.6498297287920755e-07, + "loss": 0.629, + "step": 16836 + }, + { + "epoch": 1.4347677886663825, + "grad_norm": 12.1875, + "learning_rate": 4.648522143129927e-07, + "loss": 0.4961, + "step": 16837 + }, + { + "epoch": 1.4348530038346825, + "grad_norm": 14.125, + "learning_rate": 4.647214699351883e-07, + "loss": 0.5789, + "step": 16838 + }, + { + "epoch": 1.4349382190029825, + "grad_norm": 17.125, + "learning_rate": 4.6459073974815624e-07, + "loss": 0.8912, + "step": 16839 + }, + { + "epoch": 1.4350234341712824, + "grad_norm": 16.0, + "learning_rate": 4.644600237542597e-07, + "loss": 0.8127, + "step": 16840 + }, + { + "epoch": 1.4351086493395824, + "grad_norm": 11.125, + "learning_rate": 4.6432932195586006e-07, + "loss": 0.3616, + "step": 16841 + }, + { + "epoch": 1.4351938645078823, + "grad_norm": 13.5, + "learning_rate": 4.641986343553201e-07, + "loss": 0.4179, + "step": 16842 + }, + { + "epoch": 1.4352790796761823, + "grad_norm": 17.625, + "learning_rate": 4.6406796095500044e-07, + "loss": 0.5292, + "step": 16843 + }, + { + "epoch": 1.4353642948444822, + "grad_norm": 17.25, + "learning_rate": 4.639373017572631e-07, + "loss": 0.7788, + "step": 16844 + }, + { + "epoch": 1.4354495100127822, + "grad_norm": 13.625, + "learning_rate": 4.6380665676446917e-07, + "loss": 0.6699, + "step": 16845 + }, + { + "epoch": 1.4355347251810822, + "grad_norm": 15.0625, + "learning_rate": 4.6367602597897994e-07, + "loss": 0.5409, + "step": 16846 + }, + { + "epoch": 1.4356199403493821, + "grad_norm": 14.3125, + "learning_rate": 4.635454094031551e-07, + "loss": 0.5046, + "step": 16847 + }, + { + "epoch": 1.435705155517682, + "grad_norm": 14.0, + "learning_rate": 4.6341480703935613e-07, + "loss": 0.4164, + "step": 16848 + }, + { + "epoch": 1.435790370685982, + "grad_norm": 18.5, + "learning_rate": 4.632842188899421e-07, + "loss": 0.9258, + "step": 16849 + }, + { + "epoch": 1.435875585854282, + "grad_norm": 17.0, + "learning_rate": 4.631536449572734e-07, + "loss": 0.4308, + "step": 16850 + }, + { + "epoch": 1.435960801022582, + "grad_norm": 11.75, + "learning_rate": 4.630230852437098e-07, + "loss": 0.3904, + "step": 16851 + }, + { + "epoch": 1.436046016190882, + "grad_norm": 14.75, + "learning_rate": 4.628925397516101e-07, + "loss": 0.7777, + "step": 16852 + }, + { + "epoch": 1.4361312313591819, + "grad_norm": 12.6875, + "learning_rate": 4.627620084833338e-07, + "loss": 0.4798, + "step": 16853 + }, + { + "epoch": 1.4362164465274818, + "grad_norm": 12.4375, + "learning_rate": 4.6263149144123954e-07, + "loss": 0.401, + "step": 16854 + }, + { + "epoch": 1.4363016616957818, + "grad_norm": 15.625, + "learning_rate": 4.6250098862768657e-07, + "loss": 0.9116, + "step": 16855 + }, + { + "epoch": 1.4363868768640817, + "grad_norm": 15.75, + "learning_rate": 4.6237050004503197e-07, + "loss": 0.6061, + "step": 16856 + }, + { + "epoch": 1.4364720920323817, + "grad_norm": 16.125, + "learning_rate": 4.62240025695635e-07, + "loss": 0.5902, + "step": 16857 + }, + { + "epoch": 1.4365573072006816, + "grad_norm": 18.125, + "learning_rate": 4.621095655818525e-07, + "loss": 0.7063, + "step": 16858 + }, + { + "epoch": 1.4366425223689816, + "grad_norm": 16.75, + "learning_rate": 4.619791197060429e-07, + "loss": 0.6096, + "step": 16859 + }, + { + "epoch": 1.4367277375372816, + "grad_norm": 16.125, + "learning_rate": 4.618486880705625e-07, + "loss": 0.7369, + "step": 16860 + }, + { + "epoch": 1.4368129527055815, + "grad_norm": 13.6875, + "learning_rate": 4.617182706777688e-07, + "loss": 0.5422, + "step": 16861 + }, + { + "epoch": 1.4368981678738815, + "grad_norm": 23.75, + "learning_rate": 4.615878675300188e-07, + "loss": 0.6886, + "step": 16862 + }, + { + "epoch": 1.4369833830421814, + "grad_norm": 58.5, + "learning_rate": 4.614574786296691e-07, + "loss": 0.9406, + "step": 16863 + }, + { + "epoch": 1.4370685982104814, + "grad_norm": 21.5, + "learning_rate": 4.6132710397907537e-07, + "loss": 0.8374, + "step": 16864 + }, + { + "epoch": 1.4371538133787813, + "grad_norm": 14.0, + "learning_rate": 4.6119674358059446e-07, + "loss": 0.4439, + "step": 16865 + }, + { + "epoch": 1.4372390285470813, + "grad_norm": 9.8125, + "learning_rate": 4.61066397436581e-07, + "loss": 0.5845, + "step": 16866 + }, + { + "epoch": 1.4373242437153813, + "grad_norm": 15.6875, + "learning_rate": 4.609360655493912e-07, + "loss": 0.6767, + "step": 16867 + }, + { + "epoch": 1.4374094588836812, + "grad_norm": 14.8125, + "learning_rate": 4.608057479213805e-07, + "loss": 0.6349, + "step": 16868 + }, + { + "epoch": 1.4374946740519812, + "grad_norm": 23.375, + "learning_rate": 4.6067544455490315e-07, + "loss": 0.8008, + "step": 16869 + }, + { + "epoch": 1.4375798892202813, + "grad_norm": 20.375, + "learning_rate": 4.6054515545231425e-07, + "loss": 0.9809, + "step": 16870 + }, + { + "epoch": 1.4376651043885813, + "grad_norm": 13.1875, + "learning_rate": 4.604148806159682e-07, + "loss": 0.6414, + "step": 16871 + }, + { + "epoch": 1.4377503195568813, + "grad_norm": 13.0625, + "learning_rate": 4.602846200482197e-07, + "loss": 0.3736, + "step": 16872 + }, + { + "epoch": 1.4378355347251812, + "grad_norm": 13.6875, + "learning_rate": 4.601543737514222e-07, + "loss": 0.351, + "step": 16873 + }, + { + "epoch": 1.4379207498934812, + "grad_norm": 15.1875, + "learning_rate": 4.6002414172792916e-07, + "loss": 0.6214, + "step": 16874 + }, + { + "epoch": 1.4380059650617811, + "grad_norm": 16.875, + "learning_rate": 4.598939239800941e-07, + "loss": 0.7619, + "step": 16875 + }, + { + "epoch": 1.438091180230081, + "grad_norm": 18.375, + "learning_rate": 4.597637205102708e-07, + "loss": 0.5886, + "step": 16876 + }, + { + "epoch": 1.438176395398381, + "grad_norm": 16.125, + "learning_rate": 4.596335313208113e-07, + "loss": 0.7868, + "step": 16877 + }, + { + "epoch": 1.438261610566681, + "grad_norm": 31.125, + "learning_rate": 4.5950335641406857e-07, + "loss": 0.9351, + "step": 16878 + }, + { + "epoch": 1.438346825734981, + "grad_norm": 15.0, + "learning_rate": 4.5937319579239526e-07, + "loss": 0.6491, + "step": 16879 + }, + { + "epoch": 1.438432040903281, + "grad_norm": 13.6875, + "learning_rate": 4.5924304945814363e-07, + "loss": 0.4664, + "step": 16880 + }, + { + "epoch": 1.4385172560715809, + "grad_norm": 13.1875, + "learning_rate": 4.5911291741366535e-07, + "loss": 0.4907, + "step": 16881 + }, + { + "epoch": 1.4386024712398808, + "grad_norm": 12.4375, + "learning_rate": 4.5898279966131154e-07, + "loss": 0.3438, + "step": 16882 + }, + { + "epoch": 1.4386876864081808, + "grad_norm": 14.3125, + "learning_rate": 4.5885269620343386e-07, + "loss": 0.3246, + "step": 16883 + }, + { + "epoch": 1.4387729015764807, + "grad_norm": 13.25, + "learning_rate": 4.587226070423836e-07, + "loss": 0.406, + "step": 16884 + }, + { + "epoch": 1.4388581167447807, + "grad_norm": 17.5, + "learning_rate": 4.5859253218051194e-07, + "loss": 0.7256, + "step": 16885 + }, + { + "epoch": 1.4389433319130807, + "grad_norm": 13.3125, + "learning_rate": 4.5846247162016857e-07, + "loss": 0.44, + "step": 16886 + }, + { + "epoch": 1.4390285470813806, + "grad_norm": 29.875, + "learning_rate": 4.5833242536370436e-07, + "loss": 0.7225, + "step": 16887 + }, + { + "epoch": 1.4391137622496806, + "grad_norm": 78.0, + "learning_rate": 4.5820239341346925e-07, + "loss": 0.4877, + "step": 16888 + }, + { + "epoch": 1.4391989774179805, + "grad_norm": 14.25, + "learning_rate": 4.5807237577181405e-07, + "loss": 0.4459, + "step": 16889 + }, + { + "epoch": 1.4392841925862805, + "grad_norm": 16.25, + "learning_rate": 4.5794237244108644e-07, + "loss": 0.6603, + "step": 16890 + }, + { + "epoch": 1.4393694077545804, + "grad_norm": 15.625, + "learning_rate": 4.5781238342363655e-07, + "loss": 0.5501, + "step": 16891 + }, + { + "epoch": 1.4394546229228804, + "grad_norm": 20.0, + "learning_rate": 4.576824087218136e-07, + "loss": 0.9311, + "step": 16892 + }, + { + "epoch": 1.4395398380911804, + "grad_norm": 13.75, + "learning_rate": 4.5755244833796673e-07, + "loss": 0.531, + "step": 16893 + }, + { + "epoch": 1.4396250532594803, + "grad_norm": 14.0, + "learning_rate": 4.5742250227444354e-07, + "loss": 0.6114, + "step": 16894 + }, + { + "epoch": 1.4397102684277803, + "grad_norm": 20.25, + "learning_rate": 4.572925705335926e-07, + "loss": 0.6188, + "step": 16895 + }, + { + "epoch": 1.4397954835960802, + "grad_norm": 18.0, + "learning_rate": 4.5716265311776224e-07, + "loss": 0.812, + "step": 16896 + }, + { + "epoch": 1.4398806987643802, + "grad_norm": 14.75, + "learning_rate": 4.570327500293002e-07, + "loss": 0.4128, + "step": 16897 + }, + { + "epoch": 1.4399659139326801, + "grad_norm": 16.125, + "learning_rate": 4.5690286127055394e-07, + "loss": 0.588, + "step": 16898 + }, + { + "epoch": 1.44005112910098, + "grad_norm": 17.75, + "learning_rate": 4.5677298684387e-07, + "loss": 0.7883, + "step": 16899 + }, + { + "epoch": 1.44013634426928, + "grad_norm": 17.0, + "learning_rate": 4.5664312675159584e-07, + "loss": 0.576, + "step": 16900 + }, + { + "epoch": 1.44022155943758, + "grad_norm": 14.0625, + "learning_rate": 4.565132809960782e-07, + "loss": 0.2517, + "step": 16901 + }, + { + "epoch": 1.44030677460588, + "grad_norm": 11.75, + "learning_rate": 4.5638344957966385e-07, + "loss": 0.4348, + "step": 16902 + }, + { + "epoch": 1.44039198977418, + "grad_norm": 17.625, + "learning_rate": 4.562536325046983e-07, + "loss": 0.4361, + "step": 16903 + }, + { + "epoch": 1.4404772049424799, + "grad_norm": 17.125, + "learning_rate": 4.5612382977352764e-07, + "loss": 0.4496, + "step": 16904 + }, + { + "epoch": 1.4405624201107798, + "grad_norm": 11.9375, + "learning_rate": 4.559940413884982e-07, + "loss": 0.4086, + "step": 16905 + }, + { + "epoch": 1.4406476352790798, + "grad_norm": 18.5, + "learning_rate": 4.558642673519549e-07, + "loss": 0.4921, + "step": 16906 + }, + { + "epoch": 1.4407328504473798, + "grad_norm": 13.75, + "learning_rate": 4.557345076662423e-07, + "loss": 0.4418, + "step": 16907 + }, + { + "epoch": 1.4408180656156797, + "grad_norm": 16.625, + "learning_rate": 4.5560476233370593e-07, + "loss": 0.3985, + "step": 16908 + }, + { + "epoch": 1.4409032807839797, + "grad_norm": 13.9375, + "learning_rate": 4.5547503135669026e-07, + "loss": 0.7148, + "step": 16909 + }, + { + "epoch": 1.4409884959522796, + "grad_norm": 14.875, + "learning_rate": 4.553453147375403e-07, + "loss": 0.7468, + "step": 16910 + }, + { + "epoch": 1.4410737111205796, + "grad_norm": 17.5, + "learning_rate": 4.5521561247859907e-07, + "loss": 0.566, + "step": 16911 + }, + { + "epoch": 1.4411589262888795, + "grad_norm": 14.5, + "learning_rate": 4.550859245822109e-07, + "loss": 0.6, + "step": 16912 + }, + { + "epoch": 1.4412441414571795, + "grad_norm": 22.75, + "learning_rate": 4.549562510507198e-07, + "loss": 0.7223, + "step": 16913 + }, + { + "epoch": 1.4413293566254795, + "grad_norm": 24.0, + "learning_rate": 4.5482659188646844e-07, + "loss": 0.7964, + "step": 16914 + }, + { + "epoch": 1.4414145717937794, + "grad_norm": 18.375, + "learning_rate": 4.546969470918006e-07, + "loss": 0.7927, + "step": 16915 + }, + { + "epoch": 1.4414997869620794, + "grad_norm": 21.625, + "learning_rate": 4.545673166690581e-07, + "loss": 0.9297, + "step": 16916 + }, + { + "epoch": 1.4415850021303793, + "grad_norm": 15.6875, + "learning_rate": 4.544377006205841e-07, + "loss": 0.7559, + "step": 16917 + }, + { + "epoch": 1.4416702172986793, + "grad_norm": 14.3125, + "learning_rate": 4.54308098948721e-07, + "loss": 0.585, + "step": 16918 + }, + { + "epoch": 1.4417554324669792, + "grad_norm": 17.5, + "learning_rate": 4.5417851165581087e-07, + "loss": 0.9007, + "step": 16919 + }, + { + "epoch": 1.4418406476352792, + "grad_norm": 11.5, + "learning_rate": 4.54048938744195e-07, + "loss": 0.5278, + "step": 16920 + }, + { + "epoch": 1.4419258628035792, + "grad_norm": 20.875, + "learning_rate": 4.539193802162152e-07, + "loss": 0.5611, + "step": 16921 + }, + { + "epoch": 1.442011077971879, + "grad_norm": 14.25, + "learning_rate": 4.5378983607421324e-07, + "loss": 0.4949, + "step": 16922 + }, + { + "epoch": 1.442096293140179, + "grad_norm": 14.0625, + "learning_rate": 4.5366030632052944e-07, + "loss": 0.4909, + "step": 16923 + }, + { + "epoch": 1.442181508308479, + "grad_norm": 15.1875, + "learning_rate": 4.535307909575043e-07, + "loss": 0.5904, + "step": 16924 + }, + { + "epoch": 1.442266723476779, + "grad_norm": 16.25, + "learning_rate": 4.5340128998747873e-07, + "loss": 0.8062, + "step": 16925 + }, + { + "epoch": 1.442351938645079, + "grad_norm": 20.25, + "learning_rate": 4.5327180341279284e-07, + "loss": 0.6103, + "step": 16926 + }, + { + "epoch": 1.442437153813379, + "grad_norm": 14.375, + "learning_rate": 4.53142331235787e-07, + "loss": 0.5257, + "step": 16927 + }, + { + "epoch": 1.4425223689816788, + "grad_norm": 22.125, + "learning_rate": 4.5301287345880016e-07, + "loss": 0.8956, + "step": 16928 + }, + { + "epoch": 1.4426075841499788, + "grad_norm": 17.375, + "learning_rate": 4.528834300841722e-07, + "loss": 0.6664, + "step": 16929 + }, + { + "epoch": 1.4426927993182788, + "grad_norm": 14.5625, + "learning_rate": 4.5275400111424255e-07, + "loss": 0.4301, + "step": 16930 + }, + { + "epoch": 1.4427780144865787, + "grad_norm": 16.875, + "learning_rate": 4.526245865513494e-07, + "loss": 0.7455, + "step": 16931 + }, + { + "epoch": 1.4428632296548787, + "grad_norm": 15.25, + "learning_rate": 4.5249518639783233e-07, + "loss": 0.7453, + "step": 16932 + }, + { + "epoch": 1.4429484448231786, + "grad_norm": 16.875, + "learning_rate": 4.523658006560286e-07, + "loss": 0.6358, + "step": 16933 + }, + { + "epoch": 1.4430336599914786, + "grad_norm": 14.875, + "learning_rate": 4.5223642932827704e-07, + "loss": 0.6633, + "step": 16934 + }, + { + "epoch": 1.4431188751597785, + "grad_norm": 22.0, + "learning_rate": 4.5210707241691577e-07, + "loss": 0.846, + "step": 16935 + }, + { + "epoch": 1.4432040903280785, + "grad_norm": 26.5, + "learning_rate": 4.519777299242817e-07, + "loss": 0.8921, + "step": 16936 + }, + { + "epoch": 1.4432893054963785, + "grad_norm": 14.0, + "learning_rate": 4.518484018527125e-07, + "loss": 0.6387, + "step": 16937 + }, + { + "epoch": 1.4433745206646784, + "grad_norm": 20.5, + "learning_rate": 4.5171908820454564e-07, + "loss": 0.8986, + "step": 16938 + }, + { + "epoch": 1.4434597358329784, + "grad_norm": 16.5, + "learning_rate": 4.515897889821172e-07, + "loss": 0.7503, + "step": 16939 + }, + { + "epoch": 1.4435449510012783, + "grad_norm": 11.5625, + "learning_rate": 4.514605041877647e-07, + "loss": 0.3722, + "step": 16940 + }, + { + "epoch": 1.4436301661695783, + "grad_norm": 18.5, + "learning_rate": 4.513312338238232e-07, + "loss": 0.4259, + "step": 16941 + }, + { + "epoch": 1.4437153813378782, + "grad_norm": 19.25, + "learning_rate": 4.5120197789262967e-07, + "loss": 0.8735, + "step": 16942 + }, + { + "epoch": 1.4438005965061782, + "grad_norm": 12.25, + "learning_rate": 4.5107273639651947e-07, + "loss": 0.3001, + "step": 16943 + }, + { + "epoch": 1.4438858116744782, + "grad_norm": 12.375, + "learning_rate": 4.509435093378288e-07, + "loss": 0.4209, + "step": 16944 + }, + { + "epoch": 1.4439710268427781, + "grad_norm": 19.25, + "learning_rate": 4.508142967188921e-07, + "loss": 0.8443, + "step": 16945 + }, + { + "epoch": 1.444056242011078, + "grad_norm": 11.5625, + "learning_rate": 4.5068509854204494e-07, + "loss": 0.4116, + "step": 16946 + }, + { + "epoch": 1.444141457179378, + "grad_norm": 15.0, + "learning_rate": 4.5055591480962155e-07, + "loss": 0.5546, + "step": 16947 + }, + { + "epoch": 1.444226672347678, + "grad_norm": 15.875, + "learning_rate": 4.504267455239567e-07, + "loss": 0.5437, + "step": 16948 + }, + { + "epoch": 1.444311887515978, + "grad_norm": 13.4375, + "learning_rate": 4.502975906873848e-07, + "loss": 0.673, + "step": 16949 + }, + { + "epoch": 1.444397102684278, + "grad_norm": 15.3125, + "learning_rate": 4.5016845030223936e-07, + "loss": 0.6639, + "step": 16950 + }, + { + "epoch": 1.4444823178525779, + "grad_norm": 32.5, + "learning_rate": 4.500393243708542e-07, + "loss": 0.9389, + "step": 16951 + }, + { + "epoch": 1.4445675330208778, + "grad_norm": 17.875, + "learning_rate": 4.4991021289556345e-07, + "loss": 0.7824, + "step": 16952 + }, + { + "epoch": 1.4446527481891778, + "grad_norm": 16.25, + "learning_rate": 4.4978111587869924e-07, + "loss": 0.6268, + "step": 16953 + }, + { + "epoch": 1.4447379633574777, + "grad_norm": 11.9375, + "learning_rate": 4.496520333225954e-07, + "loss": 0.2983, + "step": 16954 + }, + { + "epoch": 1.4448231785257777, + "grad_norm": 15.25, + "learning_rate": 4.495229652295835e-07, + "loss": 0.5117, + "step": 16955 + }, + { + "epoch": 1.4449083936940776, + "grad_norm": 14.25, + "learning_rate": 4.493939116019967e-07, + "loss": 0.5626, + "step": 16956 + }, + { + "epoch": 1.4449936088623776, + "grad_norm": 15.0625, + "learning_rate": 4.492648724421675e-07, + "loss": 0.923, + "step": 16957 + }, + { + "epoch": 1.4450788240306776, + "grad_norm": 13.4375, + "learning_rate": 4.491358477524267e-07, + "loss": 0.5398, + "step": 16958 + }, + { + "epoch": 1.4451640391989775, + "grad_norm": 12.75, + "learning_rate": 4.490068375351066e-07, + "loss": 0.2582, + "step": 16959 + }, + { + "epoch": 1.4452492543672775, + "grad_norm": 26.375, + "learning_rate": 4.488778417925382e-07, + "loss": 1.026, + "step": 16960 + }, + { + "epoch": 1.4453344695355774, + "grad_norm": 11.75, + "learning_rate": 4.487488605270533e-07, + "loss": 0.4216, + "step": 16961 + }, + { + "epoch": 1.4454196847038774, + "grad_norm": 14.125, + "learning_rate": 4.4861989374098234e-07, + "loss": 0.5933, + "step": 16962 + }, + { + "epoch": 1.4455048998721773, + "grad_norm": 13.5625, + "learning_rate": 4.484909414366552e-07, + "loss": 0.4097, + "step": 16963 + }, + { + "epoch": 1.4455901150404773, + "grad_norm": 17.5, + "learning_rate": 4.483620036164027e-07, + "loss": 0.6403, + "step": 16964 + }, + { + "epoch": 1.4456753302087773, + "grad_norm": 14.625, + "learning_rate": 4.4823308028255546e-07, + "loss": 0.5542, + "step": 16965 + }, + { + "epoch": 1.4457605453770772, + "grad_norm": 20.875, + "learning_rate": 4.481041714374422e-07, + "loss": 0.7441, + "step": 16966 + }, + { + "epoch": 1.4458457605453772, + "grad_norm": 11.625, + "learning_rate": 4.479752770833928e-07, + "loss": 0.4468, + "step": 16967 + }, + { + "epoch": 1.4459309757136771, + "grad_norm": 17.375, + "learning_rate": 4.478463972227369e-07, + "loss": 0.7209, + "step": 16968 + }, + { + "epoch": 1.446016190881977, + "grad_norm": 15.625, + "learning_rate": 4.477175318578035e-07, + "loss": 0.63, + "step": 16969 + }, + { + "epoch": 1.446101406050277, + "grad_norm": 17.375, + "learning_rate": 4.4758868099092106e-07, + "loss": 0.8403, + "step": 16970 + }, + { + "epoch": 1.446186621218577, + "grad_norm": 18.875, + "learning_rate": 4.474598446244176e-07, + "loss": 0.8098, + "step": 16971 + }, + { + "epoch": 1.446271836386877, + "grad_norm": 17.0, + "learning_rate": 4.473310227606219e-07, + "loss": 0.7286, + "step": 16972 + }, + { + "epoch": 1.446357051555177, + "grad_norm": 15.625, + "learning_rate": 4.472022154018615e-07, + "loss": 0.6459, + "step": 16973 + }, + { + "epoch": 1.4464422667234769, + "grad_norm": 16.875, + "learning_rate": 4.4707342255046497e-07, + "loss": 0.6156, + "step": 16974 + }, + { + "epoch": 1.4465274818917768, + "grad_norm": 15.5, + "learning_rate": 4.469446442087587e-07, + "loss": 0.4726, + "step": 16975 + }, + { + "epoch": 1.4466126970600768, + "grad_norm": 19.25, + "learning_rate": 4.4681588037907023e-07, + "loss": 0.7002, + "step": 16976 + }, + { + "epoch": 1.4466979122283767, + "grad_norm": 29.125, + "learning_rate": 4.4668713106372637e-07, + "loss": 0.7807, + "step": 16977 + }, + { + "epoch": 1.4467831273966767, + "grad_norm": 17.0, + "learning_rate": 4.4655839626505424e-07, + "loss": 0.7902, + "step": 16978 + }, + { + "epoch": 1.4468683425649767, + "grad_norm": 13.75, + "learning_rate": 4.4642967598537994e-07, + "loss": 0.6579, + "step": 16979 + }, + { + "epoch": 1.4469535577332766, + "grad_norm": 17.375, + "learning_rate": 4.463009702270289e-07, + "loss": 0.8732, + "step": 16980 + }, + { + "epoch": 1.4470387729015766, + "grad_norm": 14.25, + "learning_rate": 4.461722789923277e-07, + "loss": 0.7309, + "step": 16981 + }, + { + "epoch": 1.4471239880698765, + "grad_norm": 16.625, + "learning_rate": 4.4604360228360207e-07, + "loss": 0.8281, + "step": 16982 + }, + { + "epoch": 1.4472092032381765, + "grad_norm": 15.875, + "learning_rate": 4.459149401031765e-07, + "loss": 0.7871, + "step": 16983 + }, + { + "epoch": 1.4472944184064764, + "grad_norm": 14.8125, + "learning_rate": 4.457862924533765e-07, + "loss": 0.5367, + "step": 16984 + }, + { + "epoch": 1.4473796335747764, + "grad_norm": 15.5625, + "learning_rate": 4.4565765933652694e-07, + "loss": 0.4747, + "step": 16985 + }, + { + "epoch": 1.4474648487430763, + "grad_norm": 13.4375, + "learning_rate": 4.4552904075495285e-07, + "loss": 0.5666, + "step": 16986 + }, + { + "epoch": 1.4475500639113763, + "grad_norm": 10.9375, + "learning_rate": 4.454004367109778e-07, + "loss": 0.2091, + "step": 16987 + }, + { + "epoch": 1.4476352790796763, + "grad_norm": 16.875, + "learning_rate": 4.452718472069255e-07, + "loss": 0.6444, + "step": 16988 + }, + { + "epoch": 1.4477204942479762, + "grad_norm": 19.0, + "learning_rate": 4.451432722451203e-07, + "loss": 0.5903, + "step": 16989 + }, + { + "epoch": 1.4478057094162762, + "grad_norm": 17.125, + "learning_rate": 4.4501471182788537e-07, + "loss": 0.6806, + "step": 16990 + }, + { + "epoch": 1.4478909245845761, + "grad_norm": 17.5, + "learning_rate": 4.4488616595754456e-07, + "loss": 0.5342, + "step": 16991 + }, + { + "epoch": 1.447976139752876, + "grad_norm": 16.25, + "learning_rate": 4.447576346364199e-07, + "loss": 0.4602, + "step": 16992 + }, + { + "epoch": 1.448061354921176, + "grad_norm": 19.5, + "learning_rate": 4.446291178668345e-07, + "loss": 0.8305, + "step": 16993 + }, + { + "epoch": 1.448146570089476, + "grad_norm": 16.125, + "learning_rate": 4.4450061565111084e-07, + "loss": 0.775, + "step": 16994 + }, + { + "epoch": 1.448231785257776, + "grad_norm": 17.5, + "learning_rate": 4.443721279915719e-07, + "loss": 0.7017, + "step": 16995 + }, + { + "epoch": 1.448317000426076, + "grad_norm": 14.625, + "learning_rate": 4.442436548905378e-07, + "loss": 0.5601, + "step": 16996 + }, + { + "epoch": 1.4484022155943759, + "grad_norm": 15.625, + "learning_rate": 4.441151963503312e-07, + "loss": 0.6108, + "step": 16997 + }, + { + "epoch": 1.4484874307626758, + "grad_norm": 15.3125, + "learning_rate": 4.4398675237327325e-07, + "loss": 0.5857, + "step": 16998 + }, + { + "epoch": 1.4485726459309758, + "grad_norm": 19.0, + "learning_rate": 4.438583229616858e-07, + "loss": 0.938, + "step": 16999 + }, + { + "epoch": 1.4486578610992757, + "grad_norm": 15.5625, + "learning_rate": 4.437299081178886e-07, + "loss": 0.5075, + "step": 17000 + }, + { + "epoch": 1.4487430762675757, + "grad_norm": 15.625, + "learning_rate": 4.436015078442027e-07, + "loss": 0.5078, + "step": 17001 + }, + { + "epoch": 1.4488282914358757, + "grad_norm": 16.375, + "learning_rate": 4.434731221429485e-07, + "loss": 0.4683, + "step": 17002 + }, + { + "epoch": 1.4489135066041756, + "grad_norm": 21.375, + "learning_rate": 4.433447510164464e-07, + "loss": 1.0, + "step": 17003 + }, + { + "epoch": 1.4489987217724756, + "grad_norm": 18.0, + "learning_rate": 4.4321639446701576e-07, + "loss": 0.7153, + "step": 17004 + }, + { + "epoch": 1.4490839369407755, + "grad_norm": 18.0, + "learning_rate": 4.430880524969758e-07, + "loss": 0.5156, + "step": 17005 + }, + { + "epoch": 1.4491691521090755, + "grad_norm": 16.375, + "learning_rate": 4.4295972510864606e-07, + "loss": 0.7852, + "step": 17006 + }, + { + "epoch": 1.4492543672773754, + "grad_norm": 19.375, + "learning_rate": 4.428314123043456e-07, + "loss": 0.7611, + "step": 17007 + }, + { + "epoch": 1.4493395824456754, + "grad_norm": 21.75, + "learning_rate": 4.427031140863937e-07, + "loss": 0.6999, + "step": 17008 + }, + { + "epoch": 1.4494247976139754, + "grad_norm": 13.8125, + "learning_rate": 4.42574830457108e-07, + "loss": 0.4359, + "step": 17009 + }, + { + "epoch": 1.4495100127822753, + "grad_norm": 14.25, + "learning_rate": 4.424465614188068e-07, + "loss": 0.6387, + "step": 17010 + }, + { + "epoch": 1.4495952279505753, + "grad_norm": 14.0, + "learning_rate": 4.423183069738089e-07, + "loss": 0.3323, + "step": 17011 + }, + { + "epoch": 1.4496804431188752, + "grad_norm": 18.125, + "learning_rate": 4.421900671244314e-07, + "loss": 0.8071, + "step": 17012 + }, + { + "epoch": 1.4497656582871752, + "grad_norm": 17.25, + "learning_rate": 4.420618418729912e-07, + "loss": 0.7116, + "step": 17013 + }, + { + "epoch": 1.4498508734554751, + "grad_norm": 15.8125, + "learning_rate": 4.4193363122180604e-07, + "loss": 0.7531, + "step": 17014 + }, + { + "epoch": 1.449936088623775, + "grad_norm": 14.5625, + "learning_rate": 4.4180543517319264e-07, + "loss": 0.5508, + "step": 17015 + }, + { + "epoch": 1.450021303792075, + "grad_norm": 15.75, + "learning_rate": 4.416772537294682e-07, + "loss": 0.767, + "step": 17016 + }, + { + "epoch": 1.450106518960375, + "grad_norm": 19.5, + "learning_rate": 4.415490868929481e-07, + "loss": 0.6647, + "step": 17017 + }, + { + "epoch": 1.450191734128675, + "grad_norm": 12.625, + "learning_rate": 4.414209346659491e-07, + "loss": 0.3735, + "step": 17018 + }, + { + "epoch": 1.450276949296975, + "grad_norm": 14.25, + "learning_rate": 4.412927970507873e-07, + "loss": 0.7287, + "step": 17019 + }, + { + "epoch": 1.4503621644652749, + "grad_norm": 14.3125, + "learning_rate": 4.411646740497773e-07, + "loss": 0.712, + "step": 17020 + }, + { + "epoch": 1.4504473796335748, + "grad_norm": 18.375, + "learning_rate": 4.410365656652356e-07, + "loss": 1.0275, + "step": 17021 + }, + { + "epoch": 1.4505325948018748, + "grad_norm": 13.25, + "learning_rate": 4.4090847189947605e-07, + "loss": 0.4065, + "step": 17022 + }, + { + "epoch": 1.4506178099701748, + "grad_norm": 32.25, + "learning_rate": 4.407803927548142e-07, + "loss": 0.4734, + "step": 17023 + }, + { + "epoch": 1.4507030251384747, + "grad_norm": 15.1875, + "learning_rate": 4.4065232823356425e-07, + "loss": 0.6846, + "step": 17024 + }, + { + "epoch": 1.4507882403067747, + "grad_norm": 19.125, + "learning_rate": 4.4052427833804104e-07, + "loss": 0.8041, + "step": 17025 + }, + { + "epoch": 1.4508734554750746, + "grad_norm": 16.125, + "learning_rate": 4.403962430705577e-07, + "loss": 0.7057, + "step": 17026 + }, + { + "epoch": 1.4509586706433746, + "grad_norm": 16.875, + "learning_rate": 4.4026822243342884e-07, + "loss": 0.4326, + "step": 17027 + }, + { + "epoch": 1.4510438858116745, + "grad_norm": 10.9375, + "learning_rate": 4.40140216428967e-07, + "loss": 0.2751, + "step": 17028 + }, + { + "epoch": 1.4511291009799745, + "grad_norm": 14.625, + "learning_rate": 4.400122250594861e-07, + "loss": 0.31, + "step": 17029 + }, + { + "epoch": 1.4512143161482745, + "grad_norm": 17.5, + "learning_rate": 4.398842483272986e-07, + "loss": 0.5284, + "step": 17030 + }, + { + "epoch": 1.4512995313165744, + "grad_norm": 28.875, + "learning_rate": 4.3975628623471727e-07, + "loss": 0.7022, + "step": 17031 + }, + { + "epoch": 1.4513847464848744, + "grad_norm": 17.875, + "learning_rate": 4.3962833878405455e-07, + "loss": 0.4179, + "step": 17032 + }, + { + "epoch": 1.4514699616531743, + "grad_norm": 15.625, + "learning_rate": 4.395004059776231e-07, + "loss": 0.5578, + "step": 17033 + }, + { + "epoch": 1.4515551768214743, + "grad_norm": 25.0, + "learning_rate": 4.39372487817734e-07, + "loss": 0.6071, + "step": 17034 + }, + { + "epoch": 1.4516403919897742, + "grad_norm": 22.75, + "learning_rate": 4.3924458430669947e-07, + "loss": 0.5828, + "step": 17035 + }, + { + "epoch": 1.4517256071580742, + "grad_norm": 31.0, + "learning_rate": 4.3911669544683027e-07, + "loss": 0.8433, + "step": 17036 + }, + { + "epoch": 1.4518108223263742, + "grad_norm": 13.1875, + "learning_rate": 4.3898882124043777e-07, + "loss": 0.5233, + "step": 17037 + }, + { + "epoch": 1.451896037494674, + "grad_norm": 16.5, + "learning_rate": 4.388609616898333e-07, + "loss": 0.6398, + "step": 17038 + }, + { + "epoch": 1.451981252662974, + "grad_norm": 12.875, + "learning_rate": 4.3873311679732634e-07, + "loss": 0.4237, + "step": 17039 + }, + { + "epoch": 1.452066467831274, + "grad_norm": 20.0, + "learning_rate": 4.386052865652278e-07, + "loss": 0.604, + "step": 17040 + }, + { + "epoch": 1.452151682999574, + "grad_norm": 18.0, + "learning_rate": 4.3847747099584766e-07, + "loss": 0.8829, + "step": 17041 + }, + { + "epoch": 1.452236898167874, + "grad_norm": 21.5, + "learning_rate": 4.3834967009149615e-07, + "loss": 0.714, + "step": 17042 + }, + { + "epoch": 1.452322113336174, + "grad_norm": 15.4375, + "learning_rate": 4.3822188385448214e-07, + "loss": 0.6461, + "step": 17043 + }, + { + "epoch": 1.4524073285044738, + "grad_norm": 16.125, + "learning_rate": 4.3809411228711475e-07, + "loss": 0.7615, + "step": 17044 + }, + { + "epoch": 1.4524925436727738, + "grad_norm": 14.9375, + "learning_rate": 4.379663553917031e-07, + "loss": 0.6451, + "step": 17045 + }, + { + "epoch": 1.4525777588410738, + "grad_norm": 15.1875, + "learning_rate": 4.3783861317055644e-07, + "loss": 0.9955, + "step": 17046 + }, + { + "epoch": 1.4526629740093737, + "grad_norm": 11.25, + "learning_rate": 4.3771088562598223e-07, + "loss": 0.2094, + "step": 17047 + }, + { + "epoch": 1.4527481891776737, + "grad_norm": 16.375, + "learning_rate": 4.3758317276028915e-07, + "loss": 0.6751, + "step": 17048 + }, + { + "epoch": 1.4528334043459736, + "grad_norm": 13.5, + "learning_rate": 4.3745547457578515e-07, + "loss": 0.3459, + "step": 17049 + }, + { + "epoch": 1.4529186195142736, + "grad_norm": 12.4375, + "learning_rate": 4.3732779107477807e-07, + "loss": 0.2961, + "step": 17050 + }, + { + "epoch": 1.4530038346825735, + "grad_norm": 16.375, + "learning_rate": 4.3720012225957525e-07, + "loss": 0.7253, + "step": 17051 + }, + { + "epoch": 1.4530890498508735, + "grad_norm": 15.8125, + "learning_rate": 4.37072468132483e-07, + "loss": 0.7138, + "step": 17052 + }, + { + "epoch": 1.4531742650191735, + "grad_norm": 11.5, + "learning_rate": 4.3694482869580865e-07, + "loss": 0.292, + "step": 17053 + }, + { + "epoch": 1.4532594801874734, + "grad_norm": 16.375, + "learning_rate": 4.36817203951859e-07, + "loss": 0.5646, + "step": 17054 + }, + { + "epoch": 1.4533446953557734, + "grad_norm": 16.875, + "learning_rate": 4.3668959390294044e-07, + "loss": 0.7833, + "step": 17055 + }, + { + "epoch": 1.4534299105240733, + "grad_norm": 13.375, + "learning_rate": 4.3656199855135834e-07, + "loss": 0.4672, + "step": 17056 + }, + { + "epoch": 1.4535151256923733, + "grad_norm": 14.375, + "learning_rate": 4.36434417899419e-07, + "loss": 0.6292, + "step": 17057 + }, + { + "epoch": 1.4536003408606732, + "grad_norm": 19.5, + "learning_rate": 4.363068519494283e-07, + "loss": 0.774, + "step": 17058 + }, + { + "epoch": 1.4536855560289732, + "grad_norm": 16.875, + "learning_rate": 4.3617930070369044e-07, + "loss": 0.6635, + "step": 17059 + }, + { + "epoch": 1.4537707711972732, + "grad_norm": 16.5, + "learning_rate": 4.360517641645115e-07, + "loss": 0.5252, + "step": 17060 + }, + { + "epoch": 1.4538559863655731, + "grad_norm": 14.625, + "learning_rate": 4.359242423341953e-07, + "loss": 0.4916, + "step": 17061 + }, + { + "epoch": 1.453941201533873, + "grad_norm": 13.6875, + "learning_rate": 4.357967352150466e-07, + "loss": 0.4099, + "step": 17062 + }, + { + "epoch": 1.454026416702173, + "grad_norm": 18.375, + "learning_rate": 4.3566924280937004e-07, + "loss": 0.8215, + "step": 17063 + }, + { + "epoch": 1.454111631870473, + "grad_norm": 15.25, + "learning_rate": 4.3554176511946864e-07, + "loss": 0.554, + "step": 17064 + }, + { + "epoch": 1.454196847038773, + "grad_norm": 11.5, + "learning_rate": 4.3541430214764673e-07, + "loss": 0.3404, + "step": 17065 + }, + { + "epoch": 1.454282062207073, + "grad_norm": 11.75, + "learning_rate": 4.352868538962074e-07, + "loss": 0.2362, + "step": 17066 + }, + { + "epoch": 1.4543672773753729, + "grad_norm": 15.8125, + "learning_rate": 4.3515942036745433e-07, + "loss": 0.6827, + "step": 17067 + }, + { + "epoch": 1.4544524925436728, + "grad_norm": 17.875, + "learning_rate": 4.3503200156368984e-07, + "loss": 0.404, + "step": 17068 + }, + { + "epoch": 1.4545377077119728, + "grad_norm": 18.625, + "learning_rate": 4.3490459748721627e-07, + "loss": 0.7319, + "step": 17069 + }, + { + "epoch": 1.4546229228802727, + "grad_norm": 15.0, + "learning_rate": 4.3477720814033617e-07, + "loss": 0.5502, + "step": 17070 + }, + { + "epoch": 1.4547081380485727, + "grad_norm": 15.4375, + "learning_rate": 4.346498335253517e-07, + "loss": 0.4936, + "step": 17071 + }, + { + "epoch": 1.4547933532168726, + "grad_norm": 10.125, + "learning_rate": 4.3452247364456526e-07, + "loss": 0.1563, + "step": 17072 + }, + { + "epoch": 1.4548785683851726, + "grad_norm": 36.5, + "learning_rate": 4.34395128500277e-07, + "loss": 0.7963, + "step": 17073 + }, + { + "epoch": 1.4549637835534726, + "grad_norm": 12.1875, + "learning_rate": 4.3426779809478917e-07, + "loss": 0.3361, + "step": 17074 + }, + { + "epoch": 1.4550489987217725, + "grad_norm": 10.875, + "learning_rate": 4.3414048243040285e-07, + "loss": 0.2579, + "step": 17075 + }, + { + "epoch": 1.4551342138900725, + "grad_norm": 17.125, + "learning_rate": 4.340131815094184e-07, + "loss": 0.7359, + "step": 17076 + }, + { + "epoch": 1.4552194290583724, + "grad_norm": 15.75, + "learning_rate": 4.33885895334136e-07, + "loss": 0.816, + "step": 17077 + }, + { + "epoch": 1.4553046442266724, + "grad_norm": 19.5, + "learning_rate": 4.33758623906856e-07, + "loss": 0.6487, + "step": 17078 + }, + { + "epoch": 1.4553898593949723, + "grad_norm": 11.4375, + "learning_rate": 4.336313672298785e-07, + "loss": 0.3345, + "step": 17079 + }, + { + "epoch": 1.4554750745632723, + "grad_norm": 16.625, + "learning_rate": 4.335041253055035e-07, + "loss": 0.5754, + "step": 17080 + }, + { + "epoch": 1.4555602897315723, + "grad_norm": 20.125, + "learning_rate": 4.3337689813602963e-07, + "loss": 1.0234, + "step": 17081 + }, + { + "epoch": 1.4556455048998722, + "grad_norm": 13.25, + "learning_rate": 4.332496857237564e-07, + "loss": 0.4606, + "step": 17082 + }, + { + "epoch": 1.4557307200681722, + "grad_norm": 21.375, + "learning_rate": 4.3312248807098266e-07, + "loss": 0.4684, + "step": 17083 + }, + { + "epoch": 1.4558159352364721, + "grad_norm": 15.6875, + "learning_rate": 4.329953051800073e-07, + "loss": 0.5979, + "step": 17084 + }, + { + "epoch": 1.455901150404772, + "grad_norm": 19.75, + "learning_rate": 4.3286813705312845e-07, + "loss": 0.6487, + "step": 17085 + }, + { + "epoch": 1.455986365573072, + "grad_norm": 18.0, + "learning_rate": 4.327409836926437e-07, + "loss": 0.8636, + "step": 17086 + }, + { + "epoch": 1.456071580741372, + "grad_norm": 15.125, + "learning_rate": 4.32613845100851e-07, + "loss": 0.5902, + "step": 17087 + }, + { + "epoch": 1.456156795909672, + "grad_norm": 20.625, + "learning_rate": 4.324867212800482e-07, + "loss": 0.9098, + "step": 17088 + }, + { + "epoch": 1.456242011077972, + "grad_norm": 16.625, + "learning_rate": 4.3235961223253284e-07, + "loss": 0.917, + "step": 17089 + }, + { + "epoch": 1.4563272262462719, + "grad_norm": 15.875, + "learning_rate": 4.3223251796060103e-07, + "loss": 0.3108, + "step": 17090 + }, + { + "epoch": 1.4564124414145718, + "grad_norm": 11.6875, + "learning_rate": 4.321054384665499e-07, + "loss": 0.2126, + "step": 17091 + }, + { + "epoch": 1.4564976565828718, + "grad_norm": 21.125, + "learning_rate": 4.3197837375267645e-07, + "loss": 1.1543, + "step": 17092 + }, + { + "epoch": 1.4565828717511717, + "grad_norm": 19.625, + "learning_rate": 4.318513238212764e-07, + "loss": 0.7058, + "step": 17093 + }, + { + "epoch": 1.4566680869194717, + "grad_norm": 15.6875, + "learning_rate": 4.3172428867464517e-07, + "loss": 0.6422, + "step": 17094 + }, + { + "epoch": 1.4567533020877717, + "grad_norm": 13.0625, + "learning_rate": 4.315972683150789e-07, + "loss": 0.3985, + "step": 17095 + }, + { + "epoch": 1.4568385172560716, + "grad_norm": 15.1875, + "learning_rate": 4.3147026274487287e-07, + "loss": 0.6573, + "step": 17096 + }, + { + "epoch": 1.4569237324243716, + "grad_norm": 16.875, + "learning_rate": 4.313432719663228e-07, + "loss": 0.9987, + "step": 17097 + }, + { + "epoch": 1.4570089475926715, + "grad_norm": 20.25, + "learning_rate": 4.312162959817226e-07, + "loss": 0.8918, + "step": 17098 + }, + { + "epoch": 1.4570941627609715, + "grad_norm": 11.375, + "learning_rate": 4.3108933479336727e-07, + "loss": 0.3491, + "step": 17099 + }, + { + "epoch": 1.4571793779292714, + "grad_norm": 18.0, + "learning_rate": 4.309623884035516e-07, + "loss": 0.6478, + "step": 17100 + }, + { + "epoch": 1.4572645930975714, + "grad_norm": 19.5, + "learning_rate": 4.3083545681456866e-07, + "loss": 0.7967, + "step": 17101 + }, + { + "epoch": 1.4573498082658713, + "grad_norm": 14.875, + "learning_rate": 4.307085400287132e-07, + "loss": 0.756, + "step": 17102 + }, + { + "epoch": 1.4574350234341713, + "grad_norm": 19.5, + "learning_rate": 4.30581638048278e-07, + "loss": 0.3875, + "step": 17103 + }, + { + "epoch": 1.4575202386024713, + "grad_norm": 12.5625, + "learning_rate": 4.304547508755565e-07, + "loss": 0.2657, + "step": 17104 + }, + { + "epoch": 1.4576054537707712, + "grad_norm": 13.6875, + "learning_rate": 4.303278785128422e-07, + "loss": 0.7003, + "step": 17105 + }, + { + "epoch": 1.4576906689390712, + "grad_norm": 16.0, + "learning_rate": 4.3020102096242686e-07, + "loss": 0.5702, + "step": 17106 + }, + { + "epoch": 1.4577758841073711, + "grad_norm": 17.125, + "learning_rate": 4.3007417822660355e-07, + "loss": 0.8113, + "step": 17107 + }, + { + "epoch": 1.457861099275671, + "grad_norm": 23.125, + "learning_rate": 4.299473503076647e-07, + "loss": 0.6691, + "step": 17108 + }, + { + "epoch": 1.457946314443971, + "grad_norm": 15.875, + "learning_rate": 4.2982053720790144e-07, + "loss": 0.6064, + "step": 17109 + }, + { + "epoch": 1.458031529612271, + "grad_norm": 13.8125, + "learning_rate": 4.2969373892960615e-07, + "loss": 0.4852, + "step": 17110 + }, + { + "epoch": 1.458116744780571, + "grad_norm": 12.0625, + "learning_rate": 4.295669554750696e-07, + "loss": 0.4579, + "step": 17111 + }, + { + "epoch": 1.458201959948871, + "grad_norm": 15.375, + "learning_rate": 4.294401868465832e-07, + "loss": 0.5323, + "step": 17112 + }, + { + "epoch": 1.4582871751171709, + "grad_norm": 16.625, + "learning_rate": 4.2931343304643766e-07, + "loss": 0.513, + "step": 17113 + }, + { + "epoch": 1.4583723902854708, + "grad_norm": 24.375, + "learning_rate": 4.2918669407692416e-07, + "loss": 0.7668, + "step": 17114 + }, + { + "epoch": 1.4584576054537708, + "grad_norm": 15.5625, + "learning_rate": 4.290599699403321e-07, + "loss": 0.5011, + "step": 17115 + }, + { + "epoch": 1.4585428206220707, + "grad_norm": 16.5, + "learning_rate": 4.289332606389522e-07, + "loss": 0.8973, + "step": 17116 + }, + { + "epoch": 1.4586280357903707, + "grad_norm": 23.0, + "learning_rate": 4.288065661750736e-07, + "loss": 0.5071, + "step": 17117 + }, + { + "epoch": 1.4587132509586707, + "grad_norm": 11.3125, + "learning_rate": 4.28679886550986e-07, + "loss": 0.2919, + "step": 17118 + }, + { + "epoch": 1.4587984661269706, + "grad_norm": 28.625, + "learning_rate": 4.2855322176897917e-07, + "loss": 0.9846, + "step": 17119 + }, + { + "epoch": 1.4588836812952706, + "grad_norm": 16.25, + "learning_rate": 4.2842657183134143e-07, + "loss": 0.8774, + "step": 17120 + }, + { + "epoch": 1.4589688964635705, + "grad_norm": 13.0, + "learning_rate": 4.2829993674036145e-07, + "loss": 0.4633, + "step": 17121 + }, + { + "epoch": 1.4590541116318705, + "grad_norm": 13.4375, + "learning_rate": 4.2817331649832845e-07, + "loss": 0.5004, + "step": 17122 + }, + { + "epoch": 1.4591393268001704, + "grad_norm": 17.25, + "learning_rate": 4.2804671110752967e-07, + "loss": 0.7934, + "step": 17123 + }, + { + "epoch": 1.4592245419684704, + "grad_norm": 15.1875, + "learning_rate": 4.2792012057025323e-07, + "loss": 0.668, + "step": 17124 + }, + { + "epoch": 1.4593097571367704, + "grad_norm": 13.125, + "learning_rate": 4.2779354488878737e-07, + "loss": 0.4754, + "step": 17125 + }, + { + "epoch": 1.4593949723050703, + "grad_norm": 16.125, + "learning_rate": 4.276669840654185e-07, + "loss": 0.5229, + "step": 17126 + }, + { + "epoch": 1.4594801874733703, + "grad_norm": 19.875, + "learning_rate": 4.2754043810243464e-07, + "loss": 0.8423, + "step": 17127 + }, + { + "epoch": 1.4595654026416702, + "grad_norm": 20.5, + "learning_rate": 4.274139070021216e-07, + "loss": 0.4509, + "step": 17128 + }, + { + "epoch": 1.4596506178099702, + "grad_norm": 34.0, + "learning_rate": 4.2728739076676643e-07, + "loss": 1.3164, + "step": 17129 + }, + { + "epoch": 1.4597358329782701, + "grad_norm": 10.75, + "learning_rate": 4.271608893986555e-07, + "loss": 0.3478, + "step": 17130 + }, + { + "epoch": 1.45982104814657, + "grad_norm": 35.0, + "learning_rate": 4.27034402900075e-07, + "loss": 0.9832, + "step": 17131 + }, + { + "epoch": 1.45990626331487, + "grad_norm": 19.625, + "learning_rate": 4.269079312733101e-07, + "loss": 0.6734, + "step": 17132 + }, + { + "epoch": 1.45999147848317, + "grad_norm": 18.5, + "learning_rate": 4.267814745206469e-07, + "loss": 0.8575, + "step": 17133 + }, + { + "epoch": 1.46007669365147, + "grad_norm": 37.75, + "learning_rate": 4.2665503264436995e-07, + "loss": 0.8007, + "step": 17134 + }, + { + "epoch": 1.46016190881977, + "grad_norm": 14.0, + "learning_rate": 4.265286056467649e-07, + "loss": 0.6154, + "step": 17135 + }, + { + "epoch": 1.4602471239880699, + "grad_norm": 16.25, + "learning_rate": 4.264021935301156e-07, + "loss": 0.5894, + "step": 17136 + }, + { + "epoch": 1.4603323391563698, + "grad_norm": 18.0, + "learning_rate": 4.2627579629670677e-07, + "loss": 0.8026, + "step": 17137 + }, + { + "epoch": 1.4604175543246698, + "grad_norm": 17.875, + "learning_rate": 4.2614941394882257e-07, + "loss": 0.7687, + "step": 17138 + }, + { + "epoch": 1.4605027694929698, + "grad_norm": 19.125, + "learning_rate": 4.260230464887474e-07, + "loss": 0.8157, + "step": 17139 + }, + { + "epoch": 1.4605879846612697, + "grad_norm": 25.0, + "learning_rate": 4.258966939187641e-07, + "loss": 1.0111, + "step": 17140 + }, + { + "epoch": 1.4606731998295697, + "grad_norm": 11.0, + "learning_rate": 4.2577035624115653e-07, + "loss": 0.328, + "step": 17141 + }, + { + "epoch": 1.4607584149978696, + "grad_norm": 18.375, + "learning_rate": 4.25644033458207e-07, + "loss": 0.4941, + "step": 17142 + }, + { + "epoch": 1.4608436301661696, + "grad_norm": 15.75, + "learning_rate": 4.2551772557219873e-07, + "loss": 0.342, + "step": 17143 + }, + { + "epoch": 1.4609288453344695, + "grad_norm": 14.3125, + "learning_rate": 4.2539143258541457e-07, + "loss": 0.3734, + "step": 17144 + }, + { + "epoch": 1.4610140605027695, + "grad_norm": 16.125, + "learning_rate": 4.2526515450013616e-07, + "loss": 0.584, + "step": 17145 + }, + { + "epoch": 1.4610992756710695, + "grad_norm": 17.5, + "learning_rate": 4.2513889131864576e-07, + "loss": 0.9631, + "step": 17146 + }, + { + "epoch": 1.4611844908393694, + "grad_norm": 21.375, + "learning_rate": 4.2501264304322496e-07, + "loss": 0.7708, + "step": 17147 + }, + { + "epoch": 1.4612697060076694, + "grad_norm": 18.25, + "learning_rate": 4.248864096761557e-07, + "loss": 0.8378, + "step": 17148 + }, + { + "epoch": 1.4613549211759693, + "grad_norm": 11.125, + "learning_rate": 4.2476019121971887e-07, + "loss": 0.2903, + "step": 17149 + }, + { + "epoch": 1.4614401363442693, + "grad_norm": 14.5625, + "learning_rate": 4.2463398767619473e-07, + "loss": 0.4683, + "step": 17150 + }, + { + "epoch": 1.4615253515125692, + "grad_norm": 12.875, + "learning_rate": 4.245077990478644e-07, + "loss": 0.5628, + "step": 17151 + }, + { + "epoch": 1.4616105666808692, + "grad_norm": 17.0, + "learning_rate": 4.2438162533700853e-07, + "loss": 0.8071, + "step": 17152 + }, + { + "epoch": 1.4616957818491692, + "grad_norm": 15.4375, + "learning_rate": 4.2425546654590673e-07, + "loss": 0.4636, + "step": 17153 + }, + { + "epoch": 1.461780997017469, + "grad_norm": 15.25, + "learning_rate": 4.241293226768388e-07, + "loss": 0.4963, + "step": 17154 + }, + { + "epoch": 1.461866212185769, + "grad_norm": 11.5, + "learning_rate": 4.240031937320846e-07, + "loss": 0.2781, + "step": 17155 + }, + { + "epoch": 1.461951427354069, + "grad_norm": 22.5, + "learning_rate": 4.238770797139237e-07, + "loss": 0.5648, + "step": 17156 + }, + { + "epoch": 1.462036642522369, + "grad_norm": 13.9375, + "learning_rate": 4.237509806246347e-07, + "loss": 0.6565, + "step": 17157 + }, + { + "epoch": 1.462121857690669, + "grad_norm": 13.875, + "learning_rate": 4.236248964664959e-07, + "loss": 0.3532, + "step": 17158 + }, + { + "epoch": 1.462207072858969, + "grad_norm": 15.875, + "learning_rate": 4.2349882724178617e-07, + "loss": 0.6163, + "step": 17159 + }, + { + "epoch": 1.4622922880272688, + "grad_norm": 14.125, + "learning_rate": 4.233727729527837e-07, + "loss": 0.4131, + "step": 17160 + }, + { + "epoch": 1.4623775031955688, + "grad_norm": 16.375, + "learning_rate": 4.2324673360176703e-07, + "loss": 0.602, + "step": 17161 + }, + { + "epoch": 1.4624627183638688, + "grad_norm": 16.75, + "learning_rate": 4.231207091910128e-07, + "loss": 0.6405, + "step": 17162 + }, + { + "epoch": 1.4625479335321687, + "grad_norm": 12.875, + "learning_rate": 4.229946997227988e-07, + "loss": 0.6455, + "step": 17163 + }, + { + "epoch": 1.4626331487004687, + "grad_norm": 10.375, + "learning_rate": 4.228687051994021e-07, + "loss": 0.3049, + "step": 17164 + }, + { + "epoch": 1.4627183638687686, + "grad_norm": 25.625, + "learning_rate": 4.2274272562310065e-07, + "loss": 1.061, + "step": 17165 + }, + { + "epoch": 1.4628035790370686, + "grad_norm": 17.25, + "learning_rate": 4.22616760996169e-07, + "loss": 0.6431, + "step": 17166 + }, + { + "epoch": 1.4628887942053685, + "grad_norm": 14.5, + "learning_rate": 4.224908113208846e-07, + "loss": 0.6035, + "step": 17167 + }, + { + "epoch": 1.4629740093736685, + "grad_norm": 13.5625, + "learning_rate": 4.2236487659952334e-07, + "loss": 0.4164, + "step": 17168 + }, + { + "epoch": 1.4630592245419685, + "grad_norm": 12.5625, + "learning_rate": 4.2223895683436157e-07, + "loss": 0.5536, + "step": 17169 + }, + { + "epoch": 1.4631444397102684, + "grad_norm": 27.125, + "learning_rate": 4.2211305202767364e-07, + "loss": 0.6174, + "step": 17170 + }, + { + "epoch": 1.4632296548785684, + "grad_norm": 10.875, + "learning_rate": 4.219871621817354e-07, + "loss": 0.4187, + "step": 17171 + }, + { + "epoch": 1.4633148700468683, + "grad_norm": 17.75, + "learning_rate": 4.2186128729882174e-07, + "loss": 0.6627, + "step": 17172 + }, + { + "epoch": 1.4634000852151683, + "grad_norm": 22.75, + "learning_rate": 4.2173542738120786e-07, + "loss": 0.3031, + "step": 17173 + }, + { + "epoch": 1.4634853003834682, + "grad_norm": 11.625, + "learning_rate": 4.2160958243116766e-07, + "loss": 0.3564, + "step": 17174 + }, + { + "epoch": 1.4635705155517682, + "grad_norm": 18.0, + "learning_rate": 4.214837524509749e-07, + "loss": 0.6727, + "step": 17175 + }, + { + "epoch": 1.4636557307200682, + "grad_norm": 17.75, + "learning_rate": 4.213579374429039e-07, + "loss": 0.4214, + "step": 17176 + }, + { + "epoch": 1.4637409458883681, + "grad_norm": 9.75, + "learning_rate": 4.212321374092282e-07, + "loss": 0.228, + "step": 17177 + }, + { + "epoch": 1.463826161056668, + "grad_norm": 9.3125, + "learning_rate": 4.211063523522216e-07, + "loss": 0.1683, + "step": 17178 + }, + { + "epoch": 1.463911376224968, + "grad_norm": 13.3125, + "learning_rate": 4.2098058227415624e-07, + "loss": 0.3877, + "step": 17179 + }, + { + "epoch": 1.463996591393268, + "grad_norm": 15.6875, + "learning_rate": 4.2085482717730554e-07, + "loss": 0.449, + "step": 17180 + }, + { + "epoch": 1.464081806561568, + "grad_norm": 13.6875, + "learning_rate": 4.2072908706394215e-07, + "loss": 0.5855, + "step": 17181 + }, + { + "epoch": 1.464167021729868, + "grad_norm": 12.5, + "learning_rate": 4.2060336193633813e-07, + "loss": 0.4783, + "step": 17182 + }, + { + "epoch": 1.4642522368981679, + "grad_norm": 24.75, + "learning_rate": 4.20477651796765e-07, + "loss": 0.8068, + "step": 17183 + }, + { + "epoch": 1.4643374520664678, + "grad_norm": 13.6875, + "learning_rate": 4.203519566474948e-07, + "loss": 0.3929, + "step": 17184 + }, + { + "epoch": 1.4644226672347678, + "grad_norm": 11.6875, + "learning_rate": 4.20226276490799e-07, + "loss": 0.257, + "step": 17185 + }, + { + "epoch": 1.4645078824030677, + "grad_norm": 20.25, + "learning_rate": 4.201006113289492e-07, + "loss": 0.6328, + "step": 17186 + }, + { + "epoch": 1.4645930975713677, + "grad_norm": 18.5, + "learning_rate": 4.199749611642155e-07, + "loss": 0.6912, + "step": 17187 + }, + { + "epoch": 1.4646783127396676, + "grad_norm": 13.0, + "learning_rate": 4.19849325998869e-07, + "loss": 0.4515, + "step": 17188 + }, + { + "epoch": 1.4647635279079676, + "grad_norm": 20.375, + "learning_rate": 4.1972370583517983e-07, + "loss": 0.4455, + "step": 17189 + }, + { + "epoch": 1.4648487430762676, + "grad_norm": 19.0, + "learning_rate": 4.1959810067541863e-07, + "loss": 0.6437, + "step": 17190 + }, + { + "epoch": 1.4649339582445675, + "grad_norm": 17.5, + "learning_rate": 4.1947251052185486e-07, + "loss": 0.7048, + "step": 17191 + }, + { + "epoch": 1.4650191734128675, + "grad_norm": 15.8125, + "learning_rate": 4.193469353767575e-07, + "loss": 0.8161, + "step": 17192 + }, + { + "epoch": 1.4651043885811674, + "grad_norm": 16.125, + "learning_rate": 4.1922137524239624e-07, + "loss": 0.5937, + "step": 17193 + }, + { + "epoch": 1.4651896037494674, + "grad_norm": 12.9375, + "learning_rate": 4.190958301210403e-07, + "loss": 0.5203, + "step": 17194 + }, + { + "epoch": 1.4652748189177673, + "grad_norm": 16.625, + "learning_rate": 4.1897030001495863e-07, + "loss": 0.5454, + "step": 17195 + }, + { + "epoch": 1.4653600340860673, + "grad_norm": 12.6875, + "learning_rate": 4.188447849264188e-07, + "loss": 0.4373, + "step": 17196 + }, + { + "epoch": 1.4654452492543673, + "grad_norm": 14.125, + "learning_rate": 4.187192848576896e-07, + "loss": 0.3678, + "step": 17197 + }, + { + "epoch": 1.4655304644226672, + "grad_norm": 16.75, + "learning_rate": 4.1859379981103914e-07, + "loss": 0.9469, + "step": 17198 + }, + { + "epoch": 1.4656156795909672, + "grad_norm": 15.25, + "learning_rate": 4.184683297887349e-07, + "loss": 0.5068, + "step": 17199 + }, + { + "epoch": 1.4657008947592671, + "grad_norm": 29.5, + "learning_rate": 4.1834287479304346e-07, + "loss": 0.6375, + "step": 17200 + }, + { + "epoch": 1.465786109927567, + "grad_norm": 13.3125, + "learning_rate": 4.1821743482623284e-07, + "loss": 0.4469, + "step": 17201 + }, + { + "epoch": 1.465871325095867, + "grad_norm": 14.125, + "learning_rate": 4.180920098905694e-07, + "loss": 0.411, + "step": 17202 + }, + { + "epoch": 1.465956540264167, + "grad_norm": 21.875, + "learning_rate": 4.179665999883205e-07, + "loss": 0.7702, + "step": 17203 + }, + { + "epoch": 1.466041755432467, + "grad_norm": 16.625, + "learning_rate": 4.178412051217512e-07, + "loss": 0.5243, + "step": 17204 + }, + { + "epoch": 1.466126970600767, + "grad_norm": 9.125, + "learning_rate": 4.177158252931282e-07, + "loss": 0.2955, + "step": 17205 + }, + { + "epoch": 1.4662121857690669, + "grad_norm": 15.0, + "learning_rate": 4.1759046050471765e-07, + "loss": 0.4988, + "step": 17206 + }, + { + "epoch": 1.4662974009373668, + "grad_norm": 12.375, + "learning_rate": 4.1746511075878407e-07, + "loss": 0.4868, + "step": 17207 + }, + { + "epoch": 1.4663826161056668, + "grad_norm": 17.75, + "learning_rate": 4.173397760575934e-07, + "loss": 0.545, + "step": 17208 + }, + { + "epoch": 1.4664678312739667, + "grad_norm": 15.6875, + "learning_rate": 4.1721445640341e-07, + "loss": 0.4494, + "step": 17209 + }, + { + "epoch": 1.4665530464422667, + "grad_norm": 19.25, + "learning_rate": 4.1708915179849873e-07, + "loss": 0.5919, + "step": 17210 + }, + { + "epoch": 1.4666382616105667, + "grad_norm": 15.75, + "learning_rate": 4.1696386224512416e-07, + "loss": 0.7615, + "step": 17211 + }, + { + "epoch": 1.4667234767788666, + "grad_norm": 17.75, + "learning_rate": 4.1683858774555065e-07, + "loss": 0.7595, + "step": 17212 + }, + { + "epoch": 1.4668086919471666, + "grad_norm": 17.375, + "learning_rate": 4.1671332830204123e-07, + "loss": 0.5004, + "step": 17213 + }, + { + "epoch": 1.4668939071154665, + "grad_norm": 17.375, + "learning_rate": 4.1658808391686026e-07, + "loss": 0.6223, + "step": 17214 + }, + { + "epoch": 1.4669791222837665, + "grad_norm": 18.375, + "learning_rate": 4.164628545922704e-07, + "loss": 0.8085, + "step": 17215 + }, + { + "epoch": 1.4670643374520664, + "grad_norm": 16.125, + "learning_rate": 4.1633764033053523e-07, + "loss": 0.7252, + "step": 17216 + }, + { + "epoch": 1.4671495526203664, + "grad_norm": 18.875, + "learning_rate": 4.162124411339169e-07, + "loss": 0.7536, + "step": 17217 + }, + { + "epoch": 1.4672347677886663, + "grad_norm": 18.25, + "learning_rate": 4.1608725700467813e-07, + "loss": 1.0706, + "step": 17218 + }, + { + "epoch": 1.4673199829569663, + "grad_norm": 13.5625, + "learning_rate": 4.1596208794508127e-07, + "loss": 0.4806, + "step": 17219 + }, + { + "epoch": 1.4674051981252663, + "grad_norm": 15.5, + "learning_rate": 4.158369339573884e-07, + "loss": 0.5664, + "step": 17220 + }, + { + "epoch": 1.4674904132935662, + "grad_norm": 17.875, + "learning_rate": 4.1571179504386064e-07, + "loss": 0.5412, + "step": 17221 + }, + { + "epoch": 1.4675756284618662, + "grad_norm": 19.125, + "learning_rate": 4.1558667120676024e-07, + "loss": 0.7067, + "step": 17222 + }, + { + "epoch": 1.4676608436301661, + "grad_norm": 13.0, + "learning_rate": 4.1546156244834713e-07, + "loss": 0.1982, + "step": 17223 + }, + { + "epoch": 1.467746058798466, + "grad_norm": 21.625, + "learning_rate": 4.153364687708829e-07, + "loss": 0.6483, + "step": 17224 + }, + { + "epoch": 1.467831273966766, + "grad_norm": 16.75, + "learning_rate": 4.152113901766283e-07, + "loss": 0.7269, + "step": 17225 + }, + { + "epoch": 1.467916489135066, + "grad_norm": 13.25, + "learning_rate": 4.1508632666784303e-07, + "loss": 0.4097, + "step": 17226 + }, + { + "epoch": 1.468001704303366, + "grad_norm": 11.1875, + "learning_rate": 4.1496127824678737e-07, + "loss": 0.3734, + "step": 17227 + }, + { + "epoch": 1.468086919471666, + "grad_norm": 47.5, + "learning_rate": 4.1483624491572156e-07, + "loss": 0.8318, + "step": 17228 + }, + { + "epoch": 1.4681721346399659, + "grad_norm": 10.9375, + "learning_rate": 4.147112266769042e-07, + "loss": 0.3817, + "step": 17229 + }, + { + "epoch": 1.4682573498082658, + "grad_norm": 13.375, + "learning_rate": 4.145862235325952e-07, + "loss": 0.561, + "step": 17230 + }, + { + "epoch": 1.4683425649765658, + "grad_norm": 16.75, + "learning_rate": 4.14461235485053e-07, + "loss": 0.5147, + "step": 17231 + }, + { + "epoch": 1.4684277801448657, + "grad_norm": 18.625, + "learning_rate": 4.1433626253653633e-07, + "loss": 0.8197, + "step": 17232 + }, + { + "epoch": 1.4685129953131657, + "grad_norm": 13.1875, + "learning_rate": 4.1421130468930426e-07, + "loss": 0.4453, + "step": 17233 + }, + { + "epoch": 1.4685982104814657, + "grad_norm": 21.125, + "learning_rate": 4.140863619456138e-07, + "loss": 0.9174, + "step": 17234 + }, + { + "epoch": 1.4686834256497656, + "grad_norm": 19.125, + "learning_rate": 4.1396143430772347e-07, + "loss": 0.8058, + "step": 17235 + }, + { + "epoch": 1.4687686408180656, + "grad_norm": 20.875, + "learning_rate": 4.138365217778907e-07, + "loss": 0.6684, + "step": 17236 + }, + { + "epoch": 1.4688538559863655, + "grad_norm": 14.9375, + "learning_rate": 4.1371162435837316e-07, + "loss": 0.5723, + "step": 17237 + }, + { + "epoch": 1.4689390711546655, + "grad_norm": 15.5, + "learning_rate": 4.1358674205142763e-07, + "loss": 0.511, + "step": 17238 + }, + { + "epoch": 1.4690242863229654, + "grad_norm": 19.0, + "learning_rate": 4.134618748593103e-07, + "loss": 0.4162, + "step": 17239 + }, + { + "epoch": 1.4691095014912654, + "grad_norm": 14.5, + "learning_rate": 4.13337022784278e-07, + "loss": 0.4359, + "step": 17240 + }, + { + "epoch": 1.4691947166595654, + "grad_norm": 12.875, + "learning_rate": 4.132121858285872e-07, + "loss": 0.6238, + "step": 17241 + }, + { + "epoch": 1.4692799318278653, + "grad_norm": 14.375, + "learning_rate": 4.1308736399449394e-07, + "loss": 0.7326, + "step": 17242 + }, + { + "epoch": 1.4693651469961653, + "grad_norm": 17.625, + "learning_rate": 4.129625572842531e-07, + "loss": 0.5964, + "step": 17243 + }, + { + "epoch": 1.4694503621644652, + "grad_norm": 14.375, + "learning_rate": 4.128377657001208e-07, + "loss": 0.477, + "step": 17244 + }, + { + "epoch": 1.4695355773327652, + "grad_norm": 15.25, + "learning_rate": 4.127129892443521e-07, + "loss": 0.6399, + "step": 17245 + }, + { + "epoch": 1.4696207925010651, + "grad_norm": 11.8125, + "learning_rate": 4.1258822791920173e-07, + "loss": 0.3021, + "step": 17246 + }, + { + "epoch": 1.469706007669365, + "grad_norm": 14.75, + "learning_rate": 4.124634817269238e-07, + "loss": 0.4937, + "step": 17247 + }, + { + "epoch": 1.469791222837665, + "grad_norm": 14.1875, + "learning_rate": 4.123387506697729e-07, + "loss": 0.5869, + "step": 17248 + }, + { + "epoch": 1.469876438005965, + "grad_norm": 19.0, + "learning_rate": 4.122140347500031e-07, + "loss": 0.8371, + "step": 17249 + }, + { + "epoch": 1.469961653174265, + "grad_norm": 12.6875, + "learning_rate": 4.1208933396986865e-07, + "loss": 0.4551, + "step": 17250 + }, + { + "epoch": 1.470046868342565, + "grad_norm": 13.875, + "learning_rate": 4.11964648331622e-07, + "loss": 0.4806, + "step": 17251 + }, + { + "epoch": 1.4701320835108649, + "grad_norm": 13.375, + "learning_rate": 4.1183997783751694e-07, + "loss": 0.4856, + "step": 17252 + }, + { + "epoch": 1.4702172986791648, + "grad_norm": 15.3125, + "learning_rate": 4.117153224898063e-07, + "loss": 0.5222, + "step": 17253 + }, + { + "epoch": 1.4703025138474648, + "grad_norm": 17.75, + "learning_rate": 4.1159068229074314e-07, + "loss": 0.6354, + "step": 17254 + }, + { + "epoch": 1.4703877290157648, + "grad_norm": 14.4375, + "learning_rate": 4.114660572425794e-07, + "loss": 0.4421, + "step": 17255 + }, + { + "epoch": 1.4704729441840647, + "grad_norm": 12.5, + "learning_rate": 4.113414473475669e-07, + "loss": 0.3916, + "step": 17256 + }, + { + "epoch": 1.4705581593523647, + "grad_norm": 17.125, + "learning_rate": 4.112168526079577e-07, + "loss": 0.6804, + "step": 17257 + }, + { + "epoch": 1.4706433745206646, + "grad_norm": 15.25, + "learning_rate": 4.1109227302600375e-07, + "loss": 0.451, + "step": 17258 + }, + { + "epoch": 1.4707285896889646, + "grad_norm": 19.875, + "learning_rate": 4.109677086039558e-07, + "loss": 0.5363, + "step": 17259 + }, + { + "epoch": 1.4708138048572645, + "grad_norm": 19.25, + "learning_rate": 4.1084315934406496e-07, + "loss": 0.4129, + "step": 17260 + }, + { + "epoch": 1.4708990200255645, + "grad_norm": 19.0, + "learning_rate": 4.1071862524858215e-07, + "loss": 0.7315, + "step": 17261 + }, + { + "epoch": 1.4709842351938645, + "grad_norm": 15.9375, + "learning_rate": 4.105941063197581e-07, + "loss": 0.7197, + "step": 17262 + }, + { + "epoch": 1.4710694503621644, + "grad_norm": 13.5625, + "learning_rate": 4.104696025598426e-07, + "loss": 0.397, + "step": 17263 + }, + { + "epoch": 1.4711546655304644, + "grad_norm": 18.375, + "learning_rate": 4.1034511397108524e-07, + "loss": 0.7737, + "step": 17264 + }, + { + "epoch": 1.4712398806987643, + "grad_norm": 12.75, + "learning_rate": 4.10220640555736e-07, + "loss": 0.5017, + "step": 17265 + }, + { + "epoch": 1.4713250958670643, + "grad_norm": 12.8125, + "learning_rate": 4.1009618231604416e-07, + "loss": 0.4812, + "step": 17266 + }, + { + "epoch": 1.4714103110353642, + "grad_norm": 16.375, + "learning_rate": 4.0997173925425946e-07, + "loss": 0.6555, + "step": 17267 + }, + { + "epoch": 1.4714955262036642, + "grad_norm": 16.125, + "learning_rate": 4.0984731137262967e-07, + "loss": 0.4136, + "step": 17268 + }, + { + "epoch": 1.4715807413719642, + "grad_norm": 11.25, + "learning_rate": 4.0972289867340376e-07, + "loss": 0.2908, + "step": 17269 + }, + { + "epoch": 1.471665956540264, + "grad_norm": 24.5, + "learning_rate": 4.095985011588301e-07, + "loss": 0.523, + "step": 17270 + }, + { + "epoch": 1.471751171708564, + "grad_norm": 14.5, + "learning_rate": 4.09474118831157e-07, + "loss": 0.5015, + "step": 17271 + }, + { + "epoch": 1.471836386876864, + "grad_norm": 18.25, + "learning_rate": 4.0934975169263175e-07, + "loss": 0.7505, + "step": 17272 + }, + { + "epoch": 1.471921602045164, + "grad_norm": 13.5, + "learning_rate": 4.0922539974550144e-07, + "loss": 0.6751, + "step": 17273 + }, + { + "epoch": 1.472006817213464, + "grad_norm": 19.375, + "learning_rate": 4.091010629920136e-07, + "loss": 0.7152, + "step": 17274 + }, + { + "epoch": 1.472092032381764, + "grad_norm": 18.625, + "learning_rate": 4.089767414344156e-07, + "loss": 0.9618, + "step": 17275 + }, + { + "epoch": 1.4721772475500638, + "grad_norm": 21.0, + "learning_rate": 4.0885243507495315e-07, + "loss": 0.5924, + "step": 17276 + }, + { + "epoch": 1.4722624627183638, + "grad_norm": 18.625, + "learning_rate": 4.0872814391587303e-07, + "loss": 0.5115, + "step": 17277 + }, + { + "epoch": 1.4723476778866638, + "grad_norm": 14.5, + "learning_rate": 4.0860386795942135e-07, + "loss": 0.5079, + "step": 17278 + }, + { + "epoch": 1.4724328930549637, + "grad_norm": 19.875, + "learning_rate": 4.084796072078442e-07, + "loss": 0.6159, + "step": 17279 + }, + { + "epoch": 1.4725181082232637, + "grad_norm": 18.75, + "learning_rate": 4.083553616633869e-07, + "loss": 0.7739, + "step": 17280 + }, + { + "epoch": 1.4726033233915636, + "grad_norm": 14.375, + "learning_rate": 4.082311313282941e-07, + "loss": 0.4308, + "step": 17281 + }, + { + "epoch": 1.4726885385598636, + "grad_norm": 14.125, + "learning_rate": 4.081069162048111e-07, + "loss": 0.4191, + "step": 17282 + }, + { + "epoch": 1.4727737537281635, + "grad_norm": 13.5625, + "learning_rate": 4.0798271629518287e-07, + "loss": 0.4707, + "step": 17283 + }, + { + "epoch": 1.4728589688964635, + "grad_norm": 17.25, + "learning_rate": 4.0785853160165393e-07, + "loss": 0.746, + "step": 17284 + }, + { + "epoch": 1.4729441840647635, + "grad_norm": 16.5, + "learning_rate": 4.077343621264679e-07, + "loss": 0.6376, + "step": 17285 + }, + { + "epoch": 1.4730293992330634, + "grad_norm": 16.5, + "learning_rate": 4.0761020787186884e-07, + "loss": 0.882, + "step": 17286 + }, + { + "epoch": 1.4731146144013634, + "grad_norm": 14.375, + "learning_rate": 4.074860688401008e-07, + "loss": 0.7019, + "step": 17287 + }, + { + "epoch": 1.4731998295696633, + "grad_norm": 17.25, + "learning_rate": 4.073619450334068e-07, + "loss": 0.7243, + "step": 17288 + }, + { + "epoch": 1.4732850447379633, + "grad_norm": 24.5, + "learning_rate": 4.0723783645402927e-07, + "loss": 0.8805, + "step": 17289 + }, + { + "epoch": 1.4733702599062632, + "grad_norm": 12.875, + "learning_rate": 4.071137431042116e-07, + "loss": 0.4146, + "step": 17290 + }, + { + "epoch": 1.4734554750745632, + "grad_norm": 20.625, + "learning_rate": 4.069896649861961e-07, + "loss": 0.7767, + "step": 17291 + }, + { + "epoch": 1.4735406902428632, + "grad_norm": 19.625, + "learning_rate": 4.0686560210222547e-07, + "loss": 0.7326, + "step": 17292 + }, + { + "epoch": 1.4736259054111631, + "grad_norm": 25.875, + "learning_rate": 4.0674155445454083e-07, + "loss": 0.8807, + "step": 17293 + }, + { + "epoch": 1.473711120579463, + "grad_norm": 14.8125, + "learning_rate": 4.0661752204538427e-07, + "loss": 0.3739, + "step": 17294 + }, + { + "epoch": 1.473796335747763, + "grad_norm": 12.375, + "learning_rate": 4.0649350487699757e-07, + "loss": 0.3865, + "step": 17295 + }, + { + "epoch": 1.473881550916063, + "grad_norm": 16.625, + "learning_rate": 4.0636950295162113e-07, + "loss": 0.5387, + "step": 17296 + }, + { + "epoch": 1.473966766084363, + "grad_norm": 12.9375, + "learning_rate": 4.062455162714965e-07, + "loss": 0.3619, + "step": 17297 + }, + { + "epoch": 1.474051981252663, + "grad_norm": 13.3125, + "learning_rate": 4.0612154483886334e-07, + "loss": 0.3844, + "step": 17298 + }, + { + "epoch": 1.4741371964209629, + "grad_norm": 17.375, + "learning_rate": 4.0599758865596264e-07, + "loss": 0.7273, + "step": 17299 + }, + { + "epoch": 1.4742224115892628, + "grad_norm": 12.375, + "learning_rate": 4.05873647725034e-07, + "loss": 0.5801, + "step": 17300 + }, + { + "epoch": 1.4743076267575628, + "grad_norm": 11.875, + "learning_rate": 4.057497220483179e-07, + "loss": 0.4331, + "step": 17301 + }, + { + "epoch": 1.4743928419258627, + "grad_norm": 17.75, + "learning_rate": 4.0562581162805287e-07, + "loss": 0.5393, + "step": 17302 + }, + { + "epoch": 1.4744780570941627, + "grad_norm": 14.25, + "learning_rate": 4.055019164664789e-07, + "loss": 0.7151, + "step": 17303 + }, + { + "epoch": 1.4745632722624626, + "grad_norm": 20.125, + "learning_rate": 4.053780365658341e-07, + "loss": 0.7055, + "step": 17304 + }, + { + "epoch": 1.4746484874307626, + "grad_norm": 21.625, + "learning_rate": 4.05254171928358e-07, + "loss": 1.0267, + "step": 17305 + }, + { + "epoch": 1.4747337025990626, + "grad_norm": 12.9375, + "learning_rate": 4.05130322556288e-07, + "loss": 0.4165, + "step": 17306 + }, + { + "epoch": 1.4748189177673625, + "grad_norm": 16.625, + "learning_rate": 4.0500648845186255e-07, + "loss": 0.7469, + "step": 17307 + }, + { + "epoch": 1.4749041329356625, + "grad_norm": 11.25, + "learning_rate": 4.048826696173197e-07, + "loss": 0.3703, + "step": 17308 + }, + { + "epoch": 1.4749893481039624, + "grad_norm": 13.5, + "learning_rate": 4.047588660548972e-07, + "loss": 0.4605, + "step": 17309 + }, + { + "epoch": 1.4750745632722624, + "grad_norm": 13.375, + "learning_rate": 4.046350777668316e-07, + "loss": 0.5429, + "step": 17310 + }, + { + "epoch": 1.4751597784405623, + "grad_norm": 23.125, + "learning_rate": 4.045113047553606e-07, + "loss": 0.8238, + "step": 17311 + }, + { + "epoch": 1.4752449936088623, + "grad_norm": 11.75, + "learning_rate": 4.0438754702272e-07, + "loss": 0.2821, + "step": 17312 + }, + { + "epoch": 1.4753302087771623, + "grad_norm": 12.5, + "learning_rate": 4.0426380457114693e-07, + "loss": 0.5182, + "step": 17313 + }, + { + "epoch": 1.4754154239454622, + "grad_norm": 16.625, + "learning_rate": 4.0414007740287765e-07, + "loss": 0.6123, + "step": 17314 + }, + { + "epoch": 1.4755006391137622, + "grad_norm": 13.4375, + "learning_rate": 4.0401636552014746e-07, + "loss": 0.6747, + "step": 17315 + }, + { + "epoch": 1.4755858542820621, + "grad_norm": 13.625, + "learning_rate": 4.0389266892519205e-07, + "loss": 0.345, + "step": 17316 + }, + { + "epoch": 1.475671069450362, + "grad_norm": 27.0, + "learning_rate": 4.037689876202472e-07, + "loss": 0.3827, + "step": 17317 + }, + { + "epoch": 1.475756284618662, + "grad_norm": 19.25, + "learning_rate": 4.0364532160754795e-07, + "loss": 0.178, + "step": 17318 + }, + { + "epoch": 1.475841499786962, + "grad_norm": 12.5, + "learning_rate": 4.035216708893289e-07, + "loss": 0.4519, + "step": 17319 + }, + { + "epoch": 1.475926714955262, + "grad_norm": 12.0625, + "learning_rate": 4.0339803546782396e-07, + "loss": 0.3261, + "step": 17320 + }, + { + "epoch": 1.476011930123562, + "grad_norm": 21.75, + "learning_rate": 4.032744153452678e-07, + "loss": 0.6774, + "step": 17321 + }, + { + "epoch": 1.4760971452918619, + "grad_norm": 13.8125, + "learning_rate": 4.031508105238949e-07, + "loss": 0.6829, + "step": 17322 + }, + { + "epoch": 1.4761823604601618, + "grad_norm": 19.75, + "learning_rate": 4.0302722100593797e-07, + "loss": 0.5042, + "step": 17323 + }, + { + "epoch": 1.4762675756284618, + "grad_norm": 20.25, + "learning_rate": 4.029036467936308e-07, + "loss": 0.6105, + "step": 17324 + }, + { + "epoch": 1.4763527907967617, + "grad_norm": 11.875, + "learning_rate": 4.027800878892067e-07, + "loss": 0.257, + "step": 17325 + }, + { + "epoch": 1.4764380059650617, + "grad_norm": 17.5, + "learning_rate": 4.0265654429489864e-07, + "loss": 0.5321, + "step": 17326 + }, + { + "epoch": 1.4765232211333617, + "grad_norm": 13.0625, + "learning_rate": 4.0253301601293856e-07, + "loss": 0.4372, + "step": 17327 + }, + { + "epoch": 1.4766084363016616, + "grad_norm": 13.375, + "learning_rate": 4.0240950304555953e-07, + "loss": 0.4607, + "step": 17328 + }, + { + "epoch": 1.4766936514699616, + "grad_norm": 13.5, + "learning_rate": 4.022860053949926e-07, + "loss": 0.5283, + "step": 17329 + }, + { + "epoch": 1.4767788666382615, + "grad_norm": 17.5, + "learning_rate": 4.0216252306347e-07, + "loss": 0.6731, + "step": 17330 + }, + { + "epoch": 1.4768640818065615, + "grad_norm": 21.5, + "learning_rate": 4.020390560532236e-07, + "loss": 0.5663, + "step": 17331 + }, + { + "epoch": 1.4769492969748614, + "grad_norm": 11.5625, + "learning_rate": 4.0191560436648376e-07, + "loss": 0.3791, + "step": 17332 + }, + { + "epoch": 1.4770345121431614, + "grad_norm": 13.1875, + "learning_rate": 4.0179216800548176e-07, + "loss": 0.4408, + "step": 17333 + }, + { + "epoch": 1.4771197273114614, + "grad_norm": 16.25, + "learning_rate": 4.016687469724481e-07, + "loss": 0.7263, + "step": 17334 + }, + { + "epoch": 1.4772049424797613, + "grad_norm": 15.8125, + "learning_rate": 4.015453412696138e-07, + "loss": 0.6094, + "step": 17335 + }, + { + "epoch": 1.4772901576480613, + "grad_norm": 12.125, + "learning_rate": 4.014219508992083e-07, + "loss": 0.3427, + "step": 17336 + }, + { + "epoch": 1.4773753728163612, + "grad_norm": 16.0, + "learning_rate": 4.012985758634612e-07, + "loss": 0.7118, + "step": 17337 + }, + { + "epoch": 1.4774605879846612, + "grad_norm": 13.875, + "learning_rate": 4.0117521616460207e-07, + "loss": 0.4285, + "step": 17338 + }, + { + "epoch": 1.4775458031529611, + "grad_norm": 16.375, + "learning_rate": 4.010518718048609e-07, + "loss": 0.7916, + "step": 17339 + }, + { + "epoch": 1.477631018321261, + "grad_norm": 14.75, + "learning_rate": 4.009285427864655e-07, + "loss": 0.5064, + "step": 17340 + }, + { + "epoch": 1.477716233489561, + "grad_norm": 13.875, + "learning_rate": 4.008052291116453e-07, + "loss": 0.6246, + "step": 17341 + }, + { + "epoch": 1.477801448657861, + "grad_norm": 10.5, + "learning_rate": 4.0068193078262845e-07, + "loss": 0.3465, + "step": 17342 + }, + { + "epoch": 1.477886663826161, + "grad_norm": 14.75, + "learning_rate": 4.005586478016436e-07, + "loss": 0.3951, + "step": 17343 + }, + { + "epoch": 1.477971878994461, + "grad_norm": 13.4375, + "learning_rate": 4.004353801709182e-07, + "loss": 0.2934, + "step": 17344 + }, + { + "epoch": 1.4780570941627609, + "grad_norm": 20.0, + "learning_rate": 4.0031212789267945e-07, + "loss": 0.6358, + "step": 17345 + }, + { + "epoch": 1.4781423093310608, + "grad_norm": 28.0, + "learning_rate": 4.001888909691548e-07, + "loss": 1.1585, + "step": 17346 + }, + { + "epoch": 1.4782275244993608, + "grad_norm": 17.25, + "learning_rate": 4.000656694025716e-07, + "loss": 0.6532, + "step": 17347 + }, + { + "epoch": 1.4783127396676607, + "grad_norm": 23.125, + "learning_rate": 3.9994246319515677e-07, + "loss": 1.0379, + "step": 17348 + }, + { + "epoch": 1.4783979548359607, + "grad_norm": 12.0, + "learning_rate": 3.998192723491359e-07, + "loss": 0.457, + "step": 17349 + }, + { + "epoch": 1.4784831700042607, + "grad_norm": 15.875, + "learning_rate": 3.9969609686673587e-07, + "loss": 0.4298, + "step": 17350 + }, + { + "epoch": 1.4785683851725606, + "grad_norm": 15.0, + "learning_rate": 3.9957293675018244e-07, + "loss": 0.3518, + "step": 17351 + }, + { + "epoch": 1.4786536003408606, + "grad_norm": 13.375, + "learning_rate": 3.9944979200170193e-07, + "loss": 0.3844, + "step": 17352 + }, + { + "epoch": 1.4787388155091605, + "grad_norm": 13.8125, + "learning_rate": 3.9932666262351814e-07, + "loss": 0.3707, + "step": 17353 + }, + { + "epoch": 1.4788240306774605, + "grad_norm": 15.6875, + "learning_rate": 3.9920354861785705e-07, + "loss": 0.6726, + "step": 17354 + }, + { + "epoch": 1.4789092458457604, + "grad_norm": 18.75, + "learning_rate": 3.990804499869433e-07, + "loss": 0.7034, + "step": 17355 + }, + { + "epoch": 1.4789944610140604, + "grad_norm": 16.5, + "learning_rate": 3.98957366733002e-07, + "loss": 0.4547, + "step": 17356 + }, + { + "epoch": 1.4790796761823604, + "grad_norm": 17.5, + "learning_rate": 3.9883429885825635e-07, + "loss": 0.7266, + "step": 17357 + }, + { + "epoch": 1.4791648913506603, + "grad_norm": 15.875, + "learning_rate": 3.9871124636493093e-07, + "loss": 0.4778, + "step": 17358 + }, + { + "epoch": 1.4792501065189603, + "grad_norm": 19.25, + "learning_rate": 3.985882092552493e-07, + "loss": 0.6309, + "step": 17359 + }, + { + "epoch": 1.4793353216872602, + "grad_norm": 17.125, + "learning_rate": 3.984651875314352e-07, + "loss": 0.7441, + "step": 17360 + }, + { + "epoch": 1.4794205368555602, + "grad_norm": 13.5625, + "learning_rate": 3.9834218119571154e-07, + "loss": 0.5085, + "step": 17361 + }, + { + "epoch": 1.4795057520238601, + "grad_norm": 14.625, + "learning_rate": 3.982191902503006e-07, + "loss": 0.4947, + "step": 17362 + }, + { + "epoch": 1.47959096719216, + "grad_norm": 17.625, + "learning_rate": 3.980962146974254e-07, + "loss": 0.6811, + "step": 17363 + }, + { + "epoch": 1.47967618236046, + "grad_norm": 20.5, + "learning_rate": 3.9797325453930837e-07, + "loss": 0.732, + "step": 17364 + }, + { + "epoch": 1.47976139752876, + "grad_norm": 13.5625, + "learning_rate": 3.9785030977817175e-07, + "loss": 0.3419, + "step": 17365 + }, + { + "epoch": 1.47984661269706, + "grad_norm": 14.375, + "learning_rate": 3.977273804162365e-07, + "loss": 0.575, + "step": 17366 + }, + { + "epoch": 1.47993182786536, + "grad_norm": 14.875, + "learning_rate": 3.9760446645572465e-07, + "loss": 0.7468, + "step": 17367 + }, + { + "epoch": 1.4800170430336599, + "grad_norm": 13.5, + "learning_rate": 3.9748156789885766e-07, + "loss": 0.3805, + "step": 17368 + }, + { + "epoch": 1.4801022582019598, + "grad_norm": 14.375, + "learning_rate": 3.973586847478558e-07, + "loss": 0.469, + "step": 17369 + }, + { + "epoch": 1.4801874733702598, + "grad_norm": 16.75, + "learning_rate": 3.972358170049397e-07, + "loss": 0.6803, + "step": 17370 + }, + { + "epoch": 1.4802726885385598, + "grad_norm": 18.875, + "learning_rate": 3.971129646723297e-07, + "loss": 0.9482, + "step": 17371 + }, + { + "epoch": 1.4803579037068597, + "grad_norm": 17.5, + "learning_rate": 3.969901277522463e-07, + "loss": 0.8263, + "step": 17372 + }, + { + "epoch": 1.4804431188751597, + "grad_norm": 51.75, + "learning_rate": 3.968673062469093e-07, + "loss": 1.3426, + "step": 17373 + }, + { + "epoch": 1.4805283340434596, + "grad_norm": 15.4375, + "learning_rate": 3.967445001585377e-07, + "loss": 0.5326, + "step": 17374 + }, + { + "epoch": 1.4806135492117596, + "grad_norm": 16.125, + "learning_rate": 3.966217094893508e-07, + "loss": 0.3408, + "step": 17375 + }, + { + "epoch": 1.4806987643800595, + "grad_norm": 15.3125, + "learning_rate": 3.9649893424156827e-07, + "loss": 0.5664, + "step": 17376 + }, + { + "epoch": 1.4807839795483595, + "grad_norm": 17.625, + "learning_rate": 3.963761744174079e-07, + "loss": 0.4569, + "step": 17377 + }, + { + "epoch": 1.4808691947166595, + "grad_norm": 16.25, + "learning_rate": 3.962534300190887e-07, + "loss": 0.8735, + "step": 17378 + }, + { + "epoch": 1.4809544098849594, + "grad_norm": 14.0625, + "learning_rate": 3.9613070104882826e-07, + "loss": 0.5867, + "step": 17379 + }, + { + "epoch": 1.4810396250532594, + "grad_norm": 17.5, + "learning_rate": 3.9600798750884463e-07, + "loss": 0.6037, + "step": 17380 + }, + { + "epoch": 1.4811248402215593, + "grad_norm": 19.125, + "learning_rate": 3.9588528940135554e-07, + "loss": 0.6245, + "step": 17381 + }, + { + "epoch": 1.4812100553898593, + "grad_norm": 12.75, + "learning_rate": 3.9576260672857843e-07, + "loss": 0.2978, + "step": 17382 + }, + { + "epoch": 1.4812952705581592, + "grad_norm": 19.75, + "learning_rate": 3.956399394927296e-07, + "loss": 0.7487, + "step": 17383 + }, + { + "epoch": 1.4813804857264592, + "grad_norm": 17.0, + "learning_rate": 3.9551728769602655e-07, + "loss": 0.75, + "step": 17384 + }, + { + "epoch": 1.4814657008947592, + "grad_norm": 13.25, + "learning_rate": 3.9539465134068505e-07, + "loss": 0.5412, + "step": 17385 + }, + { + "epoch": 1.481550916063059, + "grad_norm": 16.375, + "learning_rate": 3.9527203042892205e-07, + "loss": 0.9773, + "step": 17386 + }, + { + "epoch": 1.481636131231359, + "grad_norm": 13.375, + "learning_rate": 3.9514942496295244e-07, + "loss": 0.6105, + "step": 17387 + }, + { + "epoch": 1.481721346399659, + "grad_norm": 12.3125, + "learning_rate": 3.950268349449923e-07, + "loss": 0.3127, + "step": 17388 + }, + { + "epoch": 1.481806561567959, + "grad_norm": 18.875, + "learning_rate": 3.949042603772571e-07, + "loss": 0.6285, + "step": 17389 + }, + { + "epoch": 1.481891776736259, + "grad_norm": 15.9375, + "learning_rate": 3.9478170126196217e-07, + "loss": 0.6967, + "step": 17390 + }, + { + "epoch": 1.481976991904559, + "grad_norm": 15.4375, + "learning_rate": 3.946591576013215e-07, + "loss": 0.7812, + "step": 17391 + }, + { + "epoch": 1.4820622070728589, + "grad_norm": 16.75, + "learning_rate": 3.9453662939754987e-07, + "loss": 0.4848, + "step": 17392 + }, + { + "epoch": 1.4821474222411588, + "grad_norm": 29.75, + "learning_rate": 3.94414116652862e-07, + "loss": 0.7147, + "step": 17393 + }, + { + "epoch": 1.4822326374094588, + "grad_norm": 15.875, + "learning_rate": 3.9429161936947113e-07, + "loss": 0.2916, + "step": 17394 + }, + { + "epoch": 1.4823178525777587, + "grad_norm": 12.9375, + "learning_rate": 3.941691375495914e-07, + "loss": 0.4166, + "step": 17395 + }, + { + "epoch": 1.4824030677460587, + "grad_norm": 18.75, + "learning_rate": 3.9404667119543565e-07, + "loss": 0.6084, + "step": 17396 + }, + { + "epoch": 1.4824882829143586, + "grad_norm": 12.5, + "learning_rate": 3.9392422030921724e-07, + "loss": 0.5985, + "step": 17397 + }, + { + "epoch": 1.4825734980826586, + "grad_norm": 13.9375, + "learning_rate": 3.938017848931494e-07, + "loss": 0.5318, + "step": 17398 + }, + { + "epoch": 1.4826587132509585, + "grad_norm": 15.625, + "learning_rate": 3.93679364949444e-07, + "loss": 0.4099, + "step": 17399 + }, + { + "epoch": 1.4827439284192585, + "grad_norm": 13.25, + "learning_rate": 3.935569604803134e-07, + "loss": 0.3588, + "step": 17400 + }, + { + "epoch": 1.4828291435875585, + "grad_norm": 17.75, + "learning_rate": 3.9343457148797034e-07, + "loss": 0.4108, + "step": 17401 + }, + { + "epoch": 1.4829143587558584, + "grad_norm": 16.125, + "learning_rate": 3.9331219797462543e-07, + "loss": 0.4985, + "step": 17402 + }, + { + "epoch": 1.4829995739241584, + "grad_norm": 20.875, + "learning_rate": 3.931898399424909e-07, + "loss": 0.7963, + "step": 17403 + }, + { + "epoch": 1.4830847890924583, + "grad_norm": 16.125, + "learning_rate": 3.930674973937773e-07, + "loss": 0.6314, + "step": 17404 + }, + { + "epoch": 1.4831700042607583, + "grad_norm": 16.875, + "learning_rate": 3.929451703306957e-07, + "loss": 0.9827, + "step": 17405 + }, + { + "epoch": 1.4832552194290582, + "grad_norm": 15.0625, + "learning_rate": 3.9282285875545655e-07, + "loss": 0.6056, + "step": 17406 + }, + { + "epoch": 1.4833404345973582, + "grad_norm": 13.75, + "learning_rate": 3.927005626702708e-07, + "loss": 0.4519, + "step": 17407 + }, + { + "epoch": 1.4834256497656582, + "grad_norm": 31.0, + "learning_rate": 3.9257828207734756e-07, + "loss": 0.8683, + "step": 17408 + }, + { + "epoch": 1.4835108649339581, + "grad_norm": 17.875, + "learning_rate": 3.924560169788974e-07, + "loss": 0.5357, + "step": 17409 + }, + { + "epoch": 1.483596080102258, + "grad_norm": 17.0, + "learning_rate": 3.9233376737712895e-07, + "loss": 0.7233, + "step": 17410 + }, + { + "epoch": 1.483681295270558, + "grad_norm": 14.125, + "learning_rate": 3.922115332742517e-07, + "loss": 0.4108, + "step": 17411 + }, + { + "epoch": 1.483766510438858, + "grad_norm": 16.5, + "learning_rate": 3.920893146724751e-07, + "loss": 0.5881, + "step": 17412 + }, + { + "epoch": 1.483851725607158, + "grad_norm": 17.875, + "learning_rate": 3.9196711157400674e-07, + "loss": 0.6182, + "step": 17413 + }, + { + "epoch": 1.483936940775458, + "grad_norm": 14.4375, + "learning_rate": 3.9184492398105565e-07, + "loss": 0.4522, + "step": 17414 + }, + { + "epoch": 1.4840221559437579, + "grad_norm": 13.4375, + "learning_rate": 3.9172275189583e-07, + "loss": 0.6417, + "step": 17415 + }, + { + "epoch": 1.4841073711120578, + "grad_norm": 16.0, + "learning_rate": 3.9160059532053693e-07, + "loss": 0.4736, + "step": 17416 + }, + { + "epoch": 1.4841925862803578, + "grad_norm": 18.125, + "learning_rate": 3.9147845425738474e-07, + "loss": 0.6005, + "step": 17417 + }, + { + "epoch": 1.4842778014486577, + "grad_norm": 19.125, + "learning_rate": 3.9135632870857975e-07, + "loss": 0.8761, + "step": 17418 + }, + { + "epoch": 1.4843630166169577, + "grad_norm": 16.75, + "learning_rate": 3.9123421867632933e-07, + "loss": 0.7302, + "step": 17419 + }, + { + "epoch": 1.4844482317852576, + "grad_norm": 13.375, + "learning_rate": 3.911121241628406e-07, + "loss": 0.3297, + "step": 17420 + }, + { + "epoch": 1.4845334469535576, + "grad_norm": 10.75, + "learning_rate": 3.90990045170319e-07, + "loss": 0.2252, + "step": 17421 + }, + { + "epoch": 1.4846186621218576, + "grad_norm": 15.9375, + "learning_rate": 3.908679817009711e-07, + "loss": 0.5195, + "step": 17422 + }, + { + "epoch": 1.4847038772901575, + "grad_norm": 16.5, + "learning_rate": 3.907459337570027e-07, + "loss": 0.5478, + "step": 17423 + }, + { + "epoch": 1.4847890924584575, + "grad_norm": 14.1875, + "learning_rate": 3.9062390134061974e-07, + "loss": 0.4883, + "step": 17424 + }, + { + "epoch": 1.4848743076267574, + "grad_norm": 22.0, + "learning_rate": 3.9050188445402704e-07, + "loss": 0.4488, + "step": 17425 + }, + { + "epoch": 1.4849595227950574, + "grad_norm": 12.0625, + "learning_rate": 3.9037988309942915e-07, + "loss": 0.8827, + "step": 17426 + }, + { + "epoch": 1.4850447379633573, + "grad_norm": 10.875, + "learning_rate": 3.9025789727903123e-07, + "loss": 0.2778, + "step": 17427 + }, + { + "epoch": 1.4851299531316573, + "grad_norm": 13.875, + "learning_rate": 3.9013592699503805e-07, + "loss": 0.4557, + "step": 17428 + }, + { + "epoch": 1.4852151682999573, + "grad_norm": 12.5625, + "learning_rate": 3.9001397224965295e-07, + "loss": 0.4385, + "step": 17429 + }, + { + "epoch": 1.4853003834682572, + "grad_norm": 16.375, + "learning_rate": 3.8989203304508007e-07, + "loss": 0.8618, + "step": 17430 + }, + { + "epoch": 1.4853855986365572, + "grad_norm": 15.8125, + "learning_rate": 3.8977010938352307e-07, + "loss": 0.5813, + "step": 17431 + }, + { + "epoch": 1.4854708138048571, + "grad_norm": 14.5, + "learning_rate": 3.896482012671857e-07, + "loss": 0.6423, + "step": 17432 + }, + { + "epoch": 1.485556028973157, + "grad_norm": 15.125, + "learning_rate": 3.8952630869827037e-07, + "loss": 0.5752, + "step": 17433 + }, + { + "epoch": 1.4856412441414573, + "grad_norm": 14.6875, + "learning_rate": 3.8940443167897956e-07, + "loss": 0.6695, + "step": 17434 + }, + { + "epoch": 1.4857264593097572, + "grad_norm": 15.4375, + "learning_rate": 3.8928257021151605e-07, + "loss": 0.531, + "step": 17435 + }, + { + "epoch": 1.4858116744780572, + "grad_norm": 15.125, + "learning_rate": 3.89160724298082e-07, + "loss": 0.5529, + "step": 17436 + }, + { + "epoch": 1.4858968896463571, + "grad_norm": 14.5625, + "learning_rate": 3.8903889394087986e-07, + "loss": 0.3, + "step": 17437 + }, + { + "epoch": 1.485982104814657, + "grad_norm": 15.75, + "learning_rate": 3.889170791421101e-07, + "loss": 0.4806, + "step": 17438 + }, + { + "epoch": 1.486067319982957, + "grad_norm": 14.6875, + "learning_rate": 3.8879527990397475e-07, + "loss": 0.6601, + "step": 17439 + }, + { + "epoch": 1.486152535151257, + "grad_norm": 13.3125, + "learning_rate": 3.886734962286745e-07, + "loss": 0.6116, + "step": 17440 + }, + { + "epoch": 1.486237750319557, + "grad_norm": 15.0625, + "learning_rate": 3.885517281184109e-07, + "loss": 0.634, + "step": 17441 + }, + { + "epoch": 1.486322965487857, + "grad_norm": 18.25, + "learning_rate": 3.884299755753837e-07, + "loss": 0.8881, + "step": 17442 + }, + { + "epoch": 1.4864081806561569, + "grad_norm": 18.5, + "learning_rate": 3.883082386017928e-07, + "loss": 0.5247, + "step": 17443 + }, + { + "epoch": 1.4864933958244568, + "grad_norm": 14.625, + "learning_rate": 3.8818651719983863e-07, + "loss": 0.5927, + "step": 17444 + }, + { + "epoch": 1.4865786109927568, + "grad_norm": 15.0625, + "learning_rate": 3.8806481137172113e-07, + "loss": 0.6384, + "step": 17445 + }, + { + "epoch": 1.4866638261610567, + "grad_norm": 19.5, + "learning_rate": 3.879431211196387e-07, + "loss": 0.8251, + "step": 17446 + }, + { + "epoch": 1.4867490413293567, + "grad_norm": 14.0, + "learning_rate": 3.8782144644579105e-07, + "loss": 0.5723, + "step": 17447 + }, + { + "epoch": 1.4868342564976567, + "grad_norm": 14.75, + "learning_rate": 3.8769978735237675e-07, + "loss": 0.6051, + "step": 17448 + }, + { + "epoch": 1.4869194716659566, + "grad_norm": 21.25, + "learning_rate": 3.8757814384159486e-07, + "loss": 0.7784, + "step": 17449 + }, + { + "epoch": 1.4870046868342566, + "grad_norm": 12.9375, + "learning_rate": 3.874565159156432e-07, + "loss": 0.8535, + "step": 17450 + }, + { + "epoch": 1.4870899020025565, + "grad_norm": 12.0, + "learning_rate": 3.873349035767192e-07, + "loss": 0.326, + "step": 17451 + }, + { + "epoch": 1.4871751171708565, + "grad_norm": 14.875, + "learning_rate": 3.872133068270209e-07, + "loss": 0.298, + "step": 17452 + }, + { + "epoch": 1.4872603323391564, + "grad_norm": 14.9375, + "learning_rate": 3.8709172566874576e-07, + "loss": 0.7943, + "step": 17453 + }, + { + "epoch": 1.4873455475074564, + "grad_norm": 15.75, + "learning_rate": 3.869701601040914e-07, + "loss": 0.6357, + "step": 17454 + }, + { + "epoch": 1.4874307626757564, + "grad_norm": 16.5, + "learning_rate": 3.868486101352537e-07, + "loss": 0.5098, + "step": 17455 + }, + { + "epoch": 1.4875159778440563, + "grad_norm": 15.1875, + "learning_rate": 3.867270757644294e-07, + "loss": 0.5143, + "step": 17456 + }, + { + "epoch": 1.4876011930123563, + "grad_norm": 17.25, + "learning_rate": 3.866055569938154e-07, + "loss": 0.8338, + "step": 17457 + }, + { + "epoch": 1.4876864081806562, + "grad_norm": 17.875, + "learning_rate": 3.864840538256072e-07, + "loss": 0.7286, + "step": 17458 + }, + { + "epoch": 1.4877716233489562, + "grad_norm": 20.375, + "learning_rate": 3.8636256626200007e-07, + "loss": 0.7703, + "step": 17459 + }, + { + "epoch": 1.4878568385172561, + "grad_norm": 12.8125, + "learning_rate": 3.862410943051899e-07, + "loss": 0.2893, + "step": 17460 + }, + { + "epoch": 1.487942053685556, + "grad_norm": 22.5, + "learning_rate": 3.861196379573716e-07, + "loss": 0.7677, + "step": 17461 + }, + { + "epoch": 1.488027268853856, + "grad_norm": 18.125, + "learning_rate": 3.8599819722074066e-07, + "loss": 0.591, + "step": 17462 + }, + { + "epoch": 1.488112484022156, + "grad_norm": 17.5, + "learning_rate": 3.858767720974907e-07, + "loss": 0.7926, + "step": 17463 + }, + { + "epoch": 1.488197699190456, + "grad_norm": 16.75, + "learning_rate": 3.8575536258981624e-07, + "loss": 0.8462, + "step": 17464 + }, + { + "epoch": 1.488282914358756, + "grad_norm": 14.4375, + "learning_rate": 3.856339686999117e-07, + "loss": 0.512, + "step": 17465 + }, + { + "epoch": 1.4883681295270559, + "grad_norm": 15.4375, + "learning_rate": 3.855125904299707e-07, + "loss": 0.6381, + "step": 17466 + }, + { + "epoch": 1.4884533446953558, + "grad_norm": 17.375, + "learning_rate": 3.853912277821867e-07, + "loss": 0.7897, + "step": 17467 + }, + { + "epoch": 1.4885385598636558, + "grad_norm": 20.0, + "learning_rate": 3.852698807587521e-07, + "loss": 0.9579, + "step": 17468 + }, + { + "epoch": 1.4886237750319558, + "grad_norm": 15.5625, + "learning_rate": 3.8514854936186037e-07, + "loss": 0.8539, + "step": 17469 + }, + { + "epoch": 1.4887089902002557, + "grad_norm": 14.625, + "learning_rate": 3.85027233593704e-07, + "loss": 0.7443, + "step": 17470 + }, + { + "epoch": 1.4887942053685557, + "grad_norm": 20.25, + "learning_rate": 3.849059334564759e-07, + "loss": 0.5436, + "step": 17471 + }, + { + "epoch": 1.4888794205368556, + "grad_norm": 35.5, + "learning_rate": 3.8478464895236685e-07, + "loss": 0.8486, + "step": 17472 + }, + { + "epoch": 1.4889646357051556, + "grad_norm": 14.1875, + "learning_rate": 3.846633800835693e-07, + "loss": 0.6118, + "step": 17473 + }, + { + "epoch": 1.4890498508734555, + "grad_norm": 13.125, + "learning_rate": 3.8454212685227503e-07, + "loss": 0.4509, + "step": 17474 + }, + { + "epoch": 1.4891350660417555, + "grad_norm": 15.5, + "learning_rate": 3.844208892606749e-07, + "loss": 0.3828, + "step": 17475 + }, + { + "epoch": 1.4892202812100555, + "grad_norm": 19.125, + "learning_rate": 3.842996673109593e-07, + "loss": 1.0877, + "step": 17476 + }, + { + "epoch": 1.4893054963783554, + "grad_norm": 14.375, + "learning_rate": 3.841784610053191e-07, + "loss": 0.403, + "step": 17477 + }, + { + "epoch": 1.4893907115466554, + "grad_norm": 16.75, + "learning_rate": 3.84057270345945e-07, + "loss": 0.6259, + "step": 17478 + }, + { + "epoch": 1.4894759267149553, + "grad_norm": 15.625, + "learning_rate": 3.83936095335027e-07, + "loss": 0.4996, + "step": 17479 + }, + { + "epoch": 1.4895611418832553, + "grad_norm": 11.625, + "learning_rate": 3.8381493597475424e-07, + "loss": 0.4583, + "step": 17480 + }, + { + "epoch": 1.4896463570515552, + "grad_norm": 13.25, + "learning_rate": 3.836937922673167e-07, + "loss": 0.5134, + "step": 17481 + }, + { + "epoch": 1.4897315722198552, + "grad_norm": 21.0, + "learning_rate": 3.8357266421490387e-07, + "loss": 0.6506, + "step": 17482 + }, + { + "epoch": 1.4898167873881552, + "grad_norm": 17.25, + "learning_rate": 3.8345155181970377e-07, + "loss": 0.6902, + "step": 17483 + }, + { + "epoch": 1.489902002556455, + "grad_norm": 21.625, + "learning_rate": 3.8333045508390607e-07, + "loss": 1.1121, + "step": 17484 + }, + { + "epoch": 1.489987217724755, + "grad_norm": 14.4375, + "learning_rate": 3.83209374009698e-07, + "loss": 0.5483, + "step": 17485 + }, + { + "epoch": 1.490072432893055, + "grad_norm": 15.125, + "learning_rate": 3.8308830859926816e-07, + "loss": 0.6629, + "step": 17486 + }, + { + "epoch": 1.490157648061355, + "grad_norm": 14.5625, + "learning_rate": 3.829672588548045e-07, + "loss": 0.5512, + "step": 17487 + }, + { + "epoch": 1.490242863229655, + "grad_norm": 17.875, + "learning_rate": 3.828462247784946e-07, + "loss": 0.6615, + "step": 17488 + }, + { + "epoch": 1.490328078397955, + "grad_norm": 14.6875, + "learning_rate": 3.8272520637252495e-07, + "loss": 0.3605, + "step": 17489 + }, + { + "epoch": 1.4904132935662548, + "grad_norm": 14.875, + "learning_rate": 3.8260420363908355e-07, + "loss": 0.693, + "step": 17490 + }, + { + "epoch": 1.4904985087345548, + "grad_norm": 22.625, + "learning_rate": 3.8248321658035573e-07, + "loss": 0.8082, + "step": 17491 + }, + { + "epoch": 1.4905837239028548, + "grad_norm": 31.75, + "learning_rate": 3.8236224519852915e-07, + "loss": 0.822, + "step": 17492 + }, + { + "epoch": 1.4906689390711547, + "grad_norm": 35.75, + "learning_rate": 3.8224128949578885e-07, + "loss": 0.765, + "step": 17493 + }, + { + "epoch": 1.4907541542394547, + "grad_norm": 12.5, + "learning_rate": 3.82120349474321e-07, + "loss": 0.6235, + "step": 17494 + }, + { + "epoch": 1.4908393694077546, + "grad_norm": 63.5, + "learning_rate": 3.8199942513631125e-07, + "loss": 0.544, + "step": 17495 + }, + { + "epoch": 1.4909245845760546, + "grad_norm": 9.875, + "learning_rate": 3.81878516483945e-07, + "loss": 0.4154, + "step": 17496 + }, + { + "epoch": 1.4910097997443545, + "grad_norm": 19.875, + "learning_rate": 3.8175762351940664e-07, + "loss": 0.3964, + "step": 17497 + }, + { + "epoch": 1.4910950149126545, + "grad_norm": 19.5, + "learning_rate": 3.8163674624488154e-07, + "loss": 0.7185, + "step": 17498 + }, + { + "epoch": 1.4911802300809545, + "grad_norm": 19.125, + "learning_rate": 3.8151588466255313e-07, + "loss": 0.8671, + "step": 17499 + }, + { + "epoch": 1.4912654452492544, + "grad_norm": 17.5, + "learning_rate": 3.81395038774606e-07, + "loss": 0.5918, + "step": 17500 + }, + { + "epoch": 1.4913506604175544, + "grad_norm": 12.5625, + "learning_rate": 3.8127420858322437e-07, + "loss": 0.4712, + "step": 17501 + }, + { + "epoch": 1.4914358755858543, + "grad_norm": 12.875, + "learning_rate": 3.811533940905911e-07, + "loss": 0.479, + "step": 17502 + }, + { + "epoch": 1.4915210907541543, + "grad_norm": 13.0625, + "learning_rate": 3.8103259529888956e-07, + "loss": 0.527, + "step": 17503 + }, + { + "epoch": 1.4916063059224542, + "grad_norm": 16.875, + "learning_rate": 3.809118122103028e-07, + "loss": 0.6575, + "step": 17504 + }, + { + "epoch": 1.4916915210907542, + "grad_norm": 11.4375, + "learning_rate": 3.80791044827014e-07, + "loss": 0.309, + "step": 17505 + }, + { + "epoch": 1.4917767362590542, + "grad_norm": 17.125, + "learning_rate": 3.8067029315120516e-07, + "loss": 0.2089, + "step": 17506 + }, + { + "epoch": 1.4918619514273541, + "grad_norm": 15.4375, + "learning_rate": 3.8054955718505773e-07, + "loss": 0.5543, + "step": 17507 + }, + { + "epoch": 1.491947166595654, + "grad_norm": 20.125, + "learning_rate": 3.804288369307542e-07, + "loss": 0.9201, + "step": 17508 + }, + { + "epoch": 1.492032381763954, + "grad_norm": 11.0, + "learning_rate": 3.803081323904764e-07, + "loss": 0.1756, + "step": 17509 + }, + { + "epoch": 1.492117596932254, + "grad_norm": 9.125, + "learning_rate": 3.801874435664048e-07, + "loss": 0.2427, + "step": 17510 + }, + { + "epoch": 1.492202812100554, + "grad_norm": 14.125, + "learning_rate": 3.8006677046072083e-07, + "loss": 0.7541, + "step": 17511 + }, + { + "epoch": 1.492288027268854, + "grad_norm": 15.9375, + "learning_rate": 3.7994611307560494e-07, + "loss": 0.719, + "step": 17512 + }, + { + "epoch": 1.4923732424371539, + "grad_norm": 13.375, + "learning_rate": 3.798254714132382e-07, + "loss": 0.6811, + "step": 17513 + }, + { + "epoch": 1.4924584576054538, + "grad_norm": 13.9375, + "learning_rate": 3.7970484547580026e-07, + "loss": 0.7262, + "step": 17514 + }, + { + "epoch": 1.4925436727737538, + "grad_norm": 13.125, + "learning_rate": 3.7958423526547037e-07, + "loss": 0.4447, + "step": 17515 + }, + { + "epoch": 1.4926288879420537, + "grad_norm": 12.75, + "learning_rate": 3.7946364078442864e-07, + "loss": 0.5503, + "step": 17516 + }, + { + "epoch": 1.4927141031103537, + "grad_norm": 20.5, + "learning_rate": 3.793430620348544e-07, + "loss": 0.5929, + "step": 17517 + }, + { + "epoch": 1.4927993182786536, + "grad_norm": 13.5625, + "learning_rate": 3.7922249901892694e-07, + "loss": 0.5588, + "step": 17518 + }, + { + "epoch": 1.4928845334469536, + "grad_norm": 17.5, + "learning_rate": 3.7910195173882405e-07, + "loss": 0.5391, + "step": 17519 + }, + { + "epoch": 1.4929697486152536, + "grad_norm": 13.0625, + "learning_rate": 3.7898142019672487e-07, + "loss": 0.3766, + "step": 17520 + }, + { + "epoch": 1.4930549637835535, + "grad_norm": 17.0, + "learning_rate": 3.7886090439480745e-07, + "loss": 0.5782, + "step": 17521 + }, + { + "epoch": 1.4931401789518535, + "grad_norm": 11.5625, + "learning_rate": 3.7874040433524966e-07, + "loss": 0.3422, + "step": 17522 + }, + { + "epoch": 1.4932253941201534, + "grad_norm": 22.75, + "learning_rate": 3.7861992002022843e-07, + "loss": 0.8153, + "step": 17523 + }, + { + "epoch": 1.4933106092884534, + "grad_norm": 21.0, + "learning_rate": 3.784994514519215e-07, + "loss": 0.7669, + "step": 17524 + }, + { + "epoch": 1.4933958244567533, + "grad_norm": 12.625, + "learning_rate": 3.783789986325058e-07, + "loss": 0.2496, + "step": 17525 + }, + { + "epoch": 1.4934810396250533, + "grad_norm": 14.0, + "learning_rate": 3.782585615641585e-07, + "loss": 0.4726, + "step": 17526 + }, + { + "epoch": 1.4935662547933533, + "grad_norm": 14.3125, + "learning_rate": 3.781381402490551e-07, + "loss": 0.3446, + "step": 17527 + }, + { + "epoch": 1.4936514699616532, + "grad_norm": 12.4375, + "learning_rate": 3.7801773468937235e-07, + "loss": 0.3443, + "step": 17528 + }, + { + "epoch": 1.4937366851299532, + "grad_norm": 18.125, + "learning_rate": 3.7789734488728593e-07, + "loss": 0.5557, + "step": 17529 + }, + { + "epoch": 1.4938219002982531, + "grad_norm": 9.625, + "learning_rate": 3.777769708449719e-07, + "loss": 0.2419, + "step": 17530 + }, + { + "epoch": 1.493907115466553, + "grad_norm": 16.625, + "learning_rate": 3.77656612564605e-07, + "loss": 0.5952, + "step": 17531 + }, + { + "epoch": 1.493992330634853, + "grad_norm": 13.375, + "learning_rate": 3.775362700483601e-07, + "loss": 0.5698, + "step": 17532 + }, + { + "epoch": 1.494077545803153, + "grad_norm": 17.875, + "learning_rate": 3.7741594329841204e-07, + "loss": 0.743, + "step": 17533 + }, + { + "epoch": 1.494162760971453, + "grad_norm": 17.625, + "learning_rate": 3.772956323169352e-07, + "loss": 0.6814, + "step": 17534 + }, + { + "epoch": 1.494247976139753, + "grad_norm": 17.125, + "learning_rate": 3.7717533710610445e-07, + "loss": 0.667, + "step": 17535 + }, + { + "epoch": 1.4943331913080529, + "grad_norm": 15.875, + "learning_rate": 3.7705505766809256e-07, + "loss": 0.7306, + "step": 17536 + }, + { + "epoch": 1.4944184064763528, + "grad_norm": 13.6875, + "learning_rate": 3.7693479400507357e-07, + "loss": 0.6171, + "step": 17537 + }, + { + "epoch": 1.4945036216446528, + "grad_norm": 17.375, + "learning_rate": 3.768145461192213e-07, + "loss": 0.6451, + "step": 17538 + }, + { + "epoch": 1.4945888368129527, + "grad_norm": 15.0, + "learning_rate": 3.7669431401270814e-07, + "loss": 0.4427, + "step": 17539 + }, + { + "epoch": 1.4946740519812527, + "grad_norm": 18.5, + "learning_rate": 3.765740976877065e-07, + "loss": 0.7907, + "step": 17540 + }, + { + "epoch": 1.4947592671495527, + "grad_norm": 14.5, + "learning_rate": 3.764538971463892e-07, + "loss": 0.4297, + "step": 17541 + }, + { + "epoch": 1.4948444823178526, + "grad_norm": 15.75, + "learning_rate": 3.7633371239092847e-07, + "loss": 0.4595, + "step": 17542 + }, + { + "epoch": 1.4949296974861526, + "grad_norm": 12.9375, + "learning_rate": 3.7621354342349646e-07, + "loss": 0.4803, + "step": 17543 + }, + { + "epoch": 1.4950149126544525, + "grad_norm": 20.125, + "learning_rate": 3.760933902462639e-07, + "loss": 0.8995, + "step": 17544 + }, + { + "epoch": 1.4951001278227525, + "grad_norm": 14.1875, + "learning_rate": 3.759732528614027e-07, + "loss": 0.5473, + "step": 17545 + }, + { + "epoch": 1.4951853429910524, + "grad_norm": 18.75, + "learning_rate": 3.758531312710836e-07, + "loss": 0.6203, + "step": 17546 + }, + { + "epoch": 1.4952705581593524, + "grad_norm": 17.5, + "learning_rate": 3.75733025477478e-07, + "loss": 0.741, + "step": 17547 + }, + { + "epoch": 1.4953557733276523, + "grad_norm": 15.0, + "learning_rate": 3.7561293548275565e-07, + "loss": 0.7313, + "step": 17548 + }, + { + "epoch": 1.4954409884959523, + "grad_norm": 13.0625, + "learning_rate": 3.754928612890865e-07, + "loss": 0.4672, + "step": 17549 + }, + { + "epoch": 1.4955262036642523, + "grad_norm": 13.875, + "learning_rate": 3.753728028986406e-07, + "loss": 0.5798, + "step": 17550 + }, + { + "epoch": 1.4956114188325522, + "grad_norm": 13.25, + "learning_rate": 3.752527603135882e-07, + "loss": 0.5281, + "step": 17551 + }, + { + "epoch": 1.4956966340008522, + "grad_norm": 20.0, + "learning_rate": 3.7513273353609754e-07, + "loss": 0.6578, + "step": 17552 + }, + { + "epoch": 1.4957818491691521, + "grad_norm": 13.25, + "learning_rate": 3.7501272256833824e-07, + "loss": 0.5646, + "step": 17553 + }, + { + "epoch": 1.495867064337452, + "grad_norm": 13.4375, + "learning_rate": 3.7489272741247877e-07, + "loss": 0.4328, + "step": 17554 + }, + { + "epoch": 1.495952279505752, + "grad_norm": 14.8125, + "learning_rate": 3.747727480706882e-07, + "loss": 0.5498, + "step": 17555 + }, + { + "epoch": 1.496037494674052, + "grad_norm": 14.6875, + "learning_rate": 3.7465278454513423e-07, + "loss": 0.6597, + "step": 17556 + }, + { + "epoch": 1.496122709842352, + "grad_norm": 23.25, + "learning_rate": 3.745328368379841e-07, + "loss": 1.0158, + "step": 17557 + }, + { + "epoch": 1.496207925010652, + "grad_norm": 15.25, + "learning_rate": 3.744129049514061e-07, + "loss": 0.7415, + "step": 17558 + }, + { + "epoch": 1.4962931401789519, + "grad_norm": 14.4375, + "learning_rate": 3.742929888875674e-07, + "loss": 0.7127, + "step": 17559 + }, + { + "epoch": 1.4963783553472518, + "grad_norm": 14.125, + "learning_rate": 3.741730886486353e-07, + "loss": 0.4559, + "step": 17560 + }, + { + "epoch": 1.4964635705155518, + "grad_norm": 15.6875, + "learning_rate": 3.740532042367759e-07, + "loss": 0.7112, + "step": 17561 + }, + { + "epoch": 1.4965487856838517, + "grad_norm": 18.875, + "learning_rate": 3.739333356541558e-07, + "loss": 0.8795, + "step": 17562 + }, + { + "epoch": 1.4966340008521517, + "grad_norm": 16.125, + "learning_rate": 3.7381348290294174e-07, + "loss": 0.5574, + "step": 17563 + }, + { + "epoch": 1.4967192160204517, + "grad_norm": 13.1875, + "learning_rate": 3.736936459852987e-07, + "loss": 0.397, + "step": 17564 + }, + { + "epoch": 1.4968044311887516, + "grad_norm": 17.125, + "learning_rate": 3.7357382490339307e-07, + "loss": 0.45, + "step": 17565 + }, + { + "epoch": 1.4968896463570516, + "grad_norm": 21.0, + "learning_rate": 3.734540196593893e-07, + "loss": 0.8355, + "step": 17566 + }, + { + "epoch": 1.4969748615253515, + "grad_norm": 16.875, + "learning_rate": 3.73334230255453e-07, + "loss": 0.6461, + "step": 17567 + }, + { + "epoch": 1.4970600766936515, + "grad_norm": 18.75, + "learning_rate": 3.7321445669374884e-07, + "loss": 0.5792, + "step": 17568 + }, + { + "epoch": 1.4971452918619514, + "grad_norm": 11.8125, + "learning_rate": 3.73094698976441e-07, + "loss": 0.3175, + "step": 17569 + }, + { + "epoch": 1.4972305070302514, + "grad_norm": 13.6875, + "learning_rate": 3.729749571056936e-07, + "loss": 0.5115, + "step": 17570 + }, + { + "epoch": 1.4973157221985514, + "grad_norm": 18.75, + "learning_rate": 3.728552310836711e-07, + "loss": 0.7201, + "step": 17571 + }, + { + "epoch": 1.4974009373668513, + "grad_norm": 13.0625, + "learning_rate": 3.7273552091253606e-07, + "loss": 0.4319, + "step": 17572 + }, + { + "epoch": 1.4974861525351513, + "grad_norm": 13.0, + "learning_rate": 3.726158265944529e-07, + "loss": 0.4696, + "step": 17573 + }, + { + "epoch": 1.4975713677034512, + "grad_norm": 13.75, + "learning_rate": 3.724961481315835e-07, + "loss": 0.583, + "step": 17574 + }, + { + "epoch": 1.4976565828717512, + "grad_norm": 18.375, + "learning_rate": 3.723764855260913e-07, + "loss": 0.8521, + "step": 17575 + }, + { + "epoch": 1.4977417980400511, + "grad_norm": 16.125, + "learning_rate": 3.7225683878013825e-07, + "loss": 0.6204, + "step": 17576 + }, + { + "epoch": 1.497827013208351, + "grad_norm": 16.75, + "learning_rate": 3.7213720789588735e-07, + "loss": 0.7008, + "step": 17577 + }, + { + "epoch": 1.497912228376651, + "grad_norm": 15.125, + "learning_rate": 3.7201759287549947e-07, + "loss": 0.6444, + "step": 17578 + }, + { + "epoch": 1.497997443544951, + "grad_norm": 12.25, + "learning_rate": 3.7189799372113687e-07, + "loss": 0.4901, + "step": 17579 + }, + { + "epoch": 1.498082658713251, + "grad_norm": 17.25, + "learning_rate": 3.7177841043496017e-07, + "loss": 0.5006, + "step": 17580 + }, + { + "epoch": 1.498167873881551, + "grad_norm": 13.8125, + "learning_rate": 3.716588430191312e-07, + "loss": 0.5075, + "step": 17581 + }, + { + "epoch": 1.4982530890498509, + "grad_norm": 13.75, + "learning_rate": 3.7153929147580967e-07, + "loss": 0.4614, + "step": 17582 + }, + { + "epoch": 1.4983383042181508, + "grad_norm": 19.0, + "learning_rate": 3.7141975580715653e-07, + "loss": 0.5827, + "step": 17583 + }, + { + "epoch": 1.4984235193864508, + "grad_norm": 17.375, + "learning_rate": 3.7130023601533183e-07, + "loss": 0.6694, + "step": 17584 + }, + { + "epoch": 1.4985087345547508, + "grad_norm": 14.9375, + "learning_rate": 3.711807321024959e-07, + "loss": 0.5078, + "step": 17585 + }, + { + "epoch": 1.4985939497230507, + "grad_norm": 12.0, + "learning_rate": 3.710612440708075e-07, + "loss": 0.6483, + "step": 17586 + }, + { + "epoch": 1.4986791648913507, + "grad_norm": 12.4375, + "learning_rate": 3.7094177192242664e-07, + "loss": 0.4887, + "step": 17587 + }, + { + "epoch": 1.4987643800596506, + "grad_norm": 22.25, + "learning_rate": 3.708223156595114e-07, + "loss": 0.8873, + "step": 17588 + }, + { + "epoch": 1.4988495952279506, + "grad_norm": 21.75, + "learning_rate": 3.707028752842212e-07, + "loss": 1.1401, + "step": 17589 + }, + { + "epoch": 1.4989348103962505, + "grad_norm": 25.625, + "learning_rate": 3.7058345079871453e-07, + "loss": 0.8154, + "step": 17590 + }, + { + "epoch": 1.4990200255645505, + "grad_norm": 14.8125, + "learning_rate": 3.7046404220514894e-07, + "loss": 0.4349, + "step": 17591 + }, + { + "epoch": 1.4991052407328505, + "grad_norm": 20.875, + "learning_rate": 3.703446495056824e-07, + "loss": 0.6151, + "step": 17592 + }, + { + "epoch": 1.4991904559011504, + "grad_norm": 16.5, + "learning_rate": 3.702252727024727e-07, + "loss": 0.4761, + "step": 17593 + }, + { + "epoch": 1.4992756710694504, + "grad_norm": 11.875, + "learning_rate": 3.7010591179767725e-07, + "loss": 0.9119, + "step": 17594 + }, + { + "epoch": 1.4993608862377503, + "grad_norm": 13.0, + "learning_rate": 3.6998656679345253e-07, + "loss": 0.379, + "step": 17595 + }, + { + "epoch": 1.4994461014060503, + "grad_norm": 31.875, + "learning_rate": 3.698672376919558e-07, + "loss": 0.657, + "step": 17596 + }, + { + "epoch": 1.4995313165743502, + "grad_norm": 17.75, + "learning_rate": 3.6974792449534283e-07, + "loss": 0.5086, + "step": 17597 + }, + { + "epoch": 1.4996165317426502, + "grad_norm": 16.375, + "learning_rate": 3.696286272057703e-07, + "loss": 0.4567, + "step": 17598 + }, + { + "epoch": 1.4997017469109502, + "grad_norm": 12.125, + "learning_rate": 3.6950934582539346e-07, + "loss": 0.3842, + "step": 17599 + }, + { + "epoch": 1.49978696207925, + "grad_norm": 10.25, + "learning_rate": 3.693900803563681e-07, + "loss": 0.3043, + "step": 17600 + }, + { + "epoch": 1.49987217724755, + "grad_norm": 14.6875, + "learning_rate": 3.6927083080084955e-07, + "loss": 0.2383, + "step": 17601 + }, + { + "epoch": 1.49995739241585, + "grad_norm": 12.8125, + "learning_rate": 3.6915159716099313e-07, + "loss": 0.4425, + "step": 17602 + }, + { + "epoch": 1.50004260758415, + "grad_norm": 16.375, + "learning_rate": 3.690323794389526e-07, + "loss": 0.158, + "step": 17603 + }, + { + "epoch": 1.50012782275245, + "grad_norm": 12.5625, + "learning_rate": 3.689131776368833e-07, + "loss": 0.8201, + "step": 17604 + }, + { + "epoch": 1.50021303792075, + "grad_norm": 13.375, + "learning_rate": 3.687939917569386e-07, + "loss": 0.3662, + "step": 17605 + }, + { + "epoch": 1.5002982530890498, + "grad_norm": 20.75, + "learning_rate": 3.686748218012724e-07, + "loss": 1.1057, + "step": 17606 + }, + { + "epoch": 1.5003834682573498, + "grad_norm": 14.625, + "learning_rate": 3.6855566777203894e-07, + "loss": 0.4443, + "step": 17607 + }, + { + "epoch": 1.5004686834256498, + "grad_norm": 14.8125, + "learning_rate": 3.6843652967139055e-07, + "loss": 0.4159, + "step": 17608 + }, + { + "epoch": 1.5005538985939497, + "grad_norm": 12.5625, + "learning_rate": 3.683174075014805e-07, + "loss": 0.3446, + "step": 17609 + }, + { + "epoch": 1.5006391137622497, + "grad_norm": 16.0, + "learning_rate": 3.681983012644615e-07, + "loss": 0.3066, + "step": 17610 + }, + { + "epoch": 1.5007243289305496, + "grad_norm": 19.375, + "learning_rate": 3.680792109624863e-07, + "loss": 0.7687, + "step": 17611 + }, + { + "epoch": 1.5008095440988496, + "grad_norm": 11.6875, + "learning_rate": 3.679601365977066e-07, + "loss": 0.3632, + "step": 17612 + }, + { + "epoch": 1.5008947592671495, + "grad_norm": 15.8125, + "learning_rate": 3.6784107817227383e-07, + "loss": 0.6686, + "step": 17613 + }, + { + "epoch": 1.5009799744354495, + "grad_norm": 11.5, + "learning_rate": 3.677220356883397e-07, + "loss": 0.3584, + "step": 17614 + }, + { + "epoch": 1.5010651896037495, + "grad_norm": 14.625, + "learning_rate": 3.6760300914805616e-07, + "loss": 0.7623, + "step": 17615 + }, + { + "epoch": 1.5011504047720494, + "grad_norm": 13.5625, + "learning_rate": 3.6748399855357314e-07, + "loss": 0.5739, + "step": 17616 + }, + { + "epoch": 1.5012356199403494, + "grad_norm": 19.5, + "learning_rate": 3.673650039070417e-07, + "loss": 0.6488, + "step": 17617 + }, + { + "epoch": 1.5013208351086493, + "grad_norm": 19.25, + "learning_rate": 3.672460252106122e-07, + "loss": 0.7729, + "step": 17618 + }, + { + "epoch": 1.5014060502769493, + "grad_norm": 22.875, + "learning_rate": 3.6712706246643496e-07, + "loss": 0.7672, + "step": 17619 + }, + { + "epoch": 1.5014912654452492, + "grad_norm": 13.0, + "learning_rate": 3.6700811567665957e-07, + "loss": 0.4725, + "step": 17620 + }, + { + "epoch": 1.5015764806135492, + "grad_norm": 15.5, + "learning_rate": 3.6688918484343515e-07, + "loss": 0.6762, + "step": 17621 + }, + { + "epoch": 1.5016616957818492, + "grad_norm": 17.875, + "learning_rate": 3.6677026996891116e-07, + "loss": 0.6464, + "step": 17622 + }, + { + "epoch": 1.5017469109501491, + "grad_norm": 21.25, + "learning_rate": 3.6665137105523646e-07, + "loss": 0.8062, + "step": 17623 + }, + { + "epoch": 1.501832126118449, + "grad_norm": 17.375, + "learning_rate": 3.6653248810456025e-07, + "loss": 0.6739, + "step": 17624 + }, + { + "epoch": 1.501917341286749, + "grad_norm": 19.75, + "learning_rate": 3.6641362111903e-07, + "loss": 0.6772, + "step": 17625 + }, + { + "epoch": 1.502002556455049, + "grad_norm": 14.125, + "learning_rate": 3.662947701007941e-07, + "loss": 0.4247, + "step": 17626 + }, + { + "epoch": 1.502087771623349, + "grad_norm": 20.875, + "learning_rate": 3.661759350520003e-07, + "loss": 0.7413, + "step": 17627 + }, + { + "epoch": 1.502172986791649, + "grad_norm": 12.75, + "learning_rate": 3.660571159747969e-07, + "loss": 0.4522, + "step": 17628 + }, + { + "epoch": 1.5022582019599489, + "grad_norm": 16.5, + "learning_rate": 3.6593831287132953e-07, + "loss": 0.5569, + "step": 17629 + }, + { + "epoch": 1.5023434171282488, + "grad_norm": 15.375, + "learning_rate": 3.658195257437458e-07, + "loss": 0.8198, + "step": 17630 + }, + { + "epoch": 1.5024286322965488, + "grad_norm": 16.0, + "learning_rate": 3.657007545941924e-07, + "loss": 0.3184, + "step": 17631 + }, + { + "epoch": 1.5025138474648487, + "grad_norm": 11.625, + "learning_rate": 3.6558199942481595e-07, + "loss": 0.6567, + "step": 17632 + }, + { + "epoch": 1.5025990626331487, + "grad_norm": 17.0, + "learning_rate": 3.6546326023776184e-07, + "loss": 0.7709, + "step": 17633 + }, + { + "epoch": 1.5026842778014486, + "grad_norm": 15.6875, + "learning_rate": 3.653445370351759e-07, + "loss": 0.8675, + "step": 17634 + }, + { + "epoch": 1.5027694929697486, + "grad_norm": 14.5625, + "learning_rate": 3.6522582981920395e-07, + "loss": 0.5696, + "step": 17635 + }, + { + "epoch": 1.5028547081380486, + "grad_norm": 13.875, + "learning_rate": 3.651071385919912e-07, + "loss": 0.3766, + "step": 17636 + }, + { + "epoch": 1.5029399233063485, + "grad_norm": 17.625, + "learning_rate": 3.649884633556823e-07, + "loss": 0.5326, + "step": 17637 + }, + { + "epoch": 1.5030251384746485, + "grad_norm": 14.0625, + "learning_rate": 3.648698041124216e-07, + "loss": 0.5351, + "step": 17638 + }, + { + "epoch": 1.5031103536429484, + "grad_norm": 14.25, + "learning_rate": 3.6475116086435335e-07, + "loss": 0.6083, + "step": 17639 + }, + { + "epoch": 1.5031955688112484, + "grad_norm": 17.375, + "learning_rate": 3.6463253361362193e-07, + "loss": 0.619, + "step": 17640 + }, + { + "epoch": 1.5032807839795483, + "grad_norm": 25.375, + "learning_rate": 3.6451392236237145e-07, + "loss": 0.7584, + "step": 17641 + }, + { + "epoch": 1.5033659991478483, + "grad_norm": 14.5, + "learning_rate": 3.6439532711274426e-07, + "loss": 0.5143, + "step": 17642 + }, + { + "epoch": 1.5034512143161483, + "grad_norm": 13.875, + "learning_rate": 3.6427674786688405e-07, + "loss": 0.3843, + "step": 17643 + }, + { + "epoch": 1.5035364294844482, + "grad_norm": 12.1875, + "learning_rate": 3.6415818462693417e-07, + "loss": 0.4416, + "step": 17644 + }, + { + "epoch": 1.5036216446527482, + "grad_norm": 16.125, + "learning_rate": 3.6403963739503666e-07, + "loss": 0.7131, + "step": 17645 + }, + { + "epoch": 1.5037068598210481, + "grad_norm": 24.625, + "learning_rate": 3.639211061733333e-07, + "loss": 0.713, + "step": 17646 + }, + { + "epoch": 1.503792074989348, + "grad_norm": 13.5, + "learning_rate": 3.638025909639667e-07, + "loss": 0.5319, + "step": 17647 + }, + { + "epoch": 1.503877290157648, + "grad_norm": 16.625, + "learning_rate": 3.6368409176907844e-07, + "loss": 0.4675, + "step": 17648 + }, + { + "epoch": 1.503962505325948, + "grad_norm": 11.875, + "learning_rate": 3.6356560859081016e-07, + "loss": 0.4492, + "step": 17649 + }, + { + "epoch": 1.504047720494248, + "grad_norm": 14.4375, + "learning_rate": 3.6344714143130234e-07, + "loss": 0.3993, + "step": 17650 + }, + { + "epoch": 1.504132935662548, + "grad_norm": 18.375, + "learning_rate": 3.6332869029269623e-07, + "loss": 0.7939, + "step": 17651 + }, + { + "epoch": 1.5042181508308479, + "grad_norm": 24.0, + "learning_rate": 3.632102551771327e-07, + "loss": 0.9664, + "step": 17652 + }, + { + "epoch": 1.5043033659991478, + "grad_norm": 14.0, + "learning_rate": 3.630918360867512e-07, + "loss": 0.673, + "step": 17653 + }, + { + "epoch": 1.5043885811674478, + "grad_norm": 12.75, + "learning_rate": 3.629734330236924e-07, + "loss": 0.4649, + "step": 17654 + }, + { + "epoch": 1.5044737963357477, + "grad_norm": 13.125, + "learning_rate": 3.6285504599009536e-07, + "loss": 0.4627, + "step": 17655 + }, + { + "epoch": 1.5045590115040477, + "grad_norm": 16.75, + "learning_rate": 3.6273667498809965e-07, + "loss": 0.5896, + "step": 17656 + }, + { + "epoch": 1.5046442266723477, + "grad_norm": 14.75, + "learning_rate": 3.626183200198445e-07, + "loss": 0.7516, + "step": 17657 + }, + { + "epoch": 1.5047294418406476, + "grad_norm": 22.375, + "learning_rate": 3.6249998108746897e-07, + "loss": 1.1351, + "step": 17658 + }, + { + "epoch": 1.5048146570089476, + "grad_norm": 12.125, + "learning_rate": 3.6238165819311083e-07, + "loss": 0.3761, + "step": 17659 + }, + { + "epoch": 1.5048998721772475, + "grad_norm": 14.9375, + "learning_rate": 3.6226335133890915e-07, + "loss": 0.6368, + "step": 17660 + }, + { + "epoch": 1.5049850873455475, + "grad_norm": 13.3125, + "learning_rate": 3.621450605270009e-07, + "loss": 0.4275, + "step": 17661 + }, + { + "epoch": 1.5050703025138474, + "grad_norm": 21.25, + "learning_rate": 3.620267857595247e-07, + "loss": 0.6888, + "step": 17662 + }, + { + "epoch": 1.5051555176821474, + "grad_norm": 15.3125, + "learning_rate": 3.61908527038617e-07, + "loss": 0.4342, + "step": 17663 + }, + { + "epoch": 1.5052407328504473, + "grad_norm": 25.25, + "learning_rate": 3.617902843664153e-07, + "loss": 0.7437, + "step": 17664 + }, + { + "epoch": 1.5053259480187473, + "grad_norm": 17.0, + "learning_rate": 3.6167205774505625e-07, + "loss": 0.5054, + "step": 17665 + }, + { + "epoch": 1.5054111631870473, + "grad_norm": 18.625, + "learning_rate": 3.6155384717667676e-07, + "loss": 0.9845, + "step": 17666 + }, + { + "epoch": 1.5054963783553472, + "grad_norm": 17.625, + "learning_rate": 3.6143565266341234e-07, + "loss": 0.7661, + "step": 17667 + }, + { + "epoch": 1.5055815935236472, + "grad_norm": 16.625, + "learning_rate": 3.6131747420739925e-07, + "loss": 0.6002, + "step": 17668 + }, + { + "epoch": 1.5056668086919471, + "grad_norm": 26.0, + "learning_rate": 3.611993118107733e-07, + "loss": 0.5808, + "step": 17669 + }, + { + "epoch": 1.505752023860247, + "grad_norm": 17.625, + "learning_rate": 3.610811654756692e-07, + "loss": 0.9035, + "step": 17670 + }, + { + "epoch": 1.505837239028547, + "grad_norm": 12.0625, + "learning_rate": 3.6096303520422274e-07, + "loss": 0.365, + "step": 17671 + }, + { + "epoch": 1.505922454196847, + "grad_norm": 14.5625, + "learning_rate": 3.608449209985679e-07, + "loss": 0.4467, + "step": 17672 + }, + { + "epoch": 1.506007669365147, + "grad_norm": 15.875, + "learning_rate": 3.607268228608393e-07, + "loss": 0.5869, + "step": 17673 + }, + { + "epoch": 1.506092884533447, + "grad_norm": 18.0, + "learning_rate": 3.606087407931713e-07, + "loss": 0.5201, + "step": 17674 + }, + { + "epoch": 1.5061780997017469, + "grad_norm": 24.25, + "learning_rate": 3.604906747976981e-07, + "loss": 0.8088, + "step": 17675 + }, + { + "epoch": 1.5062633148700468, + "grad_norm": 15.9375, + "learning_rate": 3.6037262487655255e-07, + "loss": 0.7864, + "step": 17676 + }, + { + "epoch": 1.5063485300383468, + "grad_norm": 11.6875, + "learning_rate": 3.602545910318686e-07, + "loss": 0.6319, + "step": 17677 + }, + { + "epoch": 1.5064337452066467, + "grad_norm": 11.9375, + "learning_rate": 3.6013657326577856e-07, + "loss": 0.3707, + "step": 17678 + }, + { + "epoch": 1.5065189603749467, + "grad_norm": 16.625, + "learning_rate": 3.600185715804158e-07, + "loss": 0.6445, + "step": 17679 + }, + { + "epoch": 1.5066041755432467, + "grad_norm": 14.9375, + "learning_rate": 3.59900585977912e-07, + "loss": 0.5241, + "step": 17680 + }, + { + "epoch": 1.5066893907115466, + "grad_norm": 14.8125, + "learning_rate": 3.597826164603997e-07, + "loss": 0.5009, + "step": 17681 + }, + { + "epoch": 1.5067746058798466, + "grad_norm": 15.6875, + "learning_rate": 3.596646630300106e-07, + "loss": 0.7159, + "step": 17682 + }, + { + "epoch": 1.5068598210481465, + "grad_norm": 15.25, + "learning_rate": 3.5954672568887685e-07, + "loss": 0.6728, + "step": 17683 + }, + { + "epoch": 1.5069450362164465, + "grad_norm": 25.5, + "learning_rate": 3.5942880443912876e-07, + "loss": 0.8292, + "step": 17684 + }, + { + "epoch": 1.5070302513847464, + "grad_norm": 20.375, + "learning_rate": 3.5931089928289824e-07, + "loss": 0.975, + "step": 17685 + }, + { + "epoch": 1.5071154665530464, + "grad_norm": 13.875, + "learning_rate": 3.5919301022231485e-07, + "loss": 0.5244, + "step": 17686 + }, + { + "epoch": 1.5072006817213464, + "grad_norm": 20.375, + "learning_rate": 3.5907513725950956e-07, + "loss": 0.8668, + "step": 17687 + }, + { + "epoch": 1.5072858968896463, + "grad_norm": 17.0, + "learning_rate": 3.589572803966129e-07, + "loss": 0.5268, + "step": 17688 + }, + { + "epoch": 1.5073711120579463, + "grad_norm": 15.625, + "learning_rate": 3.588394396357538e-07, + "loss": 0.6326, + "step": 17689 + }, + { + "epoch": 1.5074563272262462, + "grad_norm": 16.25, + "learning_rate": 3.5872161497906204e-07, + "loss": 0.7481, + "step": 17690 + }, + { + "epoch": 1.5075415423945462, + "grad_norm": 14.4375, + "learning_rate": 3.586038064286675e-07, + "loss": 0.6844, + "step": 17691 + }, + { + "epoch": 1.5076267575628461, + "grad_norm": 14.0, + "learning_rate": 3.58486013986698e-07, + "loss": 0.5031, + "step": 17692 + }, + { + "epoch": 1.507711972731146, + "grad_norm": 16.125, + "learning_rate": 3.583682376552833e-07, + "loss": 0.5117, + "step": 17693 + }, + { + "epoch": 1.507797187899446, + "grad_norm": 25.375, + "learning_rate": 3.5825047743655057e-07, + "loss": 0.6086, + "step": 17694 + }, + { + "epoch": 1.507882403067746, + "grad_norm": 17.375, + "learning_rate": 3.581327333326284e-07, + "loss": 0.5559, + "step": 17695 + }, + { + "epoch": 1.507967618236046, + "grad_norm": 14.875, + "learning_rate": 3.580150053456452e-07, + "loss": 0.6285, + "step": 17696 + }, + { + "epoch": 1.508052833404346, + "grad_norm": 12.5, + "learning_rate": 3.5789729347772724e-07, + "loss": 0.4137, + "step": 17697 + }, + { + "epoch": 1.5081380485726459, + "grad_norm": 30.875, + "learning_rate": 3.577795977310024e-07, + "loss": 0.767, + "step": 17698 + }, + { + "epoch": 1.5082232637409458, + "grad_norm": 12.3125, + "learning_rate": 3.576619181075973e-07, + "loss": 0.4889, + "step": 17699 + }, + { + "epoch": 1.5083084789092458, + "grad_norm": 13.8125, + "learning_rate": 3.5754425460963914e-07, + "loss": 0.5556, + "step": 17700 + }, + { + "epoch": 1.5083936940775458, + "grad_norm": 19.375, + "learning_rate": 3.5742660723925363e-07, + "loss": 0.6036, + "step": 17701 + }, + { + "epoch": 1.5084789092458457, + "grad_norm": 12.625, + "learning_rate": 3.5730897599856667e-07, + "loss": 0.5615, + "step": 17702 + }, + { + "epoch": 1.5085641244141457, + "grad_norm": 12.5, + "learning_rate": 3.57191360889704e-07, + "loss": 0.532, + "step": 17703 + }, + { + "epoch": 1.5086493395824456, + "grad_norm": 22.75, + "learning_rate": 3.570737619147913e-07, + "loss": 0.7152, + "step": 17704 + }, + { + "epoch": 1.5087345547507456, + "grad_norm": 20.375, + "learning_rate": 3.56956179075954e-07, + "loss": 0.5297, + "step": 17705 + }, + { + "epoch": 1.5088197699190455, + "grad_norm": 15.8125, + "learning_rate": 3.568386123753162e-07, + "loss": 0.5827, + "step": 17706 + }, + { + "epoch": 1.5089049850873455, + "grad_norm": 17.625, + "learning_rate": 3.5672106181500284e-07, + "loss": 0.9004, + "step": 17707 + }, + { + "epoch": 1.5089902002556455, + "grad_norm": 14.125, + "learning_rate": 3.5660352739713853e-07, + "loss": 0.5621, + "step": 17708 + }, + { + "epoch": 1.5090754154239454, + "grad_norm": 12.875, + "learning_rate": 3.564860091238467e-07, + "loss": 0.6179, + "step": 17709 + }, + { + "epoch": 1.5091606305922454, + "grad_norm": 20.625, + "learning_rate": 3.56368506997251e-07, + "loss": 0.879, + "step": 17710 + }, + { + "epoch": 1.5092458457605453, + "grad_norm": 13.375, + "learning_rate": 3.56251021019475e-07, + "loss": 0.5925, + "step": 17711 + }, + { + "epoch": 1.5093310609288453, + "grad_norm": 17.5, + "learning_rate": 3.561335511926417e-07, + "loss": 0.7588, + "step": 17712 + }, + { + "epoch": 1.5094162760971452, + "grad_norm": 15.1875, + "learning_rate": 3.5601609751887434e-07, + "loss": 0.5237, + "step": 17713 + }, + { + "epoch": 1.5095014912654452, + "grad_norm": 16.375, + "learning_rate": 3.558986600002948e-07, + "loss": 0.4084, + "step": 17714 + }, + { + "epoch": 1.5095867064337452, + "grad_norm": 14.1875, + "learning_rate": 3.557812386390254e-07, + "loss": 0.5346, + "step": 17715 + }, + { + "epoch": 1.5096719216020453, + "grad_norm": 14.125, + "learning_rate": 3.556638334371884e-07, + "loss": 0.6737, + "step": 17716 + }, + { + "epoch": 1.5097571367703453, + "grad_norm": 13.8125, + "learning_rate": 3.5554644439690545e-07, + "loss": 0.7111, + "step": 17717 + }, + { + "epoch": 1.5098423519386452, + "grad_norm": 16.0, + "learning_rate": 3.554290715202978e-07, + "loss": 0.6989, + "step": 17718 + }, + { + "epoch": 1.5099275671069452, + "grad_norm": 15.8125, + "learning_rate": 3.5531171480948604e-07, + "loss": 0.8963, + "step": 17719 + }, + { + "epoch": 1.5100127822752452, + "grad_norm": 21.375, + "learning_rate": 3.5519437426659126e-07, + "loss": 0.6907, + "step": 17720 + }, + { + "epoch": 1.5100979974435451, + "grad_norm": 16.75, + "learning_rate": 3.5507704989373434e-07, + "loss": 0.6058, + "step": 17721 + }, + { + "epoch": 1.510183212611845, + "grad_norm": 12.625, + "learning_rate": 3.549597416930345e-07, + "loss": 0.4691, + "step": 17722 + }, + { + "epoch": 1.510268427780145, + "grad_norm": 13.5625, + "learning_rate": 3.5484244966661224e-07, + "loss": 0.4355, + "step": 17723 + }, + { + "epoch": 1.510353642948445, + "grad_norm": 12.75, + "learning_rate": 3.5472517381658716e-07, + "loss": 0.3041, + "step": 17724 + }, + { + "epoch": 1.510438858116745, + "grad_norm": 19.0, + "learning_rate": 3.546079141450788e-07, + "loss": 0.6243, + "step": 17725 + }, + { + "epoch": 1.510524073285045, + "grad_norm": 17.0, + "learning_rate": 3.544906706542058e-07, + "loss": 0.6576, + "step": 17726 + }, + { + "epoch": 1.5106092884533449, + "grad_norm": 18.125, + "learning_rate": 3.5437344334608645e-07, + "loss": 0.9424, + "step": 17727 + }, + { + "epoch": 1.5106945036216448, + "grad_norm": 16.75, + "learning_rate": 3.5425623222283964e-07, + "loss": 0.8699, + "step": 17728 + }, + { + "epoch": 1.5107797187899448, + "grad_norm": 14.75, + "learning_rate": 3.541390372865834e-07, + "loss": 0.6933, + "step": 17729 + }, + { + "epoch": 1.5108649339582447, + "grad_norm": 13.1875, + "learning_rate": 3.5402185853943617e-07, + "loss": 0.4274, + "step": 17730 + }, + { + "epoch": 1.5109501491265447, + "grad_norm": 16.5, + "learning_rate": 3.5390469598351436e-07, + "loss": 0.5428, + "step": 17731 + }, + { + "epoch": 1.5110353642948446, + "grad_norm": 13.5, + "learning_rate": 3.5378754962093587e-07, + "loss": 0.5076, + "step": 17732 + }, + { + "epoch": 1.5111205794631446, + "grad_norm": 11.5625, + "learning_rate": 3.5367041945381765e-07, + "loss": 0.2623, + "step": 17733 + }, + { + "epoch": 1.5112057946314446, + "grad_norm": 16.0, + "learning_rate": 3.535533054842766e-07, + "loss": 0.5259, + "step": 17734 + }, + { + "epoch": 1.5112910097997445, + "grad_norm": 15.6875, + "learning_rate": 3.5343620771442864e-07, + "loss": 0.5355, + "step": 17735 + }, + { + "epoch": 1.5113762249680445, + "grad_norm": 14.8125, + "learning_rate": 3.5331912614638974e-07, + "loss": 0.5876, + "step": 17736 + }, + { + "epoch": 1.5114614401363444, + "grad_norm": 13.375, + "learning_rate": 3.5320206078227587e-07, + "loss": 0.7021, + "step": 17737 + }, + { + "epoch": 1.5115466553046444, + "grad_norm": 22.875, + "learning_rate": 3.5308501162420293e-07, + "loss": 0.6281, + "step": 17738 + }, + { + "epoch": 1.5116318704729443, + "grad_norm": 41.25, + "learning_rate": 3.529679786742854e-07, + "loss": 1.3791, + "step": 17739 + }, + { + "epoch": 1.5117170856412443, + "grad_norm": 16.625, + "learning_rate": 3.5285096193463847e-07, + "loss": 0.7039, + "step": 17740 + }, + { + "epoch": 1.5118023008095443, + "grad_norm": 15.375, + "learning_rate": 3.527339614073769e-07, + "loss": 0.6886, + "step": 17741 + }, + { + "epoch": 1.5118875159778442, + "grad_norm": 15.5, + "learning_rate": 3.5261697709461515e-07, + "loss": 0.6373, + "step": 17742 + }, + { + "epoch": 1.5119727311461442, + "grad_norm": 16.0, + "learning_rate": 3.525000089984672e-07, + "loss": 0.4837, + "step": 17743 + }, + { + "epoch": 1.5120579463144441, + "grad_norm": 12.1875, + "learning_rate": 3.52383057121046e-07, + "loss": 0.4429, + "step": 17744 + }, + { + "epoch": 1.512143161482744, + "grad_norm": 15.6875, + "learning_rate": 3.5226612146446573e-07, + "loss": 0.6118, + "step": 17745 + }, + { + "epoch": 1.512228376651044, + "grad_norm": 14.0, + "learning_rate": 3.5214920203083936e-07, + "loss": 0.3669, + "step": 17746 + }, + { + "epoch": 1.512313591819344, + "grad_norm": 16.0, + "learning_rate": 3.520322988222802e-07, + "loss": 0.6445, + "step": 17747 + }, + { + "epoch": 1.512398806987644, + "grad_norm": 14.875, + "learning_rate": 3.5191541184089987e-07, + "loss": 0.4992, + "step": 17748 + }, + { + "epoch": 1.512484022155944, + "grad_norm": 13.625, + "learning_rate": 3.5179854108881126e-07, + "loss": 0.2678, + "step": 17749 + }, + { + "epoch": 1.5125692373242439, + "grad_norm": 21.625, + "learning_rate": 3.516816865681265e-07, + "loss": 0.7409, + "step": 17750 + }, + { + "epoch": 1.5126544524925438, + "grad_norm": 17.0, + "learning_rate": 3.515648482809571e-07, + "loss": 0.4859, + "step": 17751 + }, + { + "epoch": 1.5127396676608438, + "grad_norm": 18.125, + "learning_rate": 3.5144802622941387e-07, + "loss": 0.6261, + "step": 17752 + }, + { + "epoch": 1.5128248828291437, + "grad_norm": 15.625, + "learning_rate": 3.5133122041560846e-07, + "loss": 0.5144, + "step": 17753 + }, + { + "epoch": 1.5129100979974437, + "grad_norm": 13.875, + "learning_rate": 3.5121443084165154e-07, + "loss": 0.4632, + "step": 17754 + }, + { + "epoch": 1.5129953131657436, + "grad_norm": 13.3125, + "learning_rate": 3.510976575096539e-07, + "loss": 0.5195, + "step": 17755 + }, + { + "epoch": 1.5130805283340436, + "grad_norm": 16.25, + "learning_rate": 3.509809004217253e-07, + "loss": 0.7954, + "step": 17756 + }, + { + "epoch": 1.5131657435023436, + "grad_norm": 20.375, + "learning_rate": 3.5086415957997585e-07, + "loss": 1.0288, + "step": 17757 + }, + { + "epoch": 1.5132509586706435, + "grad_norm": 18.0, + "learning_rate": 3.5074743498651545e-07, + "loss": 0.6139, + "step": 17758 + }, + { + "epoch": 1.5133361738389435, + "grad_norm": 14.6875, + "learning_rate": 3.506307266434529e-07, + "loss": 0.7012, + "step": 17759 + }, + { + "epoch": 1.5134213890072434, + "grad_norm": 12.125, + "learning_rate": 3.505140345528979e-07, + "loss": 0.4396, + "step": 17760 + }, + { + "epoch": 1.5135066041755434, + "grad_norm": 16.375, + "learning_rate": 3.5039735871695835e-07, + "loss": 0.5636, + "step": 17761 + }, + { + "epoch": 1.5135918193438433, + "grad_norm": 18.5, + "learning_rate": 3.502806991377432e-07, + "loss": 0.8029, + "step": 17762 + }, + { + "epoch": 1.5136770345121433, + "grad_norm": 20.5, + "learning_rate": 3.501640558173604e-07, + "loss": 0.9982, + "step": 17763 + }, + { + "epoch": 1.5137622496804433, + "grad_norm": 13.3125, + "learning_rate": 3.500474287579185e-07, + "loss": 0.3789, + "step": 17764 + }, + { + "epoch": 1.5138474648487432, + "grad_norm": 21.0, + "learning_rate": 3.4993081796152407e-07, + "loss": 0.6428, + "step": 17765 + }, + { + "epoch": 1.5139326800170432, + "grad_norm": 15.4375, + "learning_rate": 3.498142234302854e-07, + "loss": 0.5053, + "step": 17766 + }, + { + "epoch": 1.5140178951853431, + "grad_norm": 18.25, + "learning_rate": 3.496976451663084e-07, + "loss": 0.7665, + "step": 17767 + }, + { + "epoch": 1.514103110353643, + "grad_norm": 17.75, + "learning_rate": 3.4958108317170056e-07, + "loss": 0.6266, + "step": 17768 + }, + { + "epoch": 1.514188325521943, + "grad_norm": 14.5, + "learning_rate": 3.494645374485678e-07, + "loss": 0.7095, + "step": 17769 + }, + { + "epoch": 1.514273540690243, + "grad_norm": 15.9375, + "learning_rate": 3.4934800799901635e-07, + "loss": 0.6648, + "step": 17770 + }, + { + "epoch": 1.514358755858543, + "grad_norm": 23.125, + "learning_rate": 3.49231494825152e-07, + "loss": 0.7907, + "step": 17771 + }, + { + "epoch": 1.514443971026843, + "grad_norm": 14.625, + "learning_rate": 3.4911499792908074e-07, + "loss": 0.7316, + "step": 17772 + }, + { + "epoch": 1.5145291861951429, + "grad_norm": 12.3125, + "learning_rate": 3.489985173129071e-07, + "loss": 0.3217, + "step": 17773 + }, + { + "epoch": 1.5146144013634428, + "grad_norm": 22.0, + "learning_rate": 3.4888205297873653e-07, + "loss": 0.8005, + "step": 17774 + }, + { + "epoch": 1.5146996165317428, + "grad_norm": 17.5, + "learning_rate": 3.4876560492867306e-07, + "loss": 0.602, + "step": 17775 + }, + { + "epoch": 1.5147848317000427, + "grad_norm": 18.125, + "learning_rate": 3.4864917316482145e-07, + "loss": 0.7116, + "step": 17776 + }, + { + "epoch": 1.5148700468683427, + "grad_norm": 15.5, + "learning_rate": 3.4853275768928593e-07, + "loss": 0.4481, + "step": 17777 + }, + { + "epoch": 1.5149552620366427, + "grad_norm": 18.25, + "learning_rate": 3.484163585041696e-07, + "loss": 0.7989, + "step": 17778 + }, + { + "epoch": 1.5150404772049426, + "grad_norm": 12.875, + "learning_rate": 3.482999756115764e-07, + "loss": 0.2885, + "step": 17779 + }, + { + "epoch": 1.5151256923732426, + "grad_norm": 33.5, + "learning_rate": 3.481836090136093e-07, + "loss": 1.0799, + "step": 17780 + }, + { + "epoch": 1.5152109075415425, + "grad_norm": 13.875, + "learning_rate": 3.480672587123715e-07, + "loss": 0.3723, + "step": 17781 + }, + { + "epoch": 1.5152961227098425, + "grad_norm": 17.5, + "learning_rate": 3.479509247099655e-07, + "loss": 0.7683, + "step": 17782 + }, + { + "epoch": 1.5153813378781424, + "grad_norm": 15.6875, + "learning_rate": 3.478346070084929e-07, + "loss": 0.5179, + "step": 17783 + }, + { + "epoch": 1.5154665530464424, + "grad_norm": 23.75, + "learning_rate": 3.477183056100561e-07, + "loss": 0.6655, + "step": 17784 + }, + { + "epoch": 1.5155517682147424, + "grad_norm": 12.875, + "learning_rate": 3.4760202051675726e-07, + "loss": 0.4075, + "step": 17785 + }, + { + "epoch": 1.5156369833830423, + "grad_norm": 17.125, + "learning_rate": 3.474857517306969e-07, + "loss": 0.4393, + "step": 17786 + }, + { + "epoch": 1.5157221985513423, + "grad_norm": 16.125, + "learning_rate": 3.473694992539767e-07, + "loss": 0.936, + "step": 17787 + }, + { + "epoch": 1.5158074137196422, + "grad_norm": 25.125, + "learning_rate": 3.4725326308869713e-07, + "loss": 0.8175, + "step": 17788 + }, + { + "epoch": 1.5158926288879422, + "grad_norm": 19.0, + "learning_rate": 3.4713704323695947e-07, + "loss": 0.6298, + "step": 17789 + }, + { + "epoch": 1.5159778440562421, + "grad_norm": 15.4375, + "learning_rate": 3.4702083970086314e-07, + "loss": 0.6322, + "step": 17790 + }, + { + "epoch": 1.516063059224542, + "grad_norm": 20.0, + "learning_rate": 3.4690465248250803e-07, + "loss": 0.7886, + "step": 17791 + }, + { + "epoch": 1.516148274392842, + "grad_norm": 14.5, + "learning_rate": 3.467884815839938e-07, + "loss": 0.3825, + "step": 17792 + }, + { + "epoch": 1.516233489561142, + "grad_norm": 17.25, + "learning_rate": 3.4667232700742016e-07, + "loss": 0.3798, + "step": 17793 + }, + { + "epoch": 1.516318704729442, + "grad_norm": 19.25, + "learning_rate": 3.4655618875488613e-07, + "loss": 0.6367, + "step": 17794 + }, + { + "epoch": 1.516403919897742, + "grad_norm": 11.9375, + "learning_rate": 3.464400668284899e-07, + "loss": 0.4919, + "step": 17795 + }, + { + "epoch": 1.5164891350660419, + "grad_norm": 17.375, + "learning_rate": 3.463239612303304e-07, + "loss": 0.9935, + "step": 17796 + }, + { + "epoch": 1.5165743502343418, + "grad_norm": 11.8125, + "learning_rate": 3.462078719625056e-07, + "loss": 0.5642, + "step": 17797 + }, + { + "epoch": 1.5166595654026418, + "grad_norm": 14.875, + "learning_rate": 3.460917990271137e-07, + "loss": 0.6166, + "step": 17798 + }, + { + "epoch": 1.5167447805709418, + "grad_norm": 12.6875, + "learning_rate": 3.459757424262519e-07, + "loss": 0.3854, + "step": 17799 + }, + { + "epoch": 1.5168299957392417, + "grad_norm": 14.25, + "learning_rate": 3.458597021620172e-07, + "loss": 0.5625, + "step": 17800 + }, + { + "epoch": 1.5169152109075417, + "grad_norm": 17.75, + "learning_rate": 3.4574367823650687e-07, + "loss": 0.5641, + "step": 17801 + }, + { + "epoch": 1.5170004260758416, + "grad_norm": 16.5, + "learning_rate": 3.4562767065181803e-07, + "loss": 0.6121, + "step": 17802 + }, + { + "epoch": 1.5170856412441416, + "grad_norm": 13.875, + "learning_rate": 3.4551167941004603e-07, + "loss": 0.478, + "step": 17803 + }, + { + "epoch": 1.5171708564124415, + "grad_norm": 15.5625, + "learning_rate": 3.453957045132876e-07, + "loss": 0.681, + "step": 17804 + }, + { + "epoch": 1.5172560715807415, + "grad_norm": 20.375, + "learning_rate": 3.4527974596363853e-07, + "loss": 0.9487, + "step": 17805 + }, + { + "epoch": 1.5173412867490415, + "grad_norm": 16.375, + "learning_rate": 3.451638037631945e-07, + "loss": 0.4665, + "step": 17806 + }, + { + "epoch": 1.5174265019173414, + "grad_norm": 14.8125, + "learning_rate": 3.4504787791405036e-07, + "loss": 0.7874, + "step": 17807 + }, + { + "epoch": 1.5175117170856414, + "grad_norm": 14.0625, + "learning_rate": 3.4493196841830087e-07, + "loss": 0.5522, + "step": 17808 + }, + { + "epoch": 1.5175969322539413, + "grad_norm": 19.75, + "learning_rate": 3.4481607527804067e-07, + "loss": 0.5713, + "step": 17809 + }, + { + "epoch": 1.5176821474222413, + "grad_norm": 16.75, + "learning_rate": 3.4470019849536434e-07, + "loss": 0.8379, + "step": 17810 + }, + { + "epoch": 1.5177673625905412, + "grad_norm": 16.375, + "learning_rate": 3.44584338072366e-07, + "loss": 0.7655, + "step": 17811 + }, + { + "epoch": 1.5178525777588412, + "grad_norm": 18.125, + "learning_rate": 3.444684940111388e-07, + "loss": 0.5344, + "step": 17812 + }, + { + "epoch": 1.5179377929271411, + "grad_norm": 19.625, + "learning_rate": 3.4435266631377665e-07, + "loss": 0.6214, + "step": 17813 + }, + { + "epoch": 1.518023008095441, + "grad_norm": 18.0, + "learning_rate": 3.442368549823727e-07, + "loss": 0.5239, + "step": 17814 + }, + { + "epoch": 1.518108223263741, + "grad_norm": 11.875, + "learning_rate": 3.441210600190198e-07, + "loss": 0.2436, + "step": 17815 + }, + { + "epoch": 1.518193438432041, + "grad_norm": 14.375, + "learning_rate": 3.4400528142580976e-07, + "loss": 0.491, + "step": 17816 + }, + { + "epoch": 1.518278653600341, + "grad_norm": 24.375, + "learning_rate": 3.4388951920483527e-07, + "loss": 0.9086, + "step": 17817 + }, + { + "epoch": 1.518363868768641, + "grad_norm": 12.5, + "learning_rate": 3.4377377335818846e-07, + "loss": 0.4767, + "step": 17818 + }, + { + "epoch": 1.518449083936941, + "grad_norm": 12.1875, + "learning_rate": 3.4365804388796125e-07, + "loss": 0.3505, + "step": 17819 + }, + { + "epoch": 1.5185342991052408, + "grad_norm": 17.375, + "learning_rate": 3.4354233079624406e-07, + "loss": 0.7322, + "step": 17820 + }, + { + "epoch": 1.5186195142735408, + "grad_norm": 14.9375, + "learning_rate": 3.4342663408512863e-07, + "loss": 0.6824, + "step": 17821 + }, + { + "epoch": 1.5187047294418408, + "grad_norm": 16.125, + "learning_rate": 3.4331095375670534e-07, + "loss": 0.633, + "step": 17822 + }, + { + "epoch": 1.5187899446101407, + "grad_norm": 16.75, + "learning_rate": 3.431952898130654e-07, + "loss": 0.4536, + "step": 17823 + }, + { + "epoch": 1.5188751597784407, + "grad_norm": 17.25, + "learning_rate": 3.4307964225629835e-07, + "loss": 0.4228, + "step": 17824 + }, + { + "epoch": 1.5189603749467406, + "grad_norm": 18.75, + "learning_rate": 3.429640110884938e-07, + "loss": 0.7009, + "step": 17825 + }, + { + "epoch": 1.5190455901150406, + "grad_norm": 26.0, + "learning_rate": 3.4284839631174167e-07, + "loss": 0.6222, + "step": 17826 + }, + { + "epoch": 1.5191308052833405, + "grad_norm": 11.75, + "learning_rate": 3.4273279792813125e-07, + "loss": 0.4099, + "step": 17827 + }, + { + "epoch": 1.5192160204516405, + "grad_norm": 13.3125, + "learning_rate": 3.42617215939752e-07, + "loss": 0.3199, + "step": 17828 + }, + { + "epoch": 1.5193012356199405, + "grad_norm": 16.75, + "learning_rate": 3.425016503486915e-07, + "loss": 0.5446, + "step": 17829 + }, + { + "epoch": 1.5193864507882404, + "grad_norm": 15.75, + "learning_rate": 3.423861011570388e-07, + "loss": 0.3547, + "step": 17830 + }, + { + "epoch": 1.5194716659565404, + "grad_norm": 18.375, + "learning_rate": 3.422705683668824e-07, + "loss": 0.6353, + "step": 17831 + }, + { + "epoch": 1.5195568811248403, + "grad_norm": 13.8125, + "learning_rate": 3.421550519803095e-07, + "loss": 0.5042, + "step": 17832 + }, + { + "epoch": 1.5196420962931403, + "grad_norm": 16.5, + "learning_rate": 3.420395519994074e-07, + "loss": 0.4955, + "step": 17833 + }, + { + "epoch": 1.5197273114614402, + "grad_norm": 12.4375, + "learning_rate": 3.419240684262637e-07, + "loss": 0.5612, + "step": 17834 + }, + { + "epoch": 1.5198125266297402, + "grad_norm": 18.75, + "learning_rate": 3.418086012629651e-07, + "loss": 0.2756, + "step": 17835 + }, + { + "epoch": 1.5198977417980402, + "grad_norm": 12.5625, + "learning_rate": 3.4169315051159877e-07, + "loss": 0.4937, + "step": 17836 + }, + { + "epoch": 1.5199829569663401, + "grad_norm": 17.125, + "learning_rate": 3.415777161742502e-07, + "loss": 0.7251, + "step": 17837 + }, + { + "epoch": 1.52006817213464, + "grad_norm": 15.75, + "learning_rate": 3.414622982530058e-07, + "loss": 0.8248, + "step": 17838 + }, + { + "epoch": 1.52015338730294, + "grad_norm": 14.875, + "learning_rate": 3.413468967499516e-07, + "loss": 0.6812, + "step": 17839 + }, + { + "epoch": 1.52023860247124, + "grad_norm": 18.125, + "learning_rate": 3.412315116671723e-07, + "loss": 0.4737, + "step": 17840 + }, + { + "epoch": 1.52032381763954, + "grad_norm": 12.25, + "learning_rate": 3.411161430067539e-07, + "loss": 0.3894, + "step": 17841 + }, + { + "epoch": 1.52040903280784, + "grad_norm": 13.625, + "learning_rate": 3.4100079077078034e-07, + "loss": 0.5425, + "step": 17842 + }, + { + "epoch": 1.5204942479761399, + "grad_norm": 48.5, + "learning_rate": 3.408854549613365e-07, + "loss": 0.3435, + "step": 17843 + }, + { + "epoch": 1.5205794631444398, + "grad_norm": 67.0, + "learning_rate": 3.407701355805071e-07, + "loss": 0.6264, + "step": 17844 + }, + { + "epoch": 1.5206646783127398, + "grad_norm": 15.8125, + "learning_rate": 3.4065483263037534e-07, + "loss": 0.4741, + "step": 17845 + }, + { + "epoch": 1.5207498934810397, + "grad_norm": 25.0, + "learning_rate": 3.405395461130251e-07, + "loss": 0.7344, + "step": 17846 + }, + { + "epoch": 1.5208351086493397, + "grad_norm": 17.375, + "learning_rate": 3.404242760305401e-07, + "loss": 0.5876, + "step": 17847 + }, + { + "epoch": 1.5209203238176396, + "grad_norm": 18.625, + "learning_rate": 3.403090223850028e-07, + "loss": 0.3864, + "step": 17848 + }, + { + "epoch": 1.5210055389859396, + "grad_norm": 19.75, + "learning_rate": 3.4019378517849657e-07, + "loss": 0.6581, + "step": 17849 + }, + { + "epoch": 1.5210907541542396, + "grad_norm": 19.75, + "learning_rate": 3.4007856441310315e-07, + "loss": 0.7212, + "step": 17850 + }, + { + "epoch": 1.5211759693225395, + "grad_norm": 24.0, + "learning_rate": 3.399633600909051e-07, + "loss": 0.8558, + "step": 17851 + }, + { + "epoch": 1.5212611844908395, + "grad_norm": 15.5, + "learning_rate": 3.398481722139842e-07, + "loss": 0.5803, + "step": 17852 + }, + { + "epoch": 1.5213463996591394, + "grad_norm": 14.625, + "learning_rate": 3.397330007844225e-07, + "loss": 0.8765, + "step": 17853 + }, + { + "epoch": 1.5214316148274394, + "grad_norm": 30.25, + "learning_rate": 3.3961784580430033e-07, + "loss": 0.7677, + "step": 17854 + }, + { + "epoch": 1.5215168299957393, + "grad_norm": 16.25, + "learning_rate": 3.395027072756997e-07, + "loss": 0.516, + "step": 17855 + }, + { + "epoch": 1.5216020451640393, + "grad_norm": 19.5, + "learning_rate": 3.3938758520070017e-07, + "loss": 0.6145, + "step": 17856 + }, + { + "epoch": 1.5216872603323393, + "grad_norm": 20.0, + "learning_rate": 3.3927247958138275e-07, + "loss": 0.8716, + "step": 17857 + }, + { + "epoch": 1.5217724755006392, + "grad_norm": 15.625, + "learning_rate": 3.3915739041982784e-07, + "loss": 0.7272, + "step": 17858 + }, + { + "epoch": 1.5218576906689392, + "grad_norm": 19.125, + "learning_rate": 3.390423177181143e-07, + "loss": 0.7388, + "step": 17859 + }, + { + "epoch": 1.5219429058372391, + "grad_norm": 13.5625, + "learning_rate": 3.389272614783222e-07, + "loss": 0.3343, + "step": 17860 + }, + { + "epoch": 1.522028121005539, + "grad_norm": 17.625, + "learning_rate": 3.388122217025311e-07, + "loss": 0.5333, + "step": 17861 + }, + { + "epoch": 1.522113336173839, + "grad_norm": 18.0, + "learning_rate": 3.3869719839281887e-07, + "loss": 0.5761, + "step": 17862 + }, + { + "epoch": 1.522198551342139, + "grad_norm": 18.5, + "learning_rate": 3.385821915512652e-07, + "loss": 0.6957, + "step": 17863 + }, + { + "epoch": 1.522283766510439, + "grad_norm": 17.875, + "learning_rate": 3.384672011799474e-07, + "loss": 0.7111, + "step": 17864 + }, + { + "epoch": 1.522368981678739, + "grad_norm": 24.0, + "learning_rate": 3.3835222728094393e-07, + "loss": 0.5566, + "step": 17865 + }, + { + "epoch": 1.5224541968470389, + "grad_norm": 17.375, + "learning_rate": 3.382372698563327e-07, + "loss": 0.6995, + "step": 17866 + }, + { + "epoch": 1.5225394120153388, + "grad_norm": 18.875, + "learning_rate": 3.3812232890819057e-07, + "loss": 0.6477, + "step": 17867 + }, + { + "epoch": 1.5226246271836388, + "grad_norm": 18.875, + "learning_rate": 3.380074044385949e-07, + "loss": 0.7522, + "step": 17868 + }, + { + "epoch": 1.5227098423519387, + "grad_norm": 20.0, + "learning_rate": 3.3789249644962247e-07, + "loss": 0.5775, + "step": 17869 + }, + { + "epoch": 1.5227950575202387, + "grad_norm": 18.625, + "learning_rate": 3.377776049433502e-07, + "loss": 0.8255, + "step": 17870 + }, + { + "epoch": 1.5228802726885386, + "grad_norm": 15.125, + "learning_rate": 3.3766272992185366e-07, + "loss": 0.759, + "step": 17871 + }, + { + "epoch": 1.5229654878568386, + "grad_norm": 12.0, + "learning_rate": 3.375478713872092e-07, + "loss": 0.2591, + "step": 17872 + }, + { + "epoch": 1.5230507030251386, + "grad_norm": 12.8125, + "learning_rate": 3.374330293414921e-07, + "loss": 0.3141, + "step": 17873 + }, + { + "epoch": 1.5231359181934385, + "grad_norm": 30.25, + "learning_rate": 3.3731820378677793e-07, + "loss": 1.1391, + "step": 17874 + }, + { + "epoch": 1.5232211333617385, + "grad_norm": 18.125, + "learning_rate": 3.372033947251414e-07, + "loss": 0.6465, + "step": 17875 + }, + { + "epoch": 1.5233063485300384, + "grad_norm": 16.25, + "learning_rate": 3.370886021586574e-07, + "loss": 0.9742, + "step": 17876 + }, + { + "epoch": 1.5233915636983384, + "grad_norm": 14.9375, + "learning_rate": 3.3697382608940024e-07, + "loss": 0.3066, + "step": 17877 + }, + { + "epoch": 1.5234767788666383, + "grad_norm": 11.6875, + "learning_rate": 3.3685906651944477e-07, + "loss": 0.3006, + "step": 17878 + }, + { + "epoch": 1.5235619940349383, + "grad_norm": 15.1875, + "learning_rate": 3.3674432345086356e-07, + "loss": 0.6079, + "step": 17879 + }, + { + "epoch": 1.5236472092032383, + "grad_norm": 21.125, + "learning_rate": 3.366295968857314e-07, + "loss": 0.7518, + "step": 17880 + }, + { + "epoch": 1.5237324243715382, + "grad_norm": 13.875, + "learning_rate": 3.365148868261203e-07, + "loss": 0.5284, + "step": 17881 + }, + { + "epoch": 1.5238176395398382, + "grad_norm": 17.375, + "learning_rate": 3.3640019327410387e-07, + "loss": 0.7063, + "step": 17882 + }, + { + "epoch": 1.5239028547081381, + "grad_norm": 15.0, + "learning_rate": 3.362855162317551e-07, + "loss": 0.5995, + "step": 17883 + }, + { + "epoch": 1.523988069876438, + "grad_norm": 12.125, + "learning_rate": 3.3617085570114534e-07, + "loss": 0.3809, + "step": 17884 + }, + { + "epoch": 1.524073285044738, + "grad_norm": 17.125, + "learning_rate": 3.3605621168434717e-07, + "loss": 0.5204, + "step": 17885 + }, + { + "epoch": 1.524158500213038, + "grad_norm": 12.75, + "learning_rate": 3.3594158418343235e-07, + "loss": 0.2736, + "step": 17886 + }, + { + "epoch": 1.524243715381338, + "grad_norm": 20.375, + "learning_rate": 3.358269732004726e-07, + "loss": 0.6167, + "step": 17887 + }, + { + "epoch": 1.524328930549638, + "grad_norm": 11.5, + "learning_rate": 3.3571237873753883e-07, + "loss": 0.3267, + "step": 17888 + }, + { + "epoch": 1.5244141457179379, + "grad_norm": 15.1875, + "learning_rate": 3.3559780079670117e-07, + "loss": 0.5807, + "step": 17889 + }, + { + "epoch": 1.5244993608862378, + "grad_norm": 10.875, + "learning_rate": 3.3548323938003084e-07, + "loss": 0.2132, + "step": 17890 + }, + { + "epoch": 1.5245845760545378, + "grad_norm": 18.125, + "learning_rate": 3.3536869448959843e-07, + "loss": 0.4648, + "step": 17891 + }, + { + "epoch": 1.5246697912228377, + "grad_norm": 14.375, + "learning_rate": 3.35254166127473e-07, + "loss": 0.471, + "step": 17892 + }, + { + "epoch": 1.5247550063911377, + "grad_norm": 19.75, + "learning_rate": 3.351396542957247e-07, + "loss": 0.8795, + "step": 17893 + }, + { + "epoch": 1.5248402215594377, + "grad_norm": 12.5, + "learning_rate": 3.350251589964229e-07, + "loss": 0.5426, + "step": 17894 + }, + { + "epoch": 1.5249254367277376, + "grad_norm": 18.5, + "learning_rate": 3.3491068023163683e-07, + "loss": 0.9085, + "step": 17895 + }, + { + "epoch": 1.5250106518960376, + "grad_norm": 18.75, + "learning_rate": 3.347962180034349e-07, + "loss": 0.7864, + "step": 17896 + }, + { + "epoch": 1.5250958670643375, + "grad_norm": 13.5625, + "learning_rate": 3.3468177231388526e-07, + "loss": 0.403, + "step": 17897 + }, + { + "epoch": 1.5251810822326375, + "grad_norm": 19.75, + "learning_rate": 3.3456734316505645e-07, + "loss": 0.5928, + "step": 17898 + }, + { + "epoch": 1.5252662974009374, + "grad_norm": 17.375, + "learning_rate": 3.3445293055901636e-07, + "loss": 0.4978, + "step": 17899 + }, + { + "epoch": 1.5253515125692374, + "grad_norm": 12.8125, + "learning_rate": 3.343385344978327e-07, + "loss": 0.5398, + "step": 17900 + }, + { + "epoch": 1.5254367277375374, + "grad_norm": 18.75, + "learning_rate": 3.342241549835722e-07, + "loss": 0.4468, + "step": 17901 + }, + { + "epoch": 1.5255219429058373, + "grad_norm": 26.625, + "learning_rate": 3.34109792018302e-07, + "loss": 0.6135, + "step": 17902 + }, + { + "epoch": 1.5256071580741373, + "grad_norm": 20.875, + "learning_rate": 3.3399544560408907e-07, + "loss": 0.8641, + "step": 17903 + }, + { + "epoch": 1.5256923732424372, + "grad_norm": 11.375, + "learning_rate": 3.338811157430001e-07, + "loss": 0.3021, + "step": 17904 + }, + { + "epoch": 1.5257775884107372, + "grad_norm": 14.125, + "learning_rate": 3.337668024370999e-07, + "loss": 0.5622, + "step": 17905 + }, + { + "epoch": 1.5258628035790371, + "grad_norm": 12.875, + "learning_rate": 3.33652505688455e-07, + "loss": 0.6561, + "step": 17906 + }, + { + "epoch": 1.525948018747337, + "grad_norm": 16.625, + "learning_rate": 3.3353822549913066e-07, + "loss": 0.4569, + "step": 17907 + }, + { + "epoch": 1.526033233915637, + "grad_norm": 27.125, + "learning_rate": 3.3342396187119266e-07, + "loss": 0.9026, + "step": 17908 + }, + { + "epoch": 1.526118449083937, + "grad_norm": 18.25, + "learning_rate": 3.3330971480670495e-07, + "loss": 0.781, + "step": 17909 + }, + { + "epoch": 1.526203664252237, + "grad_norm": 13.75, + "learning_rate": 3.3319548430773244e-07, + "loss": 0.6328, + "step": 17910 + }, + { + "epoch": 1.526288879420537, + "grad_norm": 15.25, + "learning_rate": 3.3308127037633945e-07, + "loss": 0.7338, + "step": 17911 + }, + { + "epoch": 1.5263740945888369, + "grad_norm": 23.125, + "learning_rate": 3.329670730145904e-07, + "loss": 0.9208, + "step": 17912 + }, + { + "epoch": 1.5264593097571368, + "grad_norm": 15.1875, + "learning_rate": 3.3285289222454846e-07, + "loss": 0.5795, + "step": 17913 + }, + { + "epoch": 1.5265445249254368, + "grad_norm": 19.75, + "learning_rate": 3.327387280082765e-07, + "loss": 0.7229, + "step": 17914 + }, + { + "epoch": 1.5266297400937368, + "grad_norm": 14.5, + "learning_rate": 3.3262458036783837e-07, + "loss": 0.3938, + "step": 17915 + }, + { + "epoch": 1.5267149552620367, + "grad_norm": 18.625, + "learning_rate": 3.3251044930529634e-07, + "loss": 0.9254, + "step": 17916 + }, + { + "epoch": 1.5268001704303367, + "grad_norm": 15.375, + "learning_rate": 3.3239633482271355e-07, + "loss": 0.4848, + "step": 17917 + }, + { + "epoch": 1.5268853855986366, + "grad_norm": 18.375, + "learning_rate": 3.3228223692215134e-07, + "loss": 0.8506, + "step": 17918 + }, + { + "epoch": 1.5269706007669366, + "grad_norm": 12.625, + "learning_rate": 3.321681556056719e-07, + "loss": 0.3717, + "step": 17919 + }, + { + "epoch": 1.5270558159352365, + "grad_norm": 17.375, + "learning_rate": 3.3205409087533727e-07, + "loss": 0.4057, + "step": 17920 + }, + { + "epoch": 1.5271410311035365, + "grad_norm": 17.5, + "learning_rate": 3.319400427332083e-07, + "loss": 0.7842, + "step": 17921 + }, + { + "epoch": 1.5272262462718365, + "grad_norm": 18.875, + "learning_rate": 3.3182601118134553e-07, + "loss": 0.5639, + "step": 17922 + }, + { + "epoch": 1.5273114614401364, + "grad_norm": 17.375, + "learning_rate": 3.3171199622181e-07, + "loss": 0.782, + "step": 17923 + }, + { + "epoch": 1.5273966766084364, + "grad_norm": 32.75, + "learning_rate": 3.315979978566622e-07, + "loss": 1.1505, + "step": 17924 + }, + { + "epoch": 1.5274818917767363, + "grad_norm": 16.0, + "learning_rate": 3.314840160879623e-07, + "loss": 0.5953, + "step": 17925 + }, + { + "epoch": 1.5275671069450363, + "grad_norm": 15.125, + "learning_rate": 3.3137005091776973e-07, + "loss": 0.5799, + "step": 17926 + }, + { + "epoch": 1.5276523221133362, + "grad_norm": 17.0, + "learning_rate": 3.3125610234814395e-07, + "loss": 0.514, + "step": 17927 + }, + { + "epoch": 1.5277375372816362, + "grad_norm": 11.625, + "learning_rate": 3.311421703811446e-07, + "loss": 0.2853, + "step": 17928 + }, + { + "epoch": 1.5278227524499362, + "grad_norm": 13.9375, + "learning_rate": 3.3102825501883e-07, + "loss": 0.6798, + "step": 17929 + }, + { + "epoch": 1.527907967618236, + "grad_norm": 15.75, + "learning_rate": 3.309143562632591e-07, + "loss": 0.5874, + "step": 17930 + }, + { + "epoch": 1.527993182786536, + "grad_norm": 12.9375, + "learning_rate": 3.3080047411648984e-07, + "loss": 0.4893, + "step": 17931 + }, + { + "epoch": 1.528078397954836, + "grad_norm": 15.1875, + "learning_rate": 3.3068660858058013e-07, + "loss": 0.6891, + "step": 17932 + }, + { + "epoch": 1.528163613123136, + "grad_norm": 19.0, + "learning_rate": 3.305727596575881e-07, + "loss": 0.8052, + "step": 17933 + }, + { + "epoch": 1.528248828291436, + "grad_norm": 12.125, + "learning_rate": 3.3045892734957106e-07, + "loss": 0.4717, + "step": 17934 + }, + { + "epoch": 1.528334043459736, + "grad_norm": 14.1875, + "learning_rate": 3.303451116585855e-07, + "loss": 0.517, + "step": 17935 + }, + { + "epoch": 1.5284192586280358, + "grad_norm": 14.25, + "learning_rate": 3.302313125866885e-07, + "loss": 0.473, + "step": 17936 + }, + { + "epoch": 1.5285044737963358, + "grad_norm": 19.75, + "learning_rate": 3.30117530135937e-07, + "loss": 0.5991, + "step": 17937 + }, + { + "epoch": 1.5285896889646358, + "grad_norm": 25.75, + "learning_rate": 3.3000376430838687e-07, + "loss": 0.6874, + "step": 17938 + }, + { + "epoch": 1.5286749041329357, + "grad_norm": 17.375, + "learning_rate": 3.298900151060934e-07, + "loss": 0.588, + "step": 17939 + }, + { + "epoch": 1.5287601193012357, + "grad_norm": 14.25, + "learning_rate": 3.297762825311125e-07, + "loss": 0.5701, + "step": 17940 + }, + { + "epoch": 1.5288453344695356, + "grad_norm": 16.125, + "learning_rate": 3.2966256658549966e-07, + "loss": 0.4914, + "step": 17941 + }, + { + "epoch": 1.5289305496378356, + "grad_norm": 10.75, + "learning_rate": 3.2954886727131e-07, + "loss": 0.2853, + "step": 17942 + }, + { + "epoch": 1.5290157648061355, + "grad_norm": 15.4375, + "learning_rate": 3.2943518459059755e-07, + "loss": 0.606, + "step": 17943 + }, + { + "epoch": 1.5291009799744355, + "grad_norm": 19.5, + "learning_rate": 3.293215185454171e-07, + "loss": 0.7778, + "step": 17944 + }, + { + "epoch": 1.5291861951427355, + "grad_norm": 18.25, + "learning_rate": 3.292078691378228e-07, + "loss": 0.7281, + "step": 17945 + }, + { + "epoch": 1.5292714103110354, + "grad_norm": 15.4375, + "learning_rate": 3.2909423636986806e-07, + "loss": 0.963, + "step": 17946 + }, + { + "epoch": 1.5293566254793354, + "grad_norm": 13.0625, + "learning_rate": 3.289806202436067e-07, + "loss": 0.3774, + "step": 17947 + }, + { + "epoch": 1.5294418406476353, + "grad_norm": 14.875, + "learning_rate": 3.2886702076109155e-07, + "loss": 0.5016, + "step": 17948 + }, + { + "epoch": 1.5295270558159353, + "grad_norm": 16.75, + "learning_rate": 3.2875343792437537e-07, + "loss": 0.6474, + "step": 17949 + }, + { + "epoch": 1.5296122709842352, + "grad_norm": 17.75, + "learning_rate": 3.28639871735511e-07, + "loss": 0.9139, + "step": 17950 + }, + { + "epoch": 1.5296974861525352, + "grad_norm": 24.5, + "learning_rate": 3.2852632219655123e-07, + "loss": 0.6568, + "step": 17951 + }, + { + "epoch": 1.5297827013208352, + "grad_norm": 13.5, + "learning_rate": 3.284127893095468e-07, + "loss": 0.315, + "step": 17952 + }, + { + "epoch": 1.5298679164891351, + "grad_norm": 12.5, + "learning_rate": 3.282992730765505e-07, + "loss": 0.2429, + "step": 17953 + }, + { + "epoch": 1.529953131657435, + "grad_norm": 22.0, + "learning_rate": 3.2818577349961274e-07, + "loss": 0.8803, + "step": 17954 + }, + { + "epoch": 1.530038346825735, + "grad_norm": 22.25, + "learning_rate": 3.280722905807854e-07, + "loss": 0.6947, + "step": 17955 + }, + { + "epoch": 1.530123561994035, + "grad_norm": 18.625, + "learning_rate": 3.2795882432211853e-07, + "loss": 0.8352, + "step": 17956 + }, + { + "epoch": 1.530208777162335, + "grad_norm": 24.0, + "learning_rate": 3.2784537472566283e-07, + "loss": 0.7527, + "step": 17957 + }, + { + "epoch": 1.530293992330635, + "grad_norm": 14.1875, + "learning_rate": 3.277319417934685e-07, + "loss": 0.6465, + "step": 17958 + }, + { + "epoch": 1.5303792074989349, + "grad_norm": 17.625, + "learning_rate": 3.2761852552758587e-07, + "loss": 0.9668, + "step": 17959 + }, + { + "epoch": 1.5304644226672348, + "grad_norm": 13.3125, + "learning_rate": 3.275051259300635e-07, + "loss": 0.4368, + "step": 17960 + }, + { + "epoch": 1.5305496378355348, + "grad_norm": 14.125, + "learning_rate": 3.273917430029515e-07, + "loss": 0.7976, + "step": 17961 + }, + { + "epoch": 1.5306348530038347, + "grad_norm": 19.0, + "learning_rate": 3.2727837674829817e-07, + "loss": 0.8612, + "step": 17962 + }, + { + "epoch": 1.5307200681721347, + "grad_norm": 16.0, + "learning_rate": 3.2716502716815245e-07, + "loss": 0.6077, + "step": 17963 + }, + { + "epoch": 1.5308052833404346, + "grad_norm": 36.0, + "learning_rate": 3.2705169426456297e-07, + "loss": 0.6216, + "step": 17964 + }, + { + "epoch": 1.5308904985087346, + "grad_norm": 14.4375, + "learning_rate": 3.269383780395773e-07, + "loss": 0.481, + "step": 17965 + }, + { + "epoch": 1.5309757136770346, + "grad_norm": 19.375, + "learning_rate": 3.268250784952431e-07, + "loss": 0.4499, + "step": 17966 + }, + { + "epoch": 1.5310609288453345, + "grad_norm": 15.1875, + "learning_rate": 3.267117956336081e-07, + "loss": 0.3486, + "step": 17967 + }, + { + "epoch": 1.5311461440136345, + "grad_norm": 13.9375, + "learning_rate": 3.265985294567199e-07, + "loss": 0.5763, + "step": 17968 + }, + { + "epoch": 1.5312313591819344, + "grad_norm": 13.5625, + "learning_rate": 3.264852799666247e-07, + "loss": 0.5514, + "step": 17969 + }, + { + "epoch": 1.5313165743502344, + "grad_norm": 17.25, + "learning_rate": 3.263720471653688e-07, + "loss": 0.7823, + "step": 17970 + }, + { + "epoch": 1.5314017895185343, + "grad_norm": 15.75, + "learning_rate": 3.2625883105499886e-07, + "loss": 0.4922, + "step": 17971 + }, + { + "epoch": 1.5314870046868343, + "grad_norm": 16.375, + "learning_rate": 3.261456316375611e-07, + "loss": 0.5687, + "step": 17972 + }, + { + "epoch": 1.5315722198551343, + "grad_norm": 17.5, + "learning_rate": 3.2603244891510025e-07, + "loss": 0.8644, + "step": 17973 + }, + { + "epoch": 1.5316574350234342, + "grad_norm": 10.8125, + "learning_rate": 3.2591928288966235e-07, + "loss": 0.2395, + "step": 17974 + }, + { + "epoch": 1.5317426501917342, + "grad_norm": 14.0625, + "learning_rate": 3.2580613356329214e-07, + "loss": 0.6047, + "step": 17975 + }, + { + "epoch": 1.5318278653600341, + "grad_norm": 24.0, + "learning_rate": 3.256930009380349e-07, + "loss": 0.8693, + "step": 17976 + }, + { + "epoch": 1.531913080528334, + "grad_norm": 21.375, + "learning_rate": 3.255798850159346e-07, + "loss": 0.8947, + "step": 17977 + }, + { + "epoch": 1.531998295696634, + "grad_norm": 21.5, + "learning_rate": 3.2546678579903487e-07, + "loss": 0.919, + "step": 17978 + }, + { + "epoch": 1.532083510864934, + "grad_norm": 14.3125, + "learning_rate": 3.2535370328938015e-07, + "loss": 0.7489, + "step": 17979 + }, + { + "epoch": 1.532168726033234, + "grad_norm": 14.0, + "learning_rate": 3.2524063748901367e-07, + "loss": 0.4281, + "step": 17980 + }, + { + "epoch": 1.532253941201534, + "grad_norm": 17.375, + "learning_rate": 3.251275883999792e-07, + "loss": 0.8333, + "step": 17981 + }, + { + "epoch": 1.5323391563698339, + "grad_norm": 14.875, + "learning_rate": 3.25014556024319e-07, + "loss": 0.4825, + "step": 17982 + }, + { + "epoch": 1.5324243715381338, + "grad_norm": 16.75, + "learning_rate": 3.249015403640758e-07, + "loss": 0.6794, + "step": 17983 + }, + { + "epoch": 1.5325095867064338, + "grad_norm": 20.25, + "learning_rate": 3.247885414212924e-07, + "loss": 0.5464, + "step": 17984 + }, + { + "epoch": 1.5325948018747337, + "grad_norm": 14.1875, + "learning_rate": 3.2467555919801037e-07, + "loss": 0.5698, + "step": 17985 + }, + { + "epoch": 1.5326800170430337, + "grad_norm": 12.5, + "learning_rate": 3.245625936962712e-07, + "loss": 0.267, + "step": 17986 + }, + { + "epoch": 1.5327652322113337, + "grad_norm": 16.125, + "learning_rate": 3.2444964491811654e-07, + "loss": 0.5837, + "step": 17987 + }, + { + "epoch": 1.5328504473796336, + "grad_norm": 12.9375, + "learning_rate": 3.243367128655875e-07, + "loss": 0.3568, + "step": 17988 + }, + { + "epoch": 1.5329356625479336, + "grad_norm": 16.25, + "learning_rate": 3.2422379754072533e-07, + "loss": 0.6231, + "step": 17989 + }, + { + "epoch": 1.5330208777162335, + "grad_norm": 12.75, + "learning_rate": 3.2411089894556965e-07, + "loss": 0.4442, + "step": 17990 + }, + { + "epoch": 1.5331060928845335, + "grad_norm": 18.25, + "learning_rate": 3.239980170821611e-07, + "loss": 0.9889, + "step": 17991 + }, + { + "epoch": 1.5331913080528334, + "grad_norm": 13.125, + "learning_rate": 3.238851519525395e-07, + "loss": 0.7207, + "step": 17992 + }, + { + "epoch": 1.5332765232211334, + "grad_norm": 11.6875, + "learning_rate": 3.237723035587448e-07, + "loss": 0.4521, + "step": 17993 + }, + { + "epoch": 1.5333617383894333, + "grad_norm": 16.375, + "learning_rate": 3.2365947190281606e-07, + "loss": 0.4908, + "step": 17994 + }, + { + "epoch": 1.5334469535577333, + "grad_norm": 16.375, + "learning_rate": 3.235466569867918e-07, + "loss": 0.7526, + "step": 17995 + }, + { + "epoch": 1.5335321687260333, + "grad_norm": 13.125, + "learning_rate": 3.23433858812711e-07, + "loss": 0.5891, + "step": 17996 + }, + { + "epoch": 1.5336173838943332, + "grad_norm": 14.4375, + "learning_rate": 3.233210773826122e-07, + "loss": 0.7766, + "step": 17997 + }, + { + "epoch": 1.5337025990626332, + "grad_norm": 17.5, + "learning_rate": 3.232083126985336e-07, + "loss": 0.7337, + "step": 17998 + }, + { + "epoch": 1.5337878142309331, + "grad_norm": 17.875, + "learning_rate": 3.2309556476251254e-07, + "loss": 0.6895, + "step": 17999 + }, + { + "epoch": 1.533873029399233, + "grad_norm": 11.375, + "learning_rate": 3.2298283357658666e-07, + "loss": 0.3635, + "step": 18000 + }, + { + "epoch": 1.533958244567533, + "grad_norm": 17.375, + "learning_rate": 3.2287011914279354e-07, + "loss": 0.486, + "step": 18001 + }, + { + "epoch": 1.534043459735833, + "grad_norm": 42.0, + "learning_rate": 3.2275742146316955e-07, + "loss": 0.7138, + "step": 18002 + }, + { + "epoch": 1.534128674904133, + "grad_norm": 17.625, + "learning_rate": 3.226447405397512e-07, + "loss": 0.9913, + "step": 18003 + }, + { + "epoch": 1.534213890072433, + "grad_norm": 14.4375, + "learning_rate": 3.2253207637457473e-07, + "loss": 0.6176, + "step": 18004 + }, + { + "epoch": 1.5342991052407329, + "grad_norm": 20.25, + "learning_rate": 3.2241942896967637e-07, + "loss": 0.7447, + "step": 18005 + }, + { + "epoch": 1.5343843204090328, + "grad_norm": 20.0, + "learning_rate": 3.2230679832709205e-07, + "loss": 0.7433, + "step": 18006 + }, + { + "epoch": 1.5344695355773328, + "grad_norm": 16.875, + "learning_rate": 3.2219418444885646e-07, + "loss": 0.4164, + "step": 18007 + }, + { + "epoch": 1.5345547507456327, + "grad_norm": 15.5625, + "learning_rate": 3.220815873370048e-07, + "loss": 0.728, + "step": 18008 + }, + { + "epoch": 1.5346399659139327, + "grad_norm": 13.25, + "learning_rate": 3.21969006993572e-07, + "loss": 0.327, + "step": 18009 + }, + { + "epoch": 1.5347251810822327, + "grad_norm": 18.5, + "learning_rate": 3.2185644342059285e-07, + "loss": 0.8874, + "step": 18010 + }, + { + "epoch": 1.5348103962505326, + "grad_norm": 18.0, + "learning_rate": 3.21743896620101e-07, + "loss": 1.0342, + "step": 18011 + }, + { + "epoch": 1.5348956114188326, + "grad_norm": 21.0, + "learning_rate": 3.2163136659413007e-07, + "loss": 0.7385, + "step": 18012 + }, + { + "epoch": 1.5349808265871325, + "grad_norm": 41.0, + "learning_rate": 3.215188533447139e-07, + "loss": 0.3905, + "step": 18013 + }, + { + "epoch": 1.5350660417554325, + "grad_norm": 15.875, + "learning_rate": 3.214063568738861e-07, + "loss": 1.0007, + "step": 18014 + }, + { + "epoch": 1.5351512569237324, + "grad_norm": 15.75, + "learning_rate": 3.2129387718367875e-07, + "loss": 0.5405, + "step": 18015 + }, + { + "epoch": 1.5352364720920324, + "grad_norm": 17.25, + "learning_rate": 3.211814142761249e-07, + "loss": 0.461, + "step": 18016 + }, + { + "epoch": 1.5353216872603324, + "grad_norm": 14.8125, + "learning_rate": 3.210689681532571e-07, + "loss": 0.926, + "step": 18017 + }, + { + "epoch": 1.5354069024286323, + "grad_norm": 18.5, + "learning_rate": 3.2095653881710735e-07, + "loss": 0.8262, + "step": 18018 + }, + { + "epoch": 1.5354921175969323, + "grad_norm": 13.0625, + "learning_rate": 3.208441262697072e-07, + "loss": 0.5203, + "step": 18019 + }, + { + "epoch": 1.5355773327652322, + "grad_norm": 12.5, + "learning_rate": 3.2073173051308774e-07, + "loss": 0.5053, + "step": 18020 + }, + { + "epoch": 1.5356625479335322, + "grad_norm": 18.875, + "learning_rate": 3.206193515492803e-07, + "loss": 0.5634, + "step": 18021 + }, + { + "epoch": 1.5357477631018321, + "grad_norm": 17.75, + "learning_rate": 3.2050698938031577e-07, + "loss": 0.7365, + "step": 18022 + }, + { + "epoch": 1.535832978270132, + "grad_norm": 15.625, + "learning_rate": 3.2039464400822494e-07, + "loss": 0.7466, + "step": 18023 + }, + { + "epoch": 1.535918193438432, + "grad_norm": 23.625, + "learning_rate": 3.202823154350372e-07, + "loss": 0.678, + "step": 18024 + }, + { + "epoch": 1.536003408606732, + "grad_norm": 31.875, + "learning_rate": 3.201700036627832e-07, + "loss": 0.9351, + "step": 18025 + }, + { + "epoch": 1.536088623775032, + "grad_norm": 13.75, + "learning_rate": 3.2005770869349236e-07, + "loss": 0.3708, + "step": 18026 + }, + { + "epoch": 1.536173838943332, + "grad_norm": 16.875, + "learning_rate": 3.199454305291937e-07, + "loss": 0.7273, + "step": 18027 + }, + { + "epoch": 1.5362590541116319, + "grad_norm": 14.1875, + "learning_rate": 3.1983316917191654e-07, + "loss": 0.5334, + "step": 18028 + }, + { + "epoch": 1.5363442692799318, + "grad_norm": 11.625, + "learning_rate": 3.19720924623689e-07, + "loss": 0.3952, + "step": 18029 + }, + { + "epoch": 1.5364294844482318, + "grad_norm": 15.3125, + "learning_rate": 3.1960869688654e-07, + "loss": 0.6862, + "step": 18030 + }, + { + "epoch": 1.5365146996165318, + "grad_norm": 14.25, + "learning_rate": 3.194964859624977e-07, + "loss": 0.3309, + "step": 18031 + }, + { + "epoch": 1.5365999147848317, + "grad_norm": 18.25, + "learning_rate": 3.1938429185358913e-07, + "loss": 1.137, + "step": 18032 + }, + { + "epoch": 1.5366851299531317, + "grad_norm": 13.0, + "learning_rate": 3.192721145618424e-07, + "loss": 0.3966, + "step": 18033 + }, + { + "epoch": 1.5367703451214316, + "grad_norm": 16.0, + "learning_rate": 3.191599540892848e-07, + "loss": 0.5956, + "step": 18034 + }, + { + "epoch": 1.5368555602897316, + "grad_norm": 17.375, + "learning_rate": 3.1904781043794244e-07, + "loss": 0.7526, + "step": 18035 + }, + { + "epoch": 1.5369407754580315, + "grad_norm": 17.625, + "learning_rate": 3.189356836098426e-07, + "loss": 0.5607, + "step": 18036 + }, + { + "epoch": 1.5370259906263315, + "grad_norm": 15.3125, + "learning_rate": 3.1882357360701095e-07, + "loss": 0.6523, + "step": 18037 + }, + { + "epoch": 1.5371112057946315, + "grad_norm": 12.5625, + "learning_rate": 3.1871148043147366e-07, + "loss": 0.2046, + "step": 18038 + }, + { + "epoch": 1.5371964209629314, + "grad_norm": 17.25, + "learning_rate": 3.1859940408525637e-07, + "loss": 0.5863, + "step": 18039 + }, + { + "epoch": 1.5372816361312314, + "grad_norm": 19.375, + "learning_rate": 3.1848734457038483e-07, + "loss": 0.4456, + "step": 18040 + }, + { + "epoch": 1.5373668512995313, + "grad_norm": 19.5, + "learning_rate": 3.183753018888833e-07, + "loss": 0.4274, + "step": 18041 + }, + { + "epoch": 1.5374520664678313, + "grad_norm": 13.1875, + "learning_rate": 3.182632760427773e-07, + "loss": 0.3521, + "step": 18042 + }, + { + "epoch": 1.5375372816361312, + "grad_norm": 16.625, + "learning_rate": 3.181512670340904e-07, + "loss": 0.7194, + "step": 18043 + }, + { + "epoch": 1.5376224968044312, + "grad_norm": 14.125, + "learning_rate": 3.1803927486484755e-07, + "loss": 0.5185, + "step": 18044 + }, + { + "epoch": 1.5377077119727312, + "grad_norm": 13.875, + "learning_rate": 3.1792729953707176e-07, + "loss": 0.6, + "step": 18045 + }, + { + "epoch": 1.537792927141031, + "grad_norm": 17.25, + "learning_rate": 3.1781534105278705e-07, + "loss": 0.6751, + "step": 18046 + }, + { + "epoch": 1.537878142309331, + "grad_norm": 18.375, + "learning_rate": 3.1770339941401646e-07, + "loss": 0.7212, + "step": 18047 + }, + { + "epoch": 1.537963357477631, + "grad_norm": 15.375, + "learning_rate": 3.1759147462278333e-07, + "loss": 0.6568, + "step": 18048 + }, + { + "epoch": 1.538048572645931, + "grad_norm": 11.5, + "learning_rate": 3.174795666811094e-07, + "loss": 0.4066, + "step": 18049 + }, + { + "epoch": 1.538133787814231, + "grad_norm": 17.625, + "learning_rate": 3.173676755910178e-07, + "loss": 0.5937, + "step": 18050 + }, + { + "epoch": 1.538219002982531, + "grad_norm": 10.5625, + "learning_rate": 3.1725580135452986e-07, + "loss": 0.285, + "step": 18051 + }, + { + "epoch": 1.5383042181508308, + "grad_norm": 16.25, + "learning_rate": 3.1714394397366745e-07, + "loss": 0.3227, + "step": 18052 + }, + { + "epoch": 1.5383894333191308, + "grad_norm": 18.0, + "learning_rate": 3.170321034504525e-07, + "loss": 0.6471, + "step": 18053 + }, + { + "epoch": 1.5384746484874308, + "grad_norm": 18.75, + "learning_rate": 3.1692027978690516e-07, + "loss": 0.7491, + "step": 18054 + }, + { + "epoch": 1.5385598636557307, + "grad_norm": 23.25, + "learning_rate": 3.1680847298504665e-07, + "loss": 0.7991, + "step": 18055 + }, + { + "epoch": 1.5386450788240307, + "grad_norm": 14.4375, + "learning_rate": 3.1669668304689747e-07, + "loss": 0.7255, + "step": 18056 + }, + { + "epoch": 1.5387302939923306, + "grad_norm": 22.875, + "learning_rate": 3.1658490997447807e-07, + "loss": 0.8376, + "step": 18057 + }, + { + "epoch": 1.5388155091606306, + "grad_norm": 16.875, + "learning_rate": 3.1647315376980803e-07, + "loss": 0.6293, + "step": 18058 + }, + { + "epoch": 1.5389007243289305, + "grad_norm": 21.75, + "learning_rate": 3.163614144349064e-07, + "loss": 0.9939, + "step": 18059 + }, + { + "epoch": 1.5389859394972305, + "grad_norm": 17.625, + "learning_rate": 3.1624969197179294e-07, + "loss": 0.5389, + "step": 18060 + }, + { + "epoch": 1.5390711546655305, + "grad_norm": 13.25, + "learning_rate": 3.1613798638248675e-07, + "loss": 0.3476, + "step": 18061 + }, + { + "epoch": 1.5391563698338304, + "grad_norm": 12.0625, + "learning_rate": 3.1602629766900596e-07, + "loss": 0.2934, + "step": 18062 + }, + { + "epoch": 1.5392415850021304, + "grad_norm": 17.75, + "learning_rate": 3.159146258333691e-07, + "loss": 0.8337, + "step": 18063 + }, + { + "epoch": 1.5393268001704303, + "grad_norm": 14.9375, + "learning_rate": 3.158029708775942e-07, + "loss": 0.6701, + "step": 18064 + }, + { + "epoch": 1.5394120153387303, + "grad_norm": 16.125, + "learning_rate": 3.156913328036995e-07, + "loss": 0.3361, + "step": 18065 + }, + { + "epoch": 1.5394972305070302, + "grad_norm": 13.5625, + "learning_rate": 3.155797116137019e-07, + "loss": 0.5023, + "step": 18066 + }, + { + "epoch": 1.5395824456753302, + "grad_norm": 13.0625, + "learning_rate": 3.1546810730961814e-07, + "loss": 0.3776, + "step": 18067 + }, + { + "epoch": 1.5396676608436302, + "grad_norm": 11.625, + "learning_rate": 3.1535651989346535e-07, + "loss": 0.3814, + "step": 18068 + }, + { + "epoch": 1.5397528760119301, + "grad_norm": 17.25, + "learning_rate": 3.152449493672603e-07, + "loss": 0.8499, + "step": 18069 + }, + { + "epoch": 1.53983809118023, + "grad_norm": 22.125, + "learning_rate": 3.151333957330192e-07, + "loss": 0.9117, + "step": 18070 + }, + { + "epoch": 1.53992330634853, + "grad_norm": 13.5, + "learning_rate": 3.1502185899275744e-07, + "loss": 0.6264, + "step": 18071 + }, + { + "epoch": 1.54000852151683, + "grad_norm": 17.125, + "learning_rate": 3.1491033914849077e-07, + "loss": 0.7371, + "step": 18072 + }, + { + "epoch": 1.54009373668513, + "grad_norm": 14.9375, + "learning_rate": 3.1479883620223473e-07, + "loss": 0.6366, + "step": 18073 + }, + { + "epoch": 1.54017895185343, + "grad_norm": 15.375, + "learning_rate": 3.146873501560044e-07, + "loss": 0.5138, + "step": 18074 + }, + { + "epoch": 1.5402641670217299, + "grad_norm": 15.3125, + "learning_rate": 3.145758810118142e-07, + "loss": 0.8285, + "step": 18075 + }, + { + "epoch": 1.5403493821900298, + "grad_norm": 13.375, + "learning_rate": 3.1446442877167803e-07, + "loss": 0.3618, + "step": 18076 + }, + { + "epoch": 1.5404345973583298, + "grad_norm": 11.5625, + "learning_rate": 3.143529934376105e-07, + "loss": 0.4399, + "step": 18077 + }, + { + "epoch": 1.5405198125266297, + "grad_norm": 16.875, + "learning_rate": 3.1424157501162567e-07, + "loss": 0.7838, + "step": 18078 + }, + { + "epoch": 1.5406050276949297, + "grad_norm": 14.5, + "learning_rate": 3.1413017349573595e-07, + "loss": 0.504, + "step": 18079 + }, + { + "epoch": 1.5406902428632296, + "grad_norm": 16.25, + "learning_rate": 3.1401878889195534e-07, + "loss": 0.7607, + "step": 18080 + }, + { + "epoch": 1.5407754580315296, + "grad_norm": 17.5, + "learning_rate": 3.1390742120229635e-07, + "loss": 0.638, + "step": 18081 + }, + { + "epoch": 1.5408606731998296, + "grad_norm": 21.25, + "learning_rate": 3.137960704287718e-07, + "loss": 0.7269, + "step": 18082 + }, + { + "epoch": 1.5409458883681295, + "grad_norm": 13.0, + "learning_rate": 3.1368473657339373e-07, + "loss": 0.5328, + "step": 18083 + }, + { + "epoch": 1.5410311035364295, + "grad_norm": 16.625, + "learning_rate": 3.135734196381737e-07, + "loss": 0.8019, + "step": 18084 + }, + { + "epoch": 1.5411163187047294, + "grad_norm": 14.875, + "learning_rate": 3.134621196251236e-07, + "loss": 0.8248, + "step": 18085 + }, + { + "epoch": 1.5412015338730294, + "grad_norm": 16.375, + "learning_rate": 3.133508365362549e-07, + "loss": 0.6918, + "step": 18086 + }, + { + "epoch": 1.5412867490413293, + "grad_norm": 14.0, + "learning_rate": 3.132395703735787e-07, + "loss": 0.5365, + "step": 18087 + }, + { + "epoch": 1.5413719642096293, + "grad_norm": 11.6875, + "learning_rate": 3.131283211391051e-07, + "loss": 0.3397, + "step": 18088 + }, + { + "epoch": 1.5414571793779293, + "grad_norm": 17.25, + "learning_rate": 3.130170888348448e-07, + "loss": 0.6985, + "step": 18089 + }, + { + "epoch": 1.5415423945462292, + "grad_norm": 13.3125, + "learning_rate": 3.129058734628079e-07, + "loss": 0.5279, + "step": 18090 + }, + { + "epoch": 1.5416276097145292, + "grad_norm": 13.875, + "learning_rate": 3.1279467502500494e-07, + "loss": 0.69, + "step": 18091 + }, + { + "epoch": 1.5417128248828291, + "grad_norm": 13.0, + "learning_rate": 3.1268349352344395e-07, + "loss": 0.5498, + "step": 18092 + }, + { + "epoch": 1.541798040051129, + "grad_norm": 18.875, + "learning_rate": 3.125723289601347e-07, + "loss": 0.7817, + "step": 18093 + }, + { + "epoch": 1.541883255219429, + "grad_norm": 15.4375, + "learning_rate": 3.124611813370862e-07, + "loss": 0.6948, + "step": 18094 + }, + { + "epoch": 1.541968470387729, + "grad_norm": 15.5, + "learning_rate": 3.123500506563072e-07, + "loss": 0.5398, + "step": 18095 + }, + { + "epoch": 1.542053685556029, + "grad_norm": 15.8125, + "learning_rate": 3.1223893691980536e-07, + "loss": 0.4303, + "step": 18096 + }, + { + "epoch": 1.542138900724329, + "grad_norm": 25.5, + "learning_rate": 3.121278401295889e-07, + "loss": 0.498, + "step": 18097 + }, + { + "epoch": 1.5422241158926289, + "grad_norm": 22.375, + "learning_rate": 3.120167602876656e-07, + "loss": 0.8328, + "step": 18098 + }, + { + "epoch": 1.5423093310609288, + "grad_norm": 14.875, + "learning_rate": 3.1190569739604285e-07, + "loss": 0.4627, + "step": 18099 + }, + { + "epoch": 1.5423945462292288, + "grad_norm": 17.25, + "learning_rate": 3.117946514567276e-07, + "loss": 0.9619, + "step": 18100 + }, + { + "epoch": 1.5424797613975287, + "grad_norm": 13.0625, + "learning_rate": 3.1168362247172594e-07, + "loss": 0.4909, + "step": 18101 + }, + { + "epoch": 1.5425649765658287, + "grad_norm": 15.5625, + "learning_rate": 3.1157261044304484e-07, + "loss": 0.8078, + "step": 18102 + }, + { + "epoch": 1.5426501917341287, + "grad_norm": 19.5, + "learning_rate": 3.114616153726903e-07, + "loss": 0.8505, + "step": 18103 + }, + { + "epoch": 1.5427354069024286, + "grad_norm": 15.1875, + "learning_rate": 3.113506372626686e-07, + "loss": 0.6811, + "step": 18104 + }, + { + "epoch": 1.5428206220707286, + "grad_norm": 18.375, + "learning_rate": 3.1123967611498426e-07, + "loss": 0.826, + "step": 18105 + }, + { + "epoch": 1.5429058372390285, + "grad_norm": 19.0, + "learning_rate": 3.111287319316431e-07, + "loss": 0.4378, + "step": 18106 + }, + { + "epoch": 1.5429910524073285, + "grad_norm": 17.625, + "learning_rate": 3.110178047146502e-07, + "loss": 0.706, + "step": 18107 + }, + { + "epoch": 1.5430762675756284, + "grad_norm": 16.375, + "learning_rate": 3.109068944660097e-07, + "loss": 0.6275, + "step": 18108 + }, + { + "epoch": 1.5431614827439284, + "grad_norm": 21.75, + "learning_rate": 3.1079600118772564e-07, + "loss": 1.0648, + "step": 18109 + }, + { + "epoch": 1.5432466979122283, + "grad_norm": 21.25, + "learning_rate": 3.1068512488180225e-07, + "loss": 0.5861, + "step": 18110 + }, + { + "epoch": 1.5433319130805283, + "grad_norm": 16.375, + "learning_rate": 3.105742655502433e-07, + "loss": 0.4717, + "step": 18111 + }, + { + "epoch": 1.5434171282488283, + "grad_norm": 10.5625, + "learning_rate": 3.1046342319505233e-07, + "loss": 0.29, + "step": 18112 + }, + { + "epoch": 1.5435023434171282, + "grad_norm": 36.5, + "learning_rate": 3.1035259781823173e-07, + "loss": 1.2628, + "step": 18113 + }, + { + "epoch": 1.5435875585854282, + "grad_norm": 10.3125, + "learning_rate": 3.1024178942178464e-07, + "loss": 0.346, + "step": 18114 + }, + { + "epoch": 1.5436727737537281, + "grad_norm": 22.5, + "learning_rate": 3.101309980077137e-07, + "loss": 0.7548, + "step": 18115 + }, + { + "epoch": 1.543757988922028, + "grad_norm": 16.0, + "learning_rate": 3.100202235780203e-07, + "loss": 0.9395, + "step": 18116 + }, + { + "epoch": 1.543843204090328, + "grad_norm": 14.5625, + "learning_rate": 3.0990946613470727e-07, + "loss": 0.5107, + "step": 18117 + }, + { + "epoch": 1.543928419258628, + "grad_norm": 11.6875, + "learning_rate": 3.097987256797749e-07, + "loss": 0.3576, + "step": 18118 + }, + { + "epoch": 1.544013634426928, + "grad_norm": 13.0625, + "learning_rate": 3.0968800221522516e-07, + "loss": 0.3867, + "step": 18119 + }, + { + "epoch": 1.544098849595228, + "grad_norm": 22.0, + "learning_rate": 3.0957729574305873e-07, + "loss": 0.748, + "step": 18120 + }, + { + "epoch": 1.5441840647635279, + "grad_norm": 14.625, + "learning_rate": 3.0946660626527664e-07, + "loss": 0.5932, + "step": 18121 + }, + { + "epoch": 1.5442692799318278, + "grad_norm": 22.75, + "learning_rate": 3.0935593378387836e-07, + "loss": 0.3873, + "step": 18122 + }, + { + "epoch": 1.5443544951001278, + "grad_norm": 15.75, + "learning_rate": 3.092452783008647e-07, + "loss": 0.6452, + "step": 18123 + }, + { + "epoch": 1.5444397102684277, + "grad_norm": 16.625, + "learning_rate": 3.091346398182343e-07, + "loss": 0.6708, + "step": 18124 + }, + { + "epoch": 1.5445249254367277, + "grad_norm": 18.5, + "learning_rate": 3.0902401833798765e-07, + "loss": 0.4371, + "step": 18125 + }, + { + "epoch": 1.5446101406050277, + "grad_norm": 25.5, + "learning_rate": 3.089134138621226e-07, + "loss": 1.1384, + "step": 18126 + }, + { + "epoch": 1.5446953557733276, + "grad_norm": 15.75, + "learning_rate": 3.088028263926388e-07, + "loss": 0.4547, + "step": 18127 + }, + { + "epoch": 1.5447805709416276, + "grad_norm": 13.0625, + "learning_rate": 3.086922559315342e-07, + "loss": 0.8119, + "step": 18128 + }, + { + "epoch": 1.5448657861099275, + "grad_norm": 11.0, + "learning_rate": 3.085817024808074e-07, + "loss": 0.4128, + "step": 18129 + }, + { + "epoch": 1.5449510012782275, + "grad_norm": 12.625, + "learning_rate": 3.084711660424557e-07, + "loss": 0.5205, + "step": 18130 + }, + { + "epoch": 1.5450362164465274, + "grad_norm": 11.4375, + "learning_rate": 3.08360646618477e-07, + "loss": 0.3793, + "step": 18131 + }, + { + "epoch": 1.5451214316148274, + "grad_norm": 19.125, + "learning_rate": 3.08250144210868e-07, + "loss": 0.8055, + "step": 18132 + }, + { + "epoch": 1.5452066467831274, + "grad_norm": 21.375, + "learning_rate": 3.0813965882162596e-07, + "loss": 1.1525, + "step": 18133 + }, + { + "epoch": 1.5452918619514273, + "grad_norm": 10.1875, + "learning_rate": 3.080291904527477e-07, + "loss": 0.3026, + "step": 18134 + }, + { + "epoch": 1.5453770771197273, + "grad_norm": 19.875, + "learning_rate": 3.0791873910622894e-07, + "loss": 1.0445, + "step": 18135 + }, + { + "epoch": 1.5454622922880272, + "grad_norm": 16.0, + "learning_rate": 3.078083047840658e-07, + "loss": 0.7081, + "step": 18136 + }, + { + "epoch": 1.5455475074563272, + "grad_norm": 18.375, + "learning_rate": 3.076978874882544e-07, + "loss": 0.8116, + "step": 18137 + }, + { + "epoch": 1.5456327226246271, + "grad_norm": 14.0625, + "learning_rate": 3.075874872207894e-07, + "loss": 0.4058, + "step": 18138 + }, + { + "epoch": 1.545717937792927, + "grad_norm": 18.5, + "learning_rate": 3.0747710398366616e-07, + "loss": 0.5943, + "step": 18139 + }, + { + "epoch": 1.545803152961227, + "grad_norm": 17.375, + "learning_rate": 3.0736673777887983e-07, + "loss": 0.3744, + "step": 18140 + }, + { + "epoch": 1.545888368129527, + "grad_norm": 13.375, + "learning_rate": 3.072563886084241e-07, + "loss": 0.6065, + "step": 18141 + }, + { + "epoch": 1.545973583297827, + "grad_norm": 10.875, + "learning_rate": 3.071460564742937e-07, + "loss": 0.3043, + "step": 18142 + }, + { + "epoch": 1.546058798466127, + "grad_norm": 14.6875, + "learning_rate": 3.070357413784819e-07, + "loss": 0.6737, + "step": 18143 + }, + { + "epoch": 1.5461440136344269, + "grad_norm": 19.375, + "learning_rate": 3.0692544332298234e-07, + "loss": 0.8263, + "step": 18144 + }, + { + "epoch": 1.5462292288027268, + "grad_norm": 13.875, + "learning_rate": 3.0681516230978855e-07, + "loss": 0.4584, + "step": 18145 + }, + { + "epoch": 1.5463144439710268, + "grad_norm": 20.75, + "learning_rate": 3.067048983408935e-07, + "loss": 0.5865, + "step": 18146 + }, + { + "epoch": 1.5463996591393268, + "grad_norm": 11.125, + "learning_rate": 3.065946514182891e-07, + "loss": 0.3218, + "step": 18147 + }, + { + "epoch": 1.5464848743076267, + "grad_norm": 12.0, + "learning_rate": 3.064844215439683e-07, + "loss": 0.4027, + "step": 18148 + }, + { + "epoch": 1.5465700894759267, + "grad_norm": 17.0, + "learning_rate": 3.0637420871992255e-07, + "loss": 0.7887, + "step": 18149 + }, + { + "epoch": 1.5466553046442266, + "grad_norm": 13.1875, + "learning_rate": 3.062640129481437e-07, + "loss": 0.4705, + "step": 18150 + }, + { + "epoch": 1.5467405198125266, + "grad_norm": 14.8125, + "learning_rate": 3.0615383423062347e-07, + "loss": 0.603, + "step": 18151 + }, + { + "epoch": 1.5468257349808265, + "grad_norm": 15.0, + "learning_rate": 3.060436725693522e-07, + "loss": 0.6974, + "step": 18152 + }, + { + "epoch": 1.5469109501491265, + "grad_norm": 9.75, + "learning_rate": 3.05933527966321e-07, + "loss": 0.2105, + "step": 18153 + }, + { + "epoch": 1.5469961653174265, + "grad_norm": 17.125, + "learning_rate": 3.058234004235207e-07, + "loss": 0.4721, + "step": 18154 + }, + { + "epoch": 1.5470813804857264, + "grad_norm": 13.5625, + "learning_rate": 3.057132899429406e-07, + "loss": 0.4539, + "step": 18155 + }, + { + "epoch": 1.5471665956540264, + "grad_norm": 14.375, + "learning_rate": 3.056031965265713e-07, + "loss": 0.6603, + "step": 18156 + }, + { + "epoch": 1.5472518108223263, + "grad_norm": 41.5, + "learning_rate": 3.054931201764015e-07, + "loss": 0.8076, + "step": 18157 + }, + { + "epoch": 1.5473370259906263, + "grad_norm": 14.75, + "learning_rate": 3.053830608944208e-07, + "loss": 0.5673, + "step": 18158 + }, + { + "epoch": 1.5474222411589262, + "grad_norm": 17.625, + "learning_rate": 3.052730186826185e-07, + "loss": 0.9168, + "step": 18159 + }, + { + "epoch": 1.5475074563272262, + "grad_norm": 16.875, + "learning_rate": 3.051629935429823e-07, + "loss": 0.2476, + "step": 18160 + }, + { + "epoch": 1.5475926714955262, + "grad_norm": 17.625, + "learning_rate": 3.050529854775008e-07, + "loss": 1.1721, + "step": 18161 + }, + { + "epoch": 1.547677886663826, + "grad_norm": 20.125, + "learning_rate": 3.049429944881623e-07, + "loss": 0.996, + "step": 18162 + }, + { + "epoch": 1.547763101832126, + "grad_norm": 14.875, + "learning_rate": 3.0483302057695454e-07, + "loss": 0.6625, + "step": 18163 + }, + { + "epoch": 1.547848317000426, + "grad_norm": 13.0625, + "learning_rate": 3.047230637458644e-07, + "loss": 0.7355, + "step": 18164 + }, + { + "epoch": 1.547933532168726, + "grad_norm": 19.0, + "learning_rate": 3.046131239968787e-07, + "loss": 0.4716, + "step": 18165 + }, + { + "epoch": 1.548018747337026, + "grad_norm": 13.75, + "learning_rate": 3.0450320133198474e-07, + "loss": 0.6945, + "step": 18166 + }, + { + "epoch": 1.548103962505326, + "grad_norm": 15.5625, + "learning_rate": 3.043932957531688e-07, + "loss": 0.5598, + "step": 18167 + }, + { + "epoch": 1.5481891776736258, + "grad_norm": 16.0, + "learning_rate": 3.042834072624166e-07, + "loss": 0.5223, + "step": 18168 + }, + { + "epoch": 1.5482743928419258, + "grad_norm": 15.6875, + "learning_rate": 3.0417353586171427e-07, + "loss": 0.8531, + "step": 18169 + }, + { + "epoch": 1.5483596080102258, + "grad_norm": 13.6875, + "learning_rate": 3.040636815530472e-07, + "loss": 0.45, + "step": 18170 + }, + { + "epoch": 1.5484448231785257, + "grad_norm": 15.6875, + "learning_rate": 3.0395384433840103e-07, + "loss": 0.4013, + "step": 18171 + }, + { + "epoch": 1.5485300383468257, + "grad_norm": 16.0, + "learning_rate": 3.038440242197602e-07, + "loss": 0.6198, + "step": 18172 + }, + { + "epoch": 1.5486152535151256, + "grad_norm": 13.6875, + "learning_rate": 3.0373422119910885e-07, + "loss": 0.4915, + "step": 18173 + }, + { + "epoch": 1.5487004686834256, + "grad_norm": 13.3125, + "learning_rate": 3.0362443527843176e-07, + "loss": 0.6626, + "step": 18174 + }, + { + "epoch": 1.5487856838517255, + "grad_norm": 18.25, + "learning_rate": 3.0351466645971287e-07, + "loss": 0.7268, + "step": 18175 + }, + { + "epoch": 1.5488708990200255, + "grad_norm": 16.875, + "learning_rate": 3.0340491474493585e-07, + "loss": 0.612, + "step": 18176 + }, + { + "epoch": 1.5489561141883255, + "grad_norm": 14.0, + "learning_rate": 3.032951801360837e-07, + "loss": 0.3543, + "step": 18177 + }, + { + "epoch": 1.5490413293566254, + "grad_norm": 29.5, + "learning_rate": 3.031854626351395e-07, + "loss": 0.7592, + "step": 18178 + }, + { + "epoch": 1.5491265445249254, + "grad_norm": 16.0, + "learning_rate": 3.030757622440862e-07, + "loss": 0.673, + "step": 18179 + }, + { + "epoch": 1.5492117596932253, + "grad_norm": 11.625, + "learning_rate": 3.0296607896490635e-07, + "loss": 0.4766, + "step": 18180 + }, + { + "epoch": 1.5492969748615253, + "grad_norm": 19.375, + "learning_rate": 3.0285641279958185e-07, + "loss": 0.5371, + "step": 18181 + }, + { + "epoch": 1.5493821900298252, + "grad_norm": 14.625, + "learning_rate": 3.02746763750094e-07, + "loss": 0.6792, + "step": 18182 + }, + { + "epoch": 1.5494674051981252, + "grad_norm": 14.0, + "learning_rate": 3.026371318184247e-07, + "loss": 0.5179, + "step": 18183 + }, + { + "epoch": 1.5495526203664252, + "grad_norm": 16.0, + "learning_rate": 3.0252751700655544e-07, + "loss": 0.4493, + "step": 18184 + }, + { + "epoch": 1.5496378355347251, + "grad_norm": 17.125, + "learning_rate": 3.0241791931646644e-07, + "loss": 0.6506, + "step": 18185 + }, + { + "epoch": 1.549723050703025, + "grad_norm": 16.625, + "learning_rate": 3.0230833875013827e-07, + "loss": 0.7194, + "step": 18186 + }, + { + "epoch": 1.549808265871325, + "grad_norm": 15.6875, + "learning_rate": 3.0219877530955147e-07, + "loss": 0.592, + "step": 18187 + }, + { + "epoch": 1.549893481039625, + "grad_norm": 15.75, + "learning_rate": 3.0208922899668634e-07, + "loss": 0.672, + "step": 18188 + }, + { + "epoch": 1.549978696207925, + "grad_norm": 13.75, + "learning_rate": 3.01979699813522e-07, + "loss": 0.4713, + "step": 18189 + }, + { + "epoch": 1.550063911376225, + "grad_norm": 17.75, + "learning_rate": 3.018701877620374e-07, + "loss": 0.8693, + "step": 18190 + }, + { + "epoch": 1.5501491265445249, + "grad_norm": 17.375, + "learning_rate": 3.0176069284421195e-07, + "loss": 0.6229, + "step": 18191 + }, + { + "epoch": 1.5502343417128248, + "grad_norm": 17.5, + "learning_rate": 3.016512150620243e-07, + "loss": 0.839, + "step": 18192 + }, + { + "epoch": 1.5503195568811248, + "grad_norm": 12.125, + "learning_rate": 3.015417544174533e-07, + "loss": 0.3162, + "step": 18193 + }, + { + "epoch": 1.5504047720494247, + "grad_norm": 14.5, + "learning_rate": 3.01432310912476e-07, + "loss": 0.4831, + "step": 18194 + }, + { + "epoch": 1.5504899872177247, + "grad_norm": 17.625, + "learning_rate": 3.013228845490708e-07, + "loss": 0.5869, + "step": 18195 + }, + { + "epoch": 1.5505752023860246, + "grad_norm": 18.125, + "learning_rate": 3.0121347532921533e-07, + "loss": 0.518, + "step": 18196 + }, + { + "epoch": 1.5506604175543246, + "grad_norm": 13.3125, + "learning_rate": 3.0110408325488614e-07, + "loss": 0.638, + "step": 18197 + }, + { + "epoch": 1.5507456327226246, + "grad_norm": 16.625, + "learning_rate": 3.009947083280608e-07, + "loss": 0.5814, + "step": 18198 + }, + { + "epoch": 1.5508308478909245, + "grad_norm": 12.375, + "learning_rate": 3.008853505507149e-07, + "loss": 0.5837, + "step": 18199 + }, + { + "epoch": 1.5509160630592245, + "grad_norm": 18.25, + "learning_rate": 3.007760099248251e-07, + "loss": 0.77, + "step": 18200 + }, + { + "epoch": 1.5510012782275244, + "grad_norm": 16.875, + "learning_rate": 3.0066668645236774e-07, + "loss": 0.9167, + "step": 18201 + }, + { + "epoch": 1.5510864933958244, + "grad_norm": 12.5, + "learning_rate": 3.0055738013531776e-07, + "loss": 0.5106, + "step": 18202 + }, + { + "epoch": 1.5511717085641243, + "grad_norm": 15.5, + "learning_rate": 3.004480909756505e-07, + "loss": 0.8513, + "step": 18203 + }, + { + "epoch": 1.5512569237324243, + "grad_norm": 14.25, + "learning_rate": 3.00338818975341e-07, + "loss": 0.375, + "step": 18204 + }, + { + "epoch": 1.5513421389007243, + "grad_norm": 18.625, + "learning_rate": 3.002295641363645e-07, + "loss": 0.731, + "step": 18205 + }, + { + "epoch": 1.5514273540690242, + "grad_norm": 11.375, + "learning_rate": 3.0012032646069485e-07, + "loss": 0.2764, + "step": 18206 + }, + { + "epoch": 1.5515125692373242, + "grad_norm": 14.0, + "learning_rate": 3.000111059503057e-07, + "loss": 0.5755, + "step": 18207 + }, + { + "epoch": 1.5515977844056241, + "grad_norm": 13.4375, + "learning_rate": 2.9990190260717105e-07, + "loss": 0.389, + "step": 18208 + }, + { + "epoch": 1.551682999573924, + "grad_norm": 26.375, + "learning_rate": 2.997927164332645e-07, + "loss": 0.7195, + "step": 18209 + }, + { + "epoch": 1.551768214742224, + "grad_norm": 9.0625, + "learning_rate": 2.996835474305594e-07, + "loss": 0.7298, + "step": 18210 + }, + { + "epoch": 1.551853429910524, + "grad_norm": 53.25, + "learning_rate": 2.995743956010279e-07, + "loss": 0.8251, + "step": 18211 + }, + { + "epoch": 1.551938645078824, + "grad_norm": 16.75, + "learning_rate": 2.994652609466428e-07, + "loss": 0.6133, + "step": 18212 + }, + { + "epoch": 1.552023860247124, + "grad_norm": 14.4375, + "learning_rate": 2.993561434693766e-07, + "loss": 0.6612, + "step": 18213 + }, + { + "epoch": 1.5521090754154239, + "grad_norm": 11.625, + "learning_rate": 2.992470431712009e-07, + "loss": 0.4034, + "step": 18214 + }, + { + "epoch": 1.5521942905837238, + "grad_norm": 18.375, + "learning_rate": 2.991379600540867e-07, + "loss": 0.4562, + "step": 18215 + }, + { + "epoch": 1.5522795057520238, + "grad_norm": 18.875, + "learning_rate": 2.990288941200058e-07, + "loss": 0.5251, + "step": 18216 + }, + { + "epoch": 1.5523647209203237, + "grad_norm": 12.8125, + "learning_rate": 2.989198453709291e-07, + "loss": 0.4322, + "step": 18217 + }, + { + "epoch": 1.5524499360886237, + "grad_norm": 12.0, + "learning_rate": 2.9881081380882747e-07, + "loss": 0.3745, + "step": 18218 + }, + { + "epoch": 1.5525351512569237, + "grad_norm": 16.875, + "learning_rate": 2.9870179943567056e-07, + "loss": 0.7131, + "step": 18219 + }, + { + "epoch": 1.5526203664252236, + "grad_norm": 14.5625, + "learning_rate": 2.985928022534289e-07, + "loss": 0.4414, + "step": 18220 + }, + { + "epoch": 1.5527055815935236, + "grad_norm": 15.625, + "learning_rate": 2.984838222640722e-07, + "loss": 0.5665, + "step": 18221 + }, + { + "epoch": 1.5527907967618235, + "grad_norm": 30.75, + "learning_rate": 2.9837485946956927e-07, + "loss": 0.994, + "step": 18222 + }, + { + "epoch": 1.5528760119301235, + "grad_norm": 10.8125, + "learning_rate": 2.9826591387189006e-07, + "loss": 0.3535, + "step": 18223 + }, + { + "epoch": 1.5529612270984234, + "grad_norm": 12.5625, + "learning_rate": 2.9815698547300235e-07, + "loss": 0.6345, + "step": 18224 + }, + { + "epoch": 1.5530464422667234, + "grad_norm": 9.5, + "learning_rate": 2.980480742748752e-07, + "loss": 0.2725, + "step": 18225 + }, + { + "epoch": 1.5531316574350233, + "grad_norm": 17.125, + "learning_rate": 2.979391802794765e-07, + "loss": 0.6158, + "step": 18226 + }, + { + "epoch": 1.5532168726033233, + "grad_norm": 18.25, + "learning_rate": 2.978303034887746e-07, + "loss": 0.8655, + "step": 18227 + }, + { + "epoch": 1.5533020877716233, + "grad_norm": 15.375, + "learning_rate": 2.9772144390473616e-07, + "loss": 0.4809, + "step": 18228 + }, + { + "epoch": 1.5533873029399232, + "grad_norm": 22.125, + "learning_rate": 2.976126015293293e-07, + "loss": 0.7585, + "step": 18229 + }, + { + "epoch": 1.5534725181082232, + "grad_norm": 13.4375, + "learning_rate": 2.9750377636452e-07, + "loss": 0.5745, + "step": 18230 + }, + { + "epoch": 1.5535577332765231, + "grad_norm": 14.3125, + "learning_rate": 2.9739496841227567e-07, + "loss": 0.466, + "step": 18231 + }, + { + "epoch": 1.553642948444823, + "grad_norm": 15.9375, + "learning_rate": 2.9728617767456167e-07, + "loss": 0.6781, + "step": 18232 + }, + { + "epoch": 1.553728163613123, + "grad_norm": 15.0, + "learning_rate": 2.9717740415334454e-07, + "loss": 0.5491, + "step": 18233 + }, + { + "epoch": 1.553813378781423, + "grad_norm": 26.5, + "learning_rate": 2.970686478505899e-07, + "loss": 0.5826, + "step": 18234 + }, + { + "epoch": 1.553898593949723, + "grad_norm": 14.8125, + "learning_rate": 2.969599087682634e-07, + "loss": 0.3537, + "step": 18235 + }, + { + "epoch": 1.553983809118023, + "grad_norm": 17.125, + "learning_rate": 2.9685118690832947e-07, + "loss": 0.395, + "step": 18236 + }, + { + "epoch": 1.5540690242863229, + "grad_norm": 17.125, + "learning_rate": 2.9674248227275324e-07, + "loss": 0.6078, + "step": 18237 + }, + { + "epoch": 1.5541542394546228, + "grad_norm": 16.75, + "learning_rate": 2.9663379486349855e-07, + "loss": 0.4337, + "step": 18238 + }, + { + "epoch": 1.5542394546229228, + "grad_norm": 14.5625, + "learning_rate": 2.9652512468253e-07, + "loss": 0.6212, + "step": 18239 + }, + { + "epoch": 1.5543246697912227, + "grad_norm": 12.5625, + "learning_rate": 2.9641647173181144e-07, + "loss": 0.4846, + "step": 18240 + }, + { + "epoch": 1.5544098849595227, + "grad_norm": 15.3125, + "learning_rate": 2.963078360133059e-07, + "loss": 0.4001, + "step": 18241 + }, + { + "epoch": 1.5544951001278227, + "grad_norm": 11.375, + "learning_rate": 2.961992175289768e-07, + "loss": 0.364, + "step": 18242 + }, + { + "epoch": 1.5545803152961226, + "grad_norm": 10.0625, + "learning_rate": 2.9609061628078693e-07, + "loss": 0.39, + "step": 18243 + }, + { + "epoch": 1.5546655304644226, + "grad_norm": 18.25, + "learning_rate": 2.959820322706991e-07, + "loss": 0.8767, + "step": 18244 + }, + { + "epoch": 1.5547507456327225, + "grad_norm": 16.0, + "learning_rate": 2.958734655006755e-07, + "loss": 0.3333, + "step": 18245 + }, + { + "epoch": 1.5548359608010225, + "grad_norm": 17.625, + "learning_rate": 2.9576491597267735e-07, + "loss": 0.6168, + "step": 18246 + }, + { + "epoch": 1.5549211759693224, + "grad_norm": 16.5, + "learning_rate": 2.956563836886667e-07, + "loss": 0.3021, + "step": 18247 + }, + { + "epoch": 1.5550063911376224, + "grad_norm": 11.5, + "learning_rate": 2.9554786865060523e-07, + "loss": 0.3279, + "step": 18248 + }, + { + "epoch": 1.5550916063059224, + "grad_norm": 22.75, + "learning_rate": 2.9543937086045317e-07, + "loss": 0.7473, + "step": 18249 + }, + { + "epoch": 1.5551768214742223, + "grad_norm": 11.9375, + "learning_rate": 2.9533089032017166e-07, + "loss": 0.3894, + "step": 18250 + }, + { + "epoch": 1.5552620366425223, + "grad_norm": 20.75, + "learning_rate": 2.9522242703172084e-07, + "loss": 1.0367, + "step": 18251 + }, + { + "epoch": 1.5553472518108222, + "grad_norm": 38.5, + "learning_rate": 2.951139809970613e-07, + "loss": 1.6008, + "step": 18252 + }, + { + "epoch": 1.5554324669791222, + "grad_norm": 24.125, + "learning_rate": 2.9500555221815233e-07, + "loss": 0.8166, + "step": 18253 + }, + { + "epoch": 1.5555176821474221, + "grad_norm": 11.625, + "learning_rate": 2.948971406969529e-07, + "loss": 0.2904, + "step": 18254 + }, + { + "epoch": 1.555602897315722, + "grad_norm": 11.25, + "learning_rate": 2.9478874643542265e-07, + "loss": 0.3079, + "step": 18255 + }, + { + "epoch": 1.555688112484022, + "grad_norm": 15.6875, + "learning_rate": 2.946803694355203e-07, + "loss": 0.4667, + "step": 18256 + }, + { + "epoch": 1.555773327652322, + "grad_norm": 11.4375, + "learning_rate": 2.945720096992047e-07, + "loss": 0.317, + "step": 18257 + }, + { + "epoch": 1.555858542820622, + "grad_norm": 19.25, + "learning_rate": 2.944636672284333e-07, + "loss": 0.9343, + "step": 18258 + }, + { + "epoch": 1.555943757988922, + "grad_norm": 12.0, + "learning_rate": 2.943553420251642e-07, + "loss": 0.4059, + "step": 18259 + }, + { + "epoch": 1.5560289731572219, + "grad_norm": 36.75, + "learning_rate": 2.9424703409135517e-07, + "loss": 0.6904, + "step": 18260 + }, + { + "epoch": 1.5561141883255218, + "grad_norm": 16.0, + "learning_rate": 2.9413874342896385e-07, + "loss": 0.5821, + "step": 18261 + }, + { + "epoch": 1.5561994034938218, + "grad_norm": 15.875, + "learning_rate": 2.94030470039946e-07, + "loss": 0.4557, + "step": 18262 + }, + { + "epoch": 1.5562846186621218, + "grad_norm": 15.3125, + "learning_rate": 2.939222139262589e-07, + "loss": 0.3218, + "step": 18263 + }, + { + "epoch": 1.5563698338304217, + "grad_norm": 14.3125, + "learning_rate": 2.938139750898589e-07, + "loss": 0.36, + "step": 18264 + }, + { + "epoch": 1.5564550489987217, + "grad_norm": 29.75, + "learning_rate": 2.9370575353270204e-07, + "loss": 1.1815, + "step": 18265 + }, + { + "epoch": 1.5565402641670216, + "grad_norm": 14.125, + "learning_rate": 2.935975492567436e-07, + "loss": 0.5413, + "step": 18266 + }, + { + "epoch": 1.5566254793353216, + "grad_norm": 13.9375, + "learning_rate": 2.934893622639391e-07, + "loss": 0.5462, + "step": 18267 + }, + { + "epoch": 1.5567106945036215, + "grad_norm": 11.5, + "learning_rate": 2.933811925562438e-07, + "loss": 0.2213, + "step": 18268 + }, + { + "epoch": 1.5567959096719215, + "grad_norm": 14.5625, + "learning_rate": 2.932730401356126e-07, + "loss": 0.3461, + "step": 18269 + }, + { + "epoch": 1.5568811248402215, + "grad_norm": 16.625, + "learning_rate": 2.931649050039996e-07, + "loss": 0.4414, + "step": 18270 + }, + { + "epoch": 1.5569663400085214, + "grad_norm": 21.5, + "learning_rate": 2.9305678716335856e-07, + "loss": 0.4036, + "step": 18271 + }, + { + "epoch": 1.5570515551768214, + "grad_norm": 27.75, + "learning_rate": 2.9294868661564375e-07, + "loss": 0.8809, + "step": 18272 + }, + { + "epoch": 1.5571367703451213, + "grad_norm": 22.25, + "learning_rate": 2.9284060336280855e-07, + "loss": 0.7155, + "step": 18273 + }, + { + "epoch": 1.5572219855134213, + "grad_norm": 26.375, + "learning_rate": 2.927325374068064e-07, + "loss": 0.3937, + "step": 18274 + }, + { + "epoch": 1.5573072006817212, + "grad_norm": 18.75, + "learning_rate": 2.9262448874958965e-07, + "loss": 1.0006, + "step": 18275 + }, + { + "epoch": 1.5573924158500212, + "grad_norm": 18.25, + "learning_rate": 2.925164573931111e-07, + "loss": 0.574, + "step": 18276 + }, + { + "epoch": 1.5574776310183212, + "grad_norm": 15.0, + "learning_rate": 2.9240844333932333e-07, + "loss": 0.6314, + "step": 18277 + }, + { + "epoch": 1.557562846186621, + "grad_norm": 19.375, + "learning_rate": 2.9230044659017785e-07, + "loss": 0.3731, + "step": 18278 + }, + { + "epoch": 1.557648061354921, + "grad_norm": 18.0, + "learning_rate": 2.92192467147626e-07, + "loss": 0.7989, + "step": 18279 + }, + { + "epoch": 1.557733276523221, + "grad_norm": 13.625, + "learning_rate": 2.9208450501361954e-07, + "loss": 0.3797, + "step": 18280 + }, + { + "epoch": 1.557818491691521, + "grad_norm": 11.875, + "learning_rate": 2.9197656019010925e-07, + "loss": 0.419, + "step": 18281 + }, + { + "epoch": 1.557903706859821, + "grad_norm": 15.875, + "learning_rate": 2.918686326790461e-07, + "loss": 0.6322, + "step": 18282 + }, + { + "epoch": 1.557988922028121, + "grad_norm": 12.375, + "learning_rate": 2.917607224823801e-07, + "loss": 0.3976, + "step": 18283 + }, + { + "epoch": 1.5580741371964208, + "grad_norm": 19.25, + "learning_rate": 2.9165282960206126e-07, + "loss": 0.5949, + "step": 18284 + }, + { + "epoch": 1.5581593523647208, + "grad_norm": 14.0, + "learning_rate": 2.915449540400396e-07, + "loss": 0.3596, + "step": 18285 + }, + { + "epoch": 1.5582445675330208, + "grad_norm": 19.875, + "learning_rate": 2.914370957982647e-07, + "loss": 0.4221, + "step": 18286 + }, + { + "epoch": 1.5583297827013207, + "grad_norm": 14.9375, + "learning_rate": 2.9132925487868556e-07, + "loss": 0.7852, + "step": 18287 + }, + { + "epoch": 1.5584149978696207, + "grad_norm": 16.875, + "learning_rate": 2.9122143128325024e-07, + "loss": 0.403, + "step": 18288 + }, + { + "epoch": 1.5585002130379206, + "grad_norm": 14.1875, + "learning_rate": 2.9111362501390795e-07, + "loss": 0.4639, + "step": 18289 + }, + { + "epoch": 1.5585854282062206, + "grad_norm": 13.9375, + "learning_rate": 2.910058360726066e-07, + "loss": 0.495, + "step": 18290 + }, + { + "epoch": 1.5586706433745205, + "grad_norm": 14.9375, + "learning_rate": 2.9089806446129436e-07, + "loss": 0.8543, + "step": 18291 + }, + { + "epoch": 1.5587558585428205, + "grad_norm": 17.625, + "learning_rate": 2.9079031018191847e-07, + "loss": 0.5971, + "step": 18292 + }, + { + "epoch": 1.5588410737111205, + "grad_norm": 16.125, + "learning_rate": 2.9068257323642603e-07, + "loss": 0.7796, + "step": 18293 + }, + { + "epoch": 1.5589262888794204, + "grad_norm": 10.75, + "learning_rate": 2.9057485362676453e-07, + "loss": 0.9313, + "step": 18294 + }, + { + "epoch": 1.5590115040477204, + "grad_norm": 15.625, + "learning_rate": 2.9046715135488025e-07, + "loss": 0.3727, + "step": 18295 + }, + { + "epoch": 1.5590967192160203, + "grad_norm": 15.875, + "learning_rate": 2.9035946642271913e-07, + "loss": 0.6305, + "step": 18296 + }, + { + "epoch": 1.5591819343843203, + "grad_norm": 13.5625, + "learning_rate": 2.902517988322273e-07, + "loss": 0.5754, + "step": 18297 + }, + { + "epoch": 1.5592671495526202, + "grad_norm": 16.75, + "learning_rate": 2.901441485853507e-07, + "loss": 0.5426, + "step": 18298 + }, + { + "epoch": 1.5593523647209202, + "grad_norm": 15.75, + "learning_rate": 2.9003651568403483e-07, + "loss": 0.7613, + "step": 18299 + }, + { + "epoch": 1.5594375798892202, + "grad_norm": 14.9375, + "learning_rate": 2.899289001302241e-07, + "loss": 0.4962, + "step": 18300 + }, + { + "epoch": 1.5595227950575201, + "grad_norm": 18.25, + "learning_rate": 2.898213019258636e-07, + "loss": 0.8255, + "step": 18301 + }, + { + "epoch": 1.55960801022582, + "grad_norm": 20.875, + "learning_rate": 2.8971372107289816e-07, + "loss": 0.9768, + "step": 18302 + }, + { + "epoch": 1.55969322539412, + "grad_norm": 16.125, + "learning_rate": 2.896061575732709e-07, + "loss": 0.6189, + "step": 18303 + }, + { + "epoch": 1.55977844056242, + "grad_norm": 12.25, + "learning_rate": 2.894986114289265e-07, + "loss": 0.3068, + "step": 18304 + }, + { + "epoch": 1.55986365573072, + "grad_norm": 14.75, + "learning_rate": 2.893910826418077e-07, + "loss": 0.346, + "step": 18305 + }, + { + "epoch": 1.55994887089902, + "grad_norm": 15.4375, + "learning_rate": 2.892835712138581e-07, + "loss": 0.7032, + "step": 18306 + }, + { + "epoch": 1.5600340860673199, + "grad_norm": 14.1875, + "learning_rate": 2.8917607714702063e-07, + "loss": 0.5629, + "step": 18307 + }, + { + "epoch": 1.5601193012356198, + "grad_norm": 13.375, + "learning_rate": 2.8906860044323734e-07, + "loss": 0.5289, + "step": 18308 + }, + { + "epoch": 1.5602045164039198, + "grad_norm": 17.5, + "learning_rate": 2.889611411044509e-07, + "loss": 0.5603, + "step": 18309 + }, + { + "epoch": 1.5602897315722197, + "grad_norm": 14.375, + "learning_rate": 2.8885369913260324e-07, + "loss": 0.4301, + "step": 18310 + }, + { + "epoch": 1.5603749467405197, + "grad_norm": 15.5, + "learning_rate": 2.8874627452963533e-07, + "loss": 0.7183, + "step": 18311 + }, + { + "epoch": 1.5604601619088196, + "grad_norm": 14.25, + "learning_rate": 2.8863886729748934e-07, + "loss": 0.4862, + "step": 18312 + }, + { + "epoch": 1.5605453770771196, + "grad_norm": 14.125, + "learning_rate": 2.8853147743810525e-07, + "loss": 0.4342, + "step": 18313 + }, + { + "epoch": 1.5606305922454196, + "grad_norm": 16.25, + "learning_rate": 2.8842410495342436e-07, + "loss": 0.5707, + "step": 18314 + }, + { + "epoch": 1.5607158074137195, + "grad_norm": 17.75, + "learning_rate": 2.883167498453868e-07, + "loss": 0.356, + "step": 18315 + }, + { + "epoch": 1.5608010225820195, + "grad_norm": 13.3125, + "learning_rate": 2.8820941211593276e-07, + "loss": 0.5358, + "step": 18316 + }, + { + "epoch": 1.5608862377503194, + "grad_norm": 14.6875, + "learning_rate": 2.881020917670016e-07, + "loss": 0.5579, + "step": 18317 + }, + { + "epoch": 1.5609714529186194, + "grad_norm": 13.5625, + "learning_rate": 2.8799478880053325e-07, + "loss": 0.5988, + "step": 18318 + }, + { + "epoch": 1.5610566680869193, + "grad_norm": 15.625, + "learning_rate": 2.878875032184662e-07, + "loss": 0.5068, + "step": 18319 + }, + { + "epoch": 1.5611418832552193, + "grad_norm": 14.0625, + "learning_rate": 2.8778023502273935e-07, + "loss": 0.4186, + "step": 18320 + }, + { + "epoch": 1.5612270984235193, + "grad_norm": 15.5625, + "learning_rate": 2.876729842152916e-07, + "loss": 0.7328, + "step": 18321 + }, + { + "epoch": 1.5613123135918192, + "grad_norm": 21.125, + "learning_rate": 2.875657507980603e-07, + "loss": 0.9197, + "step": 18322 + }, + { + "epoch": 1.5613975287601192, + "grad_norm": 12.5625, + "learning_rate": 2.874585347729837e-07, + "loss": 0.3865, + "step": 18323 + }, + { + "epoch": 1.5614827439284191, + "grad_norm": 19.25, + "learning_rate": 2.873513361419997e-07, + "loss": 0.6067, + "step": 18324 + }, + { + "epoch": 1.561567959096719, + "grad_norm": 15.3125, + "learning_rate": 2.872441549070447e-07, + "loss": 0.4283, + "step": 18325 + }, + { + "epoch": 1.561653174265019, + "grad_norm": 16.75, + "learning_rate": 2.8713699107005625e-07, + "loss": 0.7615, + "step": 18326 + }, + { + "epoch": 1.561738389433319, + "grad_norm": 15.4375, + "learning_rate": 2.870298446329704e-07, + "loss": 0.5689, + "step": 18327 + }, + { + "epoch": 1.561823604601619, + "grad_norm": 13.1875, + "learning_rate": 2.869227155977235e-07, + "loss": 0.6617, + "step": 18328 + }, + { + "epoch": 1.561908819769919, + "grad_norm": 21.875, + "learning_rate": 2.868156039662519e-07, + "loss": 0.5673, + "step": 18329 + }, + { + "epoch": 1.5619940349382189, + "grad_norm": 12.3125, + "learning_rate": 2.867085097404905e-07, + "loss": 0.3517, + "step": 18330 + }, + { + "epoch": 1.5620792501065188, + "grad_norm": 14.5, + "learning_rate": 2.8660143292237514e-07, + "loss": 0.403, + "step": 18331 + }, + { + "epoch": 1.5621644652748188, + "grad_norm": 22.0, + "learning_rate": 2.8649437351384063e-07, + "loss": 0.6531, + "step": 18332 + }, + { + "epoch": 1.5622496804431187, + "grad_norm": 13.8125, + "learning_rate": 2.86387331516822e-07, + "loss": 0.3981, + "step": 18333 + }, + { + "epoch": 1.5623348956114187, + "grad_norm": 26.875, + "learning_rate": 2.8628030693325316e-07, + "loss": 1.0107, + "step": 18334 + }, + { + "epoch": 1.5624201107797187, + "grad_norm": 14.875, + "learning_rate": 2.8617329976506803e-07, + "loss": 0.649, + "step": 18335 + }, + { + "epoch": 1.5625053259480186, + "grad_norm": 13.75, + "learning_rate": 2.8606631001420054e-07, + "loss": 0.5941, + "step": 18336 + }, + { + "epoch": 1.5625905411163186, + "grad_norm": 12.5, + "learning_rate": 2.859593376825845e-07, + "loss": 0.3596, + "step": 18337 + }, + { + "epoch": 1.5626757562846185, + "grad_norm": 17.75, + "learning_rate": 2.8585238277215226e-07, + "loss": 0.4282, + "step": 18338 + }, + { + "epoch": 1.5627609714529185, + "grad_norm": 17.625, + "learning_rate": 2.857454452848371e-07, + "loss": 0.8426, + "step": 18339 + }, + { + "epoch": 1.5628461866212184, + "grad_norm": 15.0625, + "learning_rate": 2.856385252225713e-07, + "loss": 0.4469, + "step": 18340 + }, + { + "epoch": 1.5629314017895184, + "grad_norm": 50.5, + "learning_rate": 2.8553162258728743e-07, + "loss": 0.6292, + "step": 18341 + }, + { + "epoch": 1.5630166169578183, + "grad_norm": 11.3125, + "learning_rate": 2.854247373809166e-07, + "loss": 0.5563, + "step": 18342 + }, + { + "epoch": 1.5631018321261183, + "grad_norm": 16.375, + "learning_rate": 2.8531786960539114e-07, + "loss": 0.895, + "step": 18343 + }, + { + "epoch": 1.5631870472944183, + "grad_norm": 18.625, + "learning_rate": 2.8521101926264145e-07, + "loss": 0.8199, + "step": 18344 + }, + { + "epoch": 1.5632722624627182, + "grad_norm": 21.875, + "learning_rate": 2.8510418635459893e-07, + "loss": 0.7596, + "step": 18345 + }, + { + "epoch": 1.5633574776310182, + "grad_norm": 23.25, + "learning_rate": 2.849973708831942e-07, + "loss": 1.0762, + "step": 18346 + }, + { + "epoch": 1.5634426927993181, + "grad_norm": 13.5, + "learning_rate": 2.848905728503572e-07, + "loss": 0.5918, + "step": 18347 + }, + { + "epoch": 1.563527907967618, + "grad_norm": 17.375, + "learning_rate": 2.8478379225801797e-07, + "loss": 0.6026, + "step": 18348 + }, + { + "epoch": 1.563613123135918, + "grad_norm": 11.8125, + "learning_rate": 2.846770291081061e-07, + "loss": 0.5549, + "step": 18349 + }, + { + "epoch": 1.563698338304218, + "grad_norm": 19.5, + "learning_rate": 2.8457028340255154e-07, + "loss": 0.8989, + "step": 18350 + }, + { + "epoch": 1.563783553472518, + "grad_norm": 18.125, + "learning_rate": 2.844635551432827e-07, + "loss": 0.5535, + "step": 18351 + }, + { + "epoch": 1.563868768640818, + "grad_norm": 20.75, + "learning_rate": 2.8435684433222795e-07, + "loss": 0.7023, + "step": 18352 + }, + { + "epoch": 1.5639539838091179, + "grad_norm": 15.375, + "learning_rate": 2.8425015097131606e-07, + "loss": 0.6022, + "step": 18353 + }, + { + "epoch": 1.5640391989774178, + "grad_norm": 13.625, + "learning_rate": 2.841434750624755e-07, + "loss": 0.4727, + "step": 18354 + }, + { + "epoch": 1.5641244141457178, + "grad_norm": 17.625, + "learning_rate": 2.840368166076332e-07, + "loss": 0.6407, + "step": 18355 + }, + { + "epoch": 1.5642096293140177, + "grad_norm": 17.5, + "learning_rate": 2.839301756087169e-07, + "loss": 0.8453, + "step": 18356 + }, + { + "epoch": 1.5642948444823177, + "grad_norm": 11.9375, + "learning_rate": 2.83823552067654e-07, + "loss": 0.2862, + "step": 18357 + }, + { + "epoch": 1.5643800596506177, + "grad_norm": 13.25, + "learning_rate": 2.837169459863712e-07, + "loss": 0.6984, + "step": 18358 + }, + { + "epoch": 1.5644652748189176, + "grad_norm": 10.6875, + "learning_rate": 2.836103573667949e-07, + "loss": 0.2799, + "step": 18359 + }, + { + "epoch": 1.5645504899872176, + "grad_norm": 21.375, + "learning_rate": 2.83503786210851e-07, + "loss": 0.7326, + "step": 18360 + }, + { + "epoch": 1.5646357051555175, + "grad_norm": 23.375, + "learning_rate": 2.833972325204655e-07, + "loss": 0.7949, + "step": 18361 + }, + { + "epoch": 1.5647209203238175, + "grad_norm": 15.4375, + "learning_rate": 2.83290696297564e-07, + "loss": 0.6501, + "step": 18362 + }, + { + "epoch": 1.5648061354921174, + "grad_norm": 15.4375, + "learning_rate": 2.8318417754407206e-07, + "loss": 0.471, + "step": 18363 + }, + { + "epoch": 1.5648913506604174, + "grad_norm": 26.5, + "learning_rate": 2.830776762619139e-07, + "loss": 1.012, + "step": 18364 + }, + { + "epoch": 1.5649765658287174, + "grad_norm": 18.0, + "learning_rate": 2.8297119245301443e-07, + "loss": 0.7449, + "step": 18365 + }, + { + "epoch": 1.5650617809970173, + "grad_norm": 15.3125, + "learning_rate": 2.8286472611929794e-07, + "loss": 0.6536, + "step": 18366 + }, + { + "epoch": 1.5651469961653173, + "grad_norm": 13.625, + "learning_rate": 2.8275827726268897e-07, + "loss": 0.5513, + "step": 18367 + }, + { + "epoch": 1.5652322113336172, + "grad_norm": 16.25, + "learning_rate": 2.8265184588511007e-07, + "loss": 0.6335, + "step": 18368 + }, + { + "epoch": 1.5653174265019172, + "grad_norm": 20.375, + "learning_rate": 2.825454319884849e-07, + "loss": 0.8774, + "step": 18369 + }, + { + "epoch": 1.5654026416702171, + "grad_norm": 16.5, + "learning_rate": 2.8243903557473666e-07, + "loss": 0.5222, + "step": 18370 + }, + { + "epoch": 1.565487856838517, + "grad_norm": 13.3125, + "learning_rate": 2.8233265664578847e-07, + "loss": 0.4482, + "step": 18371 + }, + { + "epoch": 1.565573072006817, + "grad_norm": 13.3125, + "learning_rate": 2.822262952035618e-07, + "loss": 0.5475, + "step": 18372 + }, + { + "epoch": 1.565658287175117, + "grad_norm": 22.625, + "learning_rate": 2.821199512499792e-07, + "loss": 0.7336, + "step": 18373 + }, + { + "epoch": 1.565743502343417, + "grad_norm": 14.625, + "learning_rate": 2.820136247869623e-07, + "loss": 0.7298, + "step": 18374 + }, + { + "epoch": 1.565828717511717, + "grad_norm": 16.5, + "learning_rate": 2.819073158164329e-07, + "loss": 0.674, + "step": 18375 + }, + { + "epoch": 1.5659139326800169, + "grad_norm": 22.5, + "learning_rate": 2.8180102434031185e-07, + "loss": 0.6195, + "step": 18376 + }, + { + "epoch": 1.5659991478483168, + "grad_norm": 12.5625, + "learning_rate": 2.8169475036051956e-07, + "loss": 0.5279, + "step": 18377 + }, + { + "epoch": 1.5660843630166168, + "grad_norm": 20.0, + "learning_rate": 2.8158849387897674e-07, + "loss": 0.4234, + "step": 18378 + }, + { + "epoch": 1.5661695781849168, + "grad_norm": 16.5, + "learning_rate": 2.814822548976037e-07, + "loss": 0.5858, + "step": 18379 + }, + { + "epoch": 1.5662547933532167, + "grad_norm": 29.0, + "learning_rate": 2.813760334183206e-07, + "loss": 1.2449, + "step": 18380 + }, + { + "epoch": 1.5663400085215167, + "grad_norm": 13.875, + "learning_rate": 2.812698294430463e-07, + "loss": 0.3739, + "step": 18381 + }, + { + "epoch": 1.5664252236898166, + "grad_norm": 17.75, + "learning_rate": 2.8116364297370025e-07, + "loss": 0.9608, + "step": 18382 + }, + { + "epoch": 1.5665104388581166, + "grad_norm": 11.375, + "learning_rate": 2.810574740122018e-07, + "loss": 0.5017, + "step": 18383 + }, + { + "epoch": 1.5665956540264165, + "grad_norm": 13.8125, + "learning_rate": 2.8095132256046917e-07, + "loss": 0.6418, + "step": 18384 + }, + { + "epoch": 1.5666808691947165, + "grad_norm": 14.125, + "learning_rate": 2.808451886204203e-07, + "loss": 0.5668, + "step": 18385 + }, + { + "epoch": 1.5667660843630165, + "grad_norm": 16.25, + "learning_rate": 2.8073907219397317e-07, + "loss": 0.7968, + "step": 18386 + }, + { + "epoch": 1.5668512995313164, + "grad_norm": 13.75, + "learning_rate": 2.8063297328304594e-07, + "loss": 0.6196, + "step": 18387 + }, + { + "epoch": 1.5669365146996164, + "grad_norm": 16.25, + "learning_rate": 2.8052689188955584e-07, + "loss": 0.4646, + "step": 18388 + }, + { + "epoch": 1.5670217298679163, + "grad_norm": 13.75, + "learning_rate": 2.8042082801541945e-07, + "loss": 0.4478, + "step": 18389 + }, + { + "epoch": 1.5671069450362163, + "grad_norm": 24.75, + "learning_rate": 2.803147816625536e-07, + "loss": 0.9445, + "step": 18390 + }, + { + "epoch": 1.5671921602045162, + "grad_norm": 13.625, + "learning_rate": 2.802087528328751e-07, + "loss": 0.4044, + "step": 18391 + }, + { + "epoch": 1.5672773753728162, + "grad_norm": 16.875, + "learning_rate": 2.8010274152829934e-07, + "loss": 0.8602, + "step": 18392 + }, + { + "epoch": 1.5673625905411164, + "grad_norm": 19.25, + "learning_rate": 2.799967477507426e-07, + "loss": 0.4261, + "step": 18393 + }, + { + "epoch": 1.5674478057094163, + "grad_norm": 12.8125, + "learning_rate": 2.7989077150211985e-07, + "loss": 0.4724, + "step": 18394 + }, + { + "epoch": 1.5675330208777163, + "grad_norm": 15.25, + "learning_rate": 2.7978481278434623e-07, + "loss": 0.7191, + "step": 18395 + }, + { + "epoch": 1.5676182360460162, + "grad_norm": 13.4375, + "learning_rate": 2.7967887159933675e-07, + "loss": 0.5552, + "step": 18396 + }, + { + "epoch": 1.5677034512143162, + "grad_norm": 13.75, + "learning_rate": 2.7957294794900614e-07, + "loss": 0.3924, + "step": 18397 + }, + { + "epoch": 1.5677886663826162, + "grad_norm": 16.625, + "learning_rate": 2.7946704183526783e-07, + "loss": 0.7877, + "step": 18398 + }, + { + "epoch": 1.5678738815509161, + "grad_norm": 16.25, + "learning_rate": 2.793611532600363e-07, + "loss": 0.7158, + "step": 18399 + }, + { + "epoch": 1.567959096719216, + "grad_norm": 20.5, + "learning_rate": 2.7925528222522454e-07, + "loss": 0.848, + "step": 18400 + }, + { + "epoch": 1.568044311887516, + "grad_norm": 21.5, + "learning_rate": 2.7914942873274634e-07, + "loss": 0.8966, + "step": 18401 + }, + { + "epoch": 1.568129527055816, + "grad_norm": 15.625, + "learning_rate": 2.790435927845138e-07, + "loss": 0.6607, + "step": 18402 + }, + { + "epoch": 1.568214742224116, + "grad_norm": 15.125, + "learning_rate": 2.7893777438243994e-07, + "loss": 0.5567, + "step": 18403 + }, + { + "epoch": 1.568299957392416, + "grad_norm": 16.125, + "learning_rate": 2.78831973528437e-07, + "loss": 0.4404, + "step": 18404 + }, + { + "epoch": 1.5683851725607159, + "grad_norm": 14.25, + "learning_rate": 2.787261902244172e-07, + "loss": 0.4489, + "step": 18405 + }, + { + "epoch": 1.5684703877290158, + "grad_norm": 13.875, + "learning_rate": 2.786204244722915e-07, + "loss": 0.7203, + "step": 18406 + }, + { + "epoch": 1.5685556028973158, + "grad_norm": 16.125, + "learning_rate": 2.7851467627397153e-07, + "loss": 0.5947, + "step": 18407 + }, + { + "epoch": 1.5686408180656157, + "grad_norm": 14.4375, + "learning_rate": 2.784089456313686e-07, + "loss": 0.3658, + "step": 18408 + }, + { + "epoch": 1.5687260332339157, + "grad_norm": 18.5, + "learning_rate": 2.7830323254639276e-07, + "loss": 0.6681, + "step": 18409 + }, + { + "epoch": 1.5688112484022156, + "grad_norm": 13.375, + "learning_rate": 2.7819753702095486e-07, + "loss": 0.5404, + "step": 18410 + }, + { + "epoch": 1.5688964635705156, + "grad_norm": 17.875, + "learning_rate": 2.780918590569644e-07, + "loss": 0.964, + "step": 18411 + }, + { + "epoch": 1.5689816787388156, + "grad_norm": 15.1875, + "learning_rate": 2.7798619865633136e-07, + "loss": 0.6041, + "step": 18412 + }, + { + "epoch": 1.5690668939071155, + "grad_norm": 13.6875, + "learning_rate": 2.778805558209652e-07, + "loss": 0.5975, + "step": 18413 + }, + { + "epoch": 1.5691521090754155, + "grad_norm": 17.375, + "learning_rate": 2.777749305527752e-07, + "loss": 0.4763, + "step": 18414 + }, + { + "epoch": 1.5692373242437154, + "grad_norm": 11.125, + "learning_rate": 2.776693228536696e-07, + "loss": 0.4782, + "step": 18415 + }, + { + "epoch": 1.5693225394120154, + "grad_norm": 16.5, + "learning_rate": 2.775637327255573e-07, + "loss": 0.6821, + "step": 18416 + }, + { + "epoch": 1.5694077545803153, + "grad_norm": 19.0, + "learning_rate": 2.7745816017034605e-07, + "loss": 0.7034, + "step": 18417 + }, + { + "epoch": 1.5694929697486153, + "grad_norm": 16.25, + "learning_rate": 2.77352605189944e-07, + "loss": 0.5299, + "step": 18418 + }, + { + "epoch": 1.5695781849169153, + "grad_norm": 12.5625, + "learning_rate": 2.772470677862582e-07, + "loss": 0.3171, + "step": 18419 + }, + { + "epoch": 1.5696634000852152, + "grad_norm": 16.75, + "learning_rate": 2.7714154796119595e-07, + "loss": 0.6154, + "step": 18420 + }, + { + "epoch": 1.5697486152535152, + "grad_norm": 16.625, + "learning_rate": 2.770360457166643e-07, + "loss": 0.5214, + "step": 18421 + }, + { + "epoch": 1.5698338304218151, + "grad_norm": 15.3125, + "learning_rate": 2.7693056105457e-07, + "loss": 0.4467, + "step": 18422 + }, + { + "epoch": 1.569919045590115, + "grad_norm": 13.3125, + "learning_rate": 2.768250939768187e-07, + "loss": 0.4451, + "step": 18423 + }, + { + "epoch": 1.570004260758415, + "grad_norm": 15.1875, + "learning_rate": 2.767196444853168e-07, + "loss": 0.4872, + "step": 18424 + }, + { + "epoch": 1.570089475926715, + "grad_norm": 15.625, + "learning_rate": 2.766142125819693e-07, + "loss": 0.5646, + "step": 18425 + }, + { + "epoch": 1.570174691095015, + "grad_norm": 14.875, + "learning_rate": 2.765087982686819e-07, + "loss": 0.6746, + "step": 18426 + }, + { + "epoch": 1.570259906263315, + "grad_norm": 22.0, + "learning_rate": 2.7640340154735964e-07, + "loss": 0.5188, + "step": 18427 + }, + { + "epoch": 1.5703451214316149, + "grad_norm": 16.5, + "learning_rate": 2.762980224199066e-07, + "loss": 0.5773, + "step": 18428 + }, + { + "epoch": 1.5704303365999148, + "grad_norm": 16.5, + "learning_rate": 2.761926608882276e-07, + "loss": 0.6369, + "step": 18429 + }, + { + "epoch": 1.5705155517682148, + "grad_norm": 15.4375, + "learning_rate": 2.7608731695422664e-07, + "loss": 0.6914, + "step": 18430 + }, + { + "epoch": 1.5706007669365147, + "grad_norm": 12.9375, + "learning_rate": 2.759819906198069e-07, + "loss": 0.3968, + "step": 18431 + }, + { + "epoch": 1.5706859821048147, + "grad_norm": 13.125, + "learning_rate": 2.7587668188687244e-07, + "loss": 0.3522, + "step": 18432 + }, + { + "epoch": 1.5707711972731146, + "grad_norm": 16.5, + "learning_rate": 2.757713907573255e-07, + "loss": 0.4076, + "step": 18433 + }, + { + "epoch": 1.5708564124414146, + "grad_norm": 12.5625, + "learning_rate": 2.756661172330692e-07, + "loss": 0.421, + "step": 18434 + }, + { + "epoch": 1.5709416276097146, + "grad_norm": 14.5625, + "learning_rate": 2.755608613160063e-07, + "loss": 0.3806, + "step": 18435 + }, + { + "epoch": 1.5710268427780145, + "grad_norm": 16.375, + "learning_rate": 2.754556230080381e-07, + "loss": 0.7098, + "step": 18436 + }, + { + "epoch": 1.5711120579463145, + "grad_norm": 12.5625, + "learning_rate": 2.753504023110667e-07, + "loss": 0.3752, + "step": 18437 + }, + { + "epoch": 1.5711972731146144, + "grad_norm": 14.9375, + "learning_rate": 2.7524519922699375e-07, + "loss": 0.5133, + "step": 18438 + }, + { + "epoch": 1.5712824882829144, + "grad_norm": 16.75, + "learning_rate": 2.7514001375772064e-07, + "loss": 0.6404, + "step": 18439 + }, + { + "epoch": 1.5713677034512143, + "grad_norm": 18.75, + "learning_rate": 2.750348459051476e-07, + "loss": 0.773, + "step": 18440 + }, + { + "epoch": 1.5714529186195143, + "grad_norm": 12.625, + "learning_rate": 2.7492969567117514e-07, + "loss": 0.5183, + "step": 18441 + }, + { + "epoch": 1.5715381337878143, + "grad_norm": 14.25, + "learning_rate": 2.7482456305770344e-07, + "loss": 0.6035, + "step": 18442 + }, + { + "epoch": 1.5716233489561142, + "grad_norm": 18.0, + "learning_rate": 2.7471944806663266e-07, + "loss": 0.9852, + "step": 18443 + }, + { + "epoch": 1.5717085641244142, + "grad_norm": 15.6875, + "learning_rate": 2.7461435069986235e-07, + "loss": 0.684, + "step": 18444 + }, + { + "epoch": 1.5717937792927141, + "grad_norm": 13.8125, + "learning_rate": 2.7450927095929127e-07, + "loss": 0.3859, + "step": 18445 + }, + { + "epoch": 1.571878994461014, + "grad_norm": 25.25, + "learning_rate": 2.7440420884681863e-07, + "loss": 0.4881, + "step": 18446 + }, + { + "epoch": 1.571964209629314, + "grad_norm": 20.125, + "learning_rate": 2.742991643643432e-07, + "loss": 0.7466, + "step": 18447 + }, + { + "epoch": 1.572049424797614, + "grad_norm": 12.6875, + "learning_rate": 2.7419413751376314e-07, + "loss": 0.3455, + "step": 18448 + }, + { + "epoch": 1.572134639965914, + "grad_norm": 17.5, + "learning_rate": 2.7408912829697584e-07, + "loss": 0.6424, + "step": 18449 + }, + { + "epoch": 1.572219855134214, + "grad_norm": 18.25, + "learning_rate": 2.739841367158795e-07, + "loss": 0.893, + "step": 18450 + }, + { + "epoch": 1.5723050703025139, + "grad_norm": 11.0, + "learning_rate": 2.7387916277237106e-07, + "loss": 0.3577, + "step": 18451 + }, + { + "epoch": 1.5723902854708138, + "grad_norm": 14.5625, + "learning_rate": 2.7377420646834824e-07, + "loss": 0.5559, + "step": 18452 + }, + { + "epoch": 1.5724755006391138, + "grad_norm": 17.625, + "learning_rate": 2.7366926780570666e-07, + "loss": 0.9137, + "step": 18453 + }, + { + "epoch": 1.5725607158074137, + "grad_norm": 15.5, + "learning_rate": 2.735643467863433e-07, + "loss": 0.6162, + "step": 18454 + }, + { + "epoch": 1.5726459309757137, + "grad_norm": 19.5, + "learning_rate": 2.73459443412154e-07, + "loss": 0.788, + "step": 18455 + }, + { + "epoch": 1.5727311461440137, + "grad_norm": 20.5, + "learning_rate": 2.7335455768503494e-07, + "loss": 0.7222, + "step": 18456 + }, + { + "epoch": 1.5728163613123136, + "grad_norm": 29.75, + "learning_rate": 2.732496896068812e-07, + "loss": 0.6256, + "step": 18457 + }, + { + "epoch": 1.5729015764806136, + "grad_norm": 38.75, + "learning_rate": 2.7314483917958724e-07, + "loss": 1.3189, + "step": 18458 + }, + { + "epoch": 1.5729867916489135, + "grad_norm": 16.25, + "learning_rate": 2.7304000640504855e-07, + "loss": 0.6024, + "step": 18459 + }, + { + "epoch": 1.5730720068172135, + "grad_norm": 15.5625, + "learning_rate": 2.729351912851595e-07, + "loss": 0.5381, + "step": 18460 + }, + { + "epoch": 1.5731572219855134, + "grad_norm": 23.75, + "learning_rate": 2.728303938218138e-07, + "loss": 0.7995, + "step": 18461 + }, + { + "epoch": 1.5732424371538134, + "grad_norm": 12.0625, + "learning_rate": 2.7272561401690557e-07, + "loss": 0.3877, + "step": 18462 + }, + { + "epoch": 1.5733276523221134, + "grad_norm": 12.0625, + "learning_rate": 2.726208518723282e-07, + "loss": 0.3317, + "step": 18463 + }, + { + "epoch": 1.5734128674904133, + "grad_norm": 11.0, + "learning_rate": 2.7251610738997515e-07, + "loss": 0.4387, + "step": 18464 + }, + { + "epoch": 1.5734980826587133, + "grad_norm": 18.25, + "learning_rate": 2.724113805717389e-07, + "loss": 0.9723, + "step": 18465 + }, + { + "epoch": 1.5735832978270132, + "grad_norm": 15.1875, + "learning_rate": 2.723066714195119e-07, + "loss": 0.3894, + "step": 18466 + }, + { + "epoch": 1.5736685129953132, + "grad_norm": 18.125, + "learning_rate": 2.722019799351863e-07, + "loss": 0.4327, + "step": 18467 + }, + { + "epoch": 1.5737537281636131, + "grad_norm": 10.0, + "learning_rate": 2.720973061206543e-07, + "loss": 0.3784, + "step": 18468 + }, + { + "epoch": 1.573838943331913, + "grad_norm": 12.25, + "learning_rate": 2.7199264997780765e-07, + "loss": 0.3998, + "step": 18469 + }, + { + "epoch": 1.573924158500213, + "grad_norm": 19.875, + "learning_rate": 2.718880115085369e-07, + "loss": 0.7107, + "step": 18470 + }, + { + "epoch": 1.574009373668513, + "grad_norm": 16.125, + "learning_rate": 2.7178339071473337e-07, + "loss": 0.8626, + "step": 18471 + }, + { + "epoch": 1.574094588836813, + "grad_norm": 14.8125, + "learning_rate": 2.7167878759828764e-07, + "loss": 0.7088, + "step": 18472 + }, + { + "epoch": 1.574179804005113, + "grad_norm": 20.875, + "learning_rate": 2.7157420216109035e-07, + "loss": 0.9969, + "step": 18473 + }, + { + "epoch": 1.5742650191734129, + "grad_norm": 10.5625, + "learning_rate": 2.714696344050312e-07, + "loss": 0.3811, + "step": 18474 + }, + { + "epoch": 1.5743502343417128, + "grad_norm": 16.125, + "learning_rate": 2.7136508433199926e-07, + "loss": 0.7363, + "step": 18475 + }, + { + "epoch": 1.5744354495100128, + "grad_norm": 12.0, + "learning_rate": 2.7126055194388433e-07, + "loss": 0.3281, + "step": 18476 + }, + { + "epoch": 1.5745206646783128, + "grad_norm": 14.5, + "learning_rate": 2.7115603724257586e-07, + "loss": 0.3907, + "step": 18477 + }, + { + "epoch": 1.5746058798466127, + "grad_norm": 16.375, + "learning_rate": 2.7105154022996167e-07, + "loss": 0.6436, + "step": 18478 + }, + { + "epoch": 1.5746910950149127, + "grad_norm": 13.625, + "learning_rate": 2.7094706090793065e-07, + "loss": 0.6307, + "step": 18479 + }, + { + "epoch": 1.5747763101832126, + "grad_norm": 21.75, + "learning_rate": 2.708425992783707e-07, + "loss": 0.7852, + "step": 18480 + }, + { + "epoch": 1.5748615253515126, + "grad_norm": 12.1875, + "learning_rate": 2.7073815534317e-07, + "loss": 0.4404, + "step": 18481 + }, + { + "epoch": 1.5749467405198125, + "grad_norm": 18.125, + "learning_rate": 2.706337291042155e-07, + "loss": 0.6337, + "step": 18482 + }, + { + "epoch": 1.5750319556881125, + "grad_norm": 18.75, + "learning_rate": 2.7052932056339396e-07, + "loss": 0.9609, + "step": 18483 + }, + { + "epoch": 1.5751171708564125, + "grad_norm": 13.625, + "learning_rate": 2.7042492972259263e-07, + "loss": 0.589, + "step": 18484 + }, + { + "epoch": 1.5752023860247124, + "grad_norm": 13.1875, + "learning_rate": 2.7032055658369777e-07, + "loss": 0.5916, + "step": 18485 + }, + { + "epoch": 1.5752876011930124, + "grad_norm": 12.4375, + "learning_rate": 2.702162011485961e-07, + "loss": 0.2893, + "step": 18486 + }, + { + "epoch": 1.5753728163613123, + "grad_norm": 13.125, + "learning_rate": 2.7011186341917237e-07, + "loss": 0.4326, + "step": 18487 + }, + { + "epoch": 1.5754580315296123, + "grad_norm": 19.125, + "learning_rate": 2.7000754339731277e-07, + "loss": 0.6215, + "step": 18488 + }, + { + "epoch": 1.5755432466979122, + "grad_norm": 11.9375, + "learning_rate": 2.6990324108490256e-07, + "loss": 0.5567, + "step": 18489 + }, + { + "epoch": 1.5756284618662122, + "grad_norm": 13.0625, + "learning_rate": 2.6979895648382603e-07, + "loss": 0.5002, + "step": 18490 + }, + { + "epoch": 1.5757136770345122, + "grad_norm": 14.875, + "learning_rate": 2.6969468959596845e-07, + "loss": 0.4885, + "step": 18491 + }, + { + "epoch": 1.575798892202812, + "grad_norm": 11.125, + "learning_rate": 2.6959044042321316e-07, + "loss": 0.4522, + "step": 18492 + }, + { + "epoch": 1.575884107371112, + "grad_norm": 22.0, + "learning_rate": 2.694862089674445e-07, + "loss": 0.6917, + "step": 18493 + }, + { + "epoch": 1.575969322539412, + "grad_norm": 11.75, + "learning_rate": 2.693819952305464e-07, + "loss": 0.3125, + "step": 18494 + }, + { + "epoch": 1.576054537707712, + "grad_norm": 12.375, + "learning_rate": 2.6927779921440145e-07, + "loss": 0.5023, + "step": 18495 + }, + { + "epoch": 1.576139752876012, + "grad_norm": 15.1875, + "learning_rate": 2.691736209208928e-07, + "loss": 0.3659, + "step": 18496 + }, + { + "epoch": 1.576224968044312, + "grad_norm": 14.625, + "learning_rate": 2.690694603519034e-07, + "loss": 0.5715, + "step": 18497 + }, + { + "epoch": 1.5763101832126118, + "grad_norm": 17.0, + "learning_rate": 2.68965317509315e-07, + "loss": 0.7785, + "step": 18498 + }, + { + "epoch": 1.5763953983809118, + "grad_norm": 11.625, + "learning_rate": 2.688611923950102e-07, + "loss": 0.4601, + "step": 18499 + }, + { + "epoch": 1.5764806135492118, + "grad_norm": 19.75, + "learning_rate": 2.6875708501086976e-07, + "loss": 0.8054, + "step": 18500 + }, + { + "epoch": 1.5765658287175117, + "grad_norm": 12.9375, + "learning_rate": 2.686529953587756e-07, + "loss": 0.4458, + "step": 18501 + }, + { + "epoch": 1.5766510438858117, + "grad_norm": 11.3125, + "learning_rate": 2.685489234406087e-07, + "loss": 0.3875, + "step": 18502 + }, + { + "epoch": 1.5767362590541116, + "grad_norm": 12.875, + "learning_rate": 2.684448692582499e-07, + "loss": 0.589, + "step": 18503 + }, + { + "epoch": 1.5768214742224116, + "grad_norm": 16.375, + "learning_rate": 2.6834083281357904e-07, + "loss": 0.518, + "step": 18504 + }, + { + "epoch": 1.5769066893907115, + "grad_norm": 18.875, + "learning_rate": 2.682368141084769e-07, + "loss": 0.691, + "step": 18505 + }, + { + "epoch": 1.5769919045590115, + "grad_norm": 31.0, + "learning_rate": 2.681328131448223e-07, + "loss": 0.6377, + "step": 18506 + }, + { + "epoch": 1.5770771197273115, + "grad_norm": 17.5, + "learning_rate": 2.680288299244956e-07, + "loss": 0.7858, + "step": 18507 + }, + { + "epoch": 1.5771623348956114, + "grad_norm": 18.0, + "learning_rate": 2.679248644493748e-07, + "loss": 0.732, + "step": 18508 + }, + { + "epoch": 1.5772475500639114, + "grad_norm": 12.6875, + "learning_rate": 2.6782091672133957e-07, + "loss": 0.4285, + "step": 18509 + }, + { + "epoch": 1.5773327652322113, + "grad_norm": 14.9375, + "learning_rate": 2.677169867422678e-07, + "loss": 0.4973, + "step": 18510 + }, + { + "epoch": 1.5774179804005113, + "grad_norm": 18.375, + "learning_rate": 2.6761307451403833e-07, + "loss": 0.6311, + "step": 18511 + }, + { + "epoch": 1.5775031955688112, + "grad_norm": 16.875, + "learning_rate": 2.6750918003852796e-07, + "loss": 0.7236, + "step": 18512 + }, + { + "epoch": 1.5775884107371112, + "grad_norm": 18.25, + "learning_rate": 2.6740530331761506e-07, + "loss": 0.7436, + "step": 18513 + }, + { + "epoch": 1.5776736259054112, + "grad_norm": 11.1875, + "learning_rate": 2.6730144435317617e-07, + "loss": 0.2203, + "step": 18514 + }, + { + "epoch": 1.5777588410737111, + "grad_norm": 14.4375, + "learning_rate": 2.671976031470883e-07, + "loss": 0.6666, + "step": 18515 + }, + { + "epoch": 1.577844056242011, + "grad_norm": 15.5, + "learning_rate": 2.670937797012284e-07, + "loss": 0.6546, + "step": 18516 + }, + { + "epoch": 1.577929271410311, + "grad_norm": 18.25, + "learning_rate": 2.6698997401747196e-07, + "loss": 0.6795, + "step": 18517 + }, + { + "epoch": 1.578014486578611, + "grad_norm": 15.1875, + "learning_rate": 2.668861860976951e-07, + "loss": 0.5499, + "step": 18518 + }, + { + "epoch": 1.578099701746911, + "grad_norm": 16.875, + "learning_rate": 2.6678241594377346e-07, + "loss": 0.6227, + "step": 18519 + }, + { + "epoch": 1.578184916915211, + "grad_norm": 17.0, + "learning_rate": 2.666786635575827e-07, + "loss": 0.8236, + "step": 18520 + }, + { + "epoch": 1.5782701320835109, + "grad_norm": 13.1875, + "learning_rate": 2.665749289409972e-07, + "loss": 0.49, + "step": 18521 + }, + { + "epoch": 1.5783553472518108, + "grad_norm": 15.5, + "learning_rate": 2.6647121209589136e-07, + "loss": 0.6541, + "step": 18522 + }, + { + "epoch": 1.5784405624201108, + "grad_norm": 19.25, + "learning_rate": 2.663675130241397e-07, + "loss": 0.8065, + "step": 18523 + }, + { + "epoch": 1.5785257775884107, + "grad_norm": 11.9375, + "learning_rate": 2.662638317276166e-07, + "loss": 0.3863, + "step": 18524 + }, + { + "epoch": 1.5786109927567107, + "grad_norm": 18.0, + "learning_rate": 2.661601682081949e-07, + "loss": 0.5858, + "step": 18525 + }, + { + "epoch": 1.5786962079250106, + "grad_norm": 10.8125, + "learning_rate": 2.660565224677482e-07, + "loss": 0.2623, + "step": 18526 + }, + { + "epoch": 1.5787814230933106, + "grad_norm": 12.0, + "learning_rate": 2.6595289450814966e-07, + "loss": 0.3062, + "step": 18527 + }, + { + "epoch": 1.5788666382616106, + "grad_norm": 15.8125, + "learning_rate": 2.6584928433127227e-07, + "loss": 0.7284, + "step": 18528 + }, + { + "epoch": 1.5789518534299105, + "grad_norm": 18.625, + "learning_rate": 2.657456919389878e-07, + "loss": 0.6329, + "step": 18529 + }, + { + "epoch": 1.5790370685982105, + "grad_norm": 10.75, + "learning_rate": 2.6564211733316817e-07, + "loss": 0.247, + "step": 18530 + }, + { + "epoch": 1.5791222837665104, + "grad_norm": 14.0, + "learning_rate": 2.655385605156853e-07, + "loss": 0.4894, + "step": 18531 + }, + { + "epoch": 1.5792074989348104, + "grad_norm": 17.0, + "learning_rate": 2.6543502148841065e-07, + "loss": 0.8102, + "step": 18532 + }, + { + "epoch": 1.5792927141031103, + "grad_norm": 14.9375, + "learning_rate": 2.6533150025321547e-07, + "loss": 0.7557, + "step": 18533 + }, + { + "epoch": 1.5793779292714103, + "grad_norm": 14.0, + "learning_rate": 2.652279968119699e-07, + "loss": 0.5465, + "step": 18534 + }, + { + "epoch": 1.5794631444397103, + "grad_norm": 14.8125, + "learning_rate": 2.651245111665449e-07, + "loss": 0.5156, + "step": 18535 + }, + { + "epoch": 1.5795483596080102, + "grad_norm": 16.75, + "learning_rate": 2.6502104331881007e-07, + "loss": 0.5967, + "step": 18536 + }, + { + "epoch": 1.5796335747763102, + "grad_norm": 11.9375, + "learning_rate": 2.649175932706363e-07, + "loss": 0.4047, + "step": 18537 + }, + { + "epoch": 1.5797187899446101, + "grad_norm": 16.75, + "learning_rate": 2.6481416102389147e-07, + "loss": 0.4963, + "step": 18538 + }, + { + "epoch": 1.57980400511291, + "grad_norm": 14.75, + "learning_rate": 2.6471074658044537e-07, + "loss": 0.49, + "step": 18539 + }, + { + "epoch": 1.57988922028121, + "grad_norm": 18.75, + "learning_rate": 2.64607349942167e-07, + "loss": 0.5962, + "step": 18540 + }, + { + "epoch": 1.57997443544951, + "grad_norm": 22.125, + "learning_rate": 2.6450397111092497e-07, + "loss": 0.5567, + "step": 18541 + }, + { + "epoch": 1.58005965061781, + "grad_norm": 19.875, + "learning_rate": 2.644006100885868e-07, + "loss": 0.7394, + "step": 18542 + }, + { + "epoch": 1.58014486578611, + "grad_norm": 13.25, + "learning_rate": 2.642972668770208e-07, + "loss": 0.5319, + "step": 18543 + }, + { + "epoch": 1.5802300809544099, + "grad_norm": 11.3125, + "learning_rate": 2.6419394147809434e-07, + "loss": 1.1369, + "step": 18544 + }, + { + "epoch": 1.5803152961227098, + "grad_norm": 13.75, + "learning_rate": 2.6409063389367497e-07, + "loss": 0.5735, + "step": 18545 + }, + { + "epoch": 1.5804005112910098, + "grad_norm": 19.375, + "learning_rate": 2.6398734412562925e-07, + "loss": 0.6798, + "step": 18546 + }, + { + "epoch": 1.5804857264593097, + "grad_norm": 12.3125, + "learning_rate": 2.6388407217582345e-07, + "loss": 0.3749, + "step": 18547 + }, + { + "epoch": 1.5805709416276097, + "grad_norm": 17.5, + "learning_rate": 2.6378081804612407e-07, + "loss": 0.557, + "step": 18548 + }, + { + "epoch": 1.5806561567959097, + "grad_norm": 12.4375, + "learning_rate": 2.63677581738397e-07, + "loss": 0.3307, + "step": 18549 + }, + { + "epoch": 1.5807413719642096, + "grad_norm": 35.0, + "learning_rate": 2.6357436325450817e-07, + "loss": 0.8709, + "step": 18550 + }, + { + "epoch": 1.5808265871325096, + "grad_norm": 16.25, + "learning_rate": 2.6347116259632226e-07, + "loss": 0.6508, + "step": 18551 + }, + { + "epoch": 1.5809118023008095, + "grad_norm": 14.5625, + "learning_rate": 2.633679797657043e-07, + "loss": 0.3805, + "step": 18552 + }, + { + "epoch": 1.5809970174691095, + "grad_norm": 15.0625, + "learning_rate": 2.632648147645193e-07, + "loss": 0.502, + "step": 18553 + }, + { + "epoch": 1.5810822326374094, + "grad_norm": 11.8125, + "learning_rate": 2.631616675946319e-07, + "loss": 0.291, + "step": 18554 + }, + { + "epoch": 1.5811674478057094, + "grad_norm": 17.25, + "learning_rate": 2.630585382579048e-07, + "loss": 0.7447, + "step": 18555 + }, + { + "epoch": 1.5812526629740093, + "grad_norm": 15.0, + "learning_rate": 2.629554267562025e-07, + "loss": 0.6566, + "step": 18556 + }, + { + "epoch": 1.5813378781423093, + "grad_norm": 16.5, + "learning_rate": 2.6285233309138807e-07, + "loss": 0.7805, + "step": 18557 + }, + { + "epoch": 1.5814230933106093, + "grad_norm": 15.9375, + "learning_rate": 2.627492572653251e-07, + "loss": 0.7582, + "step": 18558 + }, + { + "epoch": 1.5815083084789092, + "grad_norm": 14.5625, + "learning_rate": 2.6264619927987527e-07, + "loss": 0.5567, + "step": 18559 + }, + { + "epoch": 1.5815935236472092, + "grad_norm": 16.5, + "learning_rate": 2.625431591369017e-07, + "loss": 0.8404, + "step": 18560 + }, + { + "epoch": 1.5816787388155091, + "grad_norm": 16.75, + "learning_rate": 2.6244013683826603e-07, + "loss": 0.7119, + "step": 18561 + }, + { + "epoch": 1.581763953983809, + "grad_norm": 15.75, + "learning_rate": 2.623371323858305e-07, + "loss": 0.4319, + "step": 18562 + }, + { + "epoch": 1.581849169152109, + "grad_norm": 13.5, + "learning_rate": 2.6223414578145633e-07, + "loss": 0.4016, + "step": 18563 + }, + { + "epoch": 1.581934384320409, + "grad_norm": 11.6875, + "learning_rate": 2.6213117702700395e-07, + "loss": 0.3826, + "step": 18564 + }, + { + "epoch": 1.582019599488709, + "grad_norm": 11.4375, + "learning_rate": 2.6202822612433453e-07, + "loss": 0.3079, + "step": 18565 + }, + { + "epoch": 1.582104814657009, + "grad_norm": 17.0, + "learning_rate": 2.619252930753087e-07, + "loss": 0.6956, + "step": 18566 + }, + { + "epoch": 1.5821900298253089, + "grad_norm": 17.75, + "learning_rate": 2.618223778817866e-07, + "loss": 0.6405, + "step": 18567 + }, + { + "epoch": 1.5822752449936088, + "grad_norm": 14.3125, + "learning_rate": 2.617194805456276e-07, + "loss": 0.5696, + "step": 18568 + }, + { + "epoch": 1.5823604601619088, + "grad_norm": 15.375, + "learning_rate": 2.616166010686913e-07, + "loss": 0.6368, + "step": 18569 + }, + { + "epoch": 1.5824456753302087, + "grad_norm": 14.5, + "learning_rate": 2.6151373945283734e-07, + "loss": 0.476, + "step": 18570 + }, + { + "epoch": 1.5825308904985087, + "grad_norm": 15.625, + "learning_rate": 2.614108956999241e-07, + "loss": 0.5464, + "step": 18571 + }, + { + "epoch": 1.5826161056668087, + "grad_norm": 15.75, + "learning_rate": 2.6130806981180965e-07, + "loss": 0.5787, + "step": 18572 + }, + { + "epoch": 1.5827013208351086, + "grad_norm": 13.0625, + "learning_rate": 2.6120526179035263e-07, + "loss": 0.3786, + "step": 18573 + }, + { + "epoch": 1.5827865360034086, + "grad_norm": 15.75, + "learning_rate": 2.611024716374108e-07, + "loss": 0.587, + "step": 18574 + }, + { + "epoch": 1.5828717511717085, + "grad_norm": 14.5625, + "learning_rate": 2.6099969935484205e-07, + "loss": 0.5273, + "step": 18575 + }, + { + "epoch": 1.5829569663400085, + "grad_norm": 19.0, + "learning_rate": 2.60896944944503e-07, + "loss": 0.5657, + "step": 18576 + }, + { + "epoch": 1.5830421815083084, + "grad_norm": 16.625, + "learning_rate": 2.6079420840825073e-07, + "loss": 0.655, + "step": 18577 + }, + { + "epoch": 1.5831273966766084, + "grad_norm": 15.4375, + "learning_rate": 2.606914897479422e-07, + "loss": 0.691, + "step": 18578 + }, + { + "epoch": 1.5832126118449084, + "grad_norm": 26.875, + "learning_rate": 2.6058878896543294e-07, + "loss": 0.8032, + "step": 18579 + }, + { + "epoch": 1.5832978270132083, + "grad_norm": 12.375, + "learning_rate": 2.6048610606257946e-07, + "loss": 0.3089, + "step": 18580 + }, + { + "epoch": 1.5833830421815083, + "grad_norm": 13.875, + "learning_rate": 2.6038344104123673e-07, + "loss": 0.4432, + "step": 18581 + }, + { + "epoch": 1.5834682573498082, + "grad_norm": 25.5, + "learning_rate": 2.602807939032605e-07, + "loss": 0.8797, + "step": 18582 + }, + { + "epoch": 1.5835534725181082, + "grad_norm": 22.0, + "learning_rate": 2.601781646505054e-07, + "loss": 0.7887, + "step": 18583 + }, + { + "epoch": 1.5836386876864081, + "grad_norm": 13.6875, + "learning_rate": 2.6007555328482663e-07, + "loss": 0.2908, + "step": 18584 + }, + { + "epoch": 1.583723902854708, + "grad_norm": 19.5, + "learning_rate": 2.5997295980807775e-07, + "loss": 0.7696, + "step": 18585 + }, + { + "epoch": 1.583809118023008, + "grad_norm": 19.75, + "learning_rate": 2.598703842221133e-07, + "loss": 0.6926, + "step": 18586 + }, + { + "epoch": 1.583894333191308, + "grad_norm": 15.0, + "learning_rate": 2.597678265287865e-07, + "loss": 0.5404, + "step": 18587 + }, + { + "epoch": 1.583979548359608, + "grad_norm": 22.0, + "learning_rate": 2.5966528672995097e-07, + "loss": 0.3412, + "step": 18588 + }, + { + "epoch": 1.584064763527908, + "grad_norm": 14.4375, + "learning_rate": 2.595627648274593e-07, + "loss": 0.7037, + "step": 18589 + }, + { + "epoch": 1.5841499786962079, + "grad_norm": 15.0, + "learning_rate": 2.594602608231643e-07, + "loss": 0.6279, + "step": 18590 + }, + { + "epoch": 1.5842351938645078, + "grad_norm": 16.875, + "learning_rate": 2.5935777471891867e-07, + "loss": 0.947, + "step": 18591 + }, + { + "epoch": 1.5843204090328078, + "grad_norm": 11.6875, + "learning_rate": 2.592553065165744e-07, + "loss": 0.3969, + "step": 18592 + }, + { + "epoch": 1.5844056242011078, + "grad_norm": 16.0, + "learning_rate": 2.591528562179828e-07, + "loss": 0.564, + "step": 18593 + }, + { + "epoch": 1.5844908393694077, + "grad_norm": 22.25, + "learning_rate": 2.5905042382499567e-07, + "loss": 0.9315, + "step": 18594 + }, + { + "epoch": 1.5845760545377077, + "grad_norm": 18.75, + "learning_rate": 2.5894800933946365e-07, + "loss": 0.5244, + "step": 18595 + }, + { + "epoch": 1.5846612697060076, + "grad_norm": 21.125, + "learning_rate": 2.588456127632376e-07, + "loss": 0.4891, + "step": 18596 + }, + { + "epoch": 1.5847464848743076, + "grad_norm": 18.5, + "learning_rate": 2.5874323409816824e-07, + "loss": 0.6796, + "step": 18597 + }, + { + "epoch": 1.5848317000426075, + "grad_norm": 25.75, + "learning_rate": 2.586408733461053e-07, + "loss": 0.7024, + "step": 18598 + }, + { + "epoch": 1.5849169152109075, + "grad_norm": 14.375, + "learning_rate": 2.5853853050889845e-07, + "loss": 0.58, + "step": 18599 + }, + { + "epoch": 1.5850021303792075, + "grad_norm": 13.5, + "learning_rate": 2.584362055883979e-07, + "loss": 0.5231, + "step": 18600 + }, + { + "epoch": 1.5850873455475074, + "grad_norm": 14.6875, + "learning_rate": 2.583338985864517e-07, + "loss": 0.6363, + "step": 18601 + }, + { + "epoch": 1.5851725607158074, + "grad_norm": 17.5, + "learning_rate": 2.5823160950490963e-07, + "loss": 0.7232, + "step": 18602 + }, + { + "epoch": 1.5852577758841073, + "grad_norm": 14.75, + "learning_rate": 2.5812933834561933e-07, + "loss": 0.5057, + "step": 18603 + }, + { + "epoch": 1.5853429910524073, + "grad_norm": 25.125, + "learning_rate": 2.580270851104291e-07, + "loss": 1.0878, + "step": 18604 + }, + { + "epoch": 1.5854282062207072, + "grad_norm": 18.375, + "learning_rate": 2.5792484980118743e-07, + "loss": 0.8084, + "step": 18605 + }, + { + "epoch": 1.5855134213890072, + "grad_norm": 12.375, + "learning_rate": 2.5782263241974107e-07, + "loss": 0.4712, + "step": 18606 + }, + { + "epoch": 1.5855986365573072, + "grad_norm": 12.5, + "learning_rate": 2.577204329679374e-07, + "loss": 0.2318, + "step": 18607 + }, + { + "epoch": 1.585683851725607, + "grad_norm": 16.5, + "learning_rate": 2.576182514476232e-07, + "loss": 0.6264, + "step": 18608 + }, + { + "epoch": 1.585769066893907, + "grad_norm": 23.25, + "learning_rate": 2.575160878606456e-07, + "loss": 0.8851, + "step": 18609 + }, + { + "epoch": 1.585854282062207, + "grad_norm": 11.875, + "learning_rate": 2.5741394220884993e-07, + "loss": 0.3612, + "step": 18610 + }, + { + "epoch": 1.585939497230507, + "grad_norm": 16.375, + "learning_rate": 2.5731181449408266e-07, + "loss": 0.789, + "step": 18611 + }, + { + "epoch": 1.586024712398807, + "grad_norm": 19.125, + "learning_rate": 2.57209704718189e-07, + "loss": 0.4576, + "step": 18612 + }, + { + "epoch": 1.586109927567107, + "grad_norm": 14.625, + "learning_rate": 2.5710761288301414e-07, + "loss": 0.7679, + "step": 18613 + }, + { + "epoch": 1.5861951427354068, + "grad_norm": 15.4375, + "learning_rate": 2.570055389904036e-07, + "loss": 0.6355, + "step": 18614 + }, + { + "epoch": 1.5862803579037068, + "grad_norm": 17.125, + "learning_rate": 2.569034830422011e-07, + "loss": 0.3793, + "step": 18615 + }, + { + "epoch": 1.5863655730720068, + "grad_norm": 13.5625, + "learning_rate": 2.568014450402513e-07, + "loss": 0.5668, + "step": 18616 + }, + { + "epoch": 1.5864507882403067, + "grad_norm": 16.875, + "learning_rate": 2.566994249863984e-07, + "loss": 0.7687, + "step": 18617 + }, + { + "epoch": 1.5865360034086067, + "grad_norm": 16.625, + "learning_rate": 2.5659742288248554e-07, + "loss": 0.4177, + "step": 18618 + }, + { + "epoch": 1.5866212185769069, + "grad_norm": 14.6875, + "learning_rate": 2.564954387303564e-07, + "loss": 0.4472, + "step": 18619 + }, + { + "epoch": 1.5867064337452068, + "grad_norm": 13.6875, + "learning_rate": 2.5639347253185345e-07, + "loss": 0.5648, + "step": 18620 + }, + { + "epoch": 1.5867916489135068, + "grad_norm": 22.375, + "learning_rate": 2.562915242888196e-07, + "loss": 0.832, + "step": 18621 + }, + { + "epoch": 1.5868768640818067, + "grad_norm": 17.875, + "learning_rate": 2.561895940030976e-07, + "loss": 1.0256, + "step": 18622 + }, + { + "epoch": 1.5869620792501067, + "grad_norm": 14.0625, + "learning_rate": 2.560876816765284e-07, + "loss": 0.4695, + "step": 18623 + }, + { + "epoch": 1.5870472944184066, + "grad_norm": 12.75, + "learning_rate": 2.5598578731095445e-07, + "loss": 0.4743, + "step": 18624 + }, + { + "epoch": 1.5871325095867066, + "grad_norm": 16.125, + "learning_rate": 2.5588391090821687e-07, + "loss": 0.6687, + "step": 18625 + }, + { + "epoch": 1.5872177247550066, + "grad_norm": 13.125, + "learning_rate": 2.557820524701569e-07, + "loss": 0.4836, + "step": 18626 + }, + { + "epoch": 1.5873029399233065, + "grad_norm": 10.125, + "learning_rate": 2.5568021199861515e-07, + "loss": 0.2403, + "step": 18627 + }, + { + "epoch": 1.5873881550916065, + "grad_norm": 17.375, + "learning_rate": 2.555783894954314e-07, + "loss": 0.6275, + "step": 18628 + }, + { + "epoch": 1.5874733702599064, + "grad_norm": 14.0, + "learning_rate": 2.5547658496244606e-07, + "loss": 0.4196, + "step": 18629 + }, + { + "epoch": 1.5875585854282064, + "grad_norm": 13.25, + "learning_rate": 2.553747984014994e-07, + "loss": 0.4438, + "step": 18630 + }, + { + "epoch": 1.5876438005965063, + "grad_norm": 14.375, + "learning_rate": 2.5527302981442987e-07, + "loss": 0.3848, + "step": 18631 + }, + { + "epoch": 1.5877290157648063, + "grad_norm": 19.375, + "learning_rate": 2.5517127920307695e-07, + "loss": 0.7541, + "step": 18632 + }, + { + "epoch": 1.5878142309331063, + "grad_norm": 14.5625, + "learning_rate": 2.550695465692794e-07, + "loss": 0.6279, + "step": 18633 + }, + { + "epoch": 1.5878994461014062, + "grad_norm": 12.0, + "learning_rate": 2.5496783191487603e-07, + "loss": 0.48, + "step": 18634 + }, + { + "epoch": 1.5879846612697062, + "grad_norm": 12.375, + "learning_rate": 2.548661352417045e-07, + "loss": 0.3846, + "step": 18635 + }, + { + "epoch": 1.5880698764380061, + "grad_norm": 16.25, + "learning_rate": 2.547644565516022e-07, + "loss": 0.622, + "step": 18636 + }, + { + "epoch": 1.588155091606306, + "grad_norm": 17.875, + "learning_rate": 2.5466279584640705e-07, + "loss": 0.5017, + "step": 18637 + }, + { + "epoch": 1.588240306774606, + "grad_norm": 20.625, + "learning_rate": 2.54561153127956e-07, + "loss": 0.8189, + "step": 18638 + }, + { + "epoch": 1.588325521942906, + "grad_norm": 15.5, + "learning_rate": 2.544595283980862e-07, + "loss": 0.4061, + "step": 18639 + }, + { + "epoch": 1.588410737111206, + "grad_norm": 15.125, + "learning_rate": 2.5435792165863353e-07, + "loss": 0.6542, + "step": 18640 + }, + { + "epoch": 1.588495952279506, + "grad_norm": 18.375, + "learning_rate": 2.542563329114345e-07, + "loss": 0.5429, + "step": 18641 + }, + { + "epoch": 1.5885811674478059, + "grad_norm": 16.625, + "learning_rate": 2.541547621583247e-07, + "loss": 0.4337, + "step": 18642 + }, + { + "epoch": 1.5886663826161058, + "grad_norm": 13.6875, + "learning_rate": 2.540532094011401e-07, + "loss": 0.3385, + "step": 18643 + }, + { + "epoch": 1.5887515977844058, + "grad_norm": 12.125, + "learning_rate": 2.5395167464171545e-07, + "loss": 0.2758, + "step": 18644 + }, + { + "epoch": 1.5888368129527057, + "grad_norm": 14.625, + "learning_rate": 2.538501578818854e-07, + "loss": 0.7563, + "step": 18645 + }, + { + "epoch": 1.5889220281210057, + "grad_norm": 14.1875, + "learning_rate": 2.537486591234847e-07, + "loss": 0.5784, + "step": 18646 + }, + { + "epoch": 1.5890072432893056, + "grad_norm": 19.125, + "learning_rate": 2.536471783683479e-07, + "loss": 1.0476, + "step": 18647 + }, + { + "epoch": 1.5890924584576056, + "grad_norm": 14.125, + "learning_rate": 2.535457156183081e-07, + "loss": 0.5884, + "step": 18648 + }, + { + "epoch": 1.5891776736259056, + "grad_norm": 13.875, + "learning_rate": 2.534442708751993e-07, + "loss": 0.4893, + "step": 18649 + }, + { + "epoch": 1.5892628887942055, + "grad_norm": 14.875, + "learning_rate": 2.533428441408546e-07, + "loss": 0.7088, + "step": 18650 + }, + { + "epoch": 1.5893481039625055, + "grad_norm": 36.0, + "learning_rate": 2.5324143541710727e-07, + "loss": 0.6284, + "step": 18651 + }, + { + "epoch": 1.5894333191308054, + "grad_norm": 14.4375, + "learning_rate": 2.531400447057895e-07, + "loss": 0.7716, + "step": 18652 + }, + { + "epoch": 1.5895185342991054, + "grad_norm": 19.0, + "learning_rate": 2.530386720087333e-07, + "loss": 0.5941, + "step": 18653 + }, + { + "epoch": 1.5896037494674053, + "grad_norm": 16.75, + "learning_rate": 2.529373173277706e-07, + "loss": 0.3928, + "step": 18654 + }, + { + "epoch": 1.5896889646357053, + "grad_norm": 17.25, + "learning_rate": 2.528359806647333e-07, + "loss": 0.7314, + "step": 18655 + }, + { + "epoch": 1.5897741798040053, + "grad_norm": 27.875, + "learning_rate": 2.5273466202145295e-07, + "loss": 0.7509, + "step": 18656 + }, + { + "epoch": 1.5898593949723052, + "grad_norm": 19.875, + "learning_rate": 2.5263336139975976e-07, + "loss": 0.6531, + "step": 18657 + }, + { + "epoch": 1.5899446101406052, + "grad_norm": 17.25, + "learning_rate": 2.525320788014847e-07, + "loss": 0.6442, + "step": 18658 + }, + { + "epoch": 1.5900298253089051, + "grad_norm": 16.5, + "learning_rate": 2.5243081422845816e-07, + "loss": 0.6333, + "step": 18659 + }, + { + "epoch": 1.590115040477205, + "grad_norm": 13.375, + "learning_rate": 2.523295676825101e-07, + "loss": 0.4523, + "step": 18660 + }, + { + "epoch": 1.590200255645505, + "grad_norm": 14.5, + "learning_rate": 2.5222833916546967e-07, + "loss": 0.8157, + "step": 18661 + }, + { + "epoch": 1.590285470813805, + "grad_norm": 12.8125, + "learning_rate": 2.521271286791663e-07, + "loss": 0.2267, + "step": 18662 + }, + { + "epoch": 1.590370685982105, + "grad_norm": 16.375, + "learning_rate": 2.5202593622542934e-07, + "loss": 0.7226, + "step": 18663 + }, + { + "epoch": 1.590455901150405, + "grad_norm": 23.625, + "learning_rate": 2.5192476180608744e-07, + "loss": 0.8938, + "step": 18664 + }, + { + "epoch": 1.5905411163187049, + "grad_norm": 14.9375, + "learning_rate": 2.518236054229685e-07, + "loss": 0.4867, + "step": 18665 + }, + { + "epoch": 1.5906263314870048, + "grad_norm": 20.0, + "learning_rate": 2.5172246707790076e-07, + "loss": 0.8384, + "step": 18666 + }, + { + "epoch": 1.5907115466553048, + "grad_norm": 16.625, + "learning_rate": 2.5162134677271214e-07, + "loss": 0.7599, + "step": 18667 + }, + { + "epoch": 1.5907967618236047, + "grad_norm": 13.0, + "learning_rate": 2.515202445092295e-07, + "loss": 0.4949, + "step": 18668 + }, + { + "epoch": 1.5908819769919047, + "grad_norm": 16.25, + "learning_rate": 2.514191602892804e-07, + "loss": 0.6829, + "step": 18669 + }, + { + "epoch": 1.5909671921602047, + "grad_norm": 11.25, + "learning_rate": 2.513180941146909e-07, + "loss": 0.3216, + "step": 18670 + }, + { + "epoch": 1.5910524073285046, + "grad_norm": 13.8125, + "learning_rate": 2.512170459872876e-07, + "loss": 0.6503, + "step": 18671 + }, + { + "epoch": 1.5911376224968046, + "grad_norm": 13.25, + "learning_rate": 2.511160159088967e-07, + "loss": 0.3412, + "step": 18672 + }, + { + "epoch": 1.5912228376651045, + "grad_norm": 17.875, + "learning_rate": 2.5101500388134415e-07, + "loss": 0.7667, + "step": 18673 + }, + { + "epoch": 1.5913080528334045, + "grad_norm": 20.75, + "learning_rate": 2.509140099064547e-07, + "loss": 0.9123, + "step": 18674 + }, + { + "epoch": 1.5913932680017044, + "grad_norm": 14.3125, + "learning_rate": 2.5081303398605387e-07, + "loss": 0.5971, + "step": 18675 + }, + { + "epoch": 1.5914784831700044, + "grad_norm": 11.1875, + "learning_rate": 2.507120761219664e-07, + "loss": 0.2272, + "step": 18676 + }, + { + "epoch": 1.5915636983383044, + "grad_norm": 14.0, + "learning_rate": 2.506111363160167e-07, + "loss": 0.4967, + "step": 18677 + }, + { + "epoch": 1.5916489135066043, + "grad_norm": 13.1875, + "learning_rate": 2.5051021457002847e-07, + "loss": 0.5229, + "step": 18678 + }, + { + "epoch": 1.5917341286749043, + "grad_norm": 12.875, + "learning_rate": 2.5040931088582564e-07, + "loss": 0.4295, + "step": 18679 + }, + { + "epoch": 1.5918193438432042, + "grad_norm": 14.875, + "learning_rate": 2.503084252652317e-07, + "loss": 0.57, + "step": 18680 + }, + { + "epoch": 1.5919045590115042, + "grad_norm": 19.375, + "learning_rate": 2.5020755771007015e-07, + "loss": 0.7784, + "step": 18681 + }, + { + "epoch": 1.5919897741798041, + "grad_norm": 16.125, + "learning_rate": 2.501067082221631e-07, + "loss": 0.6915, + "step": 18682 + }, + { + "epoch": 1.592074989348104, + "grad_norm": 17.5, + "learning_rate": 2.500058768033332e-07, + "loss": 0.7322, + "step": 18683 + }, + { + "epoch": 1.592160204516404, + "grad_norm": 21.0, + "learning_rate": 2.4990506345540316e-07, + "loss": 1.0505, + "step": 18684 + }, + { + "epoch": 1.592245419684704, + "grad_norm": 13.4375, + "learning_rate": 2.4980426818019376e-07, + "loss": 0.5769, + "step": 18685 + }, + { + "epoch": 1.592330634853004, + "grad_norm": 17.5, + "learning_rate": 2.4970349097952746e-07, + "loss": 0.7815, + "step": 18686 + }, + { + "epoch": 1.592415850021304, + "grad_norm": 13.5625, + "learning_rate": 2.496027318552247e-07, + "loss": 0.3663, + "step": 18687 + }, + { + "epoch": 1.5925010651896039, + "grad_norm": 13.0, + "learning_rate": 2.495019908091065e-07, + "loss": 0.6091, + "step": 18688 + }, + { + "epoch": 1.5925862803579038, + "grad_norm": 21.625, + "learning_rate": 2.4940126784299343e-07, + "loss": 0.7644, + "step": 18689 + }, + { + "epoch": 1.5926714955262038, + "grad_norm": 11.875, + "learning_rate": 2.4930056295870584e-07, + "loss": 0.3038, + "step": 18690 + }, + { + "epoch": 1.5927567106945038, + "grad_norm": 15.1875, + "learning_rate": 2.4919987615806313e-07, + "loss": 0.6516, + "step": 18691 + }, + { + "epoch": 1.5928419258628037, + "grad_norm": 15.4375, + "learning_rate": 2.4909920744288537e-07, + "loss": 0.5797, + "step": 18692 + }, + { + "epoch": 1.5929271410311037, + "grad_norm": 15.8125, + "learning_rate": 2.489985568149911e-07, + "loss": 0.5171, + "step": 18693 + }, + { + "epoch": 1.5930123561994036, + "grad_norm": 16.625, + "learning_rate": 2.4889792427619966e-07, + "loss": 0.8088, + "step": 18694 + }, + { + "epoch": 1.5930975713677036, + "grad_norm": 11.5, + "learning_rate": 2.4879730982832933e-07, + "loss": 0.366, + "step": 18695 + }, + { + "epoch": 1.5931827865360035, + "grad_norm": 13.0625, + "learning_rate": 2.486967134731982e-07, + "loss": 0.5143, + "step": 18696 + }, + { + "epoch": 1.5932680017043035, + "grad_norm": 14.4375, + "learning_rate": 2.4859613521262443e-07, + "loss": 0.6527, + "step": 18697 + }, + { + "epoch": 1.5933532168726035, + "grad_norm": 14.3125, + "learning_rate": 2.484955750484258e-07, + "loss": 0.6022, + "step": 18698 + }, + { + "epoch": 1.5934384320409034, + "grad_norm": 15.3125, + "learning_rate": 2.483950329824189e-07, + "loss": 0.7407, + "step": 18699 + }, + { + "epoch": 1.5935236472092034, + "grad_norm": 18.5, + "learning_rate": 2.482945090164213e-07, + "loss": 0.7191, + "step": 18700 + }, + { + "epoch": 1.5936088623775033, + "grad_norm": 11.9375, + "learning_rate": 2.481940031522488e-07, + "loss": 0.2934, + "step": 18701 + }, + { + "epoch": 1.5936940775458033, + "grad_norm": 16.625, + "learning_rate": 2.4809351539171815e-07, + "loss": 0.5601, + "step": 18702 + }, + { + "epoch": 1.5937792927141032, + "grad_norm": 16.375, + "learning_rate": 2.479930457366454e-07, + "loss": 0.4799, + "step": 18703 + }, + { + "epoch": 1.5938645078824032, + "grad_norm": 13.0625, + "learning_rate": 2.4789259418884564e-07, + "loss": 0.3071, + "step": 18704 + }, + { + "epoch": 1.5939497230507031, + "grad_norm": 24.25, + "learning_rate": 2.4779216075013447e-07, + "loss": 0.8387, + "step": 18705 + }, + { + "epoch": 1.594034938219003, + "grad_norm": 11.4375, + "learning_rate": 2.476917454223267e-07, + "loss": 0.4521, + "step": 18706 + }, + { + "epoch": 1.594120153387303, + "grad_norm": 19.25, + "learning_rate": 2.475913482072373e-07, + "loss": 0.648, + "step": 18707 + }, + { + "epoch": 1.594205368555603, + "grad_norm": 12.75, + "learning_rate": 2.474909691066803e-07, + "loss": 0.3155, + "step": 18708 + }, + { + "epoch": 1.594290583723903, + "grad_norm": 17.625, + "learning_rate": 2.473906081224692e-07, + "loss": 0.9932, + "step": 18709 + }, + { + "epoch": 1.594375798892203, + "grad_norm": 17.125, + "learning_rate": 2.47290265256418e-07, + "loss": 0.5818, + "step": 18710 + }, + { + "epoch": 1.594461014060503, + "grad_norm": 14.1875, + "learning_rate": 2.4718994051034034e-07, + "loss": 0.7625, + "step": 18711 + }, + { + "epoch": 1.5945462292288028, + "grad_norm": 14.6875, + "learning_rate": 2.4708963388604847e-07, + "loss": 0.5002, + "step": 18712 + }, + { + "epoch": 1.5946314443971028, + "grad_norm": 16.625, + "learning_rate": 2.469893453853556e-07, + "loss": 1.0032, + "step": 18713 + }, + { + "epoch": 1.5947166595654028, + "grad_norm": 16.625, + "learning_rate": 2.4688907501007375e-07, + "loss": 0.5558, + "step": 18714 + }, + { + "epoch": 1.5948018747337027, + "grad_norm": 19.625, + "learning_rate": 2.4678882276201537e-07, + "loss": 0.5802, + "step": 18715 + }, + { + "epoch": 1.5948870899020027, + "grad_norm": 20.75, + "learning_rate": 2.4668858864299166e-07, + "loss": 0.7337, + "step": 18716 + }, + { + "epoch": 1.5949723050703026, + "grad_norm": 17.0, + "learning_rate": 2.4658837265481376e-07, + "loss": 0.7707, + "step": 18717 + }, + { + "epoch": 1.5950575202386026, + "grad_norm": 16.875, + "learning_rate": 2.464881747992931e-07, + "loss": 0.573, + "step": 18718 + }, + { + "epoch": 1.5951427354069025, + "grad_norm": 15.5625, + "learning_rate": 2.463879950782402e-07, + "loss": 0.4791, + "step": 18719 + }, + { + "epoch": 1.5952279505752025, + "grad_norm": 19.25, + "learning_rate": 2.4628783349346566e-07, + "loss": 0.8522, + "step": 18720 + }, + { + "epoch": 1.5953131657435025, + "grad_norm": 19.0, + "learning_rate": 2.4618769004677914e-07, + "loss": 0.5529, + "step": 18721 + }, + { + "epoch": 1.5953983809118024, + "grad_norm": 17.375, + "learning_rate": 2.4608756473999033e-07, + "loss": 0.7735, + "step": 18722 + }, + { + "epoch": 1.5954835960801024, + "grad_norm": 12.4375, + "learning_rate": 2.459874575749091e-07, + "loss": 0.4652, + "step": 18723 + }, + { + "epoch": 1.5955688112484023, + "grad_norm": 12.1875, + "learning_rate": 2.4588736855334415e-07, + "loss": 0.4393, + "step": 18724 + }, + { + "epoch": 1.5956540264167023, + "grad_norm": 12.0, + "learning_rate": 2.4578729767710376e-07, + "loss": 0.4578, + "step": 18725 + }, + { + "epoch": 1.5957392415850022, + "grad_norm": 13.5625, + "learning_rate": 2.4568724494799677e-07, + "loss": 0.5, + "step": 18726 + }, + { + "epoch": 1.5958244567533022, + "grad_norm": 14.0625, + "learning_rate": 2.4558721036783127e-07, + "loss": 0.3416, + "step": 18727 + }, + { + "epoch": 1.5959096719216022, + "grad_norm": 14.0625, + "learning_rate": 2.45487193938415e-07, + "loss": 0.6808, + "step": 18728 + }, + { + "epoch": 1.5959948870899021, + "grad_norm": 24.5, + "learning_rate": 2.453871956615551e-07, + "loss": 0.9701, + "step": 18729 + }, + { + "epoch": 1.596080102258202, + "grad_norm": 14.9375, + "learning_rate": 2.452872155390587e-07, + "loss": 0.4936, + "step": 18730 + }, + { + "epoch": 1.596165317426502, + "grad_norm": 13.0625, + "learning_rate": 2.4518725357273274e-07, + "loss": 0.3895, + "step": 18731 + }, + { + "epoch": 1.596250532594802, + "grad_norm": 13.3125, + "learning_rate": 2.450873097643838e-07, + "loss": 0.5221, + "step": 18732 + }, + { + "epoch": 1.596335747763102, + "grad_norm": 15.0, + "learning_rate": 2.4498738411581763e-07, + "loss": 0.7583, + "step": 18733 + }, + { + "epoch": 1.596420962931402, + "grad_norm": 14.375, + "learning_rate": 2.4488747662883977e-07, + "loss": 0.5887, + "step": 18734 + }, + { + "epoch": 1.5965061780997019, + "grad_norm": 15.9375, + "learning_rate": 2.447875873052559e-07, + "loss": 0.428, + "step": 18735 + }, + { + "epoch": 1.5965913932680018, + "grad_norm": 26.875, + "learning_rate": 2.4468771614687125e-07, + "loss": 0.7837, + "step": 18736 + }, + { + "epoch": 1.5966766084363018, + "grad_norm": 14.3125, + "learning_rate": 2.445878631554907e-07, + "loss": 0.49, + "step": 18737 + }, + { + "epoch": 1.5967618236046017, + "grad_norm": 17.25, + "learning_rate": 2.444880283329182e-07, + "loss": 1.0393, + "step": 18738 + }, + { + "epoch": 1.5968470387729017, + "grad_norm": 12.1875, + "learning_rate": 2.4438821168095816e-07, + "loss": 0.306, + "step": 18739 + }, + { + "epoch": 1.5969322539412016, + "grad_norm": 19.25, + "learning_rate": 2.4428841320141463e-07, + "loss": 1.0694, + "step": 18740 + }, + { + "epoch": 1.5970174691095016, + "grad_norm": 16.75, + "learning_rate": 2.4418863289609075e-07, + "loss": 0.5005, + "step": 18741 + }, + { + "epoch": 1.5971026842778016, + "grad_norm": 15.5625, + "learning_rate": 2.440888707667895e-07, + "loss": 0.6171, + "step": 18742 + }, + { + "epoch": 1.5971878994461015, + "grad_norm": 29.25, + "learning_rate": 2.4398912681531385e-07, + "loss": 0.5317, + "step": 18743 + }, + { + "epoch": 1.5972731146144015, + "grad_norm": 23.5, + "learning_rate": 2.438894010434663e-07, + "loss": 0.937, + "step": 18744 + }, + { + "epoch": 1.5973583297827014, + "grad_norm": 15.125, + "learning_rate": 2.4378969345304927e-07, + "loss": 0.7253, + "step": 18745 + }, + { + "epoch": 1.5974435449510014, + "grad_norm": 13.8125, + "learning_rate": 2.4369000404586397e-07, + "loss": 0.3986, + "step": 18746 + }, + { + "epoch": 1.5975287601193013, + "grad_norm": 11.8125, + "learning_rate": 2.435903328237123e-07, + "loss": 0.2375, + "step": 18747 + }, + { + "epoch": 1.5976139752876013, + "grad_norm": 14.5, + "learning_rate": 2.434906797883953e-07, + "loss": 0.5873, + "step": 18748 + }, + { + "epoch": 1.5976991904559013, + "grad_norm": 16.125, + "learning_rate": 2.43391044941714e-07, + "loss": 0.8699, + "step": 18749 + }, + { + "epoch": 1.5977844056242012, + "grad_norm": 23.875, + "learning_rate": 2.43291428285469e-07, + "loss": 0.8254, + "step": 18750 + }, + { + "epoch": 1.5978696207925012, + "grad_norm": 28.125, + "learning_rate": 2.431918298214597e-07, + "loss": 0.8934, + "step": 18751 + }, + { + "epoch": 1.5979548359608011, + "grad_norm": 14.6875, + "learning_rate": 2.4309224955148663e-07, + "loss": 0.526, + "step": 18752 + }, + { + "epoch": 1.598040051129101, + "grad_norm": 15.25, + "learning_rate": 2.42992687477349e-07, + "loss": 0.658, + "step": 18753 + }, + { + "epoch": 1.598125266297401, + "grad_norm": 19.75, + "learning_rate": 2.428931436008465e-07, + "loss": 0.7171, + "step": 18754 + }, + { + "epoch": 1.598210481465701, + "grad_norm": 16.75, + "learning_rate": 2.427936179237773e-07, + "loss": 0.6928, + "step": 18755 + }, + { + "epoch": 1.598295696634001, + "grad_norm": 17.0, + "learning_rate": 2.4269411044794024e-07, + "loss": 0.4457, + "step": 18756 + }, + { + "epoch": 1.598380911802301, + "grad_norm": 18.0, + "learning_rate": 2.4259462117513384e-07, + "loss": 0.5815, + "step": 18757 + }, + { + "epoch": 1.5984661269706009, + "grad_norm": 20.0, + "learning_rate": 2.424951501071557e-07, + "loss": 0.7767, + "step": 18758 + }, + { + "epoch": 1.5985513421389008, + "grad_norm": 18.0, + "learning_rate": 2.4239569724580304e-07, + "loss": 0.9372, + "step": 18759 + }, + { + "epoch": 1.5986365573072008, + "grad_norm": 14.3125, + "learning_rate": 2.422962625928732e-07, + "loss": 0.6742, + "step": 18760 + }, + { + "epoch": 1.5987217724755007, + "grad_norm": 13.6875, + "learning_rate": 2.4219684615016336e-07, + "loss": 0.654, + "step": 18761 + }, + { + "epoch": 1.5988069876438007, + "grad_norm": 14.875, + "learning_rate": 2.4209744791947013e-07, + "loss": 0.5966, + "step": 18762 + }, + { + "epoch": 1.5988922028121006, + "grad_norm": 13.4375, + "learning_rate": 2.4199806790258936e-07, + "loss": 0.3408, + "step": 18763 + }, + { + "epoch": 1.5989774179804006, + "grad_norm": 17.125, + "learning_rate": 2.4189870610131714e-07, + "loss": 0.7999, + "step": 18764 + }, + { + "epoch": 1.5990626331487006, + "grad_norm": 13.125, + "learning_rate": 2.4179936251744917e-07, + "loss": 0.5337, + "step": 18765 + }, + { + "epoch": 1.5991478483170005, + "grad_norm": 14.375, + "learning_rate": 2.417000371527804e-07, + "loss": 0.7124, + "step": 18766 + }, + { + "epoch": 1.5992330634853005, + "grad_norm": 11.875, + "learning_rate": 2.4160073000910613e-07, + "loss": 0.3955, + "step": 18767 + }, + { + "epoch": 1.5993182786536004, + "grad_norm": 13.1875, + "learning_rate": 2.4150144108822036e-07, + "loss": 0.4551, + "step": 18768 + }, + { + "epoch": 1.5994034938219004, + "grad_norm": 21.5, + "learning_rate": 2.4140217039191766e-07, + "loss": 0.5333, + "step": 18769 + }, + { + "epoch": 1.5994887089902003, + "grad_norm": 11.625, + "learning_rate": 2.4130291792199236e-07, + "loss": 0.2712, + "step": 18770 + }, + { + "epoch": 1.5995739241585003, + "grad_norm": 22.875, + "learning_rate": 2.4120368368023727e-07, + "loss": 0.9137, + "step": 18771 + }, + { + "epoch": 1.5996591393268003, + "grad_norm": 15.8125, + "learning_rate": 2.4110446766844596e-07, + "loss": 0.3845, + "step": 18772 + }, + { + "epoch": 1.5997443544951002, + "grad_norm": 15.0625, + "learning_rate": 2.410052698884119e-07, + "loss": 0.4656, + "step": 18773 + }, + { + "epoch": 1.5998295696634002, + "grad_norm": 13.1875, + "learning_rate": 2.4090609034192693e-07, + "loss": 0.5611, + "step": 18774 + }, + { + "epoch": 1.5999147848317001, + "grad_norm": 18.5, + "learning_rate": 2.408069290307838e-07, + "loss": 0.6448, + "step": 18775 + }, + { + "epoch": 1.6, + "grad_norm": 21.125, + "learning_rate": 2.407077859567741e-07, + "loss": 0.5192, + "step": 18776 + }, + { + "epoch": 1.6000852151683, + "grad_norm": 18.25, + "learning_rate": 2.4060866112168957e-07, + "loss": 0.8147, + "step": 18777 + }, + { + "epoch": 1.6001704303366, + "grad_norm": 20.0, + "learning_rate": 2.405095545273216e-07, + "loss": 0.6353, + "step": 18778 + }, + { + "epoch": 1.6002556455049, + "grad_norm": 24.375, + "learning_rate": 2.4041046617546136e-07, + "loss": 0.7615, + "step": 18779 + }, + { + "epoch": 1.6003408606732, + "grad_norm": 14.5, + "learning_rate": 2.403113960678989e-07, + "loss": 0.4928, + "step": 18780 + }, + { + "epoch": 1.6004260758414999, + "grad_norm": 9.9375, + "learning_rate": 2.4021234420642525e-07, + "loss": 0.1688, + "step": 18781 + }, + { + "epoch": 1.6005112910097998, + "grad_norm": 11.0, + "learning_rate": 2.401133105928295e-07, + "loss": 0.4235, + "step": 18782 + }, + { + "epoch": 1.6005965061780998, + "grad_norm": 15.8125, + "learning_rate": 2.400142952289018e-07, + "loss": 0.5192, + "step": 18783 + }, + { + "epoch": 1.6006817213463997, + "grad_norm": 13.9375, + "learning_rate": 2.399152981164318e-07, + "loss": 0.5804, + "step": 18784 + }, + { + "epoch": 1.6007669365146997, + "grad_norm": 17.875, + "learning_rate": 2.3981631925720767e-07, + "loss": 1.0167, + "step": 18785 + }, + { + "epoch": 1.6008521516829997, + "grad_norm": 13.9375, + "learning_rate": 2.3971735865301855e-07, + "loss": 0.594, + "step": 18786 + }, + { + "epoch": 1.6009373668512996, + "grad_norm": 15.3125, + "learning_rate": 2.3961841630565295e-07, + "loss": 0.4664, + "step": 18787 + }, + { + "epoch": 1.6010225820195996, + "grad_norm": 15.5, + "learning_rate": 2.395194922168982e-07, + "loss": 0.898, + "step": 18788 + }, + { + "epoch": 1.6011077971878995, + "grad_norm": 30.25, + "learning_rate": 2.394205863885428e-07, + "loss": 0.8917, + "step": 18789 + }, + { + "epoch": 1.6011930123561995, + "grad_norm": 13.125, + "learning_rate": 2.3932169882237314e-07, + "loss": 0.5951, + "step": 18790 + }, + { + "epoch": 1.6012782275244994, + "grad_norm": 14.9375, + "learning_rate": 2.392228295201769e-07, + "loss": 0.7615, + "step": 18791 + }, + { + "epoch": 1.6013634426927994, + "grad_norm": 13.8125, + "learning_rate": 2.391239784837407e-07, + "loss": 0.4873, + "step": 18792 + }, + { + "epoch": 1.6014486578610994, + "grad_norm": 17.75, + "learning_rate": 2.390251457148504e-07, + "loss": 0.7804, + "step": 18793 + }, + { + "epoch": 1.6015338730293993, + "grad_norm": 12.6875, + "learning_rate": 2.3892633121529244e-07, + "loss": 0.3813, + "step": 18794 + }, + { + "epoch": 1.6016190881976993, + "grad_norm": 22.125, + "learning_rate": 2.388275349868524e-07, + "loss": 0.7056, + "step": 18795 + }, + { + "epoch": 1.6017043033659992, + "grad_norm": 17.25, + "learning_rate": 2.387287570313158e-07, + "loss": 0.9108, + "step": 18796 + }, + { + "epoch": 1.6017895185342992, + "grad_norm": 13.0, + "learning_rate": 2.3862999735046756e-07, + "loss": 0.6095, + "step": 18797 + }, + { + "epoch": 1.6018747337025991, + "grad_norm": 13.6875, + "learning_rate": 2.38531255946092e-07, + "loss": 0.5042, + "step": 18798 + }, + { + "epoch": 1.601959948870899, + "grad_norm": 19.125, + "learning_rate": 2.3843253281997366e-07, + "loss": 0.6324, + "step": 18799 + }, + { + "epoch": 1.602045164039199, + "grad_norm": 15.25, + "learning_rate": 2.38333827973897e-07, + "loss": 0.6849, + "step": 18800 + }, + { + "epoch": 1.602130379207499, + "grad_norm": 19.5, + "learning_rate": 2.382351414096451e-07, + "loss": 0.985, + "step": 18801 + }, + { + "epoch": 1.602215594375799, + "grad_norm": 9.6875, + "learning_rate": 2.381364731290016e-07, + "loss": 0.2165, + "step": 18802 + }, + { + "epoch": 1.602300809544099, + "grad_norm": 14.5, + "learning_rate": 2.380378231337495e-07, + "loss": 0.7024, + "step": 18803 + }, + { + "epoch": 1.6023860247123989, + "grad_norm": 15.6875, + "learning_rate": 2.3793919142567178e-07, + "loss": 0.7469, + "step": 18804 + }, + { + "epoch": 1.6024712398806988, + "grad_norm": 14.875, + "learning_rate": 2.3784057800655068e-07, + "loss": 0.6323, + "step": 18805 + }, + { + "epoch": 1.6025564550489988, + "grad_norm": 16.125, + "learning_rate": 2.3774198287816781e-07, + "loss": 0.6103, + "step": 18806 + }, + { + "epoch": 1.6026416702172988, + "grad_norm": 13.9375, + "learning_rate": 2.3764340604230524e-07, + "loss": 0.6129, + "step": 18807 + }, + { + "epoch": 1.6027268853855987, + "grad_norm": 19.5, + "learning_rate": 2.3754484750074428e-07, + "loss": 0.6792, + "step": 18808 + }, + { + "epoch": 1.6028121005538987, + "grad_norm": 17.5, + "learning_rate": 2.3744630725526642e-07, + "loss": 0.8101, + "step": 18809 + }, + { + "epoch": 1.6028973157221986, + "grad_norm": 13.625, + "learning_rate": 2.373477853076517e-07, + "loss": 0.5101, + "step": 18810 + }, + { + "epoch": 1.6029825308904986, + "grad_norm": 11.5, + "learning_rate": 2.3724928165968066e-07, + "loss": 0.3176, + "step": 18811 + }, + { + "epoch": 1.6030677460587985, + "grad_norm": 14.625, + "learning_rate": 2.371507963131335e-07, + "loss": 0.4864, + "step": 18812 + }, + { + "epoch": 1.6031529612270985, + "grad_norm": 21.625, + "learning_rate": 2.3705232926979045e-07, + "loss": 1.0411, + "step": 18813 + }, + { + "epoch": 1.6032381763953985, + "grad_norm": 32.25, + "learning_rate": 2.3695388053143017e-07, + "loss": 0.9117, + "step": 18814 + }, + { + "epoch": 1.6033233915636984, + "grad_norm": 15.4375, + "learning_rate": 2.3685545009983186e-07, + "loss": 0.6072, + "step": 18815 + }, + { + "epoch": 1.6034086067319984, + "grad_norm": 19.75, + "learning_rate": 2.3675703797677418e-07, + "loss": 0.7835, + "step": 18816 + }, + { + "epoch": 1.6034938219002983, + "grad_norm": 30.125, + "learning_rate": 2.3665864416403617e-07, + "loss": 0.8277, + "step": 18817 + }, + { + "epoch": 1.6035790370685983, + "grad_norm": 15.75, + "learning_rate": 2.3656026866339513e-07, + "loss": 0.9781, + "step": 18818 + }, + { + "epoch": 1.6036642522368982, + "grad_norm": 13.0625, + "learning_rate": 2.3646191147662908e-07, + "loss": 0.5267, + "step": 18819 + }, + { + "epoch": 1.6037494674051982, + "grad_norm": 16.125, + "learning_rate": 2.3636357260551553e-07, + "loss": 0.5873, + "step": 18820 + }, + { + "epoch": 1.6038346825734981, + "grad_norm": 13.4375, + "learning_rate": 2.3626525205183174e-07, + "loss": 0.4482, + "step": 18821 + }, + { + "epoch": 1.603919897741798, + "grad_norm": 14.5, + "learning_rate": 2.3616694981735424e-07, + "loss": 0.4091, + "step": 18822 + }, + { + "epoch": 1.604005112910098, + "grad_norm": 18.0, + "learning_rate": 2.3606866590385914e-07, + "loss": 0.9149, + "step": 18823 + }, + { + "epoch": 1.604090328078398, + "grad_norm": 17.5, + "learning_rate": 2.3597040031312286e-07, + "loss": 0.5366, + "step": 18824 + }, + { + "epoch": 1.604175543246698, + "grad_norm": 13.375, + "learning_rate": 2.3587215304692116e-07, + "loss": 0.5246, + "step": 18825 + }, + { + "epoch": 1.604260758414998, + "grad_norm": 19.75, + "learning_rate": 2.3577392410702975e-07, + "loss": 0.6386, + "step": 18826 + }, + { + "epoch": 1.604345973583298, + "grad_norm": 30.75, + "learning_rate": 2.3567571349522312e-07, + "loss": 0.3958, + "step": 18827 + }, + { + "epoch": 1.6044311887515978, + "grad_norm": 10.625, + "learning_rate": 2.355775212132763e-07, + "loss": 0.8644, + "step": 18828 + }, + { + "epoch": 1.6045164039198978, + "grad_norm": 14.375, + "learning_rate": 2.3547934726296366e-07, + "loss": 0.5665, + "step": 18829 + }, + { + "epoch": 1.6046016190881978, + "grad_norm": 13.25, + "learning_rate": 2.353811916460602e-07, + "loss": 0.4578, + "step": 18830 + }, + { + "epoch": 1.6046868342564977, + "grad_norm": 14.625, + "learning_rate": 2.3528305436433808e-07, + "loss": 0.4603, + "step": 18831 + }, + { + "epoch": 1.6047720494247977, + "grad_norm": 14.5, + "learning_rate": 2.3518493541957163e-07, + "loss": 0.6296, + "step": 18832 + }, + { + "epoch": 1.6048572645930976, + "grad_norm": 16.25, + "learning_rate": 2.3508683481353393e-07, + "loss": 0.5881, + "step": 18833 + }, + { + "epoch": 1.6049424797613976, + "grad_norm": 23.125, + "learning_rate": 2.3498875254799796e-07, + "loss": 0.5631, + "step": 18834 + }, + { + "epoch": 1.6050276949296975, + "grad_norm": 17.375, + "learning_rate": 2.3489068862473554e-07, + "loss": 0.8261, + "step": 18835 + }, + { + "epoch": 1.6051129100979975, + "grad_norm": 16.75, + "learning_rate": 2.3479264304551913e-07, + "loss": 0.6818, + "step": 18836 + }, + { + "epoch": 1.6051981252662975, + "grad_norm": 19.375, + "learning_rate": 2.3469461581212064e-07, + "loss": 0.7859, + "step": 18837 + }, + { + "epoch": 1.6052833404345974, + "grad_norm": 14.0, + "learning_rate": 2.345966069263117e-07, + "loss": 0.5161, + "step": 18838 + }, + { + "epoch": 1.6053685556028974, + "grad_norm": 19.25, + "learning_rate": 2.3449861638986314e-07, + "loss": 0.6446, + "step": 18839 + }, + { + "epoch": 1.6054537707711973, + "grad_norm": 20.75, + "learning_rate": 2.3440064420454545e-07, + "loss": 0.5604, + "step": 18840 + }, + { + "epoch": 1.6055389859394973, + "grad_norm": 13.5, + "learning_rate": 2.3430269037212933e-07, + "loss": 0.6236, + "step": 18841 + }, + { + "epoch": 1.6056242011077972, + "grad_norm": 16.875, + "learning_rate": 2.3420475489438487e-07, + "loss": 0.5447, + "step": 18842 + }, + { + "epoch": 1.6057094162760972, + "grad_norm": 15.375, + "learning_rate": 2.3410683777308235e-07, + "loss": 0.5969, + "step": 18843 + }, + { + "epoch": 1.6057946314443972, + "grad_norm": 13.625, + "learning_rate": 2.340089390099906e-07, + "loss": 0.4832, + "step": 18844 + }, + { + "epoch": 1.6058798466126971, + "grad_norm": 16.625, + "learning_rate": 2.3391105860687894e-07, + "loss": 0.6368, + "step": 18845 + }, + { + "epoch": 1.605965061780997, + "grad_norm": 16.0, + "learning_rate": 2.3381319656551647e-07, + "loss": 0.8314, + "step": 18846 + }, + { + "epoch": 1.606050276949297, + "grad_norm": 15.25, + "learning_rate": 2.3371535288767137e-07, + "loss": 0.5549, + "step": 18847 + }, + { + "epoch": 1.606135492117597, + "grad_norm": 23.25, + "learning_rate": 2.3361752757511152e-07, + "loss": 0.6812, + "step": 18848 + }, + { + "epoch": 1.606220707285897, + "grad_norm": 15.0, + "learning_rate": 2.33519720629605e-07, + "loss": 0.6904, + "step": 18849 + }, + { + "epoch": 1.606305922454197, + "grad_norm": 11.625, + "learning_rate": 2.3342193205291936e-07, + "loss": 0.2608, + "step": 18850 + }, + { + "epoch": 1.6063911376224969, + "grad_norm": 16.75, + "learning_rate": 2.3332416184682183e-07, + "loss": 0.6077, + "step": 18851 + }, + { + "epoch": 1.6064763527907968, + "grad_norm": 15.6875, + "learning_rate": 2.3322641001307878e-07, + "loss": 0.5989, + "step": 18852 + }, + { + "epoch": 1.6065615679590968, + "grad_norm": 16.5, + "learning_rate": 2.3312867655345698e-07, + "loss": 0.635, + "step": 18853 + }, + { + "epoch": 1.6066467831273967, + "grad_norm": 24.25, + "learning_rate": 2.330309614697228e-07, + "loss": 0.6872, + "step": 18854 + }, + { + "epoch": 1.6067319982956967, + "grad_norm": 11.9375, + "learning_rate": 2.3293326476364162e-07, + "loss": 0.2417, + "step": 18855 + }, + { + "epoch": 1.6068172134639966, + "grad_norm": 15.6875, + "learning_rate": 2.3283558643697926e-07, + "loss": 0.6158, + "step": 18856 + }, + { + "epoch": 1.6069024286322966, + "grad_norm": 13.9375, + "learning_rate": 2.327379264915003e-07, + "loss": 0.2301, + "step": 18857 + }, + { + "epoch": 1.6069876438005966, + "grad_norm": 18.625, + "learning_rate": 2.326402849289701e-07, + "loss": 0.5618, + "step": 18858 + }, + { + "epoch": 1.6070728589688965, + "grad_norm": 21.0, + "learning_rate": 2.3254266175115295e-07, + "loss": 0.5473, + "step": 18859 + }, + { + "epoch": 1.6071580741371965, + "grad_norm": 17.625, + "learning_rate": 2.324450569598133e-07, + "loss": 0.4874, + "step": 18860 + }, + { + "epoch": 1.6072432893054964, + "grad_norm": 13.3125, + "learning_rate": 2.323474705567144e-07, + "loss": 0.385, + "step": 18861 + }, + { + "epoch": 1.6073285044737964, + "grad_norm": 17.875, + "learning_rate": 2.3224990254362034e-07, + "loss": 0.8734, + "step": 18862 + }, + { + "epoch": 1.6074137196420963, + "grad_norm": 13.9375, + "learning_rate": 2.3215235292229368e-07, + "loss": 0.4393, + "step": 18863 + }, + { + "epoch": 1.6074989348103963, + "grad_norm": 12.6875, + "learning_rate": 2.3205482169449789e-07, + "loss": 0.3763, + "step": 18864 + }, + { + "epoch": 1.6075841499786963, + "grad_norm": 13.9375, + "learning_rate": 2.3195730886199475e-07, + "loss": 0.6207, + "step": 18865 + }, + { + "epoch": 1.6076693651469962, + "grad_norm": 17.875, + "learning_rate": 2.3185981442654673e-07, + "loss": 0.6964, + "step": 18866 + }, + { + "epoch": 1.6077545803152962, + "grad_norm": 20.125, + "learning_rate": 2.3176233838991562e-07, + "loss": 0.4676, + "step": 18867 + }, + { + "epoch": 1.6078397954835961, + "grad_norm": 11.5625, + "learning_rate": 2.3166488075386336e-07, + "loss": 0.3362, + "step": 18868 + }, + { + "epoch": 1.607925010651896, + "grad_norm": 13.9375, + "learning_rate": 2.3156744152015048e-07, + "loss": 0.6524, + "step": 18869 + }, + { + "epoch": 1.608010225820196, + "grad_norm": 17.125, + "learning_rate": 2.3147002069053833e-07, + "loss": 0.5693, + "step": 18870 + }, + { + "epoch": 1.608095440988496, + "grad_norm": 18.625, + "learning_rate": 2.3137261826678677e-07, + "loss": 0.4716, + "step": 18871 + }, + { + "epoch": 1.608180656156796, + "grad_norm": 18.0, + "learning_rate": 2.312752342506562e-07, + "loss": 0.7448, + "step": 18872 + }, + { + "epoch": 1.608265871325096, + "grad_norm": 18.875, + "learning_rate": 2.3117786864390701e-07, + "loss": 0.8164, + "step": 18873 + }, + { + "epoch": 1.6083510864933959, + "grad_norm": 13.0625, + "learning_rate": 2.3108052144829792e-07, + "loss": 0.3929, + "step": 18874 + }, + { + "epoch": 1.6084363016616958, + "grad_norm": 20.0, + "learning_rate": 2.3098319266558838e-07, + "loss": 0.7361, + "step": 18875 + }, + { + "epoch": 1.6085215168299958, + "grad_norm": 22.75, + "learning_rate": 2.3088588229753712e-07, + "loss": 0.7829, + "step": 18876 + }, + { + "epoch": 1.6086067319982957, + "grad_norm": 35.25, + "learning_rate": 2.3078859034590325e-07, + "loss": 0.7101, + "step": 18877 + }, + { + "epoch": 1.6086919471665957, + "grad_norm": 21.5, + "learning_rate": 2.3069131681244403e-07, + "loss": 0.5062, + "step": 18878 + }, + { + "epoch": 1.6087771623348956, + "grad_norm": 12.6875, + "learning_rate": 2.3059406169891816e-07, + "loss": 0.393, + "step": 18879 + }, + { + "epoch": 1.6088623775031956, + "grad_norm": 20.625, + "learning_rate": 2.3049682500708228e-07, + "loss": 0.5155, + "step": 18880 + }, + { + "epoch": 1.6089475926714956, + "grad_norm": 16.0, + "learning_rate": 2.3039960673869433e-07, + "loss": 0.6625, + "step": 18881 + }, + { + "epoch": 1.6090328078397955, + "grad_norm": 26.875, + "learning_rate": 2.3030240689551037e-07, + "loss": 0.6586, + "step": 18882 + }, + { + "epoch": 1.6091180230080955, + "grad_norm": 10.4375, + "learning_rate": 2.3020522547928733e-07, + "loss": 0.4726, + "step": 18883 + }, + { + "epoch": 1.6092032381763954, + "grad_norm": 19.125, + "learning_rate": 2.3010806249178132e-07, + "loss": 0.983, + "step": 18884 + }, + { + "epoch": 1.6092884533446954, + "grad_norm": 13.25, + "learning_rate": 2.3001091793474858e-07, + "loss": 0.5155, + "step": 18885 + }, + { + "epoch": 1.6093736685129953, + "grad_norm": 14.75, + "learning_rate": 2.2991379180994395e-07, + "loss": 0.4991, + "step": 18886 + }, + { + "epoch": 1.6094588836812953, + "grad_norm": 14.375, + "learning_rate": 2.2981668411912308e-07, + "loss": 0.5437, + "step": 18887 + }, + { + "epoch": 1.6095440988495953, + "grad_norm": 17.75, + "learning_rate": 2.2971959486404045e-07, + "loss": 0.5197, + "step": 18888 + }, + { + "epoch": 1.6096293140178952, + "grad_norm": 18.5, + "learning_rate": 2.2962252404645058e-07, + "loss": 0.9474, + "step": 18889 + }, + { + "epoch": 1.6097145291861952, + "grad_norm": 16.375, + "learning_rate": 2.2952547166810806e-07, + "loss": 0.4562, + "step": 18890 + }, + { + "epoch": 1.6097997443544951, + "grad_norm": 15.75, + "learning_rate": 2.2942843773076622e-07, + "loss": 0.657, + "step": 18891 + }, + { + "epoch": 1.609884959522795, + "grad_norm": 13.375, + "learning_rate": 2.2933142223617865e-07, + "loss": 0.3246, + "step": 18892 + }, + { + "epoch": 1.609970174691095, + "grad_norm": 16.375, + "learning_rate": 2.292344251860991e-07, + "loss": 0.5714, + "step": 18893 + }, + { + "epoch": 1.610055389859395, + "grad_norm": 15.0625, + "learning_rate": 2.291374465822796e-07, + "loss": 0.736, + "step": 18894 + }, + { + "epoch": 1.610140605027695, + "grad_norm": 16.5, + "learning_rate": 2.2904048642647323e-07, + "loss": 0.5963, + "step": 18895 + }, + { + "epoch": 1.610225820195995, + "grad_norm": 13.3125, + "learning_rate": 2.2894354472043162e-07, + "loss": 0.5651, + "step": 18896 + }, + { + "epoch": 1.6103110353642949, + "grad_norm": 15.0625, + "learning_rate": 2.28846621465907e-07, + "loss": 0.5897, + "step": 18897 + }, + { + "epoch": 1.6103962505325948, + "grad_norm": 14.25, + "learning_rate": 2.28749716664651e-07, + "loss": 0.5992, + "step": 18898 + }, + { + "epoch": 1.6104814657008948, + "grad_norm": 13.4375, + "learning_rate": 2.2865283031841434e-07, + "loss": 0.6712, + "step": 18899 + }, + { + "epoch": 1.6105666808691947, + "grad_norm": 15.9375, + "learning_rate": 2.2855596242894796e-07, + "loss": 0.494, + "step": 18900 + }, + { + "epoch": 1.6106518960374947, + "grad_norm": 13.125, + "learning_rate": 2.2845911299800252e-07, + "loss": 0.4914, + "step": 18901 + }, + { + "epoch": 1.6107371112057947, + "grad_norm": 17.0, + "learning_rate": 2.283622820273286e-07, + "loss": 0.4247, + "step": 18902 + }, + { + "epoch": 1.6108223263740946, + "grad_norm": 12.625, + "learning_rate": 2.2826546951867548e-07, + "loss": 0.4652, + "step": 18903 + }, + { + "epoch": 1.6109075415423946, + "grad_norm": 16.5, + "learning_rate": 2.2816867547379232e-07, + "loss": 0.7342, + "step": 18904 + }, + { + "epoch": 1.6109927567106945, + "grad_norm": 14.0, + "learning_rate": 2.2807189989442895e-07, + "loss": 0.4671, + "step": 18905 + }, + { + "epoch": 1.6110779718789945, + "grad_norm": 20.0, + "learning_rate": 2.2797514278233388e-07, + "loss": 0.8333, + "step": 18906 + }, + { + "epoch": 1.6111631870472944, + "grad_norm": 18.0, + "learning_rate": 2.2787840413925593e-07, + "loss": 0.7056, + "step": 18907 + }, + { + "epoch": 1.6112484022155944, + "grad_norm": 13.5625, + "learning_rate": 2.2778168396694278e-07, + "loss": 0.367, + "step": 18908 + }, + { + "epoch": 1.6113336173838944, + "grad_norm": 18.375, + "learning_rate": 2.2768498226714258e-07, + "loss": 0.3515, + "step": 18909 + }, + { + "epoch": 1.6114188325521943, + "grad_norm": 12.0625, + "learning_rate": 2.27588299041603e-07, + "loss": 0.3745, + "step": 18910 + }, + { + "epoch": 1.6115040477204943, + "grad_norm": 16.125, + "learning_rate": 2.2749163429207105e-07, + "loss": 0.5195, + "step": 18911 + }, + { + "epoch": 1.6115892628887942, + "grad_norm": 18.25, + "learning_rate": 2.2739498802029318e-07, + "loss": 0.516, + "step": 18912 + }, + { + "epoch": 1.6116744780570942, + "grad_norm": 15.0625, + "learning_rate": 2.2729836022801612e-07, + "loss": 0.4664, + "step": 18913 + }, + { + "epoch": 1.6117596932253941, + "grad_norm": 10.875, + "learning_rate": 2.2720175091698603e-07, + "loss": 0.3204, + "step": 18914 + }, + { + "epoch": 1.611844908393694, + "grad_norm": 17.625, + "learning_rate": 2.2710516008894928e-07, + "loss": 0.6475, + "step": 18915 + }, + { + "epoch": 1.611930123561994, + "grad_norm": 18.625, + "learning_rate": 2.2700858774565044e-07, + "loss": 0.7274, + "step": 18916 + }, + { + "epoch": 1.612015338730294, + "grad_norm": 15.75, + "learning_rate": 2.269120338888352e-07, + "loss": 0.7989, + "step": 18917 + }, + { + "epoch": 1.612100553898594, + "grad_norm": 14.375, + "learning_rate": 2.2681549852024823e-07, + "loss": 0.4669, + "step": 18918 + }, + { + "epoch": 1.612185769066894, + "grad_norm": 15.5625, + "learning_rate": 2.267189816416346e-07, + "loss": 0.3724, + "step": 18919 + }, + { + "epoch": 1.6122709842351939, + "grad_norm": 17.75, + "learning_rate": 2.266224832547377e-07, + "loss": 0.7062, + "step": 18920 + }, + { + "epoch": 1.6123561994034938, + "grad_norm": 11.6875, + "learning_rate": 2.265260033613015e-07, + "loss": 0.2924, + "step": 18921 + }, + { + "epoch": 1.6124414145717938, + "grad_norm": 18.375, + "learning_rate": 2.2642954196306953e-07, + "loss": 0.8064, + "step": 18922 + }, + { + "epoch": 1.6125266297400938, + "grad_norm": 12.0625, + "learning_rate": 2.2633309906178526e-07, + "loss": 0.4289, + "step": 18923 + }, + { + "epoch": 1.6126118449083937, + "grad_norm": 13.75, + "learning_rate": 2.2623667465919098e-07, + "loss": 0.4171, + "step": 18924 + }, + { + "epoch": 1.6126970600766937, + "grad_norm": 9.875, + "learning_rate": 2.261402687570295e-07, + "loss": 0.2209, + "step": 18925 + }, + { + "epoch": 1.6127822752449936, + "grad_norm": 14.0625, + "learning_rate": 2.2604388135704286e-07, + "loss": 0.4362, + "step": 18926 + }, + { + "epoch": 1.6128674904132936, + "grad_norm": 15.8125, + "learning_rate": 2.2594751246097334e-07, + "loss": 0.46, + "step": 18927 + }, + { + "epoch": 1.6129527055815935, + "grad_norm": 14.9375, + "learning_rate": 2.258511620705621e-07, + "loss": 0.5378, + "step": 18928 + }, + { + "epoch": 1.6130379207498935, + "grad_norm": 15.0625, + "learning_rate": 2.257548301875498e-07, + "loss": 0.3998, + "step": 18929 + }, + { + "epoch": 1.6131231359181935, + "grad_norm": 19.875, + "learning_rate": 2.2565851681367768e-07, + "loss": 0.8357, + "step": 18930 + }, + { + "epoch": 1.6132083510864934, + "grad_norm": 13.5, + "learning_rate": 2.255622219506863e-07, + "loss": 0.3133, + "step": 18931 + }, + { + "epoch": 1.6132935662547934, + "grad_norm": 12.0, + "learning_rate": 2.2546594560031597e-07, + "loss": 0.3148, + "step": 18932 + }, + { + "epoch": 1.6133787814230933, + "grad_norm": 15.0, + "learning_rate": 2.2536968776430593e-07, + "loss": 0.7453, + "step": 18933 + }, + { + "epoch": 1.6134639965913933, + "grad_norm": 23.125, + "learning_rate": 2.252734484443961e-07, + "loss": 0.8936, + "step": 18934 + }, + { + "epoch": 1.6135492117596932, + "grad_norm": 15.25, + "learning_rate": 2.2517722764232558e-07, + "loss": 0.5723, + "step": 18935 + }, + { + "epoch": 1.6136344269279932, + "grad_norm": 14.125, + "learning_rate": 2.250810253598329e-07, + "loss": 0.6366, + "step": 18936 + }, + { + "epoch": 1.6137196420962931, + "grad_norm": 13.875, + "learning_rate": 2.2498484159865706e-07, + "loss": 0.6842, + "step": 18937 + }, + { + "epoch": 1.613804857264593, + "grad_norm": 13.3125, + "learning_rate": 2.2488867636053554e-07, + "loss": 0.409, + "step": 18938 + }, + { + "epoch": 1.613890072432893, + "grad_norm": 23.5, + "learning_rate": 2.2479252964720658e-07, + "loss": 1.0129, + "step": 18939 + }, + { + "epoch": 1.613975287601193, + "grad_norm": 11.5, + "learning_rate": 2.2469640146040763e-07, + "loss": 0.3593, + "step": 18940 + }, + { + "epoch": 1.614060502769493, + "grad_norm": 14.25, + "learning_rate": 2.2460029180187554e-07, + "loss": 0.6708, + "step": 18941 + }, + { + "epoch": 1.614145717937793, + "grad_norm": 13.375, + "learning_rate": 2.2450420067334735e-07, + "loss": 0.4351, + "step": 18942 + }, + { + "epoch": 1.614230933106093, + "grad_norm": 14.6875, + "learning_rate": 2.2440812807655962e-07, + "loss": 0.6171, + "step": 18943 + }, + { + "epoch": 1.6143161482743928, + "grad_norm": 18.875, + "learning_rate": 2.2431207401324816e-07, + "loss": 0.7788, + "step": 18944 + }, + { + "epoch": 1.6144013634426928, + "grad_norm": 20.25, + "learning_rate": 2.2421603848514924e-07, + "loss": 0.4486, + "step": 18945 + }, + { + "epoch": 1.6144865786109928, + "grad_norm": 15.0, + "learning_rate": 2.2412002149399758e-07, + "loss": 0.7137, + "step": 18946 + }, + { + "epoch": 1.6145717937792927, + "grad_norm": 11.5625, + "learning_rate": 2.2402402304152886e-07, + "loss": 0.2267, + "step": 18947 + }, + { + "epoch": 1.6146570089475927, + "grad_norm": 19.75, + "learning_rate": 2.2392804312947768e-07, + "loss": 0.9764, + "step": 18948 + }, + { + "epoch": 1.6147422241158926, + "grad_norm": 13.875, + "learning_rate": 2.238320817595789e-07, + "loss": 0.4837, + "step": 18949 + }, + { + "epoch": 1.6148274392841926, + "grad_norm": 22.75, + "learning_rate": 2.2373613893356584e-07, + "loss": 1.129, + "step": 18950 + }, + { + "epoch": 1.6149126544524925, + "grad_norm": 18.0, + "learning_rate": 2.2364021465317285e-07, + "loss": 0.8148, + "step": 18951 + }, + { + "epoch": 1.6149978696207925, + "grad_norm": 14.3125, + "learning_rate": 2.2354430892013362e-07, + "loss": 0.7075, + "step": 18952 + }, + { + "epoch": 1.6150830847890925, + "grad_norm": 16.125, + "learning_rate": 2.2344842173618084e-07, + "loss": 0.6647, + "step": 18953 + }, + { + "epoch": 1.6151682999573924, + "grad_norm": 18.625, + "learning_rate": 2.2335255310304714e-07, + "loss": 0.9228, + "step": 18954 + }, + { + "epoch": 1.6152535151256924, + "grad_norm": 9.6875, + "learning_rate": 2.2325670302246512e-07, + "loss": 0.2184, + "step": 18955 + }, + { + "epoch": 1.6153387302939923, + "grad_norm": 15.0, + "learning_rate": 2.2316087149616694e-07, + "loss": 0.5058, + "step": 18956 + }, + { + "epoch": 1.6154239454622923, + "grad_norm": 11.75, + "learning_rate": 2.2306505852588463e-07, + "loss": 0.3358, + "step": 18957 + }, + { + "epoch": 1.6155091606305922, + "grad_norm": 20.625, + "learning_rate": 2.229692641133492e-07, + "loss": 0.6196, + "step": 18958 + }, + { + "epoch": 1.6155943757988922, + "grad_norm": 13.625, + "learning_rate": 2.2287348826029175e-07, + "loss": 0.3448, + "step": 18959 + }, + { + "epoch": 1.6156795909671922, + "grad_norm": 10.25, + "learning_rate": 2.227777309684437e-07, + "loss": 0.4847, + "step": 18960 + }, + { + "epoch": 1.6157648061354921, + "grad_norm": 15.3125, + "learning_rate": 2.2268199223953457e-07, + "loss": 0.6079, + "step": 18961 + }, + { + "epoch": 1.615850021303792, + "grad_norm": 11.8125, + "learning_rate": 2.2258627207529526e-07, + "loss": 0.4541, + "step": 18962 + }, + { + "epoch": 1.615935236472092, + "grad_norm": 12.5625, + "learning_rate": 2.2249057047745493e-07, + "loss": 0.3674, + "step": 18963 + }, + { + "epoch": 1.616020451640392, + "grad_norm": 32.5, + "learning_rate": 2.2239488744774314e-07, + "loss": 1.1958, + "step": 18964 + }, + { + "epoch": 1.616105666808692, + "grad_norm": 14.125, + "learning_rate": 2.2229922298788898e-07, + "loss": 0.5229, + "step": 18965 + }, + { + "epoch": 1.616190881976992, + "grad_norm": 15.4375, + "learning_rate": 2.222035770996217e-07, + "loss": 0.4778, + "step": 18966 + }, + { + "epoch": 1.6162760971452919, + "grad_norm": 13.4375, + "learning_rate": 2.2210794978466913e-07, + "loss": 0.4959, + "step": 18967 + }, + { + "epoch": 1.6163613123135918, + "grad_norm": 21.875, + "learning_rate": 2.2201234104475959e-07, + "loss": 0.8219, + "step": 18968 + }, + { + "epoch": 1.6164465274818918, + "grad_norm": 14.5, + "learning_rate": 2.2191675088162057e-07, + "loss": 0.6508, + "step": 18969 + }, + { + "epoch": 1.6165317426501917, + "grad_norm": 21.125, + "learning_rate": 2.2182117929698002e-07, + "loss": 0.6952, + "step": 18970 + }, + { + "epoch": 1.6166169578184917, + "grad_norm": 22.375, + "learning_rate": 2.2172562629256434e-07, + "loss": 1.0118, + "step": 18971 + }, + { + "epoch": 1.6167021729867916, + "grad_norm": 13.3125, + "learning_rate": 2.2163009187010059e-07, + "loss": 0.5616, + "step": 18972 + }, + { + "epoch": 1.6167873881550916, + "grad_norm": 20.5, + "learning_rate": 2.215345760313152e-07, + "loss": 0.7652, + "step": 18973 + }, + { + "epoch": 1.6168726033233916, + "grad_norm": 28.25, + "learning_rate": 2.2143907877793456e-07, + "loss": 0.7624, + "step": 18974 + }, + { + "epoch": 1.6169578184916915, + "grad_norm": 11.8125, + "learning_rate": 2.213436001116838e-07, + "loss": 0.647, + "step": 18975 + }, + { + "epoch": 1.6170430336599915, + "grad_norm": 13.5, + "learning_rate": 2.2124814003428877e-07, + "loss": 0.3172, + "step": 18976 + }, + { + "epoch": 1.6171282488282914, + "grad_norm": 11.3125, + "learning_rate": 2.2115269854747407e-07, + "loss": 0.4153, + "step": 18977 + }, + { + "epoch": 1.6172134639965914, + "grad_norm": 17.875, + "learning_rate": 2.2105727565296469e-07, + "loss": 0.7889, + "step": 18978 + }, + { + "epoch": 1.6172986791648913, + "grad_norm": 17.875, + "learning_rate": 2.209618713524854e-07, + "loss": 0.7738, + "step": 18979 + }, + { + "epoch": 1.6173838943331913, + "grad_norm": 14.5625, + "learning_rate": 2.208664856477595e-07, + "loss": 0.6177, + "step": 18980 + }, + { + "epoch": 1.6174691095014913, + "grad_norm": 15.375, + "learning_rate": 2.2077111854051106e-07, + "loss": 0.6108, + "step": 18981 + }, + { + "epoch": 1.6175543246697912, + "grad_norm": 19.25, + "learning_rate": 2.2067577003246354e-07, + "loss": 0.7637, + "step": 18982 + }, + { + "epoch": 1.6176395398380912, + "grad_norm": 18.375, + "learning_rate": 2.2058044012534017e-07, + "loss": 0.8011, + "step": 18983 + }, + { + "epoch": 1.6177247550063911, + "grad_norm": 42.5, + "learning_rate": 2.2048512882086343e-07, + "loss": 1.0336, + "step": 18984 + }, + { + "epoch": 1.617809970174691, + "grad_norm": 18.0, + "learning_rate": 2.2038983612075542e-07, + "loss": 0.8031, + "step": 18985 + }, + { + "epoch": 1.617895185342991, + "grad_norm": 13.125, + "learning_rate": 2.2029456202673826e-07, + "loss": 0.5233, + "step": 18986 + }, + { + "epoch": 1.617980400511291, + "grad_norm": 16.25, + "learning_rate": 2.201993065405343e-07, + "loss": 0.624, + "step": 18987 + }, + { + "epoch": 1.618065615679591, + "grad_norm": 9.375, + "learning_rate": 2.20104069663864e-07, + "loss": 0.2774, + "step": 18988 + }, + { + "epoch": 1.618150830847891, + "grad_norm": 18.25, + "learning_rate": 2.2000885139844886e-07, + "loss": 0.6238, + "step": 18989 + }, + { + "epoch": 1.6182360460161909, + "grad_norm": 19.0, + "learning_rate": 2.1991365174600944e-07, + "loss": 0.7261, + "step": 18990 + }, + { + "epoch": 1.6183212611844908, + "grad_norm": 14.6875, + "learning_rate": 2.1981847070826664e-07, + "loss": 0.469, + "step": 18991 + }, + { + "epoch": 1.6184064763527908, + "grad_norm": 22.75, + "learning_rate": 2.197233082869399e-07, + "loss": 0.8964, + "step": 18992 + }, + { + "epoch": 1.6184916915210907, + "grad_norm": 9.9375, + "learning_rate": 2.1962816448374864e-07, + "loss": 0.2543, + "step": 18993 + }, + { + "epoch": 1.6185769066893907, + "grad_norm": 16.5, + "learning_rate": 2.1953303930041262e-07, + "loss": 0.6343, + "step": 18994 + }, + { + "epoch": 1.6186621218576907, + "grad_norm": 13.75, + "learning_rate": 2.1943793273865074e-07, + "loss": 0.5399, + "step": 18995 + }, + { + "epoch": 1.6187473370259906, + "grad_norm": 14.5625, + "learning_rate": 2.1934284480018207e-07, + "loss": 0.4913, + "step": 18996 + }, + { + "epoch": 1.6188325521942906, + "grad_norm": 20.0, + "learning_rate": 2.1924777548672436e-07, + "loss": 0.701, + "step": 18997 + }, + { + "epoch": 1.6189177673625905, + "grad_norm": 14.9375, + "learning_rate": 2.1915272479999568e-07, + "loss": 0.3874, + "step": 18998 + }, + { + "epoch": 1.6190029825308905, + "grad_norm": 13.5, + "learning_rate": 2.19057692741714e-07, + "loss": 0.5942, + "step": 18999 + }, + { + "epoch": 1.6190881976991904, + "grad_norm": 19.625, + "learning_rate": 2.1896267931359711e-07, + "loss": 0.5702, + "step": 19000 + }, + { + "epoch": 1.6191734128674904, + "grad_norm": 13.9375, + "learning_rate": 2.1886768451736067e-07, + "loss": 0.358, + "step": 19001 + }, + { + "epoch": 1.6192586280357903, + "grad_norm": 22.25, + "learning_rate": 2.1877270835472225e-07, + "loss": 0.79, + "step": 19002 + }, + { + "epoch": 1.6193438432040903, + "grad_norm": 12.9375, + "learning_rate": 2.186777508273978e-07, + "loss": 0.4998, + "step": 19003 + }, + { + "epoch": 1.6194290583723903, + "grad_norm": 17.375, + "learning_rate": 2.185828119371039e-07, + "loss": 0.8208, + "step": 19004 + }, + { + "epoch": 1.6195142735406902, + "grad_norm": 14.8125, + "learning_rate": 2.1848789168555527e-07, + "loss": 0.4181, + "step": 19005 + }, + { + "epoch": 1.6195994887089902, + "grad_norm": 10.0625, + "learning_rate": 2.1839299007446789e-07, + "loss": 0.8593, + "step": 19006 + }, + { + "epoch": 1.6196847038772901, + "grad_norm": 11.5, + "learning_rate": 2.1829810710555648e-07, + "loss": 0.3258, + "step": 19007 + }, + { + "epoch": 1.61976991904559, + "grad_norm": 26.75, + "learning_rate": 2.1820324278053598e-07, + "loss": 0.609, + "step": 19008 + }, + { + "epoch": 1.61985513421389, + "grad_norm": 30.625, + "learning_rate": 2.1810839710112053e-07, + "loss": 1.1832, + "step": 19009 + }, + { + "epoch": 1.61994034938219, + "grad_norm": 16.75, + "learning_rate": 2.1801357006902364e-07, + "loss": 0.3867, + "step": 19010 + }, + { + "epoch": 1.62002556455049, + "grad_norm": 16.875, + "learning_rate": 2.179187616859592e-07, + "loss": 0.6775, + "step": 19011 + }, + { + "epoch": 1.62011077971879, + "grad_norm": 14.0625, + "learning_rate": 2.1782397195364068e-07, + "loss": 0.3081, + "step": 19012 + }, + { + "epoch": 1.6201959948870899, + "grad_norm": 11.75, + "learning_rate": 2.177292008737812e-07, + "loss": 0.5052, + "step": 19013 + }, + { + "epoch": 1.6202812100553898, + "grad_norm": 16.625, + "learning_rate": 2.1763444844809272e-07, + "loss": 0.5751, + "step": 19014 + }, + { + "epoch": 1.6203664252236898, + "grad_norm": 17.0, + "learning_rate": 2.1753971467828802e-07, + "loss": 0.6328, + "step": 19015 + }, + { + "epoch": 1.6204516403919897, + "grad_norm": 18.625, + "learning_rate": 2.1744499956607905e-07, + "loss": 0.5177, + "step": 19016 + }, + { + "epoch": 1.6205368555602897, + "grad_norm": 16.25, + "learning_rate": 2.173503031131774e-07, + "loss": 0.5115, + "step": 19017 + }, + { + "epoch": 1.6206220707285897, + "grad_norm": 12.375, + "learning_rate": 2.1725562532129374e-07, + "loss": 0.3917, + "step": 19018 + }, + { + "epoch": 1.6207072858968896, + "grad_norm": 22.375, + "learning_rate": 2.171609661921395e-07, + "loss": 0.7338, + "step": 19019 + }, + { + "epoch": 1.6207925010651896, + "grad_norm": 23.375, + "learning_rate": 2.1706632572742513e-07, + "loss": 0.8863, + "step": 19020 + }, + { + "epoch": 1.6208777162334895, + "grad_norm": 14.8125, + "learning_rate": 2.169717039288613e-07, + "loss": 0.4285, + "step": 19021 + }, + { + "epoch": 1.6209629314017895, + "grad_norm": 24.0, + "learning_rate": 2.1687710079815738e-07, + "loss": 1.2616, + "step": 19022 + }, + { + "epoch": 1.6210481465700894, + "grad_norm": 16.25, + "learning_rate": 2.1678251633702296e-07, + "loss": 0.6176, + "step": 19023 + }, + { + "epoch": 1.6211333617383894, + "grad_norm": 16.25, + "learning_rate": 2.1668795054716753e-07, + "loss": 0.7763, + "step": 19024 + }, + { + "epoch": 1.6212185769066894, + "grad_norm": 13.5625, + "learning_rate": 2.1659340343030022e-07, + "loss": 0.599, + "step": 19025 + }, + { + "epoch": 1.6213037920749893, + "grad_norm": 18.625, + "learning_rate": 2.164988749881293e-07, + "loss": 0.6235, + "step": 19026 + }, + { + "epoch": 1.6213890072432893, + "grad_norm": 19.25, + "learning_rate": 2.1640436522236268e-07, + "loss": 0.8484, + "step": 19027 + }, + { + "epoch": 1.6214742224115892, + "grad_norm": 13.1875, + "learning_rate": 2.1630987413470847e-07, + "loss": 0.5622, + "step": 19028 + }, + { + "epoch": 1.6215594375798892, + "grad_norm": 22.125, + "learning_rate": 2.162154017268743e-07, + "loss": 0.7588, + "step": 19029 + }, + { + "epoch": 1.6216446527481891, + "grad_norm": 15.0625, + "learning_rate": 2.161209480005677e-07, + "loss": 0.7635, + "step": 19030 + }, + { + "epoch": 1.621729867916489, + "grad_norm": 13.1875, + "learning_rate": 2.1602651295749497e-07, + "loss": 0.2991, + "step": 19031 + }, + { + "epoch": 1.621815083084789, + "grad_norm": 16.125, + "learning_rate": 2.1593209659936277e-07, + "loss": 0.6824, + "step": 19032 + }, + { + "epoch": 1.621900298253089, + "grad_norm": 14.25, + "learning_rate": 2.158376989278778e-07, + "loss": 0.605, + "step": 19033 + }, + { + "epoch": 1.621985513421389, + "grad_norm": 22.75, + "learning_rate": 2.1574331994474547e-07, + "loss": 1.0675, + "step": 19034 + }, + { + "epoch": 1.622070728589689, + "grad_norm": 18.5, + "learning_rate": 2.1564895965167116e-07, + "loss": 0.5317, + "step": 19035 + }, + { + "epoch": 1.6221559437579889, + "grad_norm": 39.5, + "learning_rate": 2.1555461805036023e-07, + "loss": 0.7862, + "step": 19036 + }, + { + "epoch": 1.6222411589262888, + "grad_norm": 12.5, + "learning_rate": 2.1546029514251762e-07, + "loss": 0.3959, + "step": 19037 + }, + { + "epoch": 1.6223263740945888, + "grad_norm": 12.25, + "learning_rate": 2.1536599092984806e-07, + "loss": 0.387, + "step": 19038 + }, + { + "epoch": 1.6224115892628888, + "grad_norm": 15.3125, + "learning_rate": 2.1527170541405518e-07, + "loss": 0.4703, + "step": 19039 + }, + { + "epoch": 1.6224968044311887, + "grad_norm": 17.75, + "learning_rate": 2.151774385968429e-07, + "loss": 0.7273, + "step": 19040 + }, + { + "epoch": 1.6225820195994887, + "grad_norm": 16.0, + "learning_rate": 2.1508319047991542e-07, + "loss": 0.8319, + "step": 19041 + }, + { + "epoch": 1.6226672347677886, + "grad_norm": 14.5, + "learning_rate": 2.1498896106497496e-07, + "loss": 0.6021, + "step": 19042 + }, + { + "epoch": 1.6227524499360886, + "grad_norm": 11.8125, + "learning_rate": 2.1489475035372507e-07, + "loss": 0.3779, + "step": 19043 + }, + { + "epoch": 1.6228376651043885, + "grad_norm": 19.375, + "learning_rate": 2.1480055834786767e-07, + "loss": 0.5258, + "step": 19044 + }, + { + "epoch": 1.6229228802726885, + "grad_norm": 18.25, + "learning_rate": 2.1470638504910505e-07, + "loss": 0.6699, + "step": 19045 + }, + { + "epoch": 1.6230080954409885, + "grad_norm": 15.9375, + "learning_rate": 2.146122304591393e-07, + "loss": 0.5442, + "step": 19046 + }, + { + "epoch": 1.6230933106092884, + "grad_norm": 19.625, + "learning_rate": 2.1451809457967183e-07, + "loss": 0.5053, + "step": 19047 + }, + { + "epoch": 1.6231785257775884, + "grad_norm": 12.625, + "learning_rate": 2.1442397741240352e-07, + "loss": 0.4152, + "step": 19048 + }, + { + "epoch": 1.6232637409458883, + "grad_norm": 40.0, + "learning_rate": 2.1432987895903547e-07, + "loss": 0.4815, + "step": 19049 + }, + { + "epoch": 1.6233489561141883, + "grad_norm": 16.125, + "learning_rate": 2.1423579922126764e-07, + "loss": 0.488, + "step": 19050 + }, + { + "epoch": 1.6234341712824882, + "grad_norm": 16.625, + "learning_rate": 2.1414173820080085e-07, + "loss": 0.9597, + "step": 19051 + }, + { + "epoch": 1.6235193864507882, + "grad_norm": 19.375, + "learning_rate": 2.14047695899334e-07, + "loss": 0.7146, + "step": 19052 + }, + { + "epoch": 1.6236046016190882, + "grad_norm": 12.6875, + "learning_rate": 2.1395367231856715e-07, + "loss": 0.3526, + "step": 19053 + }, + { + "epoch": 1.623689816787388, + "grad_norm": 17.75, + "learning_rate": 2.1385966746019935e-07, + "loss": 0.5707, + "step": 19054 + }, + { + "epoch": 1.623775031955688, + "grad_norm": 10.125, + "learning_rate": 2.137656813259295e-07, + "loss": 0.2695, + "step": 19055 + }, + { + "epoch": 1.623860247123988, + "grad_norm": 20.0, + "learning_rate": 2.1367171391745545e-07, + "loss": 0.8899, + "step": 19056 + }, + { + "epoch": 1.623945462292288, + "grad_norm": 15.1875, + "learning_rate": 2.1357776523647593e-07, + "loss": 0.806, + "step": 19057 + }, + { + "epoch": 1.624030677460588, + "grad_norm": 14.5625, + "learning_rate": 2.1348383528468824e-07, + "loss": 0.6306, + "step": 19058 + }, + { + "epoch": 1.624115892628888, + "grad_norm": 14.8125, + "learning_rate": 2.1338992406378976e-07, + "loss": 0.725, + "step": 19059 + }, + { + "epoch": 1.6242011077971878, + "grad_norm": 16.75, + "learning_rate": 2.1329603157547811e-07, + "loss": 0.5382, + "step": 19060 + }, + { + "epoch": 1.6242863229654878, + "grad_norm": 16.375, + "learning_rate": 2.132021578214495e-07, + "loss": 0.4606, + "step": 19061 + }, + { + "epoch": 1.6243715381337878, + "grad_norm": 17.625, + "learning_rate": 2.1310830280340034e-07, + "loss": 0.8391, + "step": 19062 + }, + { + "epoch": 1.6244567533020877, + "grad_norm": 10.8125, + "learning_rate": 2.1301446652302715e-07, + "loss": 0.7079, + "step": 19063 + }, + { + "epoch": 1.6245419684703877, + "grad_norm": 23.25, + "learning_rate": 2.12920648982025e-07, + "loss": 0.8438, + "step": 19064 + }, + { + "epoch": 1.6246271836386876, + "grad_norm": 17.125, + "learning_rate": 2.1282685018208986e-07, + "loss": 0.7646, + "step": 19065 + }, + { + "epoch": 1.6247123988069876, + "grad_norm": 14.9375, + "learning_rate": 2.1273307012491626e-07, + "loss": 0.517, + "step": 19066 + }, + { + "epoch": 1.6247976139752875, + "grad_norm": 29.75, + "learning_rate": 2.1263930881219904e-07, + "loss": 0.6678, + "step": 19067 + }, + { + "epoch": 1.6248828291435875, + "grad_norm": 15.625, + "learning_rate": 2.1254556624563314e-07, + "loss": 0.5106, + "step": 19068 + }, + { + "epoch": 1.6249680443118875, + "grad_norm": 14.9375, + "learning_rate": 2.124518424269116e-07, + "loss": 0.4258, + "step": 19069 + }, + { + "epoch": 1.6250532594801874, + "grad_norm": 20.5, + "learning_rate": 2.1235813735772866e-07, + "loss": 0.7577, + "step": 19070 + }, + { + "epoch": 1.6251384746484874, + "grad_norm": 15.625, + "learning_rate": 2.122644510397777e-07, + "loss": 0.8039, + "step": 19071 + }, + { + "epoch": 1.6252236898167873, + "grad_norm": 15.375, + "learning_rate": 2.121707834747519e-07, + "loss": 0.646, + "step": 19072 + }, + { + "epoch": 1.6253089049850873, + "grad_norm": 13.875, + "learning_rate": 2.1207713466434354e-07, + "loss": 0.7292, + "step": 19073 + }, + { + "epoch": 1.6253941201533872, + "grad_norm": 21.0, + "learning_rate": 2.119835046102449e-07, + "loss": 1.0844, + "step": 19074 + }, + { + "epoch": 1.6254793353216872, + "grad_norm": 15.3125, + "learning_rate": 2.1188989331414806e-07, + "loss": 0.5018, + "step": 19075 + }, + { + "epoch": 1.6255645504899872, + "grad_norm": 17.625, + "learning_rate": 2.117963007777449e-07, + "loss": 0.7229, + "step": 19076 + }, + { + "epoch": 1.6256497656582871, + "grad_norm": 14.9375, + "learning_rate": 2.1170272700272664e-07, + "loss": 0.6906, + "step": 19077 + }, + { + "epoch": 1.625734980826587, + "grad_norm": 15.375, + "learning_rate": 2.1160917199078408e-07, + "loss": 0.6723, + "step": 19078 + }, + { + "epoch": 1.625820195994887, + "grad_norm": 14.125, + "learning_rate": 2.1151563574360795e-07, + "loss": 0.5626, + "step": 19079 + }, + { + "epoch": 1.625905411163187, + "grad_norm": 16.625, + "learning_rate": 2.114221182628888e-07, + "loss": 0.4812, + "step": 19080 + }, + { + "epoch": 1.625990626331487, + "grad_norm": 17.0, + "learning_rate": 2.11328619550316e-07, + "loss": 0.5573, + "step": 19081 + }, + { + "epoch": 1.626075841499787, + "grad_norm": 11.4375, + "learning_rate": 2.1123513960757986e-07, + "loss": 0.1548, + "step": 19082 + }, + { + "epoch": 1.6261610566680869, + "grad_norm": 12.0, + "learning_rate": 2.11141678436369e-07, + "loss": 0.3418, + "step": 19083 + }, + { + "epoch": 1.6262462718363868, + "grad_norm": 20.375, + "learning_rate": 2.110482360383727e-07, + "loss": 0.8664, + "step": 19084 + }, + { + "epoch": 1.6263314870046868, + "grad_norm": 12.625, + "learning_rate": 2.109548124152798e-07, + "loss": 0.419, + "step": 19085 + }, + { + "epoch": 1.6264167021729867, + "grad_norm": 15.6875, + "learning_rate": 2.10861407568778e-07, + "loss": 0.6381, + "step": 19086 + }, + { + "epoch": 1.6265019173412867, + "grad_norm": 11.3125, + "learning_rate": 2.1076802150055544e-07, + "loss": 0.2664, + "step": 19087 + }, + { + "epoch": 1.6265871325095866, + "grad_norm": 15.1875, + "learning_rate": 2.1067465421229977e-07, + "loss": 0.6256, + "step": 19088 + }, + { + "epoch": 1.6266723476778866, + "grad_norm": 21.375, + "learning_rate": 2.1058130570569866e-07, + "loss": 0.6922, + "step": 19089 + }, + { + "epoch": 1.6267575628461866, + "grad_norm": 20.625, + "learning_rate": 2.104879759824384e-07, + "loss": 1.0169, + "step": 19090 + }, + { + "epoch": 1.6268427780144865, + "grad_norm": 17.125, + "learning_rate": 2.103946650442054e-07, + "loss": 0.6213, + "step": 19091 + }, + { + "epoch": 1.6269279931827865, + "grad_norm": 15.8125, + "learning_rate": 2.103013728926863e-07, + "loss": 0.6833, + "step": 19092 + }, + { + "epoch": 1.6270132083510864, + "grad_norm": 18.875, + "learning_rate": 2.1020809952956732e-07, + "loss": 0.3505, + "step": 19093 + }, + { + "epoch": 1.6270984235193864, + "grad_norm": 18.125, + "learning_rate": 2.101148449565331e-07, + "loss": 0.5466, + "step": 19094 + }, + { + "epoch": 1.6271836386876863, + "grad_norm": 17.875, + "learning_rate": 2.1002160917526923e-07, + "loss": 0.5418, + "step": 19095 + }, + { + "epoch": 1.6272688538559863, + "grad_norm": 14.75, + "learning_rate": 2.0992839218746077e-07, + "loss": 0.6152, + "step": 19096 + }, + { + "epoch": 1.6273540690242863, + "grad_norm": 15.625, + "learning_rate": 2.0983519399479248e-07, + "loss": 0.7784, + "step": 19097 + }, + { + "epoch": 1.6274392841925862, + "grad_norm": 18.875, + "learning_rate": 2.0974201459894816e-07, + "loss": 0.707, + "step": 19098 + }, + { + "epoch": 1.6275244993608862, + "grad_norm": 15.3125, + "learning_rate": 2.0964885400161145e-07, + "loss": 0.6867, + "step": 19099 + }, + { + "epoch": 1.6276097145291861, + "grad_norm": 17.125, + "learning_rate": 2.0955571220446599e-07, + "loss": 0.7038, + "step": 19100 + }, + { + "epoch": 1.627694929697486, + "grad_norm": 12.5625, + "learning_rate": 2.094625892091952e-07, + "loss": 0.4516, + "step": 19101 + }, + { + "epoch": 1.627780144865786, + "grad_norm": 14.125, + "learning_rate": 2.0936948501748198e-07, + "loss": 0.5599, + "step": 19102 + }, + { + "epoch": 1.627865360034086, + "grad_norm": 15.1875, + "learning_rate": 2.0927639963100836e-07, + "loss": 0.4817, + "step": 19103 + }, + { + "epoch": 1.627950575202386, + "grad_norm": 16.75, + "learning_rate": 2.0918333305145658e-07, + "loss": 0.6433, + "step": 19104 + }, + { + "epoch": 1.628035790370686, + "grad_norm": 10.4375, + "learning_rate": 2.0909028528050878e-07, + "loss": 0.2253, + "step": 19105 + }, + { + "epoch": 1.6281210055389859, + "grad_norm": 18.25, + "learning_rate": 2.0899725631984642e-07, + "loss": 0.6834, + "step": 19106 + }, + { + "epoch": 1.6282062207072858, + "grad_norm": 16.25, + "learning_rate": 2.0890424617115047e-07, + "loss": 0.5394, + "step": 19107 + }, + { + "epoch": 1.6282914358755858, + "grad_norm": 13.8125, + "learning_rate": 2.088112548361014e-07, + "loss": 0.5533, + "step": 19108 + }, + { + "epoch": 1.6283766510438857, + "grad_norm": 13.0, + "learning_rate": 2.0871828231637998e-07, + "loss": 0.5744, + "step": 19109 + }, + { + "epoch": 1.6284618662121857, + "grad_norm": 14.375, + "learning_rate": 2.086253286136665e-07, + "loss": 0.713, + "step": 19110 + }, + { + "epoch": 1.6285470813804857, + "grad_norm": 18.625, + "learning_rate": 2.0853239372964032e-07, + "loss": 0.7767, + "step": 19111 + }, + { + "epoch": 1.6286322965487856, + "grad_norm": 13.5625, + "learning_rate": 2.0843947766598107e-07, + "loss": 0.672, + "step": 19112 + }, + { + "epoch": 1.6287175117170856, + "grad_norm": 18.75, + "learning_rate": 2.083465804243677e-07, + "loss": 0.5927, + "step": 19113 + }, + { + "epoch": 1.6288027268853855, + "grad_norm": 16.0, + "learning_rate": 2.0825370200647955e-07, + "loss": 0.7323, + "step": 19114 + }, + { + "epoch": 1.6288879420536855, + "grad_norm": 23.375, + "learning_rate": 2.0816084241399444e-07, + "loss": 0.7287, + "step": 19115 + }, + { + "epoch": 1.6289731572219854, + "grad_norm": 13.25, + "learning_rate": 2.0806800164859032e-07, + "loss": 0.4175, + "step": 19116 + }, + { + "epoch": 1.6290583723902854, + "grad_norm": 24.625, + "learning_rate": 2.0797517971194503e-07, + "loss": 0.8377, + "step": 19117 + }, + { + "epoch": 1.6291435875585853, + "grad_norm": 13.375, + "learning_rate": 2.0788237660573626e-07, + "loss": 0.6084, + "step": 19118 + }, + { + "epoch": 1.6292288027268853, + "grad_norm": 13.5625, + "learning_rate": 2.07789592331641e-07, + "loss": 0.4501, + "step": 19119 + }, + { + "epoch": 1.6293140178951853, + "grad_norm": 16.0, + "learning_rate": 2.076968268913357e-07, + "loss": 0.5525, + "step": 19120 + }, + { + "epoch": 1.6293992330634852, + "grad_norm": 14.8125, + "learning_rate": 2.0760408028649675e-07, + "loss": 0.5471, + "step": 19121 + }, + { + "epoch": 1.6294844482317852, + "grad_norm": 12.75, + "learning_rate": 2.0751135251880063e-07, + "loss": 0.2358, + "step": 19122 + }, + { + "epoch": 1.6295696634000851, + "grad_norm": 21.25, + "learning_rate": 2.0741864358992268e-07, + "loss": 0.8726, + "step": 19123 + }, + { + "epoch": 1.629654878568385, + "grad_norm": 15.875, + "learning_rate": 2.073259535015379e-07, + "loss": 0.4473, + "step": 19124 + }, + { + "epoch": 1.629740093736685, + "grad_norm": 15.875, + "learning_rate": 2.0723328225532162e-07, + "loss": 0.6868, + "step": 19125 + }, + { + "epoch": 1.629825308904985, + "grad_norm": 12.75, + "learning_rate": 2.0714062985294854e-07, + "loss": 0.549, + "step": 19126 + }, + { + "epoch": 1.629910524073285, + "grad_norm": 11.375, + "learning_rate": 2.070479962960932e-07, + "loss": 0.74, + "step": 19127 + }, + { + "epoch": 1.629995739241585, + "grad_norm": 16.375, + "learning_rate": 2.0695538158642904e-07, + "loss": 0.6589, + "step": 19128 + }, + { + "epoch": 1.6300809544098849, + "grad_norm": 18.25, + "learning_rate": 2.0686278572563002e-07, + "loss": 0.7531, + "step": 19129 + }, + { + "epoch": 1.6301661695781848, + "grad_norm": 19.25, + "learning_rate": 2.0677020871536973e-07, + "loss": 0.9152, + "step": 19130 + }, + { + "epoch": 1.6302513847464848, + "grad_norm": 14.75, + "learning_rate": 2.0667765055732052e-07, + "loss": 0.8161, + "step": 19131 + }, + { + "epoch": 1.6303365999147847, + "grad_norm": 22.0, + "learning_rate": 2.065851112531557e-07, + "loss": 1.0395, + "step": 19132 + }, + { + "epoch": 1.6304218150830847, + "grad_norm": 18.25, + "learning_rate": 2.0649259080454674e-07, + "loss": 0.8496, + "step": 19133 + }, + { + "epoch": 1.6305070302513847, + "grad_norm": 19.125, + "learning_rate": 2.0640008921316595e-07, + "loss": 0.5147, + "step": 19134 + }, + { + "epoch": 1.6305922454196846, + "grad_norm": 22.875, + "learning_rate": 2.06307606480685e-07, + "loss": 0.5134, + "step": 19135 + }, + { + "epoch": 1.6306774605879846, + "grad_norm": 11.75, + "learning_rate": 2.062151426087755e-07, + "loss": 0.4841, + "step": 19136 + }, + { + "epoch": 1.6307626757562845, + "grad_norm": 15.5625, + "learning_rate": 2.0612269759910763e-07, + "loss": 0.5701, + "step": 19137 + }, + { + "epoch": 1.6308478909245845, + "grad_norm": 21.0, + "learning_rate": 2.0603027145335268e-07, + "loss": 1.0739, + "step": 19138 + }, + { + "epoch": 1.6309331060928844, + "grad_norm": 13.375, + "learning_rate": 2.0593786417318005e-07, + "loss": 0.4565, + "step": 19139 + }, + { + "epoch": 1.6310183212611844, + "grad_norm": 18.125, + "learning_rate": 2.0584547576026057e-07, + "loss": 0.6731, + "step": 19140 + }, + { + "epoch": 1.6311035364294844, + "grad_norm": 26.75, + "learning_rate": 2.0575310621626293e-07, + "loss": 0.8974, + "step": 19141 + }, + { + "epoch": 1.6311887515977843, + "grad_norm": 12.875, + "learning_rate": 2.056607555428568e-07, + "loss": 0.5142, + "step": 19142 + }, + { + "epoch": 1.6312739667660843, + "grad_norm": 11.0625, + "learning_rate": 2.0556842374171092e-07, + "loss": 0.3983, + "step": 19143 + }, + { + "epoch": 1.6313591819343842, + "grad_norm": 15.625, + "learning_rate": 2.054761108144941e-07, + "loss": 0.8841, + "step": 19144 + }, + { + "epoch": 1.6314443971026842, + "grad_norm": 11.3125, + "learning_rate": 2.053838167628741e-07, + "loss": 0.2462, + "step": 19145 + }, + { + "epoch": 1.6315296122709841, + "grad_norm": 14.875, + "learning_rate": 2.052915415885194e-07, + "loss": 0.6933, + "step": 19146 + }, + { + "epoch": 1.631614827439284, + "grad_norm": 15.625, + "learning_rate": 2.051992852930966e-07, + "loss": 0.5042, + "step": 19147 + }, + { + "epoch": 1.631700042607584, + "grad_norm": 17.0, + "learning_rate": 2.0510704787827337e-07, + "loss": 0.8469, + "step": 19148 + }, + { + "epoch": 1.631785257775884, + "grad_norm": 14.0625, + "learning_rate": 2.0501482934571687e-07, + "loss": 0.5831, + "step": 19149 + }, + { + "epoch": 1.631870472944184, + "grad_norm": 17.5, + "learning_rate": 2.0492262969709283e-07, + "loss": 0.5598, + "step": 19150 + }, + { + "epoch": 1.631955688112484, + "grad_norm": 29.25, + "learning_rate": 2.0483044893406786e-07, + "loss": 1.0469, + "step": 19151 + }, + { + "epoch": 1.6320409032807839, + "grad_norm": 17.375, + "learning_rate": 2.047382870583077e-07, + "loss": 0.8309, + "step": 19152 + }, + { + "epoch": 1.6321261184490838, + "grad_norm": 14.0, + "learning_rate": 2.0464614407147795e-07, + "loss": 0.5365, + "step": 19153 + }, + { + "epoch": 1.6322113336173838, + "grad_norm": 19.875, + "learning_rate": 2.045540199752434e-07, + "loss": 0.6484, + "step": 19154 + }, + { + "epoch": 1.6322965487856838, + "grad_norm": 23.0, + "learning_rate": 2.0446191477126927e-07, + "loss": 0.6926, + "step": 19155 + }, + { + "epoch": 1.6323817639539837, + "grad_norm": 20.875, + "learning_rate": 2.043698284612193e-07, + "loss": 0.8586, + "step": 19156 + }, + { + "epoch": 1.6324669791222837, + "grad_norm": 15.0625, + "learning_rate": 2.0427776104675834e-07, + "loss": 0.6592, + "step": 19157 + }, + { + "epoch": 1.6325521942905836, + "grad_norm": 17.0, + "learning_rate": 2.041857125295496e-07, + "loss": 0.4598, + "step": 19158 + }, + { + "epoch": 1.6326374094588836, + "grad_norm": 14.1875, + "learning_rate": 2.040936829112565e-07, + "loss": 0.545, + "step": 19159 + }, + { + "epoch": 1.6327226246271835, + "grad_norm": 18.75, + "learning_rate": 2.0400167219354243e-07, + "loss": 0.7947, + "step": 19160 + }, + { + "epoch": 1.6328078397954835, + "grad_norm": 16.125, + "learning_rate": 2.0390968037807017e-07, + "loss": 0.7596, + "step": 19161 + }, + { + "epoch": 1.6328930549637835, + "grad_norm": 17.875, + "learning_rate": 2.0381770746650168e-07, + "loss": 0.5368, + "step": 19162 + }, + { + "epoch": 1.6329782701320834, + "grad_norm": 12.75, + "learning_rate": 2.0372575346049958e-07, + "loss": 0.3962, + "step": 19163 + }, + { + "epoch": 1.6330634853003834, + "grad_norm": 14.5, + "learning_rate": 2.0363381836172493e-07, + "loss": 0.4097, + "step": 19164 + }, + { + "epoch": 1.6331487004686833, + "grad_norm": 14.25, + "learning_rate": 2.0354190217183935e-07, + "loss": 0.6555, + "step": 19165 + }, + { + "epoch": 1.6332339156369833, + "grad_norm": 14.4375, + "learning_rate": 2.0345000489250408e-07, + "loss": 0.6969, + "step": 19166 + }, + { + "epoch": 1.6333191308052832, + "grad_norm": 21.0, + "learning_rate": 2.0335812652537938e-07, + "loss": 0.9809, + "step": 19167 + }, + { + "epoch": 1.6334043459735832, + "grad_norm": 24.5, + "learning_rate": 2.032662670721258e-07, + "loss": 0.7171, + "step": 19168 + }, + { + "epoch": 1.6334895611418832, + "grad_norm": 18.75, + "learning_rate": 2.031744265344035e-07, + "loss": 0.874, + "step": 19169 + }, + { + "epoch": 1.633574776310183, + "grad_norm": 12.25, + "learning_rate": 2.0308260491387204e-07, + "loss": 0.388, + "step": 19170 + }, + { + "epoch": 1.633659991478483, + "grad_norm": 14.5, + "learning_rate": 2.0299080221219082e-07, + "loss": 0.4038, + "step": 19171 + }, + { + "epoch": 1.633745206646783, + "grad_norm": 22.5, + "learning_rate": 2.028990184310184e-07, + "loss": 0.8247, + "step": 19172 + }, + { + "epoch": 1.633830421815083, + "grad_norm": 11.5, + "learning_rate": 2.0280725357201354e-07, + "loss": 0.2662, + "step": 19173 + }, + { + "epoch": 1.633915636983383, + "grad_norm": 15.375, + "learning_rate": 2.0271550763683493e-07, + "loss": 0.7404, + "step": 19174 + }, + { + "epoch": 1.634000852151683, + "grad_norm": 15.875, + "learning_rate": 2.0262378062714e-07, + "loss": 0.7432, + "step": 19175 + }, + { + "epoch": 1.6340860673199828, + "grad_norm": 25.625, + "learning_rate": 2.0253207254458647e-07, + "loss": 0.6961, + "step": 19176 + }, + { + "epoch": 1.6341712824882828, + "grad_norm": 17.25, + "learning_rate": 2.0244038339083185e-07, + "loss": 0.7056, + "step": 19177 + }, + { + "epoch": 1.6342564976565828, + "grad_norm": 14.1875, + "learning_rate": 2.0234871316753308e-07, + "loss": 0.6173, + "step": 19178 + }, + { + "epoch": 1.6343417128248827, + "grad_norm": 14.3125, + "learning_rate": 2.0225706187634655e-07, + "loss": 0.454, + "step": 19179 + }, + { + "epoch": 1.6344269279931827, + "grad_norm": 34.75, + "learning_rate": 2.0216542951892833e-07, + "loss": 0.7762, + "step": 19180 + }, + { + "epoch": 1.6345121431614826, + "grad_norm": 12.1875, + "learning_rate": 2.0207381609693433e-07, + "loss": 0.4802, + "step": 19181 + }, + { + "epoch": 1.6345973583297826, + "grad_norm": 15.75, + "learning_rate": 2.0198222161202042e-07, + "loss": 0.6747, + "step": 19182 + }, + { + "epoch": 1.6346825734980825, + "grad_norm": 26.5, + "learning_rate": 2.018906460658418e-07, + "loss": 1.0639, + "step": 19183 + }, + { + "epoch": 1.6347677886663825, + "grad_norm": 11.9375, + "learning_rate": 2.0179908946005287e-07, + "loss": 0.3698, + "step": 19184 + }, + { + "epoch": 1.6348530038346825, + "grad_norm": 17.5, + "learning_rate": 2.017075517963084e-07, + "loss": 0.512, + "step": 19185 + }, + { + "epoch": 1.6349382190029824, + "grad_norm": 18.0, + "learning_rate": 2.0161603307626302e-07, + "loss": 0.7948, + "step": 19186 + }, + { + "epoch": 1.6350234341712824, + "grad_norm": 12.8125, + "learning_rate": 2.0152453330157001e-07, + "loss": 0.3664, + "step": 19187 + }, + { + "epoch": 1.6351086493395823, + "grad_norm": 16.375, + "learning_rate": 2.0143305247388278e-07, + "loss": 0.5645, + "step": 19188 + }, + { + "epoch": 1.6351938645078823, + "grad_norm": 17.625, + "learning_rate": 2.0134159059485472e-07, + "loss": 0.7854, + "step": 19189 + }, + { + "epoch": 1.6352790796761822, + "grad_norm": 15.8125, + "learning_rate": 2.0125014766613867e-07, + "loss": 0.4956, + "step": 19190 + }, + { + "epoch": 1.6353642948444822, + "grad_norm": 10.375, + "learning_rate": 2.0115872368938734e-07, + "loss": 0.2675, + "step": 19191 + }, + { + "epoch": 1.6354495100127822, + "grad_norm": 14.5625, + "learning_rate": 2.010673186662522e-07, + "loss": 0.4401, + "step": 19192 + }, + { + "epoch": 1.6355347251810821, + "grad_norm": 14.75, + "learning_rate": 2.0097593259838537e-07, + "loss": 0.5151, + "step": 19193 + }, + { + "epoch": 1.635619940349382, + "grad_norm": 14.625, + "learning_rate": 2.0088456548743848e-07, + "loss": 0.6012, + "step": 19194 + }, + { + "epoch": 1.635705155517682, + "grad_norm": 17.125, + "learning_rate": 2.0079321733506255e-07, + "loss": 0.7165, + "step": 19195 + }, + { + "epoch": 1.635790370685982, + "grad_norm": 22.0, + "learning_rate": 2.0070188814290834e-07, + "loss": 0.6291, + "step": 19196 + }, + { + "epoch": 1.635875585854282, + "grad_norm": 14.6875, + "learning_rate": 2.0061057791262593e-07, + "loss": 0.7295, + "step": 19197 + }, + { + "epoch": 1.635960801022582, + "grad_norm": 19.125, + "learning_rate": 2.005192866458655e-07, + "loss": 0.7184, + "step": 19198 + }, + { + "epoch": 1.6360460161908819, + "grad_norm": 13.9375, + "learning_rate": 2.0042801434427702e-07, + "loss": 0.4732, + "step": 19199 + }, + { + "epoch": 1.6361312313591818, + "grad_norm": 14.8125, + "learning_rate": 2.0033676100951e-07, + "loss": 0.5279, + "step": 19200 + }, + { + "epoch": 1.6362164465274818, + "grad_norm": 14.625, + "learning_rate": 2.0024552664321282e-07, + "loss": 0.5978, + "step": 19201 + }, + { + "epoch": 1.6363016616957817, + "grad_norm": 16.875, + "learning_rate": 2.0015431124703472e-07, + "loss": 0.5472, + "step": 19202 + }, + { + "epoch": 1.6363868768640817, + "grad_norm": 18.375, + "learning_rate": 2.000631148226241e-07, + "loss": 0.977, + "step": 19203 + }, + { + "epoch": 1.6364720920323816, + "grad_norm": 14.875, + "learning_rate": 1.9997193737162885e-07, + "loss": 0.6837, + "step": 19204 + }, + { + "epoch": 1.6365573072006816, + "grad_norm": 16.5, + "learning_rate": 1.9988077889569634e-07, + "loss": 0.8697, + "step": 19205 + }, + { + "epoch": 1.6366425223689816, + "grad_norm": 13.9375, + "learning_rate": 1.9978963939647405e-07, + "loss": 0.4756, + "step": 19206 + }, + { + "epoch": 1.6367277375372815, + "grad_norm": 20.875, + "learning_rate": 1.9969851887560897e-07, + "loss": 1.0981, + "step": 19207 + }, + { + "epoch": 1.6368129527055815, + "grad_norm": 14.0625, + "learning_rate": 1.996074173347483e-07, + "loss": 0.6102, + "step": 19208 + }, + { + "epoch": 1.6368981678738814, + "grad_norm": 25.5, + "learning_rate": 1.9951633477553746e-07, + "loss": 0.5971, + "step": 19209 + }, + { + "epoch": 1.6369833830421814, + "grad_norm": 10.8125, + "learning_rate": 1.994252711996228e-07, + "loss": 0.2685, + "step": 19210 + }, + { + "epoch": 1.6370685982104813, + "grad_norm": 18.5, + "learning_rate": 1.9933422660865024e-07, + "loss": 1.0529, + "step": 19211 + }, + { + "epoch": 1.6371538133787813, + "grad_norm": 18.75, + "learning_rate": 1.9924320100426428e-07, + "loss": 0.6649, + "step": 19212 + }, + { + "epoch": 1.6372390285470813, + "grad_norm": 13.4375, + "learning_rate": 1.9915219438811084e-07, + "loss": 0.5465, + "step": 19213 + }, + { + "epoch": 1.6373242437153812, + "grad_norm": 14.875, + "learning_rate": 1.9906120676183347e-07, + "loss": 0.3934, + "step": 19214 + }, + { + "epoch": 1.6374094588836812, + "grad_norm": 14.625, + "learning_rate": 1.9897023812707694e-07, + "loss": 0.67, + "step": 19215 + }, + { + "epoch": 1.6374946740519811, + "grad_norm": 17.5, + "learning_rate": 1.988792884854855e-07, + "loss": 0.4604, + "step": 19216 + }, + { + "epoch": 1.637579889220281, + "grad_norm": 12.5625, + "learning_rate": 1.987883578387019e-07, + "loss": 0.353, + "step": 19217 + }, + { + "epoch": 1.637665104388581, + "grad_norm": 15.625, + "learning_rate": 1.9869744618836978e-07, + "loss": 0.4494, + "step": 19218 + }, + { + "epoch": 1.637750319556881, + "grad_norm": 17.625, + "learning_rate": 1.9860655353613188e-07, + "loss": 0.5565, + "step": 19219 + }, + { + "epoch": 1.637835534725181, + "grad_norm": 16.875, + "learning_rate": 1.9851567988363128e-07, + "loss": 0.9302, + "step": 19220 + }, + { + "epoch": 1.637920749893481, + "grad_norm": 12.9375, + "learning_rate": 1.9842482523250952e-07, + "loss": 0.3044, + "step": 19221 + }, + { + "epoch": 1.6380059650617809, + "grad_norm": 21.25, + "learning_rate": 1.9833398958440844e-07, + "loss": 0.9634, + "step": 19222 + }, + { + "epoch": 1.6380911802300808, + "grad_norm": 15.3125, + "learning_rate": 1.9824317294096964e-07, + "loss": 0.6796, + "step": 19223 + }, + { + "epoch": 1.6381763953983808, + "grad_norm": 13.6875, + "learning_rate": 1.9815237530383431e-07, + "loss": 0.4766, + "step": 19224 + }, + { + "epoch": 1.6382616105666807, + "grad_norm": 19.75, + "learning_rate": 1.9806159667464367e-07, + "loss": 0.8462, + "step": 19225 + }, + { + "epoch": 1.6383468257349807, + "grad_norm": 14.625, + "learning_rate": 1.9797083705503749e-07, + "loss": 0.7259, + "step": 19226 + }, + { + "epoch": 1.6384320409032807, + "grad_norm": 16.75, + "learning_rate": 1.978800964466561e-07, + "loss": 0.7671, + "step": 19227 + }, + { + "epoch": 1.6385172560715806, + "grad_norm": 16.625, + "learning_rate": 1.977893748511396e-07, + "loss": 0.5474, + "step": 19228 + }, + { + "epoch": 1.6386024712398806, + "grad_norm": 12.625, + "learning_rate": 1.97698672270127e-07, + "loss": 0.5465, + "step": 19229 + }, + { + "epoch": 1.6386876864081805, + "grad_norm": 20.75, + "learning_rate": 1.9760798870525788e-07, + "loss": 0.5657, + "step": 19230 + }, + { + "epoch": 1.6387729015764805, + "grad_norm": 20.75, + "learning_rate": 1.975173241581703e-07, + "loss": 0.3315, + "step": 19231 + }, + { + "epoch": 1.6388581167447804, + "grad_norm": 20.375, + "learning_rate": 1.974266786305029e-07, + "loss": 1.0918, + "step": 19232 + }, + { + "epoch": 1.6389433319130804, + "grad_norm": 13.875, + "learning_rate": 1.9733605212389428e-07, + "loss": 0.4472, + "step": 19233 + }, + { + "epoch": 1.6390285470813803, + "grad_norm": 15.8125, + "learning_rate": 1.9724544463998158e-07, + "loss": 0.9391, + "step": 19234 + }, + { + "epoch": 1.6391137622496803, + "grad_norm": 18.0, + "learning_rate": 1.9715485618040208e-07, + "loss": 0.7093, + "step": 19235 + }, + { + "epoch": 1.6391989774179803, + "grad_norm": 18.875, + "learning_rate": 1.9706428674679354e-07, + "loss": 0.8956, + "step": 19236 + }, + { + "epoch": 1.6392841925862802, + "grad_norm": 16.25, + "learning_rate": 1.969737363407917e-07, + "loss": 0.5911, + "step": 19237 + }, + { + "epoch": 1.6393694077545802, + "grad_norm": 18.5, + "learning_rate": 1.9688320496403374e-07, + "loss": 0.4869, + "step": 19238 + }, + { + "epoch": 1.6394546229228801, + "grad_norm": 17.25, + "learning_rate": 1.9679269261815488e-07, + "loss": 0.4933, + "step": 19239 + }, + { + "epoch": 1.63953983809118, + "grad_norm": 16.375, + "learning_rate": 1.9670219930479114e-07, + "loss": 0.9891, + "step": 19240 + }, + { + "epoch": 1.63962505325948, + "grad_norm": 16.5, + "learning_rate": 1.9661172502557793e-07, + "loss": 0.5033, + "step": 19241 + }, + { + "epoch": 1.63971026842778, + "grad_norm": 19.375, + "learning_rate": 1.9652126978215032e-07, + "loss": 0.881, + "step": 19242 + }, + { + "epoch": 1.63979548359608, + "grad_norm": 15.6875, + "learning_rate": 1.9643083357614252e-07, + "loss": 0.641, + "step": 19243 + }, + { + "epoch": 1.63988069876438, + "grad_norm": 13.125, + "learning_rate": 1.9634041640918923e-07, + "loss": 0.4672, + "step": 19244 + }, + { + "epoch": 1.6399659139326799, + "grad_norm": 17.75, + "learning_rate": 1.9625001828292384e-07, + "loss": 0.4367, + "step": 19245 + }, + { + "epoch": 1.6400511291009798, + "grad_norm": 10.75, + "learning_rate": 1.9615963919898047e-07, + "loss": 0.2846, + "step": 19246 + }, + { + "epoch": 1.6401363442692798, + "grad_norm": 20.375, + "learning_rate": 1.9606927915899201e-07, + "loss": 0.6163, + "step": 19247 + }, + { + "epoch": 1.6402215594375797, + "grad_norm": 18.625, + "learning_rate": 1.9597893816459144e-07, + "loss": 0.6651, + "step": 19248 + }, + { + "epoch": 1.6403067746058797, + "grad_norm": 12.0, + "learning_rate": 1.958886162174113e-07, + "loss": 0.2826, + "step": 19249 + }, + { + "epoch": 1.6403919897741797, + "grad_norm": 17.375, + "learning_rate": 1.9579831331908426e-07, + "loss": 0.6941, + "step": 19250 + }, + { + "epoch": 1.6404772049424796, + "grad_norm": 17.25, + "learning_rate": 1.957080294712413e-07, + "loss": 0.4887, + "step": 19251 + }, + { + "epoch": 1.6405624201107796, + "grad_norm": 12.5, + "learning_rate": 1.9561776467551488e-07, + "loss": 0.3961, + "step": 19252 + }, + { + "epoch": 1.6406476352790795, + "grad_norm": 12.75, + "learning_rate": 1.955275189335354e-07, + "loss": 0.4397, + "step": 19253 + }, + { + "epoch": 1.6407328504473795, + "grad_norm": 32.5, + "learning_rate": 1.9543729224693385e-07, + "loss": 0.5277, + "step": 19254 + }, + { + "epoch": 1.6408180656156794, + "grad_norm": 18.0, + "learning_rate": 1.9534708461734138e-07, + "loss": 0.6993, + "step": 19255 + }, + { + "epoch": 1.6409032807839794, + "grad_norm": 15.4375, + "learning_rate": 1.9525689604638716e-07, + "loss": 0.504, + "step": 19256 + }, + { + "epoch": 1.6409884959522794, + "grad_norm": 15.0, + "learning_rate": 1.9516672653570157e-07, + "loss": 0.4653, + "step": 19257 + }, + { + "epoch": 1.6410737111205793, + "grad_norm": 17.5, + "learning_rate": 1.9507657608691388e-07, + "loss": 0.8008, + "step": 19258 + }, + { + "epoch": 1.6411589262888793, + "grad_norm": 15.9375, + "learning_rate": 1.9498644470165363e-07, + "loss": 0.7395, + "step": 19259 + }, + { + "epoch": 1.6412441414571792, + "grad_norm": 11.6875, + "learning_rate": 1.948963323815492e-07, + "loss": 0.2985, + "step": 19260 + }, + { + "epoch": 1.6413293566254792, + "grad_norm": 19.5, + "learning_rate": 1.9480623912822877e-07, + "loss": 0.6626, + "step": 19261 + }, + { + "epoch": 1.6414145717937791, + "grad_norm": 19.25, + "learning_rate": 1.947161649433206e-07, + "loss": 0.6038, + "step": 19262 + }, + { + "epoch": 1.641499786962079, + "grad_norm": 24.0, + "learning_rate": 1.9462610982845284e-07, + "loss": 0.9866, + "step": 19263 + }, + { + "epoch": 1.641585002130379, + "grad_norm": 14.625, + "learning_rate": 1.9453607378525223e-07, + "loss": 0.594, + "step": 19264 + }, + { + "epoch": 1.641670217298679, + "grad_norm": 20.125, + "learning_rate": 1.9444605681534612e-07, + "loss": 0.9843, + "step": 19265 + }, + { + "epoch": 1.641755432466979, + "grad_norm": 12.5625, + "learning_rate": 1.9435605892036122e-07, + "loss": 0.4618, + "step": 19266 + }, + { + "epoch": 1.641840647635279, + "grad_norm": 21.0, + "learning_rate": 1.9426608010192416e-07, + "loss": 0.5147, + "step": 19267 + }, + { + "epoch": 1.6419258628035789, + "grad_norm": 14.3125, + "learning_rate": 1.9417612036166059e-07, + "loss": 0.5629, + "step": 19268 + }, + { + "epoch": 1.6420110779718788, + "grad_norm": 15.4375, + "learning_rate": 1.94086179701196e-07, + "loss": 0.6224, + "step": 19269 + }, + { + "epoch": 1.6420962931401788, + "grad_norm": 13.4375, + "learning_rate": 1.939962581221559e-07, + "loss": 0.4071, + "step": 19270 + }, + { + "epoch": 1.6421815083084788, + "grad_norm": 15.1875, + "learning_rate": 1.939063556261654e-07, + "loss": 0.6087, + "step": 19271 + }, + { + "epoch": 1.6422667234767787, + "grad_norm": 39.75, + "learning_rate": 1.938164722148493e-07, + "loss": 0.5465, + "step": 19272 + }, + { + "epoch": 1.6423519386450787, + "grad_norm": 15.6875, + "learning_rate": 1.937266078898313e-07, + "loss": 0.8606, + "step": 19273 + }, + { + "epoch": 1.6424371538133786, + "grad_norm": 14.125, + "learning_rate": 1.9363676265273565e-07, + "loss": 0.4762, + "step": 19274 + }, + { + "epoch": 1.6425223689816786, + "grad_norm": 13.1875, + "learning_rate": 1.9354693650518592e-07, + "loss": 0.4164, + "step": 19275 + }, + { + "epoch": 1.6426075841499785, + "grad_norm": 10.9375, + "learning_rate": 1.9345712944880607e-07, + "loss": 0.6559, + "step": 19276 + }, + { + "epoch": 1.6426927993182785, + "grad_norm": 12.625, + "learning_rate": 1.9336734148521775e-07, + "loss": 0.3196, + "step": 19277 + }, + { + "epoch": 1.6427780144865785, + "grad_norm": 17.25, + "learning_rate": 1.9327757261604394e-07, + "loss": 0.6215, + "step": 19278 + }, + { + "epoch": 1.6428632296548784, + "grad_norm": 16.875, + "learning_rate": 1.9318782284290711e-07, + "loss": 0.5282, + "step": 19279 + }, + { + "epoch": 1.6429484448231784, + "grad_norm": 16.375, + "learning_rate": 1.9309809216742942e-07, + "loss": 0.4708, + "step": 19280 + }, + { + "epoch": 1.6430336599914783, + "grad_norm": 21.125, + "learning_rate": 1.9300838059123154e-07, + "loss": 0.7098, + "step": 19281 + }, + { + "epoch": 1.6431188751597783, + "grad_norm": 15.4375, + "learning_rate": 1.929186881159352e-07, + "loss": 0.7243, + "step": 19282 + }, + { + "epoch": 1.6432040903280782, + "grad_norm": 11.4375, + "learning_rate": 1.928290147431612e-07, + "loss": 0.3831, + "step": 19283 + }, + { + "epoch": 1.6432893054963782, + "grad_norm": 31.25, + "learning_rate": 1.927393604745302e-07, + "loss": 1.1392, + "step": 19284 + }, + { + "epoch": 1.6433745206646782, + "grad_norm": 16.375, + "learning_rate": 1.9264972531166202e-07, + "loss": 0.5971, + "step": 19285 + }, + { + "epoch": 1.643459735832978, + "grad_norm": 21.25, + "learning_rate": 1.9256010925617643e-07, + "loss": 0.7387, + "step": 19286 + }, + { + "epoch": 1.643544951001278, + "grad_norm": 18.375, + "learning_rate": 1.9247051230969288e-07, + "loss": 0.7771, + "step": 19287 + }, + { + "epoch": 1.643630166169578, + "grad_norm": 13.9375, + "learning_rate": 1.923809344738306e-07, + "loss": 0.4712, + "step": 19288 + }, + { + "epoch": 1.643715381337878, + "grad_norm": 16.125, + "learning_rate": 1.9229137575020874e-07, + "loss": 0.4438, + "step": 19289 + }, + { + "epoch": 1.643800596506178, + "grad_norm": 12.3125, + "learning_rate": 1.92201836140445e-07, + "loss": 0.4888, + "step": 19290 + }, + { + "epoch": 1.643885811674478, + "grad_norm": 13.8125, + "learning_rate": 1.9211231564615768e-07, + "loss": 0.4234, + "step": 19291 + }, + { + "epoch": 1.6439710268427778, + "grad_norm": 20.0, + "learning_rate": 1.920228142689648e-07, + "loss": 0.5276, + "step": 19292 + }, + { + "epoch": 1.6440562420110778, + "grad_norm": 14.875, + "learning_rate": 1.9193333201048395e-07, + "loss": 0.805, + "step": 19293 + }, + { + "epoch": 1.6441414571793778, + "grad_norm": 14.1875, + "learning_rate": 1.918438688723312e-07, + "loss": 0.5112, + "step": 19294 + }, + { + "epoch": 1.6442266723476777, + "grad_norm": 16.375, + "learning_rate": 1.9175442485612385e-07, + "loss": 0.8751, + "step": 19295 + }, + { + "epoch": 1.644311887515978, + "grad_norm": 12.1875, + "learning_rate": 1.9166499996347828e-07, + "loss": 0.302, + "step": 19296 + }, + { + "epoch": 1.6443971026842779, + "grad_norm": 12.8125, + "learning_rate": 1.9157559419601055e-07, + "loss": 0.486, + "step": 19297 + }, + { + "epoch": 1.6444823178525778, + "grad_norm": 18.375, + "learning_rate": 1.9148620755533588e-07, + "loss": 0.6681, + "step": 19298 + }, + { + "epoch": 1.6445675330208778, + "grad_norm": 16.75, + "learning_rate": 1.913968400430699e-07, + "loss": 0.5504, + "step": 19299 + }, + { + "epoch": 1.6446527481891777, + "grad_norm": 20.125, + "learning_rate": 1.9130749166082749e-07, + "loss": 0.9259, + "step": 19300 + }, + { + "epoch": 1.6447379633574777, + "grad_norm": 23.875, + "learning_rate": 1.912181624102237e-07, + "loss": 0.8031, + "step": 19301 + }, + { + "epoch": 1.6448231785257776, + "grad_norm": 14.0, + "learning_rate": 1.9112885229287227e-07, + "loss": 1.0276, + "step": 19302 + }, + { + "epoch": 1.6449083936940776, + "grad_norm": 14.5, + "learning_rate": 1.91039561310387e-07, + "loss": 0.5001, + "step": 19303 + }, + { + "epoch": 1.6449936088623776, + "grad_norm": 16.0, + "learning_rate": 1.909502894643818e-07, + "loss": 0.7078, + "step": 19304 + }, + { + "epoch": 1.6450788240306775, + "grad_norm": 14.8125, + "learning_rate": 1.9086103675646978e-07, + "loss": 0.5766, + "step": 19305 + }, + { + "epoch": 1.6451640391989775, + "grad_norm": 17.125, + "learning_rate": 1.907718031882641e-07, + "loss": 0.5363, + "step": 19306 + }, + { + "epoch": 1.6452492543672774, + "grad_norm": 10.625, + "learning_rate": 1.9068258876137682e-07, + "loss": 0.3351, + "step": 19307 + }, + { + "epoch": 1.6453344695355774, + "grad_norm": 12.0, + "learning_rate": 1.905933934774204e-07, + "loss": 0.2664, + "step": 19308 + }, + { + "epoch": 1.6454196847038773, + "grad_norm": 16.125, + "learning_rate": 1.90504217338007e-07, + "loss": 0.8447, + "step": 19309 + }, + { + "epoch": 1.6455048998721773, + "grad_norm": 14.625, + "learning_rate": 1.904150603447477e-07, + "loss": 0.8133, + "step": 19310 + }, + { + "epoch": 1.6455901150404773, + "grad_norm": 20.375, + "learning_rate": 1.903259224992536e-07, + "loss": 0.1961, + "step": 19311 + }, + { + "epoch": 1.6456753302087772, + "grad_norm": 11.5, + "learning_rate": 1.902368038031356e-07, + "loss": 0.3076, + "step": 19312 + }, + { + "epoch": 1.6457605453770772, + "grad_norm": 13.625, + "learning_rate": 1.9014770425800425e-07, + "loss": 0.6216, + "step": 19313 + }, + { + "epoch": 1.6458457605453771, + "grad_norm": 14.9375, + "learning_rate": 1.900586238654699e-07, + "loss": 0.654, + "step": 19314 + }, + { + "epoch": 1.645930975713677, + "grad_norm": 13.5, + "learning_rate": 1.8996956262714166e-07, + "loss": 0.4777, + "step": 19315 + }, + { + "epoch": 1.646016190881977, + "grad_norm": 13.125, + "learning_rate": 1.8988052054462954e-07, + "loss": 0.5721, + "step": 19316 + }, + { + "epoch": 1.646101406050277, + "grad_norm": 18.375, + "learning_rate": 1.897914976195428e-07, + "loss": 0.8751, + "step": 19317 + }, + { + "epoch": 1.646186621218577, + "grad_norm": 16.75, + "learning_rate": 1.897024938534893e-07, + "loss": 0.5409, + "step": 19318 + }, + { + "epoch": 1.646271836386877, + "grad_norm": 21.25, + "learning_rate": 1.8961350924807844e-07, + "loss": 0.6097, + "step": 19319 + }, + { + "epoch": 1.6463570515551769, + "grad_norm": 17.75, + "learning_rate": 1.8952454380491744e-07, + "loss": 0.6218, + "step": 19320 + }, + { + "epoch": 1.6464422667234768, + "grad_norm": 13.3125, + "learning_rate": 1.8943559752561417e-07, + "loss": 0.4256, + "step": 19321 + }, + { + "epoch": 1.6465274818917768, + "grad_norm": 18.625, + "learning_rate": 1.8934667041177633e-07, + "loss": 0.9468, + "step": 19322 + }, + { + "epoch": 1.6466126970600767, + "grad_norm": 13.6875, + "learning_rate": 1.892577624650109e-07, + "loss": 0.5616, + "step": 19323 + }, + { + "epoch": 1.6466979122283767, + "grad_norm": 13.8125, + "learning_rate": 1.8916887368692406e-07, + "loss": 0.2948, + "step": 19324 + }, + { + "epoch": 1.6467831273966766, + "grad_norm": 11.0, + "learning_rate": 1.8908000407912284e-07, + "loss": 0.3557, + "step": 19325 + }, + { + "epoch": 1.6468683425649766, + "grad_norm": 18.375, + "learning_rate": 1.8899115364321236e-07, + "loss": 0.7896, + "step": 19326 + }, + { + "epoch": 1.6469535577332766, + "grad_norm": 13.3125, + "learning_rate": 1.8890232238079897e-07, + "loss": 0.2526, + "step": 19327 + }, + { + "epoch": 1.6470387729015765, + "grad_norm": 19.875, + "learning_rate": 1.8881351029348737e-07, + "loss": 0.9509, + "step": 19328 + }, + { + "epoch": 1.6471239880698765, + "grad_norm": 13.375, + "learning_rate": 1.8872471738288282e-07, + "loss": 0.3303, + "step": 19329 + }, + { + "epoch": 1.6472092032381764, + "grad_norm": 16.125, + "learning_rate": 1.8863594365058968e-07, + "loss": 0.714, + "step": 19330 + }, + { + "epoch": 1.6472944184064764, + "grad_norm": 17.375, + "learning_rate": 1.8854718909821267e-07, + "loss": 0.5752, + "step": 19331 + }, + { + "epoch": 1.6473796335747763, + "grad_norm": 20.5, + "learning_rate": 1.8845845372735511e-07, + "loss": 0.6939, + "step": 19332 + }, + { + "epoch": 1.6474648487430763, + "grad_norm": 16.0, + "learning_rate": 1.8836973753962096e-07, + "loss": 0.6664, + "step": 19333 + }, + { + "epoch": 1.6475500639113763, + "grad_norm": 14.0625, + "learning_rate": 1.8828104053661298e-07, + "loss": 0.5217, + "step": 19334 + }, + { + "epoch": 1.6476352790796762, + "grad_norm": 18.5, + "learning_rate": 1.8819236271993418e-07, + "loss": 0.9461, + "step": 19335 + }, + { + "epoch": 1.6477204942479762, + "grad_norm": 18.875, + "learning_rate": 1.8810370409118734e-07, + "loss": 0.7008, + "step": 19336 + }, + { + "epoch": 1.6478057094162761, + "grad_norm": 11.5625, + "learning_rate": 1.880150646519742e-07, + "loss": 0.2939, + "step": 19337 + }, + { + "epoch": 1.647890924584576, + "grad_norm": 14.625, + "learning_rate": 1.8792644440389668e-07, + "loss": 0.5609, + "step": 19338 + }, + { + "epoch": 1.647976139752876, + "grad_norm": 16.375, + "learning_rate": 1.8783784334855642e-07, + "loss": 0.4804, + "step": 19339 + }, + { + "epoch": 1.648061354921176, + "grad_norm": 13.3125, + "learning_rate": 1.8774926148755464e-07, + "loss": 0.4724, + "step": 19340 + }, + { + "epoch": 1.648146570089476, + "grad_norm": 16.625, + "learning_rate": 1.8766069882249188e-07, + "loss": 0.4475, + "step": 19341 + }, + { + "epoch": 1.648231785257776, + "grad_norm": 21.5, + "learning_rate": 1.8757215535496823e-07, + "loss": 0.8342, + "step": 19342 + }, + { + "epoch": 1.6483170004260759, + "grad_norm": 22.25, + "learning_rate": 1.8748363108658408e-07, + "loss": 0.9996, + "step": 19343 + }, + { + "epoch": 1.6484022155943758, + "grad_norm": 17.5, + "learning_rate": 1.8739512601893952e-07, + "loss": 0.696, + "step": 19344 + }, + { + "epoch": 1.6484874307626758, + "grad_norm": 27.25, + "learning_rate": 1.8730664015363332e-07, + "loss": 0.7625, + "step": 19345 + }, + { + "epoch": 1.6485726459309757, + "grad_norm": 14.1875, + "learning_rate": 1.8721817349226457e-07, + "loss": 0.5164, + "step": 19346 + }, + { + "epoch": 1.6486578610992757, + "grad_norm": 12.75, + "learning_rate": 1.8712972603643228e-07, + "loss": 0.4156, + "step": 19347 + }, + { + "epoch": 1.6487430762675757, + "grad_norm": 22.25, + "learning_rate": 1.870412977877349e-07, + "loss": 1.2572, + "step": 19348 + }, + { + "epoch": 1.6488282914358756, + "grad_norm": 16.25, + "learning_rate": 1.869528887477702e-07, + "loss": 0.7531, + "step": 19349 + }, + { + "epoch": 1.6489135066041756, + "grad_norm": 13.1875, + "learning_rate": 1.8686449891813536e-07, + "loss": 0.5164, + "step": 19350 + }, + { + "epoch": 1.6489987217724755, + "grad_norm": 16.5, + "learning_rate": 1.8677612830042823e-07, + "loss": 0.3172, + "step": 19351 + }, + { + "epoch": 1.6490839369407755, + "grad_norm": 11.875, + "learning_rate": 1.866877768962455e-07, + "loss": 0.4066, + "step": 19352 + }, + { + "epoch": 1.6491691521090754, + "grad_norm": 18.5, + "learning_rate": 1.8659944470718422e-07, + "loss": 0.7412, + "step": 19353 + }, + { + "epoch": 1.6492543672773754, + "grad_norm": 15.0625, + "learning_rate": 1.8651113173484004e-07, + "loss": 0.6349, + "step": 19354 + }, + { + "epoch": 1.6493395824456754, + "grad_norm": 16.875, + "learning_rate": 1.8642283798080917e-07, + "loss": 0.6759, + "step": 19355 + }, + { + "epoch": 1.6494247976139753, + "grad_norm": 17.25, + "learning_rate": 1.8633456344668733e-07, + "loss": 0.718, + "step": 19356 + }, + { + "epoch": 1.6495100127822753, + "grad_norm": 12.25, + "learning_rate": 1.862463081340693e-07, + "loss": 0.2772, + "step": 19357 + }, + { + "epoch": 1.6495952279505752, + "grad_norm": 17.0, + "learning_rate": 1.8615807204455035e-07, + "loss": 0.8108, + "step": 19358 + }, + { + "epoch": 1.6496804431188752, + "grad_norm": 19.75, + "learning_rate": 1.8606985517972464e-07, + "loss": 0.7674, + "step": 19359 + }, + { + "epoch": 1.6497656582871751, + "grad_norm": 20.75, + "learning_rate": 1.8598165754118643e-07, + "loss": 0.8205, + "step": 19360 + }, + { + "epoch": 1.649850873455475, + "grad_norm": 21.0, + "learning_rate": 1.8589347913053003e-07, + "loss": 0.8357, + "step": 19361 + }, + { + "epoch": 1.649936088623775, + "grad_norm": 16.875, + "learning_rate": 1.8580531994934813e-07, + "loss": 0.5318, + "step": 19362 + }, + { + "epoch": 1.650021303792075, + "grad_norm": 24.75, + "learning_rate": 1.8571717999923424e-07, + "loss": 0.4585, + "step": 19363 + }, + { + "epoch": 1.650106518960375, + "grad_norm": 16.625, + "learning_rate": 1.856290592817811e-07, + "loss": 0.4659, + "step": 19364 + }, + { + "epoch": 1.650191734128675, + "grad_norm": 14.9375, + "learning_rate": 1.8554095779858145e-07, + "loss": 0.7044, + "step": 19365 + }, + { + "epoch": 1.6502769492969749, + "grad_norm": 23.5, + "learning_rate": 1.8545287555122706e-07, + "loss": 0.7706, + "step": 19366 + }, + { + "epoch": 1.6503621644652748, + "grad_norm": 14.5, + "learning_rate": 1.8536481254130948e-07, + "loss": 0.5212, + "step": 19367 + }, + { + "epoch": 1.6504473796335748, + "grad_norm": 17.75, + "learning_rate": 1.852767687704203e-07, + "loss": 0.7551, + "step": 19368 + }, + { + "epoch": 1.6505325948018748, + "grad_norm": 10.5, + "learning_rate": 1.8518874424015065e-07, + "loss": 0.1772, + "step": 19369 + }, + { + "epoch": 1.6506178099701747, + "grad_norm": 16.75, + "learning_rate": 1.8510073895209133e-07, + "loss": 0.5679, + "step": 19370 + }, + { + "epoch": 1.6507030251384747, + "grad_norm": 18.375, + "learning_rate": 1.8501275290783217e-07, + "loss": 0.4533, + "step": 19371 + }, + { + "epoch": 1.6507882403067746, + "grad_norm": 13.0625, + "learning_rate": 1.8492478610896355e-07, + "loss": 0.519, + "step": 19372 + }, + { + "epoch": 1.6508734554750746, + "grad_norm": 13.3125, + "learning_rate": 1.8483683855707536e-07, + "loss": 0.4941, + "step": 19373 + }, + { + "epoch": 1.6509586706433745, + "grad_norm": 14.9375, + "learning_rate": 1.8474891025375654e-07, + "loss": 0.6649, + "step": 19374 + }, + { + "epoch": 1.6510438858116745, + "grad_norm": 14.8125, + "learning_rate": 1.8466100120059587e-07, + "loss": 0.6033, + "step": 19375 + }, + { + "epoch": 1.6511291009799745, + "grad_norm": 20.75, + "learning_rate": 1.8457311139918222e-07, + "loss": 0.7452, + "step": 19376 + }, + { + "epoch": 1.6512143161482744, + "grad_norm": 14.75, + "learning_rate": 1.8448524085110373e-07, + "loss": 0.8775, + "step": 19377 + }, + { + "epoch": 1.6512995313165744, + "grad_norm": 16.5, + "learning_rate": 1.8439738955794872e-07, + "loss": 0.5007, + "step": 19378 + }, + { + "epoch": 1.6513847464848743, + "grad_norm": 11.25, + "learning_rate": 1.8430955752130415e-07, + "loss": 0.3371, + "step": 19379 + }, + { + "epoch": 1.6514699616531743, + "grad_norm": 12.625, + "learning_rate": 1.842217447427576e-07, + "loss": 0.6475, + "step": 19380 + }, + { + "epoch": 1.6515551768214742, + "grad_norm": 18.25, + "learning_rate": 1.8413395122389572e-07, + "loss": 0.6553, + "step": 19381 + }, + { + "epoch": 1.6516403919897742, + "grad_norm": 11.0, + "learning_rate": 1.8404617696630544e-07, + "loss": 0.2985, + "step": 19382 + }, + { + "epoch": 1.6517256071580741, + "grad_norm": 20.625, + "learning_rate": 1.8395842197157273e-07, + "loss": 0.7027, + "step": 19383 + }, + { + "epoch": 1.651810822326374, + "grad_norm": 18.5, + "learning_rate": 1.83870686241283e-07, + "loss": 0.5273, + "step": 19384 + }, + { + "epoch": 1.651896037494674, + "grad_norm": 12.4375, + "learning_rate": 1.8378296977702201e-07, + "loss": 0.2964, + "step": 19385 + }, + { + "epoch": 1.651981252662974, + "grad_norm": 19.0, + "learning_rate": 1.836952725803752e-07, + "loss": 0.7612, + "step": 19386 + }, + { + "epoch": 1.652066467831274, + "grad_norm": 13.6875, + "learning_rate": 1.8360759465292685e-07, + "loss": 0.3732, + "step": 19387 + }, + { + "epoch": 1.652151682999574, + "grad_norm": 20.375, + "learning_rate": 1.8351993599626154e-07, + "loss": 0.5553, + "step": 19388 + }, + { + "epoch": 1.652236898167874, + "grad_norm": 18.375, + "learning_rate": 1.8343229661196337e-07, + "loss": 0.6769, + "step": 19389 + }, + { + "epoch": 1.6523221133361738, + "grad_norm": 19.0, + "learning_rate": 1.8334467650161656e-07, + "loss": 0.8116, + "step": 19390 + }, + { + "epoch": 1.6524073285044738, + "grad_norm": 13.0625, + "learning_rate": 1.8325707566680394e-07, + "loss": 0.6021, + "step": 19391 + }, + { + "epoch": 1.6524925436727738, + "grad_norm": 21.25, + "learning_rate": 1.8316949410910833e-07, + "loss": 0.8711, + "step": 19392 + }, + { + "epoch": 1.6525777588410737, + "grad_norm": 18.875, + "learning_rate": 1.830819318301129e-07, + "loss": 0.7389, + "step": 19393 + }, + { + "epoch": 1.6526629740093737, + "grad_norm": 19.875, + "learning_rate": 1.8299438883139967e-07, + "loss": 0.8028, + "step": 19394 + }, + { + "epoch": 1.6527481891776736, + "grad_norm": 16.75, + "learning_rate": 1.8290686511455117e-07, + "loss": 0.4817, + "step": 19395 + }, + { + "epoch": 1.6528334043459736, + "grad_norm": 16.0, + "learning_rate": 1.8281936068114832e-07, + "loss": 0.5762, + "step": 19396 + }, + { + "epoch": 1.6529186195142735, + "grad_norm": 20.375, + "learning_rate": 1.8273187553277284e-07, + "loss": 0.7924, + "step": 19397 + }, + { + "epoch": 1.6530038346825735, + "grad_norm": 30.0, + "learning_rate": 1.8264440967100574e-07, + "loss": 0.8572, + "step": 19398 + }, + { + "epoch": 1.6530890498508735, + "grad_norm": 12.0625, + "learning_rate": 1.8255696309742722e-07, + "loss": 0.4286, + "step": 19399 + }, + { + "epoch": 1.6531742650191734, + "grad_norm": 15.3125, + "learning_rate": 1.8246953581361819e-07, + "loss": 0.5601, + "step": 19400 + }, + { + "epoch": 1.6532594801874734, + "grad_norm": 14.9375, + "learning_rate": 1.8238212782115766e-07, + "loss": 0.6186, + "step": 19401 + }, + { + "epoch": 1.6533446953557733, + "grad_norm": 15.6875, + "learning_rate": 1.822947391216258e-07, + "loss": 0.635, + "step": 19402 + }, + { + "epoch": 1.6534299105240733, + "grad_norm": 17.625, + "learning_rate": 1.8220736971660186e-07, + "loss": 0.7179, + "step": 19403 + }, + { + "epoch": 1.6535151256923732, + "grad_norm": 20.125, + "learning_rate": 1.8212001960766433e-07, + "loss": 0.7307, + "step": 19404 + }, + { + "epoch": 1.6536003408606732, + "grad_norm": 19.25, + "learning_rate": 1.8203268879639176e-07, + "loss": 1.0239, + "step": 19405 + }, + { + "epoch": 1.6536855560289732, + "grad_norm": 16.0, + "learning_rate": 1.8194537728436267e-07, + "loss": 0.6888, + "step": 19406 + }, + { + "epoch": 1.6537707711972731, + "grad_norm": 14.6875, + "learning_rate": 1.8185808507315453e-07, + "loss": 0.65, + "step": 19407 + }, + { + "epoch": 1.653855986365573, + "grad_norm": 16.875, + "learning_rate": 1.817708121643451e-07, + "loss": 0.5776, + "step": 19408 + }, + { + "epoch": 1.653941201533873, + "grad_norm": 20.5, + "learning_rate": 1.8168355855951085e-07, + "loss": 0.8526, + "step": 19409 + }, + { + "epoch": 1.654026416702173, + "grad_norm": 22.75, + "learning_rate": 1.8159632426022908e-07, + "loss": 0.6855, + "step": 19410 + }, + { + "epoch": 1.654111631870473, + "grad_norm": 13.25, + "learning_rate": 1.815091092680761e-07, + "loss": 0.413, + "step": 19411 + }, + { + "epoch": 1.654196847038773, + "grad_norm": 13.6875, + "learning_rate": 1.8142191358462833e-07, + "loss": 0.5441, + "step": 19412 + }, + { + "epoch": 1.6542820622070729, + "grad_norm": 14.0, + "learning_rate": 1.8133473721146072e-07, + "loss": 0.6318, + "step": 19413 + }, + { + "epoch": 1.6543672773753728, + "grad_norm": 12.5625, + "learning_rate": 1.812475801501494e-07, + "loss": 0.2501, + "step": 19414 + }, + { + "epoch": 1.6544524925436728, + "grad_norm": 15.375, + "learning_rate": 1.8116044240226866e-07, + "loss": 0.7859, + "step": 19415 + }, + { + "epoch": 1.6545377077119727, + "grad_norm": 17.0, + "learning_rate": 1.810733239693939e-07, + "loss": 0.8148, + "step": 19416 + }, + { + "epoch": 1.6546229228802727, + "grad_norm": 16.875, + "learning_rate": 1.8098622485309873e-07, + "loss": 0.5387, + "step": 19417 + }, + { + "epoch": 1.6547081380485726, + "grad_norm": 12.75, + "learning_rate": 1.8089914505495744e-07, + "loss": 0.3067, + "step": 19418 + }, + { + "epoch": 1.6547933532168726, + "grad_norm": 13.9375, + "learning_rate": 1.8081208457654362e-07, + "loss": 0.4237, + "step": 19419 + }, + { + "epoch": 1.6548785683851726, + "grad_norm": 17.5, + "learning_rate": 1.8072504341943089e-07, + "loss": 0.6383, + "step": 19420 + }, + { + "epoch": 1.6549637835534725, + "grad_norm": 25.625, + "learning_rate": 1.8063802158519161e-07, + "loss": 0.6936, + "step": 19421 + }, + { + "epoch": 1.6550489987217725, + "grad_norm": 12.0625, + "learning_rate": 1.805510190753987e-07, + "loss": 0.3968, + "step": 19422 + }, + { + "epoch": 1.6551342138900724, + "grad_norm": 13.5, + "learning_rate": 1.8046403589162446e-07, + "loss": 0.5926, + "step": 19423 + }, + { + "epoch": 1.6552194290583724, + "grad_norm": 12.5, + "learning_rate": 1.803770720354403e-07, + "loss": 0.397, + "step": 19424 + }, + { + "epoch": 1.6553046442266723, + "grad_norm": 18.125, + "learning_rate": 1.802901275084183e-07, + "loss": 0.7631, + "step": 19425 + }, + { + "epoch": 1.6553898593949723, + "grad_norm": 14.1875, + "learning_rate": 1.8020320231212919e-07, + "loss": 0.6995, + "step": 19426 + }, + { + "epoch": 1.6554750745632723, + "grad_norm": 13.6875, + "learning_rate": 1.80116296448144e-07, + "loss": 0.6697, + "step": 19427 + }, + { + "epoch": 1.6555602897315722, + "grad_norm": 12.3125, + "learning_rate": 1.8002940991803303e-07, + "loss": 0.3863, + "step": 19428 + }, + { + "epoch": 1.6556455048998722, + "grad_norm": 12.5, + "learning_rate": 1.7994254272336698e-07, + "loss": 0.398, + "step": 19429 + }, + { + "epoch": 1.6557307200681721, + "grad_norm": 15.6875, + "learning_rate": 1.7985569486571486e-07, + "loss": 0.6979, + "step": 19430 + }, + { + "epoch": 1.655815935236472, + "grad_norm": 10.75, + "learning_rate": 1.7976886634664682e-07, + "loss": 0.3375, + "step": 19431 + }, + { + "epoch": 1.655901150404772, + "grad_norm": 17.125, + "learning_rate": 1.7968205716773116e-07, + "loss": 0.5716, + "step": 19432 + }, + { + "epoch": 1.655986365573072, + "grad_norm": 15.0, + "learning_rate": 1.7959526733053733e-07, + "loss": 0.5077, + "step": 19433 + }, + { + "epoch": 1.656071580741372, + "grad_norm": 25.375, + "learning_rate": 1.7950849683663312e-07, + "loss": 0.6731, + "step": 19434 + }, + { + "epoch": 1.656156795909672, + "grad_norm": 14.8125, + "learning_rate": 1.7942174568758686e-07, + "loss": 0.4194, + "step": 19435 + }, + { + "epoch": 1.6562420110779719, + "grad_norm": 24.625, + "learning_rate": 1.7933501388496616e-07, + "loss": 1.1108, + "step": 19436 + }, + { + "epoch": 1.6563272262462718, + "grad_norm": 15.6875, + "learning_rate": 1.7924830143033866e-07, + "loss": 0.7367, + "step": 19437 + }, + { + "epoch": 1.6564124414145718, + "grad_norm": 13.9375, + "learning_rate": 1.791616083252709e-07, + "loss": 0.6487, + "step": 19438 + }, + { + "epoch": 1.6564976565828717, + "grad_norm": 15.625, + "learning_rate": 1.7907493457132995e-07, + "loss": 0.696, + "step": 19439 + }, + { + "epoch": 1.6565828717511717, + "grad_norm": 19.5, + "learning_rate": 1.7898828017008151e-07, + "loss": 0.7518, + "step": 19440 + }, + { + "epoch": 1.6566680869194716, + "grad_norm": 14.0625, + "learning_rate": 1.7890164512309184e-07, + "loss": 0.3174, + "step": 19441 + }, + { + "epoch": 1.6567533020877716, + "grad_norm": 15.0625, + "learning_rate": 1.7881502943192672e-07, + "loss": 0.5122, + "step": 19442 + }, + { + "epoch": 1.6568385172560716, + "grad_norm": 13.25, + "learning_rate": 1.7872843309815092e-07, + "loss": 0.3762, + "step": 19443 + }, + { + "epoch": 1.6569237324243715, + "grad_norm": 18.5, + "learning_rate": 1.786418561233297e-07, + "loss": 0.6564, + "step": 19444 + }, + { + "epoch": 1.6570089475926715, + "grad_norm": 14.875, + "learning_rate": 1.7855529850902733e-07, + "loss": 0.382, + "step": 19445 + }, + { + "epoch": 1.6570941627609714, + "grad_norm": 11.5, + "learning_rate": 1.784687602568086e-07, + "loss": 0.3701, + "step": 19446 + }, + { + "epoch": 1.6571793779292714, + "grad_norm": 14.5, + "learning_rate": 1.7838224136823666e-07, + "loss": 0.4403, + "step": 19447 + }, + { + "epoch": 1.6572645930975713, + "grad_norm": 15.8125, + "learning_rate": 1.7829574184487511e-07, + "loss": 0.6807, + "step": 19448 + }, + { + "epoch": 1.6573498082658713, + "grad_norm": 12.5, + "learning_rate": 1.7820926168828703e-07, + "loss": 0.4887, + "step": 19449 + }, + { + "epoch": 1.6574350234341713, + "grad_norm": 22.625, + "learning_rate": 1.7812280090003574e-07, + "loss": 0.7665, + "step": 19450 + }, + { + "epoch": 1.6575202386024712, + "grad_norm": 12.0, + "learning_rate": 1.7803635948168307e-07, + "loss": 0.4414, + "step": 19451 + }, + { + "epoch": 1.6576054537707712, + "grad_norm": 19.875, + "learning_rate": 1.779499374347912e-07, + "loss": 1.1401, + "step": 19452 + }, + { + "epoch": 1.6576906689390711, + "grad_norm": 16.125, + "learning_rate": 1.7786353476092214e-07, + "loss": 0.6111, + "step": 19453 + }, + { + "epoch": 1.657775884107371, + "grad_norm": 13.0, + "learning_rate": 1.777771514616372e-07, + "loss": 0.6027, + "step": 19454 + }, + { + "epoch": 1.657861099275671, + "grad_norm": 16.75, + "learning_rate": 1.7769078753849756e-07, + "loss": 0.6748, + "step": 19455 + }, + { + "epoch": 1.657946314443971, + "grad_norm": 12.3125, + "learning_rate": 1.7760444299306317e-07, + "loss": 0.3051, + "step": 19456 + }, + { + "epoch": 1.658031529612271, + "grad_norm": 16.5, + "learning_rate": 1.7751811782689505e-07, + "loss": 1.0536, + "step": 19457 + }, + { + "epoch": 1.658116744780571, + "grad_norm": 17.25, + "learning_rate": 1.7743181204155287e-07, + "loss": 0.5162, + "step": 19458 + }, + { + "epoch": 1.6582019599488709, + "grad_norm": 29.625, + "learning_rate": 1.7734552563859683e-07, + "loss": 0.791, + "step": 19459 + }, + { + "epoch": 1.6582871751171708, + "grad_norm": 15.375, + "learning_rate": 1.7725925861958538e-07, + "loss": 0.6744, + "step": 19460 + }, + { + "epoch": 1.6583723902854708, + "grad_norm": 24.875, + "learning_rate": 1.7717301098607794e-07, + "loss": 0.2824, + "step": 19461 + }, + { + "epoch": 1.6584576054537707, + "grad_norm": 30.375, + "learning_rate": 1.770867827396329e-07, + "loss": 0.9232, + "step": 19462 + }, + { + "epoch": 1.6585428206220707, + "grad_norm": 10.875, + "learning_rate": 1.7700057388180926e-07, + "loss": 0.2941, + "step": 19463 + }, + { + "epoch": 1.6586280357903707, + "grad_norm": 13.625, + "learning_rate": 1.7691438441416368e-07, + "loss": 0.3183, + "step": 19464 + }, + { + "epoch": 1.6587132509586706, + "grad_norm": 16.375, + "learning_rate": 1.768282143382541e-07, + "loss": 0.622, + "step": 19465 + }, + { + "epoch": 1.6587984661269706, + "grad_norm": 11.4375, + "learning_rate": 1.7674206365563802e-07, + "loss": 0.3902, + "step": 19466 + }, + { + "epoch": 1.6588836812952705, + "grad_norm": 13.625, + "learning_rate": 1.7665593236787224e-07, + "loss": 0.483, + "step": 19467 + }, + { + "epoch": 1.6589688964635705, + "grad_norm": 15.375, + "learning_rate": 1.765698204765129e-07, + "loss": 0.7258, + "step": 19468 + }, + { + "epoch": 1.6590541116318704, + "grad_norm": 16.75, + "learning_rate": 1.7648372798311622e-07, + "loss": 0.4711, + "step": 19469 + }, + { + "epoch": 1.6591393268001704, + "grad_norm": 20.125, + "learning_rate": 1.7639765488923806e-07, + "loss": 0.931, + "step": 19470 + }, + { + "epoch": 1.6592245419684704, + "grad_norm": 16.25, + "learning_rate": 1.7631160119643428e-07, + "loss": 0.4046, + "step": 19471 + }, + { + "epoch": 1.6593097571367703, + "grad_norm": 18.625, + "learning_rate": 1.7622556690625944e-07, + "loss": 0.8349, + "step": 19472 + }, + { + "epoch": 1.6593949723050703, + "grad_norm": 21.5, + "learning_rate": 1.7613955202026815e-07, + "loss": 0.5973, + "step": 19473 + }, + { + "epoch": 1.6594801874733702, + "grad_norm": 11.0, + "learning_rate": 1.760535565400151e-07, + "loss": 0.295, + "step": 19474 + }, + { + "epoch": 1.6595654026416702, + "grad_norm": 15.375, + "learning_rate": 1.759675804670541e-07, + "loss": 0.7016, + "step": 19475 + }, + { + "epoch": 1.6596506178099701, + "grad_norm": 16.375, + "learning_rate": 1.7588162380293913e-07, + "loss": 0.5547, + "step": 19476 + }, + { + "epoch": 1.65973583297827, + "grad_norm": 17.125, + "learning_rate": 1.7579568654922317e-07, + "loss": 0.8985, + "step": 19477 + }, + { + "epoch": 1.65982104814657, + "grad_norm": 20.875, + "learning_rate": 1.7570976870745937e-07, + "loss": 0.7554, + "step": 19478 + }, + { + "epoch": 1.65990626331487, + "grad_norm": 12.375, + "learning_rate": 1.7562387027920055e-07, + "loss": 0.5377, + "step": 19479 + }, + { + "epoch": 1.65999147848317, + "grad_norm": 24.625, + "learning_rate": 1.755379912659988e-07, + "loss": 0.5473, + "step": 19480 + }, + { + "epoch": 1.66007669365147, + "grad_norm": 14.5, + "learning_rate": 1.7545213166940577e-07, + "loss": 0.3896, + "step": 19481 + }, + { + "epoch": 1.6601619088197699, + "grad_norm": 18.125, + "learning_rate": 1.7536629149097318e-07, + "loss": 0.8575, + "step": 19482 + }, + { + "epoch": 1.6602471239880698, + "grad_norm": 13.0625, + "learning_rate": 1.752804707322524e-07, + "loss": 0.5296, + "step": 19483 + }, + { + "epoch": 1.6603323391563698, + "grad_norm": 12.6875, + "learning_rate": 1.7519466939479437e-07, + "loss": 0.4234, + "step": 19484 + }, + { + "epoch": 1.6604175543246698, + "grad_norm": 11.5625, + "learning_rate": 1.751088874801493e-07, + "loss": 0.3034, + "step": 19485 + }, + { + "epoch": 1.6605027694929697, + "grad_norm": 12.5625, + "learning_rate": 1.7502312498986747e-07, + "loss": 0.4663, + "step": 19486 + }, + { + "epoch": 1.6605879846612697, + "grad_norm": 13.3125, + "learning_rate": 1.7493738192549888e-07, + "loss": 0.5164, + "step": 19487 + }, + { + "epoch": 1.6606731998295696, + "grad_norm": 12.125, + "learning_rate": 1.7485165828859298e-07, + "loss": 0.3834, + "step": 19488 + }, + { + "epoch": 1.6607584149978696, + "grad_norm": 15.1875, + "learning_rate": 1.7476595408069882e-07, + "loss": 0.6588, + "step": 19489 + }, + { + "epoch": 1.6608436301661695, + "grad_norm": 13.5625, + "learning_rate": 1.7468026930336474e-07, + "loss": 0.6609, + "step": 19490 + }, + { + "epoch": 1.6609288453344695, + "grad_norm": 18.5, + "learning_rate": 1.7459460395813963e-07, + "loss": 0.7481, + "step": 19491 + }, + { + "epoch": 1.6610140605027695, + "grad_norm": 14.25, + "learning_rate": 1.7450895804657143e-07, + "loss": 0.5069, + "step": 19492 + }, + { + "epoch": 1.6610992756710694, + "grad_norm": 17.5, + "learning_rate": 1.7442333157020806e-07, + "loss": 0.4555, + "step": 19493 + }, + { + "epoch": 1.6611844908393694, + "grad_norm": 16.25, + "learning_rate": 1.7433772453059633e-07, + "loss": 0.4616, + "step": 19494 + }, + { + "epoch": 1.6612697060076693, + "grad_norm": 15.5625, + "learning_rate": 1.7425213692928377e-07, + "loss": 0.5307, + "step": 19495 + }, + { + "epoch": 1.6613549211759693, + "grad_norm": 18.375, + "learning_rate": 1.7416656876781694e-07, + "loss": 0.5097, + "step": 19496 + }, + { + "epoch": 1.6614401363442692, + "grad_norm": 17.5, + "learning_rate": 1.740810200477422e-07, + "loss": 0.528, + "step": 19497 + }, + { + "epoch": 1.6615253515125692, + "grad_norm": 22.0, + "learning_rate": 1.7399549077060484e-07, + "loss": 0.9126, + "step": 19498 + }, + { + "epoch": 1.6616105666808691, + "grad_norm": 13.875, + "learning_rate": 1.7390998093795117e-07, + "loss": 0.4675, + "step": 19499 + }, + { + "epoch": 1.661695781849169, + "grad_norm": 30.5, + "learning_rate": 1.7382449055132616e-07, + "loss": 0.8561, + "step": 19500 + }, + { + "epoch": 1.661780997017469, + "grad_norm": 14.4375, + "learning_rate": 1.7373901961227513e-07, + "loss": 0.4115, + "step": 19501 + }, + { + "epoch": 1.661866212185769, + "grad_norm": 14.9375, + "learning_rate": 1.7365356812234197e-07, + "loss": 0.7084, + "step": 19502 + }, + { + "epoch": 1.661951427354069, + "grad_norm": 14.3125, + "learning_rate": 1.7356813608307114e-07, + "loss": 0.5722, + "step": 19503 + }, + { + "epoch": 1.662036642522369, + "grad_norm": 14.4375, + "learning_rate": 1.7348272349600682e-07, + "loss": 0.5985, + "step": 19504 + }, + { + "epoch": 1.662121857690669, + "grad_norm": 17.125, + "learning_rate": 1.7339733036269194e-07, + "loss": 0.5964, + "step": 19505 + }, + { + "epoch": 1.6622070728589688, + "grad_norm": 11.375, + "learning_rate": 1.7331195668467013e-07, + "loss": 0.8222, + "step": 19506 + }, + { + "epoch": 1.6622922880272688, + "grad_norm": 13.625, + "learning_rate": 1.7322660246348378e-07, + "loss": 0.3949, + "step": 19507 + }, + { + "epoch": 1.6623775031955688, + "grad_norm": 14.125, + "learning_rate": 1.7314126770067552e-07, + "loss": 0.4898, + "step": 19508 + }, + { + "epoch": 1.6624627183638687, + "grad_norm": 26.375, + "learning_rate": 1.730559523977876e-07, + "loss": 1.0845, + "step": 19509 + }, + { + "epoch": 1.6625479335321687, + "grad_norm": 14.25, + "learning_rate": 1.729706565563613e-07, + "loss": 0.6164, + "step": 19510 + }, + { + "epoch": 1.6626331487004686, + "grad_norm": 15.625, + "learning_rate": 1.728853801779383e-07, + "loss": 0.4716, + "step": 19511 + }, + { + "epoch": 1.6627183638687686, + "grad_norm": 20.625, + "learning_rate": 1.7280012326405986e-07, + "loss": 0.646, + "step": 19512 + }, + { + "epoch": 1.6628035790370685, + "grad_norm": 14.3125, + "learning_rate": 1.7271488581626625e-07, + "loss": 0.4666, + "step": 19513 + }, + { + "epoch": 1.6628887942053685, + "grad_norm": 14.875, + "learning_rate": 1.7262966783609808e-07, + "loss": 0.4652, + "step": 19514 + }, + { + "epoch": 1.6629740093736685, + "grad_norm": 20.375, + "learning_rate": 1.7254446932509494e-07, + "loss": 0.7882, + "step": 19515 + }, + { + "epoch": 1.6630592245419684, + "grad_norm": 17.375, + "learning_rate": 1.724592902847967e-07, + "loss": 0.5199, + "step": 19516 + }, + { + "epoch": 1.6631444397102684, + "grad_norm": 22.5, + "learning_rate": 1.723741307167427e-07, + "loss": 0.6446, + "step": 19517 + }, + { + "epoch": 1.6632296548785683, + "grad_norm": 25.875, + "learning_rate": 1.72288990622472e-07, + "loss": 1.0557, + "step": 19518 + }, + { + "epoch": 1.6633148700468683, + "grad_norm": 15.625, + "learning_rate": 1.7220387000352265e-07, + "loss": 0.5639, + "step": 19519 + }, + { + "epoch": 1.6634000852151682, + "grad_norm": 30.5, + "learning_rate": 1.7211876886143355e-07, + "loss": 0.9343, + "step": 19520 + }, + { + "epoch": 1.6634853003834684, + "grad_norm": 14.0625, + "learning_rate": 1.7203368719774182e-07, + "loss": 0.4428, + "step": 19521 + }, + { + "epoch": 1.6635705155517684, + "grad_norm": 14.75, + "learning_rate": 1.719486250139854e-07, + "loss": 0.6097, + "step": 19522 + }, + { + "epoch": 1.6636557307200683, + "grad_norm": 12.375, + "learning_rate": 1.7186358231170167e-07, + "loss": 0.5344, + "step": 19523 + }, + { + "epoch": 1.6637409458883683, + "grad_norm": 20.625, + "learning_rate": 1.717785590924269e-07, + "loss": 0.6887, + "step": 19524 + }, + { + "epoch": 1.6638261610566683, + "grad_norm": 12.75, + "learning_rate": 1.7169355535769775e-07, + "loss": 0.4937, + "step": 19525 + }, + { + "epoch": 1.6639113762249682, + "grad_norm": 16.0, + "learning_rate": 1.716085711090508e-07, + "loss": 0.4135, + "step": 19526 + }, + { + "epoch": 1.6639965913932682, + "grad_norm": 17.75, + "learning_rate": 1.7152360634802092e-07, + "loss": 1.01, + "step": 19527 + }, + { + "epoch": 1.6640818065615681, + "grad_norm": 15.0625, + "learning_rate": 1.7143866107614426e-07, + "loss": 0.6504, + "step": 19528 + }, + { + "epoch": 1.664167021729868, + "grad_norm": 11.8125, + "learning_rate": 1.713537352949554e-07, + "loss": 0.4485, + "step": 19529 + }, + { + "epoch": 1.664252236898168, + "grad_norm": 16.125, + "learning_rate": 1.7126882900598912e-07, + "loss": 0.688, + "step": 19530 + }, + { + "epoch": 1.664337452066468, + "grad_norm": 23.0, + "learning_rate": 1.7118394221078013e-07, + "loss": 0.9478, + "step": 19531 + }, + { + "epoch": 1.664422667234768, + "grad_norm": 14.8125, + "learning_rate": 1.7109907491086195e-07, + "loss": 0.6269, + "step": 19532 + }, + { + "epoch": 1.664507882403068, + "grad_norm": 16.875, + "learning_rate": 1.7101422710776832e-07, + "loss": 0.499, + "step": 19533 + }, + { + "epoch": 1.6645930975713679, + "grad_norm": 13.125, + "learning_rate": 1.7092939880303263e-07, + "loss": 0.6598, + "step": 19534 + }, + { + "epoch": 1.6646783127396678, + "grad_norm": 29.5, + "learning_rate": 1.7084458999818795e-07, + "loss": 0.5009, + "step": 19535 + }, + { + "epoch": 1.6647635279079678, + "grad_norm": 16.25, + "learning_rate": 1.7075980069476693e-07, + "loss": 0.5432, + "step": 19536 + }, + { + "epoch": 1.6648487430762677, + "grad_norm": 17.875, + "learning_rate": 1.7067503089430116e-07, + "loss": 0.6029, + "step": 19537 + }, + { + "epoch": 1.6649339582445677, + "grad_norm": 22.625, + "learning_rate": 1.7059028059832298e-07, + "loss": 0.4113, + "step": 19538 + }, + { + "epoch": 1.6650191734128676, + "grad_norm": 16.25, + "learning_rate": 1.705055498083641e-07, + "loss": 0.8155, + "step": 19539 + }, + { + "epoch": 1.6651043885811676, + "grad_norm": 9.25, + "learning_rate": 1.7042083852595509e-07, + "loss": 0.1954, + "step": 19540 + }, + { + "epoch": 1.6651896037494676, + "grad_norm": 21.375, + "learning_rate": 1.7033614675262726e-07, + "loss": 0.8048, + "step": 19541 + }, + { + "epoch": 1.6652748189177675, + "grad_norm": 11.9375, + "learning_rate": 1.7025147448991087e-07, + "loss": 0.294, + "step": 19542 + }, + { + "epoch": 1.6653600340860675, + "grad_norm": 14.0625, + "learning_rate": 1.7016682173933652e-07, + "loss": 0.5275, + "step": 19543 + }, + { + "epoch": 1.6654452492543674, + "grad_norm": 18.75, + "learning_rate": 1.7008218850243342e-07, + "loss": 1.06, + "step": 19544 + }, + { + "epoch": 1.6655304644226674, + "grad_norm": 13.875, + "learning_rate": 1.6999757478073086e-07, + "loss": 0.5648, + "step": 19545 + }, + { + "epoch": 1.6656156795909673, + "grad_norm": 24.5, + "learning_rate": 1.6991298057575807e-07, + "loss": 0.497, + "step": 19546 + }, + { + "epoch": 1.6657008947592673, + "grad_norm": 15.9375, + "learning_rate": 1.6982840588904397e-07, + "loss": 0.4411, + "step": 19547 + }, + { + "epoch": 1.6657861099275673, + "grad_norm": 16.375, + "learning_rate": 1.6974385072211705e-07, + "loss": 0.7338, + "step": 19548 + }, + { + "epoch": 1.6658713250958672, + "grad_norm": 12.875, + "learning_rate": 1.6965931507650468e-07, + "loss": 0.7609, + "step": 19549 + }, + { + "epoch": 1.6659565402641672, + "grad_norm": 15.25, + "learning_rate": 1.6957479895373482e-07, + "loss": 0.7155, + "step": 19550 + }, + { + "epoch": 1.6660417554324671, + "grad_norm": 11.6875, + "learning_rate": 1.6949030235533488e-07, + "loss": 0.2808, + "step": 19551 + }, + { + "epoch": 1.666126970600767, + "grad_norm": 15.5625, + "learning_rate": 1.6940582528283206e-07, + "loss": 0.669, + "step": 19552 + }, + { + "epoch": 1.666212185769067, + "grad_norm": 13.375, + "learning_rate": 1.693213677377524e-07, + "loss": 0.6048, + "step": 19553 + }, + { + "epoch": 1.666297400937367, + "grad_norm": 14.5625, + "learning_rate": 1.6923692972162218e-07, + "loss": 0.5538, + "step": 19554 + }, + { + "epoch": 1.666382616105667, + "grad_norm": 20.375, + "learning_rate": 1.6915251123596737e-07, + "loss": 0.4744, + "step": 19555 + }, + { + "epoch": 1.666467831273967, + "grad_norm": 19.0, + "learning_rate": 1.6906811228231388e-07, + "loss": 0.815, + "step": 19556 + }, + { + "epoch": 1.6665530464422669, + "grad_norm": 13.625, + "learning_rate": 1.6898373286218627e-07, + "loss": 0.3131, + "step": 19557 + }, + { + "epoch": 1.6666382616105668, + "grad_norm": 12.625, + "learning_rate": 1.6889937297710977e-07, + "loss": 0.3792, + "step": 19558 + }, + { + "epoch": 1.6667234767788668, + "grad_norm": 15.5625, + "learning_rate": 1.688150326286085e-07, + "loss": 0.5057, + "step": 19559 + }, + { + "epoch": 1.6668086919471667, + "grad_norm": 13.5, + "learning_rate": 1.687307118182073e-07, + "loss": 0.4515, + "step": 19560 + }, + { + "epoch": 1.6668939071154667, + "grad_norm": 15.375, + "learning_rate": 1.6864641054742934e-07, + "loss": 0.4161, + "step": 19561 + }, + { + "epoch": 1.6669791222837667, + "grad_norm": 10.9375, + "learning_rate": 1.6856212881779786e-07, + "loss": 0.307, + "step": 19562 + }, + { + "epoch": 1.6670643374520666, + "grad_norm": 15.875, + "learning_rate": 1.6847786663083607e-07, + "loss": 0.7221, + "step": 19563 + }, + { + "epoch": 1.6671495526203666, + "grad_norm": 17.625, + "learning_rate": 1.6839362398806695e-07, + "loss": 0.7561, + "step": 19564 + }, + { + "epoch": 1.6672347677886665, + "grad_norm": 41.25, + "learning_rate": 1.6830940089101285e-07, + "loss": 1.2363, + "step": 19565 + }, + { + "epoch": 1.6673199829569665, + "grad_norm": 23.125, + "learning_rate": 1.682251973411954e-07, + "loss": 0.7129, + "step": 19566 + }, + { + "epoch": 1.6674051981252664, + "grad_norm": 15.125, + "learning_rate": 1.681410133401365e-07, + "loss": 0.6204, + "step": 19567 + }, + { + "epoch": 1.6674904132935664, + "grad_norm": 18.25, + "learning_rate": 1.6805684888935718e-07, + "loss": 0.9542, + "step": 19568 + }, + { + "epoch": 1.6675756284618664, + "grad_norm": 12.6875, + "learning_rate": 1.6797270399037932e-07, + "loss": 0.5049, + "step": 19569 + }, + { + "epoch": 1.6676608436301663, + "grad_norm": 14.9375, + "learning_rate": 1.6788857864472207e-07, + "loss": 0.6156, + "step": 19570 + }, + { + "epoch": 1.6677460587984663, + "grad_norm": 15.3125, + "learning_rate": 1.6780447285390646e-07, + "loss": 0.4022, + "step": 19571 + }, + { + "epoch": 1.6678312739667662, + "grad_norm": 16.75, + "learning_rate": 1.6772038661945226e-07, + "loss": 0.6107, + "step": 19572 + }, + { + "epoch": 1.6679164891350662, + "grad_norm": 13.9375, + "learning_rate": 1.6763631994287917e-07, + "loss": 0.5795, + "step": 19573 + }, + { + "epoch": 1.6680017043033661, + "grad_norm": 15.9375, + "learning_rate": 1.675522728257059e-07, + "loss": 0.6579, + "step": 19574 + }, + { + "epoch": 1.668086919471666, + "grad_norm": 17.5, + "learning_rate": 1.6746824526945163e-07, + "loss": 0.4819, + "step": 19575 + }, + { + "epoch": 1.668172134639966, + "grad_norm": 11.5625, + "learning_rate": 1.673842372756347e-07, + "loss": 0.3505, + "step": 19576 + }, + { + "epoch": 1.668257349808266, + "grad_norm": 14.25, + "learning_rate": 1.6730024884577354e-07, + "loss": 0.7329, + "step": 19577 + }, + { + "epoch": 1.668342564976566, + "grad_norm": 16.25, + "learning_rate": 1.6721627998138562e-07, + "loss": 0.453, + "step": 19578 + }, + { + "epoch": 1.668427780144866, + "grad_norm": 13.3125, + "learning_rate": 1.671323306839881e-07, + "loss": 0.4497, + "step": 19579 + }, + { + "epoch": 1.6685129953131659, + "grad_norm": 13.4375, + "learning_rate": 1.6704840095509823e-07, + "loss": 0.425, + "step": 19580 + }, + { + "epoch": 1.6685982104814658, + "grad_norm": 12.5625, + "learning_rate": 1.6696449079623282e-07, + "loss": 0.5173, + "step": 19581 + }, + { + "epoch": 1.6686834256497658, + "grad_norm": 19.25, + "learning_rate": 1.6688060020890837e-07, + "loss": 0.8994, + "step": 19582 + }, + { + "epoch": 1.6687686408180658, + "grad_norm": 16.875, + "learning_rate": 1.6679672919464026e-07, + "loss": 0.5687, + "step": 19583 + }, + { + "epoch": 1.6688538559863657, + "grad_norm": 20.0, + "learning_rate": 1.6671287775494469e-07, + "loss": 0.4264, + "step": 19584 + }, + { + "epoch": 1.6689390711546657, + "grad_norm": 12.875, + "learning_rate": 1.666290458913368e-07, + "loss": 0.4121, + "step": 19585 + }, + { + "epoch": 1.6690242863229656, + "grad_norm": 19.125, + "learning_rate": 1.6654523360533166e-07, + "loss": 0.645, + "step": 19586 + }, + { + "epoch": 1.6691095014912656, + "grad_norm": 11.3125, + "learning_rate": 1.664614408984433e-07, + "loss": 0.3593, + "step": 19587 + }, + { + "epoch": 1.6691947166595655, + "grad_norm": 11.1875, + "learning_rate": 1.6637766777218622e-07, + "loss": 0.3309, + "step": 19588 + }, + { + "epoch": 1.6692799318278655, + "grad_norm": 20.875, + "learning_rate": 1.662939142280745e-07, + "loss": 0.6345, + "step": 19589 + }, + { + "epoch": 1.6693651469961654, + "grad_norm": 16.75, + "learning_rate": 1.6621018026762164e-07, + "loss": 0.5042, + "step": 19590 + }, + { + "epoch": 1.6694503621644654, + "grad_norm": 12.25, + "learning_rate": 1.6612646589234046e-07, + "loss": 0.6791, + "step": 19591 + }, + { + "epoch": 1.6695355773327654, + "grad_norm": 17.375, + "learning_rate": 1.6604277110374392e-07, + "loss": 0.6099, + "step": 19592 + }, + { + "epoch": 1.6696207925010653, + "grad_norm": 26.5, + "learning_rate": 1.659590959033447e-07, + "loss": 0.9235, + "step": 19593 + }, + { + "epoch": 1.6697060076693653, + "grad_norm": 16.5, + "learning_rate": 1.6587544029265462e-07, + "loss": 0.8553, + "step": 19594 + }, + { + "epoch": 1.6697912228376652, + "grad_norm": 21.5, + "learning_rate": 1.6579180427318555e-07, + "loss": 0.877, + "step": 19595 + }, + { + "epoch": 1.6698764380059652, + "grad_norm": 16.125, + "learning_rate": 1.657081878464488e-07, + "loss": 0.8051, + "step": 19596 + }, + { + "epoch": 1.6699616531742651, + "grad_norm": 20.0, + "learning_rate": 1.6562459101395535e-07, + "loss": 0.9766, + "step": 19597 + }, + { + "epoch": 1.670046868342565, + "grad_norm": 18.625, + "learning_rate": 1.6554101377721592e-07, + "loss": 0.8149, + "step": 19598 + }, + { + "epoch": 1.670132083510865, + "grad_norm": 21.625, + "learning_rate": 1.6545745613774117e-07, + "loss": 0.8536, + "step": 19599 + }, + { + "epoch": 1.670217298679165, + "grad_norm": 13.125, + "learning_rate": 1.6537391809704039e-07, + "loss": 0.4464, + "step": 19600 + }, + { + "epoch": 1.670302513847465, + "grad_norm": 32.25, + "learning_rate": 1.6529039965662408e-07, + "loss": 0.4111, + "step": 19601 + }, + { + "epoch": 1.670387729015765, + "grad_norm": 21.75, + "learning_rate": 1.652069008180006e-07, + "loss": 0.7956, + "step": 19602 + }, + { + "epoch": 1.670472944184065, + "grad_norm": 19.125, + "learning_rate": 1.6512342158267956e-07, + "loss": 0.5453, + "step": 19603 + }, + { + "epoch": 1.6705581593523648, + "grad_norm": 12.625, + "learning_rate": 1.6503996195216892e-07, + "loss": 0.4224, + "step": 19604 + }, + { + "epoch": 1.6706433745206648, + "grad_norm": 11.875, + "learning_rate": 1.6495652192797724e-07, + "loss": 0.3406, + "step": 19605 + }, + { + "epoch": 1.6707285896889648, + "grad_norm": 13.0625, + "learning_rate": 1.6487310151161229e-07, + "loss": 0.5211, + "step": 19606 + }, + { + "epoch": 1.6708138048572647, + "grad_norm": 19.875, + "learning_rate": 1.6478970070458178e-07, + "loss": 0.6165, + "step": 19607 + }, + { + "epoch": 1.6708990200255647, + "grad_norm": 16.125, + "learning_rate": 1.6470631950839242e-07, + "loss": 0.8308, + "step": 19608 + }, + { + "epoch": 1.6709842351938646, + "grad_norm": 29.375, + "learning_rate": 1.646229579245516e-07, + "loss": 0.5105, + "step": 19609 + }, + { + "epoch": 1.6710694503621646, + "grad_norm": 17.875, + "learning_rate": 1.6453961595456494e-07, + "loss": 0.4058, + "step": 19610 + }, + { + "epoch": 1.6711546655304645, + "grad_norm": 20.875, + "learning_rate": 1.6445629359993901e-07, + "loss": 1.0022, + "step": 19611 + }, + { + "epoch": 1.6712398806987645, + "grad_norm": 14.5625, + "learning_rate": 1.6437299086217967e-07, + "loss": 0.5625, + "step": 19612 + }, + { + "epoch": 1.6713250958670645, + "grad_norm": 12.1875, + "learning_rate": 1.6428970774279185e-07, + "loss": 0.4387, + "step": 19613 + }, + { + "epoch": 1.6714103110353644, + "grad_norm": 22.25, + "learning_rate": 1.6420644424328073e-07, + "loss": 0.8694, + "step": 19614 + }, + { + "epoch": 1.6714955262036644, + "grad_norm": 12.4375, + "learning_rate": 1.6412320036515093e-07, + "loss": 0.4558, + "step": 19615 + }, + { + "epoch": 1.6715807413719643, + "grad_norm": 12.875, + "learning_rate": 1.640399761099072e-07, + "loss": 0.4908, + "step": 19616 + }, + { + "epoch": 1.6716659565402643, + "grad_norm": 20.875, + "learning_rate": 1.6395677147905312e-07, + "loss": 1.0584, + "step": 19617 + }, + { + "epoch": 1.6717511717085642, + "grad_norm": 20.875, + "learning_rate": 1.6387358647409203e-07, + "loss": 0.857, + "step": 19618 + }, + { + "epoch": 1.6718363868768642, + "grad_norm": 12.25, + "learning_rate": 1.6379042109652743e-07, + "loss": 0.4299, + "step": 19619 + }, + { + "epoch": 1.6719216020451642, + "grad_norm": 13.8125, + "learning_rate": 1.6370727534786247e-07, + "loss": 0.4542, + "step": 19620 + }, + { + "epoch": 1.6720068172134641, + "grad_norm": 11.9375, + "learning_rate": 1.6362414922959897e-07, + "loss": 0.3407, + "step": 19621 + }, + { + "epoch": 1.672092032381764, + "grad_norm": 41.75, + "learning_rate": 1.6354104274323977e-07, + "loss": 0.7355, + "step": 19622 + }, + { + "epoch": 1.672177247550064, + "grad_norm": 20.75, + "learning_rate": 1.634579558902863e-07, + "loss": 0.8174, + "step": 19623 + }, + { + "epoch": 1.672262462718364, + "grad_norm": 16.5, + "learning_rate": 1.633748886722404e-07, + "loss": 0.83, + "step": 19624 + }, + { + "epoch": 1.672347677886664, + "grad_norm": 14.3125, + "learning_rate": 1.6329184109060273e-07, + "loss": 0.3509, + "step": 19625 + }, + { + "epoch": 1.672432893054964, + "grad_norm": 17.5, + "learning_rate": 1.6320881314687453e-07, + "loss": 0.7026, + "step": 19626 + }, + { + "epoch": 1.6725181082232639, + "grad_norm": 19.875, + "learning_rate": 1.6312580484255562e-07, + "loss": 0.9518, + "step": 19627 + }, + { + "epoch": 1.6726033233915638, + "grad_norm": 14.0, + "learning_rate": 1.6304281617914642e-07, + "loss": 0.3117, + "step": 19628 + }, + { + "epoch": 1.6726885385598638, + "grad_norm": 13.75, + "learning_rate": 1.6295984715814687e-07, + "loss": 0.4702, + "step": 19629 + }, + { + "epoch": 1.6727737537281637, + "grad_norm": 16.375, + "learning_rate": 1.6287689778105563e-07, + "loss": 0.484, + "step": 19630 + }, + { + "epoch": 1.6728589688964637, + "grad_norm": 18.125, + "learning_rate": 1.6279396804937203e-07, + "loss": 0.8856, + "step": 19631 + }, + { + "epoch": 1.6729441840647636, + "grad_norm": 11.875, + "learning_rate": 1.6271105796459464e-07, + "loss": 0.3434, + "step": 19632 + }, + { + "epoch": 1.6730293992330636, + "grad_norm": 13.1875, + "learning_rate": 1.6262816752822223e-07, + "loss": 0.5004, + "step": 19633 + }, + { + "epoch": 1.6731146144013636, + "grad_norm": 21.125, + "learning_rate": 1.6254529674175206e-07, + "loss": 1.0232, + "step": 19634 + }, + { + "epoch": 1.6731998295696635, + "grad_norm": 14.125, + "learning_rate": 1.6246244560668186e-07, + "loss": 0.5048, + "step": 19635 + }, + { + "epoch": 1.6732850447379635, + "grad_norm": 18.25, + "learning_rate": 1.6237961412450877e-07, + "loss": 0.9212, + "step": 19636 + }, + { + "epoch": 1.6733702599062634, + "grad_norm": 14.75, + "learning_rate": 1.6229680229672987e-07, + "loss": 0.6671, + "step": 19637 + }, + { + "epoch": 1.6734554750745634, + "grad_norm": 13.25, + "learning_rate": 1.6221401012484136e-07, + "loss": 0.5043, + "step": 19638 + }, + { + "epoch": 1.6735406902428633, + "grad_norm": 12.875, + "learning_rate": 1.6213123761033957e-07, + "loss": 0.6146, + "step": 19639 + }, + { + "epoch": 1.6736259054111633, + "grad_norm": 14.1875, + "learning_rate": 1.620484847547203e-07, + "loss": 0.4952, + "step": 19640 + }, + { + "epoch": 1.6737111205794633, + "grad_norm": 26.125, + "learning_rate": 1.6196575155947914e-07, + "loss": 0.9596, + "step": 19641 + }, + { + "epoch": 1.6737963357477632, + "grad_norm": 19.25, + "learning_rate": 1.618830380261109e-07, + "loss": 0.6887, + "step": 19642 + }, + { + "epoch": 1.6738815509160632, + "grad_norm": 26.25, + "learning_rate": 1.6180034415611014e-07, + "loss": 0.6722, + "step": 19643 + }, + { + "epoch": 1.6739667660843631, + "grad_norm": 12.9375, + "learning_rate": 1.617176699509715e-07, + "loss": 0.4559, + "step": 19644 + }, + { + "epoch": 1.674051981252663, + "grad_norm": 9.9375, + "learning_rate": 1.6163501541218874e-07, + "loss": 0.4614, + "step": 19645 + }, + { + "epoch": 1.674137196420963, + "grad_norm": 15.25, + "learning_rate": 1.615523805412561e-07, + "loss": 0.64, + "step": 19646 + }, + { + "epoch": 1.674222411589263, + "grad_norm": 14.8125, + "learning_rate": 1.6146976533966616e-07, + "loss": 0.5161, + "step": 19647 + }, + { + "epoch": 1.674307626757563, + "grad_norm": 23.375, + "learning_rate": 1.6138716980891214e-07, + "loss": 0.9481, + "step": 19648 + }, + { + "epoch": 1.674392841925863, + "grad_norm": 27.5, + "learning_rate": 1.6130459395048688e-07, + "loss": 0.8486, + "step": 19649 + }, + { + "epoch": 1.6744780570941629, + "grad_norm": 19.0, + "learning_rate": 1.6122203776588241e-07, + "loss": 0.6844, + "step": 19650 + }, + { + "epoch": 1.6745632722624628, + "grad_norm": 16.75, + "learning_rate": 1.6113950125659028e-07, + "loss": 0.5532, + "step": 19651 + }, + { + "epoch": 1.6746484874307628, + "grad_norm": 10.1875, + "learning_rate": 1.6105698442410236e-07, + "loss": 0.3066, + "step": 19652 + }, + { + "epoch": 1.6747337025990627, + "grad_norm": 14.8125, + "learning_rate": 1.6097448726990968e-07, + "loss": 0.7653, + "step": 19653 + }, + { + "epoch": 1.6748189177673627, + "grad_norm": 15.25, + "learning_rate": 1.6089200979550342e-07, + "loss": 0.7557, + "step": 19654 + }, + { + "epoch": 1.6749041329356626, + "grad_norm": 10.5, + "learning_rate": 1.608095520023735e-07, + "loss": 0.1761, + "step": 19655 + }, + { + "epoch": 1.6749893481039626, + "grad_norm": 21.375, + "learning_rate": 1.6072711389201023e-07, + "loss": 0.8089, + "step": 19656 + }, + { + "epoch": 1.6750745632722626, + "grad_norm": 11.0, + "learning_rate": 1.6064469546590327e-07, + "loss": 0.3148, + "step": 19657 + }, + { + "epoch": 1.6751597784405625, + "grad_norm": 17.25, + "learning_rate": 1.6056229672554242e-07, + "loss": 0.6035, + "step": 19658 + }, + { + "epoch": 1.6752449936088625, + "grad_norm": 11.6875, + "learning_rate": 1.6047991767241645e-07, + "loss": 0.4517, + "step": 19659 + }, + { + "epoch": 1.6753302087771624, + "grad_norm": 12.625, + "learning_rate": 1.6039755830801362e-07, + "loss": 0.4029, + "step": 19660 + }, + { + "epoch": 1.6754154239454624, + "grad_norm": 14.875, + "learning_rate": 1.6031521863382249e-07, + "loss": 0.5574, + "step": 19661 + }, + { + "epoch": 1.6755006391137623, + "grad_norm": 14.8125, + "learning_rate": 1.6023289865133128e-07, + "loss": 0.412, + "step": 19662 + }, + { + "epoch": 1.6755858542820623, + "grad_norm": 17.375, + "learning_rate": 1.6015059836202759e-07, + "loss": 0.3952, + "step": 19663 + }, + { + "epoch": 1.6756710694503623, + "grad_norm": 15.125, + "learning_rate": 1.600683177673984e-07, + "loss": 0.5004, + "step": 19664 + }, + { + "epoch": 1.6757562846186622, + "grad_norm": 15.5, + "learning_rate": 1.5998605686893055e-07, + "loss": 0.7931, + "step": 19665 + }, + { + "epoch": 1.6758414997869622, + "grad_norm": 18.75, + "learning_rate": 1.5990381566811112e-07, + "loss": 0.6097, + "step": 19666 + }, + { + "epoch": 1.6759267149552621, + "grad_norm": 19.0, + "learning_rate": 1.5982159416642593e-07, + "loss": 0.7413, + "step": 19667 + }, + { + "epoch": 1.676011930123562, + "grad_norm": 23.125, + "learning_rate": 1.5973939236536048e-07, + "loss": 0.9017, + "step": 19668 + }, + { + "epoch": 1.676097145291862, + "grad_norm": 15.5625, + "learning_rate": 1.5965721026640055e-07, + "loss": 0.4956, + "step": 19669 + }, + { + "epoch": 1.676182360460162, + "grad_norm": 20.625, + "learning_rate": 1.595750478710313e-07, + "loss": 0.7898, + "step": 19670 + }, + { + "epoch": 1.676267575628462, + "grad_norm": 20.25, + "learning_rate": 1.594929051807377e-07, + "loss": 1.1608, + "step": 19671 + }, + { + "epoch": 1.676352790796762, + "grad_norm": 17.5, + "learning_rate": 1.5941078219700353e-07, + "loss": 0.6206, + "step": 19672 + }, + { + "epoch": 1.6764380059650619, + "grad_norm": 15.75, + "learning_rate": 1.5932867892131332e-07, + "loss": 0.51, + "step": 19673 + }, + { + "epoch": 1.6765232211333618, + "grad_norm": 25.375, + "learning_rate": 1.5924659535515087e-07, + "loss": 0.9831, + "step": 19674 + }, + { + "epoch": 1.6766084363016618, + "grad_norm": 19.0, + "learning_rate": 1.59164531499999e-07, + "loss": 0.7045, + "step": 19675 + }, + { + "epoch": 1.6766936514699617, + "grad_norm": 28.125, + "learning_rate": 1.5908248735734114e-07, + "loss": 0.5495, + "step": 19676 + }, + { + "epoch": 1.6767788666382617, + "grad_norm": 15.5, + "learning_rate": 1.5900046292865957e-07, + "loss": 0.4374, + "step": 19677 + }, + { + "epoch": 1.6768640818065617, + "grad_norm": 17.125, + "learning_rate": 1.589184582154367e-07, + "loss": 0.8149, + "step": 19678 + }, + { + "epoch": 1.6769492969748616, + "grad_norm": 15.0625, + "learning_rate": 1.588364732191547e-07, + "loss": 0.4686, + "step": 19679 + }, + { + "epoch": 1.6770345121431616, + "grad_norm": 17.125, + "learning_rate": 1.5875450794129472e-07, + "loss": 0.677, + "step": 19680 + }, + { + "epoch": 1.6771197273114615, + "grad_norm": 11.125, + "learning_rate": 1.5867256238333795e-07, + "loss": 0.2074, + "step": 19681 + }, + { + "epoch": 1.6772049424797615, + "grad_norm": 24.25, + "learning_rate": 1.5859063654676585e-07, + "loss": 0.997, + "step": 19682 + }, + { + "epoch": 1.6772901576480614, + "grad_norm": 12.4375, + "learning_rate": 1.5850873043305807e-07, + "loss": 0.2566, + "step": 19683 + }, + { + "epoch": 1.6773753728163614, + "grad_norm": 14.875, + "learning_rate": 1.5842684404369534e-07, + "loss": 0.4794, + "step": 19684 + }, + { + "epoch": 1.6774605879846614, + "grad_norm": 14.375, + "learning_rate": 1.5834497738015694e-07, + "loss": 0.3119, + "step": 19685 + }, + { + "epoch": 1.6775458031529613, + "grad_norm": 13.5625, + "learning_rate": 1.5826313044392262e-07, + "loss": 0.5822, + "step": 19686 + }, + { + "epoch": 1.6776310183212613, + "grad_norm": 22.875, + "learning_rate": 1.5818130323647124e-07, + "loss": 0.8634, + "step": 19687 + }, + { + "epoch": 1.6777162334895612, + "grad_norm": 13.3125, + "learning_rate": 1.58099495759282e-07, + "loss": 0.6015, + "step": 19688 + }, + { + "epoch": 1.6778014486578612, + "grad_norm": 15.4375, + "learning_rate": 1.5801770801383258e-07, + "loss": 0.6848, + "step": 19689 + }, + { + "epoch": 1.6778866638261611, + "grad_norm": 12.3125, + "learning_rate": 1.5793594000160117e-07, + "loss": 0.4451, + "step": 19690 + }, + { + "epoch": 1.677971878994461, + "grad_norm": 15.8125, + "learning_rate": 1.5785419172406585e-07, + "loss": 0.7328, + "step": 19691 + }, + { + "epoch": 1.678057094162761, + "grad_norm": 20.0, + "learning_rate": 1.5777246318270324e-07, + "loss": 0.8237, + "step": 19692 + }, + { + "epoch": 1.678142309331061, + "grad_norm": 12.3125, + "learning_rate": 1.5769075437899072e-07, + "loss": 0.4048, + "step": 19693 + }, + { + "epoch": 1.678227524499361, + "grad_norm": 14.9375, + "learning_rate": 1.576090653144044e-07, + "loss": 0.5528, + "step": 19694 + }, + { + "epoch": 1.678312739667661, + "grad_norm": 23.0, + "learning_rate": 1.5752739599042084e-07, + "loss": 1.0978, + "step": 19695 + }, + { + "epoch": 1.6783979548359609, + "grad_norm": 32.25, + "learning_rate": 1.5744574640851607e-07, + "loss": 0.7942, + "step": 19696 + }, + { + "epoch": 1.6784831700042608, + "grad_norm": 12.4375, + "learning_rate": 1.5736411657016504e-07, + "loss": 0.4339, + "step": 19697 + }, + { + "epoch": 1.6785683851725608, + "grad_norm": 19.25, + "learning_rate": 1.572825064768431e-07, + "loss": 0.9457, + "step": 19698 + }, + { + "epoch": 1.6786536003408608, + "grad_norm": 15.5625, + "learning_rate": 1.572009161300253e-07, + "loss": 0.714, + "step": 19699 + }, + { + "epoch": 1.6787388155091607, + "grad_norm": 12.9375, + "learning_rate": 1.5711934553118577e-07, + "loss": 0.595, + "step": 19700 + }, + { + "epoch": 1.6788240306774607, + "grad_norm": 22.75, + "learning_rate": 1.570377946817987e-07, + "loss": 0.7886, + "step": 19701 + }, + { + "epoch": 1.6789092458457606, + "grad_norm": 20.125, + "learning_rate": 1.5695626358333758e-07, + "loss": 0.7049, + "step": 19702 + }, + { + "epoch": 1.6789944610140606, + "grad_norm": 11.75, + "learning_rate": 1.5687475223727585e-07, + "loss": 0.5709, + "step": 19703 + }, + { + "epoch": 1.6790796761823605, + "grad_norm": 13.5, + "learning_rate": 1.567932606450867e-07, + "loss": 0.5358, + "step": 19704 + }, + { + "epoch": 1.6791648913506605, + "grad_norm": 12.875, + "learning_rate": 1.5671178880824284e-07, + "loss": 0.2842, + "step": 19705 + }, + { + "epoch": 1.6792501065189605, + "grad_norm": 13.0625, + "learning_rate": 1.5663033672821598e-07, + "loss": 0.5186, + "step": 19706 + }, + { + "epoch": 1.6793353216872604, + "grad_norm": 17.375, + "learning_rate": 1.565489044064787e-07, + "loss": 0.9706, + "step": 19707 + }, + { + "epoch": 1.6794205368555604, + "grad_norm": 14.375, + "learning_rate": 1.5646749184450207e-07, + "loss": 0.6231, + "step": 19708 + }, + { + "epoch": 1.6795057520238603, + "grad_norm": 18.875, + "learning_rate": 1.5638609904375782e-07, + "loss": 0.7224, + "step": 19709 + }, + { + "epoch": 1.6795909671921603, + "grad_norm": 18.25, + "learning_rate": 1.56304726005716e-07, + "loss": 0.5201, + "step": 19710 + }, + { + "epoch": 1.6796761823604602, + "grad_norm": 17.625, + "learning_rate": 1.562233727318478e-07, + "loss": 0.5264, + "step": 19711 + }, + { + "epoch": 1.6797613975287602, + "grad_norm": 14.4375, + "learning_rate": 1.5614203922362292e-07, + "loss": 0.4242, + "step": 19712 + }, + { + "epoch": 1.6798466126970601, + "grad_norm": 15.3125, + "learning_rate": 1.560607254825118e-07, + "loss": 0.5372, + "step": 19713 + }, + { + "epoch": 1.67993182786536, + "grad_norm": 12.9375, + "learning_rate": 1.5597943150998302e-07, + "loss": 0.3765, + "step": 19714 + }, + { + "epoch": 1.68001704303366, + "grad_norm": 13.75, + "learning_rate": 1.5589815730750634e-07, + "loss": 0.4476, + "step": 19715 + }, + { + "epoch": 1.68010225820196, + "grad_norm": 16.25, + "learning_rate": 1.5581690287654994e-07, + "loss": 0.5117, + "step": 19716 + }, + { + "epoch": 1.68018747337026, + "grad_norm": 16.875, + "learning_rate": 1.5573566821858233e-07, + "loss": 0.7862, + "step": 19717 + }, + { + "epoch": 1.68027268853856, + "grad_norm": 18.875, + "learning_rate": 1.5565445333507191e-07, + "loss": 0.4951, + "step": 19718 + }, + { + "epoch": 1.68035790370686, + "grad_norm": 11.6875, + "learning_rate": 1.5557325822748562e-07, + "loss": 0.3699, + "step": 19719 + }, + { + "epoch": 1.6804431188751598, + "grad_norm": 14.25, + "learning_rate": 1.5549208289729112e-07, + "loss": 0.9187, + "step": 19720 + }, + { + "epoch": 1.6805283340434598, + "grad_norm": 15.5, + "learning_rate": 1.5541092734595532e-07, + "loss": 0.6602, + "step": 19721 + }, + { + "epoch": 1.6806135492117598, + "grad_norm": 15.5, + "learning_rate": 1.5532979157494493e-07, + "loss": 0.5393, + "step": 19722 + }, + { + "epoch": 1.6806987643800597, + "grad_norm": 15.3125, + "learning_rate": 1.5524867558572617e-07, + "loss": 0.4877, + "step": 19723 + }, + { + "epoch": 1.6807839795483597, + "grad_norm": 16.375, + "learning_rate": 1.5516757937976427e-07, + "loss": 0.5432, + "step": 19724 + }, + { + "epoch": 1.6808691947166596, + "grad_norm": 16.25, + "learning_rate": 1.5508650295852523e-07, + "loss": 0.6153, + "step": 19725 + }, + { + "epoch": 1.6809544098849596, + "grad_norm": 30.125, + "learning_rate": 1.5500544632347432e-07, + "loss": 1.0579, + "step": 19726 + }, + { + "epoch": 1.6810396250532595, + "grad_norm": 13.0625, + "learning_rate": 1.5492440947607573e-07, + "loss": 0.5595, + "step": 19727 + }, + { + "epoch": 1.6811248402215595, + "grad_norm": 13.8125, + "learning_rate": 1.5484339241779445e-07, + "loss": 0.6792, + "step": 19728 + }, + { + "epoch": 1.6812100553898595, + "grad_norm": 19.75, + "learning_rate": 1.5476239515009413e-07, + "loss": 0.8399, + "step": 19729 + }, + { + "epoch": 1.6812952705581594, + "grad_norm": 14.375, + "learning_rate": 1.546814176744392e-07, + "loss": 0.6664, + "step": 19730 + }, + { + "epoch": 1.6813804857264594, + "grad_norm": 18.25, + "learning_rate": 1.546004599922922e-07, + "loss": 0.9092, + "step": 19731 + }, + { + "epoch": 1.6814657008947593, + "grad_norm": 18.375, + "learning_rate": 1.5451952210511628e-07, + "loss": 0.8535, + "step": 19732 + }, + { + "epoch": 1.6815509160630593, + "grad_norm": 16.5, + "learning_rate": 1.5443860401437417e-07, + "loss": 0.4278, + "step": 19733 + }, + { + "epoch": 1.6816361312313592, + "grad_norm": 13.6875, + "learning_rate": 1.5435770572152813e-07, + "loss": 0.6655, + "step": 19734 + }, + { + "epoch": 1.6817213463996592, + "grad_norm": 20.375, + "learning_rate": 1.5427682722804052e-07, + "loss": 0.5804, + "step": 19735 + }, + { + "epoch": 1.6818065615679592, + "grad_norm": 11.625, + "learning_rate": 1.5419596853537207e-07, + "loss": 0.2571, + "step": 19736 + }, + { + "epoch": 1.6818917767362591, + "grad_norm": 13.375, + "learning_rate": 1.541151296449843e-07, + "loss": 0.7082, + "step": 19737 + }, + { + "epoch": 1.681976991904559, + "grad_norm": 19.0, + "learning_rate": 1.5403431055833834e-07, + "loss": 0.7951, + "step": 19738 + }, + { + "epoch": 1.682062207072859, + "grad_norm": 15.75, + "learning_rate": 1.5395351127689486e-07, + "loss": 0.6159, + "step": 19739 + }, + { + "epoch": 1.682147422241159, + "grad_norm": 15.0, + "learning_rate": 1.5387273180211312e-07, + "loss": 0.3712, + "step": 19740 + }, + { + "epoch": 1.682232637409459, + "grad_norm": 13.5, + "learning_rate": 1.5379197213545334e-07, + "loss": 0.2122, + "step": 19741 + }, + { + "epoch": 1.682317852577759, + "grad_norm": 16.25, + "learning_rate": 1.5371123227837488e-07, + "loss": 0.9449, + "step": 19742 + }, + { + "epoch": 1.6824030677460589, + "grad_norm": 17.875, + "learning_rate": 1.5363051223233712e-07, + "loss": 0.8562, + "step": 19743 + }, + { + "epoch": 1.6824882829143588, + "grad_norm": 11.125, + "learning_rate": 1.535498119987984e-07, + "loss": 0.3278, + "step": 19744 + }, + { + "epoch": 1.6825734980826588, + "grad_norm": 16.5, + "learning_rate": 1.5346913157921694e-07, + "loss": 0.6628, + "step": 19745 + }, + { + "epoch": 1.6826587132509587, + "grad_norm": 13.8125, + "learning_rate": 1.5338847097505105e-07, + "loss": 0.4233, + "step": 19746 + }, + { + "epoch": 1.6827439284192587, + "grad_norm": 10.5625, + "learning_rate": 1.5330783018775844e-07, + "loss": 0.2306, + "step": 19747 + }, + { + "epoch": 1.6828291435875586, + "grad_norm": 12.375, + "learning_rate": 1.5322720921879614e-07, + "loss": 0.2267, + "step": 19748 + }, + { + "epoch": 1.6829143587558586, + "grad_norm": 21.75, + "learning_rate": 1.5314660806962076e-07, + "loss": 1.1723, + "step": 19749 + }, + { + "epoch": 1.6829995739241586, + "grad_norm": 12.3125, + "learning_rate": 1.530660267416892e-07, + "loss": 0.3432, + "step": 19750 + }, + { + "epoch": 1.6830847890924585, + "grad_norm": 12.75, + "learning_rate": 1.5298546523645762e-07, + "loss": 0.4834, + "step": 19751 + }, + { + "epoch": 1.6831700042607585, + "grad_norm": 16.625, + "learning_rate": 1.52904923555382e-07, + "loss": 0.4959, + "step": 19752 + }, + { + "epoch": 1.6832552194290584, + "grad_norm": 21.75, + "learning_rate": 1.528244016999174e-07, + "loss": 0.6294, + "step": 19753 + }, + { + "epoch": 1.6833404345973584, + "grad_norm": 17.25, + "learning_rate": 1.5274389967151916e-07, + "loss": 0.6183, + "step": 19754 + }, + { + "epoch": 1.6834256497656583, + "grad_norm": 17.625, + "learning_rate": 1.5266341747164224e-07, + "loss": 0.6248, + "step": 19755 + }, + { + "epoch": 1.6835108649339583, + "grad_norm": 18.375, + "learning_rate": 1.525829551017409e-07, + "loss": 0.6868, + "step": 19756 + }, + { + "epoch": 1.6835960801022583, + "grad_norm": 11.5625, + "learning_rate": 1.5250251256326884e-07, + "loss": 0.5363, + "step": 19757 + }, + { + "epoch": 1.6836812952705582, + "grad_norm": 12.5625, + "learning_rate": 1.524220898576799e-07, + "loss": 0.333, + "step": 19758 + }, + { + "epoch": 1.6837665104388582, + "grad_norm": 13.9375, + "learning_rate": 1.523416869864275e-07, + "loss": 0.7298, + "step": 19759 + }, + { + "epoch": 1.6838517256071581, + "grad_norm": 15.5, + "learning_rate": 1.5226130395096503e-07, + "loss": 0.5665, + "step": 19760 + }, + { + "epoch": 1.683936940775458, + "grad_norm": 16.875, + "learning_rate": 1.5218094075274428e-07, + "loss": 0.7613, + "step": 19761 + }, + { + "epoch": 1.684022155943758, + "grad_norm": 22.625, + "learning_rate": 1.5210059739321783e-07, + "loss": 0.7901, + "step": 19762 + }, + { + "epoch": 1.684107371112058, + "grad_norm": 18.25, + "learning_rate": 1.520202738738376e-07, + "loss": 0.6076, + "step": 19763 + }, + { + "epoch": 1.684192586280358, + "grad_norm": 15.9375, + "learning_rate": 1.519399701960554e-07, + "loss": 0.4777, + "step": 19764 + }, + { + "epoch": 1.684277801448658, + "grad_norm": 11.5625, + "learning_rate": 1.5185968636132227e-07, + "loss": 0.4876, + "step": 19765 + }, + { + "epoch": 1.6843630166169579, + "grad_norm": 13.75, + "learning_rate": 1.5177942237108846e-07, + "loss": 0.5104, + "step": 19766 + }, + { + "epoch": 1.6844482317852578, + "grad_norm": 14.625, + "learning_rate": 1.5169917822680485e-07, + "loss": 0.5821, + "step": 19767 + }, + { + "epoch": 1.6845334469535578, + "grad_norm": 29.25, + "learning_rate": 1.5161895392992162e-07, + "loss": 0.5703, + "step": 19768 + }, + { + "epoch": 1.6846186621218577, + "grad_norm": 10.5, + "learning_rate": 1.515387494818886e-07, + "loss": 0.3133, + "step": 19769 + }, + { + "epoch": 1.6847038772901577, + "grad_norm": 15.9375, + "learning_rate": 1.514585648841546e-07, + "loss": 0.5234, + "step": 19770 + }, + { + "epoch": 1.6847890924584576, + "grad_norm": 17.375, + "learning_rate": 1.513784001381692e-07, + "loss": 0.847, + "step": 19771 + }, + { + "epoch": 1.6848743076267576, + "grad_norm": 19.625, + "learning_rate": 1.5129825524538095e-07, + "loss": 1.0173, + "step": 19772 + }, + { + "epoch": 1.6849595227950576, + "grad_norm": 20.5, + "learning_rate": 1.5121813020723812e-07, + "loss": 0.675, + "step": 19773 + }, + { + "epoch": 1.6850447379633575, + "grad_norm": 12.9375, + "learning_rate": 1.511380250251883e-07, + "loss": 0.456, + "step": 19774 + }, + { + "epoch": 1.6851299531316575, + "grad_norm": 15.0, + "learning_rate": 1.5105793970067936e-07, + "loss": 0.4614, + "step": 19775 + }, + { + "epoch": 1.6852151682999574, + "grad_norm": 13.5625, + "learning_rate": 1.5097787423515864e-07, + "loss": 0.4237, + "step": 19776 + }, + { + "epoch": 1.6853003834682574, + "grad_norm": 11.8125, + "learning_rate": 1.5089782863007302e-07, + "loss": 0.4394, + "step": 19777 + }, + { + "epoch": 1.6853855986365573, + "grad_norm": 14.875, + "learning_rate": 1.508178028868687e-07, + "loss": 0.5848, + "step": 19778 + }, + { + "epoch": 1.6854708138048573, + "grad_norm": 13.0625, + "learning_rate": 1.507377970069919e-07, + "loss": 0.405, + "step": 19779 + }, + { + "epoch": 1.6855560289731573, + "grad_norm": 17.75, + "learning_rate": 1.5065781099188881e-07, + "loss": 0.4306, + "step": 19780 + }, + { + "epoch": 1.6856412441414572, + "grad_norm": 24.0, + "learning_rate": 1.5057784484300425e-07, + "loss": 0.7947, + "step": 19781 + }, + { + "epoch": 1.6857264593097572, + "grad_norm": 18.125, + "learning_rate": 1.5049789856178388e-07, + "loss": 0.7904, + "step": 19782 + }, + { + "epoch": 1.6858116744780571, + "grad_norm": 17.375, + "learning_rate": 1.504179721496718e-07, + "loss": 0.7298, + "step": 19783 + }, + { + "epoch": 1.685896889646357, + "grad_norm": 23.5, + "learning_rate": 1.503380656081127e-07, + "loss": 0.94, + "step": 19784 + }, + { + "epoch": 1.685982104814657, + "grad_norm": 14.0, + "learning_rate": 1.502581789385507e-07, + "loss": 0.3489, + "step": 19785 + }, + { + "epoch": 1.686067319982957, + "grad_norm": 14.6875, + "learning_rate": 1.501783121424294e-07, + "loss": 0.6618, + "step": 19786 + }, + { + "epoch": 1.686152535151257, + "grad_norm": 16.625, + "learning_rate": 1.5009846522119177e-07, + "loss": 0.3954, + "step": 19787 + }, + { + "epoch": 1.686237750319557, + "grad_norm": 26.25, + "learning_rate": 1.500186381762811e-07, + "loss": 0.6644, + "step": 19788 + }, + { + "epoch": 1.6863229654878569, + "grad_norm": 16.375, + "learning_rate": 1.4993883100913946e-07, + "loss": 0.5749, + "step": 19789 + }, + { + "epoch": 1.6864081806561568, + "grad_norm": 18.0, + "learning_rate": 1.4985904372120969e-07, + "loss": 0.8556, + "step": 19790 + }, + { + "epoch": 1.6864933958244568, + "grad_norm": 14.6875, + "learning_rate": 1.4977927631393302e-07, + "loss": 0.5212, + "step": 19791 + }, + { + "epoch": 1.6865786109927567, + "grad_norm": 13.25, + "learning_rate": 1.4969952878875108e-07, + "loss": 0.843, + "step": 19792 + }, + { + "epoch": 1.6866638261610567, + "grad_norm": 14.8125, + "learning_rate": 1.496198011471052e-07, + "loss": 0.3448, + "step": 19793 + }, + { + "epoch": 1.6867490413293567, + "grad_norm": 21.875, + "learning_rate": 1.4954009339043618e-07, + "loss": 0.7323, + "step": 19794 + }, + { + "epoch": 1.6868342564976566, + "grad_norm": 13.1875, + "learning_rate": 1.494604055201841e-07, + "loss": 0.5456, + "step": 19795 + }, + { + "epoch": 1.6869194716659566, + "grad_norm": 14.5625, + "learning_rate": 1.493807375377894e-07, + "loss": 0.6925, + "step": 19796 + }, + { + "epoch": 1.6870046868342565, + "grad_norm": 16.25, + "learning_rate": 1.4930108944469126e-07, + "loss": 0.5471, + "step": 19797 + }, + { + "epoch": 1.6870899020025565, + "grad_norm": 25.0, + "learning_rate": 1.4922146124232929e-07, + "loss": 0.8355, + "step": 19798 + }, + { + "epoch": 1.6871751171708564, + "grad_norm": 14.9375, + "learning_rate": 1.491418529321427e-07, + "loss": 0.4994, + "step": 19799 + }, + { + "epoch": 1.6872603323391564, + "grad_norm": 14.875, + "learning_rate": 1.4906226451556968e-07, + "loss": 0.3923, + "step": 19800 + }, + { + "epoch": 1.6873455475074564, + "grad_norm": 17.625, + "learning_rate": 1.489826959940484e-07, + "loss": 1.0337, + "step": 19801 + }, + { + "epoch": 1.6874307626757563, + "grad_norm": 17.0, + "learning_rate": 1.4890314736901737e-07, + "loss": 0.7032, + "step": 19802 + }, + { + "epoch": 1.6875159778440563, + "grad_norm": 16.875, + "learning_rate": 1.488236186419134e-07, + "loss": 0.7413, + "step": 19803 + }, + { + "epoch": 1.6876011930123562, + "grad_norm": 19.5, + "learning_rate": 1.4874410981417426e-07, + "loss": 0.7288, + "step": 19804 + }, + { + "epoch": 1.6876864081806562, + "grad_norm": 16.0, + "learning_rate": 1.486646208872361e-07, + "loss": 0.6496, + "step": 19805 + }, + { + "epoch": 1.6877716233489561, + "grad_norm": 17.25, + "learning_rate": 1.4858515186253587e-07, + "loss": 0.5152, + "step": 19806 + }, + { + "epoch": 1.687856838517256, + "grad_norm": 13.4375, + "learning_rate": 1.485057027415096e-07, + "loss": 0.3354, + "step": 19807 + }, + { + "epoch": 1.687942053685556, + "grad_norm": 19.5, + "learning_rate": 1.4842627352559272e-07, + "loss": 0.8337, + "step": 19808 + }, + { + "epoch": 1.688027268853856, + "grad_norm": 13.6875, + "learning_rate": 1.4834686421622086e-07, + "loss": 0.5229, + "step": 19809 + }, + { + "epoch": 1.688112484022156, + "grad_norm": 17.0, + "learning_rate": 1.4826747481482885e-07, + "loss": 0.6045, + "step": 19810 + }, + { + "epoch": 1.688197699190456, + "grad_norm": 13.8125, + "learning_rate": 1.4818810532285178e-07, + "loss": 0.3666, + "step": 19811 + }, + { + "epoch": 1.6882829143587559, + "grad_norm": 14.875, + "learning_rate": 1.4810875574172352e-07, + "loss": 0.4309, + "step": 19812 + }, + { + "epoch": 1.6883681295270558, + "grad_norm": 13.875, + "learning_rate": 1.480294260728779e-07, + "loss": 0.4722, + "step": 19813 + }, + { + "epoch": 1.6884533446953558, + "grad_norm": 17.125, + "learning_rate": 1.4795011631774866e-07, + "loss": 0.8184, + "step": 19814 + }, + { + "epoch": 1.6885385598636558, + "grad_norm": 18.125, + "learning_rate": 1.4787082647776908e-07, + "loss": 0.9364, + "step": 19815 + }, + { + "epoch": 1.6886237750319557, + "grad_norm": 15.3125, + "learning_rate": 1.477915565543721e-07, + "loss": 0.607, + "step": 19816 + }, + { + "epoch": 1.6887089902002557, + "grad_norm": 12.4375, + "learning_rate": 1.477123065489898e-07, + "loss": 0.4657, + "step": 19817 + }, + { + "epoch": 1.6887942053685556, + "grad_norm": 15.75, + "learning_rate": 1.476330764630546e-07, + "loss": 0.5009, + "step": 19818 + }, + { + "epoch": 1.6888794205368556, + "grad_norm": 14.0625, + "learning_rate": 1.475538662979985e-07, + "loss": 0.6768, + "step": 19819 + }, + { + "epoch": 1.6889646357051555, + "grad_norm": 17.125, + "learning_rate": 1.4747467605525257e-07, + "loss": 0.3975, + "step": 19820 + }, + { + "epoch": 1.6890498508734555, + "grad_norm": 16.375, + "learning_rate": 1.4739550573624764e-07, + "loss": 0.888, + "step": 19821 + }, + { + "epoch": 1.6891350660417555, + "grad_norm": 12.0625, + "learning_rate": 1.4731635534241477e-07, + "loss": 0.2654, + "step": 19822 + }, + { + "epoch": 1.6892202812100554, + "grad_norm": 11.375, + "learning_rate": 1.4723722487518412e-07, + "loss": 0.3611, + "step": 19823 + }, + { + "epoch": 1.6893054963783554, + "grad_norm": 18.75, + "learning_rate": 1.4715811433598598e-07, + "loss": 0.7624, + "step": 19824 + }, + { + "epoch": 1.6893907115466553, + "grad_norm": 27.5, + "learning_rate": 1.4707902372624954e-07, + "loss": 0.9125, + "step": 19825 + }, + { + "epoch": 1.6894759267149553, + "grad_norm": 14.6875, + "learning_rate": 1.469999530474041e-07, + "loss": 0.7287, + "step": 19826 + }, + { + "epoch": 1.6895611418832552, + "grad_norm": 14.25, + "learning_rate": 1.4692090230087871e-07, + "loss": 0.7356, + "step": 19827 + }, + { + "epoch": 1.6896463570515552, + "grad_norm": 15.125, + "learning_rate": 1.4684187148810217e-07, + "loss": 0.6066, + "step": 19828 + }, + { + "epoch": 1.6897315722198551, + "grad_norm": 15.8125, + "learning_rate": 1.467628606105022e-07, + "loss": 0.549, + "step": 19829 + }, + { + "epoch": 1.689816787388155, + "grad_norm": 16.625, + "learning_rate": 1.466838696695065e-07, + "loss": 0.6604, + "step": 19830 + }, + { + "epoch": 1.689902002556455, + "grad_norm": 14.5, + "learning_rate": 1.466048986665429e-07, + "loss": 0.4092, + "step": 19831 + }, + { + "epoch": 1.689987217724755, + "grad_norm": 20.625, + "learning_rate": 1.465259476030384e-07, + "loss": 0.7878, + "step": 19832 + }, + { + "epoch": 1.690072432893055, + "grad_norm": 21.375, + "learning_rate": 1.464470164804195e-07, + "loss": 0.7124, + "step": 19833 + }, + { + "epoch": 1.690157648061355, + "grad_norm": 18.125, + "learning_rate": 1.4636810530011258e-07, + "loss": 0.7486, + "step": 19834 + }, + { + "epoch": 1.690242863229655, + "grad_norm": 15.4375, + "learning_rate": 1.46289214063544e-07, + "loss": 0.8039, + "step": 19835 + }, + { + "epoch": 1.6903280783979548, + "grad_norm": 15.0, + "learning_rate": 1.4621034277213925e-07, + "loss": 0.7368, + "step": 19836 + }, + { + "epoch": 1.6904132935662548, + "grad_norm": 15.625, + "learning_rate": 1.4613149142732374e-07, + "loss": 0.7952, + "step": 19837 + }, + { + "epoch": 1.6904985087345548, + "grad_norm": 19.625, + "learning_rate": 1.4605266003052178e-07, + "loss": 0.5878, + "step": 19838 + }, + { + "epoch": 1.6905837239028547, + "grad_norm": 16.25, + "learning_rate": 1.459738485831584e-07, + "loss": 0.7525, + "step": 19839 + }, + { + "epoch": 1.6906689390711547, + "grad_norm": 16.75, + "learning_rate": 1.4589505708665782e-07, + "loss": 0.6464, + "step": 19840 + }, + { + "epoch": 1.6907541542394546, + "grad_norm": 16.625, + "learning_rate": 1.4581628554244402e-07, + "loss": 0.6383, + "step": 19841 + }, + { + "epoch": 1.6908393694077546, + "grad_norm": 15.5, + "learning_rate": 1.4573753395194013e-07, + "loss": 0.7119, + "step": 19842 + }, + { + "epoch": 1.6909245845760545, + "grad_norm": 20.75, + "learning_rate": 1.4565880231656934e-07, + "loss": 0.4111, + "step": 19843 + }, + { + "epoch": 1.6910097997443545, + "grad_norm": 14.3125, + "learning_rate": 1.4558009063775442e-07, + "loss": 0.5719, + "step": 19844 + }, + { + "epoch": 1.6910950149126545, + "grad_norm": 17.625, + "learning_rate": 1.4550139891691824e-07, + "loss": 0.6986, + "step": 19845 + }, + { + "epoch": 1.6911802300809544, + "grad_norm": 13.3125, + "learning_rate": 1.454227271554824e-07, + "loss": 0.4287, + "step": 19846 + }, + { + "epoch": 1.6912654452492544, + "grad_norm": 29.125, + "learning_rate": 1.4534407535486836e-07, + "loss": 0.7954, + "step": 19847 + }, + { + "epoch": 1.6913506604175543, + "grad_norm": 11.5625, + "learning_rate": 1.4526544351649768e-07, + "loss": 0.3777, + "step": 19848 + }, + { + "epoch": 1.6914358755858543, + "grad_norm": 16.25, + "learning_rate": 1.4518683164179147e-07, + "loss": 0.3116, + "step": 19849 + }, + { + "epoch": 1.6915210907541542, + "grad_norm": 13.9375, + "learning_rate": 1.4510823973217014e-07, + "loss": 0.5906, + "step": 19850 + }, + { + "epoch": 1.6916063059224542, + "grad_norm": 17.0, + "learning_rate": 1.4502966778905379e-07, + "loss": 0.4817, + "step": 19851 + }, + { + "epoch": 1.6916915210907542, + "grad_norm": 17.0, + "learning_rate": 1.4495111581386264e-07, + "loss": 0.4834, + "step": 19852 + }, + { + "epoch": 1.6917767362590541, + "grad_norm": 11.3125, + "learning_rate": 1.4487258380801617e-07, + "loss": 0.3179, + "step": 19853 + }, + { + "epoch": 1.691861951427354, + "grad_norm": 17.0, + "learning_rate": 1.4479407177293348e-07, + "loss": 0.5432, + "step": 19854 + }, + { + "epoch": 1.691947166595654, + "grad_norm": 10.25, + "learning_rate": 1.4471557971003304e-07, + "loss": 0.2705, + "step": 19855 + }, + { + "epoch": 1.692032381763954, + "grad_norm": 17.875, + "learning_rate": 1.446371076207334e-07, + "loss": 0.7952, + "step": 19856 + }, + { + "epoch": 1.692117596932254, + "grad_norm": 10.375, + "learning_rate": 1.4455865550645292e-07, + "loss": 0.2833, + "step": 19857 + }, + { + "epoch": 1.692202812100554, + "grad_norm": 16.75, + "learning_rate": 1.4448022336860933e-07, + "loss": 0.821, + "step": 19858 + }, + { + "epoch": 1.6922880272688539, + "grad_norm": 16.25, + "learning_rate": 1.4440181120861942e-07, + "loss": 0.391, + "step": 19859 + }, + { + "epoch": 1.6923732424371538, + "grad_norm": 21.625, + "learning_rate": 1.4432341902790068e-07, + "loss": 0.9421, + "step": 19860 + }, + { + "epoch": 1.6924584576054538, + "grad_norm": 16.5, + "learning_rate": 1.4424504682786985e-07, + "loss": 0.6095, + "step": 19861 + }, + { + "epoch": 1.6925436727737537, + "grad_norm": 12.4375, + "learning_rate": 1.4416669460994265e-07, + "loss": 0.4578, + "step": 19862 + }, + { + "epoch": 1.6926288879420537, + "grad_norm": 16.625, + "learning_rate": 1.4408836237553555e-07, + "loss": 0.4686, + "step": 19863 + }, + { + "epoch": 1.6927141031103536, + "grad_norm": 15.125, + "learning_rate": 1.4401005012606357e-07, + "loss": 0.6618, + "step": 19864 + }, + { + "epoch": 1.6927993182786536, + "grad_norm": 16.625, + "learning_rate": 1.439317578629422e-07, + "loss": 0.5734, + "step": 19865 + }, + { + "epoch": 1.6928845334469536, + "grad_norm": 14.375, + "learning_rate": 1.438534855875863e-07, + "loss": 0.5537, + "step": 19866 + }, + { + "epoch": 1.6929697486152535, + "grad_norm": 10.375, + "learning_rate": 1.4377523330141016e-07, + "loss": 0.2787, + "step": 19867 + }, + { + "epoch": 1.6930549637835535, + "grad_norm": 20.0, + "learning_rate": 1.436970010058278e-07, + "loss": 0.6065, + "step": 19868 + }, + { + "epoch": 1.6931401789518534, + "grad_norm": 11.125, + "learning_rate": 1.4361878870225345e-07, + "loss": 0.6043, + "step": 19869 + }, + { + "epoch": 1.6932253941201534, + "grad_norm": 15.375, + "learning_rate": 1.435405963920998e-07, + "loss": 0.5849, + "step": 19870 + }, + { + "epoch": 1.6933106092884533, + "grad_norm": 14.125, + "learning_rate": 1.4346242407678054e-07, + "loss": 0.5293, + "step": 19871 + }, + { + "epoch": 1.6933958244567533, + "grad_norm": 17.875, + "learning_rate": 1.4338427175770775e-07, + "loss": 0.5661, + "step": 19872 + }, + { + "epoch": 1.6934810396250533, + "grad_norm": 12.375, + "learning_rate": 1.4330613943629391e-07, + "loss": 0.407, + "step": 19873 + }, + { + "epoch": 1.6935662547933532, + "grad_norm": 15.9375, + "learning_rate": 1.4322802711395106e-07, + "loss": 0.6975, + "step": 19874 + }, + { + "epoch": 1.6936514699616532, + "grad_norm": 16.0, + "learning_rate": 1.4314993479209088e-07, + "loss": 0.3579, + "step": 19875 + }, + { + "epoch": 1.6937366851299531, + "grad_norm": 13.875, + "learning_rate": 1.4307186247212418e-07, + "loss": 0.4404, + "step": 19876 + }, + { + "epoch": 1.693821900298253, + "grad_norm": 18.0, + "learning_rate": 1.429938101554622e-07, + "loss": 0.7064, + "step": 19877 + }, + { + "epoch": 1.693907115466553, + "grad_norm": 13.5, + "learning_rate": 1.4291577784351516e-07, + "loss": 0.5175, + "step": 19878 + }, + { + "epoch": 1.693992330634853, + "grad_norm": 16.25, + "learning_rate": 1.4283776553769336e-07, + "loss": 0.5695, + "step": 19879 + }, + { + "epoch": 1.694077545803153, + "grad_norm": 17.0, + "learning_rate": 1.4275977323940636e-07, + "loss": 0.6726, + "step": 19880 + }, + { + "epoch": 1.694162760971453, + "grad_norm": 18.875, + "learning_rate": 1.4268180095006357e-07, + "loss": 0.7951, + "step": 19881 + }, + { + "epoch": 1.6942479761397529, + "grad_norm": 13.5, + "learning_rate": 1.4260384867107416e-07, + "loss": 0.6943, + "step": 19882 + }, + { + "epoch": 1.6943331913080528, + "grad_norm": 16.5, + "learning_rate": 1.4252591640384713e-07, + "loss": 0.7309, + "step": 19883 + }, + { + "epoch": 1.6944184064763528, + "grad_norm": 19.375, + "learning_rate": 1.4244800414979e-07, + "loss": 0.8296, + "step": 19884 + }, + { + "epoch": 1.6945036216446527, + "grad_norm": 60.25, + "learning_rate": 1.4237011191031158e-07, + "loss": 1.2402, + "step": 19885 + }, + { + "epoch": 1.6945888368129527, + "grad_norm": 20.125, + "learning_rate": 1.422922396868187e-07, + "loss": 0.7189, + "step": 19886 + }, + { + "epoch": 1.6946740519812526, + "grad_norm": 18.375, + "learning_rate": 1.42214387480719e-07, + "loss": 0.6563, + "step": 19887 + }, + { + "epoch": 1.6947592671495526, + "grad_norm": 17.375, + "learning_rate": 1.4213655529341952e-07, + "loss": 0.9295, + "step": 19888 + }, + { + "epoch": 1.6948444823178526, + "grad_norm": 11.9375, + "learning_rate": 1.4205874312632624e-07, + "loss": 0.313, + "step": 19889 + }, + { + "epoch": 1.6949296974861525, + "grad_norm": 14.6875, + "learning_rate": 1.4198095098084563e-07, + "loss": 0.5131, + "step": 19890 + }, + { + "epoch": 1.6950149126544525, + "grad_norm": 18.125, + "learning_rate": 1.4190317885838343e-07, + "loss": 0.9255, + "step": 19891 + }, + { + "epoch": 1.6951001278227524, + "grad_norm": 9.3125, + "learning_rate": 1.4182542676034526e-07, + "loss": 0.2048, + "step": 19892 + }, + { + "epoch": 1.6951853429910524, + "grad_norm": 16.375, + "learning_rate": 1.4174769468813577e-07, + "loss": 0.569, + "step": 19893 + }, + { + "epoch": 1.6952705581593523, + "grad_norm": 16.625, + "learning_rate": 1.4166998264316015e-07, + "loss": 0.6468, + "step": 19894 + }, + { + "epoch": 1.6953557733276523, + "grad_norm": 16.0, + "learning_rate": 1.4159229062682218e-07, + "loss": 0.7157, + "step": 19895 + }, + { + "epoch": 1.6954409884959523, + "grad_norm": 12.3125, + "learning_rate": 1.415146186405264e-07, + "loss": 0.4289, + "step": 19896 + }, + { + "epoch": 1.6955262036642522, + "grad_norm": 13.875, + "learning_rate": 1.4143696668567588e-07, + "loss": 0.7149, + "step": 19897 + }, + { + "epoch": 1.6956114188325522, + "grad_norm": 12.875, + "learning_rate": 1.413593347636741e-07, + "loss": 0.3794, + "step": 19898 + }, + { + "epoch": 1.6956966340008521, + "grad_norm": 14.125, + "learning_rate": 1.412817228759239e-07, + "loss": 0.5518, + "step": 19899 + }, + { + "epoch": 1.695781849169152, + "grad_norm": 17.25, + "learning_rate": 1.412041310238281e-07, + "loss": 0.5134, + "step": 19900 + }, + { + "epoch": 1.695867064337452, + "grad_norm": 15.5, + "learning_rate": 1.411265592087885e-07, + "loss": 0.685, + "step": 19901 + }, + { + "epoch": 1.695952279505752, + "grad_norm": 12.9375, + "learning_rate": 1.4104900743220725e-07, + "loss": 0.4813, + "step": 19902 + }, + { + "epoch": 1.696037494674052, + "grad_norm": 18.125, + "learning_rate": 1.4097147569548525e-07, + "loss": 0.8228, + "step": 19903 + }, + { + "epoch": 1.696122709842352, + "grad_norm": 13.875, + "learning_rate": 1.4089396400002397e-07, + "loss": 0.467, + "step": 19904 + }, + { + "epoch": 1.6962079250106519, + "grad_norm": 17.25, + "learning_rate": 1.4081647234722427e-07, + "loss": 0.93, + "step": 19905 + }, + { + "epoch": 1.6962931401789518, + "grad_norm": 14.3125, + "learning_rate": 1.407390007384861e-07, + "loss": 0.7149, + "step": 19906 + }, + { + "epoch": 1.6963783553472518, + "grad_norm": 16.5, + "learning_rate": 1.406615491752096e-07, + "loss": 0.5214, + "step": 19907 + }, + { + "epoch": 1.6964635705155517, + "grad_norm": 14.5, + "learning_rate": 1.4058411765879436e-07, + "loss": 0.4017, + "step": 19908 + }, + { + "epoch": 1.6965487856838517, + "grad_norm": 16.875, + "learning_rate": 1.405067061906401e-07, + "loss": 0.2838, + "step": 19909 + }, + { + "epoch": 1.6966340008521517, + "grad_norm": 15.9375, + "learning_rate": 1.4042931477214528e-07, + "loss": 0.5578, + "step": 19910 + }, + { + "epoch": 1.6967192160204516, + "grad_norm": 13.375, + "learning_rate": 1.403519434047082e-07, + "loss": 0.5497, + "step": 19911 + }, + { + "epoch": 1.6968044311887516, + "grad_norm": 18.875, + "learning_rate": 1.4027459208972734e-07, + "loss": 0.3461, + "step": 19912 + }, + { + "epoch": 1.6968896463570515, + "grad_norm": 15.625, + "learning_rate": 1.401972608286009e-07, + "loss": 0.5493, + "step": 19913 + }, + { + "epoch": 1.6969748615253515, + "grad_norm": 23.25, + "learning_rate": 1.4011994962272553e-07, + "loss": 0.5026, + "step": 19914 + }, + { + "epoch": 1.6970600766936514, + "grad_norm": 15.0625, + "learning_rate": 1.4004265847349886e-07, + "loss": 0.6886, + "step": 19915 + }, + { + "epoch": 1.6971452918619514, + "grad_norm": 14.125, + "learning_rate": 1.3996538738231742e-07, + "loss": 0.5768, + "step": 19916 + }, + { + "epoch": 1.6972305070302514, + "grad_norm": 10.8125, + "learning_rate": 1.3988813635057802e-07, + "loss": 0.3552, + "step": 19917 + }, + { + "epoch": 1.6973157221985513, + "grad_norm": 11.125, + "learning_rate": 1.3981090537967617e-07, + "loss": 0.2984, + "step": 19918 + }, + { + "epoch": 1.6974009373668513, + "grad_norm": 17.5, + "learning_rate": 1.3973369447100747e-07, + "loss": 1.0995, + "step": 19919 + }, + { + "epoch": 1.6974861525351512, + "grad_norm": 17.25, + "learning_rate": 1.396565036259673e-07, + "loss": 0.4521, + "step": 19920 + }, + { + "epoch": 1.6975713677034512, + "grad_norm": 14.8125, + "learning_rate": 1.3957933284595078e-07, + "loss": 0.5068, + "step": 19921 + }, + { + "epoch": 1.6976565828717511, + "grad_norm": 17.0, + "learning_rate": 1.395021821323525e-07, + "loss": 0.6094, + "step": 19922 + }, + { + "epoch": 1.697741798040051, + "grad_norm": 21.0, + "learning_rate": 1.394250514865661e-07, + "loss": 0.8253, + "step": 19923 + }, + { + "epoch": 1.697827013208351, + "grad_norm": 15.125, + "learning_rate": 1.39347940909986e-07, + "loss": 0.3387, + "step": 19924 + }, + { + "epoch": 1.697912228376651, + "grad_norm": 17.625, + "learning_rate": 1.3927085040400538e-07, + "loss": 0.7344, + "step": 19925 + }, + { + "epoch": 1.697997443544951, + "grad_norm": 15.625, + "learning_rate": 1.3919377997001786e-07, + "loss": 0.8281, + "step": 19926 + }, + { + "epoch": 1.698082658713251, + "grad_norm": 17.0, + "learning_rate": 1.3911672960941523e-07, + "loss": 0.3442, + "step": 19927 + }, + { + "epoch": 1.6981678738815509, + "grad_norm": 19.875, + "learning_rate": 1.3903969932359054e-07, + "loss": 0.4838, + "step": 19928 + }, + { + "epoch": 1.6982530890498508, + "grad_norm": 34.75, + "learning_rate": 1.389626891139355e-07, + "loss": 0.6817, + "step": 19929 + }, + { + "epoch": 1.6983383042181508, + "grad_norm": 15.625, + "learning_rate": 1.388856989818421e-07, + "loss": 0.4962, + "step": 19930 + }, + { + "epoch": 1.6984235193864508, + "grad_norm": 16.875, + "learning_rate": 1.388087289287013e-07, + "loss": 0.833, + "step": 19931 + }, + { + "epoch": 1.6985087345547507, + "grad_norm": 16.875, + "learning_rate": 1.387317789559041e-07, + "loss": 0.9758, + "step": 19932 + }, + { + "epoch": 1.6985939497230507, + "grad_norm": 22.25, + "learning_rate": 1.3865484906484122e-07, + "loss": 0.8172, + "step": 19933 + }, + { + "epoch": 1.6986791648913506, + "grad_norm": 18.625, + "learning_rate": 1.3857793925690282e-07, + "loss": 0.4925, + "step": 19934 + }, + { + "epoch": 1.6987643800596506, + "grad_norm": 19.25, + "learning_rate": 1.385010495334789e-07, + "loss": 0.6085, + "step": 19935 + }, + { + "epoch": 1.6988495952279505, + "grad_norm": 20.5, + "learning_rate": 1.3842417989595832e-07, + "loss": 0.9952, + "step": 19936 + }, + { + "epoch": 1.6989348103962505, + "grad_norm": 15.5, + "learning_rate": 1.3834733034573055e-07, + "loss": 0.6747, + "step": 19937 + }, + { + "epoch": 1.6990200255645505, + "grad_norm": 20.0, + "learning_rate": 1.3827050088418447e-07, + "loss": 0.8391, + "step": 19938 + }, + { + "epoch": 1.6991052407328504, + "grad_norm": 14.125, + "learning_rate": 1.3819369151270856e-07, + "loss": 0.4728, + "step": 19939 + }, + { + "epoch": 1.6991904559011504, + "grad_norm": 14.75, + "learning_rate": 1.3811690223269044e-07, + "loss": 0.2329, + "step": 19940 + }, + { + "epoch": 1.6992756710694503, + "grad_norm": 10.75, + "learning_rate": 1.3804013304551806e-07, + "loss": 0.4992, + "step": 19941 + }, + { + "epoch": 1.6993608862377503, + "grad_norm": 20.0, + "learning_rate": 1.3796338395257875e-07, + "loss": 0.8773, + "step": 19942 + }, + { + "epoch": 1.6994461014060502, + "grad_norm": 17.375, + "learning_rate": 1.3788665495525934e-07, + "loss": 0.435, + "step": 19943 + }, + { + "epoch": 1.6995313165743502, + "grad_norm": 18.75, + "learning_rate": 1.3780994605494618e-07, + "loss": 0.6134, + "step": 19944 + }, + { + "epoch": 1.6996165317426501, + "grad_norm": 18.375, + "learning_rate": 1.3773325725302572e-07, + "loss": 0.9821, + "step": 19945 + }, + { + "epoch": 1.69970174691095, + "grad_norm": 14.8125, + "learning_rate": 1.3765658855088377e-07, + "loss": 0.7667, + "step": 19946 + }, + { + "epoch": 1.69978696207925, + "grad_norm": 13.6875, + "learning_rate": 1.37579939949906e-07, + "loss": 0.6059, + "step": 19947 + }, + { + "epoch": 1.69987217724755, + "grad_norm": 17.625, + "learning_rate": 1.375033114514772e-07, + "loss": 0.5839, + "step": 19948 + }, + { + "epoch": 1.69995739241585, + "grad_norm": 15.125, + "learning_rate": 1.3742670305698217e-07, + "loss": 0.5672, + "step": 19949 + }, + { + "epoch": 1.70004260758415, + "grad_norm": 14.75, + "learning_rate": 1.3735011476780565e-07, + "loss": 0.5276, + "step": 19950 + }, + { + "epoch": 1.70012782275245, + "grad_norm": 16.0, + "learning_rate": 1.3727354658533114e-07, + "loss": 0.6352, + "step": 19951 + }, + { + "epoch": 1.7002130379207498, + "grad_norm": 17.625, + "learning_rate": 1.3719699851094295e-07, + "loss": 0.9877, + "step": 19952 + }, + { + "epoch": 1.7002982530890498, + "grad_norm": 22.875, + "learning_rate": 1.371204705460237e-07, + "loss": 1.067, + "step": 19953 + }, + { + "epoch": 1.7003834682573498, + "grad_norm": 15.625, + "learning_rate": 1.370439626919566e-07, + "loss": 0.4755, + "step": 19954 + }, + { + "epoch": 1.7004686834256497, + "grad_norm": 16.75, + "learning_rate": 1.3696747495012417e-07, + "loss": 0.7257, + "step": 19955 + }, + { + "epoch": 1.7005538985939497, + "grad_norm": 11.8125, + "learning_rate": 1.3689100732190916e-07, + "loss": 0.3891, + "step": 19956 + }, + { + "epoch": 1.7006391137622496, + "grad_norm": 14.0, + "learning_rate": 1.3681455980869257e-07, + "loss": 0.6662, + "step": 19957 + }, + { + "epoch": 1.7007243289305496, + "grad_norm": 19.5, + "learning_rate": 1.3673813241185634e-07, + "loss": 0.5472, + "step": 19958 + }, + { + "epoch": 1.7008095440988495, + "grad_norm": 11.875, + "learning_rate": 1.3666172513278175e-07, + "loss": 0.3451, + "step": 19959 + }, + { + "epoch": 1.7008947592671495, + "grad_norm": 12.4375, + "learning_rate": 1.3658533797284935e-07, + "loss": 0.344, + "step": 19960 + }, + { + "epoch": 1.7009799744354495, + "grad_norm": 14.9375, + "learning_rate": 1.3650897093343927e-07, + "loss": 0.6649, + "step": 19961 + }, + { + "epoch": 1.7010651896037494, + "grad_norm": 15.1875, + "learning_rate": 1.3643262401593166e-07, + "loss": 0.5444, + "step": 19962 + }, + { + "epoch": 1.7011504047720494, + "grad_norm": 11.3125, + "learning_rate": 1.3635629722170653e-07, + "loss": 0.3413, + "step": 19963 + }, + { + "epoch": 1.7012356199403493, + "grad_norm": 27.0, + "learning_rate": 1.3627999055214308e-07, + "loss": 0.6247, + "step": 19964 + }, + { + "epoch": 1.7013208351086493, + "grad_norm": 16.625, + "learning_rate": 1.3620370400861992e-07, + "loss": 0.6945, + "step": 19965 + }, + { + "epoch": 1.7014060502769492, + "grad_norm": 13.5, + "learning_rate": 1.3612743759251575e-07, + "loss": 0.4382, + "step": 19966 + }, + { + "epoch": 1.7014912654452492, + "grad_norm": 22.0, + "learning_rate": 1.3605119130520927e-07, + "loss": 0.7941, + "step": 19967 + }, + { + "epoch": 1.7015764806135492, + "grad_norm": 17.75, + "learning_rate": 1.3597496514807754e-07, + "loss": 0.4649, + "step": 19968 + }, + { + "epoch": 1.7016616957818491, + "grad_norm": 13.0625, + "learning_rate": 1.3589875912249874e-07, + "loss": 0.3829, + "step": 19969 + }, + { + "epoch": 1.701746910950149, + "grad_norm": 16.75, + "learning_rate": 1.3582257322984946e-07, + "loss": 0.4994, + "step": 19970 + }, + { + "epoch": 1.701832126118449, + "grad_norm": 19.125, + "learning_rate": 1.357464074715066e-07, + "loss": 0.8673, + "step": 19971 + }, + { + "epoch": 1.701917341286749, + "grad_norm": 11.8125, + "learning_rate": 1.3567026184884702e-07, + "loss": 0.5102, + "step": 19972 + }, + { + "epoch": 1.702002556455049, + "grad_norm": 17.5, + "learning_rate": 1.3559413636324598e-07, + "loss": 0.7309, + "step": 19973 + }, + { + "epoch": 1.702087771623349, + "grad_norm": 19.125, + "learning_rate": 1.355180310160796e-07, + "loss": 0.6753, + "step": 19974 + }, + { + "epoch": 1.7021729867916489, + "grad_norm": 15.75, + "learning_rate": 1.354419458087232e-07, + "loss": 0.6555, + "step": 19975 + }, + { + "epoch": 1.7022582019599488, + "grad_norm": 16.75, + "learning_rate": 1.3536588074255147e-07, + "loss": 0.6903, + "step": 19976 + }, + { + "epoch": 1.7023434171282488, + "grad_norm": 23.625, + "learning_rate": 1.3528983581893946e-07, + "loss": 0.7415, + "step": 19977 + }, + { + "epoch": 1.7024286322965487, + "grad_norm": 16.375, + "learning_rate": 1.3521381103926062e-07, + "loss": 0.5858, + "step": 19978 + }, + { + "epoch": 1.7025138474648487, + "grad_norm": 13.125, + "learning_rate": 1.351378064048893e-07, + "loss": 0.3025, + "step": 19979 + }, + { + "epoch": 1.7025990626331486, + "grad_norm": 13.8125, + "learning_rate": 1.3506182191719884e-07, + "loss": 0.5214, + "step": 19980 + }, + { + "epoch": 1.7026842778014486, + "grad_norm": 17.5, + "learning_rate": 1.3498585757756283e-07, + "loss": 0.5411, + "step": 19981 + }, + { + "epoch": 1.7027694929697486, + "grad_norm": 15.5, + "learning_rate": 1.3490991338735313e-07, + "loss": 0.5804, + "step": 19982 + }, + { + "epoch": 1.7028547081380485, + "grad_norm": 14.8125, + "learning_rate": 1.3483398934794307e-07, + "loss": 0.4496, + "step": 19983 + }, + { + "epoch": 1.7029399233063485, + "grad_norm": 20.75, + "learning_rate": 1.3475808546070377e-07, + "loss": 0.8078, + "step": 19984 + }, + { + "epoch": 1.7030251384746484, + "grad_norm": 22.875, + "learning_rate": 1.3468220172700745e-07, + "loss": 0.6641, + "step": 19985 + }, + { + "epoch": 1.7031103536429484, + "grad_norm": 22.125, + "learning_rate": 1.3460633814822557e-07, + "loss": 0.9287, + "step": 19986 + }, + { + "epoch": 1.7031955688112483, + "grad_norm": 16.75, + "learning_rate": 1.3453049472572837e-07, + "loss": 0.2935, + "step": 19987 + }, + { + "epoch": 1.7032807839795483, + "grad_norm": 19.625, + "learning_rate": 1.3445467146088687e-07, + "loss": 0.6302, + "step": 19988 + }, + { + "epoch": 1.7033659991478483, + "grad_norm": 18.0, + "learning_rate": 1.3437886835507162e-07, + "loss": 0.6597, + "step": 19989 + }, + { + "epoch": 1.7034512143161482, + "grad_norm": 20.625, + "learning_rate": 1.3430308540965165e-07, + "loss": 1.0542, + "step": 19990 + }, + { + "epoch": 1.7035364294844482, + "grad_norm": 17.125, + "learning_rate": 1.3422732262599714e-07, + "loss": 0.7308, + "step": 19991 + }, + { + "epoch": 1.7036216446527481, + "grad_norm": 16.25, + "learning_rate": 1.341515800054767e-07, + "loss": 0.8165, + "step": 19992 + }, + { + "epoch": 1.703706859821048, + "grad_norm": 15.25, + "learning_rate": 1.3407585754945926e-07, + "loss": 0.4318, + "step": 19993 + }, + { + "epoch": 1.703792074989348, + "grad_norm": 13.5, + "learning_rate": 1.3400015525931343e-07, + "loss": 0.3637, + "step": 19994 + }, + { + "epoch": 1.703877290157648, + "grad_norm": 15.5, + "learning_rate": 1.3392447313640684e-07, + "loss": 0.6671, + "step": 19995 + }, + { + "epoch": 1.703962505325948, + "grad_norm": 38.0, + "learning_rate": 1.338488111821072e-07, + "loss": 0.6747, + "step": 19996 + }, + { + "epoch": 1.704047720494248, + "grad_norm": 38.75, + "learning_rate": 1.3377316939778202e-07, + "loss": 0.8417, + "step": 19997 + }, + { + "epoch": 1.7041329356625479, + "grad_norm": 16.0, + "learning_rate": 1.3369754778479835e-07, + "loss": 0.7618, + "step": 19998 + }, + { + "epoch": 1.7042181508308478, + "grad_norm": 13.875, + "learning_rate": 1.336219463445225e-07, + "loss": 0.3476, + "step": 19999 + }, + { + "epoch": 1.7043033659991478, + "grad_norm": 15.3125, + "learning_rate": 1.3354636507832045e-07, + "loss": 0.6015, + "step": 20000 + }, + { + "epoch": 1.7043885811674477, + "grad_norm": 16.0, + "learning_rate": 1.3347080398755833e-07, + "loss": 1.1305, + "step": 20001 + }, + { + "epoch": 1.7044737963357477, + "grad_norm": 11.5, + "learning_rate": 1.333952630736017e-07, + "loss": 0.4524, + "step": 20002 + }, + { + "epoch": 1.7045590115040476, + "grad_norm": 14.9375, + "learning_rate": 1.3331974233781534e-07, + "loss": 0.465, + "step": 20003 + }, + { + "epoch": 1.7046442266723476, + "grad_norm": 19.75, + "learning_rate": 1.3324424178156408e-07, + "loss": 1.0665, + "step": 20004 + }, + { + "epoch": 1.7047294418406476, + "grad_norm": 17.25, + "learning_rate": 1.331687614062123e-07, + "loss": 0.3595, + "step": 20005 + }, + { + "epoch": 1.7048146570089475, + "grad_norm": 22.0, + "learning_rate": 1.3309330121312441e-07, + "loss": 0.9665, + "step": 20006 + }, + { + "epoch": 1.7048998721772475, + "grad_norm": 14.5625, + "learning_rate": 1.3301786120366368e-07, + "loss": 0.6113, + "step": 20007 + }, + { + "epoch": 1.7049850873455474, + "grad_norm": 21.5, + "learning_rate": 1.3294244137919313e-07, + "loss": 0.7563, + "step": 20008 + }, + { + "epoch": 1.7050703025138474, + "grad_norm": 10.3125, + "learning_rate": 1.3286704174107602e-07, + "loss": 0.3996, + "step": 20009 + }, + { + "epoch": 1.7051555176821473, + "grad_norm": 17.5, + "learning_rate": 1.327916622906747e-07, + "loss": 0.3564, + "step": 20010 + }, + { + "epoch": 1.7052407328504473, + "grad_norm": 22.25, + "learning_rate": 1.3271630302935186e-07, + "loss": 1.038, + "step": 20011 + }, + { + "epoch": 1.7053259480187473, + "grad_norm": 18.5, + "learning_rate": 1.326409639584686e-07, + "loss": 0.9324, + "step": 20012 + }, + { + "epoch": 1.7054111631870472, + "grad_norm": 19.875, + "learning_rate": 1.3256564507938677e-07, + "loss": 0.5797, + "step": 20013 + }, + { + "epoch": 1.7054963783553472, + "grad_norm": 11.75, + "learning_rate": 1.3249034639346737e-07, + "loss": 0.3314, + "step": 20014 + }, + { + "epoch": 1.7055815935236471, + "grad_norm": 12.0625, + "learning_rate": 1.3241506790207136e-07, + "loss": 0.4046, + "step": 20015 + }, + { + "epoch": 1.705666808691947, + "grad_norm": 13.9375, + "learning_rate": 1.3233980960655894e-07, + "loss": 0.5557, + "step": 20016 + }, + { + "epoch": 1.705752023860247, + "grad_norm": 14.3125, + "learning_rate": 1.3226457150828987e-07, + "loss": 0.6009, + "step": 20017 + }, + { + "epoch": 1.705837239028547, + "grad_norm": 17.25, + "learning_rate": 1.32189353608624e-07, + "loss": 0.8705, + "step": 20018 + }, + { + "epoch": 1.705922454196847, + "grad_norm": 15.3125, + "learning_rate": 1.3211415590892064e-07, + "loss": 0.4909, + "step": 20019 + }, + { + "epoch": 1.706007669365147, + "grad_norm": 21.0, + "learning_rate": 1.3203897841053845e-07, + "loss": 0.6456, + "step": 20020 + }, + { + "epoch": 1.7060928845334469, + "grad_norm": 17.625, + "learning_rate": 1.3196382111483608e-07, + "loss": 0.7092, + "step": 20021 + }, + { + "epoch": 1.7061780997017468, + "grad_norm": 13.6875, + "learning_rate": 1.3188868402317174e-07, + "loss": 0.3371, + "step": 20022 + }, + { + "epoch": 1.7062633148700468, + "grad_norm": 17.625, + "learning_rate": 1.3181356713690348e-07, + "loss": 0.8493, + "step": 20023 + }, + { + "epoch": 1.7063485300383467, + "grad_norm": 16.125, + "learning_rate": 1.3173847045738847e-07, + "loss": 0.4968, + "step": 20024 + }, + { + "epoch": 1.7064337452066467, + "grad_norm": 14.9375, + "learning_rate": 1.3166339398598362e-07, + "loss": 0.4469, + "step": 20025 + }, + { + "epoch": 1.7065189603749467, + "grad_norm": 20.5, + "learning_rate": 1.3158833772404567e-07, + "loss": 1.081, + "step": 20026 + }, + { + "epoch": 1.7066041755432466, + "grad_norm": 15.0625, + "learning_rate": 1.3151330167293128e-07, + "loss": 0.3628, + "step": 20027 + }, + { + "epoch": 1.7066893907115466, + "grad_norm": 19.25, + "learning_rate": 1.3143828583399647e-07, + "loss": 0.7509, + "step": 20028 + }, + { + "epoch": 1.7067746058798465, + "grad_norm": 11.375, + "learning_rate": 1.3136329020859624e-07, + "loss": 0.3011, + "step": 20029 + }, + { + "epoch": 1.7068598210481465, + "grad_norm": 14.3125, + "learning_rate": 1.312883147980863e-07, + "loss": 0.6595, + "step": 20030 + }, + { + "epoch": 1.7069450362164464, + "grad_norm": 14.0, + "learning_rate": 1.3121335960382147e-07, + "loss": 0.6328, + "step": 20031 + }, + { + "epoch": 1.7070302513847464, + "grad_norm": 18.25, + "learning_rate": 1.3113842462715687e-07, + "loss": 0.4145, + "step": 20032 + }, + { + "epoch": 1.7071154665530464, + "grad_norm": 12.25, + "learning_rate": 1.3106350986944545e-07, + "loss": 0.3674, + "step": 20033 + }, + { + "epoch": 1.7072006817213463, + "grad_norm": 11.125, + "learning_rate": 1.309886153320415e-07, + "loss": 0.348, + "step": 20034 + }, + { + "epoch": 1.7072858968896463, + "grad_norm": 23.125, + "learning_rate": 1.3091374101629876e-07, + "loss": 0.7777, + "step": 20035 + }, + { + "epoch": 1.7073711120579462, + "grad_norm": 17.375, + "learning_rate": 1.3083888692357008e-07, + "loss": 0.5751, + "step": 20036 + }, + { + "epoch": 1.7074563272262462, + "grad_norm": 11.625, + "learning_rate": 1.3076405305520789e-07, + "loss": 0.2671, + "step": 20037 + }, + { + "epoch": 1.7075415423945461, + "grad_norm": 19.25, + "learning_rate": 1.3068923941256487e-07, + "loss": 1.0301, + "step": 20038 + }, + { + "epoch": 1.707626757562846, + "grad_norm": 15.75, + "learning_rate": 1.306144459969927e-07, + "loss": 0.5997, + "step": 20039 + }, + { + "epoch": 1.707711972731146, + "grad_norm": 17.25, + "learning_rate": 1.3053967280984347e-07, + "loss": 0.6087, + "step": 20040 + }, + { + "epoch": 1.707797187899446, + "grad_norm": 12.0625, + "learning_rate": 1.3046491985246796e-07, + "loss": 0.3466, + "step": 20041 + }, + { + "epoch": 1.707882403067746, + "grad_norm": 16.875, + "learning_rate": 1.3039018712621697e-07, + "loss": 0.4123, + "step": 20042 + }, + { + "epoch": 1.707967618236046, + "grad_norm": 16.125, + "learning_rate": 1.3031547463244116e-07, + "loss": 0.4959, + "step": 20043 + }, + { + "epoch": 1.7080528334043459, + "grad_norm": 15.5625, + "learning_rate": 1.3024078237249064e-07, + "loss": 0.5744, + "step": 20044 + }, + { + "epoch": 1.7081380485726458, + "grad_norm": 36.75, + "learning_rate": 1.3016611034771531e-07, + "loss": 0.6897, + "step": 20045 + }, + { + "epoch": 1.7082232637409458, + "grad_norm": 15.0, + "learning_rate": 1.300914585594644e-07, + "loss": 0.555, + "step": 20046 + }, + { + "epoch": 1.7083084789092458, + "grad_norm": 15.6875, + "learning_rate": 1.300168270090868e-07, + "loss": 0.5285, + "step": 20047 + }, + { + "epoch": 1.7083936940775457, + "grad_norm": 11.0, + "learning_rate": 1.2994221569793172e-07, + "loss": 0.2655, + "step": 20048 + }, + { + "epoch": 1.7084789092458457, + "grad_norm": 12.625, + "learning_rate": 1.298676246273471e-07, + "loss": 0.3018, + "step": 20049 + }, + { + "epoch": 1.7085641244141456, + "grad_norm": 15.625, + "learning_rate": 1.2979305379868066e-07, + "loss": 0.4826, + "step": 20050 + }, + { + "epoch": 1.7086493395824456, + "grad_norm": 13.4375, + "learning_rate": 1.2971850321328014e-07, + "loss": 0.7242, + "step": 20051 + }, + { + "epoch": 1.7087345547507455, + "grad_norm": 13.4375, + "learning_rate": 1.2964397287249282e-07, + "loss": 0.5074, + "step": 20052 + }, + { + "epoch": 1.7088197699190455, + "grad_norm": 12.8125, + "learning_rate": 1.295694627776657e-07, + "loss": 0.2861, + "step": 20053 + }, + { + "epoch": 1.7089049850873455, + "grad_norm": 12.5, + "learning_rate": 1.2949497293014492e-07, + "loss": 0.509, + "step": 20054 + }, + { + "epoch": 1.7089902002556454, + "grad_norm": 17.0, + "learning_rate": 1.2942050333127676e-07, + "loss": 0.8834, + "step": 20055 + }, + { + "epoch": 1.7090754154239454, + "grad_norm": 15.6875, + "learning_rate": 1.2934605398240708e-07, + "loss": 0.6043, + "step": 20056 + }, + { + "epoch": 1.7091606305922453, + "grad_norm": 16.75, + "learning_rate": 1.292716248848809e-07, + "loss": 0.4089, + "step": 20057 + }, + { + "epoch": 1.7092458457605453, + "grad_norm": 14.625, + "learning_rate": 1.291972160400437e-07, + "loss": 0.3871, + "step": 20058 + }, + { + "epoch": 1.7093310609288452, + "grad_norm": 20.625, + "learning_rate": 1.2912282744923953e-07, + "loss": 0.4721, + "step": 20059 + }, + { + "epoch": 1.7094162760971452, + "grad_norm": 16.875, + "learning_rate": 1.2904845911381304e-07, + "loss": 0.7479, + "step": 20060 + }, + { + "epoch": 1.7095014912654451, + "grad_norm": 16.75, + "learning_rate": 1.289741110351081e-07, + "loss": 0.9102, + "step": 20061 + }, + { + "epoch": 1.709586706433745, + "grad_norm": 14.1875, + "learning_rate": 1.2889978321446853e-07, + "loss": 0.8046, + "step": 20062 + }, + { + "epoch": 1.709671921602045, + "grad_norm": 12.5625, + "learning_rate": 1.28825475653237e-07, + "loss": 0.2122, + "step": 20063 + }, + { + "epoch": 1.709757136770345, + "grad_norm": 15.625, + "learning_rate": 1.2875118835275675e-07, + "loss": 0.921, + "step": 20064 + }, + { + "epoch": 1.709842351938645, + "grad_norm": 18.25, + "learning_rate": 1.2867692131436989e-07, + "loss": 0.5957, + "step": 20065 + }, + { + "epoch": 1.709927567106945, + "grad_norm": 10.375, + "learning_rate": 1.286026745394188e-07, + "loss": 0.2149, + "step": 20066 + }, + { + "epoch": 1.710012782275245, + "grad_norm": 16.25, + "learning_rate": 1.285284480292448e-07, + "loss": 0.5347, + "step": 20067 + }, + { + "epoch": 1.7100979974435448, + "grad_norm": 16.625, + "learning_rate": 1.2845424178518945e-07, + "loss": 0.2482, + "step": 20068 + }, + { + "epoch": 1.7101832126118448, + "grad_norm": 12.375, + "learning_rate": 1.2838005580859375e-07, + "loss": 0.4436, + "step": 20069 + }, + { + "epoch": 1.7102684277801448, + "grad_norm": 13.9375, + "learning_rate": 1.2830589010079857e-07, + "loss": 0.3525, + "step": 20070 + }, + { + "epoch": 1.7103536429484447, + "grad_norm": 14.9375, + "learning_rate": 1.2823174466314369e-07, + "loss": 0.6067, + "step": 20071 + }, + { + "epoch": 1.7104388581167447, + "grad_norm": 13.4375, + "learning_rate": 1.2815761949696942e-07, + "loss": 0.7096, + "step": 20072 + }, + { + "epoch": 1.7105240732850446, + "grad_norm": 15.5, + "learning_rate": 1.2808351460361482e-07, + "loss": 0.8955, + "step": 20073 + }, + { + "epoch": 1.7106092884533446, + "grad_norm": 17.375, + "learning_rate": 1.2800942998441924e-07, + "loss": 0.7419, + "step": 20074 + }, + { + "epoch": 1.7106945036216445, + "grad_norm": 11.625, + "learning_rate": 1.2793536564072186e-07, + "loss": 0.3938, + "step": 20075 + }, + { + "epoch": 1.7107797187899445, + "grad_norm": 18.125, + "learning_rate": 1.278613215738604e-07, + "loss": 0.7994, + "step": 20076 + }, + { + "epoch": 1.7108649339582445, + "grad_norm": 16.375, + "learning_rate": 1.2778729778517339e-07, + "loss": 0.3486, + "step": 20077 + }, + { + "epoch": 1.7109501491265444, + "grad_norm": 23.5, + "learning_rate": 1.2771329427599828e-07, + "loss": 0.8718, + "step": 20078 + }, + { + "epoch": 1.7110353642948444, + "grad_norm": 12.625, + "learning_rate": 1.2763931104767272e-07, + "loss": 0.259, + "step": 20079 + }, + { + "epoch": 1.7111205794631443, + "grad_norm": 25.25, + "learning_rate": 1.2756534810153348e-07, + "loss": 1.097, + "step": 20080 + }, + { + "epoch": 1.7112057946314443, + "grad_norm": 13.8125, + "learning_rate": 1.2749140543891675e-07, + "loss": 0.5999, + "step": 20081 + }, + { + "epoch": 1.7112910097997442, + "grad_norm": 21.125, + "learning_rate": 1.2741748306115923e-07, + "loss": 0.5775, + "step": 20082 + }, + { + "epoch": 1.7113762249680442, + "grad_norm": 17.5, + "learning_rate": 1.2734358096959678e-07, + "loss": 0.9197, + "step": 20083 + }, + { + "epoch": 1.7114614401363442, + "grad_norm": 17.125, + "learning_rate": 1.272696991655646e-07, + "loss": 0.4731, + "step": 20084 + }, + { + "epoch": 1.7115466553046441, + "grad_norm": 13.5625, + "learning_rate": 1.271958376503979e-07, + "loss": 0.6585, + "step": 20085 + }, + { + "epoch": 1.711631870472944, + "grad_norm": 13.0625, + "learning_rate": 1.271219964254315e-07, + "loss": 0.5869, + "step": 20086 + }, + { + "epoch": 1.711717085641244, + "grad_norm": 18.625, + "learning_rate": 1.27048175492e-07, + "loss": 0.7434, + "step": 20087 + }, + { + "epoch": 1.711802300809544, + "grad_norm": 17.25, + "learning_rate": 1.269743748514371e-07, + "loss": 0.4438, + "step": 20088 + }, + { + "epoch": 1.711887515977844, + "grad_norm": 30.5, + "learning_rate": 1.2690059450507637e-07, + "loss": 0.9487, + "step": 20089 + }, + { + "epoch": 1.711972731146144, + "grad_norm": 15.0, + "learning_rate": 1.2682683445425128e-07, + "loss": 0.2973, + "step": 20090 + }, + { + "epoch": 1.7120579463144439, + "grad_norm": 15.0, + "learning_rate": 1.267530947002947e-07, + "loss": 0.6495, + "step": 20091 + }, + { + "epoch": 1.7121431614827438, + "grad_norm": 18.625, + "learning_rate": 1.2667937524453945e-07, + "loss": 0.9474, + "step": 20092 + }, + { + "epoch": 1.7122283766510438, + "grad_norm": 14.1875, + "learning_rate": 1.266056760883172e-07, + "loss": 0.4801, + "step": 20093 + }, + { + "epoch": 1.7123135918193437, + "grad_norm": 12.4375, + "learning_rate": 1.265319972329601e-07, + "loss": 0.3763, + "step": 20094 + }, + { + "epoch": 1.7123988069876437, + "grad_norm": 14.9375, + "learning_rate": 1.2645833867979975e-07, + "loss": 0.8136, + "step": 20095 + }, + { + "epoch": 1.7124840221559436, + "grad_norm": 19.5, + "learning_rate": 1.2638470043016674e-07, + "loss": 0.8341, + "step": 20096 + }, + { + "epoch": 1.7125692373242436, + "grad_norm": 15.625, + "learning_rate": 1.2631108248539247e-07, + "loss": 0.6708, + "step": 20097 + }, + { + "epoch": 1.7126544524925436, + "grad_norm": 16.625, + "learning_rate": 1.2623748484680647e-07, + "loss": 0.6029, + "step": 20098 + }, + { + "epoch": 1.7127396676608435, + "grad_norm": 23.375, + "learning_rate": 1.2616390751573935e-07, + "loss": 0.6127, + "step": 20099 + }, + { + "epoch": 1.7128248828291435, + "grad_norm": 13.3125, + "learning_rate": 1.2609035049352056e-07, + "loss": 0.452, + "step": 20100 + }, + { + "epoch": 1.7129100979974434, + "grad_norm": 16.0, + "learning_rate": 1.260168137814792e-07, + "loss": 0.6542, + "step": 20101 + }, + { + "epoch": 1.7129953131657434, + "grad_norm": 13.6875, + "learning_rate": 1.2594329738094422e-07, + "loss": 0.6103, + "step": 20102 + }, + { + "epoch": 1.7130805283340433, + "grad_norm": 14.875, + "learning_rate": 1.2586980129324426e-07, + "loss": 0.4936, + "step": 20103 + }, + { + "epoch": 1.7131657435023433, + "grad_norm": 24.375, + "learning_rate": 1.2579632551970756e-07, + "loss": 0.5636, + "step": 20104 + }, + { + "epoch": 1.7132509586706433, + "grad_norm": 14.9375, + "learning_rate": 1.257228700616618e-07, + "loss": 0.5469, + "step": 20105 + }, + { + "epoch": 1.7133361738389432, + "grad_norm": 13.375, + "learning_rate": 1.2564943492043402e-07, + "loss": 0.6071, + "step": 20106 + }, + { + "epoch": 1.7134213890072432, + "grad_norm": 22.125, + "learning_rate": 1.255760200973516e-07, + "loss": 0.7089, + "step": 20107 + }, + { + "epoch": 1.7135066041755431, + "grad_norm": 15.3125, + "learning_rate": 1.255026255937411e-07, + "loss": 0.6959, + "step": 20108 + }, + { + "epoch": 1.713591819343843, + "grad_norm": 15.5, + "learning_rate": 1.2542925141092926e-07, + "loss": 0.4797, + "step": 20109 + }, + { + "epoch": 1.713677034512143, + "grad_norm": 18.0, + "learning_rate": 1.2535589755024141e-07, + "loss": 1.0135, + "step": 20110 + }, + { + "epoch": 1.713762249680443, + "grad_norm": 14.5625, + "learning_rate": 1.2528256401300342e-07, + "loss": 0.5365, + "step": 20111 + }, + { + "epoch": 1.713847464848743, + "grad_norm": 27.0, + "learning_rate": 1.2520925080054076e-07, + "loss": 0.8062, + "step": 20112 + }, + { + "epoch": 1.713932680017043, + "grad_norm": 14.0625, + "learning_rate": 1.2513595791417796e-07, + "loss": 0.5805, + "step": 20113 + }, + { + "epoch": 1.7140178951853429, + "grad_norm": 12.9375, + "learning_rate": 1.2506268535523933e-07, + "loss": 0.5623, + "step": 20114 + }, + { + "epoch": 1.7141031103536428, + "grad_norm": 18.875, + "learning_rate": 1.2498943312504924e-07, + "loss": 0.6646, + "step": 20115 + }, + { + "epoch": 1.7141883255219428, + "grad_norm": 12.5, + "learning_rate": 1.2491620122493137e-07, + "loss": 0.5012, + "step": 20116 + }, + { + "epoch": 1.7142735406902427, + "grad_norm": 16.125, + "learning_rate": 1.2484298965620922e-07, + "loss": 0.5747, + "step": 20117 + }, + { + "epoch": 1.7143587558585427, + "grad_norm": 15.1875, + "learning_rate": 1.247697984202055e-07, + "loss": 0.3942, + "step": 20118 + }, + { + "epoch": 1.7144439710268427, + "grad_norm": 19.25, + "learning_rate": 1.2469662751824305e-07, + "loss": 0.8065, + "step": 20119 + }, + { + "epoch": 1.7145291861951426, + "grad_norm": 16.375, + "learning_rate": 1.246234769516441e-07, + "loss": 0.5845, + "step": 20120 + }, + { + "epoch": 1.7146144013634426, + "grad_norm": 14.875, + "learning_rate": 1.2455034672173086e-07, + "loss": 0.5019, + "step": 20121 + }, + { + "epoch": 1.7146996165317425, + "grad_norm": 13.4375, + "learning_rate": 1.2447723682982457e-07, + "loss": 0.4346, + "step": 20122 + }, + { + "epoch": 1.7147848317000425, + "grad_norm": 19.5, + "learning_rate": 1.2440414727724615e-07, + "loss": 0.7161, + "step": 20123 + }, + { + "epoch": 1.7148700468683424, + "grad_norm": 14.8125, + "learning_rate": 1.243310780653166e-07, + "loss": 0.5796, + "step": 20124 + }, + { + "epoch": 1.7149552620366424, + "grad_norm": 19.25, + "learning_rate": 1.2425802919535671e-07, + "loss": 0.6949, + "step": 20125 + }, + { + "epoch": 1.7150404772049423, + "grad_norm": 16.5, + "learning_rate": 1.2418500066868608e-07, + "loss": 0.8475, + "step": 20126 + }, + { + "epoch": 1.7151256923732423, + "grad_norm": 18.625, + "learning_rate": 1.241119924866245e-07, + "loss": 0.8156, + "step": 20127 + }, + { + "epoch": 1.7152109075415423, + "grad_norm": 12.25, + "learning_rate": 1.2403900465049148e-07, + "loss": 0.3294, + "step": 20128 + }, + { + "epoch": 1.7152961227098422, + "grad_norm": 16.0, + "learning_rate": 1.2396603716160596e-07, + "loss": 0.391, + "step": 20129 + }, + { + "epoch": 1.7153813378781422, + "grad_norm": 21.75, + "learning_rate": 1.2389309002128662e-07, + "loss": 0.743, + "step": 20130 + }, + { + "epoch": 1.7154665530464421, + "grad_norm": 21.125, + "learning_rate": 1.2382016323085125e-07, + "loss": 0.5649, + "step": 20131 + }, + { + "epoch": 1.715551768214742, + "grad_norm": 15.125, + "learning_rate": 1.23747256791618e-07, + "loss": 0.3936, + "step": 20132 + }, + { + "epoch": 1.715636983383042, + "grad_norm": 15.6875, + "learning_rate": 1.2367437070490428e-07, + "loss": 0.481, + "step": 20133 + }, + { + "epoch": 1.715722198551342, + "grad_norm": 18.125, + "learning_rate": 1.2360150497202753e-07, + "loss": 0.5826, + "step": 20134 + }, + { + "epoch": 1.715807413719642, + "grad_norm": 11.9375, + "learning_rate": 1.2352865959430415e-07, + "loss": 0.3104, + "step": 20135 + }, + { + "epoch": 1.715892628887942, + "grad_norm": 15.6875, + "learning_rate": 1.234558345730505e-07, + "loss": 0.6593, + "step": 20136 + }, + { + "epoch": 1.7159778440562419, + "grad_norm": 26.0, + "learning_rate": 1.2338302990958311e-07, + "loss": 0.4735, + "step": 20137 + }, + { + "epoch": 1.7160630592245418, + "grad_norm": 13.875, + "learning_rate": 1.2331024560521695e-07, + "loss": 0.4946, + "step": 20138 + }, + { + "epoch": 1.7161482743928418, + "grad_norm": 17.5, + "learning_rate": 1.232374816612679e-07, + "loss": 0.4564, + "step": 20139 + }, + { + "epoch": 1.7162334895611417, + "grad_norm": 14.875, + "learning_rate": 1.231647380790503e-07, + "loss": 0.3257, + "step": 20140 + }, + { + "epoch": 1.7163187047294417, + "grad_norm": 19.875, + "learning_rate": 1.230920148598791e-07, + "loss": 1.0537, + "step": 20141 + }, + { + "epoch": 1.7164039198977417, + "grad_norm": 11.9375, + "learning_rate": 1.2301931200506864e-07, + "loss": 0.4197, + "step": 20142 + }, + { + "epoch": 1.7164891350660416, + "grad_norm": 18.375, + "learning_rate": 1.229466295159322e-07, + "loss": 0.8462, + "step": 20143 + }, + { + "epoch": 1.7165743502343416, + "grad_norm": 20.875, + "learning_rate": 1.2287396739378343e-07, + "loss": 0.7014, + "step": 20144 + }, + { + "epoch": 1.7166595654026415, + "grad_norm": 15.375, + "learning_rate": 1.2280132563993575e-07, + "loss": 0.6376, + "step": 20145 + }, + { + "epoch": 1.7167447805709415, + "grad_norm": 20.125, + "learning_rate": 1.227287042557014e-07, + "loss": 0.6549, + "step": 20146 + }, + { + "epoch": 1.7168299957392414, + "grad_norm": 13.125, + "learning_rate": 1.2265610324239313e-07, + "loss": 0.4254, + "step": 20147 + }, + { + "epoch": 1.7169152109075414, + "grad_norm": 11.5625, + "learning_rate": 1.225835226013225e-07, + "loss": 0.2966, + "step": 20148 + }, + { + "epoch": 1.7170004260758414, + "grad_norm": 19.25, + "learning_rate": 1.225109623338014e-07, + "loss": 0.9581, + "step": 20149 + }, + { + "epoch": 1.7170856412441413, + "grad_norm": 12.375, + "learning_rate": 1.2243842244114082e-07, + "loss": 0.3737, + "step": 20150 + }, + { + "epoch": 1.7171708564124413, + "grad_norm": 19.375, + "learning_rate": 1.2236590292465212e-07, + "loss": 0.7292, + "step": 20151 + }, + { + "epoch": 1.7172560715807412, + "grad_norm": 14.6875, + "learning_rate": 1.2229340378564532e-07, + "loss": 0.5515, + "step": 20152 + }, + { + "epoch": 1.7173412867490412, + "grad_norm": 16.5, + "learning_rate": 1.2222092502543083e-07, + "loss": 0.4632, + "step": 20153 + }, + { + "epoch": 1.7174265019173411, + "grad_norm": 16.25, + "learning_rate": 1.2214846664531807e-07, + "loss": 0.6727, + "step": 20154 + }, + { + "epoch": 1.717511717085641, + "grad_norm": 18.0, + "learning_rate": 1.2207602864661676e-07, + "loss": 0.8742, + "step": 20155 + }, + { + "epoch": 1.717596932253941, + "grad_norm": 13.375, + "learning_rate": 1.2200361103063594e-07, + "loss": 0.3635, + "step": 20156 + }, + { + "epoch": 1.717682147422241, + "grad_norm": 18.125, + "learning_rate": 1.219312137986839e-07, + "loss": 0.6539, + "step": 20157 + }, + { + "epoch": 1.717767362590541, + "grad_norm": 12.0, + "learning_rate": 1.218588369520693e-07, + "loss": 0.3792, + "step": 20158 + }, + { + "epoch": 1.717852577758841, + "grad_norm": 22.125, + "learning_rate": 1.2178648049210012e-07, + "loss": 0.9524, + "step": 20159 + }, + { + "epoch": 1.7179377929271409, + "grad_norm": 25.0, + "learning_rate": 1.2171414442008352e-07, + "loss": 0.828, + "step": 20160 + }, + { + "epoch": 1.7180230080954408, + "grad_norm": 18.625, + "learning_rate": 1.2164182873732694e-07, + "loss": 0.7211, + "step": 20161 + }, + { + "epoch": 1.7181082232637408, + "grad_norm": 16.25, + "learning_rate": 1.2156953344513736e-07, + "loss": 0.7031, + "step": 20162 + }, + { + "epoch": 1.7181934384320408, + "grad_norm": 18.0, + "learning_rate": 1.2149725854482083e-07, + "loss": 0.7063, + "step": 20163 + }, + { + "epoch": 1.7182786536003407, + "grad_norm": 12.0625, + "learning_rate": 1.214250040376838e-07, + "loss": 0.3333, + "step": 20164 + }, + { + "epoch": 1.7183638687686407, + "grad_norm": 14.125, + "learning_rate": 1.213527699250318e-07, + "loss": 0.6548, + "step": 20165 + }, + { + "epoch": 1.7184490839369406, + "grad_norm": 13.875, + "learning_rate": 1.2128055620816996e-07, + "loss": 0.4636, + "step": 20166 + }, + { + "epoch": 1.7185342991052406, + "grad_norm": 16.5, + "learning_rate": 1.2120836288840352e-07, + "loss": 0.6505, + "step": 20167 + }, + { + "epoch": 1.7186195142735405, + "grad_norm": 17.5, + "learning_rate": 1.211361899670374e-07, + "loss": 0.9041, + "step": 20168 + }, + { + "epoch": 1.7187047294418405, + "grad_norm": 17.25, + "learning_rate": 1.2106403744537517e-07, + "loss": 0.6368, + "step": 20169 + }, + { + "epoch": 1.7187899446101405, + "grad_norm": 19.125, + "learning_rate": 1.209919053247212e-07, + "loss": 0.4667, + "step": 20170 + }, + { + "epoch": 1.7188751597784404, + "grad_norm": 21.125, + "learning_rate": 1.209197936063787e-07, + "loss": 0.4623, + "step": 20171 + }, + { + "epoch": 1.7189603749467404, + "grad_norm": 12.8125, + "learning_rate": 1.20847702291651e-07, + "loss": 0.5971, + "step": 20172 + }, + { + "epoch": 1.7190455901150403, + "grad_norm": 24.875, + "learning_rate": 1.207756313818406e-07, + "loss": 1.0578, + "step": 20173 + }, + { + "epoch": 1.7191308052833403, + "grad_norm": 17.0, + "learning_rate": 1.2070358087825e-07, + "loss": 0.6266, + "step": 20174 + }, + { + "epoch": 1.7192160204516402, + "grad_norm": 19.625, + "learning_rate": 1.2063155078218148e-07, + "loss": 0.5084, + "step": 20175 + }, + { + "epoch": 1.7193012356199402, + "grad_norm": 24.125, + "learning_rate": 1.2055954109493658e-07, + "loss": 0.5791, + "step": 20176 + }, + { + "epoch": 1.7193864507882402, + "grad_norm": 11.9375, + "learning_rate": 1.2048755181781627e-07, + "loss": 0.4565, + "step": 20177 + }, + { + "epoch": 1.71947166595654, + "grad_norm": 15.0, + "learning_rate": 1.2041558295212197e-07, + "loss": 0.5942, + "step": 20178 + }, + { + "epoch": 1.71955688112484, + "grad_norm": 13.625, + "learning_rate": 1.2034363449915377e-07, + "loss": 0.4087, + "step": 20179 + }, + { + "epoch": 1.71964209629314, + "grad_norm": 15.0625, + "learning_rate": 1.2027170646021214e-07, + "loss": 0.7813, + "step": 20180 + }, + { + "epoch": 1.71972731146144, + "grad_norm": 16.5, + "learning_rate": 1.2019979883659688e-07, + "loss": 0.5385, + "step": 20181 + }, + { + "epoch": 1.71981252662974, + "grad_norm": 11.3125, + "learning_rate": 1.2012791162960724e-07, + "loss": 0.3076, + "step": 20182 + }, + { + "epoch": 1.71989774179804, + "grad_norm": 11.0, + "learning_rate": 1.2005604484054232e-07, + "loss": 0.3014, + "step": 20183 + }, + { + "epoch": 1.7199829569663398, + "grad_norm": 22.0, + "learning_rate": 1.1998419847070092e-07, + "loss": 1.0147, + "step": 20184 + }, + { + "epoch": 1.7200681721346398, + "grad_norm": 18.5, + "learning_rate": 1.199123725213816e-07, + "loss": 1.0917, + "step": 20185 + }, + { + "epoch": 1.7201533873029398, + "grad_norm": 16.25, + "learning_rate": 1.198405669938822e-07, + "loss": 0.4462, + "step": 20186 + }, + { + "epoch": 1.7202386024712397, + "grad_norm": 20.5, + "learning_rate": 1.1976878188949992e-07, + "loss": 0.755, + "step": 20187 + }, + { + "epoch": 1.7203238176395397, + "grad_norm": 14.5625, + "learning_rate": 1.1969701720953229e-07, + "loss": 0.5837, + "step": 20188 + }, + { + "epoch": 1.7204090328078396, + "grad_norm": 11.9375, + "learning_rate": 1.1962527295527631e-07, + "loss": 0.2258, + "step": 20189 + }, + { + "epoch": 1.7204942479761396, + "grad_norm": 12.4375, + "learning_rate": 1.1955354912802823e-07, + "loss": 0.5737, + "step": 20190 + }, + { + "epoch": 1.7205794631444395, + "grad_norm": 14.5, + "learning_rate": 1.1948184572908417e-07, + "loss": 0.4786, + "step": 20191 + }, + { + "epoch": 1.7206646783127395, + "grad_norm": 16.875, + "learning_rate": 1.1941016275974007e-07, + "loss": 0.5263, + "step": 20192 + }, + { + "epoch": 1.7207498934810395, + "grad_norm": 14.75, + "learning_rate": 1.1933850022129143e-07, + "loss": 0.5117, + "step": 20193 + }, + { + "epoch": 1.7208351086493394, + "grad_norm": 13.6875, + "learning_rate": 1.1926685811503302e-07, + "loss": 0.5417, + "step": 20194 + }, + { + "epoch": 1.7209203238176394, + "grad_norm": 19.625, + "learning_rate": 1.1919523644225939e-07, + "loss": 0.7597, + "step": 20195 + }, + { + "epoch": 1.7210055389859393, + "grad_norm": 21.5, + "learning_rate": 1.1912363520426476e-07, + "loss": 0.6601, + "step": 20196 + }, + { + "epoch": 1.7210907541542393, + "grad_norm": 21.125, + "learning_rate": 1.1905205440234339e-07, + "loss": 0.4605, + "step": 20197 + }, + { + "epoch": 1.7211759693225395, + "grad_norm": 13.6875, + "learning_rate": 1.189804940377888e-07, + "loss": 0.4952, + "step": 20198 + }, + { + "epoch": 1.7212611844908394, + "grad_norm": 12.875, + "learning_rate": 1.1890895411189387e-07, + "loss": 0.4862, + "step": 20199 + }, + { + "epoch": 1.7213463996591394, + "grad_norm": 14.8125, + "learning_rate": 1.1883743462595147e-07, + "loss": 0.8428, + "step": 20200 + }, + { + "epoch": 1.7214316148274393, + "grad_norm": 16.375, + "learning_rate": 1.1876593558125412e-07, + "loss": 0.5378, + "step": 20201 + }, + { + "epoch": 1.7215168299957393, + "grad_norm": 20.5, + "learning_rate": 1.1869445697909432e-07, + "loss": 0.5653, + "step": 20202 + }, + { + "epoch": 1.7216020451640393, + "grad_norm": 15.9375, + "learning_rate": 1.1862299882076281e-07, + "loss": 0.555, + "step": 20203 + }, + { + "epoch": 1.7216872603323392, + "grad_norm": 14.0625, + "learning_rate": 1.1855156110755148e-07, + "loss": 0.2329, + "step": 20204 + }, + { + "epoch": 1.7217724755006392, + "grad_norm": 13.375, + "learning_rate": 1.184801438407511e-07, + "loss": 0.4022, + "step": 20205 + }, + { + "epoch": 1.7218576906689391, + "grad_norm": 15.625, + "learning_rate": 1.1840874702165259e-07, + "loss": 0.6301, + "step": 20206 + }, + { + "epoch": 1.721942905837239, + "grad_norm": 15.625, + "learning_rate": 1.1833737065154574e-07, + "loss": 0.4992, + "step": 20207 + }, + { + "epoch": 1.722028121005539, + "grad_norm": 12.375, + "learning_rate": 1.1826601473172064e-07, + "loss": 0.6199, + "step": 20208 + }, + { + "epoch": 1.722113336173839, + "grad_norm": 16.125, + "learning_rate": 1.1819467926346654e-07, + "loss": 0.6434, + "step": 20209 + }, + { + "epoch": 1.722198551342139, + "grad_norm": 18.0, + "learning_rate": 1.1812336424807308e-07, + "loss": 0.6117, + "step": 20210 + }, + { + "epoch": 1.722283766510439, + "grad_norm": 16.375, + "learning_rate": 1.1805206968682855e-07, + "loss": 0.4356, + "step": 20211 + }, + { + "epoch": 1.7223689816787389, + "grad_norm": 12.25, + "learning_rate": 1.179807955810211e-07, + "loss": 0.496, + "step": 20212 + }, + { + "epoch": 1.7224541968470388, + "grad_norm": 17.0, + "learning_rate": 1.1790954193193912e-07, + "loss": 0.3543, + "step": 20213 + }, + { + "epoch": 1.7225394120153388, + "grad_norm": 19.875, + "learning_rate": 1.1783830874087008e-07, + "loss": 0.806, + "step": 20214 + }, + { + "epoch": 1.7226246271836387, + "grad_norm": 15.8125, + "learning_rate": 1.1776709600910152e-07, + "loss": 0.8668, + "step": 20215 + }, + { + "epoch": 1.7227098423519387, + "grad_norm": 19.25, + "learning_rate": 1.1769590373791981e-07, + "loss": 0.7513, + "step": 20216 + }, + { + "epoch": 1.7227950575202386, + "grad_norm": 18.75, + "learning_rate": 1.1762473192861168e-07, + "loss": 0.6213, + "step": 20217 + }, + { + "epoch": 1.7228802726885386, + "grad_norm": 16.75, + "learning_rate": 1.1755358058246375e-07, + "loss": 0.6169, + "step": 20218 + }, + { + "epoch": 1.7229654878568386, + "grad_norm": 15.875, + "learning_rate": 1.1748244970076123e-07, + "loss": 0.801, + "step": 20219 + }, + { + "epoch": 1.7230507030251385, + "grad_norm": 16.125, + "learning_rate": 1.1741133928478949e-07, + "loss": 0.6383, + "step": 20220 + }, + { + "epoch": 1.7231359181934385, + "grad_norm": 18.875, + "learning_rate": 1.1734024933583373e-07, + "loss": 0.6675, + "step": 20221 + }, + { + "epoch": 1.7232211333617384, + "grad_norm": 16.0, + "learning_rate": 1.1726917985517865e-07, + "loss": 0.4613, + "step": 20222 + }, + { + "epoch": 1.7233063485300384, + "grad_norm": 14.9375, + "learning_rate": 1.1719813084410877e-07, + "loss": 0.5021, + "step": 20223 + }, + { + "epoch": 1.7233915636983383, + "grad_norm": 14.6875, + "learning_rate": 1.1712710230390764e-07, + "loss": 0.6141, + "step": 20224 + }, + { + "epoch": 1.7234767788666383, + "grad_norm": 16.375, + "learning_rate": 1.1705609423585881e-07, + "loss": 0.6159, + "step": 20225 + }, + { + "epoch": 1.7235619940349383, + "grad_norm": 15.0625, + "learning_rate": 1.1698510664124599e-07, + "loss": 0.3312, + "step": 20226 + }, + { + "epoch": 1.7236472092032382, + "grad_norm": 14.6875, + "learning_rate": 1.1691413952135133e-07, + "loss": 0.6655, + "step": 20227 + }, + { + "epoch": 1.7237324243715382, + "grad_norm": 20.0, + "learning_rate": 1.1684319287745785e-07, + "loss": 0.9766, + "step": 20228 + }, + { + "epoch": 1.7238176395398381, + "grad_norm": 22.375, + "learning_rate": 1.1677226671084716e-07, + "loss": 0.6145, + "step": 20229 + }, + { + "epoch": 1.723902854708138, + "grad_norm": 23.375, + "learning_rate": 1.16701361022801e-07, + "loss": 0.9781, + "step": 20230 + }, + { + "epoch": 1.723988069876438, + "grad_norm": 14.25, + "learning_rate": 1.1663047581460099e-07, + "loss": 0.6627, + "step": 20231 + }, + { + "epoch": 1.724073285044738, + "grad_norm": 26.75, + "learning_rate": 1.165596110875282e-07, + "loss": 0.7677, + "step": 20232 + }, + { + "epoch": 1.724158500213038, + "grad_norm": 13.0, + "learning_rate": 1.164887668428627e-07, + "loss": 0.4149, + "step": 20233 + }, + { + "epoch": 1.724243715381338, + "grad_norm": 14.25, + "learning_rate": 1.16417943081885e-07, + "loss": 0.5992, + "step": 20234 + }, + { + "epoch": 1.7243289305496379, + "grad_norm": 19.25, + "learning_rate": 1.1634713980587534e-07, + "loss": 0.9113, + "step": 20235 + }, + { + "epoch": 1.7244141457179378, + "grad_norm": 18.0, + "learning_rate": 1.1627635701611282e-07, + "loss": 0.825, + "step": 20236 + }, + { + "epoch": 1.7244993608862378, + "grad_norm": 16.5, + "learning_rate": 1.1620559471387627e-07, + "loss": 0.6716, + "step": 20237 + }, + { + "epoch": 1.7245845760545377, + "grad_norm": 16.25, + "learning_rate": 1.1613485290044468e-07, + "loss": 0.7599, + "step": 20238 + }, + { + "epoch": 1.7246697912228377, + "grad_norm": 18.5, + "learning_rate": 1.1606413157709662e-07, + "loss": 1.1758, + "step": 20239 + }, + { + "epoch": 1.7247550063911377, + "grad_norm": 12.4375, + "learning_rate": 1.1599343074511008e-07, + "loss": 0.4573, + "step": 20240 + }, + { + "epoch": 1.7248402215594376, + "grad_norm": 16.75, + "learning_rate": 1.1592275040576251e-07, + "loss": 0.7164, + "step": 20241 + }, + { + "epoch": 1.7249254367277376, + "grad_norm": 15.125, + "learning_rate": 1.1585209056033108e-07, + "loss": 0.6129, + "step": 20242 + }, + { + "epoch": 1.7250106518960375, + "grad_norm": 14.0, + "learning_rate": 1.1578145121009312e-07, + "loss": 0.4169, + "step": 20243 + }, + { + "epoch": 1.7250958670643375, + "grad_norm": 16.75, + "learning_rate": 1.1571083235632468e-07, + "loss": 0.4133, + "step": 20244 + }, + { + "epoch": 1.7251810822326374, + "grad_norm": 13.9375, + "learning_rate": 1.1564023400030239e-07, + "loss": 0.6, + "step": 20245 + }, + { + "epoch": 1.7252662974009374, + "grad_norm": 16.0, + "learning_rate": 1.1556965614330146e-07, + "loss": 0.7857, + "step": 20246 + }, + { + "epoch": 1.7253515125692374, + "grad_norm": 11.8125, + "learning_rate": 1.1549909878659754e-07, + "loss": 0.3227, + "step": 20247 + }, + { + "epoch": 1.7254367277375373, + "grad_norm": 27.875, + "learning_rate": 1.1542856193146587e-07, + "loss": 0.7216, + "step": 20248 + }, + { + "epoch": 1.7255219429058373, + "grad_norm": 27.25, + "learning_rate": 1.1535804557918112e-07, + "loss": 0.8416, + "step": 20249 + }, + { + "epoch": 1.7256071580741372, + "grad_norm": 11.3125, + "learning_rate": 1.1528754973101728e-07, + "loss": 0.462, + "step": 20250 + }, + { + "epoch": 1.7256923732424372, + "grad_norm": 18.625, + "learning_rate": 1.1521707438824872e-07, + "loss": 0.678, + "step": 20251 + }, + { + "epoch": 1.7257775884107371, + "grad_norm": 19.875, + "learning_rate": 1.1514661955214834e-07, + "loss": 0.779, + "step": 20252 + }, + { + "epoch": 1.725862803579037, + "grad_norm": 13.8125, + "learning_rate": 1.1507618522399011e-07, + "loss": 0.4549, + "step": 20253 + }, + { + "epoch": 1.725948018747337, + "grad_norm": 10.0, + "learning_rate": 1.1500577140504621e-07, + "loss": 0.2901, + "step": 20254 + }, + { + "epoch": 1.726033233915637, + "grad_norm": 17.125, + "learning_rate": 1.1493537809658936e-07, + "loss": 0.6195, + "step": 20255 + }, + { + "epoch": 1.726118449083937, + "grad_norm": 18.125, + "learning_rate": 1.1486500529989161e-07, + "loss": 0.8804, + "step": 20256 + }, + { + "epoch": 1.726203664252237, + "grad_norm": 17.25, + "learning_rate": 1.1479465301622486e-07, + "loss": 0.709, + "step": 20257 + }, + { + "epoch": 1.7262888794205369, + "grad_norm": 18.25, + "learning_rate": 1.1472432124686017e-07, + "loss": 0.6032, + "step": 20258 + }, + { + "epoch": 1.7263740945888368, + "grad_norm": 20.5, + "learning_rate": 1.1465400999306875e-07, + "loss": 0.8568, + "step": 20259 + }, + { + "epoch": 1.7264593097571368, + "grad_norm": 18.625, + "learning_rate": 1.1458371925612083e-07, + "loss": 0.5294, + "step": 20260 + }, + { + "epoch": 1.7265445249254368, + "grad_norm": 18.875, + "learning_rate": 1.1451344903728679e-07, + "loss": 0.6295, + "step": 20261 + }, + { + "epoch": 1.7266297400937367, + "grad_norm": 17.0, + "learning_rate": 1.1444319933783685e-07, + "loss": 0.4918, + "step": 20262 + }, + { + "epoch": 1.7267149552620367, + "grad_norm": 15.0, + "learning_rate": 1.1437297015903987e-07, + "loss": 0.6205, + "step": 20263 + }, + { + "epoch": 1.7268001704303366, + "grad_norm": 16.25, + "learning_rate": 1.1430276150216538e-07, + "loss": 0.5765, + "step": 20264 + }, + { + "epoch": 1.7268853855986366, + "grad_norm": 18.125, + "learning_rate": 1.1423257336848223e-07, + "loss": 0.5456, + "step": 20265 + }, + { + "epoch": 1.7269706007669365, + "grad_norm": 14.5, + "learning_rate": 1.141624057592583e-07, + "loss": 0.553, + "step": 20266 + }, + { + "epoch": 1.7270558159352365, + "grad_norm": 12.875, + "learning_rate": 1.1409225867576215e-07, + "loss": 0.5272, + "step": 20267 + }, + { + "epoch": 1.7271410311035365, + "grad_norm": 15.125, + "learning_rate": 1.1402213211926083e-07, + "loss": 0.5327, + "step": 20268 + }, + { + "epoch": 1.7272262462718364, + "grad_norm": 17.625, + "learning_rate": 1.1395202609102194e-07, + "loss": 0.6694, + "step": 20269 + }, + { + "epoch": 1.7273114614401364, + "grad_norm": 12.1875, + "learning_rate": 1.1388194059231264e-07, + "loss": 0.387, + "step": 20270 + }, + { + "epoch": 1.7273966766084363, + "grad_norm": 15.75, + "learning_rate": 1.1381187562439874e-07, + "loss": 0.6186, + "step": 20271 + }, + { + "epoch": 1.7274818917767363, + "grad_norm": 12.125, + "learning_rate": 1.1374183118854687e-07, + "loss": 0.3256, + "step": 20272 + }, + { + "epoch": 1.7275671069450362, + "grad_norm": 14.5, + "learning_rate": 1.1367180728602267e-07, + "loss": 0.6508, + "step": 20273 + }, + { + "epoch": 1.7276523221133362, + "grad_norm": 17.5, + "learning_rate": 1.1360180391809181e-07, + "loss": 0.7319, + "step": 20274 + }, + { + "epoch": 1.7277375372816361, + "grad_norm": 22.625, + "learning_rate": 1.135318210860191e-07, + "loss": 0.7254, + "step": 20275 + }, + { + "epoch": 1.727822752449936, + "grad_norm": 15.0625, + "learning_rate": 1.1346185879106908e-07, + "loss": 0.6313, + "step": 20276 + }, + { + "epoch": 1.727907967618236, + "grad_norm": 15.5625, + "learning_rate": 1.1339191703450602e-07, + "loss": 0.5576, + "step": 20277 + }, + { + "epoch": 1.727993182786536, + "grad_norm": 14.6875, + "learning_rate": 1.1332199581759393e-07, + "loss": 0.7117, + "step": 20278 + }, + { + "epoch": 1.728078397954836, + "grad_norm": 16.5, + "learning_rate": 1.1325209514159677e-07, + "loss": 0.6357, + "step": 20279 + }, + { + "epoch": 1.728163613123136, + "grad_norm": 14.625, + "learning_rate": 1.1318221500777702e-07, + "loss": 0.7452, + "step": 20280 + }, + { + "epoch": 1.728248828291436, + "grad_norm": 12.125, + "learning_rate": 1.1311235541739768e-07, + "loss": 0.7665, + "step": 20281 + }, + { + "epoch": 1.7283340434597358, + "grad_norm": 15.5, + "learning_rate": 1.1304251637172165e-07, + "loss": 0.8261, + "step": 20282 + }, + { + "epoch": 1.7284192586280358, + "grad_norm": 17.25, + "learning_rate": 1.1297269787201068e-07, + "loss": 0.5681, + "step": 20283 + }, + { + "epoch": 1.7285044737963358, + "grad_norm": 20.625, + "learning_rate": 1.1290289991952627e-07, + "loss": 0.7, + "step": 20284 + }, + { + "epoch": 1.7285896889646357, + "grad_norm": 11.25, + "learning_rate": 1.1283312251552978e-07, + "loss": 0.4455, + "step": 20285 + }, + { + "epoch": 1.7286749041329357, + "grad_norm": 15.8125, + "learning_rate": 1.1276336566128226e-07, + "loss": 0.6668, + "step": 20286 + }, + { + "epoch": 1.7287601193012356, + "grad_norm": 17.75, + "learning_rate": 1.1269362935804454e-07, + "loss": 0.8941, + "step": 20287 + }, + { + "epoch": 1.7288453344695356, + "grad_norm": 18.5, + "learning_rate": 1.1262391360707642e-07, + "loss": 0.5162, + "step": 20288 + }, + { + "epoch": 1.7289305496378355, + "grad_norm": 17.625, + "learning_rate": 1.1255421840963788e-07, + "loss": 0.7474, + "step": 20289 + }, + { + "epoch": 1.7290157648061355, + "grad_norm": 13.9375, + "learning_rate": 1.1248454376698844e-07, + "loss": 0.5555, + "step": 20290 + }, + { + "epoch": 1.7291009799744355, + "grad_norm": 12.375, + "learning_rate": 1.1241488968038727e-07, + "loss": 0.3473, + "step": 20291 + }, + { + "epoch": 1.7291861951427354, + "grad_norm": 10.9375, + "learning_rate": 1.1234525615109307e-07, + "loss": 0.2393, + "step": 20292 + }, + { + "epoch": 1.7292714103110354, + "grad_norm": 24.625, + "learning_rate": 1.1227564318036385e-07, + "loss": 0.5151, + "step": 20293 + }, + { + "epoch": 1.7293566254793353, + "grad_norm": 12.875, + "learning_rate": 1.1220605076945778e-07, + "loss": 0.5041, + "step": 20294 + }, + { + "epoch": 1.7294418406476353, + "grad_norm": 14.125, + "learning_rate": 1.1213647891963276e-07, + "loss": 0.3513, + "step": 20295 + }, + { + "epoch": 1.7295270558159352, + "grad_norm": 16.375, + "learning_rate": 1.1206692763214553e-07, + "loss": 0.4617, + "step": 20296 + }, + { + "epoch": 1.7296122709842352, + "grad_norm": 16.375, + "learning_rate": 1.1199739690825329e-07, + "loss": 0.7602, + "step": 20297 + }, + { + "epoch": 1.7296974861525352, + "grad_norm": 13.0, + "learning_rate": 1.119278867492124e-07, + "loss": 0.3296, + "step": 20298 + }, + { + "epoch": 1.7297827013208351, + "grad_norm": 20.875, + "learning_rate": 1.1185839715627923e-07, + "loss": 0.7227, + "step": 20299 + }, + { + "epoch": 1.729867916489135, + "grad_norm": 15.3125, + "learning_rate": 1.1178892813070927e-07, + "loss": 0.5868, + "step": 20300 + }, + { + "epoch": 1.729953131657435, + "grad_norm": 16.25, + "learning_rate": 1.1171947967375779e-07, + "loss": 0.5576, + "step": 20301 + }, + { + "epoch": 1.730038346825735, + "grad_norm": 18.5, + "learning_rate": 1.116500517866799e-07, + "loss": 1.0257, + "step": 20302 + }, + { + "epoch": 1.730123561994035, + "grad_norm": 28.0, + "learning_rate": 1.1158064447073028e-07, + "loss": 0.7797, + "step": 20303 + }, + { + "epoch": 1.730208777162335, + "grad_norm": 15.3125, + "learning_rate": 1.1151125772716334e-07, + "loss": 0.555, + "step": 20304 + }, + { + "epoch": 1.7302939923306349, + "grad_norm": 19.25, + "learning_rate": 1.1144189155723267e-07, + "loss": 0.5972, + "step": 20305 + }, + { + "epoch": 1.7303792074989348, + "grad_norm": 14.6875, + "learning_rate": 1.1137254596219186e-07, + "loss": 0.4612, + "step": 20306 + }, + { + "epoch": 1.7304644226672348, + "grad_norm": 12.25, + "learning_rate": 1.1130322094329418e-07, + "loss": 0.3243, + "step": 20307 + }, + { + "epoch": 1.7305496378355347, + "grad_norm": 16.0, + "learning_rate": 1.1123391650179255e-07, + "loss": 0.5448, + "step": 20308 + }, + { + "epoch": 1.7306348530038347, + "grad_norm": 16.25, + "learning_rate": 1.1116463263893915e-07, + "loss": 0.7366, + "step": 20309 + }, + { + "epoch": 1.7307200681721346, + "grad_norm": 13.875, + "learning_rate": 1.1109536935598589e-07, + "loss": 0.6221, + "step": 20310 + }, + { + "epoch": 1.7308052833404346, + "grad_norm": 14.0, + "learning_rate": 1.1102612665418441e-07, + "loss": 0.2628, + "step": 20311 + }, + { + "epoch": 1.7308904985087346, + "grad_norm": 20.0, + "learning_rate": 1.109569045347865e-07, + "loss": 0.9918, + "step": 20312 + }, + { + "epoch": 1.7309757136770345, + "grad_norm": 17.0, + "learning_rate": 1.1088770299904252e-07, + "loss": 0.7479, + "step": 20313 + }, + { + "epoch": 1.7310609288453345, + "grad_norm": 18.375, + "learning_rate": 1.1081852204820317e-07, + "loss": 0.5524, + "step": 20314 + }, + { + "epoch": 1.7311461440136344, + "grad_norm": 16.25, + "learning_rate": 1.1074936168351855e-07, + "loss": 0.6101, + "step": 20315 + }, + { + "epoch": 1.7312313591819344, + "grad_norm": 14.125, + "learning_rate": 1.106802219062389e-07, + "loss": 0.6001, + "step": 20316 + }, + { + "epoch": 1.7313165743502343, + "grad_norm": 18.125, + "learning_rate": 1.1061110271761335e-07, + "loss": 0.5876, + "step": 20317 + }, + { + "epoch": 1.7314017895185343, + "grad_norm": 19.5, + "learning_rate": 1.1054200411889066e-07, + "loss": 0.7039, + "step": 20318 + }, + { + "epoch": 1.7314870046868343, + "grad_norm": 15.9375, + "learning_rate": 1.1047292611131966e-07, + "loss": 0.55, + "step": 20319 + }, + { + "epoch": 1.7315722198551342, + "grad_norm": 12.8125, + "learning_rate": 1.1040386869614894e-07, + "loss": 0.4629, + "step": 20320 + }, + { + "epoch": 1.7316574350234342, + "grad_norm": 16.5, + "learning_rate": 1.103348318746264e-07, + "loss": 0.7466, + "step": 20321 + }, + { + "epoch": 1.7317426501917341, + "grad_norm": 19.375, + "learning_rate": 1.1026581564799923e-07, + "loss": 0.8077, + "step": 20322 + }, + { + "epoch": 1.731827865360034, + "grad_norm": 15.625, + "learning_rate": 1.1019682001751492e-07, + "loss": 0.7528, + "step": 20323 + }, + { + "epoch": 1.731913080528334, + "grad_norm": 15.9375, + "learning_rate": 1.1012784498442036e-07, + "loss": 0.6486, + "step": 20324 + }, + { + "epoch": 1.731998295696634, + "grad_norm": 16.375, + "learning_rate": 1.1005889054996194e-07, + "loss": 0.3865, + "step": 20325 + }, + { + "epoch": 1.732083510864934, + "grad_norm": 14.0625, + "learning_rate": 1.0998995671538545e-07, + "loss": 0.5874, + "step": 20326 + }, + { + "epoch": 1.732168726033234, + "grad_norm": 19.375, + "learning_rate": 1.0992104348193686e-07, + "loss": 0.7047, + "step": 20327 + }, + { + "epoch": 1.7322539412015339, + "grad_norm": 12.25, + "learning_rate": 1.098521508508614e-07, + "loss": 0.4578, + "step": 20328 + }, + { + "epoch": 1.7323391563698338, + "grad_norm": 16.75, + "learning_rate": 1.0978327882340434e-07, + "loss": 0.5496, + "step": 20329 + }, + { + "epoch": 1.7324243715381338, + "grad_norm": 13.875, + "learning_rate": 1.0971442740080968e-07, + "loss": 0.2163, + "step": 20330 + }, + { + "epoch": 1.7325095867064337, + "grad_norm": 21.875, + "learning_rate": 1.0964559658432211e-07, + "loss": 0.7502, + "step": 20331 + }, + { + "epoch": 1.7325948018747337, + "grad_norm": 12.75, + "learning_rate": 1.0957678637518538e-07, + "loss": 0.3934, + "step": 20332 + }, + { + "epoch": 1.7326800170430336, + "grad_norm": 22.5, + "learning_rate": 1.0950799677464277e-07, + "loss": 0.6742, + "step": 20333 + }, + { + "epoch": 1.7327652322113336, + "grad_norm": 15.3125, + "learning_rate": 1.0943922778393776e-07, + "loss": 0.6318, + "step": 20334 + }, + { + "epoch": 1.7328504473796336, + "grad_norm": 14.125, + "learning_rate": 1.0937047940431253e-07, + "loss": 0.5712, + "step": 20335 + }, + { + "epoch": 1.7329356625479335, + "grad_norm": 16.0, + "learning_rate": 1.0930175163700985e-07, + "loss": 0.4628, + "step": 20336 + }, + { + "epoch": 1.7330208777162335, + "grad_norm": 16.5, + "learning_rate": 1.092330444832715e-07, + "loss": 0.4205, + "step": 20337 + }, + { + "epoch": 1.7331060928845334, + "grad_norm": 15.4375, + "learning_rate": 1.0916435794433939e-07, + "loss": 0.5404, + "step": 20338 + }, + { + "epoch": 1.7331913080528334, + "grad_norm": 16.625, + "learning_rate": 1.0909569202145435e-07, + "loss": 0.8362, + "step": 20339 + }, + { + "epoch": 1.7332765232211333, + "grad_norm": 17.125, + "learning_rate": 1.090270467158576e-07, + "loss": 0.8025, + "step": 20340 + }, + { + "epoch": 1.7333617383894333, + "grad_norm": 18.625, + "learning_rate": 1.0895842202878911e-07, + "loss": 0.7452, + "step": 20341 + }, + { + "epoch": 1.7334469535577333, + "grad_norm": 15.4375, + "learning_rate": 1.0888981796148973e-07, + "loss": 0.6066, + "step": 20342 + }, + { + "epoch": 1.7335321687260332, + "grad_norm": 12.6875, + "learning_rate": 1.0882123451519857e-07, + "loss": 0.5634, + "step": 20343 + }, + { + "epoch": 1.7336173838943332, + "grad_norm": 17.0, + "learning_rate": 1.0875267169115523e-07, + "loss": 0.6802, + "step": 20344 + }, + { + "epoch": 1.7337025990626331, + "grad_norm": 20.375, + "learning_rate": 1.0868412949059868e-07, + "loss": 0.6934, + "step": 20345 + }, + { + "epoch": 1.733787814230933, + "grad_norm": 15.0625, + "learning_rate": 1.0861560791476794e-07, + "loss": 0.5575, + "step": 20346 + }, + { + "epoch": 1.733873029399233, + "grad_norm": 12.625, + "learning_rate": 1.0854710696490064e-07, + "loss": 0.3875, + "step": 20347 + }, + { + "epoch": 1.733958244567533, + "grad_norm": 15.75, + "learning_rate": 1.084786266422351e-07, + "loss": 0.3892, + "step": 20348 + }, + { + "epoch": 1.734043459735833, + "grad_norm": 25.5, + "learning_rate": 1.0841016694800838e-07, + "loss": 1.094, + "step": 20349 + }, + { + "epoch": 1.734128674904133, + "grad_norm": 20.875, + "learning_rate": 1.083417278834581e-07, + "loss": 1.0756, + "step": 20350 + }, + { + "epoch": 1.7342138900724329, + "grad_norm": 15.625, + "learning_rate": 1.082733094498209e-07, + "loss": 0.5394, + "step": 20351 + }, + { + "epoch": 1.7342991052407328, + "grad_norm": 15.1875, + "learning_rate": 1.0820491164833288e-07, + "loss": 0.845, + "step": 20352 + }, + { + "epoch": 1.7343843204090328, + "grad_norm": 17.125, + "learning_rate": 1.081365344802303e-07, + "loss": 0.727, + "step": 20353 + }, + { + "epoch": 1.7344695355773327, + "grad_norm": 13.8125, + "learning_rate": 1.0806817794674878e-07, + "loss": 0.4667, + "step": 20354 + }, + { + "epoch": 1.7345547507456327, + "grad_norm": 22.25, + "learning_rate": 1.0799984204912378e-07, + "loss": 0.8753, + "step": 20355 + }, + { + "epoch": 1.7346399659139327, + "grad_norm": 18.75, + "learning_rate": 1.0793152678859011e-07, + "loss": 0.7515, + "step": 20356 + }, + { + "epoch": 1.7347251810822326, + "grad_norm": 18.75, + "learning_rate": 1.0786323216638181e-07, + "loss": 0.6915, + "step": 20357 + }, + { + "epoch": 1.7348103962505326, + "grad_norm": 18.25, + "learning_rate": 1.0779495818373356e-07, + "loss": 0.7628, + "step": 20358 + }, + { + "epoch": 1.7348956114188325, + "grad_norm": 19.0, + "learning_rate": 1.077267048418791e-07, + "loss": 0.6592, + "step": 20359 + }, + { + "epoch": 1.7349808265871325, + "grad_norm": 15.875, + "learning_rate": 1.0765847214205149e-07, + "loss": 0.5753, + "step": 20360 + }, + { + "epoch": 1.7350660417554324, + "grad_norm": 15.625, + "learning_rate": 1.0759026008548404e-07, + "loss": 0.2975, + "step": 20361 + }, + { + "epoch": 1.7351512569237324, + "grad_norm": 27.75, + "learning_rate": 1.0752206867340951e-07, + "loss": 1.1362, + "step": 20362 + }, + { + "epoch": 1.7352364720920324, + "grad_norm": 20.875, + "learning_rate": 1.074538979070601e-07, + "loss": 0.876, + "step": 20363 + }, + { + "epoch": 1.7353216872603323, + "grad_norm": 15.1875, + "learning_rate": 1.0738574778766747e-07, + "loss": 0.3301, + "step": 20364 + }, + { + "epoch": 1.7354069024286323, + "grad_norm": 12.375, + "learning_rate": 1.073176183164637e-07, + "loss": 0.3144, + "step": 20365 + }, + { + "epoch": 1.7354921175969322, + "grad_norm": 20.5, + "learning_rate": 1.0724950949467933e-07, + "loss": 0.5909, + "step": 20366 + }, + { + "epoch": 1.7355773327652322, + "grad_norm": 16.5, + "learning_rate": 1.0718142132354531e-07, + "loss": 0.633, + "step": 20367 + }, + { + "epoch": 1.7356625479335321, + "grad_norm": 24.875, + "learning_rate": 1.071133538042926e-07, + "loss": 0.5368, + "step": 20368 + }, + { + "epoch": 1.735747763101832, + "grad_norm": 18.5, + "learning_rate": 1.0704530693815065e-07, + "loss": 0.5844, + "step": 20369 + }, + { + "epoch": 1.735832978270132, + "grad_norm": 15.5625, + "learning_rate": 1.0697728072634916e-07, + "loss": 0.4528, + "step": 20370 + }, + { + "epoch": 1.735918193438432, + "grad_norm": 19.375, + "learning_rate": 1.069092751701177e-07, + "loss": 0.7777, + "step": 20371 + }, + { + "epoch": 1.736003408606732, + "grad_norm": 15.8125, + "learning_rate": 1.0684129027068544e-07, + "loss": 0.8896, + "step": 20372 + }, + { + "epoch": 1.736088623775032, + "grad_norm": 13.625, + "learning_rate": 1.067733260292804e-07, + "loss": 0.5444, + "step": 20373 + }, + { + "epoch": 1.7361738389433319, + "grad_norm": 14.625, + "learning_rate": 1.0670538244713092e-07, + "loss": 0.4681, + "step": 20374 + }, + { + "epoch": 1.7362590541116318, + "grad_norm": 10.875, + "learning_rate": 1.0663745952546475e-07, + "loss": 0.5546, + "step": 20375 + }, + { + "epoch": 1.7363442692799318, + "grad_norm": 16.875, + "learning_rate": 1.0656955726550955e-07, + "loss": 0.4509, + "step": 20376 + }, + { + "epoch": 1.7364294844482318, + "grad_norm": 19.0, + "learning_rate": 1.0650167566849223e-07, + "loss": 0.7451, + "step": 20377 + }, + { + "epoch": 1.7365146996165317, + "grad_norm": 18.75, + "learning_rate": 1.0643381473563946e-07, + "loss": 0.8219, + "step": 20378 + }, + { + "epoch": 1.7365999147848317, + "grad_norm": 19.0, + "learning_rate": 1.0636597446817748e-07, + "loss": 0.6184, + "step": 20379 + }, + { + "epoch": 1.7366851299531316, + "grad_norm": 13.4375, + "learning_rate": 1.0629815486733267e-07, + "loss": 0.5796, + "step": 20380 + }, + { + "epoch": 1.7367703451214316, + "grad_norm": 15.5625, + "learning_rate": 1.0623035593433029e-07, + "loss": 0.8514, + "step": 20381 + }, + { + "epoch": 1.7368555602897315, + "grad_norm": 19.125, + "learning_rate": 1.0616257767039534e-07, + "loss": 0.6903, + "step": 20382 + }, + { + "epoch": 1.7369407754580315, + "grad_norm": 15.75, + "learning_rate": 1.0609482007675268e-07, + "loss": 0.7132, + "step": 20383 + }, + { + "epoch": 1.7370259906263315, + "grad_norm": 13.625, + "learning_rate": 1.0602708315462701e-07, + "loss": 0.5968, + "step": 20384 + }, + { + "epoch": 1.7371112057946314, + "grad_norm": 19.5, + "learning_rate": 1.0595936690524263e-07, + "loss": 0.8211, + "step": 20385 + }, + { + "epoch": 1.7371964209629314, + "grad_norm": 14.0625, + "learning_rate": 1.0589167132982247e-07, + "loss": 0.6062, + "step": 20386 + }, + { + "epoch": 1.7372816361312313, + "grad_norm": 18.25, + "learning_rate": 1.0582399642959041e-07, + "loss": 0.808, + "step": 20387 + }, + { + "epoch": 1.7373668512995313, + "grad_norm": 14.5625, + "learning_rate": 1.057563422057696e-07, + "loss": 0.2836, + "step": 20388 + }, + { + "epoch": 1.7374520664678312, + "grad_norm": 14.5, + "learning_rate": 1.0568870865958216e-07, + "loss": 0.3815, + "step": 20389 + }, + { + "epoch": 1.7375372816361312, + "grad_norm": 12.5625, + "learning_rate": 1.056210957922503e-07, + "loss": 0.457, + "step": 20390 + }, + { + "epoch": 1.7376224968044311, + "grad_norm": 14.625, + "learning_rate": 1.0555350360499595e-07, + "loss": 0.4201, + "step": 20391 + }, + { + "epoch": 1.737707711972731, + "grad_norm": 15.5, + "learning_rate": 1.0548593209904076e-07, + "loss": 0.5519, + "step": 20392 + }, + { + "epoch": 1.737792927141031, + "grad_norm": 18.0, + "learning_rate": 1.0541838127560572e-07, + "loss": 0.8545, + "step": 20393 + }, + { + "epoch": 1.737878142309331, + "grad_norm": 21.25, + "learning_rate": 1.0535085113591136e-07, + "loss": 0.6859, + "step": 20394 + }, + { + "epoch": 1.737963357477631, + "grad_norm": 15.0, + "learning_rate": 1.0528334168117812e-07, + "loss": 0.4384, + "step": 20395 + }, + { + "epoch": 1.738048572645931, + "grad_norm": 18.75, + "learning_rate": 1.0521585291262612e-07, + "loss": 0.9514, + "step": 20396 + }, + { + "epoch": 1.738133787814231, + "grad_norm": 12.5625, + "learning_rate": 1.0514838483147494e-07, + "loss": 0.4935, + "step": 20397 + }, + { + "epoch": 1.7382190029825308, + "grad_norm": 27.0, + "learning_rate": 1.0508093743894373e-07, + "loss": 0.7809, + "step": 20398 + }, + { + "epoch": 1.7383042181508308, + "grad_norm": 18.75, + "learning_rate": 1.0501351073625099e-07, + "loss": 0.8966, + "step": 20399 + }, + { + "epoch": 1.7383894333191308, + "grad_norm": 11.3125, + "learning_rate": 1.0494610472461544e-07, + "loss": 0.4087, + "step": 20400 + }, + { + "epoch": 1.7384746484874307, + "grad_norm": 14.25, + "learning_rate": 1.0487871940525543e-07, + "loss": 0.4796, + "step": 20401 + }, + { + "epoch": 1.7385598636557307, + "grad_norm": 20.0, + "learning_rate": 1.0481135477938844e-07, + "loss": 0.6553, + "step": 20402 + }, + { + "epoch": 1.7386450788240306, + "grad_norm": 17.375, + "learning_rate": 1.0474401084823185e-07, + "loss": 0.7148, + "step": 20403 + }, + { + "epoch": 1.7387302939923306, + "grad_norm": 14.75, + "learning_rate": 1.0467668761300245e-07, + "loss": 0.6115, + "step": 20404 + }, + { + "epoch": 1.7388155091606305, + "grad_norm": 18.75, + "learning_rate": 1.0460938507491733e-07, + "loss": 0.5769, + "step": 20405 + }, + { + "epoch": 1.7389007243289305, + "grad_norm": 13.75, + "learning_rate": 1.0454210323519245e-07, + "loss": 0.6073, + "step": 20406 + }, + { + "epoch": 1.7389859394972305, + "grad_norm": 18.875, + "learning_rate": 1.0447484209504325e-07, + "loss": 0.5419, + "step": 20407 + }, + { + "epoch": 1.7390711546655304, + "grad_norm": 15.375, + "learning_rate": 1.0440760165568568e-07, + "loss": 0.5887, + "step": 20408 + }, + { + "epoch": 1.7391563698338304, + "grad_norm": 16.5, + "learning_rate": 1.0434038191833474e-07, + "loss": 0.604, + "step": 20409 + }, + { + "epoch": 1.7392415850021303, + "grad_norm": 14.4375, + "learning_rate": 1.0427318288420529e-07, + "loss": 0.5071, + "step": 20410 + }, + { + "epoch": 1.7393268001704303, + "grad_norm": 16.5, + "learning_rate": 1.0420600455451138e-07, + "loss": 0.3532, + "step": 20411 + }, + { + "epoch": 1.7394120153387302, + "grad_norm": 23.625, + "learning_rate": 1.0413884693046703e-07, + "loss": 0.3544, + "step": 20412 + }, + { + "epoch": 1.7394972305070302, + "grad_norm": 16.25, + "learning_rate": 1.0407171001328639e-07, + "loss": 0.5435, + "step": 20413 + }, + { + "epoch": 1.7395824456753302, + "grad_norm": 13.9375, + "learning_rate": 1.0400459380418184e-07, + "loss": 0.2353, + "step": 20414 + }, + { + "epoch": 1.7396676608436301, + "grad_norm": 16.0, + "learning_rate": 1.03937498304367e-07, + "loss": 0.7209, + "step": 20415 + }, + { + "epoch": 1.73975287601193, + "grad_norm": 15.5625, + "learning_rate": 1.038704235150538e-07, + "loss": 0.4017, + "step": 20416 + }, + { + "epoch": 1.73983809118023, + "grad_norm": 16.375, + "learning_rate": 1.0380336943745462e-07, + "loss": 0.565, + "step": 20417 + }, + { + "epoch": 1.73992330634853, + "grad_norm": 16.875, + "learning_rate": 1.037363360727811e-07, + "loss": 0.7592, + "step": 20418 + }, + { + "epoch": 1.74000852151683, + "grad_norm": 11.5625, + "learning_rate": 1.036693234222448e-07, + "loss": 0.4036, + "step": 20419 + }, + { + "epoch": 1.74009373668513, + "grad_norm": 24.75, + "learning_rate": 1.0360233148705643e-07, + "loss": 0.6585, + "step": 20420 + }, + { + "epoch": 1.7401789518534299, + "grad_norm": 16.875, + "learning_rate": 1.0353536026842693e-07, + "loss": 0.6857, + "step": 20421 + }, + { + "epoch": 1.7402641670217298, + "grad_norm": 21.0, + "learning_rate": 1.0346840976756604e-07, + "loss": 0.6703, + "step": 20422 + }, + { + "epoch": 1.7403493821900298, + "grad_norm": 13.0, + "learning_rate": 1.0340147998568417e-07, + "loss": 0.5531, + "step": 20423 + }, + { + "epoch": 1.74043459735833, + "grad_norm": 15.75, + "learning_rate": 1.0333457092399024e-07, + "loss": 0.7951, + "step": 20424 + }, + { + "epoch": 1.74051981252663, + "grad_norm": 14.25, + "learning_rate": 1.0326768258369383e-07, + "loss": 0.4648, + "step": 20425 + }, + { + "epoch": 1.7406050276949299, + "grad_norm": 15.6875, + "learning_rate": 1.032008149660034e-07, + "loss": 0.9563, + "step": 20426 + }, + { + "epoch": 1.7406902428632298, + "grad_norm": 13.6875, + "learning_rate": 1.0313396807212772e-07, + "loss": 0.3751, + "step": 20427 + }, + { + "epoch": 1.7407754580315298, + "grad_norm": 11.625, + "learning_rate": 1.030671419032743e-07, + "loss": 0.4556, + "step": 20428 + }, + { + "epoch": 1.7408606731998297, + "grad_norm": 21.25, + "learning_rate": 1.0300033646065119e-07, + "loss": 1.0432, + "step": 20429 + }, + { + "epoch": 1.7409458883681297, + "grad_norm": 9.875, + "learning_rate": 1.0293355174546507e-07, + "loss": 0.1894, + "step": 20430 + }, + { + "epoch": 1.7410311035364296, + "grad_norm": 12.3125, + "learning_rate": 1.0286678775892318e-07, + "loss": 0.41, + "step": 20431 + }, + { + "epoch": 1.7411163187047296, + "grad_norm": 14.3125, + "learning_rate": 1.028000445022323e-07, + "loss": 0.7088, + "step": 20432 + }, + { + "epoch": 1.7412015338730296, + "grad_norm": 14.0625, + "learning_rate": 1.0273332197659788e-07, + "loss": 0.5009, + "step": 20433 + }, + { + "epoch": 1.7412867490413295, + "grad_norm": 13.4375, + "learning_rate": 1.0266662018322604e-07, + "loss": 0.4311, + "step": 20434 + }, + { + "epoch": 1.7413719642096295, + "grad_norm": 15.9375, + "learning_rate": 1.0259993912332233e-07, + "loss": 0.5274, + "step": 20435 + }, + { + "epoch": 1.7414571793779294, + "grad_norm": 18.0, + "learning_rate": 1.0253327879809122e-07, + "loss": 0.6928, + "step": 20436 + }, + { + "epoch": 1.7415423945462294, + "grad_norm": 12.25, + "learning_rate": 1.024666392087377e-07, + "loss": 0.5453, + "step": 20437 + }, + { + "epoch": 1.7416276097145293, + "grad_norm": 14.8125, + "learning_rate": 1.024000203564661e-07, + "loss": 0.6398, + "step": 20438 + }, + { + "epoch": 1.7417128248828293, + "grad_norm": 24.625, + "learning_rate": 1.0233342224248002e-07, + "loss": 0.5252, + "step": 20439 + }, + { + "epoch": 1.7417980400511293, + "grad_norm": 33.5, + "learning_rate": 1.0226684486798325e-07, + "loss": 0.7633, + "step": 20440 + }, + { + "epoch": 1.7418832552194292, + "grad_norm": 13.6875, + "learning_rate": 1.022002882341784e-07, + "loss": 0.513, + "step": 20441 + }, + { + "epoch": 1.7419684703877292, + "grad_norm": 14.5625, + "learning_rate": 1.0213375234226871e-07, + "loss": 0.6458, + "step": 20442 + }, + { + "epoch": 1.7420536855560291, + "grad_norm": 19.5, + "learning_rate": 1.0206723719345626e-07, + "loss": 0.6199, + "step": 20443 + }, + { + "epoch": 1.742138900724329, + "grad_norm": 10.8125, + "learning_rate": 1.0200074278894342e-07, + "loss": 0.3249, + "step": 20444 + }, + { + "epoch": 1.742224115892629, + "grad_norm": 18.5, + "learning_rate": 1.0193426912993145e-07, + "loss": 0.9058, + "step": 20445 + }, + { + "epoch": 1.742309331060929, + "grad_norm": 14.125, + "learning_rate": 1.0186781621762176e-07, + "loss": 0.6742, + "step": 20446 + }, + { + "epoch": 1.742394546229229, + "grad_norm": 18.75, + "learning_rate": 1.0180138405321505e-07, + "loss": 0.749, + "step": 20447 + }, + { + "epoch": 1.742479761397529, + "grad_norm": 13.625, + "learning_rate": 1.0173497263791188e-07, + "loss": 0.4413, + "step": 20448 + }, + { + "epoch": 1.7425649765658289, + "grad_norm": 16.0, + "learning_rate": 1.0166858197291268e-07, + "loss": 0.4983, + "step": 20449 + }, + { + "epoch": 1.7426501917341288, + "grad_norm": 25.625, + "learning_rate": 1.0160221205941665e-07, + "loss": 0.4441, + "step": 20450 + }, + { + "epoch": 1.7427354069024288, + "grad_norm": 11.6875, + "learning_rate": 1.0153586289862349e-07, + "loss": 0.4698, + "step": 20451 + }, + { + "epoch": 1.7428206220707287, + "grad_norm": 15.875, + "learning_rate": 1.0146953449173242e-07, + "loss": 0.5969, + "step": 20452 + }, + { + "epoch": 1.7429058372390287, + "grad_norm": 14.1875, + "learning_rate": 1.0140322683994161e-07, + "loss": 0.5445, + "step": 20453 + }, + { + "epoch": 1.7429910524073287, + "grad_norm": 14.375, + "learning_rate": 1.0133693994444971e-07, + "loss": 0.605, + "step": 20454 + }, + { + "epoch": 1.7430762675756286, + "grad_norm": 13.8125, + "learning_rate": 1.0127067380645409e-07, + "loss": 0.494, + "step": 20455 + }, + { + "epoch": 1.7431614827439286, + "grad_norm": 16.125, + "learning_rate": 1.0120442842715252e-07, + "loss": 0.6719, + "step": 20456 + }, + { + "epoch": 1.7432466979122285, + "grad_norm": 11.0, + "learning_rate": 1.0113820380774242e-07, + "loss": 0.2613, + "step": 20457 + }, + { + "epoch": 1.7433319130805285, + "grad_norm": 21.5, + "learning_rate": 1.0107199994941988e-07, + "loss": 0.5414, + "step": 20458 + }, + { + "epoch": 1.7434171282488284, + "grad_norm": 27.625, + "learning_rate": 1.0100581685338173e-07, + "loss": 0.8083, + "step": 20459 + }, + { + "epoch": 1.7435023434171284, + "grad_norm": 12.8125, + "learning_rate": 1.0093965452082397e-07, + "loss": 0.505, + "step": 20460 + }, + { + "epoch": 1.7435875585854284, + "grad_norm": 15.0625, + "learning_rate": 1.0087351295294229e-07, + "loss": 0.4654, + "step": 20461 + }, + { + "epoch": 1.7436727737537283, + "grad_norm": 14.1875, + "learning_rate": 1.0080739215093172e-07, + "loss": 0.4282, + "step": 20462 + }, + { + "epoch": 1.7437579889220283, + "grad_norm": 23.75, + "learning_rate": 1.0074129211598684e-07, + "loss": 0.7353, + "step": 20463 + }, + { + "epoch": 1.7438432040903282, + "grad_norm": 23.75, + "learning_rate": 1.0067521284930271e-07, + "loss": 0.6236, + "step": 20464 + }, + { + "epoch": 1.7439284192586282, + "grad_norm": 18.75, + "learning_rate": 1.0060915435207333e-07, + "loss": 0.8028, + "step": 20465 + }, + { + "epoch": 1.7440136344269281, + "grad_norm": 13.5625, + "learning_rate": 1.0054311662549208e-07, + "loss": 0.4277, + "step": 20466 + }, + { + "epoch": 1.744098849595228, + "grad_norm": 15.9375, + "learning_rate": 1.004770996707527e-07, + "loss": 0.6807, + "step": 20467 + }, + { + "epoch": 1.744184064763528, + "grad_norm": 18.375, + "learning_rate": 1.0041110348904787e-07, + "loss": 0.6957, + "step": 20468 + }, + { + "epoch": 1.744269279931828, + "grad_norm": 19.25, + "learning_rate": 1.0034512808157079e-07, + "loss": 0.8879, + "step": 20469 + }, + { + "epoch": 1.744354495100128, + "grad_norm": 17.375, + "learning_rate": 1.0027917344951329e-07, + "loss": 0.662, + "step": 20470 + }, + { + "epoch": 1.744439710268428, + "grad_norm": 14.25, + "learning_rate": 1.0021323959406692e-07, + "loss": 0.5011, + "step": 20471 + }, + { + "epoch": 1.7445249254367279, + "grad_norm": 17.0, + "learning_rate": 1.0014732651642365e-07, + "loss": 0.8457, + "step": 20472 + }, + { + "epoch": 1.7446101406050278, + "grad_norm": 21.75, + "learning_rate": 1.0008143421777444e-07, + "loss": 0.9203, + "step": 20473 + }, + { + "epoch": 1.7446953557733278, + "grad_norm": 11.75, + "learning_rate": 1.0001556269931018e-07, + "loss": 0.3661, + "step": 20474 + }, + { + "epoch": 1.7447805709416278, + "grad_norm": 18.125, + "learning_rate": 9.9949711962221e-08, + "loss": 0.7605, + "step": 20475 + }, + { + "epoch": 1.7448657861099277, + "grad_norm": 10.1875, + "learning_rate": 9.988388200769694e-08, + "loss": 0.2155, + "step": 20476 + }, + { + "epoch": 1.7449510012782277, + "grad_norm": 15.5, + "learning_rate": 9.98180728369276e-08, + "loss": 0.4357, + "step": 20477 + }, + { + "epoch": 1.7450362164465276, + "grad_norm": 15.125, + "learning_rate": 9.975228445110258e-08, + "loss": 0.6236, + "step": 20478 + }, + { + "epoch": 1.7451214316148276, + "grad_norm": 18.75, + "learning_rate": 9.968651685141038e-08, + "loss": 0.787, + "step": 20479 + }, + { + "epoch": 1.7452066467831275, + "grad_norm": 10.25, + "learning_rate": 9.962077003903933e-08, + "loss": 0.1847, + "step": 20480 + }, + { + "epoch": 1.7452918619514275, + "grad_norm": 22.625, + "learning_rate": 9.955504401517782e-08, + "loss": 0.5763, + "step": 20481 + }, + { + "epoch": 1.7453770771197274, + "grad_norm": 15.0625, + "learning_rate": 9.948933878101378e-08, + "loss": 0.5129, + "step": 20482 + }, + { + "epoch": 1.7454622922880274, + "grad_norm": 17.875, + "learning_rate": 9.942365433773401e-08, + "loss": 0.5203, + "step": 20483 + }, + { + "epoch": 1.7455475074563274, + "grad_norm": 16.0, + "learning_rate": 9.93579906865258e-08, + "loss": 0.655, + "step": 20484 + }, + { + "epoch": 1.7456327226246273, + "grad_norm": 18.25, + "learning_rate": 9.92923478285758e-08, + "loss": 0.8075, + "step": 20485 + }, + { + "epoch": 1.7457179377929273, + "grad_norm": 12.8125, + "learning_rate": 9.922672576507045e-08, + "loss": 0.5332, + "step": 20486 + }, + { + "epoch": 1.7458031529612272, + "grad_norm": 18.875, + "learning_rate": 9.916112449719517e-08, + "loss": 0.52, + "step": 20487 + }, + { + "epoch": 1.7458883681295272, + "grad_norm": 13.625, + "learning_rate": 9.909554402613555e-08, + "loss": 0.3861, + "step": 20488 + }, + { + "epoch": 1.7459735832978271, + "grad_norm": 13.5, + "learning_rate": 9.90299843530769e-08, + "loss": 0.3637, + "step": 20489 + }, + { + "epoch": 1.746058798466127, + "grad_norm": 13.5625, + "learning_rate": 9.896444547920367e-08, + "loss": 0.5383, + "step": 20490 + }, + { + "epoch": 1.746144013634427, + "grad_norm": 18.0, + "learning_rate": 9.889892740570062e-08, + "loss": 0.662, + "step": 20491 + }, + { + "epoch": 1.746229228802727, + "grad_norm": 17.5, + "learning_rate": 9.883343013375123e-08, + "loss": 0.9331, + "step": 20492 + }, + { + "epoch": 1.746314443971027, + "grad_norm": 22.125, + "learning_rate": 9.876795366453932e-08, + "loss": 0.613, + "step": 20493 + }, + { + "epoch": 1.746399659139327, + "grad_norm": 23.75, + "learning_rate": 9.870249799924847e-08, + "loss": 0.5122, + "step": 20494 + }, + { + "epoch": 1.7464848743076269, + "grad_norm": 12.8125, + "learning_rate": 9.863706313906111e-08, + "loss": 0.4902, + "step": 20495 + }, + { + "epoch": 1.7465700894759268, + "grad_norm": 13.3125, + "learning_rate": 9.85716490851596e-08, + "loss": 0.8337, + "step": 20496 + }, + { + "epoch": 1.7466553046442268, + "grad_norm": 12.9375, + "learning_rate": 9.85062558387262e-08, + "loss": 0.5055, + "step": 20497 + }, + { + "epoch": 1.7467405198125268, + "grad_norm": 16.0, + "learning_rate": 9.844088340094274e-08, + "loss": 0.5628, + "step": 20498 + }, + { + "epoch": 1.7468257349808267, + "grad_norm": 21.0, + "learning_rate": 9.837553177299063e-08, + "loss": 0.5968, + "step": 20499 + }, + { + "epoch": 1.7469109501491267, + "grad_norm": 10.3125, + "learning_rate": 9.831020095605062e-08, + "loss": 0.8684, + "step": 20500 + }, + { + "epoch": 1.7469961653174266, + "grad_norm": 12.9375, + "learning_rate": 9.82448909513034e-08, + "loss": 0.2963, + "step": 20501 + }, + { + "epoch": 1.7470813804857266, + "grad_norm": 17.625, + "learning_rate": 9.817960175992902e-08, + "loss": 0.7024, + "step": 20502 + }, + { + "epoch": 1.7471665956540265, + "grad_norm": 15.625, + "learning_rate": 9.811433338310778e-08, + "loss": 0.5766, + "step": 20503 + }, + { + "epoch": 1.7472518108223265, + "grad_norm": 20.25, + "learning_rate": 9.804908582201886e-08, + "loss": 0.8173, + "step": 20504 + }, + { + "epoch": 1.7473370259906265, + "grad_norm": 12.8125, + "learning_rate": 9.79838590778412e-08, + "loss": 0.5356, + "step": 20505 + }, + { + "epoch": 1.7474222411589264, + "grad_norm": 16.125, + "learning_rate": 9.791865315175355e-08, + "loss": 0.7171, + "step": 20506 + }, + { + "epoch": 1.7475074563272264, + "grad_norm": 12.5625, + "learning_rate": 9.785346804493432e-08, + "loss": 0.4671, + "step": 20507 + }, + { + "epoch": 1.7475926714955263, + "grad_norm": 14.4375, + "learning_rate": 9.778830375856168e-08, + "loss": 0.7069, + "step": 20508 + }, + { + "epoch": 1.7476778866638263, + "grad_norm": 15.875, + "learning_rate": 9.772316029381293e-08, + "loss": 0.7917, + "step": 20509 + }, + { + "epoch": 1.7477631018321262, + "grad_norm": 14.1875, + "learning_rate": 9.76580376518653e-08, + "loss": 0.5675, + "step": 20510 + }, + { + "epoch": 1.7478483170004262, + "grad_norm": 12.4375, + "learning_rate": 9.759293583389603e-08, + "loss": 0.4669, + "step": 20511 + }, + { + "epoch": 1.7479335321687262, + "grad_norm": 19.5, + "learning_rate": 9.752785484108115e-08, + "loss": 0.88, + "step": 20512 + }, + { + "epoch": 1.7480187473370261, + "grad_norm": 15.3125, + "learning_rate": 9.746279467459665e-08, + "loss": 0.7422, + "step": 20513 + }, + { + "epoch": 1.748103962505326, + "grad_norm": 13.6875, + "learning_rate": 9.73977553356184e-08, + "loss": 0.4824, + "step": 20514 + }, + { + "epoch": 1.748189177673626, + "grad_norm": 14.875, + "learning_rate": 9.733273682532172e-08, + "loss": 0.8472, + "step": 20515 + }, + { + "epoch": 1.748274392841926, + "grad_norm": 14.5625, + "learning_rate": 9.726773914488174e-08, + "loss": 0.5523, + "step": 20516 + }, + { + "epoch": 1.748359608010226, + "grad_norm": 12.6875, + "learning_rate": 9.720276229547271e-08, + "loss": 0.4846, + "step": 20517 + }, + { + "epoch": 1.748444823178526, + "grad_norm": 18.875, + "learning_rate": 9.713780627826908e-08, + "loss": 0.7862, + "step": 20518 + }, + { + "epoch": 1.7485300383468259, + "grad_norm": 16.625, + "learning_rate": 9.707287109444478e-08, + "loss": 0.729, + "step": 20519 + }, + { + "epoch": 1.7486152535151258, + "grad_norm": 9.625, + "learning_rate": 9.700795674517274e-08, + "loss": 0.8156, + "step": 20520 + }, + { + "epoch": 1.7487004686834258, + "grad_norm": 12.0, + "learning_rate": 9.694306323162664e-08, + "loss": 0.1883, + "step": 20521 + }, + { + "epoch": 1.7487856838517257, + "grad_norm": 14.3125, + "learning_rate": 9.687819055497857e-08, + "loss": 0.4699, + "step": 20522 + }, + { + "epoch": 1.7488708990200257, + "grad_norm": 14.4375, + "learning_rate": 9.681333871640123e-08, + "loss": 0.6269, + "step": 20523 + }, + { + "epoch": 1.7489561141883256, + "grad_norm": 13.125, + "learning_rate": 9.674850771706657e-08, + "loss": 0.5829, + "step": 20524 + }, + { + "epoch": 1.7490413293566256, + "grad_norm": 21.125, + "learning_rate": 9.668369755814616e-08, + "loss": 1.0002, + "step": 20525 + }, + { + "epoch": 1.7491265445249256, + "grad_norm": 13.875, + "learning_rate": 9.661890824081088e-08, + "loss": 0.5371, + "step": 20526 + }, + { + "epoch": 1.7492117596932255, + "grad_norm": 17.625, + "learning_rate": 9.655413976623187e-08, + "loss": 0.5174, + "step": 20527 + }, + { + "epoch": 1.7492969748615255, + "grad_norm": 18.0, + "learning_rate": 9.648939213557931e-08, + "loss": 0.7669, + "step": 20528 + }, + { + "epoch": 1.7493821900298254, + "grad_norm": 15.0, + "learning_rate": 9.642466535002365e-08, + "loss": 0.6562, + "step": 20529 + }, + { + "epoch": 1.7494674051981254, + "grad_norm": 22.0, + "learning_rate": 9.635995941073395e-08, + "loss": 0.5418, + "step": 20530 + }, + { + "epoch": 1.7495526203664253, + "grad_norm": 14.625, + "learning_rate": 9.629527431887997e-08, + "loss": 0.5061, + "step": 20531 + }, + { + "epoch": 1.7496378355347253, + "grad_norm": 17.5, + "learning_rate": 9.62306100756305e-08, + "loss": 0.8293, + "step": 20532 + }, + { + "epoch": 1.7497230507030253, + "grad_norm": 11.9375, + "learning_rate": 9.616596668215433e-08, + "loss": 0.3131, + "step": 20533 + }, + { + "epoch": 1.7498082658713252, + "grad_norm": 13.375, + "learning_rate": 9.610134413961914e-08, + "loss": 0.441, + "step": 20534 + }, + { + "epoch": 1.7498934810396252, + "grad_norm": 14.0, + "learning_rate": 9.603674244919328e-08, + "loss": 0.6186, + "step": 20535 + }, + { + "epoch": 1.7499786962079251, + "grad_norm": 20.625, + "learning_rate": 9.597216161204362e-08, + "loss": 0.7141, + "step": 20536 + }, + { + "epoch": 1.750063911376225, + "grad_norm": 17.125, + "learning_rate": 9.59076016293374e-08, + "loss": 0.5162, + "step": 20537 + }, + { + "epoch": 1.750149126544525, + "grad_norm": 27.75, + "learning_rate": 9.584306250224163e-08, + "loss": 0.5631, + "step": 20538 + }, + { + "epoch": 1.750234341712825, + "grad_norm": 31.0, + "learning_rate": 9.577854423192215e-08, + "loss": 1.4732, + "step": 20539 + }, + { + "epoch": 1.750319556881125, + "grad_norm": 13.0625, + "learning_rate": 9.5714046819545e-08, + "loss": 0.43, + "step": 20540 + }, + { + "epoch": 1.750404772049425, + "grad_norm": 14.6875, + "learning_rate": 9.564957026627575e-08, + "loss": 0.5671, + "step": 20541 + }, + { + "epoch": 1.7504899872177249, + "grad_norm": 12.625, + "learning_rate": 9.558511457327974e-08, + "loss": 0.3444, + "step": 20542 + }, + { + "epoch": 1.7505752023860248, + "grad_norm": 34.75, + "learning_rate": 9.552067974172158e-08, + "loss": 0.75, + "step": 20543 + }, + { + "epoch": 1.7506604175543248, + "grad_norm": 16.0, + "learning_rate": 9.545626577276548e-08, + "loss": 0.3507, + "step": 20544 + }, + { + "epoch": 1.7507456327226247, + "grad_norm": 14.5625, + "learning_rate": 9.539187266757566e-08, + "loss": 0.6438, + "step": 20545 + }, + { + "epoch": 1.7508308478909247, + "grad_norm": 17.125, + "learning_rate": 9.53275004273159e-08, + "loss": 0.6065, + "step": 20546 + }, + { + "epoch": 1.7509160630592246, + "grad_norm": 19.5, + "learning_rate": 9.526314905314916e-08, + "loss": 0.4324, + "step": 20547 + }, + { + "epoch": 1.7510012782275246, + "grad_norm": 13.4375, + "learning_rate": 9.519881854623838e-08, + "loss": 0.3667, + "step": 20548 + }, + { + "epoch": 1.7510864933958246, + "grad_norm": 17.75, + "learning_rate": 9.513450890774628e-08, + "loss": 0.8786, + "step": 20549 + }, + { + "epoch": 1.7511717085641245, + "grad_norm": 21.75, + "learning_rate": 9.50702201388351e-08, + "loss": 1.016, + "step": 20550 + }, + { + "epoch": 1.7512569237324245, + "grad_norm": 14.375, + "learning_rate": 9.500595224066642e-08, + "loss": 0.4673, + "step": 20551 + }, + { + "epoch": 1.7513421389007244, + "grad_norm": 12.375, + "learning_rate": 9.494170521440138e-08, + "loss": 0.3984, + "step": 20552 + }, + { + "epoch": 1.7514273540690244, + "grad_norm": 11.6875, + "learning_rate": 9.48774790612013e-08, + "loss": 0.521, + "step": 20553 + }, + { + "epoch": 1.7515125692373243, + "grad_norm": 20.5, + "learning_rate": 9.481327378222674e-08, + "loss": 1.1127, + "step": 20554 + }, + { + "epoch": 1.7515977844056243, + "grad_norm": 16.375, + "learning_rate": 9.474908937863805e-08, + "loss": 0.6917, + "step": 20555 + }, + { + "epoch": 1.7516829995739243, + "grad_norm": 17.625, + "learning_rate": 9.468492585159486e-08, + "loss": 0.7381, + "step": 20556 + }, + { + "epoch": 1.7517682147422242, + "grad_norm": 15.125, + "learning_rate": 9.462078320225693e-08, + "loss": 0.7665, + "step": 20557 + }, + { + "epoch": 1.7518534299105242, + "grad_norm": 15.5625, + "learning_rate": 9.455666143178332e-08, + "loss": 0.7582, + "step": 20558 + }, + { + "epoch": 1.7519386450788241, + "grad_norm": 13.6875, + "learning_rate": 9.449256054133286e-08, + "loss": 0.6289, + "step": 20559 + }, + { + "epoch": 1.752023860247124, + "grad_norm": 15.0625, + "learning_rate": 9.442848053206347e-08, + "loss": 0.5717, + "step": 20560 + }, + { + "epoch": 1.752109075415424, + "grad_norm": 19.875, + "learning_rate": 9.436442140513341e-08, + "loss": 0.7492, + "step": 20561 + }, + { + "epoch": 1.752194290583724, + "grad_norm": 15.3125, + "learning_rate": 9.430038316170051e-08, + "loss": 0.9068, + "step": 20562 + }, + { + "epoch": 1.752279505752024, + "grad_norm": 14.375, + "learning_rate": 9.423636580292189e-08, + "loss": 0.3273, + "step": 20563 + }, + { + "epoch": 1.752364720920324, + "grad_norm": 14.0625, + "learning_rate": 9.417236932995413e-08, + "loss": 0.34, + "step": 20564 + }, + { + "epoch": 1.7524499360886239, + "grad_norm": 12.75, + "learning_rate": 9.410839374395409e-08, + "loss": 0.559, + "step": 20565 + }, + { + "epoch": 1.7525351512569238, + "grad_norm": 16.125, + "learning_rate": 9.404443904607764e-08, + "loss": 0.7444, + "step": 20566 + }, + { + "epoch": 1.7526203664252238, + "grad_norm": 15.3125, + "learning_rate": 9.398050523748065e-08, + "loss": 0.6814, + "step": 20567 + }, + { + "epoch": 1.7527055815935237, + "grad_norm": 16.875, + "learning_rate": 9.39165923193186e-08, + "loss": 0.7218, + "step": 20568 + }, + { + "epoch": 1.7527907967618237, + "grad_norm": 15.75, + "learning_rate": 9.385270029274598e-08, + "loss": 0.6058, + "step": 20569 + }, + { + "epoch": 1.7528760119301237, + "grad_norm": 19.125, + "learning_rate": 9.378882915891758e-08, + "loss": 0.8717, + "step": 20570 + }, + { + "epoch": 1.7529612270984236, + "grad_norm": 13.625, + "learning_rate": 9.372497891898771e-08, + "loss": 0.3192, + "step": 20571 + }, + { + "epoch": 1.7530464422667236, + "grad_norm": 14.375, + "learning_rate": 9.366114957411049e-08, + "loss": 0.7445, + "step": 20572 + }, + { + "epoch": 1.7531316574350235, + "grad_norm": 18.5, + "learning_rate": 9.359734112543886e-08, + "loss": 0.7718, + "step": 20573 + }, + { + "epoch": 1.7532168726033235, + "grad_norm": 30.25, + "learning_rate": 9.353355357412608e-08, + "loss": 0.8057, + "step": 20574 + }, + { + "epoch": 1.7533020877716234, + "grad_norm": 20.75, + "learning_rate": 9.346978692132525e-08, + "loss": 0.8301, + "step": 20575 + }, + { + "epoch": 1.7533873029399234, + "grad_norm": 14.5, + "learning_rate": 9.340604116818821e-08, + "loss": 0.7265, + "step": 20576 + }, + { + "epoch": 1.7534725181082234, + "grad_norm": 14.25, + "learning_rate": 9.334231631586684e-08, + "loss": 0.4403, + "step": 20577 + }, + { + "epoch": 1.7535577332765233, + "grad_norm": 13.9375, + "learning_rate": 9.3278612365513e-08, + "loss": 0.6983, + "step": 20578 + }, + { + "epoch": 1.7536429484448233, + "grad_norm": 14.9375, + "learning_rate": 9.32149293182777e-08, + "loss": 0.4973, + "step": 20579 + }, + { + "epoch": 1.7537281636131232, + "grad_norm": 13.9375, + "learning_rate": 9.315126717531222e-08, + "loss": 0.5061, + "step": 20580 + }, + { + "epoch": 1.7538133787814232, + "grad_norm": 11.1875, + "learning_rate": 9.308762593776638e-08, + "loss": 0.301, + "step": 20581 + }, + { + "epoch": 1.7538985939497231, + "grad_norm": 16.75, + "learning_rate": 9.302400560679048e-08, + "loss": 0.5514, + "step": 20582 + }, + { + "epoch": 1.753983809118023, + "grad_norm": 20.125, + "learning_rate": 9.296040618353417e-08, + "loss": 0.6677, + "step": 20583 + }, + { + "epoch": 1.754069024286323, + "grad_norm": 15.375, + "learning_rate": 9.289682766914706e-08, + "loss": 0.651, + "step": 20584 + }, + { + "epoch": 1.754154239454623, + "grad_norm": 11.4375, + "learning_rate": 9.283327006477786e-08, + "loss": 0.3943, + "step": 20585 + }, + { + "epoch": 1.754239454622923, + "grad_norm": 18.125, + "learning_rate": 9.27697333715749e-08, + "loss": 0.5301, + "step": 20586 + }, + { + "epoch": 1.754324669791223, + "grad_norm": 16.5, + "learning_rate": 9.270621759068649e-08, + "loss": 0.9207, + "step": 20587 + }, + { + "epoch": 1.7544098849595229, + "grad_norm": 22.625, + "learning_rate": 9.264272272326083e-08, + "loss": 0.7703, + "step": 20588 + }, + { + "epoch": 1.7544951001278228, + "grad_norm": 14.625, + "learning_rate": 9.257924877044466e-08, + "loss": 0.6064, + "step": 20589 + }, + { + "epoch": 1.7545803152961228, + "grad_norm": 12.375, + "learning_rate": 9.251579573338526e-08, + "loss": 0.7159, + "step": 20590 + }, + { + "epoch": 1.7546655304644228, + "grad_norm": 23.0, + "learning_rate": 9.245236361322948e-08, + "loss": 0.8068, + "step": 20591 + }, + { + "epoch": 1.7547507456327227, + "grad_norm": 22.375, + "learning_rate": 9.238895241112363e-08, + "loss": 0.4809, + "step": 20592 + }, + { + "epoch": 1.7548359608010227, + "grad_norm": 12.9375, + "learning_rate": 9.232556212821362e-08, + "loss": 0.4832, + "step": 20593 + }, + { + "epoch": 1.7549211759693226, + "grad_norm": 15.4375, + "learning_rate": 9.226219276564447e-08, + "loss": 0.6109, + "step": 20594 + }, + { + "epoch": 1.7550063911376226, + "grad_norm": 13.5, + "learning_rate": 9.21988443245618e-08, + "loss": 0.5394, + "step": 20595 + }, + { + "epoch": 1.7550916063059225, + "grad_norm": 16.375, + "learning_rate": 9.213551680611013e-08, + "loss": 0.4728, + "step": 20596 + }, + { + "epoch": 1.7551768214742225, + "grad_norm": 16.625, + "learning_rate": 9.207221021143437e-08, + "loss": 0.7177, + "step": 20597 + }, + { + "epoch": 1.7552620366425224, + "grad_norm": 13.5625, + "learning_rate": 9.200892454167789e-08, + "loss": 0.3143, + "step": 20598 + }, + { + "epoch": 1.7553472518108224, + "grad_norm": 20.625, + "learning_rate": 9.194565979798453e-08, + "loss": 0.87, + "step": 20599 + }, + { + "epoch": 1.7554324669791224, + "grad_norm": 18.0, + "learning_rate": 9.18824159814978e-08, + "loss": 0.8074, + "step": 20600 + }, + { + "epoch": 1.7555176821474223, + "grad_norm": 13.625, + "learning_rate": 9.181919309336027e-08, + "loss": 0.5505, + "step": 20601 + }, + { + "epoch": 1.7556028973157223, + "grad_norm": 13.125, + "learning_rate": 9.175599113471475e-08, + "loss": 0.5577, + "step": 20602 + }, + { + "epoch": 1.7556881124840222, + "grad_norm": 14.8125, + "learning_rate": 9.1692810106703e-08, + "loss": 0.7381, + "step": 20603 + }, + { + "epoch": 1.7557733276523222, + "grad_norm": 18.375, + "learning_rate": 9.162965001046686e-08, + "loss": 0.5871, + "step": 20604 + }, + { + "epoch": 1.7558585428206221, + "grad_norm": 13.1875, + "learning_rate": 9.156651084714808e-08, + "loss": 0.4642, + "step": 20605 + }, + { + "epoch": 1.755943757988922, + "grad_norm": 11.9375, + "learning_rate": 9.15033926178871e-08, + "loss": 0.3324, + "step": 20606 + }, + { + "epoch": 1.756028973157222, + "grad_norm": 13.8125, + "learning_rate": 9.144029532382471e-08, + "loss": 0.7006, + "step": 20607 + }, + { + "epoch": 1.756114188325522, + "grad_norm": 14.5, + "learning_rate": 9.137721896610149e-08, + "loss": 0.4925, + "step": 20608 + }, + { + "epoch": 1.756199403493822, + "grad_norm": 16.375, + "learning_rate": 9.131416354585685e-08, + "loss": 0.6276, + "step": 20609 + }, + { + "epoch": 1.756284618662122, + "grad_norm": 20.0, + "learning_rate": 9.125112906423053e-08, + "loss": 0.9773, + "step": 20610 + }, + { + "epoch": 1.756369833830422, + "grad_norm": 15.1875, + "learning_rate": 9.118811552236124e-08, + "loss": 0.4553, + "step": 20611 + }, + { + "epoch": 1.7564550489987218, + "grad_norm": 22.625, + "learning_rate": 9.112512292138817e-08, + "loss": 0.8546, + "step": 20612 + }, + { + "epoch": 1.7565402641670218, + "grad_norm": 15.9375, + "learning_rate": 9.106215126244932e-08, + "loss": 0.6129, + "step": 20613 + }, + { + "epoch": 1.7566254793353218, + "grad_norm": 18.25, + "learning_rate": 9.09992005466831e-08, + "loss": 0.662, + "step": 20614 + }, + { + "epoch": 1.7567106945036217, + "grad_norm": 10.0625, + "learning_rate": 9.093627077522648e-08, + "loss": 0.2943, + "step": 20615 + }, + { + "epoch": 1.7567959096719217, + "grad_norm": 12.25, + "learning_rate": 9.087336194921734e-08, + "loss": 0.4908, + "step": 20616 + }, + { + "epoch": 1.7568811248402216, + "grad_norm": 12.625, + "learning_rate": 9.081047406979183e-08, + "loss": 0.5093, + "step": 20617 + }, + { + "epoch": 1.7569663400085216, + "grad_norm": 15.3125, + "learning_rate": 9.074760713808694e-08, + "loss": 0.7204, + "step": 20618 + }, + { + "epoch": 1.7570515551768215, + "grad_norm": 11.4375, + "learning_rate": 9.068476115523819e-08, + "loss": 0.3845, + "step": 20619 + }, + { + "epoch": 1.7571367703451215, + "grad_norm": 15.6875, + "learning_rate": 9.062193612238174e-08, + "loss": 0.6329, + "step": 20620 + }, + { + "epoch": 1.7572219855134215, + "grad_norm": 14.875, + "learning_rate": 9.055913204065264e-08, + "loss": 0.5527, + "step": 20621 + }, + { + "epoch": 1.7573072006817214, + "grad_norm": 14.5625, + "learning_rate": 9.049634891118625e-08, + "loss": 0.7298, + "step": 20622 + }, + { + "epoch": 1.7573924158500214, + "grad_norm": 12.0, + "learning_rate": 9.043358673511652e-08, + "loss": 0.4839, + "step": 20623 + }, + { + "epoch": 1.7574776310183213, + "grad_norm": 19.5, + "learning_rate": 9.037084551357809e-08, + "loss": 0.713, + "step": 20624 + }, + { + "epoch": 1.7575628461866213, + "grad_norm": 11.9375, + "learning_rate": 9.03081252477045e-08, + "loss": 0.4115, + "step": 20625 + }, + { + "epoch": 1.7576480613549212, + "grad_norm": 15.75, + "learning_rate": 9.024542593862915e-08, + "loss": 0.4372, + "step": 20626 + }, + { + "epoch": 1.7577332765232212, + "grad_norm": 17.75, + "learning_rate": 9.018274758748543e-08, + "loss": 0.6959, + "step": 20627 + }, + { + "epoch": 1.7578184916915212, + "grad_norm": 13.875, + "learning_rate": 9.012009019540552e-08, + "loss": 0.6051, + "step": 20628 + }, + { + "epoch": 1.7579037068598211, + "grad_norm": 13.1875, + "learning_rate": 9.005745376352182e-08, + "loss": 0.5824, + "step": 20629 + }, + { + "epoch": 1.757988922028121, + "grad_norm": 14.6875, + "learning_rate": 8.999483829296649e-08, + "loss": 0.7291, + "step": 20630 + }, + { + "epoch": 1.758074137196421, + "grad_norm": 22.625, + "learning_rate": 8.993224378487111e-08, + "loss": 0.6835, + "step": 20631 + }, + { + "epoch": 1.758159352364721, + "grad_norm": 24.125, + "learning_rate": 8.98696702403666e-08, + "loss": 0.7427, + "step": 20632 + }, + { + "epoch": 1.758244567533021, + "grad_norm": 16.75, + "learning_rate": 8.980711766058345e-08, + "loss": 0.6333, + "step": 20633 + }, + { + "epoch": 1.758329782701321, + "grad_norm": 21.875, + "learning_rate": 8.97445860466524e-08, + "loss": 0.735, + "step": 20634 + }, + { + "epoch": 1.7584149978696209, + "grad_norm": 12.75, + "learning_rate": 8.968207539970367e-08, + "loss": 0.4113, + "step": 20635 + }, + { + "epoch": 1.7585002130379208, + "grad_norm": 17.5, + "learning_rate": 8.961958572086635e-08, + "loss": 0.4289, + "step": 20636 + }, + { + "epoch": 1.7585854282062208, + "grad_norm": 15.375, + "learning_rate": 8.955711701127009e-08, + "loss": 0.7262, + "step": 20637 + }, + { + "epoch": 1.7586706433745207, + "grad_norm": 14.125, + "learning_rate": 8.949466927204358e-08, + "loss": 0.5647, + "step": 20638 + }, + { + "epoch": 1.7587558585428207, + "grad_norm": 13.25, + "learning_rate": 8.943224250431551e-08, + "loss": 0.5478, + "step": 20639 + }, + { + "epoch": 1.7588410737111206, + "grad_norm": 13.625, + "learning_rate": 8.93698367092137e-08, + "loss": 0.6549, + "step": 20640 + }, + { + "epoch": 1.7589262888794206, + "grad_norm": 23.25, + "learning_rate": 8.93074518878663e-08, + "loss": 0.8717, + "step": 20641 + }, + { + "epoch": 1.7590115040477206, + "grad_norm": 14.9375, + "learning_rate": 8.924508804140017e-08, + "loss": 0.5585, + "step": 20642 + }, + { + "epoch": 1.7590967192160205, + "grad_norm": 16.125, + "learning_rate": 8.918274517094247e-08, + "loss": 0.6258, + "step": 20643 + }, + { + "epoch": 1.7591819343843205, + "grad_norm": 16.375, + "learning_rate": 8.912042327762008e-08, + "loss": 0.4924, + "step": 20644 + }, + { + "epoch": 1.7592671495526204, + "grad_norm": 17.125, + "learning_rate": 8.90581223625589e-08, + "loss": 0.6396, + "step": 20645 + }, + { + "epoch": 1.7593523647209204, + "grad_norm": 20.5, + "learning_rate": 8.89958424268847e-08, + "loss": 0.6827, + "step": 20646 + }, + { + "epoch": 1.7594375798892203, + "grad_norm": 14.5, + "learning_rate": 8.893358347172324e-08, + "loss": 0.5003, + "step": 20647 + }, + { + "epoch": 1.7595227950575203, + "grad_norm": 10.1875, + "learning_rate": 8.887134549819959e-08, + "loss": 0.1869, + "step": 20648 + }, + { + "epoch": 1.7596080102258203, + "grad_norm": 29.625, + "learning_rate": 8.880912850743842e-08, + "loss": 1.0387, + "step": 20649 + }, + { + "epoch": 1.7596932253941202, + "grad_norm": 13.6875, + "learning_rate": 8.874693250056369e-08, + "loss": 0.6402, + "step": 20650 + }, + { + "epoch": 1.7597784405624202, + "grad_norm": 14.625, + "learning_rate": 8.868475747869962e-08, + "loss": 0.4866, + "step": 20651 + }, + { + "epoch": 1.7598636557307201, + "grad_norm": 10.75, + "learning_rate": 8.862260344297006e-08, + "loss": 0.2299, + "step": 20652 + }, + { + "epoch": 1.75994887089902, + "grad_norm": 19.0, + "learning_rate": 8.856047039449755e-08, + "loss": 0.7385, + "step": 20653 + }, + { + "epoch": 1.76003408606732, + "grad_norm": 13.625, + "learning_rate": 8.84983583344054e-08, + "loss": 0.5489, + "step": 20654 + }, + { + "epoch": 1.76011930123562, + "grad_norm": 18.125, + "learning_rate": 8.843626726381588e-08, + "loss": 0.9173, + "step": 20655 + }, + { + "epoch": 1.76020451640392, + "grad_norm": 25.25, + "learning_rate": 8.837419718385116e-08, + "loss": 0.7126, + "step": 20656 + }, + { + "epoch": 1.76028973157222, + "grad_norm": 10.375, + "learning_rate": 8.831214809563296e-08, + "loss": 0.2193, + "step": 20657 + }, + { + "epoch": 1.7603749467405199, + "grad_norm": 13.8125, + "learning_rate": 8.825012000028207e-08, + "loss": 0.3984, + "step": 20658 + }, + { + "epoch": 1.7604601619088198, + "grad_norm": 16.125, + "learning_rate": 8.818811289891982e-08, + "loss": 0.7087, + "step": 20659 + }, + { + "epoch": 1.7605453770771198, + "grad_norm": 17.25, + "learning_rate": 8.812612679266655e-08, + "loss": 0.8959, + "step": 20660 + }, + { + "epoch": 1.7606305922454197, + "grad_norm": 16.0, + "learning_rate": 8.80641616826429e-08, + "loss": 0.7694, + "step": 20661 + }, + { + "epoch": 1.7607158074137197, + "grad_norm": 25.25, + "learning_rate": 8.800221756996796e-08, + "loss": 0.9626, + "step": 20662 + }, + { + "epoch": 1.7608010225820196, + "grad_norm": 19.375, + "learning_rate": 8.794029445576142e-08, + "loss": 0.9632, + "step": 20663 + }, + { + "epoch": 1.7608862377503196, + "grad_norm": 18.375, + "learning_rate": 8.787839234114234e-08, + "loss": 0.6846, + "step": 20664 + }, + { + "epoch": 1.7609714529186196, + "grad_norm": 15.25, + "learning_rate": 8.78165112272296e-08, + "loss": 0.3888, + "step": 20665 + }, + { + "epoch": 1.7610566680869195, + "grad_norm": 14.9375, + "learning_rate": 8.775465111514087e-08, + "loss": 0.5192, + "step": 20666 + }, + { + "epoch": 1.7611418832552195, + "grad_norm": 21.625, + "learning_rate": 8.76928120059943e-08, + "loss": 0.5423, + "step": 20667 + }, + { + "epoch": 1.7612270984235194, + "grad_norm": 15.3125, + "learning_rate": 8.763099390090749e-08, + "loss": 0.6201, + "step": 20668 + }, + { + "epoch": 1.7613123135918194, + "grad_norm": 19.375, + "learning_rate": 8.756919680099757e-08, + "loss": 0.7506, + "step": 20669 + }, + { + "epoch": 1.7613975287601193, + "grad_norm": 12.375, + "learning_rate": 8.750742070738103e-08, + "loss": 0.3427, + "step": 20670 + }, + { + "epoch": 1.7614827439284193, + "grad_norm": 13.3125, + "learning_rate": 8.744566562117432e-08, + "loss": 0.3936, + "step": 20671 + }, + { + "epoch": 1.7615679590967193, + "grad_norm": 14.6875, + "learning_rate": 8.73839315434935e-08, + "loss": 0.6321, + "step": 20672 + }, + { + "epoch": 1.7616531742650192, + "grad_norm": 19.75, + "learning_rate": 8.732221847545435e-08, + "loss": 0.8804, + "step": 20673 + }, + { + "epoch": 1.7617383894333192, + "grad_norm": 13.625, + "learning_rate": 8.726052641817182e-08, + "loss": 0.4666, + "step": 20674 + }, + { + "epoch": 1.7618236046016191, + "grad_norm": 15.125, + "learning_rate": 8.719885537276069e-08, + "loss": 0.5986, + "step": 20675 + }, + { + "epoch": 1.761908819769919, + "grad_norm": 9.9375, + "learning_rate": 8.713720534033548e-08, + "loss": 0.2478, + "step": 20676 + }, + { + "epoch": 1.761994034938219, + "grad_norm": 26.625, + "learning_rate": 8.707557632201032e-08, + "loss": 0.7374, + "step": 20677 + }, + { + "epoch": 1.762079250106519, + "grad_norm": 13.0625, + "learning_rate": 8.701396831889919e-08, + "loss": 0.3632, + "step": 20678 + }, + { + "epoch": 1.762164465274819, + "grad_norm": 16.25, + "learning_rate": 8.695238133211478e-08, + "loss": 0.7837, + "step": 20679 + }, + { + "epoch": 1.762249680443119, + "grad_norm": 13.1875, + "learning_rate": 8.689081536277052e-08, + "loss": 0.3609, + "step": 20680 + }, + { + "epoch": 1.7623348956114189, + "grad_norm": 17.125, + "learning_rate": 8.682927041197886e-08, + "loss": 0.667, + "step": 20681 + }, + { + "epoch": 1.7624201107797188, + "grad_norm": 12.125, + "learning_rate": 8.67677464808521e-08, + "loss": 0.3665, + "step": 20682 + }, + { + "epoch": 1.7625053259480188, + "grad_norm": 13.1875, + "learning_rate": 8.670624357050156e-08, + "loss": 0.6295, + "step": 20683 + }, + { + "epoch": 1.7625905411163187, + "grad_norm": 19.25, + "learning_rate": 8.664476168203915e-08, + "loss": 0.9638, + "step": 20684 + }, + { + "epoch": 1.7626757562846187, + "grad_norm": 17.125, + "learning_rate": 8.658330081657562e-08, + "loss": 0.4027, + "step": 20685 + }, + { + "epoch": 1.7627609714529187, + "grad_norm": 17.0, + "learning_rate": 8.652186097522191e-08, + "loss": 0.4941, + "step": 20686 + }, + { + "epoch": 1.7628461866212186, + "grad_norm": 13.8125, + "learning_rate": 8.646044215908796e-08, + "loss": 0.4105, + "step": 20687 + }, + { + "epoch": 1.7629314017895186, + "grad_norm": 17.125, + "learning_rate": 8.639904436928387e-08, + "loss": 0.6597, + "step": 20688 + }, + { + "epoch": 1.7630166169578185, + "grad_norm": 15.1875, + "learning_rate": 8.633766760691942e-08, + "loss": 0.639, + "step": 20689 + }, + { + "epoch": 1.7631018321261185, + "grad_norm": 13.875, + "learning_rate": 8.627631187310318e-08, + "loss": 0.6132, + "step": 20690 + }, + { + "epoch": 1.7631870472944184, + "grad_norm": 25.75, + "learning_rate": 8.621497716894439e-08, + "loss": 0.6651, + "step": 20691 + }, + { + "epoch": 1.7632722624627184, + "grad_norm": 11.0, + "learning_rate": 8.615366349555107e-08, + "loss": 0.2936, + "step": 20692 + }, + { + "epoch": 1.7633574776310184, + "grad_norm": 18.75, + "learning_rate": 8.609237085403135e-08, + "loss": 1.0493, + "step": 20693 + }, + { + "epoch": 1.7634426927993183, + "grad_norm": 14.5625, + "learning_rate": 8.603109924549296e-08, + "loss": 0.6845, + "step": 20694 + }, + { + "epoch": 1.7635279079676183, + "grad_norm": 12.0625, + "learning_rate": 8.596984867104322e-08, + "loss": 0.3444, + "step": 20695 + }, + { + "epoch": 1.7636131231359182, + "grad_norm": 21.875, + "learning_rate": 8.590861913178859e-08, + "loss": 1.1164, + "step": 20696 + }, + { + "epoch": 1.7636983383042182, + "grad_norm": 8.625, + "learning_rate": 8.584741062883609e-08, + "loss": 0.4481, + "step": 20697 + }, + { + "epoch": 1.7637835534725181, + "grad_norm": 12.1875, + "learning_rate": 8.578622316329125e-08, + "loss": 0.2588, + "step": 20698 + }, + { + "epoch": 1.763868768640818, + "grad_norm": 13.875, + "learning_rate": 8.572505673626025e-08, + "loss": 0.566, + "step": 20699 + }, + { + "epoch": 1.763953983809118, + "grad_norm": 23.0, + "learning_rate": 8.566391134884819e-08, + "loss": 0.6733, + "step": 20700 + }, + { + "epoch": 1.764039198977418, + "grad_norm": 14.3125, + "learning_rate": 8.560278700216001e-08, + "loss": 0.9824, + "step": 20701 + }, + { + "epoch": 1.764124414145718, + "grad_norm": 19.375, + "learning_rate": 8.554168369730026e-08, + "loss": 1.0365, + "step": 20702 + }, + { + "epoch": 1.764209629314018, + "grad_norm": 13.6875, + "learning_rate": 8.548060143537359e-08, + "loss": 0.4187, + "step": 20703 + }, + { + "epoch": 1.7642948444823179, + "grad_norm": 13.875, + "learning_rate": 8.541954021748317e-08, + "loss": 0.3648, + "step": 20704 + }, + { + "epoch": 1.7643800596506178, + "grad_norm": 12.5625, + "learning_rate": 8.535850004473267e-08, + "loss": 0.5229, + "step": 20705 + }, + { + "epoch": 1.7644652748189178, + "grad_norm": 19.25, + "learning_rate": 8.529748091822554e-08, + "loss": 0.5447, + "step": 20706 + }, + { + "epoch": 1.7645504899872178, + "grad_norm": 11.25, + "learning_rate": 8.523648283906394e-08, + "loss": 0.3352, + "step": 20707 + }, + { + "epoch": 1.7646357051555177, + "grad_norm": 16.75, + "learning_rate": 8.517550580835046e-08, + "loss": 0.6091, + "step": 20708 + }, + { + "epoch": 1.7647209203238177, + "grad_norm": 17.125, + "learning_rate": 8.511454982718672e-08, + "loss": 0.561, + "step": 20709 + }, + { + "epoch": 1.7648061354921176, + "grad_norm": 22.375, + "learning_rate": 8.50536148966745e-08, + "loss": 0.9402, + "step": 20710 + }, + { + "epoch": 1.7648913506604176, + "grad_norm": 26.75, + "learning_rate": 8.499270101791496e-08, + "loss": 0.8066, + "step": 20711 + }, + { + "epoch": 1.7649765658287175, + "grad_norm": 15.125, + "learning_rate": 8.493180819200891e-08, + "loss": 0.3866, + "step": 20712 + }, + { + "epoch": 1.7650617809970175, + "grad_norm": 12.1875, + "learning_rate": 8.487093642005645e-08, + "loss": 0.3368, + "step": 20713 + }, + { + "epoch": 1.7651469961653175, + "grad_norm": 9.875, + "learning_rate": 8.481008570315807e-08, + "loss": 0.2732, + "step": 20714 + }, + { + "epoch": 1.7652322113336174, + "grad_norm": 14.875, + "learning_rate": 8.474925604241291e-08, + "loss": 0.4356, + "step": 20715 + }, + { + "epoch": 1.7653174265019174, + "grad_norm": 20.125, + "learning_rate": 8.468844743892063e-08, + "loss": 0.8607, + "step": 20716 + }, + { + "epoch": 1.7654026416702173, + "grad_norm": 16.875, + "learning_rate": 8.462765989377967e-08, + "loss": 0.5252, + "step": 20717 + }, + { + "epoch": 1.7654878568385173, + "grad_norm": 16.375, + "learning_rate": 8.456689340808885e-08, + "loss": 0.3658, + "step": 20718 + }, + { + "epoch": 1.7655730720068172, + "grad_norm": 30.875, + "learning_rate": 8.450614798294607e-08, + "loss": 0.6712, + "step": 20719 + }, + { + "epoch": 1.7656582871751172, + "grad_norm": 16.5, + "learning_rate": 8.44454236194496e-08, + "loss": 0.8908, + "step": 20720 + }, + { + "epoch": 1.7657435023434171, + "grad_norm": 25.375, + "learning_rate": 8.438472031869607e-08, + "loss": 0.9415, + "step": 20721 + }, + { + "epoch": 1.765828717511717, + "grad_norm": 19.625, + "learning_rate": 8.432403808178294e-08, + "loss": 0.7745, + "step": 20722 + }, + { + "epoch": 1.765913932680017, + "grad_norm": 17.75, + "learning_rate": 8.42633769098064e-08, + "loss": 0.7632, + "step": 20723 + }, + { + "epoch": 1.765999147848317, + "grad_norm": 17.0, + "learning_rate": 8.420273680386296e-08, + "loss": 0.716, + "step": 20724 + }, + { + "epoch": 1.766084363016617, + "grad_norm": 16.875, + "learning_rate": 8.414211776504852e-08, + "loss": 0.5491, + "step": 20725 + }, + { + "epoch": 1.766169578184917, + "grad_norm": 12.0625, + "learning_rate": 8.408151979445805e-08, + "loss": 0.2904, + "step": 20726 + }, + { + "epoch": 1.766254793353217, + "grad_norm": 18.0, + "learning_rate": 8.402094289318707e-08, + "loss": 0.8032, + "step": 20727 + }, + { + "epoch": 1.7663400085215168, + "grad_norm": 19.25, + "learning_rate": 8.39603870623304e-08, + "loss": 0.7255, + "step": 20728 + }, + { + "epoch": 1.7664252236898168, + "grad_norm": 13.0, + "learning_rate": 8.389985230298172e-08, + "loss": 0.5369, + "step": 20729 + }, + { + "epoch": 1.7665104388581168, + "grad_norm": 25.125, + "learning_rate": 8.383933861623562e-08, + "loss": 0.7596, + "step": 20730 + }, + { + "epoch": 1.7665956540264167, + "grad_norm": 16.125, + "learning_rate": 8.37788460031852e-08, + "loss": 0.3899, + "step": 20731 + }, + { + "epoch": 1.7666808691947167, + "grad_norm": 16.375, + "learning_rate": 8.371837446492364e-08, + "loss": 0.4507, + "step": 20732 + }, + { + "epoch": 1.7667660843630166, + "grad_norm": 19.75, + "learning_rate": 8.36579240025441e-08, + "loss": 0.838, + "step": 20733 + }, + { + "epoch": 1.7668512995313166, + "grad_norm": 16.0, + "learning_rate": 8.359749461713849e-08, + "loss": 0.4838, + "step": 20734 + }, + { + "epoch": 1.7669365146996165, + "grad_norm": 15.9375, + "learning_rate": 8.353708630979895e-08, + "loss": 0.4212, + "step": 20735 + }, + { + "epoch": 1.7670217298679165, + "grad_norm": 15.5625, + "learning_rate": 8.347669908161743e-08, + "loss": 0.5375, + "step": 20736 + }, + { + "epoch": 1.7671069450362165, + "grad_norm": 17.125, + "learning_rate": 8.341633293368512e-08, + "loss": 0.5857, + "step": 20737 + }, + { + "epoch": 1.7671921602045164, + "grad_norm": 17.75, + "learning_rate": 8.335598786709267e-08, + "loss": 0.7174, + "step": 20738 + }, + { + "epoch": 1.7672773753728164, + "grad_norm": 13.4375, + "learning_rate": 8.329566388293062e-08, + "loss": 0.3059, + "step": 20739 + }, + { + "epoch": 1.7673625905411163, + "grad_norm": 12.1875, + "learning_rate": 8.323536098228905e-08, + "loss": 0.3925, + "step": 20740 + }, + { + "epoch": 1.7674478057094163, + "grad_norm": 12.875, + "learning_rate": 8.317507916625778e-08, + "loss": 0.4589, + "step": 20741 + }, + { + "epoch": 1.7675330208777162, + "grad_norm": 15.75, + "learning_rate": 8.311481843592637e-08, + "loss": 0.8686, + "step": 20742 + }, + { + "epoch": 1.7676182360460162, + "grad_norm": 23.75, + "learning_rate": 8.305457879238338e-08, + "loss": 1.1705, + "step": 20743 + }, + { + "epoch": 1.7677034512143162, + "grad_norm": 19.375, + "learning_rate": 8.299436023671754e-08, + "loss": 1.0137, + "step": 20744 + }, + { + "epoch": 1.7677886663826161, + "grad_norm": 15.0625, + "learning_rate": 8.29341627700174e-08, + "loss": 0.388, + "step": 20745 + }, + { + "epoch": 1.767873881550916, + "grad_norm": 12.3125, + "learning_rate": 8.287398639337058e-08, + "loss": 0.3021, + "step": 20746 + }, + { + "epoch": 1.767959096719216, + "grad_norm": 13.6875, + "learning_rate": 8.281383110786411e-08, + "loss": 0.3623, + "step": 20747 + }, + { + "epoch": 1.768044311887516, + "grad_norm": 16.375, + "learning_rate": 8.275369691458562e-08, + "loss": 0.8024, + "step": 20748 + }, + { + "epoch": 1.768129527055816, + "grad_norm": 18.25, + "learning_rate": 8.269358381462145e-08, + "loss": 0.956, + "step": 20749 + }, + { + "epoch": 1.768214742224116, + "grad_norm": 13.3125, + "learning_rate": 8.263349180905821e-08, + "loss": 0.5184, + "step": 20750 + }, + { + "epoch": 1.7682999573924159, + "grad_norm": 31.0, + "learning_rate": 8.257342089898146e-08, + "loss": 0.8549, + "step": 20751 + }, + { + "epoch": 1.7683851725607158, + "grad_norm": 11.9375, + "learning_rate": 8.251337108547697e-08, + "loss": 0.223, + "step": 20752 + }, + { + "epoch": 1.7684703877290158, + "grad_norm": 14.75, + "learning_rate": 8.245334236962985e-08, + "loss": 0.609, + "step": 20753 + }, + { + "epoch": 1.7685556028973157, + "grad_norm": 12.625, + "learning_rate": 8.239333475252506e-08, + "loss": 0.3914, + "step": 20754 + }, + { + "epoch": 1.7686408180656157, + "grad_norm": 21.25, + "learning_rate": 8.233334823524689e-08, + "loss": 0.4608, + "step": 20755 + }, + { + "epoch": 1.7687260332339156, + "grad_norm": 28.25, + "learning_rate": 8.227338281887904e-08, + "loss": 0.7798, + "step": 20756 + }, + { + "epoch": 1.7688112484022156, + "grad_norm": 18.125, + "learning_rate": 8.22134385045055e-08, + "loss": 0.626, + "step": 20757 + }, + { + "epoch": 1.7688964635705156, + "grad_norm": 10.5625, + "learning_rate": 8.215351529320958e-08, + "loss": 0.2872, + "step": 20758 + }, + { + "epoch": 1.7689816787388155, + "grad_norm": 13.25, + "learning_rate": 8.209361318607376e-08, + "loss": 0.4227, + "step": 20759 + }, + { + "epoch": 1.7690668939071155, + "grad_norm": 16.875, + "learning_rate": 8.203373218418076e-08, + "loss": 0.83, + "step": 20760 + }, + { + "epoch": 1.7691521090754154, + "grad_norm": 11.125, + "learning_rate": 8.197387228861278e-08, + "loss": 0.309, + "step": 20761 + }, + { + "epoch": 1.7692373242437154, + "grad_norm": 12.5625, + "learning_rate": 8.19140335004516e-08, + "loss": 0.3202, + "step": 20762 + }, + { + "epoch": 1.7693225394120153, + "grad_norm": 13.1875, + "learning_rate": 8.185421582077843e-08, + "loss": 0.3868, + "step": 20763 + }, + { + "epoch": 1.7694077545803153, + "grad_norm": 13.125, + "learning_rate": 8.179441925067408e-08, + "loss": 0.4064, + "step": 20764 + }, + { + "epoch": 1.7694929697486153, + "grad_norm": 12.625, + "learning_rate": 8.17346437912192e-08, + "loss": 0.3826, + "step": 20765 + }, + { + "epoch": 1.7695781849169152, + "grad_norm": 18.875, + "learning_rate": 8.16748894434942e-08, + "loss": 0.4824, + "step": 20766 + }, + { + "epoch": 1.7696634000852152, + "grad_norm": 10.75, + "learning_rate": 8.161515620857902e-08, + "loss": 0.3649, + "step": 20767 + }, + { + "epoch": 1.7697486152535151, + "grad_norm": 19.25, + "learning_rate": 8.155544408755253e-08, + "loss": 0.3717, + "step": 20768 + }, + { + "epoch": 1.769833830421815, + "grad_norm": 12.0625, + "learning_rate": 8.149575308149429e-08, + "loss": 0.3922, + "step": 20769 + }, + { + "epoch": 1.769919045590115, + "grad_norm": 10.8125, + "learning_rate": 8.143608319148274e-08, + "loss": 0.2052, + "step": 20770 + }, + { + "epoch": 1.770004260758415, + "grad_norm": 14.0, + "learning_rate": 8.137643441859632e-08, + "loss": 0.4265, + "step": 20771 + }, + { + "epoch": 1.770089475926715, + "grad_norm": 22.375, + "learning_rate": 8.131680676391307e-08, + "loss": 1.121, + "step": 20772 + }, + { + "epoch": 1.770174691095015, + "grad_norm": 15.9375, + "learning_rate": 8.125720022851003e-08, + "loss": 0.7936, + "step": 20773 + }, + { + "epoch": 1.7702599062633149, + "grad_norm": 15.6875, + "learning_rate": 8.119761481346466e-08, + "loss": 0.5297, + "step": 20774 + }, + { + "epoch": 1.7703451214316148, + "grad_norm": 16.75, + "learning_rate": 8.11380505198539e-08, + "loss": 0.6804, + "step": 20775 + }, + { + "epoch": 1.7704303365999148, + "grad_norm": 13.75, + "learning_rate": 8.10785073487537e-08, + "loss": 0.4745, + "step": 20776 + }, + { + "epoch": 1.7705155517682147, + "grad_norm": 19.75, + "learning_rate": 8.101898530124039e-08, + "loss": 0.6839, + "step": 20777 + }, + { + "epoch": 1.7706007669365147, + "grad_norm": 23.0, + "learning_rate": 8.095948437838952e-08, + "loss": 0.705, + "step": 20778 + }, + { + "epoch": 1.7706859821048146, + "grad_norm": 13.0, + "learning_rate": 8.090000458127647e-08, + "loss": 0.3285, + "step": 20779 + }, + { + "epoch": 1.7707711972731146, + "grad_norm": 17.125, + "learning_rate": 8.084054591097609e-08, + "loss": 0.676, + "step": 20780 + }, + { + "epoch": 1.7708564124414146, + "grad_norm": 11.1875, + "learning_rate": 8.078110836856237e-08, + "loss": 0.3991, + "step": 20781 + }, + { + "epoch": 1.7709416276097145, + "grad_norm": 46.25, + "learning_rate": 8.072169195510987e-08, + "loss": 0.4596, + "step": 20782 + }, + { + "epoch": 1.7710268427780145, + "grad_norm": 30.625, + "learning_rate": 8.066229667169218e-08, + "loss": 0.8189, + "step": 20783 + }, + { + "epoch": 1.7711120579463144, + "grad_norm": 17.5, + "learning_rate": 8.060292251938273e-08, + "loss": 0.8743, + "step": 20784 + }, + { + "epoch": 1.7711972731146144, + "grad_norm": 21.5, + "learning_rate": 8.054356949925429e-08, + "loss": 1.1527, + "step": 20785 + }, + { + "epoch": 1.7712824882829143, + "grad_norm": 21.625, + "learning_rate": 8.048423761237947e-08, + "loss": 0.6426, + "step": 20786 + }, + { + "epoch": 1.7713677034512143, + "grad_norm": 19.25, + "learning_rate": 8.042492685983074e-08, + "loss": 0.8161, + "step": 20787 + }, + { + "epoch": 1.7714529186195143, + "grad_norm": 15.125, + "learning_rate": 8.036563724267976e-08, + "loss": 0.393, + "step": 20788 + }, + { + "epoch": 1.7715381337878142, + "grad_norm": 21.875, + "learning_rate": 8.03063687619976e-08, + "loss": 0.5933, + "step": 20789 + }, + { + "epoch": 1.7716233489561142, + "grad_norm": 12.8125, + "learning_rate": 8.024712141885563e-08, + "loss": 0.5284, + "step": 20790 + }, + { + "epoch": 1.7717085641244141, + "grad_norm": 18.0, + "learning_rate": 8.018789521432438e-08, + "loss": 0.7969, + "step": 20791 + }, + { + "epoch": 1.771793779292714, + "grad_norm": 24.125, + "learning_rate": 8.012869014947453e-08, + "loss": 0.585, + "step": 20792 + }, + { + "epoch": 1.771878994461014, + "grad_norm": 14.625, + "learning_rate": 8.006950622537535e-08, + "loss": 0.524, + "step": 20793 + }, + { + "epoch": 1.771964209629314, + "grad_norm": 16.75, + "learning_rate": 8.001034344309655e-08, + "loss": 0.8657, + "step": 20794 + }, + { + "epoch": 1.772049424797614, + "grad_norm": 13.5, + "learning_rate": 7.995120180370769e-08, + "loss": 0.5076, + "step": 20795 + }, + { + "epoch": 1.772134639965914, + "grad_norm": 18.0, + "learning_rate": 7.989208130827694e-08, + "loss": 0.5252, + "step": 20796 + }, + { + "epoch": 1.7722198551342139, + "grad_norm": 13.5625, + "learning_rate": 7.983298195787317e-08, + "loss": 0.5297, + "step": 20797 + }, + { + "epoch": 1.7723050703025138, + "grad_norm": 15.625, + "learning_rate": 7.977390375356387e-08, + "loss": 0.5128, + "step": 20798 + }, + { + "epoch": 1.7723902854708138, + "grad_norm": 13.4375, + "learning_rate": 7.97148466964169e-08, + "loss": 0.3772, + "step": 20799 + }, + { + "epoch": 1.7724755006391137, + "grad_norm": 14.5, + "learning_rate": 7.965581078749938e-08, + "loss": 0.5202, + "step": 20800 + }, + { + "epoch": 1.7725607158074137, + "grad_norm": 18.625, + "learning_rate": 7.959679602787846e-08, + "loss": 0.6212, + "step": 20801 + }, + { + "epoch": 1.7726459309757137, + "grad_norm": 12.4375, + "learning_rate": 7.953780241862024e-08, + "loss": 0.5291, + "step": 20802 + }, + { + "epoch": 1.7727311461440136, + "grad_norm": 13.9375, + "learning_rate": 7.947882996079112e-08, + "loss": 0.5405, + "step": 20803 + }, + { + "epoch": 1.7728163613123136, + "grad_norm": 31.0, + "learning_rate": 7.941987865545633e-08, + "loss": 0.9705, + "step": 20804 + }, + { + "epoch": 1.7729015764806135, + "grad_norm": 17.625, + "learning_rate": 7.93609485036817e-08, + "loss": 0.5886, + "step": 20805 + }, + { + "epoch": 1.7729867916489135, + "grad_norm": 11.1875, + "learning_rate": 7.930203950653165e-08, + "loss": 0.1844, + "step": 20806 + }, + { + "epoch": 1.7730720068172134, + "grad_norm": 16.5, + "learning_rate": 7.924315166507102e-08, + "loss": 0.7846, + "step": 20807 + }, + { + "epoch": 1.7731572219855134, + "grad_norm": 24.5, + "learning_rate": 7.918428498036395e-08, + "loss": 0.853, + "step": 20808 + }, + { + "epoch": 1.7732424371538134, + "grad_norm": 20.875, + "learning_rate": 7.912543945347434e-08, + "loss": 0.4868, + "step": 20809 + }, + { + "epoch": 1.7733276523221133, + "grad_norm": 16.25, + "learning_rate": 7.906661508546521e-08, + "loss": 0.4852, + "step": 20810 + }, + { + "epoch": 1.7734128674904133, + "grad_norm": 14.0, + "learning_rate": 7.900781187740001e-08, + "loss": 0.5662, + "step": 20811 + }, + { + "epoch": 1.7734980826587132, + "grad_norm": 17.625, + "learning_rate": 7.894902983034108e-08, + "loss": 0.6124, + "step": 20812 + }, + { + "epoch": 1.7735832978270132, + "grad_norm": 15.125, + "learning_rate": 7.889026894535065e-08, + "loss": 0.2674, + "step": 20813 + }, + { + "epoch": 1.7736685129953131, + "grad_norm": 19.0, + "learning_rate": 7.883152922349091e-08, + "loss": 0.9484, + "step": 20814 + }, + { + "epoch": 1.773753728163613, + "grad_norm": 14.625, + "learning_rate": 7.877281066582295e-08, + "loss": 0.4141, + "step": 20815 + }, + { + "epoch": 1.773838943331913, + "grad_norm": 15.0625, + "learning_rate": 7.871411327340817e-08, + "loss": 0.6194, + "step": 20816 + }, + { + "epoch": 1.773924158500213, + "grad_norm": 16.5, + "learning_rate": 7.865543704730708e-08, + "loss": 0.6546, + "step": 20817 + }, + { + "epoch": 1.774009373668513, + "grad_norm": 13.0625, + "learning_rate": 7.859678198858037e-08, + "loss": 0.3428, + "step": 20818 + }, + { + "epoch": 1.774094588836813, + "grad_norm": 18.75, + "learning_rate": 7.853814809828775e-08, + "loss": 0.8422, + "step": 20819 + }, + { + "epoch": 1.7741798040051129, + "grad_norm": 13.125, + "learning_rate": 7.847953537748851e-08, + "loss": 0.8881, + "step": 20820 + }, + { + "epoch": 1.7742650191734128, + "grad_norm": 13.0625, + "learning_rate": 7.842094382724222e-08, + "loss": 0.5303, + "step": 20821 + }, + { + "epoch": 1.7743502343417128, + "grad_norm": 13.75, + "learning_rate": 7.836237344860775e-08, + "loss": 0.5254, + "step": 20822 + }, + { + "epoch": 1.7744354495100128, + "grad_norm": 16.25, + "learning_rate": 7.830382424264316e-08, + "loss": 0.7206, + "step": 20823 + }, + { + "epoch": 1.7745206646783127, + "grad_norm": 19.625, + "learning_rate": 7.824529621040661e-08, + "loss": 0.6966, + "step": 20824 + }, + { + "epoch": 1.7746058798466127, + "grad_norm": 14.875, + "learning_rate": 7.818678935295601e-08, + "loss": 0.6469, + "step": 20825 + }, + { + "epoch": 1.7746910950149126, + "grad_norm": 17.75, + "learning_rate": 7.812830367134857e-08, + "loss": 0.785, + "step": 20826 + }, + { + "epoch": 1.7747763101832126, + "grad_norm": 14.3125, + "learning_rate": 7.80698391666411e-08, + "loss": 0.6313, + "step": 20827 + }, + { + "epoch": 1.7748615253515125, + "grad_norm": 12.9375, + "learning_rate": 7.801139583988996e-08, + "loss": 0.5272, + "step": 20828 + }, + { + "epoch": 1.7749467405198125, + "grad_norm": 14.25, + "learning_rate": 7.795297369215124e-08, + "loss": 0.413, + "step": 20829 + }, + { + "epoch": 1.7750319556881125, + "grad_norm": 15.8125, + "learning_rate": 7.789457272448106e-08, + "loss": 0.6186, + "step": 20830 + }, + { + "epoch": 1.7751171708564124, + "grad_norm": 14.625, + "learning_rate": 7.783619293793482e-08, + "loss": 0.5406, + "step": 20831 + }, + { + "epoch": 1.7752023860247124, + "grad_norm": 10.875, + "learning_rate": 7.777783433356695e-08, + "loss": 0.3311, + "step": 20832 + }, + { + "epoch": 1.7752876011930123, + "grad_norm": 13.3125, + "learning_rate": 7.771949691243244e-08, + "loss": 0.512, + "step": 20833 + }, + { + "epoch": 1.7753728163613123, + "grad_norm": 20.5, + "learning_rate": 7.766118067558545e-08, + "loss": 0.3655, + "step": 20834 + }, + { + "epoch": 1.7754580315296122, + "grad_norm": 11.9375, + "learning_rate": 7.760288562408014e-08, + "loss": 0.7185, + "step": 20835 + }, + { + "epoch": 1.7755432466979122, + "grad_norm": 21.75, + "learning_rate": 7.754461175896926e-08, + "loss": 0.8764, + "step": 20836 + }, + { + "epoch": 1.7756284618662121, + "grad_norm": 15.375, + "learning_rate": 7.748635908130616e-08, + "loss": 0.6555, + "step": 20837 + }, + { + "epoch": 1.775713677034512, + "grad_norm": 17.5, + "learning_rate": 7.742812759214375e-08, + "loss": 0.4821, + "step": 20838 + }, + { + "epoch": 1.775798892202812, + "grad_norm": 17.125, + "learning_rate": 7.736991729253434e-08, + "loss": 0.7515, + "step": 20839 + }, + { + "epoch": 1.775884107371112, + "grad_norm": 16.125, + "learning_rate": 7.73117281835295e-08, + "loss": 0.7898, + "step": 20840 + }, + { + "epoch": 1.775969322539412, + "grad_norm": 14.9375, + "learning_rate": 7.725356026618086e-08, + "loss": 0.7704, + "step": 20841 + }, + { + "epoch": 1.776054537707712, + "grad_norm": 18.125, + "learning_rate": 7.719541354153981e-08, + "loss": 0.5937, + "step": 20842 + }, + { + "epoch": 1.776139752876012, + "grad_norm": 13.6875, + "learning_rate": 7.713728801065718e-08, + "loss": 0.419, + "step": 20843 + }, + { + "epoch": 1.7762249680443118, + "grad_norm": 12.125, + "learning_rate": 7.707918367458309e-08, + "loss": 0.3469, + "step": 20844 + }, + { + "epoch": 1.7763101832126118, + "grad_norm": 10.9375, + "learning_rate": 7.70211005343674e-08, + "loss": 0.4084, + "step": 20845 + }, + { + "epoch": 1.7763953983809118, + "grad_norm": 16.375, + "learning_rate": 7.696303859105997e-08, + "loss": 0.5205, + "step": 20846 + }, + { + "epoch": 1.7764806135492117, + "grad_norm": 13.8125, + "learning_rate": 7.690499784570995e-08, + "loss": 0.5272, + "step": 20847 + }, + { + "epoch": 1.7765658287175117, + "grad_norm": 22.25, + "learning_rate": 7.684697829936652e-08, + "loss": 0.8031, + "step": 20848 + }, + { + "epoch": 1.7766510438858116, + "grad_norm": 14.9375, + "learning_rate": 7.678897995307772e-08, + "loss": 0.5615, + "step": 20849 + }, + { + "epoch": 1.7767362590541116, + "grad_norm": 16.75, + "learning_rate": 7.673100280789187e-08, + "loss": 0.4769, + "step": 20850 + }, + { + "epoch": 1.7768214742224115, + "grad_norm": 17.625, + "learning_rate": 7.667304686485661e-08, + "loss": 0.4673, + "step": 20851 + }, + { + "epoch": 1.7769066893907115, + "grad_norm": 18.5, + "learning_rate": 7.661511212501943e-08, + "loss": 0.8326, + "step": 20852 + }, + { + "epoch": 1.7769919045590115, + "grad_norm": 17.125, + "learning_rate": 7.655719858942687e-08, + "loss": 0.7633, + "step": 20853 + }, + { + "epoch": 1.7770771197273114, + "grad_norm": 18.875, + "learning_rate": 7.649930625912572e-08, + "loss": 0.6106, + "step": 20854 + }, + { + "epoch": 1.7771623348956114, + "grad_norm": 29.25, + "learning_rate": 7.644143513516209e-08, + "loss": 0.9202, + "step": 20855 + }, + { + "epoch": 1.7772475500639113, + "grad_norm": 13.625, + "learning_rate": 7.638358521858207e-08, + "loss": 0.4029, + "step": 20856 + }, + { + "epoch": 1.7773327652322113, + "grad_norm": 31.75, + "learning_rate": 7.63257565104307e-08, + "loss": 0.6591, + "step": 20857 + }, + { + "epoch": 1.7774179804005112, + "grad_norm": 27.625, + "learning_rate": 7.626794901175294e-08, + "loss": 0.7683, + "step": 20858 + }, + { + "epoch": 1.7775031955688112, + "grad_norm": 14.5, + "learning_rate": 7.62101627235938e-08, + "loss": 0.6206, + "step": 20859 + }, + { + "epoch": 1.7775884107371112, + "grad_norm": 19.25, + "learning_rate": 7.615239764699744e-08, + "loss": 0.8193, + "step": 20860 + }, + { + "epoch": 1.7776736259054111, + "grad_norm": 18.875, + "learning_rate": 7.609465378300776e-08, + "loss": 0.684, + "step": 20861 + }, + { + "epoch": 1.777758841073711, + "grad_norm": 12.375, + "learning_rate": 7.603693113266793e-08, + "loss": 0.418, + "step": 20862 + }, + { + "epoch": 1.777844056242011, + "grad_norm": 25.875, + "learning_rate": 7.597922969702118e-08, + "loss": 0.8458, + "step": 20863 + }, + { + "epoch": 1.777929271410311, + "grad_norm": 12.5, + "learning_rate": 7.59215494771104e-08, + "loss": 0.4495, + "step": 20864 + }, + { + "epoch": 1.778014486578611, + "grad_norm": 15.3125, + "learning_rate": 7.586389047397794e-08, + "loss": 0.7082, + "step": 20865 + }, + { + "epoch": 1.778099701746911, + "grad_norm": 14.875, + "learning_rate": 7.580625268866562e-08, + "loss": 0.6233, + "step": 20866 + }, + { + "epoch": 1.7781849169152109, + "grad_norm": 14.4375, + "learning_rate": 7.574863612221497e-08, + "loss": 0.6831, + "step": 20867 + }, + { + "epoch": 1.7782701320835108, + "grad_norm": 14.5, + "learning_rate": 7.569104077566739e-08, + "loss": 0.5957, + "step": 20868 + }, + { + "epoch": 1.7783553472518108, + "grad_norm": 15.8125, + "learning_rate": 7.563346665006369e-08, + "loss": 0.4627, + "step": 20869 + }, + { + "epoch": 1.7784405624201107, + "grad_norm": 15.125, + "learning_rate": 7.557591374644388e-08, + "loss": 0.5541, + "step": 20870 + }, + { + "epoch": 1.7785257775884107, + "grad_norm": 20.375, + "learning_rate": 7.551838206584838e-08, + "loss": 0.3962, + "step": 20871 + }, + { + "epoch": 1.7786109927567106, + "grad_norm": 17.375, + "learning_rate": 7.546087160931664e-08, + "loss": 0.5385, + "step": 20872 + }, + { + "epoch": 1.7786962079250106, + "grad_norm": 15.9375, + "learning_rate": 7.540338237788825e-08, + "loss": 0.645, + "step": 20873 + }, + { + "epoch": 1.7787814230933106, + "grad_norm": 14.8125, + "learning_rate": 7.53459143726018e-08, + "loss": 0.7579, + "step": 20874 + }, + { + "epoch": 1.7788666382616105, + "grad_norm": 62.5, + "learning_rate": 7.528846759449578e-08, + "loss": 0.7933, + "step": 20875 + }, + { + "epoch": 1.7789518534299105, + "grad_norm": 20.0, + "learning_rate": 7.523104204460869e-08, + "loss": 0.9118, + "step": 20876 + }, + { + "epoch": 1.7790370685982104, + "grad_norm": 14.375, + "learning_rate": 7.517363772397771e-08, + "loss": 0.4295, + "step": 20877 + }, + { + "epoch": 1.7791222837665104, + "grad_norm": 19.625, + "learning_rate": 7.511625463364064e-08, + "loss": 0.6884, + "step": 20878 + }, + { + "epoch": 1.7792074989348103, + "grad_norm": 15.5, + "learning_rate": 7.505889277463416e-08, + "loss": 0.8537, + "step": 20879 + }, + { + "epoch": 1.7792927141031103, + "grad_norm": 12.0, + "learning_rate": 7.500155214799507e-08, + "loss": 0.3129, + "step": 20880 + }, + { + "epoch": 1.7793779292714103, + "grad_norm": 15.0625, + "learning_rate": 7.494423275475949e-08, + "loss": 0.4777, + "step": 20881 + }, + { + "epoch": 1.7794631444397102, + "grad_norm": 11.8125, + "learning_rate": 7.488693459596325e-08, + "loss": 0.3009, + "step": 20882 + }, + { + "epoch": 1.7795483596080102, + "grad_norm": 15.3125, + "learning_rate": 7.482965767264164e-08, + "loss": 0.4861, + "step": 20883 + }, + { + "epoch": 1.7796335747763101, + "grad_norm": 17.0, + "learning_rate": 7.477240198583008e-08, + "loss": 0.8164, + "step": 20884 + }, + { + "epoch": 1.77971878994461, + "grad_norm": 12.6875, + "learning_rate": 7.471516753656275e-08, + "loss": 0.3189, + "step": 20885 + }, + { + "epoch": 1.77980400511291, + "grad_norm": 15.0625, + "learning_rate": 7.465795432587436e-08, + "loss": 0.688, + "step": 20886 + }, + { + "epoch": 1.77988922028121, + "grad_norm": 16.25, + "learning_rate": 7.460076235479855e-08, + "loss": 0.7007, + "step": 20887 + }, + { + "epoch": 1.77997443544951, + "grad_norm": 13.6875, + "learning_rate": 7.454359162436877e-08, + "loss": 0.3728, + "step": 20888 + }, + { + "epoch": 1.78005965061781, + "grad_norm": 15.0625, + "learning_rate": 7.44864421356184e-08, + "loss": 0.5299, + "step": 20889 + }, + { + "epoch": 1.7801448657861099, + "grad_norm": 15.6875, + "learning_rate": 7.442931388958033e-08, + "loss": 0.5082, + "step": 20890 + }, + { + "epoch": 1.7802300809544098, + "grad_norm": 15.75, + "learning_rate": 7.437220688728638e-08, + "loss": 0.5812, + "step": 20891 + }, + { + "epoch": 1.7803152961227098, + "grad_norm": 17.875, + "learning_rate": 7.431512112976908e-08, + "loss": 0.4528, + "step": 20892 + }, + { + "epoch": 1.7804005112910097, + "grad_norm": 8.8125, + "learning_rate": 7.425805661805966e-08, + "loss": 0.2022, + "step": 20893 + }, + { + "epoch": 1.7804857264593097, + "grad_norm": 14.5, + "learning_rate": 7.420101335318938e-08, + "loss": 0.2984, + "step": 20894 + }, + { + "epoch": 1.7805709416276096, + "grad_norm": 12.875, + "learning_rate": 7.414399133618924e-08, + "loss": 0.26, + "step": 20895 + }, + { + "epoch": 1.7806561567959096, + "grad_norm": 18.375, + "learning_rate": 7.408699056808952e-08, + "loss": 0.798, + "step": 20896 + }, + { + "epoch": 1.7807413719642096, + "grad_norm": 18.125, + "learning_rate": 7.403001104992036e-08, + "loss": 0.7421, + "step": 20897 + }, + { + "epoch": 1.7808265871325095, + "grad_norm": 23.375, + "learning_rate": 7.397305278271163e-08, + "loss": 0.5364, + "step": 20898 + }, + { + "epoch": 1.7809118023008095, + "grad_norm": 12.0625, + "learning_rate": 7.391611576749224e-08, + "loss": 0.3358, + "step": 20899 + }, + { + "epoch": 1.7809970174691094, + "grad_norm": 13.5, + "learning_rate": 7.385920000529137e-08, + "loss": 0.8672, + "step": 20900 + }, + { + "epoch": 1.7810822326374094, + "grad_norm": 11.875, + "learning_rate": 7.380230549713734e-08, + "loss": 0.3468, + "step": 20901 + }, + { + "epoch": 1.7811674478057093, + "grad_norm": 13.75, + "learning_rate": 7.374543224405839e-08, + "loss": 0.5031, + "step": 20902 + }, + { + "epoch": 1.7812526629740093, + "grad_norm": 12.3125, + "learning_rate": 7.368858024708256e-08, + "loss": 0.3944, + "step": 20903 + }, + { + "epoch": 1.7813378781423093, + "grad_norm": 13.75, + "learning_rate": 7.363174950723668e-08, + "loss": 0.5372, + "step": 20904 + }, + { + "epoch": 1.7814230933106092, + "grad_norm": 18.75, + "learning_rate": 7.35749400255481e-08, + "loss": 0.7873, + "step": 20905 + }, + { + "epoch": 1.7815083084789092, + "grad_norm": 19.75, + "learning_rate": 7.351815180304326e-08, + "loss": 0.5804, + "step": 20906 + }, + { + "epoch": 1.7815935236472091, + "grad_norm": 19.625, + "learning_rate": 7.346138484074867e-08, + "loss": 0.5995, + "step": 20907 + }, + { + "epoch": 1.781678738815509, + "grad_norm": 21.625, + "learning_rate": 7.340463913968976e-08, + "loss": 0.5808, + "step": 20908 + }, + { + "epoch": 1.781763953983809, + "grad_norm": 10.9375, + "learning_rate": 7.334791470089239e-08, + "loss": 0.3275, + "step": 20909 + }, + { + "epoch": 1.781849169152109, + "grad_norm": 13.125, + "learning_rate": 7.329121152538116e-08, + "loss": 0.5764, + "step": 20910 + }, + { + "epoch": 1.781934384320409, + "grad_norm": 18.875, + "learning_rate": 7.32345296141812e-08, + "loss": 0.7965, + "step": 20911 + }, + { + "epoch": 1.782019599488709, + "grad_norm": 14.5, + "learning_rate": 7.317786896831658e-08, + "loss": 0.4579, + "step": 20912 + }, + { + "epoch": 1.7821048146570089, + "grad_norm": 35.0, + "learning_rate": 7.312122958881119e-08, + "loss": 0.7094, + "step": 20913 + }, + { + "epoch": 1.7821900298253088, + "grad_norm": 33.25, + "learning_rate": 7.306461147668851e-08, + "loss": 0.7936, + "step": 20914 + }, + { + "epoch": 1.7822752449936088, + "grad_norm": 17.5, + "learning_rate": 7.300801463297202e-08, + "loss": 0.3258, + "step": 20915 + }, + { + "epoch": 1.7823604601619087, + "grad_norm": 15.0, + "learning_rate": 7.295143905868413e-08, + "loss": 0.8004, + "step": 20916 + }, + { + "epoch": 1.7824456753302087, + "grad_norm": 12.9375, + "learning_rate": 7.289488475484761e-08, + "loss": 0.4927, + "step": 20917 + }, + { + "epoch": 1.7825308904985087, + "grad_norm": 13.4375, + "learning_rate": 7.283835172248386e-08, + "loss": 0.5715, + "step": 20918 + }, + { + "epoch": 1.7826161056668086, + "grad_norm": 13.9375, + "learning_rate": 7.278183996261498e-08, + "loss": 0.4183, + "step": 20919 + }, + { + "epoch": 1.7827013208351086, + "grad_norm": 13.6875, + "learning_rate": 7.272534947626211e-08, + "loss": 0.6403, + "step": 20920 + }, + { + "epoch": 1.7827865360034085, + "grad_norm": 16.375, + "learning_rate": 7.266888026444582e-08, + "loss": 0.7423, + "step": 20921 + }, + { + "epoch": 1.7828717511717085, + "grad_norm": 12.625, + "learning_rate": 7.261243232818683e-08, + "loss": 0.424, + "step": 20922 + }, + { + "epoch": 1.7829569663400084, + "grad_norm": 16.125, + "learning_rate": 7.255600566850512e-08, + "loss": 0.658, + "step": 20923 + }, + { + "epoch": 1.7830421815083084, + "grad_norm": 15.1875, + "learning_rate": 7.249960028642047e-08, + "loss": 0.6631, + "step": 20924 + }, + { + "epoch": 1.7831273966766084, + "grad_norm": 18.5, + "learning_rate": 7.244321618295233e-08, + "loss": 0.9254, + "step": 20925 + }, + { + "epoch": 1.7832126118449083, + "grad_norm": 32.0, + "learning_rate": 7.238685335911904e-08, + "loss": 1.0861, + "step": 20926 + }, + { + "epoch": 1.7832978270132083, + "grad_norm": 17.125, + "learning_rate": 7.233051181593953e-08, + "loss": 0.7111, + "step": 20927 + }, + { + "epoch": 1.7833830421815082, + "grad_norm": 21.0, + "learning_rate": 7.227419155443213e-08, + "loss": 0.7134, + "step": 20928 + }, + { + "epoch": 1.7834682573498082, + "grad_norm": 14.4375, + "learning_rate": 7.221789257561423e-08, + "loss": 0.6712, + "step": 20929 + }, + { + "epoch": 1.7835534725181081, + "grad_norm": 13.0625, + "learning_rate": 7.216161488050321e-08, + "loss": 0.4143, + "step": 20930 + }, + { + "epoch": 1.783638687686408, + "grad_norm": 16.75, + "learning_rate": 7.210535847011632e-08, + "loss": 0.6517, + "step": 20931 + }, + { + "epoch": 1.783723902854708, + "grad_norm": 20.25, + "learning_rate": 7.204912334547023e-08, + "loss": 0.9461, + "step": 20932 + }, + { + "epoch": 1.783809118023008, + "grad_norm": 12.8125, + "learning_rate": 7.19929095075811e-08, + "loss": 0.4668, + "step": 20933 + }, + { + "epoch": 1.783894333191308, + "grad_norm": 19.125, + "learning_rate": 7.193671695746446e-08, + "loss": 0.745, + "step": 20934 + }, + { + "epoch": 1.783979548359608, + "grad_norm": 15.875, + "learning_rate": 7.188054569613592e-08, + "loss": 0.8223, + "step": 20935 + }, + { + "epoch": 1.7840647635279079, + "grad_norm": 15.4375, + "learning_rate": 7.18243957246105e-08, + "loss": 0.7636, + "step": 20936 + }, + { + "epoch": 1.7841499786962078, + "grad_norm": 12.875, + "learning_rate": 7.176826704390336e-08, + "loss": 0.3382, + "step": 20937 + }, + { + "epoch": 1.7842351938645078, + "grad_norm": 19.75, + "learning_rate": 7.171215965502825e-08, + "loss": 0.3358, + "step": 20938 + }, + { + "epoch": 1.7843204090328078, + "grad_norm": 23.25, + "learning_rate": 7.165607355899912e-08, + "loss": 0.8789, + "step": 20939 + }, + { + "epoch": 1.7844056242011077, + "grad_norm": 15.5625, + "learning_rate": 7.160000875682985e-08, + "loss": 0.7682, + "step": 20940 + }, + { + "epoch": 1.7844908393694077, + "grad_norm": 21.0, + "learning_rate": 7.154396524953367e-08, + "loss": 0.9422, + "step": 20941 + }, + { + "epoch": 1.7845760545377076, + "grad_norm": 15.375, + "learning_rate": 7.148794303812268e-08, + "loss": 0.4914, + "step": 20942 + }, + { + "epoch": 1.7846612697060076, + "grad_norm": 16.75, + "learning_rate": 7.143194212360968e-08, + "loss": 0.7679, + "step": 20943 + }, + { + "epoch": 1.7847464848743075, + "grad_norm": 12.9375, + "learning_rate": 7.137596250700666e-08, + "loss": 0.4968, + "step": 20944 + }, + { + "epoch": 1.7848317000426075, + "grad_norm": 22.125, + "learning_rate": 7.132000418932528e-08, + "loss": 0.8453, + "step": 20945 + }, + { + "epoch": 1.7849169152109075, + "grad_norm": 19.375, + "learning_rate": 7.126406717157656e-08, + "loss": 0.7881, + "step": 20946 + }, + { + "epoch": 1.7850021303792074, + "grad_norm": 19.125, + "learning_rate": 7.120815145477148e-08, + "loss": 0.626, + "step": 20947 + }, + { + "epoch": 1.7850873455475074, + "grad_norm": 12.9375, + "learning_rate": 7.115225703992035e-08, + "loss": 0.3385, + "step": 20948 + }, + { + "epoch": 1.7851725607158073, + "grad_norm": 9.8125, + "learning_rate": 7.109638392803361e-08, + "loss": 0.2594, + "step": 20949 + }, + { + "epoch": 1.7852577758841073, + "grad_norm": 12.8125, + "learning_rate": 7.104053212012074e-08, + "loss": 0.4388, + "step": 20950 + }, + { + "epoch": 1.7853429910524072, + "grad_norm": 17.125, + "learning_rate": 7.098470161719079e-08, + "loss": 0.7642, + "step": 20951 + }, + { + "epoch": 1.7854282062207072, + "grad_norm": 11.0, + "learning_rate": 7.092889242025281e-08, + "loss": 0.3661, + "step": 20952 + }, + { + "epoch": 1.7855134213890071, + "grad_norm": 15.125, + "learning_rate": 7.087310453031543e-08, + "loss": 0.414, + "step": 20953 + }, + { + "epoch": 1.785598636557307, + "grad_norm": 13.8125, + "learning_rate": 7.081733794838703e-08, + "loss": 0.4447, + "step": 20954 + }, + { + "epoch": 1.785683851725607, + "grad_norm": 15.125, + "learning_rate": 7.076159267547484e-08, + "loss": 0.6418, + "step": 20955 + }, + { + "epoch": 1.785769066893907, + "grad_norm": 14.6875, + "learning_rate": 7.07058687125864e-08, + "loss": 0.5832, + "step": 20956 + }, + { + "epoch": 1.785854282062207, + "grad_norm": 15.4375, + "learning_rate": 7.065016606072908e-08, + "loss": 0.8401, + "step": 20957 + }, + { + "epoch": 1.785939497230507, + "grad_norm": 24.0, + "learning_rate": 7.059448472090918e-08, + "loss": 0.4693, + "step": 20958 + }, + { + "epoch": 1.786024712398807, + "grad_norm": 9.9375, + "learning_rate": 7.05388246941327e-08, + "loss": 0.2564, + "step": 20959 + }, + { + "epoch": 1.7861099275671068, + "grad_norm": 15.75, + "learning_rate": 7.048318598140564e-08, + "loss": 0.5743, + "step": 20960 + }, + { + "epoch": 1.7861951427354068, + "grad_norm": 19.125, + "learning_rate": 7.042756858373356e-08, + "loss": 1.1321, + "step": 20961 + }, + { + "epoch": 1.7862803579037068, + "grad_norm": 13.375, + "learning_rate": 7.037197250212165e-08, + "loss": 0.2654, + "step": 20962 + }, + { + "epoch": 1.7863655730720067, + "grad_norm": 16.75, + "learning_rate": 7.031639773757426e-08, + "loss": 0.5771, + "step": 20963 + }, + { + "epoch": 1.7864507882403067, + "grad_norm": 12.75, + "learning_rate": 7.026084429109584e-08, + "loss": 0.4544, + "step": 20964 + }, + { + "epoch": 1.7865360034086066, + "grad_norm": 15.6875, + "learning_rate": 7.020531216369045e-08, + "loss": 0.8182, + "step": 20965 + }, + { + "epoch": 1.7866212185769066, + "grad_norm": 20.5, + "learning_rate": 7.014980135636118e-08, + "loss": 0.7598, + "step": 20966 + }, + { + "epoch": 1.7867064337452065, + "grad_norm": 15.625, + "learning_rate": 7.009431187011168e-08, + "loss": 0.6849, + "step": 20967 + }, + { + "epoch": 1.7867916489135065, + "grad_norm": 13.5625, + "learning_rate": 7.003884370594433e-08, + "loss": 0.5311, + "step": 20968 + }, + { + "epoch": 1.7868768640818065, + "grad_norm": 11.25, + "learning_rate": 6.998339686486153e-08, + "loss": 0.2957, + "step": 20969 + }, + { + "epoch": 1.7869620792501064, + "grad_norm": 16.25, + "learning_rate": 6.992797134786525e-08, + "loss": 0.6898, + "step": 20970 + }, + { + "epoch": 1.7870472944184064, + "grad_norm": 10.3125, + "learning_rate": 6.987256715595761e-08, + "loss": 0.2829, + "step": 20971 + }, + { + "epoch": 1.7871325095867063, + "grad_norm": 20.0, + "learning_rate": 6.981718429013907e-08, + "loss": 0.9671, + "step": 20972 + }, + { + "epoch": 1.7872177247550063, + "grad_norm": 28.0, + "learning_rate": 6.976182275141075e-08, + "loss": 1.1709, + "step": 20973 + }, + { + "epoch": 1.7873029399233062, + "grad_norm": 14.8125, + "learning_rate": 6.970648254077325e-08, + "loss": 0.7531, + "step": 20974 + }, + { + "epoch": 1.7873881550916062, + "grad_norm": 21.0, + "learning_rate": 6.965116365922661e-08, + "loss": 0.5028, + "step": 20975 + }, + { + "epoch": 1.7874733702599062, + "grad_norm": 15.3125, + "learning_rate": 6.959586610777016e-08, + "loss": 0.8341, + "step": 20976 + }, + { + "epoch": 1.7875585854282061, + "grad_norm": 12.125, + "learning_rate": 6.954058988740339e-08, + "loss": 0.4628, + "step": 20977 + }, + { + "epoch": 1.787643800596506, + "grad_norm": 14.75, + "learning_rate": 6.948533499912521e-08, + "loss": 0.5426, + "step": 20978 + }, + { + "epoch": 1.787729015764806, + "grad_norm": 14.3125, + "learning_rate": 6.943010144393427e-08, + "loss": 0.6175, + "step": 20979 + }, + { + "epoch": 1.787814230933106, + "grad_norm": 13.5, + "learning_rate": 6.93748892228284e-08, + "loss": 0.5818, + "step": 20980 + }, + { + "epoch": 1.787899446101406, + "grad_norm": 18.625, + "learning_rate": 6.931969833680552e-08, + "loss": 0.5297, + "step": 20981 + }, + { + "epoch": 1.787984661269706, + "grad_norm": 15.9375, + "learning_rate": 6.926452878686305e-08, + "loss": 0.6702, + "step": 20982 + }, + { + "epoch": 1.7880698764380059, + "grad_norm": 23.875, + "learning_rate": 6.920938057399782e-08, + "loss": 0.6081, + "step": 20983 + }, + { + "epoch": 1.7881550916063058, + "grad_norm": 15.0625, + "learning_rate": 6.915425369920654e-08, + "loss": 0.3354, + "step": 20984 + }, + { + "epoch": 1.7882403067746058, + "grad_norm": 26.625, + "learning_rate": 6.90991481634852e-08, + "loss": 0.8031, + "step": 20985 + }, + { + "epoch": 1.7883255219429057, + "grad_norm": 11.9375, + "learning_rate": 6.90440639678297e-08, + "loss": 0.3551, + "step": 20986 + }, + { + "epoch": 1.7884107371112057, + "grad_norm": 14.6875, + "learning_rate": 6.898900111323548e-08, + "loss": 0.488, + "step": 20987 + }, + { + "epoch": 1.7884959522795056, + "grad_norm": 13.1875, + "learning_rate": 6.893395960069784e-08, + "loss": 0.6641, + "step": 20988 + }, + { + "epoch": 1.7885811674478056, + "grad_norm": 18.75, + "learning_rate": 6.887893943121102e-08, + "loss": 1.0079, + "step": 20989 + }, + { + "epoch": 1.7886663826161056, + "grad_norm": 19.25, + "learning_rate": 6.88239406057696e-08, + "loss": 1.0023, + "step": 20990 + }, + { + "epoch": 1.7887515977844055, + "grad_norm": 24.625, + "learning_rate": 6.876896312536727e-08, + "loss": 0.7635, + "step": 20991 + }, + { + "epoch": 1.7888368129527055, + "grad_norm": 14.5625, + "learning_rate": 6.871400699099764e-08, + "loss": 0.5046, + "step": 20992 + }, + { + "epoch": 1.7889220281210054, + "grad_norm": 13.25, + "learning_rate": 6.86590722036537e-08, + "loss": 0.4322, + "step": 20993 + }, + { + "epoch": 1.7890072432893054, + "grad_norm": 14.9375, + "learning_rate": 6.860415876432824e-08, + "loss": 0.7393, + "step": 20994 + }, + { + "epoch": 1.7890924584576053, + "grad_norm": 11.8125, + "learning_rate": 6.854926667401368e-08, + "loss": 0.6324, + "step": 20995 + }, + { + "epoch": 1.7891776736259053, + "grad_norm": 15.1875, + "learning_rate": 6.8494395933702e-08, + "loss": 0.5619, + "step": 20996 + }, + { + "epoch": 1.7892628887942053, + "grad_norm": 13.625, + "learning_rate": 6.843954654438448e-08, + "loss": 0.6032, + "step": 20997 + }, + { + "epoch": 1.7893481039625052, + "grad_norm": 16.625, + "learning_rate": 6.83847185070527e-08, + "loss": 0.9303, + "step": 20998 + }, + { + "epoch": 1.7894333191308052, + "grad_norm": 16.875, + "learning_rate": 6.832991182269711e-08, + "loss": 0.6181, + "step": 20999 + }, + { + "epoch": 1.7895185342991051, + "grad_norm": 17.375, + "learning_rate": 6.827512649230831e-08, + "loss": 0.3765, + "step": 21000 + }, + { + "epoch": 1.789603749467405, + "grad_norm": 12.3125, + "learning_rate": 6.822036251687633e-08, + "loss": 0.3982, + "step": 21001 + }, + { + "epoch": 1.789688964635705, + "grad_norm": 14.75, + "learning_rate": 6.816561989739068e-08, + "loss": 0.5757, + "step": 21002 + }, + { + "epoch": 1.789774179804005, + "grad_norm": 13.3125, + "learning_rate": 6.811089863484083e-08, + "loss": 0.3254, + "step": 21003 + }, + { + "epoch": 1.789859394972305, + "grad_norm": 16.875, + "learning_rate": 6.80561987302153e-08, + "loss": 0.6562, + "step": 21004 + }, + { + "epoch": 1.789944610140605, + "grad_norm": 16.25, + "learning_rate": 6.800152018450315e-08, + "loss": 0.5072, + "step": 21005 + }, + { + "epoch": 1.7900298253089049, + "grad_norm": 12.0, + "learning_rate": 6.794686299869207e-08, + "loss": 0.455, + "step": 21006 + }, + { + "epoch": 1.7901150404772048, + "grad_norm": 15.9375, + "learning_rate": 6.789222717376962e-08, + "loss": 0.826, + "step": 21007 + }, + { + "epoch": 1.7902002556455048, + "grad_norm": 17.25, + "learning_rate": 6.783761271072331e-08, + "loss": 0.8436, + "step": 21008 + }, + { + "epoch": 1.7902854708138047, + "grad_norm": 39.25, + "learning_rate": 6.778301961054028e-08, + "loss": 0.7903, + "step": 21009 + }, + { + "epoch": 1.7903706859821047, + "grad_norm": 12.25, + "learning_rate": 6.772844787420671e-08, + "loss": 0.4044, + "step": 21010 + }, + { + "epoch": 1.7904559011504046, + "grad_norm": 13.0, + "learning_rate": 6.7673897502709e-08, + "loss": 0.5063, + "step": 21011 + }, + { + "epoch": 1.7905411163187046, + "grad_norm": 14.0, + "learning_rate": 6.76193684970329e-08, + "loss": 0.3535, + "step": 21012 + }, + { + "epoch": 1.7906263314870046, + "grad_norm": 12.1875, + "learning_rate": 6.756486085816388e-08, + "loss": 0.2655, + "step": 21013 + }, + { + "epoch": 1.7907115466553045, + "grad_norm": 15.125, + "learning_rate": 6.751037458708684e-08, + "loss": 0.6746, + "step": 21014 + }, + { + "epoch": 1.7907967618236045, + "grad_norm": 11.625, + "learning_rate": 6.745590968478627e-08, + "loss": 0.3738, + "step": 21015 + }, + { + "epoch": 1.7908819769919044, + "grad_norm": 19.5, + "learning_rate": 6.740146615224666e-08, + "loss": 0.8112, + "step": 21016 + }, + { + "epoch": 1.7909671921602044, + "grad_norm": 28.5, + "learning_rate": 6.734704399045166e-08, + "loss": 0.7921, + "step": 21017 + }, + { + "epoch": 1.7910524073285043, + "grad_norm": 13.375, + "learning_rate": 6.729264320038495e-08, + "loss": 0.3311, + "step": 21018 + }, + { + "epoch": 1.7911376224968043, + "grad_norm": 13.375, + "learning_rate": 6.723826378302947e-08, + "loss": 0.572, + "step": 21019 + }, + { + "epoch": 1.7912228376651043, + "grad_norm": 15.875, + "learning_rate": 6.718390573936778e-08, + "loss": 0.8561, + "step": 21020 + }, + { + "epoch": 1.7913080528334042, + "grad_norm": 14.25, + "learning_rate": 6.712956907038257e-08, + "loss": 0.5327, + "step": 21021 + }, + { + "epoch": 1.7913932680017042, + "grad_norm": 25.5, + "learning_rate": 6.70752537770554e-08, + "loss": 0.6842, + "step": 21022 + }, + { + "epoch": 1.7914784831700041, + "grad_norm": 18.0, + "learning_rate": 6.702095986036785e-08, + "loss": 0.732, + "step": 21023 + }, + { + "epoch": 1.791563698338304, + "grad_norm": 15.125, + "learning_rate": 6.69666873213011e-08, + "loss": 0.4281, + "step": 21024 + }, + { + "epoch": 1.791648913506604, + "grad_norm": 16.25, + "learning_rate": 6.691243616083601e-08, + "loss": 0.5987, + "step": 21025 + }, + { + "epoch": 1.791734128674904, + "grad_norm": 17.0, + "learning_rate": 6.685820637995292e-08, + "loss": 0.8103, + "step": 21026 + }, + { + "epoch": 1.791819343843204, + "grad_norm": 24.25, + "learning_rate": 6.680399797963172e-08, + "loss": 0.4756, + "step": 21027 + }, + { + "epoch": 1.791904559011504, + "grad_norm": 15.25, + "learning_rate": 6.674981096085206e-08, + "loss": 0.7943, + "step": 21028 + }, + { + "epoch": 1.7919897741798039, + "grad_norm": 19.25, + "learning_rate": 6.669564532459316e-08, + "loss": 0.7504, + "step": 21029 + }, + { + "epoch": 1.7920749893481038, + "grad_norm": 18.875, + "learning_rate": 6.664150107183395e-08, + "loss": 0.6268, + "step": 21030 + }, + { + "epoch": 1.7921602045164038, + "grad_norm": 13.9375, + "learning_rate": 6.658737820355296e-08, + "loss": 0.6233, + "step": 21031 + }, + { + "epoch": 1.7922454196847037, + "grad_norm": 10.4375, + "learning_rate": 6.653327672072773e-08, + "loss": 0.3469, + "step": 21032 + }, + { + "epoch": 1.7923306348530037, + "grad_norm": 12.25, + "learning_rate": 6.647919662433625e-08, + "loss": 0.409, + "step": 21033 + }, + { + "epoch": 1.7924158500213037, + "grad_norm": 12.625, + "learning_rate": 6.64251379153559e-08, + "loss": 0.3213, + "step": 21034 + }, + { + "epoch": 1.7925010651896036, + "grad_norm": 12.3125, + "learning_rate": 6.637110059476357e-08, + "loss": 0.6079, + "step": 21035 + }, + { + "epoch": 1.7925862803579036, + "grad_norm": 18.875, + "learning_rate": 6.631708466353554e-08, + "loss": 0.5788, + "step": 21036 + }, + { + "epoch": 1.7926714955262035, + "grad_norm": 16.0, + "learning_rate": 6.626309012264812e-08, + "loss": 0.5411, + "step": 21037 + }, + { + "epoch": 1.7927567106945035, + "grad_norm": 15.4375, + "learning_rate": 6.62091169730772e-08, + "loss": 0.6758, + "step": 21038 + }, + { + "epoch": 1.7928419258628034, + "grad_norm": 15.6875, + "learning_rate": 6.615516521579782e-08, + "loss": 0.8004, + "step": 21039 + }, + { + "epoch": 1.7929271410311034, + "grad_norm": 20.5, + "learning_rate": 6.61012348517849e-08, + "loss": 1.1118, + "step": 21040 + }, + { + "epoch": 1.7930123561994034, + "grad_norm": 15.25, + "learning_rate": 6.604732588201323e-08, + "loss": 0.6572, + "step": 21041 + }, + { + "epoch": 1.7930975713677033, + "grad_norm": 15.75, + "learning_rate": 6.599343830745688e-08, + "loss": 0.4054, + "step": 21042 + }, + { + "epoch": 1.7931827865360033, + "grad_norm": 16.625, + "learning_rate": 6.593957212908994e-08, + "loss": 0.5916, + "step": 21043 + }, + { + "epoch": 1.7932680017043032, + "grad_norm": 16.875, + "learning_rate": 6.588572734788537e-08, + "loss": 0.765, + "step": 21044 + }, + { + "epoch": 1.7933532168726032, + "grad_norm": 18.0, + "learning_rate": 6.583190396481642e-08, + "loss": 0.2296, + "step": 21045 + }, + { + "epoch": 1.7934384320409031, + "grad_norm": 11.125, + "learning_rate": 6.577810198085579e-08, + "loss": 0.311, + "step": 21046 + }, + { + "epoch": 1.793523647209203, + "grad_norm": 16.0, + "learning_rate": 6.572432139697577e-08, + "loss": 0.6527, + "step": 21047 + }, + { + "epoch": 1.793608862377503, + "grad_norm": 16.625, + "learning_rate": 6.567056221414833e-08, + "loss": 0.9143, + "step": 21048 + }, + { + "epoch": 1.793694077545803, + "grad_norm": 22.625, + "learning_rate": 6.561682443334438e-08, + "loss": 0.4783, + "step": 21049 + }, + { + "epoch": 1.793779292714103, + "grad_norm": 12.375, + "learning_rate": 6.55631080555355e-08, + "loss": 0.3378, + "step": 21050 + }, + { + "epoch": 1.793864507882403, + "grad_norm": 13.875, + "learning_rate": 6.550941308169245e-08, + "loss": 0.5618, + "step": 21051 + }, + { + "epoch": 1.7939497230507029, + "grad_norm": 19.75, + "learning_rate": 6.545573951278528e-08, + "loss": 0.779, + "step": 21052 + }, + { + "epoch": 1.7940349382190028, + "grad_norm": 12.0, + "learning_rate": 6.540208734978403e-08, + "loss": 0.3878, + "step": 21053 + }, + { + "epoch": 1.7941201533873028, + "grad_norm": 15.25, + "learning_rate": 6.534845659365823e-08, + "loss": 0.7297, + "step": 21054 + }, + { + "epoch": 1.7942053685556028, + "grad_norm": 13.875, + "learning_rate": 6.529484724537724e-08, + "loss": 0.3871, + "step": 21055 + }, + { + "epoch": 1.7942905837239027, + "grad_norm": 13.625, + "learning_rate": 6.524125930590972e-08, + "loss": 0.2855, + "step": 21056 + }, + { + "epoch": 1.7943757988922027, + "grad_norm": 14.4375, + "learning_rate": 6.518769277622378e-08, + "loss": 0.4363, + "step": 21057 + }, + { + "epoch": 1.7944610140605026, + "grad_norm": 23.875, + "learning_rate": 6.513414765728781e-08, + "loss": 0.689, + "step": 21058 + }, + { + "epoch": 1.7945462292288026, + "grad_norm": 14.5625, + "learning_rate": 6.508062395006911e-08, + "loss": 0.538, + "step": 21059 + }, + { + "epoch": 1.7946314443971025, + "grad_norm": 14.875, + "learning_rate": 6.502712165553524e-08, + "loss": 0.3263, + "step": 21060 + }, + { + "epoch": 1.7947166595654025, + "grad_norm": 19.875, + "learning_rate": 6.497364077465277e-08, + "loss": 0.508, + "step": 21061 + }, + { + "epoch": 1.7948018747337025, + "grad_norm": 18.625, + "learning_rate": 6.492018130838815e-08, + "loss": 0.7944, + "step": 21062 + }, + { + "epoch": 1.7948870899020024, + "grad_norm": 21.375, + "learning_rate": 6.486674325770783e-08, + "loss": 1.1572, + "step": 21063 + }, + { + "epoch": 1.7949723050703024, + "grad_norm": 14.75, + "learning_rate": 6.481332662357703e-08, + "loss": 0.7825, + "step": 21064 + }, + { + "epoch": 1.7950575202386023, + "grad_norm": 15.0, + "learning_rate": 6.475993140696121e-08, + "loss": 0.6307, + "step": 21065 + }, + { + "epoch": 1.7951427354069023, + "grad_norm": 15.0625, + "learning_rate": 6.47065576088253e-08, + "loss": 0.5511, + "step": 21066 + }, + { + "epoch": 1.7952279505752022, + "grad_norm": 16.375, + "learning_rate": 6.465320523013366e-08, + "loss": 0.3282, + "step": 21067 + }, + { + "epoch": 1.7953131657435022, + "grad_norm": 14.125, + "learning_rate": 6.459987427185094e-08, + "loss": 0.4313, + "step": 21068 + }, + { + "epoch": 1.7953983809118021, + "grad_norm": 17.5, + "learning_rate": 6.454656473494011e-08, + "loss": 0.6278, + "step": 21069 + }, + { + "epoch": 1.795483596080102, + "grad_norm": 14.9375, + "learning_rate": 6.4493276620365e-08, + "loss": 0.8363, + "step": 21070 + }, + { + "epoch": 1.795568811248402, + "grad_norm": 14.375, + "learning_rate": 6.444000992908871e-08, + "loss": 0.5306, + "step": 21071 + }, + { + "epoch": 1.795654026416702, + "grad_norm": 11.25, + "learning_rate": 6.438676466207341e-08, + "loss": 0.523, + "step": 21072 + }, + { + "epoch": 1.795739241585002, + "grad_norm": 13.5, + "learning_rate": 6.433354082028165e-08, + "loss": 0.296, + "step": 21073 + }, + { + "epoch": 1.795824456753302, + "grad_norm": 13.75, + "learning_rate": 6.428033840467487e-08, + "loss": 0.3835, + "step": 21074 + }, + { + "epoch": 1.795909671921602, + "grad_norm": 18.25, + "learning_rate": 6.422715741621482e-08, + "loss": 0.6709, + "step": 21075 + }, + { + "epoch": 1.7959948870899018, + "grad_norm": 16.625, + "learning_rate": 6.41739978558624e-08, + "loss": 0.5212, + "step": 21076 + }, + { + "epoch": 1.7960801022582018, + "grad_norm": 14.1875, + "learning_rate": 6.412085972457849e-08, + "loss": 0.5237, + "step": 21077 + }, + { + "epoch": 1.7961653174265018, + "grad_norm": 35.75, + "learning_rate": 6.406774302332289e-08, + "loss": 0.6927, + "step": 21078 + }, + { + "epoch": 1.7962505325948017, + "grad_norm": 17.875, + "learning_rate": 6.401464775305579e-08, + "loss": 0.7802, + "step": 21079 + }, + { + "epoch": 1.7963357477631017, + "grad_norm": 21.75, + "learning_rate": 6.39615739147366e-08, + "loss": 0.7312, + "step": 21080 + }, + { + "epoch": 1.7964209629314016, + "grad_norm": 12.375, + "learning_rate": 6.39085215093245e-08, + "loss": 0.4967, + "step": 21081 + }, + { + "epoch": 1.7965061780997016, + "grad_norm": 12.375, + "learning_rate": 6.385549053777807e-08, + "loss": 0.3718, + "step": 21082 + }, + { + "epoch": 1.7965913932680015, + "grad_norm": 16.5, + "learning_rate": 6.380248100105554e-08, + "loss": 0.4894, + "step": 21083 + }, + { + "epoch": 1.7966766084363015, + "grad_norm": 26.875, + "learning_rate": 6.374949290011507e-08, + "loss": 1.0531, + "step": 21084 + }, + { + "epoch": 1.7967618236046015, + "grad_norm": 17.75, + "learning_rate": 6.369652623591433e-08, + "loss": 0.5618, + "step": 21085 + }, + { + "epoch": 1.7968470387729014, + "grad_norm": 16.125, + "learning_rate": 6.364358100941007e-08, + "loss": 0.581, + "step": 21086 + }, + { + "epoch": 1.7969322539412014, + "grad_norm": 15.4375, + "learning_rate": 6.359065722155944e-08, + "loss": 0.475, + "step": 21087 + }, + { + "epoch": 1.7970174691095013, + "grad_norm": 18.375, + "learning_rate": 6.353775487331847e-08, + "loss": 0.5578, + "step": 21088 + }, + { + "epoch": 1.7971026842778013, + "grad_norm": 14.375, + "learning_rate": 6.348487396564352e-08, + "loss": 0.6529, + "step": 21089 + }, + { + "epoch": 1.7971878994461012, + "grad_norm": 14.4375, + "learning_rate": 6.343201449949001e-08, + "loss": 0.5464, + "step": 21090 + }, + { + "epoch": 1.7972731146144012, + "grad_norm": 16.5, + "learning_rate": 6.337917647581307e-08, + "loss": 0.632, + "step": 21091 + }, + { + "epoch": 1.7973583297827012, + "grad_norm": 19.5, + "learning_rate": 6.332635989556759e-08, + "loss": 0.9232, + "step": 21092 + }, + { + "epoch": 1.7974435449510011, + "grad_norm": 16.75, + "learning_rate": 6.327356475970811e-08, + "loss": 0.5285, + "step": 21093 + }, + { + "epoch": 1.797528760119301, + "grad_norm": 16.5, + "learning_rate": 6.322079106918886e-08, + "loss": 0.7075, + "step": 21094 + }, + { + "epoch": 1.797613975287601, + "grad_norm": 15.75, + "learning_rate": 6.316803882496336e-08, + "loss": 0.6207, + "step": 21095 + }, + { + "epoch": 1.797699190455901, + "grad_norm": 10.4375, + "learning_rate": 6.311530802798463e-08, + "loss": 0.3289, + "step": 21096 + }, + { + "epoch": 1.797784405624201, + "grad_norm": 16.375, + "learning_rate": 6.306259867920578e-08, + "loss": 0.5936, + "step": 21097 + }, + { + "epoch": 1.797869620792501, + "grad_norm": 15.0625, + "learning_rate": 6.300991077957952e-08, + "loss": 0.4197, + "step": 21098 + }, + { + "epoch": 1.7979548359608009, + "grad_norm": 11.6875, + "learning_rate": 6.29572443300576e-08, + "loss": 0.261, + "step": 21099 + }, + { + "epoch": 1.7980400511291008, + "grad_norm": 14.9375, + "learning_rate": 6.290459933159202e-08, + "loss": 0.7058, + "step": 21100 + }, + { + "epoch": 1.798125266297401, + "grad_norm": 16.625, + "learning_rate": 6.285197578513396e-08, + "loss": 0.6482, + "step": 21101 + }, + { + "epoch": 1.798210481465701, + "grad_norm": 15.875, + "learning_rate": 6.279937369163464e-08, + "loss": 0.6408, + "step": 21102 + }, + { + "epoch": 1.798295696634001, + "grad_norm": 16.75, + "learning_rate": 6.274679305204437e-08, + "loss": 0.5156, + "step": 21103 + }, + { + "epoch": 1.7983809118023009, + "grad_norm": 15.5625, + "learning_rate": 6.269423386731338e-08, + "loss": 0.4094, + "step": 21104 + }, + { + "epoch": 1.7984661269706008, + "grad_norm": 14.75, + "learning_rate": 6.264169613839147e-08, + "loss": 0.6403, + "step": 21105 + }, + { + "epoch": 1.7985513421389008, + "grad_norm": 17.125, + "learning_rate": 6.258917986622815e-08, + "loss": 0.5696, + "step": 21106 + }, + { + "epoch": 1.7986365573072007, + "grad_norm": 19.0, + "learning_rate": 6.253668505177251e-08, + "loss": 0.6961, + "step": 21107 + }, + { + "epoch": 1.7987217724755007, + "grad_norm": 13.875, + "learning_rate": 6.248421169597286e-08, + "loss": 0.5665, + "step": 21108 + }, + { + "epoch": 1.7988069876438006, + "grad_norm": 17.25, + "learning_rate": 6.243175979977769e-08, + "loss": 0.6573, + "step": 21109 + }, + { + "epoch": 1.7988922028121006, + "grad_norm": 31.625, + "learning_rate": 6.237932936413476e-08, + "loss": 0.3568, + "step": 21110 + }, + { + "epoch": 1.7989774179804006, + "grad_norm": 14.375, + "learning_rate": 6.232692038999177e-08, + "loss": 0.3256, + "step": 21111 + }, + { + "epoch": 1.7990626331487005, + "grad_norm": 16.75, + "learning_rate": 6.227453287829575e-08, + "loss": 0.6166, + "step": 21112 + }, + { + "epoch": 1.7991478483170005, + "grad_norm": 15.5625, + "learning_rate": 6.222216682999299e-08, + "loss": 0.6015, + "step": 21113 + }, + { + "epoch": 1.7992330634853004, + "grad_norm": 21.375, + "learning_rate": 6.216982224603e-08, + "loss": 0.7397, + "step": 21114 + }, + { + "epoch": 1.7993182786536004, + "grad_norm": 13.6875, + "learning_rate": 6.211749912735307e-08, + "loss": 0.5298, + "step": 21115 + }, + { + "epoch": 1.7994034938219003, + "grad_norm": 12.125, + "learning_rate": 6.20651974749073e-08, + "loss": 0.3061, + "step": 21116 + }, + { + "epoch": 1.7994887089902003, + "grad_norm": 12.875, + "learning_rate": 6.20129172896379e-08, + "loss": 0.3986, + "step": 21117 + }, + { + "epoch": 1.7995739241585003, + "grad_norm": 18.5, + "learning_rate": 6.196065857248968e-08, + "loss": 0.6202, + "step": 21118 + }, + { + "epoch": 1.7996591393268002, + "grad_norm": 16.0, + "learning_rate": 6.190842132440727e-08, + "loss": 0.5267, + "step": 21119 + }, + { + "epoch": 1.7997443544951002, + "grad_norm": 14.75, + "learning_rate": 6.185620554633426e-08, + "loss": 0.5679, + "step": 21120 + }, + { + "epoch": 1.7998295696634001, + "grad_norm": 30.375, + "learning_rate": 6.180401123921445e-08, + "loss": 1.1426, + "step": 21121 + }, + { + "epoch": 1.7999147848317, + "grad_norm": 23.5, + "learning_rate": 6.175183840399071e-08, + "loss": 0.6732, + "step": 21122 + }, + { + "epoch": 1.8, + "grad_norm": 22.625, + "learning_rate": 6.16996870416063e-08, + "loss": 0.4845, + "step": 21123 + }, + { + "epoch": 1.8000852151683, + "grad_norm": 10.875, + "learning_rate": 6.16475571530037e-08, + "loss": 0.3375, + "step": 21124 + }, + { + "epoch": 1.8001704303366, + "grad_norm": 14.9375, + "learning_rate": 6.159544873912446e-08, + "loss": 0.7536, + "step": 21125 + }, + { + "epoch": 1.8002556455049, + "grad_norm": 25.125, + "learning_rate": 6.154336180091036e-08, + "loss": 0.8698, + "step": 21126 + }, + { + "epoch": 1.8003408606731999, + "grad_norm": 22.5, + "learning_rate": 6.149129633930301e-08, + "loss": 0.3568, + "step": 21127 + }, + { + "epoch": 1.8004260758414998, + "grad_norm": 14.0625, + "learning_rate": 6.143925235524331e-08, + "loss": 0.4095, + "step": 21128 + }, + { + "epoch": 1.8005112910097998, + "grad_norm": 24.375, + "learning_rate": 6.138722984967122e-08, + "loss": 0.8236, + "step": 21129 + }, + { + "epoch": 1.8005965061780997, + "grad_norm": 21.75, + "learning_rate": 6.133522882352707e-08, + "loss": 0.3936, + "step": 21130 + }, + { + "epoch": 1.8006817213463997, + "grad_norm": 11.5, + "learning_rate": 6.12832492777507e-08, + "loss": 0.3036, + "step": 21131 + }, + { + "epoch": 1.8007669365146997, + "grad_norm": 20.625, + "learning_rate": 6.123129121328148e-08, + "loss": 0.7993, + "step": 21132 + }, + { + "epoch": 1.8008521516829996, + "grad_norm": 17.75, + "learning_rate": 6.11793546310581e-08, + "loss": 0.6057, + "step": 21133 + }, + { + "epoch": 1.8009373668512996, + "grad_norm": 15.4375, + "learning_rate": 6.112743953201924e-08, + "loss": 0.5835, + "step": 21134 + }, + { + "epoch": 1.8010225820195995, + "grad_norm": 13.25, + "learning_rate": 6.107554591710294e-08, + "loss": 0.3504, + "step": 21135 + }, + { + "epoch": 1.8011077971878995, + "grad_norm": 15.5625, + "learning_rate": 6.10236737872473e-08, + "loss": 0.5919, + "step": 21136 + }, + { + "epoch": 1.8011930123561994, + "grad_norm": 18.875, + "learning_rate": 6.097182314338951e-08, + "loss": 0.7687, + "step": 21137 + }, + { + "epoch": 1.8012782275244994, + "grad_norm": 21.375, + "learning_rate": 6.091999398646645e-08, + "loss": 0.8587, + "step": 21138 + }, + { + "epoch": 1.8013634426927994, + "grad_norm": 14.25, + "learning_rate": 6.086818631741473e-08, + "loss": 0.6482, + "step": 21139 + }, + { + "epoch": 1.8014486578610993, + "grad_norm": 17.25, + "learning_rate": 6.081640013717069e-08, + "loss": 0.9081, + "step": 21140 + }, + { + "epoch": 1.8015338730293993, + "grad_norm": 22.0, + "learning_rate": 6.076463544667024e-08, + "loss": 0.571, + "step": 21141 + }, + { + "epoch": 1.8016190881976992, + "grad_norm": 13.375, + "learning_rate": 6.071289224684848e-08, + "loss": 0.4691, + "step": 21142 + }, + { + "epoch": 1.8017043033659992, + "grad_norm": 13.5625, + "learning_rate": 6.066117053864076e-08, + "loss": 0.4281, + "step": 21143 + }, + { + "epoch": 1.8017895185342991, + "grad_norm": 12.75, + "learning_rate": 6.060947032298176e-08, + "loss": 0.4088, + "step": 21144 + }, + { + "epoch": 1.801874733702599, + "grad_norm": 17.75, + "learning_rate": 6.055779160080571e-08, + "loss": 0.5262, + "step": 21145 + }, + { + "epoch": 1.801959948870899, + "grad_norm": 15.625, + "learning_rate": 6.05061343730462e-08, + "loss": 0.5465, + "step": 21146 + }, + { + "epoch": 1.802045164039199, + "grad_norm": 11.9375, + "learning_rate": 6.045449864063704e-08, + "loss": 0.4282, + "step": 21147 + }, + { + "epoch": 1.802130379207499, + "grad_norm": 13.375, + "learning_rate": 6.040288440451126e-08, + "loss": 0.4883, + "step": 21148 + }, + { + "epoch": 1.802215594375799, + "grad_norm": 15.5, + "learning_rate": 6.035129166560183e-08, + "loss": 0.601, + "step": 21149 + }, + { + "epoch": 1.8023008095440989, + "grad_norm": 18.0, + "learning_rate": 6.029972042484053e-08, + "loss": 0.6285, + "step": 21150 + }, + { + "epoch": 1.8023860247123988, + "grad_norm": 15.5, + "learning_rate": 6.024817068315966e-08, + "loss": 0.3969, + "step": 21151 + }, + { + "epoch": 1.8024712398806988, + "grad_norm": 21.0, + "learning_rate": 6.019664244149098e-08, + "loss": 0.8103, + "step": 21152 + }, + { + "epoch": 1.8025564550489988, + "grad_norm": 18.75, + "learning_rate": 6.014513570076527e-08, + "loss": 0.9013, + "step": 21153 + }, + { + "epoch": 1.8026416702172987, + "grad_norm": 16.125, + "learning_rate": 6.009365046191345e-08, + "loss": 0.5859, + "step": 21154 + }, + { + "epoch": 1.8027268853855987, + "grad_norm": 20.5, + "learning_rate": 6.004218672586589e-08, + "loss": 0.6017, + "step": 21155 + }, + { + "epoch": 1.8028121005538986, + "grad_norm": 14.5, + "learning_rate": 5.999074449355255e-08, + "loss": 0.423, + "step": 21156 + }, + { + "epoch": 1.8028973157221986, + "grad_norm": 12.0, + "learning_rate": 5.993932376590308e-08, + "loss": 0.4374, + "step": 21157 + }, + { + "epoch": 1.8029825308904985, + "grad_norm": 16.0, + "learning_rate": 5.98879245438469e-08, + "loss": 0.5996, + "step": 21158 + }, + { + "epoch": 1.8030677460587985, + "grad_norm": 14.1875, + "learning_rate": 5.983654682831256e-08, + "loss": 0.5771, + "step": 21159 + }, + { + "epoch": 1.8031529612270984, + "grad_norm": 19.625, + "learning_rate": 5.978519062022875e-08, + "loss": 0.8037, + "step": 21160 + }, + { + "epoch": 1.8032381763953984, + "grad_norm": 24.375, + "learning_rate": 5.973385592052322e-08, + "loss": 0.5462, + "step": 21161 + }, + { + "epoch": 1.8033233915636984, + "grad_norm": 11.6875, + "learning_rate": 5.968254273012411e-08, + "loss": 0.3765, + "step": 21162 + }, + { + "epoch": 1.8034086067319983, + "grad_norm": 14.75, + "learning_rate": 5.963125104995804e-08, + "loss": 0.3729, + "step": 21163 + }, + { + "epoch": 1.8034938219002983, + "grad_norm": 26.0, + "learning_rate": 5.9579980880952456e-08, + "loss": 1.0825, + "step": 21164 + }, + { + "epoch": 1.8035790370685982, + "grad_norm": 17.75, + "learning_rate": 5.9528732224033574e-08, + "loss": 0.5638, + "step": 21165 + }, + { + "epoch": 1.8036642522368982, + "grad_norm": 23.5, + "learning_rate": 5.947750508012773e-08, + "loss": 0.7519, + "step": 21166 + }, + { + "epoch": 1.8037494674051981, + "grad_norm": 22.5, + "learning_rate": 5.942629945016043e-08, + "loss": 0.7844, + "step": 21167 + }, + { + "epoch": 1.803834682573498, + "grad_norm": 18.875, + "learning_rate": 5.937511533505732e-08, + "loss": 0.7357, + "step": 21168 + }, + { + "epoch": 1.803919897741798, + "grad_norm": 14.1875, + "learning_rate": 5.9323952735742956e-08, + "loss": 0.4265, + "step": 21169 + }, + { + "epoch": 1.804005112910098, + "grad_norm": 21.25, + "learning_rate": 5.9272811653142e-08, + "loss": 0.7067, + "step": 21170 + }, + { + "epoch": 1.804090328078398, + "grad_norm": 13.8125, + "learning_rate": 5.922169208817899e-08, + "loss": 0.778, + "step": 21171 + }, + { + "epoch": 1.804175543246698, + "grad_norm": 14.5625, + "learning_rate": 5.917059404177722e-08, + "loss": 0.4801, + "step": 21172 + }, + { + "epoch": 1.804260758414998, + "grad_norm": 13.0625, + "learning_rate": 5.911951751486025e-08, + "loss": 0.7916, + "step": 21173 + }, + { + "epoch": 1.8043459735832978, + "grad_norm": 14.5625, + "learning_rate": 5.906846250835138e-08, + "loss": 0.8196, + "step": 21174 + }, + { + "epoch": 1.8044311887515978, + "grad_norm": 23.0, + "learning_rate": 5.9017429023172644e-08, + "loss": 0.7831, + "step": 21175 + }, + { + "epoch": 1.8045164039198978, + "grad_norm": 15.375, + "learning_rate": 5.896641706024678e-08, + "loss": 0.6292, + "step": 21176 + }, + { + "epoch": 1.8046016190881977, + "grad_norm": 14.5625, + "learning_rate": 5.89154266204954e-08, + "loss": 0.7866, + "step": 21177 + }, + { + "epoch": 1.8046868342564977, + "grad_norm": 14.25, + "learning_rate": 5.8864457704840005e-08, + "loss": 0.3858, + "step": 21178 + }, + { + "epoch": 1.8047720494247976, + "grad_norm": 13.5, + "learning_rate": 5.881351031420165e-08, + "loss": 0.4504, + "step": 21179 + }, + { + "epoch": 1.8048572645930976, + "grad_norm": 18.75, + "learning_rate": 5.8762584449500865e-08, + "loss": 0.7286, + "step": 21180 + }, + { + "epoch": 1.8049424797613975, + "grad_norm": 12.5, + "learning_rate": 5.871168011165815e-08, + "loss": 0.4054, + "step": 21181 + }, + { + "epoch": 1.8050276949296975, + "grad_norm": 11.1875, + "learning_rate": 5.8660797301593323e-08, + "loss": 0.1924, + "step": 21182 + }, + { + "epoch": 1.8051129100979975, + "grad_norm": 16.25, + "learning_rate": 5.8609936020225935e-08, + "loss": 0.8235, + "step": 21183 + }, + { + "epoch": 1.8051981252662974, + "grad_norm": 39.0, + "learning_rate": 5.8559096268474834e-08, + "loss": 0.7683, + "step": 21184 + }, + { + "epoch": 1.8052833404345974, + "grad_norm": 10.0, + "learning_rate": 5.850827804725928e-08, + "loss": 0.4351, + "step": 21185 + }, + { + "epoch": 1.8053685556028973, + "grad_norm": 14.875, + "learning_rate": 5.8457481357497015e-08, + "loss": 0.3737, + "step": 21186 + }, + { + "epoch": 1.8054537707711973, + "grad_norm": 11.3125, + "learning_rate": 5.840670620010619e-08, + "loss": 0.294, + "step": 21187 + }, + { + "epoch": 1.8055389859394972, + "grad_norm": 15.1875, + "learning_rate": 5.835595257600468e-08, + "loss": 0.5477, + "step": 21188 + }, + { + "epoch": 1.8056242011077972, + "grad_norm": 10.8125, + "learning_rate": 5.830522048610926e-08, + "loss": 0.3038, + "step": 21189 + }, + { + "epoch": 1.8057094162760972, + "grad_norm": 19.25, + "learning_rate": 5.8254509931336826e-08, + "loss": 0.3682, + "step": 21190 + }, + { + "epoch": 1.8057946314443971, + "grad_norm": 11.8125, + "learning_rate": 5.820382091260388e-08, + "loss": 0.3793, + "step": 21191 + }, + { + "epoch": 1.805879846612697, + "grad_norm": 10.9375, + "learning_rate": 5.81531534308262e-08, + "loss": 0.273, + "step": 21192 + }, + { + "epoch": 1.805965061780997, + "grad_norm": 17.125, + "learning_rate": 5.810250748691973e-08, + "loss": 0.4601, + "step": 21193 + }, + { + "epoch": 1.806050276949297, + "grad_norm": 12.9375, + "learning_rate": 5.805188308179929e-08, + "loss": 0.5265, + "step": 21194 + }, + { + "epoch": 1.806135492117597, + "grad_norm": 19.0, + "learning_rate": 5.8001280216379994e-08, + "loss": 0.4475, + "step": 21195 + }, + { + "epoch": 1.806220707285897, + "grad_norm": 15.1875, + "learning_rate": 5.795069889157637e-08, + "loss": 0.6086, + "step": 21196 + }, + { + "epoch": 1.8063059224541969, + "grad_norm": 18.125, + "learning_rate": 5.7900139108302e-08, + "loss": 1.2038, + "step": 21197 + }, + { + "epoch": 1.8063911376224968, + "grad_norm": 22.375, + "learning_rate": 5.7849600867471015e-08, + "loss": 0.8019, + "step": 21198 + }, + { + "epoch": 1.8064763527907968, + "grad_norm": 23.125, + "learning_rate": 5.7799084169996575e-08, + "loss": 0.7975, + "step": 21199 + }, + { + "epoch": 1.8065615679590967, + "grad_norm": 11.625, + "learning_rate": 5.774858901679156e-08, + "loss": 0.465, + "step": 21200 + }, + { + "epoch": 1.8066467831273967, + "grad_norm": 19.625, + "learning_rate": 5.7698115408768576e-08, + "loss": 0.6513, + "step": 21201 + }, + { + "epoch": 1.8067319982956966, + "grad_norm": 19.125, + "learning_rate": 5.764766334683924e-08, + "loss": 0.6677, + "step": 21202 + }, + { + "epoch": 1.8068172134639966, + "grad_norm": 15.25, + "learning_rate": 5.759723283191576e-08, + "loss": 0.7323, + "step": 21203 + }, + { + "epoch": 1.8069024286322966, + "grad_norm": 17.5, + "learning_rate": 5.7546823864909616e-08, + "loss": 0.4889, + "step": 21204 + }, + { + "epoch": 1.8069876438005965, + "grad_norm": 17.75, + "learning_rate": 5.74964364467312e-08, + "loss": 0.4655, + "step": 21205 + }, + { + "epoch": 1.8070728589688965, + "grad_norm": 19.125, + "learning_rate": 5.7446070578291296e-08, + "loss": 0.9548, + "step": 21206 + }, + { + "epoch": 1.8071580741371964, + "grad_norm": 10.6875, + "learning_rate": 5.7395726260500025e-08, + "loss": 0.2968, + "step": 21207 + }, + { + "epoch": 1.8072432893054964, + "grad_norm": 13.4375, + "learning_rate": 5.734540349426762e-08, + "loss": 0.5287, + "step": 21208 + }, + { + "epoch": 1.8073285044737963, + "grad_norm": 12.8125, + "learning_rate": 5.7295102280502945e-08, + "loss": 0.4295, + "step": 21209 + }, + { + "epoch": 1.8074137196420963, + "grad_norm": 12.8125, + "learning_rate": 5.7244822620114995e-08, + "loss": 0.292, + "step": 21210 + }, + { + "epoch": 1.8074989348103963, + "grad_norm": 24.75, + "learning_rate": 5.719456451401248e-08, + "loss": 0.7745, + "step": 21211 + }, + { + "epoch": 1.8075841499786962, + "grad_norm": 17.125, + "learning_rate": 5.7144327963103566e-08, + "loss": 0.6955, + "step": 21212 + }, + { + "epoch": 1.8076693651469962, + "grad_norm": 12.25, + "learning_rate": 5.7094112968296425e-08, + "loss": 0.3815, + "step": 21213 + }, + { + "epoch": 1.8077545803152961, + "grad_norm": 19.0, + "learning_rate": 5.704391953049809e-08, + "loss": 0.536, + "step": 21214 + }, + { + "epoch": 1.807839795483596, + "grad_norm": 18.375, + "learning_rate": 5.699374765061563e-08, + "loss": 0.7225, + "step": 21215 + }, + { + "epoch": 1.807925010651896, + "grad_norm": 28.625, + "learning_rate": 5.6943597329555953e-08, + "loss": 0.3959, + "step": 21216 + }, + { + "epoch": 1.808010225820196, + "grad_norm": 15.5625, + "learning_rate": 5.689346856822528e-08, + "loss": 0.8343, + "step": 21217 + }, + { + "epoch": 1.808095440988496, + "grad_norm": 14.625, + "learning_rate": 5.684336136752941e-08, + "loss": 0.6874, + "step": 21218 + }, + { + "epoch": 1.808180656156796, + "grad_norm": 14.875, + "learning_rate": 5.679327572837359e-08, + "loss": 0.5423, + "step": 21219 + }, + { + "epoch": 1.8082658713250959, + "grad_norm": 16.75, + "learning_rate": 5.6743211651663076e-08, + "loss": 0.5814, + "step": 21220 + }, + { + "epoch": 1.8083510864933958, + "grad_norm": 14.4375, + "learning_rate": 5.669316913830297e-08, + "loss": 0.5748, + "step": 21221 + }, + { + "epoch": 1.8084363016616958, + "grad_norm": 15.5, + "learning_rate": 5.6643148189197e-08, + "loss": 0.6341, + "step": 21222 + }, + { + "epoch": 1.8085215168299957, + "grad_norm": 19.375, + "learning_rate": 5.659314880524944e-08, + "loss": 0.3313, + "step": 21223 + }, + { + "epoch": 1.8086067319982957, + "grad_norm": 11.125, + "learning_rate": 5.654317098736359e-08, + "loss": 0.2741, + "step": 21224 + }, + { + "epoch": 1.8086919471665956, + "grad_norm": 14.4375, + "learning_rate": 5.649321473644304e-08, + "loss": 0.6152, + "step": 21225 + }, + { + "epoch": 1.8087771623348956, + "grad_norm": 14.9375, + "learning_rate": 5.6443280053390255e-08, + "loss": 0.411, + "step": 21226 + }, + { + "epoch": 1.8088623775031956, + "grad_norm": 15.6875, + "learning_rate": 5.639336693910744e-08, + "loss": 0.6349, + "step": 21227 + }, + { + "epoch": 1.8089475926714955, + "grad_norm": 11.75, + "learning_rate": 5.634347539449664e-08, + "loss": 0.3545, + "step": 21228 + }, + { + "epoch": 1.8090328078397955, + "grad_norm": 14.625, + "learning_rate": 5.629360542045978e-08, + "loss": 0.4897, + "step": 21229 + }, + { + "epoch": 1.8091180230080954, + "grad_norm": 14.9375, + "learning_rate": 5.62437570178978e-08, + "loss": 0.7002, + "step": 21230 + }, + { + "epoch": 1.8092032381763954, + "grad_norm": 16.875, + "learning_rate": 5.619393018771152e-08, + "loss": 0.2501, + "step": 21231 + }, + { + "epoch": 1.8092884533446953, + "grad_norm": 11.25, + "learning_rate": 5.614412493080132e-08, + "loss": 0.4144, + "step": 21232 + }, + { + "epoch": 1.8093736685129953, + "grad_norm": 16.625, + "learning_rate": 5.6094341248067454e-08, + "loss": 0.6416, + "step": 21233 + }, + { + "epoch": 1.8094588836812953, + "grad_norm": 19.0, + "learning_rate": 5.604457914040934e-08, + "loss": 0.38, + "step": 21234 + }, + { + "epoch": 1.8095440988495952, + "grad_norm": 13.0625, + "learning_rate": 5.599483860872626e-08, + "loss": 0.7383, + "step": 21235 + }, + { + "epoch": 1.8096293140178952, + "grad_norm": 11.0, + "learning_rate": 5.5945119653916943e-08, + "loss": 0.3988, + "step": 21236 + }, + { + "epoch": 1.8097145291861951, + "grad_norm": 25.875, + "learning_rate": 5.589542227688025e-08, + "loss": 0.7243, + "step": 21237 + }, + { + "epoch": 1.809799744354495, + "grad_norm": 18.375, + "learning_rate": 5.5845746478514066e-08, + "loss": 0.7721, + "step": 21238 + }, + { + "epoch": 1.809884959522795, + "grad_norm": 15.0, + "learning_rate": 5.579609225971588e-08, + "loss": 0.5977, + "step": 21239 + }, + { + "epoch": 1.809970174691095, + "grad_norm": 13.75, + "learning_rate": 5.574645962138331e-08, + "loss": 0.5725, + "step": 21240 + }, + { + "epoch": 1.810055389859395, + "grad_norm": 17.125, + "learning_rate": 5.569684856441312e-08, + "loss": 0.8047, + "step": 21241 + }, + { + "epoch": 1.810140605027695, + "grad_norm": 17.375, + "learning_rate": 5.56472590897017e-08, + "loss": 0.766, + "step": 21242 + }, + { + "epoch": 1.8102258201959949, + "grad_norm": 10.5, + "learning_rate": 5.559769119814567e-08, + "loss": 0.4415, + "step": 21243 + }, + { + "epoch": 1.8103110353642948, + "grad_norm": 15.9375, + "learning_rate": 5.554814489064003e-08, + "loss": 0.4318, + "step": 21244 + }, + { + "epoch": 1.8103962505325948, + "grad_norm": 14.25, + "learning_rate": 5.549862016808072e-08, + "loss": 0.4174, + "step": 21245 + }, + { + "epoch": 1.8104814657008947, + "grad_norm": 13.5, + "learning_rate": 5.544911703136244e-08, + "loss": 0.4747, + "step": 21246 + }, + { + "epoch": 1.8105666808691947, + "grad_norm": 20.25, + "learning_rate": 5.539963548137989e-08, + "loss": 0.6882, + "step": 21247 + }, + { + "epoch": 1.8106518960374947, + "grad_norm": 21.125, + "learning_rate": 5.535017551902708e-08, + "loss": 0.5219, + "step": 21248 + }, + { + "epoch": 1.8107371112057946, + "grad_norm": 18.25, + "learning_rate": 5.5300737145198016e-08, + "loss": 0.6248, + "step": 21249 + }, + { + "epoch": 1.8108223263740946, + "grad_norm": 19.625, + "learning_rate": 5.5251320360786007e-08, + "loss": 0.847, + "step": 21250 + }, + { + "epoch": 1.8109075415423945, + "grad_norm": 13.5, + "learning_rate": 5.52019251666841e-08, + "loss": 0.5029, + "step": 21251 + }, + { + "epoch": 1.8109927567106945, + "grad_norm": 12.6875, + "learning_rate": 5.515255156378477e-08, + "loss": 0.5523, + "step": 21252 + }, + { + "epoch": 1.8110779718789944, + "grad_norm": 16.25, + "learning_rate": 5.510319955298035e-08, + "loss": 0.5491, + "step": 21253 + }, + { + "epoch": 1.8111631870472944, + "grad_norm": 18.25, + "learning_rate": 5.505386913516264e-08, + "loss": 0.9061, + "step": 21254 + }, + { + "epoch": 1.8112484022155944, + "grad_norm": 16.75, + "learning_rate": 5.5004560311223285e-08, + "loss": 0.5282, + "step": 21255 + }, + { + "epoch": 1.8113336173838943, + "grad_norm": 13.6875, + "learning_rate": 5.495527308205295e-08, + "loss": 0.5139, + "step": 21256 + }, + { + "epoch": 1.8114188325521943, + "grad_norm": 13.875, + "learning_rate": 5.4906007448542606e-08, + "loss": 0.6256, + "step": 21257 + }, + { + "epoch": 1.8115040477204942, + "grad_norm": 22.125, + "learning_rate": 5.485676341158264e-08, + "loss": 1.0493, + "step": 21258 + }, + { + "epoch": 1.8115892628887942, + "grad_norm": 20.625, + "learning_rate": 5.4807540972062616e-08, + "loss": 0.7657, + "step": 21259 + }, + { + "epoch": 1.8116744780570941, + "grad_norm": 14.5625, + "learning_rate": 5.475834013087239e-08, + "loss": 0.4227, + "step": 21260 + }, + { + "epoch": 1.811759693225394, + "grad_norm": 15.0, + "learning_rate": 5.470916088890055e-08, + "loss": 0.5205, + "step": 21261 + }, + { + "epoch": 1.811844908393694, + "grad_norm": 14.75, + "learning_rate": 5.4660003247036245e-08, + "loss": 0.7306, + "step": 21262 + }, + { + "epoch": 1.811930123561994, + "grad_norm": 19.375, + "learning_rate": 5.461086720616765e-08, + "loss": 0.7887, + "step": 21263 + }, + { + "epoch": 1.812015338730294, + "grad_norm": 14.875, + "learning_rate": 5.456175276718295e-08, + "loss": 0.6164, + "step": 21264 + }, + { + "epoch": 1.812100553898594, + "grad_norm": 17.25, + "learning_rate": 5.451265993096935e-08, + "loss": 0.9021, + "step": 21265 + }, + { + "epoch": 1.8121857690668939, + "grad_norm": 13.5625, + "learning_rate": 5.44635886984142e-08, + "loss": 0.629, + "step": 21266 + }, + { + "epoch": 1.8122709842351938, + "grad_norm": 9.9375, + "learning_rate": 5.4414539070404144e-08, + "loss": 0.3214, + "step": 21267 + }, + { + "epoch": 1.8123561994034938, + "grad_norm": 15.0, + "learning_rate": 5.4365511047825695e-08, + "loss": 0.5054, + "step": 21268 + }, + { + "epoch": 1.8124414145717938, + "grad_norm": 18.25, + "learning_rate": 5.4316504631564544e-08, + "loss": 0.7605, + "step": 21269 + }, + { + "epoch": 1.8125266297400937, + "grad_norm": 16.125, + "learning_rate": 5.4267519822506634e-08, + "loss": 0.4561, + "step": 21270 + }, + { + "epoch": 1.8126118449083937, + "grad_norm": 17.875, + "learning_rate": 5.421855662153711e-08, + "loss": 0.7088, + "step": 21271 + }, + { + "epoch": 1.8126970600766936, + "grad_norm": 13.625, + "learning_rate": 5.416961502954066e-08, + "loss": 0.5098, + "step": 21272 + }, + { + "epoch": 1.8127822752449936, + "grad_norm": 13.625, + "learning_rate": 5.412069504740172e-08, + "loss": 0.4754, + "step": 21273 + }, + { + "epoch": 1.8128674904132935, + "grad_norm": 17.625, + "learning_rate": 5.40717966760046e-08, + "loss": 0.7427, + "step": 21274 + }, + { + "epoch": 1.8129527055815935, + "grad_norm": 13.25, + "learning_rate": 5.402291991623246e-08, + "loss": 0.4288, + "step": 21275 + }, + { + "epoch": 1.8130379207498935, + "grad_norm": 20.0, + "learning_rate": 5.3974064768968907e-08, + "loss": 0.8285, + "step": 21276 + }, + { + "epoch": 1.8131231359181934, + "grad_norm": 17.375, + "learning_rate": 5.392523123509671e-08, + "loss": 0.4472, + "step": 21277 + }, + { + "epoch": 1.8132083510864934, + "grad_norm": 13.5625, + "learning_rate": 5.387641931549822e-08, + "loss": 0.4207, + "step": 21278 + }, + { + "epoch": 1.8132935662547933, + "grad_norm": 20.5, + "learning_rate": 5.382762901105565e-08, + "loss": 0.8046, + "step": 21279 + }, + { + "epoch": 1.8133787814230933, + "grad_norm": 14.375, + "learning_rate": 5.377886032265078e-08, + "loss": 0.5426, + "step": 21280 + }, + { + "epoch": 1.8134639965913932, + "grad_norm": 18.75, + "learning_rate": 5.373011325116487e-08, + "loss": 0.9003, + "step": 21281 + }, + { + "epoch": 1.8135492117596932, + "grad_norm": 12.4375, + "learning_rate": 5.3681387797478724e-08, + "loss": 0.5048, + "step": 21282 + }, + { + "epoch": 1.8136344269279931, + "grad_norm": 15.375, + "learning_rate": 5.363268396247276e-08, + "loss": 0.6907, + "step": 21283 + }, + { + "epoch": 1.813719642096293, + "grad_norm": 20.5, + "learning_rate": 5.3584001747027105e-08, + "loss": 0.6764, + "step": 21284 + }, + { + "epoch": 1.813804857264593, + "grad_norm": 20.0, + "learning_rate": 5.353534115202203e-08, + "loss": 0.6199, + "step": 21285 + }, + { + "epoch": 1.813890072432893, + "grad_norm": 16.375, + "learning_rate": 5.348670217833612e-08, + "loss": 0.7251, + "step": 21286 + }, + { + "epoch": 1.813975287601193, + "grad_norm": 15.9375, + "learning_rate": 5.343808482684884e-08, + "loss": 0.8319, + "step": 21287 + }, + { + "epoch": 1.814060502769493, + "grad_norm": 18.125, + "learning_rate": 5.33894890984385e-08, + "loss": 0.5676, + "step": 21288 + }, + { + "epoch": 1.814145717937793, + "grad_norm": 15.0625, + "learning_rate": 5.3340914993983555e-08, + "loss": 0.6181, + "step": 21289 + }, + { + "epoch": 1.8142309331060928, + "grad_norm": 20.5, + "learning_rate": 5.329236251436165e-08, + "loss": 1.2265, + "step": 21290 + }, + { + "epoch": 1.8143161482743928, + "grad_norm": 13.125, + "learning_rate": 5.324383166044986e-08, + "loss": 0.5486, + "step": 21291 + }, + { + "epoch": 1.8144013634426928, + "grad_norm": 16.25, + "learning_rate": 5.319532243312553e-08, + "loss": 0.7016, + "step": 21292 + }, + { + "epoch": 1.8144865786109927, + "grad_norm": 17.25, + "learning_rate": 5.3146834833265326e-08, + "loss": 0.8644, + "step": 21293 + }, + { + "epoch": 1.8145717937792927, + "grad_norm": 16.75, + "learning_rate": 5.3098368861745355e-08, + "loss": 0.4157, + "step": 21294 + }, + { + "epoch": 1.8146570089475926, + "grad_norm": 20.375, + "learning_rate": 5.30499245194413e-08, + "loss": 0.8387, + "step": 21295 + }, + { + "epoch": 1.8147422241158926, + "grad_norm": 16.0, + "learning_rate": 5.300150180722871e-08, + "loss": 0.864, + "step": 21296 + }, + { + "epoch": 1.8148274392841925, + "grad_norm": 17.375, + "learning_rate": 5.2953100725982724e-08, + "loss": 0.6088, + "step": 21297 + }, + { + "epoch": 1.8149126544524925, + "grad_norm": 21.25, + "learning_rate": 5.2904721276578336e-08, + "loss": 1.0939, + "step": 21298 + }, + { + "epoch": 1.8149978696207925, + "grad_norm": 14.1875, + "learning_rate": 5.285636345988887e-08, + "loss": 0.6003, + "step": 21299 + }, + { + "epoch": 1.8150830847890924, + "grad_norm": 18.75, + "learning_rate": 5.280802727678891e-08, + "loss": 0.7013, + "step": 21300 + }, + { + "epoch": 1.8151682999573924, + "grad_norm": 11.9375, + "learning_rate": 5.275971272815164e-08, + "loss": 0.2607, + "step": 21301 + }, + { + "epoch": 1.8152535151256923, + "grad_norm": 15.25, + "learning_rate": 5.271141981485053e-08, + "loss": 0.589, + "step": 21302 + }, + { + "epoch": 1.8153387302939923, + "grad_norm": 14.1875, + "learning_rate": 5.2663148537757945e-08, + "loss": 0.586, + "step": 21303 + }, + { + "epoch": 1.8154239454622922, + "grad_norm": 16.125, + "learning_rate": 5.26148988977461e-08, + "loss": 0.7015, + "step": 21304 + }, + { + "epoch": 1.8155091606305922, + "grad_norm": 15.0625, + "learning_rate": 5.2566670895687205e-08, + "loss": 0.5755, + "step": 21305 + }, + { + "epoch": 1.8155943757988922, + "grad_norm": 14.4375, + "learning_rate": 5.2518464532452795e-08, + "loss": 0.6604, + "step": 21306 + }, + { + "epoch": 1.8156795909671921, + "grad_norm": 13.1875, + "learning_rate": 5.2470279808913975e-08, + "loss": 0.5393, + "step": 21307 + }, + { + "epoch": 1.815764806135492, + "grad_norm": 17.125, + "learning_rate": 5.2422116725941165e-08, + "loss": 0.7454, + "step": 21308 + }, + { + "epoch": 1.815850021303792, + "grad_norm": 15.3125, + "learning_rate": 5.237397528440491e-08, + "loss": 0.7634, + "step": 21309 + }, + { + "epoch": 1.815935236472092, + "grad_norm": 16.625, + "learning_rate": 5.2325855485175355e-08, + "loss": 0.5728, + "step": 21310 + }, + { + "epoch": 1.816020451640392, + "grad_norm": 12.5625, + "learning_rate": 5.2277757329122075e-08, + "loss": 0.5788, + "step": 21311 + }, + { + "epoch": 1.816105666808692, + "grad_norm": 19.25, + "learning_rate": 5.2229680817113966e-08, + "loss": 0.7389, + "step": 21312 + }, + { + "epoch": 1.8161908819769919, + "grad_norm": 11.8125, + "learning_rate": 5.218162595001991e-08, + "loss": 0.252, + "step": 21313 + }, + { + "epoch": 1.8162760971452918, + "grad_norm": 14.8125, + "learning_rate": 5.2133592728708524e-08, + "loss": 0.4949, + "step": 21314 + }, + { + "epoch": 1.8163613123135918, + "grad_norm": 40.25, + "learning_rate": 5.2085581154047714e-08, + "loss": 0.6069, + "step": 21315 + }, + { + "epoch": 1.8164465274818917, + "grad_norm": 12.625, + "learning_rate": 5.203759122690485e-08, + "loss": 0.5695, + "step": 21316 + }, + { + "epoch": 1.8165317426501917, + "grad_norm": 14.8125, + "learning_rate": 5.198962294814744e-08, + "loss": 0.7667, + "step": 21317 + }, + { + "epoch": 1.8166169578184916, + "grad_norm": 17.0, + "learning_rate": 5.194167631864227e-08, + "loss": 0.7167, + "step": 21318 + }, + { + "epoch": 1.8167021729867916, + "grad_norm": 19.125, + "learning_rate": 5.189375133925575e-08, + "loss": 0.8102, + "step": 21319 + }, + { + "epoch": 1.8167873881550916, + "grad_norm": 14.3125, + "learning_rate": 5.184584801085385e-08, + "loss": 0.4576, + "step": 21320 + }, + { + "epoch": 1.8168726033233915, + "grad_norm": 14.3125, + "learning_rate": 5.17979663343024e-08, + "loss": 0.5733, + "step": 21321 + }, + { + "epoch": 1.8169578184916915, + "grad_norm": 19.125, + "learning_rate": 5.17501063104664e-08, + "loss": 0.8762, + "step": 21322 + }, + { + "epoch": 1.8170430336599914, + "grad_norm": 14.875, + "learning_rate": 5.170226794021127e-08, + "loss": 0.5276, + "step": 21323 + }, + { + "epoch": 1.8171282488282914, + "grad_norm": 13.6875, + "learning_rate": 5.165445122440091e-08, + "loss": 0.2523, + "step": 21324 + }, + { + "epoch": 1.8172134639965913, + "grad_norm": 15.5625, + "learning_rate": 5.160665616389962e-08, + "loss": 0.75, + "step": 21325 + }, + { + "epoch": 1.8172986791648913, + "grad_norm": 17.125, + "learning_rate": 5.155888275957116e-08, + "loss": 0.3432, + "step": 21326 + }, + { + "epoch": 1.8173838943331915, + "grad_norm": 14.0625, + "learning_rate": 5.1511131012278863e-08, + "loss": 0.6225, + "step": 21327 + }, + { + "epoch": 1.8174691095014914, + "grad_norm": 12.0625, + "learning_rate": 5.146340092288565e-08, + "loss": 0.4061, + "step": 21328 + }, + { + "epoch": 1.8175543246697914, + "grad_norm": 31.875, + "learning_rate": 5.1415692492253895e-08, + "loss": 0.6328, + "step": 21329 + }, + { + "epoch": 1.8176395398380913, + "grad_norm": 12.75, + "learning_rate": 5.1368005721245814e-08, + "loss": 0.3743, + "step": 21330 + }, + { + "epoch": 1.8177247550063913, + "grad_norm": 19.375, + "learning_rate": 5.1320340610723356e-08, + "loss": 0.8362, + "step": 21331 + }, + { + "epoch": 1.8178099701746913, + "grad_norm": 16.75, + "learning_rate": 5.1272697161547644e-08, + "loss": 0.7192, + "step": 21332 + }, + { + "epoch": 1.8178951853429912, + "grad_norm": 14.1875, + "learning_rate": 5.122507537457966e-08, + "loss": 0.4297, + "step": 21333 + }, + { + "epoch": 1.8179804005112912, + "grad_norm": 17.125, + "learning_rate": 5.1177475250679955e-08, + "loss": 0.4724, + "step": 21334 + }, + { + "epoch": 1.8180656156795911, + "grad_norm": 15.1875, + "learning_rate": 5.112989679070882e-08, + "loss": 0.3734, + "step": 21335 + }, + { + "epoch": 1.818150830847891, + "grad_norm": 11.0, + "learning_rate": 5.1082339995526114e-08, + "loss": 0.2715, + "step": 21336 + }, + { + "epoch": 1.818236046016191, + "grad_norm": 19.5, + "learning_rate": 5.1034804865991024e-08, + "loss": 0.5687, + "step": 21337 + }, + { + "epoch": 1.818321261184491, + "grad_norm": 13.875, + "learning_rate": 5.098729140296271e-08, + "loss": 0.5952, + "step": 21338 + }, + { + "epoch": 1.818406476352791, + "grad_norm": 13.6875, + "learning_rate": 5.09397996072998e-08, + "loss": 0.593, + "step": 21339 + }, + { + "epoch": 1.818491691521091, + "grad_norm": 14.0, + "learning_rate": 5.089232947986036e-08, + "loss": 0.4833, + "step": 21340 + }, + { + "epoch": 1.8185769066893909, + "grad_norm": 14.25, + "learning_rate": 5.084488102150245e-08, + "loss": 0.3318, + "step": 21341 + }, + { + "epoch": 1.8186621218576908, + "grad_norm": 22.25, + "learning_rate": 5.07974542330833e-08, + "loss": 0.8225, + "step": 21342 + }, + { + "epoch": 1.8187473370259908, + "grad_norm": 17.5, + "learning_rate": 5.075004911546e-08, + "loss": 0.7787, + "step": 21343 + }, + { + "epoch": 1.8188325521942907, + "grad_norm": 16.0, + "learning_rate": 5.0702665669489365e-08, + "loss": 0.553, + "step": 21344 + }, + { + "epoch": 1.8189177673625907, + "grad_norm": 33.5, + "learning_rate": 5.065530389602738e-08, + "loss": 1.2756, + "step": 21345 + }, + { + "epoch": 1.8190029825308907, + "grad_norm": 25.0, + "learning_rate": 5.060796379593016e-08, + "loss": 0.6956, + "step": 21346 + }, + { + "epoch": 1.8190881976991906, + "grad_norm": 16.125, + "learning_rate": 5.0560645370053145e-08, + "loss": 0.6878, + "step": 21347 + }, + { + "epoch": 1.8191734128674906, + "grad_norm": 18.375, + "learning_rate": 5.051334861925134e-08, + "loss": 0.6867, + "step": 21348 + }, + { + "epoch": 1.8192586280357905, + "grad_norm": 16.5, + "learning_rate": 5.046607354437974e-08, + "loss": 0.6802, + "step": 21349 + }, + { + "epoch": 1.8193438432040905, + "grad_norm": 21.0, + "learning_rate": 5.0418820146292124e-08, + "loss": 0.6712, + "step": 21350 + }, + { + "epoch": 1.8194290583723904, + "grad_norm": 16.625, + "learning_rate": 5.037158842584281e-08, + "loss": 0.7883, + "step": 21351 + }, + { + "epoch": 1.8195142735406904, + "grad_norm": 10.3125, + "learning_rate": 5.032437838388513e-08, + "loss": 0.2326, + "step": 21352 + }, + { + "epoch": 1.8195994887089904, + "grad_norm": 15.3125, + "learning_rate": 5.0277190021272434e-08, + "loss": 0.397, + "step": 21353 + }, + { + "epoch": 1.8196847038772903, + "grad_norm": 20.375, + "learning_rate": 5.0230023338857236e-08, + "loss": 0.4478, + "step": 21354 + }, + { + "epoch": 1.8197699190455903, + "grad_norm": 27.5, + "learning_rate": 5.018287833749205e-08, + "loss": 0.6683, + "step": 21355 + }, + { + "epoch": 1.8198551342138902, + "grad_norm": 15.1875, + "learning_rate": 5.01357550180287e-08, + "loss": 0.6098, + "step": 21356 + }, + { + "epoch": 1.8199403493821902, + "grad_norm": 23.125, + "learning_rate": 5.008865338131885e-08, + "loss": 0.8253, + "step": 21357 + }, + { + "epoch": 1.8200255645504901, + "grad_norm": 21.375, + "learning_rate": 5.004157342821364e-08, + "loss": 0.9422, + "step": 21358 + }, + { + "epoch": 1.82011077971879, + "grad_norm": 16.375, + "learning_rate": 4.9994515159563775e-08, + "loss": 0.774, + "step": 21359 + }, + { + "epoch": 1.82019599488709, + "grad_norm": 20.125, + "learning_rate": 4.9947478576219816e-08, + "loss": 1.0129, + "step": 21360 + }, + { + "epoch": 1.82028121005539, + "grad_norm": 14.875, + "learning_rate": 4.990046367903165e-08, + "loss": 0.5755, + "step": 21361 + }, + { + "epoch": 1.82036642522369, + "grad_norm": 14.8125, + "learning_rate": 4.9853470468848864e-08, + "loss": 0.5216, + "step": 21362 + }, + { + "epoch": 1.82045164039199, + "grad_norm": 15.875, + "learning_rate": 4.980649894652079e-08, + "loss": 0.5713, + "step": 21363 + }, + { + "epoch": 1.8205368555602899, + "grad_norm": 13.75, + "learning_rate": 4.975954911289604e-08, + "loss": 0.6019, + "step": 21364 + }, + { + "epoch": 1.8206220707285898, + "grad_norm": 13.4375, + "learning_rate": 4.971262096882326e-08, + "loss": 0.5116, + "step": 21365 + }, + { + "epoch": 1.8207072858968898, + "grad_norm": 15.3125, + "learning_rate": 4.966571451515037e-08, + "loss": 0.8255, + "step": 21366 + }, + { + "epoch": 1.8207925010651898, + "grad_norm": 15.5, + "learning_rate": 4.961882975272503e-08, + "loss": 0.6264, + "step": 21367 + }, + { + "epoch": 1.8208777162334897, + "grad_norm": 13.625, + "learning_rate": 4.9571966682394476e-08, + "loss": 0.333, + "step": 21368 + }, + { + "epoch": 1.8209629314017897, + "grad_norm": 13.6875, + "learning_rate": 4.9525125305005664e-08, + "loss": 0.6245, + "step": 21369 + }, + { + "epoch": 1.8210481465700896, + "grad_norm": 30.25, + "learning_rate": 4.947830562140501e-08, + "loss": 0.9868, + "step": 21370 + }, + { + "epoch": 1.8211333617383896, + "grad_norm": 12.3125, + "learning_rate": 4.9431507632438785e-08, + "loss": 0.39, + "step": 21371 + }, + { + "epoch": 1.8212185769066895, + "grad_norm": 23.25, + "learning_rate": 4.938473133895214e-08, + "loss": 1.0266, + "step": 21372 + }, + { + "epoch": 1.8213037920749895, + "grad_norm": 11.9375, + "learning_rate": 4.933797674179092e-08, + "loss": 0.4355, + "step": 21373 + }, + { + "epoch": 1.8213890072432894, + "grad_norm": 13.25, + "learning_rate": 4.929124384179973e-08, + "loss": 0.3653, + "step": 21374 + }, + { + "epoch": 1.8214742224115894, + "grad_norm": 12.3125, + "learning_rate": 4.9244532639823176e-08, + "loss": 0.3776, + "step": 21375 + }, + { + "epoch": 1.8215594375798894, + "grad_norm": 15.5, + "learning_rate": 4.919784313670531e-08, + "loss": 0.8436, + "step": 21376 + }, + { + "epoch": 1.8216446527481893, + "grad_norm": 12.4375, + "learning_rate": 4.915117533329003e-08, + "loss": 0.3679, + "step": 21377 + }, + { + "epoch": 1.8217298679164893, + "grad_norm": 16.75, + "learning_rate": 4.910452923042056e-08, + "loss": 0.6665, + "step": 21378 + }, + { + "epoch": 1.8218150830847892, + "grad_norm": 14.1875, + "learning_rate": 4.9057904828939825e-08, + "loss": 0.5479, + "step": 21379 + }, + { + "epoch": 1.8219002982530892, + "grad_norm": 14.8125, + "learning_rate": 4.901130212969049e-08, + "loss": 0.6039, + "step": 21380 + }, + { + "epoch": 1.8219855134213891, + "grad_norm": 14.125, + "learning_rate": 4.8964721133514385e-08, + "loss": 0.3785, + "step": 21381 + }, + { + "epoch": 1.822070728589689, + "grad_norm": 17.0, + "learning_rate": 4.891816184125361e-08, + "loss": 1.0527, + "step": 21382 + }, + { + "epoch": 1.822155943757989, + "grad_norm": 14.375, + "learning_rate": 4.8871624253749586e-08, + "loss": 0.4976, + "step": 21383 + }, + { + "epoch": 1.822241158926289, + "grad_norm": 15.6875, + "learning_rate": 4.8825108371842875e-08, + "loss": 0.7907, + "step": 21384 + }, + { + "epoch": 1.822326374094589, + "grad_norm": 27.75, + "learning_rate": 4.877861419637434e-08, + "loss": 0.9878, + "step": 21385 + }, + { + "epoch": 1.822411589262889, + "grad_norm": 18.75, + "learning_rate": 4.8732141728184277e-08, + "loss": 0.6157, + "step": 21386 + }, + { + "epoch": 1.8224968044311889, + "grad_norm": 14.25, + "learning_rate": 4.868569096811243e-08, + "loss": 0.5093, + "step": 21387 + }, + { + "epoch": 1.8225820195994888, + "grad_norm": 11.1875, + "learning_rate": 4.8639261916998276e-08, + "loss": 0.3687, + "step": 21388 + }, + { + "epoch": 1.8226672347677888, + "grad_norm": 12.125, + "learning_rate": 4.859285457568044e-08, + "loss": 0.3571, + "step": 21389 + }, + { + "epoch": 1.8227524499360888, + "grad_norm": 11.5, + "learning_rate": 4.854646894499798e-08, + "loss": 0.4654, + "step": 21390 + }, + { + "epoch": 1.8228376651043887, + "grad_norm": 22.875, + "learning_rate": 4.850010502578897e-08, + "loss": 0.7046, + "step": 21391 + }, + { + "epoch": 1.8229228802726887, + "grad_norm": 11.875, + "learning_rate": 4.845376281889122e-08, + "loss": 0.4469, + "step": 21392 + }, + { + "epoch": 1.8230080954409886, + "grad_norm": 12.8125, + "learning_rate": 4.8407442325142246e-08, + "loss": 0.6002, + "step": 21393 + }, + { + "epoch": 1.8230933106092886, + "grad_norm": 13.375, + "learning_rate": 4.836114354537902e-08, + "loss": 0.4496, + "step": 21394 + }, + { + "epoch": 1.8231785257775885, + "grad_norm": 15.3125, + "learning_rate": 4.8314866480438375e-08, + "loss": 0.5457, + "step": 21395 + }, + { + "epoch": 1.8232637409458885, + "grad_norm": 18.375, + "learning_rate": 4.826861113115658e-08, + "loss": 0.9431, + "step": 21396 + }, + { + "epoch": 1.8233489561141885, + "grad_norm": 21.875, + "learning_rate": 4.822237749836922e-08, + "loss": 0.49, + "step": 21397 + }, + { + "epoch": 1.8234341712824884, + "grad_norm": 19.25, + "learning_rate": 4.8176165582912024e-08, + "loss": 0.7673, + "step": 21398 + }, + { + "epoch": 1.8235193864507884, + "grad_norm": 14.625, + "learning_rate": 4.8129975385620005e-08, + "loss": 0.5409, + "step": 21399 + }, + { + "epoch": 1.8236046016190883, + "grad_norm": 14.75, + "learning_rate": 4.8083806907328055e-08, + "loss": 0.5689, + "step": 21400 + }, + { + "epoch": 1.8236898167873883, + "grad_norm": 13.9375, + "learning_rate": 4.80376601488701e-08, + "loss": 0.4984, + "step": 21401 + }, + { + "epoch": 1.8237750319556882, + "grad_norm": 16.625, + "learning_rate": 4.7991535111080454e-08, + "loss": 0.9298, + "step": 21402 + }, + { + "epoch": 1.8238602471239882, + "grad_norm": 19.625, + "learning_rate": 4.7945431794792354e-08, + "loss": 0.8878, + "step": 21403 + }, + { + "epoch": 1.8239454622922882, + "grad_norm": 17.0, + "learning_rate": 4.789935020083944e-08, + "loss": 0.6657, + "step": 21404 + }, + { + "epoch": 1.8240306774605881, + "grad_norm": 16.5, + "learning_rate": 4.785329033005368e-08, + "loss": 0.7463, + "step": 21405 + }, + { + "epoch": 1.824115892628888, + "grad_norm": 19.25, + "learning_rate": 4.7807252183267873e-08, + "loss": 0.7621, + "step": 21406 + }, + { + "epoch": 1.824201107797188, + "grad_norm": 13.5625, + "learning_rate": 4.7761235761314e-08, + "loss": 0.4564, + "step": 21407 + }, + { + "epoch": 1.824286322965488, + "grad_norm": 20.875, + "learning_rate": 4.7715241065023624e-08, + "loss": 0.4771, + "step": 21408 + }, + { + "epoch": 1.824371538133788, + "grad_norm": 14.1875, + "learning_rate": 4.7669268095227594e-08, + "loss": 0.6978, + "step": 21409 + }, + { + "epoch": 1.824456753302088, + "grad_norm": 13.75, + "learning_rate": 4.762331685275706e-08, + "loss": 0.5189, + "step": 21410 + }, + { + "epoch": 1.8245419684703879, + "grad_norm": 11.25, + "learning_rate": 4.7577387338442196e-08, + "loss": 0.2768, + "step": 21411 + }, + { + "epoch": 1.8246271836386878, + "grad_norm": 19.5, + "learning_rate": 4.7531479553113296e-08, + "loss": 0.5551, + "step": 21412 + }, + { + "epoch": 1.8247123988069878, + "grad_norm": 12.5, + "learning_rate": 4.748559349759957e-08, + "loss": 0.4111, + "step": 21413 + }, + { + "epoch": 1.8247976139752877, + "grad_norm": 14.5625, + "learning_rate": 4.743972917273035e-08, + "loss": 0.5345, + "step": 21414 + }, + { + "epoch": 1.8248828291435877, + "grad_norm": 21.125, + "learning_rate": 4.7393886579334554e-08, + "loss": 1.0472, + "step": 21415 + }, + { + "epoch": 1.8249680443118876, + "grad_norm": 18.875, + "learning_rate": 4.734806571824055e-08, + "loss": 0.7113, + "step": 21416 + }, + { + "epoch": 1.8250532594801876, + "grad_norm": 16.25, + "learning_rate": 4.7302266590276424e-08, + "loss": 0.4739, + "step": 21417 + }, + { + "epoch": 1.8251384746484876, + "grad_norm": 15.4375, + "learning_rate": 4.7256489196269574e-08, + "loss": 0.5918, + "step": 21418 + }, + { + "epoch": 1.8252236898167875, + "grad_norm": 14.8125, + "learning_rate": 4.7210733537047525e-08, + "loss": 0.7054, + "step": 21419 + }, + { + "epoch": 1.8253089049850875, + "grad_norm": 19.0, + "learning_rate": 4.7164999613436986e-08, + "loss": 0.6973, + "step": 21420 + }, + { + "epoch": 1.8253941201533874, + "grad_norm": 24.875, + "learning_rate": 4.711928742626465e-08, + "loss": 0.8388, + "step": 21421 + }, + { + "epoch": 1.8254793353216874, + "grad_norm": 11.5625, + "learning_rate": 4.7073596976356103e-08, + "loss": 0.3528, + "step": 21422 + }, + { + "epoch": 1.8255645504899873, + "grad_norm": 14.3125, + "learning_rate": 4.702792826453736e-08, + "loss": 0.5413, + "step": 21423 + }, + { + "epoch": 1.8256497656582873, + "grad_norm": 24.875, + "learning_rate": 4.69822812916336e-08, + "loss": 0.9731, + "step": 21424 + }, + { + "epoch": 1.8257349808265873, + "grad_norm": 20.25, + "learning_rate": 4.693665605846998e-08, + "loss": 0.9214, + "step": 21425 + }, + { + "epoch": 1.8258201959948872, + "grad_norm": 17.25, + "learning_rate": 4.689105256587043e-08, + "loss": 0.5614, + "step": 21426 + }, + { + "epoch": 1.8259054111631872, + "grad_norm": 19.75, + "learning_rate": 4.6845470814659564e-08, + "loss": 0.9505, + "step": 21427 + }, + { + "epoch": 1.8259906263314871, + "grad_norm": 10.9375, + "learning_rate": 4.67999108056609e-08, + "loss": 0.269, + "step": 21428 + }, + { + "epoch": 1.826075841499787, + "grad_norm": 11.875, + "learning_rate": 4.6754372539697666e-08, + "loss": 0.4099, + "step": 21429 + }, + { + "epoch": 1.826161056668087, + "grad_norm": 22.25, + "learning_rate": 4.670885601759309e-08, + "loss": 0.8499, + "step": 21430 + }, + { + "epoch": 1.826246271836387, + "grad_norm": 17.5, + "learning_rate": 4.666336124016929e-08, + "loss": 0.4865, + "step": 21431 + }, + { + "epoch": 1.826331487004687, + "grad_norm": 37.25, + "learning_rate": 4.6617888208248667e-08, + "loss": 1.4276, + "step": 21432 + }, + { + "epoch": 1.826416702172987, + "grad_norm": 13.0625, + "learning_rate": 4.657243692265279e-08, + "loss": 0.6957, + "step": 21433 + }, + { + "epoch": 1.8265019173412869, + "grad_norm": 10.75, + "learning_rate": 4.652700738420335e-08, + "loss": 0.9205, + "step": 21434 + }, + { + "epoch": 1.8265871325095868, + "grad_norm": 15.6875, + "learning_rate": 4.648159959372081e-08, + "loss": 0.6468, + "step": 21435 + }, + { + "epoch": 1.8266723476778868, + "grad_norm": 21.875, + "learning_rate": 4.643621355202632e-08, + "loss": 0.8943, + "step": 21436 + }, + { + "epoch": 1.8267575628461867, + "grad_norm": 15.6875, + "learning_rate": 4.63908492599395e-08, + "loss": 0.5578, + "step": 21437 + }, + { + "epoch": 1.8268427780144867, + "grad_norm": 14.3125, + "learning_rate": 4.634550671828039e-08, + "loss": 0.5269, + "step": 21438 + }, + { + "epoch": 1.8269279931827866, + "grad_norm": 19.75, + "learning_rate": 4.630018592786833e-08, + "loss": 0.8209, + "step": 21439 + }, + { + "epoch": 1.8270132083510866, + "grad_norm": 13.3125, + "learning_rate": 4.625488688952226e-08, + "loss": 0.2515, + "step": 21440 + }, + { + "epoch": 1.8270984235193866, + "grad_norm": 17.625, + "learning_rate": 4.6209609604060957e-08, + "loss": 0.2577, + "step": 21441 + }, + { + "epoch": 1.8271836386876865, + "grad_norm": 13.875, + "learning_rate": 4.6164354072302517e-08, + "loss": 0.7431, + "step": 21442 + }, + { + "epoch": 1.8272688538559865, + "grad_norm": 16.5, + "learning_rate": 4.611912029506463e-08, + "loss": 0.8397, + "step": 21443 + }, + { + "epoch": 1.8273540690242864, + "grad_norm": 16.875, + "learning_rate": 4.6073908273164966e-08, + "loss": 0.785, + "step": 21444 + }, + { + "epoch": 1.8274392841925864, + "grad_norm": 16.75, + "learning_rate": 4.602871800742037e-08, + "loss": 0.4577, + "step": 21445 + }, + { + "epoch": 1.8275244993608863, + "grad_norm": 14.1875, + "learning_rate": 4.598354949864756e-08, + "loss": 0.484, + "step": 21446 + }, + { + "epoch": 1.8276097145291863, + "grad_norm": 18.125, + "learning_rate": 4.5938402747662814e-08, + "loss": 0.6348, + "step": 21447 + }, + { + "epoch": 1.8276949296974863, + "grad_norm": 23.5, + "learning_rate": 4.589327775528174e-08, + "loss": 0.9853, + "step": 21448 + }, + { + "epoch": 1.8277801448657862, + "grad_norm": 16.375, + "learning_rate": 4.584817452231993e-08, + "loss": 0.6923, + "step": 21449 + }, + { + "epoch": 1.8278653600340862, + "grad_norm": 22.875, + "learning_rate": 4.580309304959257e-08, + "loss": 0.7489, + "step": 21450 + }, + { + "epoch": 1.8279505752023861, + "grad_norm": 17.0, + "learning_rate": 4.5758033337914284e-08, + "loss": 0.6064, + "step": 21451 + }, + { + "epoch": 1.828035790370686, + "grad_norm": 16.75, + "learning_rate": 4.571299538809914e-08, + "loss": 0.3532, + "step": 21452 + }, + { + "epoch": 1.828121005538986, + "grad_norm": 13.625, + "learning_rate": 4.566797920096136e-08, + "loss": 0.4172, + "step": 21453 + }, + { + "epoch": 1.828206220707286, + "grad_norm": 28.25, + "learning_rate": 4.562298477731417e-08, + "loss": 1.0465, + "step": 21454 + }, + { + "epoch": 1.828291435875586, + "grad_norm": 15.375, + "learning_rate": 4.55780121179708e-08, + "loss": 0.6456, + "step": 21455 + }, + { + "epoch": 1.828376651043886, + "grad_norm": 12.25, + "learning_rate": 4.553306122374382e-08, + "loss": 0.3024, + "step": 21456 + }, + { + "epoch": 1.8284618662121859, + "grad_norm": 12.5625, + "learning_rate": 4.548813209544548e-08, + "loss": 0.4537, + "step": 21457 + }, + { + "epoch": 1.8285470813804858, + "grad_norm": 11.875, + "learning_rate": 4.5443224733887905e-08, + "loss": 0.4194, + "step": 21458 + }, + { + "epoch": 1.8286322965487858, + "grad_norm": 14.0, + "learning_rate": 4.539833913988267e-08, + "loss": 0.687, + "step": 21459 + }, + { + "epoch": 1.8287175117170857, + "grad_norm": 11.0625, + "learning_rate": 4.535347531424067e-08, + "loss": 0.291, + "step": 21460 + }, + { + "epoch": 1.8288027268853857, + "grad_norm": 23.625, + "learning_rate": 4.53086332577729e-08, + "loss": 0.9811, + "step": 21461 + }, + { + "epoch": 1.8288879420536857, + "grad_norm": 11.875, + "learning_rate": 4.5263812971289286e-08, + "loss": 0.3369, + "step": 21462 + }, + { + "epoch": 1.8289731572219856, + "grad_norm": 17.0, + "learning_rate": 4.5219014455600004e-08, + "loss": 0.5036, + "step": 21463 + }, + { + "epoch": 1.8290583723902856, + "grad_norm": 21.125, + "learning_rate": 4.5174237711514966e-08, + "loss": 0.5683, + "step": 21464 + }, + { + "epoch": 1.8291435875585855, + "grad_norm": 26.75, + "learning_rate": 4.512948273984269e-08, + "loss": 0.7892, + "step": 21465 + }, + { + "epoch": 1.8292288027268855, + "grad_norm": 11.4375, + "learning_rate": 4.5084749541392245e-08, + "loss": 0.4789, + "step": 21466 + }, + { + "epoch": 1.8293140178951854, + "grad_norm": 11.625, + "learning_rate": 4.50400381169723e-08, + "loss": 0.2607, + "step": 21467 + }, + { + "epoch": 1.8293992330634854, + "grad_norm": 13.75, + "learning_rate": 4.499534846739026e-08, + "loss": 0.4236, + "step": 21468 + }, + { + "epoch": 1.8294844482317854, + "grad_norm": 14.3125, + "learning_rate": 4.4950680593454226e-08, + "loss": 0.5723, + "step": 21469 + }, + { + "epoch": 1.8295696634000853, + "grad_norm": 16.875, + "learning_rate": 4.490603449597092e-08, + "loss": 0.5243, + "step": 21470 + }, + { + "epoch": 1.8296548785683853, + "grad_norm": 12.0, + "learning_rate": 4.4861410175747464e-08, + "loss": 0.4585, + "step": 21471 + }, + { + "epoch": 1.8297400937366852, + "grad_norm": 15.6875, + "learning_rate": 4.481680763359031e-08, + "loss": 0.5439, + "step": 21472 + }, + { + "epoch": 1.8298253089049852, + "grad_norm": 17.25, + "learning_rate": 4.477222687030519e-08, + "loss": 0.4801, + "step": 21473 + }, + { + "epoch": 1.8299105240732851, + "grad_norm": 24.875, + "learning_rate": 4.4727667886697996e-08, + "loss": 0.6309, + "step": 21474 + }, + { + "epoch": 1.829995739241585, + "grad_norm": 13.625, + "learning_rate": 4.468313068357377e-08, + "loss": 0.4806, + "step": 21475 + }, + { + "epoch": 1.830080954409885, + "grad_norm": 15.0, + "learning_rate": 4.4638615261737436e-08, + "loss": 0.2977, + "step": 21476 + }, + { + "epoch": 1.830166169578185, + "grad_norm": 13.625, + "learning_rate": 4.4594121621993615e-08, + "loss": 0.6979, + "step": 21477 + }, + { + "epoch": 1.830251384746485, + "grad_norm": 15.3125, + "learning_rate": 4.4549649765145834e-08, + "loss": 0.6475, + "step": 21478 + }, + { + "epoch": 1.830336599914785, + "grad_norm": 13.8125, + "learning_rate": 4.450519969199818e-08, + "loss": 0.4471, + "step": 21479 + }, + { + "epoch": 1.8304218150830849, + "grad_norm": 16.375, + "learning_rate": 4.4460771403353615e-08, + "loss": 0.4798, + "step": 21480 + }, + { + "epoch": 1.8305070302513848, + "grad_norm": 14.1875, + "learning_rate": 4.441636490001552e-08, + "loss": 0.6174, + "step": 21481 + }, + { + "epoch": 1.8305922454196848, + "grad_norm": 13.3125, + "learning_rate": 4.437198018278577e-08, + "loss": 0.4413, + "step": 21482 + }, + { + "epoch": 1.8306774605879848, + "grad_norm": 13.9375, + "learning_rate": 4.432761725246662e-08, + "loss": 0.5577, + "step": 21483 + }, + { + "epoch": 1.8307626757562847, + "grad_norm": 15.75, + "learning_rate": 4.4283276109860076e-08, + "loss": 0.5993, + "step": 21484 + }, + { + "epoch": 1.8308478909245847, + "grad_norm": 13.3125, + "learning_rate": 4.423895675576717e-08, + "loss": 0.4681, + "step": 21485 + }, + { + "epoch": 1.8309331060928846, + "grad_norm": 13.0, + "learning_rate": 4.4194659190988645e-08, + "loss": 0.279, + "step": 21486 + }, + { + "epoch": 1.8310183212611846, + "grad_norm": 20.125, + "learning_rate": 4.4150383416325106e-08, + "loss": 0.6261, + "step": 21487 + }, + { + "epoch": 1.8311035364294845, + "grad_norm": 15.1875, + "learning_rate": 4.410612943257675e-08, + "loss": 0.7691, + "step": 21488 + }, + { + "epoch": 1.8311887515977845, + "grad_norm": 20.0, + "learning_rate": 4.406189724054349e-08, + "loss": 0.6924, + "step": 21489 + }, + { + "epoch": 1.8312739667660844, + "grad_norm": 14.1875, + "learning_rate": 4.4017686841024145e-08, + "loss": 0.6117, + "step": 21490 + }, + { + "epoch": 1.8313591819343844, + "grad_norm": 21.75, + "learning_rate": 4.397349823481792e-08, + "loss": 0.7554, + "step": 21491 + }, + { + "epoch": 1.8314443971026844, + "grad_norm": 13.9375, + "learning_rate": 4.3929331422723345e-08, + "loss": 0.5587, + "step": 21492 + }, + { + "epoch": 1.8315296122709843, + "grad_norm": 13.75, + "learning_rate": 4.388518640553868e-08, + "loss": 0.5387, + "step": 21493 + }, + { + "epoch": 1.8316148274392843, + "grad_norm": 14.5625, + "learning_rate": 4.3841063184061615e-08, + "loss": 0.4323, + "step": 21494 + }, + { + "epoch": 1.8317000426075842, + "grad_norm": 13.3125, + "learning_rate": 4.379696175908929e-08, + "loss": 0.5678, + "step": 21495 + }, + { + "epoch": 1.8317852577758842, + "grad_norm": 18.625, + "learning_rate": 4.375288213141871e-08, + "loss": 0.5798, + "step": 21496 + }, + { + "epoch": 1.8318704729441841, + "grad_norm": 20.5, + "learning_rate": 4.3708824301846723e-08, + "loss": 0.7563, + "step": 21497 + }, + { + "epoch": 1.831955688112484, + "grad_norm": 20.125, + "learning_rate": 4.36647882711691e-08, + "loss": 0.7358, + "step": 21498 + }, + { + "epoch": 1.832040903280784, + "grad_norm": 20.625, + "learning_rate": 4.3620774040181865e-08, + "loss": 0.9437, + "step": 21499 + }, + { + "epoch": 1.832126118449084, + "grad_norm": 22.0, + "learning_rate": 4.3576781609680346e-08, + "loss": 0.7043, + "step": 21500 + }, + { + "epoch": 1.832211333617384, + "grad_norm": 11.6875, + "learning_rate": 4.353281098045961e-08, + "loss": 0.4967, + "step": 21501 + }, + { + "epoch": 1.832296548785684, + "grad_norm": 22.125, + "learning_rate": 4.3488862153314296e-08, + "loss": 0.8167, + "step": 21502 + }, + { + "epoch": 1.8323817639539839, + "grad_norm": 14.75, + "learning_rate": 4.3444935129038216e-08, + "loss": 0.6233, + "step": 21503 + }, + { + "epoch": 1.8324669791222838, + "grad_norm": 13.625, + "learning_rate": 4.3401029908425596e-08, + "loss": 0.5071, + "step": 21504 + }, + { + "epoch": 1.8325521942905838, + "grad_norm": 36.25, + "learning_rate": 4.3357146492269545e-08, + "loss": 0.7682, + "step": 21505 + }, + { + "epoch": 1.8326374094588838, + "grad_norm": 15.375, + "learning_rate": 4.331328488136347e-08, + "loss": 0.366, + "step": 21506 + }, + { + "epoch": 1.8327226246271837, + "grad_norm": 11.375, + "learning_rate": 4.3269445076499636e-08, + "loss": 0.3402, + "step": 21507 + }, + { + "epoch": 1.8328078397954837, + "grad_norm": 13.875, + "learning_rate": 4.322562707847033e-08, + "loss": 0.7265, + "step": 21508 + }, + { + "epoch": 1.8328930549637836, + "grad_norm": 19.0, + "learning_rate": 4.318183088806757e-08, + "loss": 0.7686, + "step": 21509 + }, + { + "epoch": 1.8329782701320836, + "grad_norm": 14.25, + "learning_rate": 4.313805650608252e-08, + "loss": 0.6415, + "step": 21510 + }, + { + "epoch": 1.8330634853003835, + "grad_norm": 22.0, + "learning_rate": 4.309430393330663e-08, + "loss": 0.7169, + "step": 21511 + }, + { + "epoch": 1.8331487004686835, + "grad_norm": 14.3125, + "learning_rate": 4.3050573170529966e-08, + "loss": 0.506, + "step": 21512 + }, + { + "epoch": 1.8332339156369835, + "grad_norm": 15.4375, + "learning_rate": 4.3006864218543274e-08, + "loss": 0.5407, + "step": 21513 + }, + { + "epoch": 1.8333191308052834, + "grad_norm": 17.75, + "learning_rate": 4.296317707813649e-08, + "loss": 0.768, + "step": 21514 + }, + { + "epoch": 1.8334043459735834, + "grad_norm": 18.0, + "learning_rate": 4.2919511750098566e-08, + "loss": 1.1035, + "step": 21515 + }, + { + "epoch": 1.8334895611418833, + "grad_norm": 16.75, + "learning_rate": 4.2875868235218995e-08, + "loss": 0.6447, + "step": 21516 + }, + { + "epoch": 1.8335747763101833, + "grad_norm": 11.4375, + "learning_rate": 4.283224653428619e-08, + "loss": 0.2259, + "step": 21517 + }, + { + "epoch": 1.8336599914784832, + "grad_norm": 12.9375, + "learning_rate": 4.278864664808882e-08, + "loss": 0.3799, + "step": 21518 + }, + { + "epoch": 1.8337452066467832, + "grad_norm": 14.8125, + "learning_rate": 4.274506857741459e-08, + "loss": 0.6836, + "step": 21519 + }, + { + "epoch": 1.8338304218150832, + "grad_norm": 19.5, + "learning_rate": 4.270151232305078e-08, + "loss": 0.8429, + "step": 21520 + }, + { + "epoch": 1.8339156369833831, + "grad_norm": 17.375, + "learning_rate": 4.2657977885784685e-08, + "loss": 0.7382, + "step": 21521 + }, + { + "epoch": 1.834000852151683, + "grad_norm": 18.125, + "learning_rate": 4.261446526640303e-08, + "loss": 0.7791, + "step": 21522 + }, + { + "epoch": 1.834086067319983, + "grad_norm": 11.1875, + "learning_rate": 4.257097446569228e-08, + "loss": 0.357, + "step": 21523 + }, + { + "epoch": 1.834171282488283, + "grad_norm": 12.9375, + "learning_rate": 4.25275054844379e-08, + "loss": 0.508, + "step": 21524 + }, + { + "epoch": 1.834256497656583, + "grad_norm": 14.9375, + "learning_rate": 4.248405832342581e-08, + "loss": 0.4457, + "step": 21525 + }, + { + "epoch": 1.834341712824883, + "grad_norm": 12.125, + "learning_rate": 4.24406329834412e-08, + "loss": 0.3831, + "step": 21526 + }, + { + "epoch": 1.8344269279931829, + "grad_norm": 15.375, + "learning_rate": 4.2397229465268444e-08, + "loss": 0.5507, + "step": 21527 + }, + { + "epoch": 1.8345121431614828, + "grad_norm": 14.3125, + "learning_rate": 4.23538477696922e-08, + "loss": 0.484, + "step": 21528 + }, + { + "epoch": 1.8345973583297828, + "grad_norm": 14.375, + "learning_rate": 4.2310487897496284e-08, + "loss": 0.7808, + "step": 21529 + }, + { + "epoch": 1.8346825734980827, + "grad_norm": 17.75, + "learning_rate": 4.226714984946423e-08, + "loss": 0.6765, + "step": 21530 + }, + { + "epoch": 1.8347677886663827, + "grad_norm": 14.625, + "learning_rate": 4.22238336263793e-08, + "loss": 0.7523, + "step": 21531 + }, + { + "epoch": 1.8348530038346826, + "grad_norm": 14.75, + "learning_rate": 4.218053922902421e-08, + "loss": 0.463, + "step": 21532 + }, + { + "epoch": 1.8349382190029826, + "grad_norm": 15.0625, + "learning_rate": 4.213726665818124e-08, + "loss": 0.5986, + "step": 21533 + }, + { + "epoch": 1.8350234341712826, + "grad_norm": 16.75, + "learning_rate": 4.209401591463255e-08, + "loss": 0.7107, + "step": 21534 + }, + { + "epoch": 1.8351086493395825, + "grad_norm": 15.3125, + "learning_rate": 4.2050786999159596e-08, + "loss": 0.44, + "step": 21535 + }, + { + "epoch": 1.8351938645078825, + "grad_norm": 12.0625, + "learning_rate": 4.20075799125437e-08, + "loss": 0.424, + "step": 21536 + }, + { + "epoch": 1.8352790796761824, + "grad_norm": 9.6875, + "learning_rate": 4.196439465556548e-08, + "loss": 0.188, + "step": 21537 + }, + { + "epoch": 1.8353642948444824, + "grad_norm": 21.0, + "learning_rate": 4.1921231229005435e-08, + "loss": 0.6415, + "step": 21538 + }, + { + "epoch": 1.8354495100127823, + "grad_norm": 20.625, + "learning_rate": 4.187808963364348e-08, + "loss": 0.8235, + "step": 21539 + }, + { + "epoch": 1.8355347251810823, + "grad_norm": 20.125, + "learning_rate": 4.183496987025942e-08, + "loss": 0.9157, + "step": 21540 + }, + { + "epoch": 1.8356199403493823, + "grad_norm": 17.125, + "learning_rate": 4.179187193963236e-08, + "loss": 0.6801, + "step": 21541 + }, + { + "epoch": 1.8357051555176822, + "grad_norm": 14.75, + "learning_rate": 4.17487958425411e-08, + "loss": 0.4658, + "step": 21542 + }, + { + "epoch": 1.8357903706859822, + "grad_norm": 16.0, + "learning_rate": 4.1705741579764065e-08, + "loss": 0.5891, + "step": 21543 + }, + { + "epoch": 1.8358755858542821, + "grad_norm": 10.5625, + "learning_rate": 4.166270915207937e-08, + "loss": 0.2203, + "step": 21544 + }, + { + "epoch": 1.835960801022582, + "grad_norm": 17.25, + "learning_rate": 4.16196985602646e-08, + "loss": 0.4637, + "step": 21545 + }, + { + "epoch": 1.836046016190882, + "grad_norm": 18.625, + "learning_rate": 4.15767098050969e-08, + "loss": 0.7447, + "step": 21546 + }, + { + "epoch": 1.836131231359182, + "grad_norm": 15.125, + "learning_rate": 4.153374288735315e-08, + "loss": 0.672, + "step": 21547 + }, + { + "epoch": 1.836216446527482, + "grad_norm": 13.8125, + "learning_rate": 4.1490797807810096e-08, + "loss": 0.3447, + "step": 21548 + }, + { + "epoch": 1.836301661695782, + "grad_norm": 18.5, + "learning_rate": 4.144787456724336e-08, + "loss": 0.5308, + "step": 21549 + }, + { + "epoch": 1.8363868768640819, + "grad_norm": 14.875, + "learning_rate": 4.140497316642914e-08, + "loss": 0.7252, + "step": 21550 + }, + { + "epoch": 1.8364720920323818, + "grad_norm": 16.625, + "learning_rate": 4.1362093606142074e-08, + "loss": 0.5289, + "step": 21551 + }, + { + "epoch": 1.8365573072006818, + "grad_norm": 11.875, + "learning_rate": 4.131923588715739e-08, + "loss": 0.2394, + "step": 21552 + }, + { + "epoch": 1.8366425223689817, + "grad_norm": 12.4375, + "learning_rate": 4.127640001024974e-08, + "loss": 0.3451, + "step": 21553 + }, + { + "epoch": 1.8367277375372817, + "grad_norm": 14.875, + "learning_rate": 4.123358597619281e-08, + "loss": 0.6578, + "step": 21554 + }, + { + "epoch": 1.8368129527055816, + "grad_norm": 11.5, + "learning_rate": 4.1190793785760574e-08, + "loss": 0.3202, + "step": 21555 + }, + { + "epoch": 1.8368981678738816, + "grad_norm": 26.625, + "learning_rate": 4.114802343972615e-08, + "loss": 0.6101, + "step": 21556 + }, + { + "epoch": 1.8369833830421816, + "grad_norm": 16.75, + "learning_rate": 4.1105274938862685e-08, + "loss": 0.6658, + "step": 21557 + }, + { + "epoch": 1.8370685982104815, + "grad_norm": 13.8125, + "learning_rate": 4.10625482839426e-08, + "loss": 0.4154, + "step": 21558 + }, + { + "epoch": 1.8371538133787815, + "grad_norm": 12.75, + "learning_rate": 4.101984347573779e-08, + "loss": 0.4266, + "step": 21559 + }, + { + "epoch": 1.8372390285470814, + "grad_norm": 15.6875, + "learning_rate": 4.097716051501999e-08, + "loss": 0.3439, + "step": 21560 + }, + { + "epoch": 1.8373242437153814, + "grad_norm": 17.5, + "learning_rate": 4.093449940256094e-08, + "loss": 1.0169, + "step": 21561 + }, + { + "epoch": 1.8374094588836813, + "grad_norm": 17.5, + "learning_rate": 4.089186013913116e-08, + "loss": 0.4537, + "step": 21562 + }, + { + "epoch": 1.8374946740519813, + "grad_norm": 15.9375, + "learning_rate": 4.084924272550139e-08, + "loss": 0.713, + "step": 21563 + }, + { + "epoch": 1.8375798892202813, + "grad_norm": 14.4375, + "learning_rate": 4.0806647162441596e-08, + "loss": 0.6854, + "step": 21564 + }, + { + "epoch": 1.8376651043885812, + "grad_norm": 13.4375, + "learning_rate": 4.0764073450721697e-08, + "loss": 0.4051, + "step": 21565 + }, + { + "epoch": 1.8377503195568812, + "grad_norm": 20.125, + "learning_rate": 4.072152159111109e-08, + "loss": 0.8643, + "step": 21566 + }, + { + "epoch": 1.8378355347251811, + "grad_norm": 21.5, + "learning_rate": 4.067899158437846e-08, + "loss": 0.8252, + "step": 21567 + }, + { + "epoch": 1.837920749893481, + "grad_norm": 20.625, + "learning_rate": 4.0636483431292507e-08, + "loss": 0.7776, + "step": 21568 + }, + { + "epoch": 1.838005965061781, + "grad_norm": 15.4375, + "learning_rate": 4.059399713262135e-08, + "loss": 0.5886, + "step": 21569 + }, + { + "epoch": 1.838091180230081, + "grad_norm": 14.5625, + "learning_rate": 4.0551532689133004e-08, + "loss": 0.528, + "step": 21570 + }, + { + "epoch": 1.838176395398381, + "grad_norm": 16.125, + "learning_rate": 4.0509090101594486e-08, + "loss": 1.032, + "step": 21571 + }, + { + "epoch": 1.838261610566681, + "grad_norm": 16.875, + "learning_rate": 4.046666937077309e-08, + "loss": 0.2717, + "step": 21572 + }, + { + "epoch": 1.8383468257349809, + "grad_norm": 11.5, + "learning_rate": 4.0424270497435165e-08, + "loss": 0.5418, + "step": 21573 + }, + { + "epoch": 1.8384320409032808, + "grad_norm": 13.3125, + "learning_rate": 4.0381893482347166e-08, + "loss": 0.5199, + "step": 21574 + }, + { + "epoch": 1.8385172560715808, + "grad_norm": 10.25, + "learning_rate": 4.0339538326274604e-08, + "loss": 0.2337, + "step": 21575 + }, + { + "epoch": 1.8386024712398807, + "grad_norm": 14.625, + "learning_rate": 4.0297205029982825e-08, + "loss": 0.5854, + "step": 21576 + }, + { + "epoch": 1.8386876864081807, + "grad_norm": 11.5, + "learning_rate": 4.0254893594237064e-08, + "loss": 1.0806, + "step": 21577 + }, + { + "epoch": 1.8387729015764807, + "grad_norm": 19.625, + "learning_rate": 4.021260401980212e-08, + "loss": 0.6252, + "step": 21578 + }, + { + "epoch": 1.8388581167447806, + "grad_norm": 18.0, + "learning_rate": 4.017033630744169e-08, + "loss": 0.8195, + "step": 21579 + }, + { + "epoch": 1.8389433319130806, + "grad_norm": 14.1875, + "learning_rate": 4.012809045791988e-08, + "loss": 0.43, + "step": 21580 + }, + { + "epoch": 1.8390285470813805, + "grad_norm": 14.0625, + "learning_rate": 4.0085866472000106e-08, + "loss": 0.4873, + "step": 21581 + }, + { + "epoch": 1.8391137622496805, + "grad_norm": 16.875, + "learning_rate": 4.004366435044566e-08, + "loss": 0.4081, + "step": 21582 + }, + { + "epoch": 1.8391989774179804, + "grad_norm": 12.9375, + "learning_rate": 4.000148409401869e-08, + "loss": 0.4974, + "step": 21583 + }, + { + "epoch": 1.8392841925862804, + "grad_norm": 21.125, + "learning_rate": 3.9959325703481656e-08, + "loss": 0.9991, + "step": 21584 + }, + { + "epoch": 1.8393694077545804, + "grad_norm": 18.625, + "learning_rate": 3.9917189179596307e-08, + "loss": 1.0739, + "step": 21585 + }, + { + "epoch": 1.8394546229228803, + "grad_norm": 12.9375, + "learning_rate": 3.9875074523124255e-08, + "loss": 0.4647, + "step": 21586 + }, + { + "epoch": 1.8395398380911803, + "grad_norm": 14.0625, + "learning_rate": 3.983298173482655e-08, + "loss": 0.3774, + "step": 21587 + }, + { + "epoch": 1.8396250532594802, + "grad_norm": 14.25, + "learning_rate": 3.979091081546371e-08, + "loss": 0.5796, + "step": 21588 + }, + { + "epoch": 1.8397102684277802, + "grad_norm": 19.625, + "learning_rate": 3.9748861765795956e-08, + "loss": 0.9681, + "step": 21589 + }, + { + "epoch": 1.8397954835960801, + "grad_norm": 16.5, + "learning_rate": 3.9706834586583365e-08, + "loss": 0.3252, + "step": 21590 + }, + { + "epoch": 1.83988069876438, + "grad_norm": 13.4375, + "learning_rate": 3.9664829278585616e-08, + "loss": 0.3255, + "step": 21591 + }, + { + "epoch": 1.83996591393268, + "grad_norm": 16.875, + "learning_rate": 3.9622845842561266e-08, + "loss": 0.6017, + "step": 21592 + }, + { + "epoch": 1.84005112910098, + "grad_norm": 15.4375, + "learning_rate": 3.958088427926901e-08, + "loss": 0.5483, + "step": 21593 + }, + { + "epoch": 1.84013634426928, + "grad_norm": 15.0625, + "learning_rate": 3.9538944589467555e-08, + "loss": 0.6672, + "step": 21594 + }, + { + "epoch": 1.84022155943758, + "grad_norm": 12.5, + "learning_rate": 3.949702677391462e-08, + "loss": 0.2627, + "step": 21595 + }, + { + "epoch": 1.8403067746058799, + "grad_norm": 14.3125, + "learning_rate": 3.945513083336752e-08, + "loss": 0.5056, + "step": 21596 + }, + { + "epoch": 1.8403919897741798, + "grad_norm": 12.75, + "learning_rate": 3.941325676858357e-08, + "loss": 0.3087, + "step": 21597 + }, + { + "epoch": 1.8404772049424798, + "grad_norm": 16.625, + "learning_rate": 3.937140458031938e-08, + "loss": 0.5177, + "step": 21598 + }, + { + "epoch": 1.8405624201107798, + "grad_norm": 13.4375, + "learning_rate": 3.932957426933143e-08, + "loss": 0.388, + "step": 21599 + }, + { + "epoch": 1.8406476352790797, + "grad_norm": 16.75, + "learning_rate": 3.928776583637536e-08, + "loss": 0.5433, + "step": 21600 + }, + { + "epoch": 1.8407328504473797, + "grad_norm": 11.25, + "learning_rate": 3.924597928220683e-08, + "loss": 0.3943, + "step": 21601 + }, + { + "epoch": 1.8408180656156796, + "grad_norm": 20.5, + "learning_rate": 3.920421460758092e-08, + "loss": 0.8793, + "step": 21602 + }, + { + "epoch": 1.8409032807839796, + "grad_norm": 18.75, + "learning_rate": 3.9162471813252444e-08, + "loss": 0.7447, + "step": 21603 + }, + { + "epoch": 1.8409884959522795, + "grad_norm": 14.5, + "learning_rate": 3.91207508999758e-08, + "loss": 0.5494, + "step": 21604 + }, + { + "epoch": 1.8410737111205795, + "grad_norm": 10.25, + "learning_rate": 3.907905186850469e-08, + "loss": 0.3094, + "step": 21605 + }, + { + "epoch": 1.8411589262888794, + "grad_norm": 11.3125, + "learning_rate": 3.903737471959282e-08, + "loss": 0.3316, + "step": 21606 + }, + { + "epoch": 1.8412441414571794, + "grad_norm": 19.75, + "learning_rate": 3.899571945399347e-08, + "loss": 0.6314, + "step": 21607 + }, + { + "epoch": 1.8413293566254794, + "grad_norm": 17.5, + "learning_rate": 3.8954086072459096e-08, + "loss": 0.7562, + "step": 21608 + }, + { + "epoch": 1.8414145717937793, + "grad_norm": 16.5, + "learning_rate": 3.8912474575742155e-08, + "loss": 0.8842, + "step": 21609 + }, + { + "epoch": 1.8414997869620793, + "grad_norm": 12.75, + "learning_rate": 3.887088496459468e-08, + "loss": 0.3192, + "step": 21610 + }, + { + "epoch": 1.8415850021303792, + "grad_norm": 13.9375, + "learning_rate": 3.8829317239768017e-08, + "loss": 0.4573, + "step": 21611 + }, + { + "epoch": 1.8416702172986792, + "grad_norm": 18.75, + "learning_rate": 3.878777140201379e-08, + "loss": 0.7371, + "step": 21612 + }, + { + "epoch": 1.8417554324669791, + "grad_norm": 11.125, + "learning_rate": 3.874624745208236e-08, + "loss": 0.3375, + "step": 21613 + }, + { + "epoch": 1.841840647635279, + "grad_norm": 13.5625, + "learning_rate": 3.8704745390724244e-08, + "loss": 0.4617, + "step": 21614 + }, + { + "epoch": 1.841925862803579, + "grad_norm": 14.25, + "learning_rate": 3.866326521868954e-08, + "loss": 0.4934, + "step": 21615 + }, + { + "epoch": 1.842011077971879, + "grad_norm": 12.6875, + "learning_rate": 3.862180693672765e-08, + "loss": 0.4485, + "step": 21616 + }, + { + "epoch": 1.842096293140179, + "grad_norm": 13.75, + "learning_rate": 3.858037054558797e-08, + "loss": 0.3848, + "step": 21617 + }, + { + "epoch": 1.842181508308479, + "grad_norm": 21.5, + "learning_rate": 3.853895604601893e-08, + "loss": 0.9219, + "step": 21618 + }, + { + "epoch": 1.842266723476779, + "grad_norm": 16.5, + "learning_rate": 3.849756343876923e-08, + "loss": 0.9358, + "step": 21619 + }, + { + "epoch": 1.8423519386450788, + "grad_norm": 15.375, + "learning_rate": 3.8456192724586895e-08, + "loss": 0.6992, + "step": 21620 + }, + { + "epoch": 1.8424371538133788, + "grad_norm": 14.4375, + "learning_rate": 3.841484390421951e-08, + "loss": 0.7419, + "step": 21621 + }, + { + "epoch": 1.8425223689816788, + "grad_norm": 13.75, + "learning_rate": 3.8373516978414125e-08, + "loss": 0.5878, + "step": 21622 + }, + { + "epoch": 1.8426075841499787, + "grad_norm": 17.125, + "learning_rate": 3.833221194791778e-08, + "loss": 0.5279, + "step": 21623 + }, + { + "epoch": 1.8426927993182787, + "grad_norm": 16.375, + "learning_rate": 3.829092881347668e-08, + "loss": 0.5738, + "step": 21624 + }, + { + "epoch": 1.8427780144865786, + "grad_norm": 12.0625, + "learning_rate": 3.824966757583704e-08, + "loss": 0.5105, + "step": 21625 + }, + { + "epoch": 1.8428632296548786, + "grad_norm": 16.75, + "learning_rate": 3.820842823574436e-08, + "loss": 0.6726, + "step": 21626 + }, + { + "epoch": 1.8429484448231785, + "grad_norm": 13.8125, + "learning_rate": 3.8167210793943894e-08, + "loss": 0.5963, + "step": 21627 + }, + { + "epoch": 1.8430336599914785, + "grad_norm": 17.125, + "learning_rate": 3.8126015251180594e-08, + "loss": 0.5211, + "step": 21628 + }, + { + "epoch": 1.8431188751597785, + "grad_norm": 13.6875, + "learning_rate": 3.8084841608198864e-08, + "loss": 0.4285, + "step": 21629 + }, + { + "epoch": 1.8432040903280784, + "grad_norm": 16.75, + "learning_rate": 3.80436898657427e-08, + "loss": 0.6153, + "step": 21630 + }, + { + "epoch": 1.8432893054963784, + "grad_norm": 11.3125, + "learning_rate": 3.8002560024555794e-08, + "loss": 0.3088, + "step": 21631 + }, + { + "epoch": 1.8433745206646783, + "grad_norm": 18.0, + "learning_rate": 3.796145208538132e-08, + "loss": 0.9886, + "step": 21632 + }, + { + "epoch": 1.8434597358329783, + "grad_norm": 14.0, + "learning_rate": 3.792036604896229e-08, + "loss": 0.7846, + "step": 21633 + }, + { + "epoch": 1.8435449510012782, + "grad_norm": 16.25, + "learning_rate": 3.787930191604103e-08, + "loss": 0.6192, + "step": 21634 + }, + { + "epoch": 1.8436301661695782, + "grad_norm": 25.875, + "learning_rate": 3.783825968735971e-08, + "loss": 1.1105, + "step": 21635 + }, + { + "epoch": 1.8437153813378782, + "grad_norm": 20.5, + "learning_rate": 3.7797239363659836e-08, + "loss": 0.4376, + "step": 21636 + }, + { + "epoch": 1.8438005965061781, + "grad_norm": 24.0, + "learning_rate": 3.775624094568289e-08, + "loss": 0.7224, + "step": 21637 + }, + { + "epoch": 1.843885811674478, + "grad_norm": 11.625, + "learning_rate": 3.7715264434169675e-08, + "loss": 0.4368, + "step": 21638 + }, + { + "epoch": 1.843971026842778, + "grad_norm": 13.3125, + "learning_rate": 3.7674309829860836e-08, + "loss": 0.8988, + "step": 21639 + }, + { + "epoch": 1.844056242011078, + "grad_norm": 25.125, + "learning_rate": 3.763337713349607e-08, + "loss": 1.1449, + "step": 21640 + }, + { + "epoch": 1.844141457179378, + "grad_norm": 18.875, + "learning_rate": 3.759246634581534e-08, + "loss": 0.8606, + "step": 21641 + }, + { + "epoch": 1.844226672347678, + "grad_norm": 15.0625, + "learning_rate": 3.7551577467558056e-08, + "loss": 0.8007, + "step": 21642 + }, + { + "epoch": 1.8443118875159779, + "grad_norm": 17.0, + "learning_rate": 3.751071049946278e-08, + "loss": 0.7167, + "step": 21643 + }, + { + "epoch": 1.8443971026842778, + "grad_norm": 12.25, + "learning_rate": 3.7469865442268384e-08, + "loss": 0.3641, + "step": 21644 + }, + { + "epoch": 1.8444823178525778, + "grad_norm": 19.0, + "learning_rate": 3.7429042296712595e-08, + "loss": 0.7557, + "step": 21645 + }, + { + "epoch": 1.8445675330208777, + "grad_norm": 17.875, + "learning_rate": 3.738824106353373e-08, + "loss": 0.8048, + "step": 21646 + }, + { + "epoch": 1.8446527481891777, + "grad_norm": 17.625, + "learning_rate": 3.73474617434684e-08, + "loss": 0.6834, + "step": 21647 + }, + { + "epoch": 1.8447379633574776, + "grad_norm": 14.25, + "learning_rate": 3.7306704337254085e-08, + "loss": 0.5979, + "step": 21648 + }, + { + "epoch": 1.8448231785257776, + "grad_norm": 15.1875, + "learning_rate": 3.726596884562686e-08, + "loss": 0.535, + "step": 21649 + }, + { + "epoch": 1.8449083936940776, + "grad_norm": 14.8125, + "learning_rate": 3.722525526932308e-08, + "loss": 0.7474, + "step": 21650 + }, + { + "epoch": 1.8449936088623775, + "grad_norm": 14.1875, + "learning_rate": 3.718456360907868e-08, + "loss": 0.5332, + "step": 21651 + }, + { + "epoch": 1.8450788240306775, + "grad_norm": 12.875, + "learning_rate": 3.7143893865628636e-08, + "loss": 0.3849, + "step": 21652 + }, + { + "epoch": 1.8451640391989774, + "grad_norm": 13.375, + "learning_rate": 3.7103246039708055e-08, + "loss": 0.4779, + "step": 21653 + }, + { + "epoch": 1.8452492543672774, + "grad_norm": 13.6875, + "learning_rate": 3.706262013205161e-08, + "loss": 0.6061, + "step": 21654 + }, + { + "epoch": 1.8453344695355773, + "grad_norm": 21.375, + "learning_rate": 3.702201614339318e-08, + "loss": 0.6873, + "step": 21655 + }, + { + "epoch": 1.8454196847038773, + "grad_norm": 13.0625, + "learning_rate": 3.698143407446675e-08, + "loss": 0.4483, + "step": 21656 + }, + { + "epoch": 1.8455048998721773, + "grad_norm": 16.75, + "learning_rate": 3.6940873926005484e-08, + "loss": 0.568, + "step": 21657 + }, + { + "epoch": 1.8455901150404772, + "grad_norm": 19.875, + "learning_rate": 3.690033569874254e-08, + "loss": 0.6002, + "step": 21658 + }, + { + "epoch": 1.8456753302087772, + "grad_norm": 14.4375, + "learning_rate": 3.685981939341041e-08, + "loss": 0.3991, + "step": 21659 + }, + { + "epoch": 1.8457605453770771, + "grad_norm": 14.6875, + "learning_rate": 3.681932501074112e-08, + "loss": 0.8265, + "step": 21660 + }, + { + "epoch": 1.845845760545377, + "grad_norm": 11.4375, + "learning_rate": 3.6778852551466737e-08, + "loss": 0.4021, + "step": 21661 + }, + { + "epoch": 1.845930975713677, + "grad_norm": 20.625, + "learning_rate": 3.6738402016318345e-08, + "loss": 0.805, + "step": 21662 + }, + { + "epoch": 1.846016190881977, + "grad_norm": 10.875, + "learning_rate": 3.669797340602715e-08, + "loss": 0.3081, + "step": 21663 + }, + { + "epoch": 1.846101406050277, + "grad_norm": 16.875, + "learning_rate": 3.66575667213237e-08, + "loss": 0.7856, + "step": 21664 + }, + { + "epoch": 1.846186621218577, + "grad_norm": 18.875, + "learning_rate": 3.6617181962937956e-08, + "loss": 0.5548, + "step": 21665 + }, + { + "epoch": 1.8462718363868769, + "grad_norm": 36.25, + "learning_rate": 3.6576819131600024e-08, + "loss": 1.0543, + "step": 21666 + }, + { + "epoch": 1.8463570515551768, + "grad_norm": 12.125, + "learning_rate": 3.653647822803919e-08, + "loss": 0.4291, + "step": 21667 + }, + { + "epoch": 1.8464422667234768, + "grad_norm": 14.9375, + "learning_rate": 3.64961592529843e-08, + "loss": 0.6597, + "step": 21668 + }, + { + "epoch": 1.8465274818917767, + "grad_norm": 13.375, + "learning_rate": 3.64558622071641e-08, + "loss": 0.3739, + "step": 21669 + }, + { + "epoch": 1.8466126970600767, + "grad_norm": 14.1875, + "learning_rate": 3.641558709130674e-08, + "loss": 0.6106, + "step": 21670 + }, + { + "epoch": 1.8466979122283766, + "grad_norm": 16.625, + "learning_rate": 3.637533390614026e-08, + "loss": 0.4475, + "step": 21671 + }, + { + "epoch": 1.8467831273966766, + "grad_norm": 13.875, + "learning_rate": 3.6335102652391844e-08, + "loss": 0.8005, + "step": 21672 + }, + { + "epoch": 1.8468683425649766, + "grad_norm": 16.375, + "learning_rate": 3.629489333078842e-08, + "loss": 0.6222, + "step": 21673 + }, + { + "epoch": 1.8469535577332765, + "grad_norm": 23.25, + "learning_rate": 3.625470594205677e-08, + "loss": 1.2736, + "step": 21674 + }, + { + "epoch": 1.8470387729015765, + "grad_norm": 14.9375, + "learning_rate": 3.6214540486923106e-08, + "loss": 0.51, + "step": 21675 + }, + { + "epoch": 1.8471239880698764, + "grad_norm": 12.0625, + "learning_rate": 3.617439696611338e-08, + "loss": 0.3949, + "step": 21676 + }, + { + "epoch": 1.8472092032381764, + "grad_norm": 17.75, + "learning_rate": 3.6134275380352855e-08, + "loss": 0.7713, + "step": 21677 + }, + { + "epoch": 1.8472944184064763, + "grad_norm": 17.875, + "learning_rate": 3.60941757303665e-08, + "loss": 0.5781, + "step": 21678 + }, + { + "epoch": 1.8473796335747763, + "grad_norm": 14.0625, + "learning_rate": 3.605409801687901e-08, + "loss": 0.6103, + "step": 21679 + }, + { + "epoch": 1.8474648487430763, + "grad_norm": 13.5625, + "learning_rate": 3.601404224061494e-08, + "loss": 0.7507, + "step": 21680 + }, + { + "epoch": 1.8475500639113762, + "grad_norm": 15.3125, + "learning_rate": 3.597400840229789e-08, + "loss": 0.6495, + "step": 21681 + }, + { + "epoch": 1.8476352790796762, + "grad_norm": 15.5, + "learning_rate": 3.593399650265103e-08, + "loss": 0.3275, + "step": 21682 + }, + { + "epoch": 1.8477204942479761, + "grad_norm": 19.5, + "learning_rate": 3.5894006542397805e-08, + "loss": 0.9388, + "step": 21683 + }, + { + "epoch": 1.847805709416276, + "grad_norm": 14.625, + "learning_rate": 3.5854038522260964e-08, + "loss": 0.5404, + "step": 21684 + }, + { + "epoch": 1.847890924584576, + "grad_norm": 12.6875, + "learning_rate": 3.58140924429623e-08, + "loss": 0.4406, + "step": 21685 + }, + { + "epoch": 1.847976139752876, + "grad_norm": 19.625, + "learning_rate": 3.577416830522401e-08, + "loss": 0.4623, + "step": 21686 + }, + { + "epoch": 1.848061354921176, + "grad_norm": 13.125, + "learning_rate": 3.573426610976746e-08, + "loss": 0.3937, + "step": 21687 + }, + { + "epoch": 1.848146570089476, + "grad_norm": 11.8125, + "learning_rate": 3.5694385857313885e-08, + "loss": 0.3808, + "step": 21688 + }, + { + "epoch": 1.8482317852577759, + "grad_norm": 17.625, + "learning_rate": 3.565452754858381e-08, + "loss": 0.7616, + "step": 21689 + }, + { + "epoch": 1.8483170004260758, + "grad_norm": 11.625, + "learning_rate": 3.5614691184297504e-08, + "loss": 0.4579, + "step": 21690 + }, + { + "epoch": 1.8484022155943758, + "grad_norm": 12.0625, + "learning_rate": 3.5574876765174796e-08, + "loss": 0.3455, + "step": 21691 + }, + { + "epoch": 1.8484874307626757, + "grad_norm": 24.875, + "learning_rate": 3.553508429193539e-08, + "loss": 0.9388, + "step": 21692 + }, + { + "epoch": 1.8485726459309757, + "grad_norm": 13.875, + "learning_rate": 3.54953137652983e-08, + "loss": 0.6455, + "step": 21693 + }, + { + "epoch": 1.8486578610992757, + "grad_norm": 16.625, + "learning_rate": 3.545556518598198e-08, + "loss": 0.3097, + "step": 21694 + }, + { + "epoch": 1.8487430762675756, + "grad_norm": 13.0625, + "learning_rate": 3.5415838554705015e-08, + "loss": 0.547, + "step": 21695 + }, + { + "epoch": 1.8488282914358756, + "grad_norm": 16.875, + "learning_rate": 3.5376133872185304e-08, + "loss": 0.6479, + "step": 21696 + }, + { + "epoch": 1.8489135066041755, + "grad_norm": 22.125, + "learning_rate": 3.533645113914019e-08, + "loss": 0.8056, + "step": 21697 + }, + { + "epoch": 1.8489987217724755, + "grad_norm": 15.5625, + "learning_rate": 3.529679035628675e-08, + "loss": 0.632, + "step": 21698 + }, + { + "epoch": 1.8490839369407754, + "grad_norm": 16.125, + "learning_rate": 3.525715152434175e-08, + "loss": 0.5089, + "step": 21699 + }, + { + "epoch": 1.8491691521090754, + "grad_norm": 15.0625, + "learning_rate": 3.5217534644021585e-08, + "loss": 0.6917, + "step": 21700 + }, + { + "epoch": 1.8492543672773754, + "grad_norm": 13.8125, + "learning_rate": 3.517793971604233e-08, + "loss": 0.4252, + "step": 21701 + }, + { + "epoch": 1.8493395824456753, + "grad_norm": 22.0, + "learning_rate": 3.513836674111912e-08, + "loss": 0.8536, + "step": 21702 + }, + { + "epoch": 1.8494247976139753, + "grad_norm": 14.4375, + "learning_rate": 3.5098815719967205e-08, + "loss": 0.43, + "step": 21703 + }, + { + "epoch": 1.8495100127822752, + "grad_norm": 13.0625, + "learning_rate": 3.505928665330158e-08, + "loss": 0.4086, + "step": 21704 + }, + { + "epoch": 1.8495952279505752, + "grad_norm": 14.3125, + "learning_rate": 3.501977954183611e-08, + "loss": 0.4753, + "step": 21705 + }, + { + "epoch": 1.8496804431188751, + "grad_norm": 22.75, + "learning_rate": 3.498029438628523e-08, + "loss": 0.7643, + "step": 21706 + }, + { + "epoch": 1.849765658287175, + "grad_norm": 17.625, + "learning_rate": 3.4940831187362104e-08, + "loss": 0.5079, + "step": 21707 + }, + { + "epoch": 1.849850873455475, + "grad_norm": 18.5, + "learning_rate": 3.4901389945779933e-08, + "loss": 0.6738, + "step": 21708 + }, + { + "epoch": 1.849936088623775, + "grad_norm": 22.875, + "learning_rate": 3.4861970662251474e-08, + "loss": 0.7038, + "step": 21709 + }, + { + "epoch": 1.850021303792075, + "grad_norm": 12.125, + "learning_rate": 3.482257333748934e-08, + "loss": 0.4246, + "step": 21710 + }, + { + "epoch": 1.850106518960375, + "grad_norm": 12.25, + "learning_rate": 3.4783197972205194e-08, + "loss": 0.5588, + "step": 21711 + }, + { + "epoch": 1.8501917341286749, + "grad_norm": 17.75, + "learning_rate": 3.4743844567110693e-08, + "loss": 0.6004, + "step": 21712 + }, + { + "epoch": 1.8502769492969748, + "grad_norm": 11.375, + "learning_rate": 3.470451312291678e-08, + "loss": 0.2119, + "step": 21713 + }, + { + "epoch": 1.8503621644652748, + "grad_norm": 15.0625, + "learning_rate": 3.466520364033457e-08, + "loss": 0.5964, + "step": 21714 + }, + { + "epoch": 1.8504473796335748, + "grad_norm": 13.875, + "learning_rate": 3.462591612007404e-08, + "loss": 0.4294, + "step": 21715 + }, + { + "epoch": 1.8505325948018747, + "grad_norm": 14.125, + "learning_rate": 3.4586650562845464e-08, + "loss": 0.4852, + "step": 21716 + }, + { + "epoch": 1.8506178099701747, + "grad_norm": 14.375, + "learning_rate": 3.4547406969358264e-08, + "loss": 0.5655, + "step": 21717 + }, + { + "epoch": 1.8507030251384746, + "grad_norm": 17.5, + "learning_rate": 3.450818534032188e-08, + "loss": 0.575, + "step": 21718 + }, + { + "epoch": 1.8507882403067746, + "grad_norm": 16.875, + "learning_rate": 3.4468985676444635e-08, + "loss": 0.2654, + "step": 21719 + }, + { + "epoch": 1.8508734554750745, + "grad_norm": 23.125, + "learning_rate": 3.442980797843512e-08, + "loss": 0.528, + "step": 21720 + }, + { + "epoch": 1.8509586706433745, + "grad_norm": 21.25, + "learning_rate": 3.439065224700139e-08, + "loss": 1.1141, + "step": 21721 + }, + { + "epoch": 1.8510438858116744, + "grad_norm": 14.4375, + "learning_rate": 3.435151848285093e-08, + "loss": 0.5612, + "step": 21722 + }, + { + "epoch": 1.8511291009799744, + "grad_norm": 14.8125, + "learning_rate": 3.4312406686691086e-08, + "loss": 0.4325, + "step": 21723 + }, + { + "epoch": 1.8512143161482744, + "grad_norm": 13.125, + "learning_rate": 3.427331685922838e-08, + "loss": 0.5176, + "step": 21724 + }, + { + "epoch": 1.8512995313165743, + "grad_norm": 22.5, + "learning_rate": 3.423424900116934e-08, + "loss": 0.5155, + "step": 21725 + }, + { + "epoch": 1.8513847464848743, + "grad_norm": 19.125, + "learning_rate": 3.419520311322005e-08, + "loss": 0.7926, + "step": 21726 + }, + { + "epoch": 1.8514699616531742, + "grad_norm": 18.5, + "learning_rate": 3.4156179196086075e-08, + "loss": 0.5331, + "step": 21727 + }, + { + "epoch": 1.8515551768214742, + "grad_norm": 15.0625, + "learning_rate": 3.4117177250472394e-08, + "loss": 0.6379, + "step": 21728 + }, + { + "epoch": 1.8516403919897741, + "grad_norm": 14.8125, + "learning_rate": 3.407819727708414e-08, + "loss": 0.8452, + "step": 21729 + }, + { + "epoch": 1.851725607158074, + "grad_norm": 13.9375, + "learning_rate": 3.4039239276625474e-08, + "loss": 0.6309, + "step": 21730 + }, + { + "epoch": 1.851810822326374, + "grad_norm": 23.125, + "learning_rate": 3.400030324980055e-08, + "loss": 0.945, + "step": 21731 + }, + { + "epoch": 1.851896037494674, + "grad_norm": 19.625, + "learning_rate": 3.396138919731298e-08, + "loss": 0.6985, + "step": 21732 + }, + { + "epoch": 1.851981252662974, + "grad_norm": 18.0, + "learning_rate": 3.392249711986581e-08, + "loss": 0.6468, + "step": 21733 + }, + { + "epoch": 1.852066467831274, + "grad_norm": 16.125, + "learning_rate": 3.3883627018161946e-08, + "loss": 0.6337, + "step": 21734 + }, + { + "epoch": 1.852151682999574, + "grad_norm": 16.125, + "learning_rate": 3.384477889290402e-08, + "loss": 0.5205, + "step": 21735 + }, + { + "epoch": 1.8522368981678738, + "grad_norm": 13.125, + "learning_rate": 3.3805952744793704e-08, + "loss": 0.4538, + "step": 21736 + }, + { + "epoch": 1.8523221133361738, + "grad_norm": 17.0, + "learning_rate": 3.376714857453292e-08, + "loss": 0.5704, + "step": 21737 + }, + { + "epoch": 1.8524073285044738, + "grad_norm": 16.375, + "learning_rate": 3.372836638282265e-08, + "loss": 0.6977, + "step": 21738 + }, + { + "epoch": 1.8524925436727737, + "grad_norm": 17.75, + "learning_rate": 3.368960617036399e-08, + "loss": 0.6494, + "step": 21739 + }, + { + "epoch": 1.8525777588410737, + "grad_norm": 14.8125, + "learning_rate": 3.3650867937857215e-08, + "loss": 0.6914, + "step": 21740 + }, + { + "epoch": 1.8526629740093736, + "grad_norm": 14.625, + "learning_rate": 3.361215168600232e-08, + "loss": 0.4342, + "step": 21741 + }, + { + "epoch": 1.8527481891776736, + "grad_norm": 14.375, + "learning_rate": 3.3573457415499025e-08, + "loss": 0.6917, + "step": 21742 + }, + { + "epoch": 1.8528334043459735, + "grad_norm": 15.6875, + "learning_rate": 3.35347851270465e-08, + "loss": 0.5003, + "step": 21743 + }, + { + "epoch": 1.8529186195142735, + "grad_norm": 13.8125, + "learning_rate": 3.3496134821343894e-08, + "loss": 0.5738, + "step": 21744 + }, + { + "epoch": 1.8530038346825735, + "grad_norm": 9.0625, + "learning_rate": 3.345750649908927e-08, + "loss": 0.2187, + "step": 21745 + }, + { + "epoch": 1.8530890498508734, + "grad_norm": 15.375, + "learning_rate": 3.341890016098082e-08, + "loss": 0.785, + "step": 21746 + }, + { + "epoch": 1.8531742650191734, + "grad_norm": 22.875, + "learning_rate": 3.338031580771617e-08, + "loss": 0.9838, + "step": 21747 + }, + { + "epoch": 1.8532594801874733, + "grad_norm": 11.0, + "learning_rate": 3.3341753439992825e-08, + "loss": 0.3502, + "step": 21748 + }, + { + "epoch": 1.8533446953557733, + "grad_norm": 15.9375, + "learning_rate": 3.3303213058507314e-08, + "loss": 0.9607, + "step": 21749 + }, + { + "epoch": 1.8534299105240732, + "grad_norm": 14.625, + "learning_rate": 3.3264694663956155e-08, + "loss": 0.5956, + "step": 21750 + }, + { + "epoch": 1.8535151256923732, + "grad_norm": 14.625, + "learning_rate": 3.322619825703546e-08, + "loss": 0.6904, + "step": 21751 + }, + { + "epoch": 1.8536003408606732, + "grad_norm": 10.375, + "learning_rate": 3.318772383844107e-08, + "loss": 0.2875, + "step": 21752 + }, + { + "epoch": 1.8536855560289731, + "grad_norm": 15.0, + "learning_rate": 3.3149271408868116e-08, + "loss": 0.7884, + "step": 21753 + }, + { + "epoch": 1.853770771197273, + "grad_norm": 15.0625, + "learning_rate": 3.311084096901132e-08, + "loss": 0.6617, + "step": 21754 + }, + { + "epoch": 1.853855986365573, + "grad_norm": 19.625, + "learning_rate": 3.307243251956541e-08, + "loss": 1.328, + "step": 21755 + }, + { + "epoch": 1.853941201533873, + "grad_norm": 17.25, + "learning_rate": 3.303404606122426e-08, + "loss": 0.6941, + "step": 21756 + }, + { + "epoch": 1.854026416702173, + "grad_norm": 20.0, + "learning_rate": 3.2995681594681775e-08, + "loss": 0.6445, + "step": 21757 + }, + { + "epoch": 1.854111631870473, + "grad_norm": 23.25, + "learning_rate": 3.295733912063101e-08, + "loss": 0.8694, + "step": 21758 + }, + { + "epoch": 1.8541968470387729, + "grad_norm": 12.3125, + "learning_rate": 3.2919018639765015e-08, + "loss": 0.4179, + "step": 21759 + }, + { + "epoch": 1.8542820622070728, + "grad_norm": 15.625, + "learning_rate": 3.2880720152776294e-08, + "loss": 0.4169, + "step": 21760 + }, + { + "epoch": 1.8543672773753728, + "grad_norm": 19.875, + "learning_rate": 3.28424436603568e-08, + "loss": 0.7681, + "step": 21761 + }, + { + "epoch": 1.8544524925436727, + "grad_norm": 18.0, + "learning_rate": 3.280418916319833e-08, + "loss": 0.5571, + "step": 21762 + }, + { + "epoch": 1.8545377077119727, + "grad_norm": 16.5, + "learning_rate": 3.2765956661992005e-08, + "loss": 0.7276, + "step": 21763 + }, + { + "epoch": 1.8546229228802726, + "grad_norm": 20.875, + "learning_rate": 3.2727746157428934e-08, + "loss": 0.7449, + "step": 21764 + }, + { + "epoch": 1.8547081380485726, + "grad_norm": 24.25, + "learning_rate": 3.268955765019968e-08, + "loss": 1.0629, + "step": 21765 + }, + { + "epoch": 1.8547933532168726, + "grad_norm": 26.875, + "learning_rate": 3.265139114099397e-08, + "loss": 0.667, + "step": 21766 + }, + { + "epoch": 1.8548785683851725, + "grad_norm": 16.375, + "learning_rate": 3.2613246630501946e-08, + "loss": 0.451, + "step": 21767 + }, + { + "epoch": 1.8549637835534725, + "grad_norm": 17.25, + "learning_rate": 3.257512411941263e-08, + "loss": 0.8483, + "step": 21768 + }, + { + "epoch": 1.8550489987217724, + "grad_norm": 15.4375, + "learning_rate": 3.253702360841521e-08, + "loss": 0.6194, + "step": 21769 + }, + { + "epoch": 1.8551342138900724, + "grad_norm": 16.875, + "learning_rate": 3.249894509819801e-08, + "loss": 0.8664, + "step": 21770 + }, + { + "epoch": 1.8552194290583723, + "grad_norm": 16.625, + "learning_rate": 3.246088858944896e-08, + "loss": 0.6467, + "step": 21771 + }, + { + "epoch": 1.8553046442266723, + "grad_norm": 16.75, + "learning_rate": 3.242285408285611e-08, + "loss": 0.4776, + "step": 21772 + }, + { + "epoch": 1.8553898593949723, + "grad_norm": 17.375, + "learning_rate": 3.238484157910657e-08, + "loss": 1.0107, + "step": 21773 + }, + { + "epoch": 1.8554750745632722, + "grad_norm": 15.9375, + "learning_rate": 3.234685107888755e-08, + "loss": 0.6366, + "step": 21774 + }, + { + "epoch": 1.8555602897315722, + "grad_norm": 20.0, + "learning_rate": 3.230888258288517e-08, + "loss": 0.9838, + "step": 21775 + }, + { + "epoch": 1.8556455048998721, + "grad_norm": 12.75, + "learning_rate": 3.227093609178583e-08, + "loss": 0.2061, + "step": 21776 + }, + { + "epoch": 1.855730720068172, + "grad_norm": 13.125, + "learning_rate": 3.2233011606275236e-08, + "loss": 0.5362, + "step": 21777 + }, + { + "epoch": 1.855815935236472, + "grad_norm": 13.5625, + "learning_rate": 3.21951091270388e-08, + "loss": 0.5551, + "step": 21778 + }, + { + "epoch": 1.855901150404772, + "grad_norm": 18.5, + "learning_rate": 3.215722865476112e-08, + "loss": 0.5603, + "step": 21779 + }, + { + "epoch": 1.855986365573072, + "grad_norm": 13.625, + "learning_rate": 3.2119370190127064e-08, + "loss": 0.625, + "step": 21780 + }, + { + "epoch": 1.856071580741372, + "grad_norm": 17.25, + "learning_rate": 3.208153373382053e-08, + "loss": 0.5408, + "step": 21781 + }, + { + "epoch": 1.8561567959096719, + "grad_norm": 17.125, + "learning_rate": 3.2043719286525556e-08, + "loss": 0.6354, + "step": 21782 + }, + { + "epoch": 1.8562420110779718, + "grad_norm": 13.375, + "learning_rate": 3.20059268489252e-08, + "loss": 0.503, + "step": 21783 + }, + { + "epoch": 1.8563272262462718, + "grad_norm": 13.625, + "learning_rate": 3.1968156421702535e-08, + "loss": 0.485, + "step": 21784 + }, + { + "epoch": 1.8564124414145717, + "grad_norm": 14.125, + "learning_rate": 3.1930408005540204e-08, + "loss": 0.4398, + "step": 21785 + }, + { + "epoch": 1.8564976565828717, + "grad_norm": 17.5, + "learning_rate": 3.18926816011203e-08, + "loss": 0.5902, + "step": 21786 + }, + { + "epoch": 1.8565828717511716, + "grad_norm": 22.375, + "learning_rate": 3.1854977209124506e-08, + "loss": 0.6216, + "step": 21787 + }, + { + "epoch": 1.8566680869194716, + "grad_norm": 13.75, + "learning_rate": 3.1817294830234074e-08, + "loss": 0.3478, + "step": 21788 + }, + { + "epoch": 1.8567533020877716, + "grad_norm": 12.375, + "learning_rate": 3.177963446513027e-08, + "loss": 0.4178, + "step": 21789 + }, + { + "epoch": 1.8568385172560715, + "grad_norm": 12.0625, + "learning_rate": 3.174199611449338e-08, + "loss": 0.4947, + "step": 21790 + }, + { + "epoch": 1.8569237324243715, + "grad_norm": 16.125, + "learning_rate": 3.1704379779003703e-08, + "loss": 0.6826, + "step": 21791 + }, + { + "epoch": 1.8570089475926714, + "grad_norm": 14.3125, + "learning_rate": 3.166678545934096e-08, + "loss": 0.6584, + "step": 21792 + }, + { + "epoch": 1.8570941627609714, + "grad_norm": 16.625, + "learning_rate": 3.1629213156184475e-08, + "loss": 0.5517, + "step": 21793 + }, + { + "epoch": 1.8571793779292713, + "grad_norm": 15.875, + "learning_rate": 3.159166287021343e-08, + "loss": 0.6834, + "step": 21794 + }, + { + "epoch": 1.8572645930975713, + "grad_norm": 16.625, + "learning_rate": 3.15541346021063e-08, + "loss": 0.6807, + "step": 21795 + }, + { + "epoch": 1.8573498082658713, + "grad_norm": 18.625, + "learning_rate": 3.151662835254102e-08, + "loss": 0.8298, + "step": 21796 + }, + { + "epoch": 1.8574350234341712, + "grad_norm": 37.5, + "learning_rate": 3.147914412219552e-08, + "loss": 0.6353, + "step": 21797 + }, + { + "epoch": 1.8575202386024712, + "grad_norm": 15.875, + "learning_rate": 3.1441681911747315e-08, + "loss": 0.6473, + "step": 21798 + }, + { + "epoch": 1.8576054537707711, + "grad_norm": 23.375, + "learning_rate": 3.140424172187323e-08, + "loss": 0.6477, + "step": 21799 + }, + { + "epoch": 1.857690668939071, + "grad_norm": 11.6875, + "learning_rate": 3.136682355324994e-08, + "loss": 0.5029, + "step": 21800 + }, + { + "epoch": 1.857775884107371, + "grad_norm": 14.4375, + "learning_rate": 3.1329427406553424e-08, + "loss": 0.7623, + "step": 21801 + }, + { + "epoch": 1.857861099275671, + "grad_norm": 12.0, + "learning_rate": 3.129205328245982e-08, + "loss": 0.3784, + "step": 21802 + }, + { + "epoch": 1.857946314443971, + "grad_norm": 15.8125, + "learning_rate": 3.125470118164414e-08, + "loss": 0.8258, + "step": 21803 + }, + { + "epoch": 1.858031529612271, + "grad_norm": 16.75, + "learning_rate": 3.121737110478182e-08, + "loss": 0.8034, + "step": 21804 + }, + { + "epoch": 1.8581167447805709, + "grad_norm": 15.875, + "learning_rate": 3.1180063052546895e-08, + "loss": 0.3096, + "step": 21805 + }, + { + "epoch": 1.8582019599488708, + "grad_norm": 22.25, + "learning_rate": 3.1142777025613967e-08, + "loss": 0.7881, + "step": 21806 + }, + { + "epoch": 1.8582871751171708, + "grad_norm": 16.5, + "learning_rate": 3.1105513024656806e-08, + "loss": 0.7035, + "step": 21807 + }, + { + "epoch": 1.8583723902854707, + "grad_norm": 15.3125, + "learning_rate": 3.1068271050348486e-08, + "loss": 0.8079, + "step": 21808 + }, + { + "epoch": 1.8584576054537707, + "grad_norm": 14.375, + "learning_rate": 3.1031051103362353e-08, + "loss": 0.7055, + "step": 21809 + }, + { + "epoch": 1.8585428206220707, + "grad_norm": 23.75, + "learning_rate": 3.099385318437106e-08, + "loss": 1.0617, + "step": 21810 + }, + { + "epoch": 1.8586280357903706, + "grad_norm": 22.375, + "learning_rate": 3.0956677294046446e-08, + "loss": 0.6819, + "step": 21811 + }, + { + "epoch": 1.8587132509586706, + "grad_norm": 12.5625, + "learning_rate": 3.0919523433060594e-08, + "loss": 0.525, + "step": 21812 + }, + { + "epoch": 1.8587984661269705, + "grad_norm": 19.125, + "learning_rate": 3.0882391602084784e-08, + "loss": 0.6218, + "step": 21813 + }, + { + "epoch": 1.8588836812952705, + "grad_norm": 11.3125, + "learning_rate": 3.084528180179014e-08, + "loss": 0.3098, + "step": 21814 + }, + { + "epoch": 1.8589688964635704, + "grad_norm": 20.5, + "learning_rate": 3.08081940328471e-08, + "loss": 0.5908, + "step": 21815 + }, + { + "epoch": 1.8590541116318704, + "grad_norm": 15.0, + "learning_rate": 3.07711282959261e-08, + "loss": 0.5866, + "step": 21816 + }, + { + "epoch": 1.8591393268001704, + "grad_norm": 17.0, + "learning_rate": 3.073408459169675e-08, + "loss": 0.5662, + "step": 21817 + }, + { + "epoch": 1.8592245419684703, + "grad_norm": 21.5, + "learning_rate": 3.0697062920828635e-08, + "loss": 0.6827, + "step": 21818 + }, + { + "epoch": 1.8593097571367703, + "grad_norm": 11.25, + "learning_rate": 3.066006328399068e-08, + "loss": 0.338, + "step": 21819 + }, + { + "epoch": 1.8593949723050702, + "grad_norm": 18.375, + "learning_rate": 3.0623085681851517e-08, + "loss": 0.7246, + "step": 21820 + }, + { + "epoch": 1.8594801874733702, + "grad_norm": 23.125, + "learning_rate": 3.0586130115079354e-08, + "loss": 0.5349, + "step": 21821 + }, + { + "epoch": 1.8595654026416701, + "grad_norm": 23.25, + "learning_rate": 3.054919658434199e-08, + "loss": 0.6655, + "step": 21822 + }, + { + "epoch": 1.85965061780997, + "grad_norm": 22.125, + "learning_rate": 3.051228509030696e-08, + "loss": 0.7001, + "step": 21823 + }, + { + "epoch": 1.85973583297827, + "grad_norm": 19.5, + "learning_rate": 3.047539563364135e-08, + "loss": 1.0705, + "step": 21824 + }, + { + "epoch": 1.85982104814657, + "grad_norm": 14.25, + "learning_rate": 3.0438528215011446e-08, + "loss": 0.4888, + "step": 21825 + }, + { + "epoch": 1.85990626331487, + "grad_norm": 14.0625, + "learning_rate": 3.0401682835083934e-08, + "loss": 0.6391, + "step": 21826 + }, + { + "epoch": 1.85999147848317, + "grad_norm": 17.375, + "learning_rate": 3.036485949452425e-08, + "loss": 0.7349, + "step": 21827 + }, + { + "epoch": 1.8600766936514699, + "grad_norm": 14.625, + "learning_rate": 3.032805819399798e-08, + "loss": 0.5132, + "step": 21828 + }, + { + "epoch": 1.8601619088197698, + "grad_norm": 15.6875, + "learning_rate": 3.0291278934170276e-08, + "loss": 0.4225, + "step": 21829 + }, + { + "epoch": 1.8602471239880698, + "grad_norm": 17.625, + "learning_rate": 3.0254521715705476e-08, + "loss": 0.5203, + "step": 21830 + }, + { + "epoch": 1.8603323391563698, + "grad_norm": 14.875, + "learning_rate": 3.021778653926818e-08, + "loss": 0.5024, + "step": 21831 + }, + { + "epoch": 1.8604175543246697, + "grad_norm": 25.625, + "learning_rate": 3.01810734055219e-08, + "loss": 0.9732, + "step": 21832 + }, + { + "epoch": 1.8605027694929697, + "grad_norm": 20.125, + "learning_rate": 3.014438231513053e-08, + "loss": 0.4581, + "step": 21833 + }, + { + "epoch": 1.8605879846612696, + "grad_norm": 16.75, + "learning_rate": 3.010771326875675e-08, + "loss": 0.5764, + "step": 21834 + }, + { + "epoch": 1.8606731998295696, + "grad_norm": 17.875, + "learning_rate": 3.0071066267063215e-08, + "loss": 0.6923, + "step": 21835 + }, + { + "epoch": 1.8607584149978695, + "grad_norm": 18.75, + "learning_rate": 3.0034441310712174e-08, + "loss": 0.629, + "step": 21836 + }, + { + "epoch": 1.8608436301661695, + "grad_norm": 21.75, + "learning_rate": 2.999783840036574e-08, + "loss": 0.8463, + "step": 21837 + }, + { + "epoch": 1.8609288453344695, + "grad_norm": 37.25, + "learning_rate": 2.996125753668505e-08, + "loss": 0.7086, + "step": 21838 + }, + { + "epoch": 1.8610140605027694, + "grad_norm": 13.6875, + "learning_rate": 2.992469872033124e-08, + "loss": 0.3883, + "step": 21839 + }, + { + "epoch": 1.8610992756710694, + "grad_norm": 12.3125, + "learning_rate": 2.9888161951965034e-08, + "loss": 0.3862, + "step": 21840 + }, + { + "epoch": 1.8611844908393693, + "grad_norm": 16.5, + "learning_rate": 2.985164723224673e-08, + "loss": 0.6201, + "step": 21841 + }, + { + "epoch": 1.8612697060076693, + "grad_norm": 13.4375, + "learning_rate": 2.981515456183609e-08, + "loss": 0.3361, + "step": 21842 + }, + { + "epoch": 1.8613549211759692, + "grad_norm": 21.5, + "learning_rate": 2.9778683941392576e-08, + "loss": 0.6424, + "step": 21843 + }, + { + "epoch": 1.8614401363442692, + "grad_norm": 19.5, + "learning_rate": 2.974223537157525e-08, + "loss": 0.9167, + "step": 21844 + }, + { + "epoch": 1.8615253515125691, + "grad_norm": 13.625, + "learning_rate": 2.9705808853042752e-08, + "loss": 0.5752, + "step": 21845 + }, + { + "epoch": 1.861610566680869, + "grad_norm": 12.5, + "learning_rate": 2.9669404386453438e-08, + "loss": 0.472, + "step": 21846 + }, + { + "epoch": 1.861695781849169, + "grad_norm": 16.625, + "learning_rate": 2.963302197246512e-08, + "loss": 0.7601, + "step": 21847 + }, + { + "epoch": 1.861780997017469, + "grad_norm": 16.875, + "learning_rate": 2.959666161173533e-08, + "loss": 0.5139, + "step": 21848 + }, + { + "epoch": 1.861866212185769, + "grad_norm": 16.25, + "learning_rate": 2.9560323304920898e-08, + "loss": 0.7937, + "step": 21849 + }, + { + "epoch": 1.861951427354069, + "grad_norm": 13.375, + "learning_rate": 2.9524007052678937e-08, + "loss": 0.6047, + "step": 21850 + }, + { + "epoch": 1.862036642522369, + "grad_norm": 14.125, + "learning_rate": 2.9487712855665318e-08, + "loss": 0.3327, + "step": 21851 + }, + { + "epoch": 1.8621218576906688, + "grad_norm": 63.75, + "learning_rate": 2.9451440714536038e-08, + "loss": 0.2523, + "step": 21852 + }, + { + "epoch": 1.8622070728589688, + "grad_norm": 18.375, + "learning_rate": 2.941519062994655e-08, + "loss": 0.7093, + "step": 21853 + }, + { + "epoch": 1.8622922880272688, + "grad_norm": 18.125, + "learning_rate": 2.937896260255216e-08, + "loss": 0.5175, + "step": 21854 + }, + { + "epoch": 1.8623775031955687, + "grad_norm": 18.875, + "learning_rate": 2.9342756633007063e-08, + "loss": 0.6343, + "step": 21855 + }, + { + "epoch": 1.8624627183638687, + "grad_norm": 15.0625, + "learning_rate": 2.930657272196588e-08, + "loss": 0.6965, + "step": 21856 + }, + { + "epoch": 1.8625479335321686, + "grad_norm": 28.0, + "learning_rate": 2.9270410870082532e-08, + "loss": 0.7711, + "step": 21857 + }, + { + "epoch": 1.8626331487004686, + "grad_norm": 12.75, + "learning_rate": 2.9234271078010384e-08, + "loss": 0.3663, + "step": 21858 + }, + { + "epoch": 1.8627183638687685, + "grad_norm": 13.0, + "learning_rate": 2.9198153346402664e-08, + "loss": 0.3477, + "step": 21859 + }, + { + "epoch": 1.8628035790370685, + "grad_norm": 22.5, + "learning_rate": 2.9162057675911627e-08, + "loss": 0.982, + "step": 21860 + }, + { + "epoch": 1.8628887942053685, + "grad_norm": 19.25, + "learning_rate": 2.9125984067189943e-08, + "loss": 0.6279, + "step": 21861 + }, + { + "epoch": 1.8629740093736684, + "grad_norm": 22.25, + "learning_rate": 2.9089932520889458e-08, + "loss": 1.2789, + "step": 21862 + }, + { + "epoch": 1.8630592245419684, + "grad_norm": 22.625, + "learning_rate": 2.905390303766173e-08, + "loss": 0.5285, + "step": 21863 + }, + { + "epoch": 1.8631444397102683, + "grad_norm": 13.6875, + "learning_rate": 2.9017895618157626e-08, + "loss": 0.5708, + "step": 21864 + }, + { + "epoch": 1.8632296548785683, + "grad_norm": 20.625, + "learning_rate": 2.898191026302788e-08, + "loss": 0.8125, + "step": 21865 + }, + { + "epoch": 1.8633148700468682, + "grad_norm": 13.9375, + "learning_rate": 2.89459469729228e-08, + "loss": 0.5551, + "step": 21866 + }, + { + "epoch": 1.8634000852151682, + "grad_norm": 16.375, + "learning_rate": 2.8910005748492566e-08, + "loss": 0.4955, + "step": 21867 + }, + { + "epoch": 1.8634853003834682, + "grad_norm": 15.8125, + "learning_rate": 2.8874086590386234e-08, + "loss": 0.5985, + "step": 21868 + }, + { + "epoch": 1.8635705155517681, + "grad_norm": 19.375, + "learning_rate": 2.8838189499253154e-08, + "loss": 0.8152, + "step": 21869 + }, + { + "epoch": 1.863655730720068, + "grad_norm": 12.5625, + "learning_rate": 2.880231447574197e-08, + "loss": 0.4431, + "step": 21870 + }, + { + "epoch": 1.863740945888368, + "grad_norm": 12.6875, + "learning_rate": 2.8766461520500915e-08, + "loss": 0.4003, + "step": 21871 + }, + { + "epoch": 1.863826161056668, + "grad_norm": 23.125, + "learning_rate": 2.873063063417794e-08, + "loss": 0.6986, + "step": 21872 + }, + { + "epoch": 1.863911376224968, + "grad_norm": 16.25, + "learning_rate": 2.869482181742059e-08, + "loss": 0.6462, + "step": 21873 + }, + { + "epoch": 1.863996591393268, + "grad_norm": 12.0, + "learning_rate": 2.8659035070875975e-08, + "loss": 0.3209, + "step": 21874 + }, + { + "epoch": 1.8640818065615679, + "grad_norm": 17.5, + "learning_rate": 2.8623270395190667e-08, + "loss": 0.429, + "step": 21875 + }, + { + "epoch": 1.8641670217298678, + "grad_norm": 12.75, + "learning_rate": 2.8587527791011233e-08, + "loss": 0.4146, + "step": 21876 + }, + { + "epoch": 1.8642522368981678, + "grad_norm": 24.75, + "learning_rate": 2.8551807258983266e-08, + "loss": 0.6662, + "step": 21877 + }, + { + "epoch": 1.8643374520664677, + "grad_norm": 15.5625, + "learning_rate": 2.8516108799752362e-08, + "loss": 0.5689, + "step": 21878 + }, + { + "epoch": 1.8644226672347677, + "grad_norm": 24.75, + "learning_rate": 2.848043241396356e-08, + "loss": 0.5794, + "step": 21879 + }, + { + "epoch": 1.8645078824030676, + "grad_norm": 21.75, + "learning_rate": 2.84447781022619e-08, + "loss": 0.7863, + "step": 21880 + }, + { + "epoch": 1.8645930975713676, + "grad_norm": 15.5, + "learning_rate": 2.840914586529145e-08, + "loss": 0.6148, + "step": 21881 + }, + { + "epoch": 1.8646783127396676, + "grad_norm": 16.875, + "learning_rate": 2.8373535703695998e-08, + "loss": 0.6959, + "step": 21882 + }, + { + "epoch": 1.8647635279079675, + "grad_norm": 23.25, + "learning_rate": 2.8337947618119337e-08, + "loss": 0.5256, + "step": 21883 + }, + { + "epoch": 1.8648487430762675, + "grad_norm": 36.0, + "learning_rate": 2.8302381609204426e-08, + "loss": 0.9293, + "step": 21884 + }, + { + "epoch": 1.8649339582445674, + "grad_norm": 18.625, + "learning_rate": 2.8266837677593944e-08, + "loss": 0.7706, + "step": 21885 + }, + { + "epoch": 1.8650191734128674, + "grad_norm": 19.875, + "learning_rate": 2.8231315823930154e-08, + "loss": 0.5775, + "step": 21886 + }, + { + "epoch": 1.8651043885811673, + "grad_norm": 20.5, + "learning_rate": 2.8195816048855184e-08, + "loss": 0.6981, + "step": 21887 + }, + { + "epoch": 1.8651896037494673, + "grad_norm": 10.8125, + "learning_rate": 2.8160338353010464e-08, + "loss": 0.6661, + "step": 21888 + }, + { + "epoch": 1.8652748189177673, + "grad_norm": 12.4375, + "learning_rate": 2.8124882737037008e-08, + "loss": 0.5212, + "step": 21889 + }, + { + "epoch": 1.8653600340860672, + "grad_norm": 12.125, + "learning_rate": 2.8089449201575553e-08, + "loss": 0.3896, + "step": 21890 + }, + { + "epoch": 1.8654452492543672, + "grad_norm": 15.4375, + "learning_rate": 2.80540377472667e-08, + "loss": 0.6503, + "step": 21891 + }, + { + "epoch": 1.8655304644226671, + "grad_norm": 22.375, + "learning_rate": 2.8018648374749936e-08, + "loss": 0.5124, + "step": 21892 + }, + { + "epoch": 1.865615679590967, + "grad_norm": 12.625, + "learning_rate": 2.7983281084665165e-08, + "loss": 0.7948, + "step": 21893 + }, + { + "epoch": 1.865700894759267, + "grad_norm": 10.5625, + "learning_rate": 2.7947935877651182e-08, + "loss": 0.2206, + "step": 21894 + }, + { + "epoch": 1.865786109927567, + "grad_norm": 12.6875, + "learning_rate": 2.7912612754346925e-08, + "loss": 0.3345, + "step": 21895 + }, + { + "epoch": 1.865871325095867, + "grad_norm": 13.875, + "learning_rate": 2.7877311715390626e-08, + "loss": 0.4532, + "step": 21896 + }, + { + "epoch": 1.865956540264167, + "grad_norm": 24.625, + "learning_rate": 2.784203276142039e-08, + "loss": 0.5684, + "step": 21897 + }, + { + "epoch": 1.8660417554324669, + "grad_norm": 19.875, + "learning_rate": 2.7806775893073347e-08, + "loss": 0.857, + "step": 21898 + }, + { + "epoch": 1.8661269706007668, + "grad_norm": 16.125, + "learning_rate": 2.777154111098718e-08, + "loss": 0.7175, + "step": 21899 + }, + { + "epoch": 1.8662121857690668, + "grad_norm": 18.0, + "learning_rate": 2.773632841579804e-08, + "loss": 0.5918, + "step": 21900 + }, + { + "epoch": 1.8662974009373667, + "grad_norm": 14.6875, + "learning_rate": 2.7701137808142652e-08, + "loss": 0.8391, + "step": 21901 + }, + { + "epoch": 1.8663826161056667, + "grad_norm": 15.0, + "learning_rate": 2.7665969288656753e-08, + "loss": 0.6167, + "step": 21902 + }, + { + "epoch": 1.8664678312739666, + "grad_norm": 12.8125, + "learning_rate": 2.7630822857975942e-08, + "loss": 0.3063, + "step": 21903 + }, + { + "epoch": 1.8665530464422666, + "grad_norm": 9.6875, + "learning_rate": 2.7595698516735408e-08, + "loss": 0.3097, + "step": 21904 + }, + { + "epoch": 1.8666382616105666, + "grad_norm": 19.25, + "learning_rate": 2.7560596265569924e-08, + "loss": 0.2733, + "step": 21905 + }, + { + "epoch": 1.8667234767788665, + "grad_norm": 24.125, + "learning_rate": 2.752551610511356e-08, + "loss": 0.6206, + "step": 21906 + }, + { + "epoch": 1.8668086919471665, + "grad_norm": 13.4375, + "learning_rate": 2.7490458036000535e-08, + "loss": 0.5084, + "step": 21907 + }, + { + "epoch": 1.8668939071154664, + "grad_norm": 25.125, + "learning_rate": 2.745542205886409e-08, + "loss": 0.7136, + "step": 21908 + }, + { + "epoch": 1.8669791222837664, + "grad_norm": 13.4375, + "learning_rate": 2.742040817433761e-08, + "loss": 0.4616, + "step": 21909 + }, + { + "epoch": 1.8670643374520663, + "grad_norm": 14.5, + "learning_rate": 2.738541638305392e-08, + "loss": 0.391, + "step": 21910 + }, + { + "epoch": 1.8671495526203663, + "grad_norm": 23.0, + "learning_rate": 2.7350446685645017e-08, + "loss": 0.8314, + "step": 21911 + }, + { + "epoch": 1.8672347677886663, + "grad_norm": 18.125, + "learning_rate": 2.731549908274289e-08, + "loss": 0.8527, + "step": 21912 + }, + { + "epoch": 1.8673199829569662, + "grad_norm": 15.6875, + "learning_rate": 2.7280573574979402e-08, + "loss": 0.5991, + "step": 21913 + }, + { + "epoch": 1.8674051981252662, + "grad_norm": 13.125, + "learning_rate": 2.7245670162985433e-08, + "loss": 0.3721, + "step": 21914 + }, + { + "epoch": 1.8674904132935661, + "grad_norm": 17.375, + "learning_rate": 2.7210788847391867e-08, + "loss": 0.6904, + "step": 21915 + }, + { + "epoch": 1.867575628461866, + "grad_norm": 17.625, + "learning_rate": 2.7175929628828894e-08, + "loss": 0.522, + "step": 21916 + }, + { + "epoch": 1.867660843630166, + "grad_norm": 15.875, + "learning_rate": 2.714109250792643e-08, + "loss": 0.6886, + "step": 21917 + }, + { + "epoch": 1.867746058798466, + "grad_norm": 14.8125, + "learning_rate": 2.7106277485314247e-08, + "loss": 0.3627, + "step": 21918 + }, + { + "epoch": 1.867831273966766, + "grad_norm": 23.5, + "learning_rate": 2.7071484561621146e-08, + "loss": 0.6687, + "step": 21919 + }, + { + "epoch": 1.867916489135066, + "grad_norm": 15.3125, + "learning_rate": 2.703671373747621e-08, + "loss": 0.6993, + "step": 21920 + }, + { + "epoch": 1.8680017043033659, + "grad_norm": 18.125, + "learning_rate": 2.7001965013507546e-08, + "loss": 0.7583, + "step": 21921 + }, + { + "epoch": 1.8680869194716658, + "grad_norm": 14.0, + "learning_rate": 2.6967238390343403e-08, + "loss": 0.6386, + "step": 21922 + }, + { + "epoch": 1.8681721346399658, + "grad_norm": 14.5625, + "learning_rate": 2.6932533868610917e-08, + "loss": 0.3819, + "step": 21923 + }, + { + "epoch": 1.8682573498082657, + "grad_norm": 15.6875, + "learning_rate": 2.68978514489375e-08, + "loss": 0.5252, + "step": 21924 + }, + { + "epoch": 1.8683425649765657, + "grad_norm": 22.5, + "learning_rate": 2.6863191131949878e-08, + "loss": 0.9955, + "step": 21925 + }, + { + "epoch": 1.8684277801448657, + "grad_norm": 12.4375, + "learning_rate": 2.6828552918274216e-08, + "loss": 0.3316, + "step": 21926 + }, + { + "epoch": 1.8685129953131656, + "grad_norm": 19.25, + "learning_rate": 2.6793936808536813e-08, + "loss": 0.6341, + "step": 21927 + }, + { + "epoch": 1.8685982104814656, + "grad_norm": 14.9375, + "learning_rate": 2.6759342803362732e-08, + "loss": 0.4685, + "step": 21928 + }, + { + "epoch": 1.8686834256497655, + "grad_norm": 17.0, + "learning_rate": 2.6724770903377578e-08, + "loss": 0.8278, + "step": 21929 + }, + { + "epoch": 1.8687686408180655, + "grad_norm": 14.8125, + "learning_rate": 2.6690221109205854e-08, + "loss": 0.5974, + "step": 21930 + }, + { + "epoch": 1.8688538559863654, + "grad_norm": 20.25, + "learning_rate": 2.6655693421472062e-08, + "loss": 0.7381, + "step": 21931 + }, + { + "epoch": 1.8689390711546654, + "grad_norm": 14.5, + "learning_rate": 2.6621187840800006e-08, + "loss": 0.4029, + "step": 21932 + }, + { + "epoch": 1.8690242863229654, + "grad_norm": 15.375, + "learning_rate": 2.6586704367813215e-08, + "loss": 0.4434, + "step": 21933 + }, + { + "epoch": 1.8691095014912653, + "grad_norm": 14.0625, + "learning_rate": 2.655224300313494e-08, + "loss": 0.3934, + "step": 21934 + }, + { + "epoch": 1.8691947166595653, + "grad_norm": 21.75, + "learning_rate": 2.6517803747387882e-08, + "loss": 0.956, + "step": 21935 + }, + { + "epoch": 1.8692799318278652, + "grad_norm": 14.5, + "learning_rate": 2.6483386601194456e-08, + "loss": 0.4541, + "step": 21936 + }, + { + "epoch": 1.8693651469961652, + "grad_norm": 13.9375, + "learning_rate": 2.6448991565176525e-08, + "loss": 0.4694, + "step": 21937 + }, + { + "epoch": 1.8694503621644651, + "grad_norm": 16.5, + "learning_rate": 2.6414618639955675e-08, + "loss": 0.4535, + "step": 21938 + }, + { + "epoch": 1.869535577332765, + "grad_norm": 12.375, + "learning_rate": 2.6380267826153216e-08, + "loss": 0.4347, + "step": 21939 + }, + { + "epoch": 1.869620792501065, + "grad_norm": 14.25, + "learning_rate": 2.6345939124389763e-08, + "loss": 0.4251, + "step": 21940 + }, + { + "epoch": 1.869706007669365, + "grad_norm": 12.6875, + "learning_rate": 2.6311632535285653e-08, + "loss": 0.3962, + "step": 21941 + }, + { + "epoch": 1.869791222837665, + "grad_norm": 16.75, + "learning_rate": 2.6277348059460667e-08, + "loss": 0.9155, + "step": 21942 + }, + { + "epoch": 1.869876438005965, + "grad_norm": 21.0, + "learning_rate": 2.6243085697534725e-08, + "loss": 0.8807, + "step": 21943 + }, + { + "epoch": 1.8699616531742649, + "grad_norm": 12.375, + "learning_rate": 2.6208845450126778e-08, + "loss": 0.4244, + "step": 21944 + }, + { + "epoch": 1.8700468683425648, + "grad_norm": 27.625, + "learning_rate": 2.61746273178555e-08, + "loss": 0.8291, + "step": 21945 + }, + { + "epoch": 1.8701320835108648, + "grad_norm": 17.75, + "learning_rate": 2.614043130133942e-08, + "loss": 0.6901, + "step": 21946 + }, + { + "epoch": 1.8702172986791648, + "grad_norm": 14.9375, + "learning_rate": 2.6106257401196654e-08, + "loss": 0.6168, + "step": 21947 + }, + { + "epoch": 1.8703025138474647, + "grad_norm": 13.3125, + "learning_rate": 2.6072105618044354e-08, + "loss": 0.5937, + "step": 21948 + }, + { + "epoch": 1.8703877290157647, + "grad_norm": 13.25, + "learning_rate": 2.6037975952499795e-08, + "loss": 0.5186, + "step": 21949 + }, + { + "epoch": 1.8704729441840646, + "grad_norm": 11.8125, + "learning_rate": 2.6003868405179848e-08, + "loss": 0.3299, + "step": 21950 + }, + { + "epoch": 1.8705581593523646, + "grad_norm": 15.8125, + "learning_rate": 2.596978297670083e-08, + "loss": 0.5085, + "step": 21951 + }, + { + "epoch": 1.8706433745206645, + "grad_norm": 20.125, + "learning_rate": 2.593571966767877e-08, + "loss": 0.6104, + "step": 21952 + }, + { + "epoch": 1.8707285896889645, + "grad_norm": 14.0625, + "learning_rate": 2.5901678478729015e-08, + "loss": 0.4664, + "step": 21953 + }, + { + "epoch": 1.8708138048572645, + "grad_norm": 15.1875, + "learning_rate": 2.58676594104669e-08, + "loss": 0.8993, + "step": 21954 + }, + { + "epoch": 1.8708990200255644, + "grad_norm": 24.375, + "learning_rate": 2.5833662463507076e-08, + "loss": 0.5503, + "step": 21955 + }, + { + "epoch": 1.8709842351938644, + "grad_norm": 14.25, + "learning_rate": 2.579968763846405e-08, + "loss": 0.4742, + "step": 21956 + }, + { + "epoch": 1.8710694503621643, + "grad_norm": 16.0, + "learning_rate": 2.5765734935951775e-08, + "loss": 0.5133, + "step": 21957 + }, + { + "epoch": 1.8711546655304643, + "grad_norm": 17.75, + "learning_rate": 2.573180435658365e-08, + "loss": 0.5561, + "step": 21958 + }, + { + "epoch": 1.8712398806987642, + "grad_norm": 13.5625, + "learning_rate": 2.5697895900972934e-08, + "loss": 0.5137, + "step": 21959 + }, + { + "epoch": 1.8713250958670642, + "grad_norm": 18.875, + "learning_rate": 2.5664009569732334e-08, + "loss": 0.5218, + "step": 21960 + }, + { + "epoch": 1.8714103110353641, + "grad_norm": 13.9375, + "learning_rate": 2.5630145363474274e-08, + "loss": 0.5542, + "step": 21961 + }, + { + "epoch": 1.871495526203664, + "grad_norm": 17.875, + "learning_rate": 2.5596303282810626e-08, + "loss": 0.4552, + "step": 21962 + }, + { + "epoch": 1.871580741371964, + "grad_norm": 13.25, + "learning_rate": 2.5562483328352983e-08, + "loss": 0.6426, + "step": 21963 + }, + { + "epoch": 1.871665956540264, + "grad_norm": 14.3125, + "learning_rate": 2.5528685500712663e-08, + "loss": 0.524, + "step": 21964 + }, + { + "epoch": 1.871751171708564, + "grad_norm": 12.125, + "learning_rate": 2.5494909800500426e-08, + "loss": 0.4173, + "step": 21965 + }, + { + "epoch": 1.871836386876864, + "grad_norm": 14.1875, + "learning_rate": 2.5461156228326207e-08, + "loss": 0.573, + "step": 21966 + }, + { + "epoch": 1.871921602045164, + "grad_norm": 16.5, + "learning_rate": 2.542742478480034e-08, + "loss": 0.6359, + "step": 21967 + }, + { + "epoch": 1.8720068172134638, + "grad_norm": 16.125, + "learning_rate": 2.539371547053235e-08, + "loss": 0.5192, + "step": 21968 + }, + { + "epoch": 1.8720920323817638, + "grad_norm": 18.125, + "learning_rate": 2.536002828613146e-08, + "loss": 0.4916, + "step": 21969 + }, + { + "epoch": 1.8721772475500638, + "grad_norm": 13.1875, + "learning_rate": 2.532636323220608e-08, + "loss": 0.366, + "step": 21970 + }, + { + "epoch": 1.8722624627183637, + "grad_norm": 23.625, + "learning_rate": 2.5292720309365027e-08, + "loss": 0.4947, + "step": 21971 + }, + { + "epoch": 1.8723476778866637, + "grad_norm": 15.125, + "learning_rate": 2.5259099518216008e-08, + "loss": 0.8987, + "step": 21972 + }, + { + "epoch": 1.8724328930549636, + "grad_norm": 11.9375, + "learning_rate": 2.522550085936659e-08, + "loss": 0.2877, + "step": 21973 + }, + { + "epoch": 1.8725181082232636, + "grad_norm": 15.375, + "learning_rate": 2.519192433342407e-08, + "loss": 0.6671, + "step": 21974 + }, + { + "epoch": 1.8726033233915635, + "grad_norm": 14.6875, + "learning_rate": 2.5158369940995048e-08, + "loss": 0.6552, + "step": 21975 + }, + { + "epoch": 1.8726885385598635, + "grad_norm": 12.25, + "learning_rate": 2.5124837682685976e-08, + "loss": 0.3795, + "step": 21976 + }, + { + "epoch": 1.8727737537281635, + "grad_norm": 15.4375, + "learning_rate": 2.5091327559102762e-08, + "loss": 0.771, + "step": 21977 + }, + { + "epoch": 1.8728589688964634, + "grad_norm": 13.625, + "learning_rate": 2.5057839570851032e-08, + "loss": 0.5563, + "step": 21978 + }, + { + "epoch": 1.8729441840647634, + "grad_norm": 22.25, + "learning_rate": 2.502437371853586e-08, + "loss": 0.5065, + "step": 21979 + }, + { + "epoch": 1.8730293992330633, + "grad_norm": 16.0, + "learning_rate": 2.4990930002762177e-08, + "loss": 0.7986, + "step": 21980 + }, + { + "epoch": 1.8731146144013633, + "grad_norm": 14.0, + "learning_rate": 2.495750842413408e-08, + "loss": 0.5089, + "step": 21981 + }, + { + "epoch": 1.8731998295696632, + "grad_norm": 12.5, + "learning_rate": 2.4924108983255813e-08, + "loss": 0.491, + "step": 21982 + }, + { + "epoch": 1.8732850447379632, + "grad_norm": 14.8125, + "learning_rate": 2.4890731680730778e-08, + "loss": 0.6292, + "step": 21983 + }, + { + "epoch": 1.8733702599062632, + "grad_norm": 20.0, + "learning_rate": 2.4857376517162107e-08, + "loss": 0.8556, + "step": 21984 + }, + { + "epoch": 1.8734554750745631, + "grad_norm": 13.875, + "learning_rate": 2.482404349315265e-08, + "loss": 0.4589, + "step": 21985 + }, + { + "epoch": 1.873540690242863, + "grad_norm": 11.625, + "learning_rate": 2.47907326093047e-08, + "loss": 0.2202, + "step": 21986 + }, + { + "epoch": 1.873625905411163, + "grad_norm": 14.5, + "learning_rate": 2.4757443866220282e-08, + "loss": 0.425, + "step": 21987 + }, + { + "epoch": 1.873711120579463, + "grad_norm": 16.875, + "learning_rate": 2.4724177264500994e-08, + "loss": 0.7121, + "step": 21988 + }, + { + "epoch": 1.873796335747763, + "grad_norm": 28.25, + "learning_rate": 2.469093280474788e-08, + "loss": 0.3728, + "step": 21989 + }, + { + "epoch": 1.873881550916063, + "grad_norm": 12.3125, + "learning_rate": 2.4657710487561854e-08, + "loss": 0.4249, + "step": 21990 + }, + { + "epoch": 1.8739667660843629, + "grad_norm": 14.3125, + "learning_rate": 2.4624510313542986e-08, + "loss": 0.4398, + "step": 21991 + }, + { + "epoch": 1.8740519812526628, + "grad_norm": 21.625, + "learning_rate": 2.4591332283291498e-08, + "loss": 0.4363, + "step": 21992 + }, + { + "epoch": 1.8741371964209628, + "grad_norm": 21.625, + "learning_rate": 2.4558176397406902e-08, + "loss": 0.5688, + "step": 21993 + }, + { + "epoch": 1.8742224115892627, + "grad_norm": 20.625, + "learning_rate": 2.4525042656488306e-08, + "loss": 0.7042, + "step": 21994 + }, + { + "epoch": 1.8743076267575627, + "grad_norm": 17.75, + "learning_rate": 2.449193106113454e-08, + "loss": 0.7033, + "step": 21995 + }, + { + "epoch": 1.8743928419258626, + "grad_norm": 17.5, + "learning_rate": 2.4458841611943866e-08, + "loss": 1.0877, + "step": 21996 + }, + { + "epoch": 1.8744780570941626, + "grad_norm": 19.875, + "learning_rate": 2.442577430951443e-08, + "loss": 0.6551, + "step": 21997 + }, + { + "epoch": 1.8745632722624626, + "grad_norm": 15.5, + "learning_rate": 2.4392729154443516e-08, + "loss": 0.5409, + "step": 21998 + }, + { + "epoch": 1.8746484874307625, + "grad_norm": 19.875, + "learning_rate": 2.4359706147328576e-08, + "loss": 0.7905, + "step": 21999 + }, + { + "epoch": 1.8747337025990625, + "grad_norm": 18.375, + "learning_rate": 2.4326705288766074e-08, + "loss": 0.6612, + "step": 22000 + }, + { + "epoch": 1.8748189177673624, + "grad_norm": 14.25, + "learning_rate": 2.429372657935261e-08, + "loss": 0.7088, + "step": 22001 + }, + { + "epoch": 1.8749041329356624, + "grad_norm": 18.875, + "learning_rate": 2.4260770019683965e-08, + "loss": 0.6104, + "step": 22002 + }, + { + "epoch": 1.8749893481039623, + "grad_norm": 12.5625, + "learning_rate": 2.422783561035591e-08, + "loss": 0.4768, + "step": 22003 + }, + { + "epoch": 1.8750745632722625, + "grad_norm": 15.5625, + "learning_rate": 2.4194923351963383e-08, + "loss": 0.8429, + "step": 22004 + }, + { + "epoch": 1.8751597784405625, + "grad_norm": 20.25, + "learning_rate": 2.416203324510133e-08, + "loss": 0.4881, + "step": 22005 + }, + { + "epoch": 1.8752449936088624, + "grad_norm": 14.25, + "learning_rate": 2.4129165290363854e-08, + "loss": 0.4009, + "step": 22006 + }, + { + "epoch": 1.8753302087771624, + "grad_norm": 14.8125, + "learning_rate": 2.4096319488345204e-08, + "loss": 0.6384, + "step": 22007 + }, + { + "epoch": 1.8754154239454623, + "grad_norm": 11.3125, + "learning_rate": 2.4063495839638655e-08, + "loss": 0.3404, + "step": 22008 + }, + { + "epoch": 1.8755006391137623, + "grad_norm": 11.5, + "learning_rate": 2.4030694344837625e-08, + "loss": 0.3316, + "step": 22009 + }, + { + "epoch": 1.8755858542820623, + "grad_norm": 17.5, + "learning_rate": 2.3997915004534687e-08, + "loss": 0.8472, + "step": 22010 + }, + { + "epoch": 1.8756710694503622, + "grad_norm": 25.625, + "learning_rate": 2.3965157819322433e-08, + "loss": 0.4261, + "step": 22011 + }, + { + "epoch": 1.8757562846186622, + "grad_norm": 10.625, + "learning_rate": 2.3932422789792463e-08, + "loss": 0.3023, + "step": 22012 + }, + { + "epoch": 1.8758414997869621, + "grad_norm": 18.5, + "learning_rate": 2.3899709916536675e-08, + "loss": 0.5454, + "step": 22013 + }, + { + "epoch": 1.875926714955262, + "grad_norm": 14.8125, + "learning_rate": 2.3867019200145836e-08, + "loss": 0.52, + "step": 22014 + }, + { + "epoch": 1.876011930123562, + "grad_norm": 13.9375, + "learning_rate": 2.3834350641211147e-08, + "loss": 0.2894, + "step": 22015 + }, + { + "epoch": 1.876097145291862, + "grad_norm": 23.0, + "learning_rate": 2.3801704240322688e-08, + "loss": 0.7118, + "step": 22016 + }, + { + "epoch": 1.876182360460162, + "grad_norm": 20.5, + "learning_rate": 2.3769079998070404e-08, + "loss": 0.6109, + "step": 22017 + }, + { + "epoch": 1.876267575628462, + "grad_norm": 15.6875, + "learning_rate": 2.3736477915043964e-08, + "loss": 0.4807, + "step": 22018 + }, + { + "epoch": 1.8763527907967619, + "grad_norm": 16.5, + "learning_rate": 2.370389799183248e-08, + "loss": 0.5337, + "step": 22019 + }, + { + "epoch": 1.8764380059650618, + "grad_norm": 13.4375, + "learning_rate": 2.3671340229024646e-08, + "loss": 0.6018, + "step": 22020 + }, + { + "epoch": 1.8765232211333618, + "grad_norm": 17.875, + "learning_rate": 2.3638804627209022e-08, + "loss": 0.9148, + "step": 22021 + }, + { + "epoch": 1.8766084363016617, + "grad_norm": 19.25, + "learning_rate": 2.3606291186973328e-08, + "loss": 1.0585, + "step": 22022 + }, + { + "epoch": 1.8766936514699617, + "grad_norm": 16.375, + "learning_rate": 2.3573799908905013e-08, + "loss": 0.5612, + "step": 22023 + }, + { + "epoch": 1.8767788666382617, + "grad_norm": 15.0, + "learning_rate": 2.3541330793591665e-08, + "loss": 0.6944, + "step": 22024 + }, + { + "epoch": 1.8768640818065616, + "grad_norm": 14.5625, + "learning_rate": 2.3508883841619617e-08, + "loss": 0.5887, + "step": 22025 + }, + { + "epoch": 1.8769492969748616, + "grad_norm": 17.25, + "learning_rate": 2.3476459053575345e-08, + "loss": 0.8026, + "step": 22026 + }, + { + "epoch": 1.8770345121431615, + "grad_norm": 11.5, + "learning_rate": 2.344405643004477e-08, + "loss": 0.3071, + "step": 22027 + }, + { + "epoch": 1.8771197273114615, + "grad_norm": 15.375, + "learning_rate": 2.3411675971613673e-08, + "loss": 0.7051, + "step": 22028 + }, + { + "epoch": 1.8772049424797614, + "grad_norm": 11.0, + "learning_rate": 2.3379317678867002e-08, + "loss": 0.3305, + "step": 22029 + }, + { + "epoch": 1.8772901576480614, + "grad_norm": 12.0, + "learning_rate": 2.334698155238943e-08, + "loss": 0.3486, + "step": 22030 + }, + { + "epoch": 1.8773753728163614, + "grad_norm": 12.8125, + "learning_rate": 2.331466759276535e-08, + "loss": 0.5568, + "step": 22031 + }, + { + "epoch": 1.8774605879846613, + "grad_norm": 19.625, + "learning_rate": 2.3282375800578878e-08, + "loss": 0.6735, + "step": 22032 + }, + { + "epoch": 1.8775458031529613, + "grad_norm": 14.375, + "learning_rate": 2.3250106176413435e-08, + "loss": 0.4478, + "step": 22033 + }, + { + "epoch": 1.8776310183212612, + "grad_norm": 14.1875, + "learning_rate": 2.3217858720852028e-08, + "loss": 0.6183, + "step": 22034 + }, + { + "epoch": 1.8777162334895612, + "grad_norm": 18.375, + "learning_rate": 2.318563343447766e-08, + "loss": 0.7437, + "step": 22035 + }, + { + "epoch": 1.8778014486578611, + "grad_norm": 17.875, + "learning_rate": 2.315343031787237e-08, + "loss": 0.9006, + "step": 22036 + }, + { + "epoch": 1.877886663826161, + "grad_norm": 13.625, + "learning_rate": 2.3121249371618747e-08, + "loss": 0.518, + "step": 22037 + }, + { + "epoch": 1.877971878994461, + "grad_norm": 16.0, + "learning_rate": 2.3089090596297435e-08, + "loss": 0.8933, + "step": 22038 + }, + { + "epoch": 1.878057094162761, + "grad_norm": 20.625, + "learning_rate": 2.3056953992490055e-08, + "loss": 1.0285, + "step": 22039 + }, + { + "epoch": 1.878142309331061, + "grad_norm": 17.5, + "learning_rate": 2.302483956077739e-08, + "loss": 0.4511, + "step": 22040 + }, + { + "epoch": 1.878227524499361, + "grad_norm": 19.75, + "learning_rate": 2.299274730173981e-08, + "loss": 0.5839, + "step": 22041 + }, + { + "epoch": 1.8783127396676609, + "grad_norm": 14.1875, + "learning_rate": 2.2960677215956988e-08, + "loss": 0.5305, + "step": 22042 + }, + { + "epoch": 1.8783979548359608, + "grad_norm": 13.875, + "learning_rate": 2.2928629304008602e-08, + "loss": 0.5578, + "step": 22043 + }, + { + "epoch": 1.8784831700042608, + "grad_norm": 21.0, + "learning_rate": 2.2896603566473773e-08, + "loss": 0.7825, + "step": 22044 + }, + { + "epoch": 1.8785683851725608, + "grad_norm": 15.6875, + "learning_rate": 2.2864600003931343e-08, + "loss": 0.7924, + "step": 22045 + }, + { + "epoch": 1.8786536003408607, + "grad_norm": 14.5625, + "learning_rate": 2.2832618616959596e-08, + "loss": 0.6439, + "step": 22046 + }, + { + "epoch": 1.8787388155091607, + "grad_norm": 18.125, + "learning_rate": 2.2800659406136404e-08, + "loss": 0.2738, + "step": 22047 + }, + { + "epoch": 1.8788240306774606, + "grad_norm": 17.875, + "learning_rate": 2.2768722372039364e-08, + "loss": 0.4208, + "step": 22048 + }, + { + "epoch": 1.8789092458457606, + "grad_norm": 17.875, + "learning_rate": 2.273680751524551e-08, + "loss": 0.7611, + "step": 22049 + }, + { + "epoch": 1.8789944610140605, + "grad_norm": 15.8125, + "learning_rate": 2.2704914836331748e-08, + "loss": 0.4722, + "step": 22050 + }, + { + "epoch": 1.8790796761823605, + "grad_norm": 14.6875, + "learning_rate": 2.2673044335874138e-08, + "loss": 0.8036, + "step": 22051 + }, + { + "epoch": 1.8791648913506604, + "grad_norm": 18.25, + "learning_rate": 2.264119601444889e-08, + "loss": 0.5202, + "step": 22052 + }, + { + "epoch": 1.8792501065189604, + "grad_norm": 12.625, + "learning_rate": 2.2609369872631652e-08, + "loss": 0.5062, + "step": 22053 + }, + { + "epoch": 1.8793353216872604, + "grad_norm": 17.5, + "learning_rate": 2.2577565910997106e-08, + "loss": 0.5519, + "step": 22054 + }, + { + "epoch": 1.8794205368555603, + "grad_norm": 13.3125, + "learning_rate": 2.2545784130120347e-08, + "loss": 0.3928, + "step": 22055 + }, + { + "epoch": 1.8795057520238603, + "grad_norm": 17.0, + "learning_rate": 2.2514024530575357e-08, + "loss": 0.3066, + "step": 22056 + }, + { + "epoch": 1.8795909671921602, + "grad_norm": 13.75, + "learning_rate": 2.2482287112936403e-08, + "loss": 0.4484, + "step": 22057 + }, + { + "epoch": 1.8796761823604602, + "grad_norm": 12.625, + "learning_rate": 2.245057187777691e-08, + "loss": 0.4606, + "step": 22058 + }, + { + "epoch": 1.8797613975287601, + "grad_norm": 15.5, + "learning_rate": 2.2418878825669897e-08, + "loss": 0.7741, + "step": 22059 + }, + { + "epoch": 1.87984661269706, + "grad_norm": 15.6875, + "learning_rate": 2.238720795718824e-08, + "loss": 0.5149, + "step": 22060 + }, + { + "epoch": 1.87993182786536, + "grad_norm": 11.9375, + "learning_rate": 2.2355559272904116e-08, + "loss": 0.434, + "step": 22061 + }, + { + "epoch": 1.88001704303366, + "grad_norm": 14.9375, + "learning_rate": 2.2323932773389706e-08, + "loss": 0.5591, + "step": 22062 + }, + { + "epoch": 1.88010225820196, + "grad_norm": 14.0625, + "learning_rate": 2.2292328459216362e-08, + "loss": 0.5445, + "step": 22063 + }, + { + "epoch": 1.88018747337026, + "grad_norm": 19.875, + "learning_rate": 2.2260746330955014e-08, + "loss": 0.9148, + "step": 22064 + }, + { + "epoch": 1.8802726885385599, + "grad_norm": 12.3125, + "learning_rate": 2.2229186389176593e-08, + "loss": 0.5505, + "step": 22065 + }, + { + "epoch": 1.8803579037068598, + "grad_norm": 17.125, + "learning_rate": 2.219764863445134e-08, + "loss": 0.4022, + "step": 22066 + }, + { + "epoch": 1.8804431188751598, + "grad_norm": 15.3125, + "learning_rate": 2.2166133067349354e-08, + "loss": 0.719, + "step": 22067 + }, + { + "epoch": 1.8805283340434598, + "grad_norm": 16.25, + "learning_rate": 2.213463968844004e-08, + "loss": 0.5788, + "step": 22068 + }, + { + "epoch": 1.8806135492117597, + "grad_norm": 21.0, + "learning_rate": 2.2103168498292387e-08, + "loss": 0.8303, + "step": 22069 + }, + { + "epoch": 1.8806987643800597, + "grad_norm": 13.9375, + "learning_rate": 2.2071719497475384e-08, + "loss": 0.4182, + "step": 22070 + }, + { + "epoch": 1.8807839795483596, + "grad_norm": 16.5, + "learning_rate": 2.204029268655719e-08, + "loss": 0.741, + "step": 22071 + }, + { + "epoch": 1.8808691947166596, + "grad_norm": 16.625, + "learning_rate": 2.2008888066105544e-08, + "loss": 0.5421, + "step": 22072 + }, + { + "epoch": 1.8809544098849595, + "grad_norm": 14.3125, + "learning_rate": 2.1977505636688184e-08, + "loss": 0.5588, + "step": 22073 + }, + { + "epoch": 1.8810396250532595, + "grad_norm": 13.9375, + "learning_rate": 2.194614539887216e-08, + "loss": 0.4855, + "step": 22074 + }, + { + "epoch": 1.8811248402215595, + "grad_norm": 23.125, + "learning_rate": 2.1914807353224236e-08, + "loss": 0.6756, + "step": 22075 + }, + { + "epoch": 1.8812100553898594, + "grad_norm": 11.3125, + "learning_rate": 2.1883491500310634e-08, + "loss": 0.2515, + "step": 22076 + }, + { + "epoch": 1.8812952705581594, + "grad_norm": 18.375, + "learning_rate": 2.1852197840697282e-08, + "loss": 0.4925, + "step": 22077 + }, + { + "epoch": 1.8813804857264593, + "grad_norm": 19.25, + "learning_rate": 2.182092637494984e-08, + "loss": 0.636, + "step": 22078 + }, + { + "epoch": 1.8814657008947593, + "grad_norm": 15.75, + "learning_rate": 2.1789677103633278e-08, + "loss": 0.5575, + "step": 22079 + }, + { + "epoch": 1.8815509160630592, + "grad_norm": 15.3125, + "learning_rate": 2.175845002731228e-08, + "loss": 0.6264, + "step": 22080 + }, + { + "epoch": 1.8816361312313592, + "grad_norm": 14.375, + "learning_rate": 2.172724514655111e-08, + "loss": 0.5365, + "step": 22081 + }, + { + "epoch": 1.8817213463996592, + "grad_norm": 13.5625, + "learning_rate": 2.1696062461913775e-08, + "loss": 0.3233, + "step": 22082 + }, + { + "epoch": 1.8818065615679591, + "grad_norm": 18.0, + "learning_rate": 2.1664901973963704e-08, + "loss": 0.5785, + "step": 22083 + }, + { + "epoch": 1.881891776736259, + "grad_norm": 14.875, + "learning_rate": 2.1633763683264197e-08, + "loss": 0.5567, + "step": 22084 + }, + { + "epoch": 1.881976991904559, + "grad_norm": 19.125, + "learning_rate": 2.1602647590377585e-08, + "loss": 0.6331, + "step": 22085 + }, + { + "epoch": 1.882062207072859, + "grad_norm": 16.75, + "learning_rate": 2.157155369586661e-08, + "loss": 0.5978, + "step": 22086 + }, + { + "epoch": 1.882147422241159, + "grad_norm": 19.0, + "learning_rate": 2.1540482000292767e-08, + "loss": 1.0022, + "step": 22087 + }, + { + "epoch": 1.882232637409459, + "grad_norm": 16.5, + "learning_rate": 2.1509432504217832e-08, + "loss": 0.8776, + "step": 22088 + }, + { + "epoch": 1.8823178525777589, + "grad_norm": 22.625, + "learning_rate": 2.1478405208202742e-08, + "loss": 0.9097, + "step": 22089 + }, + { + "epoch": 1.8824030677460588, + "grad_norm": 19.25, + "learning_rate": 2.1447400112808157e-08, + "loss": 0.9157, + "step": 22090 + }, + { + "epoch": 1.8824882829143588, + "grad_norm": 14.9375, + "learning_rate": 2.141641721859461e-08, + "loss": 0.7165, + "step": 22091 + }, + { + "epoch": 1.8825734980826587, + "grad_norm": 13.0625, + "learning_rate": 2.1385456526121786e-08, + "loss": 0.5067, + "step": 22092 + }, + { + "epoch": 1.8826587132509587, + "grad_norm": 14.625, + "learning_rate": 2.1354518035949233e-08, + "loss": 0.5628, + "step": 22093 + }, + { + "epoch": 1.8827439284192586, + "grad_norm": 17.625, + "learning_rate": 2.1323601748636096e-08, + "loss": 0.9869, + "step": 22094 + }, + { + "epoch": 1.8828291435875586, + "grad_norm": 15.5, + "learning_rate": 2.1292707664740813e-08, + "loss": 0.5149, + "step": 22095 + }, + { + "epoch": 1.8829143587558586, + "grad_norm": 23.875, + "learning_rate": 2.1261835784821966e-08, + "loss": 0.6424, + "step": 22096 + }, + { + "epoch": 1.8829995739241585, + "grad_norm": 12.0625, + "learning_rate": 2.123098610943744e-08, + "loss": 0.3805, + "step": 22097 + }, + { + "epoch": 1.8830847890924585, + "grad_norm": 15.6875, + "learning_rate": 2.120015863914443e-08, + "loss": 0.3814, + "step": 22098 + }, + { + "epoch": 1.8831700042607584, + "grad_norm": 15.75, + "learning_rate": 2.1169353374500407e-08, + "loss": 0.5576, + "step": 22099 + }, + { + "epoch": 1.8832552194290584, + "grad_norm": 15.3125, + "learning_rate": 2.113857031606173e-08, + "loss": 0.7356, + "step": 22100 + }, + { + "epoch": 1.8833404345973583, + "grad_norm": 17.125, + "learning_rate": 2.110780946438476e-08, + "loss": 0.495, + "step": 22101 + }, + { + "epoch": 1.8834256497656583, + "grad_norm": 15.125, + "learning_rate": 2.107707082002558e-08, + "loss": 0.5506, + "step": 22102 + }, + { + "epoch": 1.8835108649339583, + "grad_norm": 11.375, + "learning_rate": 2.1046354383539442e-08, + "loss": 0.4423, + "step": 22103 + }, + { + "epoch": 1.8835960801022582, + "grad_norm": 21.625, + "learning_rate": 2.1015660155481456e-08, + "loss": 0.4508, + "step": 22104 + }, + { + "epoch": 1.8836812952705582, + "grad_norm": 16.25, + "learning_rate": 2.0984988136406313e-08, + "loss": 0.413, + "step": 22105 + }, + { + "epoch": 1.8837665104388581, + "grad_norm": 20.75, + "learning_rate": 2.0954338326868303e-08, + "loss": 0.6871, + "step": 22106 + }, + { + "epoch": 1.883851725607158, + "grad_norm": 13.6875, + "learning_rate": 2.0923710727421277e-08, + "loss": 0.5738, + "step": 22107 + }, + { + "epoch": 1.883936940775458, + "grad_norm": 13.3125, + "learning_rate": 2.0893105338618825e-08, + "loss": 0.4429, + "step": 22108 + }, + { + "epoch": 1.884022155943758, + "grad_norm": 14.6875, + "learning_rate": 2.0862522161013975e-08, + "loss": 0.5472, + "step": 22109 + }, + { + "epoch": 1.884107371112058, + "grad_norm": 15.6875, + "learning_rate": 2.0831961195159205e-08, + "loss": 0.6662, + "step": 22110 + }, + { + "epoch": 1.884192586280358, + "grad_norm": 17.875, + "learning_rate": 2.0801422441606987e-08, + "loss": 0.6349, + "step": 22111 + }, + { + "epoch": 1.8842778014486579, + "grad_norm": 16.25, + "learning_rate": 2.0770905900909104e-08, + "loss": 0.9843, + "step": 22112 + }, + { + "epoch": 1.8843630166169578, + "grad_norm": 13.4375, + "learning_rate": 2.074041157361692e-08, + "loss": 0.5592, + "step": 22113 + }, + { + "epoch": 1.8844482317852578, + "grad_norm": 14.875, + "learning_rate": 2.0709939460281798e-08, + "loss": 0.6278, + "step": 22114 + }, + { + "epoch": 1.8845334469535577, + "grad_norm": 17.0, + "learning_rate": 2.067948956145413e-08, + "loss": 0.591, + "step": 22115 + }, + { + "epoch": 1.8846186621218577, + "grad_norm": 20.125, + "learning_rate": 2.0649061877684173e-08, + "loss": 0.9596, + "step": 22116 + }, + { + "epoch": 1.8847038772901576, + "grad_norm": 14.6875, + "learning_rate": 2.061865640952204e-08, + "loss": 0.2455, + "step": 22117 + }, + { + "epoch": 1.8847890924584576, + "grad_norm": 16.375, + "learning_rate": 2.058827315751702e-08, + "loss": 0.438, + "step": 22118 + }, + { + "epoch": 1.8848743076267576, + "grad_norm": 10.75, + "learning_rate": 2.055791212221811e-08, + "loss": 0.6009, + "step": 22119 + }, + { + "epoch": 1.8849595227950575, + "grad_norm": 12.5625, + "learning_rate": 2.052757330417404e-08, + "loss": 0.431, + "step": 22120 + }, + { + "epoch": 1.8850447379633575, + "grad_norm": 37.0, + "learning_rate": 2.0497256703932988e-08, + "loss": 1.1228, + "step": 22121 + }, + { + "epoch": 1.8851299531316574, + "grad_norm": 12.9375, + "learning_rate": 2.0466962322042984e-08, + "loss": 0.3412, + "step": 22122 + }, + { + "epoch": 1.8852151682999574, + "grad_norm": 15.625, + "learning_rate": 2.0436690159051225e-08, + "loss": 0.8763, + "step": 22123 + }, + { + "epoch": 1.8853003834682573, + "grad_norm": 16.875, + "learning_rate": 2.040644021550492e-08, + "loss": 0.5438, + "step": 22124 + }, + { + "epoch": 1.8853855986365573, + "grad_norm": 14.5, + "learning_rate": 2.037621249195071e-08, + "loss": 0.5155, + "step": 22125 + }, + { + "epoch": 1.8854708138048573, + "grad_norm": 19.25, + "learning_rate": 2.0346006988934962e-08, + "loss": 0.8729, + "step": 22126 + }, + { + "epoch": 1.8855560289731572, + "grad_norm": 13.6875, + "learning_rate": 2.0315823707003358e-08, + "loss": 0.4281, + "step": 22127 + }, + { + "epoch": 1.8856412441414572, + "grad_norm": 16.375, + "learning_rate": 2.0285662646701286e-08, + "loss": 0.764, + "step": 22128 + }, + { + "epoch": 1.8857264593097571, + "grad_norm": 26.25, + "learning_rate": 2.0255523808573734e-08, + "loss": 0.5942, + "step": 22129 + }, + { + "epoch": 1.885811674478057, + "grad_norm": 38.75, + "learning_rate": 2.022540719316568e-08, + "loss": 0.916, + "step": 22130 + }, + { + "epoch": 1.885896889646357, + "grad_norm": 23.25, + "learning_rate": 2.0195312801020993e-08, + "loss": 0.9039, + "step": 22131 + }, + { + "epoch": 1.885982104814657, + "grad_norm": 12.75, + "learning_rate": 2.0165240632683825e-08, + "loss": 0.328, + "step": 22132 + }, + { + "epoch": 1.886067319982957, + "grad_norm": 16.625, + "learning_rate": 2.013519068869735e-08, + "loss": 0.7123, + "step": 22133 + }, + { + "epoch": 1.886152535151257, + "grad_norm": 16.375, + "learning_rate": 2.010516296960474e-08, + "loss": 0.6224, + "step": 22134 + }, + { + "epoch": 1.8862377503195569, + "grad_norm": 19.75, + "learning_rate": 2.0075157475948627e-08, + "loss": 0.8743, + "step": 22135 + }, + { + "epoch": 1.8863229654878568, + "grad_norm": 19.5, + "learning_rate": 2.0045174208271066e-08, + "loss": 0.7685, + "step": 22136 + }, + { + "epoch": 1.8864081806561568, + "grad_norm": 17.5, + "learning_rate": 2.0015213167114134e-08, + "loss": 0.1976, + "step": 22137 + }, + { + "epoch": 1.8864933958244567, + "grad_norm": 19.5, + "learning_rate": 1.9985274353019058e-08, + "loss": 0.9319, + "step": 22138 + }, + { + "epoch": 1.8865786109927567, + "grad_norm": 14.875, + "learning_rate": 1.9955357766527072e-08, + "loss": 0.7175, + "step": 22139 + }, + { + "epoch": 1.8866638261610567, + "grad_norm": 15.0625, + "learning_rate": 1.992546340817858e-08, + "loss": 0.6067, + "step": 22140 + }, + { + "epoch": 1.8867490413293566, + "grad_norm": 13.25, + "learning_rate": 1.9895591278513843e-08, + "loss": 0.4501, + "step": 22141 + }, + { + "epoch": 1.8868342564976566, + "grad_norm": 17.25, + "learning_rate": 1.9865741378072846e-08, + "loss": 0.9939, + "step": 22142 + }, + { + "epoch": 1.8869194716659565, + "grad_norm": 25.625, + "learning_rate": 1.9835913707394876e-08, + "loss": 0.5444, + "step": 22143 + }, + { + "epoch": 1.8870046868342565, + "grad_norm": 20.875, + "learning_rate": 1.9806108267019093e-08, + "loss": 0.7615, + "step": 22144 + }, + { + "epoch": 1.8870899020025564, + "grad_norm": 16.0, + "learning_rate": 1.9776325057483807e-08, + "loss": 0.6213, + "step": 22145 + }, + { + "epoch": 1.8871751171708564, + "grad_norm": 15.3125, + "learning_rate": 1.9746564079327486e-08, + "loss": 0.7745, + "step": 22146 + }, + { + "epoch": 1.8872603323391564, + "grad_norm": 14.4375, + "learning_rate": 1.9716825333087886e-08, + "loss": 0.7141, + "step": 22147 + }, + { + "epoch": 1.8873455475074563, + "grad_norm": 15.25, + "learning_rate": 1.968710881930236e-08, + "loss": 0.5947, + "step": 22148 + }, + { + "epoch": 1.8874307626757563, + "grad_norm": 13.375, + "learning_rate": 1.9657414538507974e-08, + "loss": 0.5496, + "step": 22149 + }, + { + "epoch": 1.8875159778440562, + "grad_norm": 13.6875, + "learning_rate": 1.9627742491241247e-08, + "loss": 0.2495, + "step": 22150 + }, + { + "epoch": 1.8876011930123562, + "grad_norm": 12.8125, + "learning_rate": 1.9598092678038556e-08, + "loss": 0.504, + "step": 22151 + }, + { + "epoch": 1.8876864081806561, + "grad_norm": 31.375, + "learning_rate": 1.956846509943558e-08, + "loss": 0.7275, + "step": 22152 + }, + { + "epoch": 1.887771623348956, + "grad_norm": 31.875, + "learning_rate": 1.9538859755967726e-08, + "loss": 0.6519, + "step": 22153 + }, + { + "epoch": 1.887856838517256, + "grad_norm": 11.6875, + "learning_rate": 1.9509276648169846e-08, + "loss": 0.2972, + "step": 22154 + }, + { + "epoch": 1.887942053685556, + "grad_norm": 19.875, + "learning_rate": 1.9479715776576785e-08, + "loss": 0.6288, + "step": 22155 + }, + { + "epoch": 1.888027268853856, + "grad_norm": 11.5, + "learning_rate": 1.9450177141722703e-08, + "loss": 0.3427, + "step": 22156 + }, + { + "epoch": 1.888112484022156, + "grad_norm": 16.5, + "learning_rate": 1.942066074414134e-08, + "loss": 0.6717, + "step": 22157 + }, + { + "epoch": 1.8881976991904559, + "grad_norm": 14.25, + "learning_rate": 1.9391166584366017e-08, + "loss": 0.3025, + "step": 22158 + }, + { + "epoch": 1.8882829143587558, + "grad_norm": 27.375, + "learning_rate": 1.936169466292992e-08, + "loss": 0.7383, + "step": 22159 + }, + { + "epoch": 1.8883681295270558, + "grad_norm": 13.0, + "learning_rate": 1.93322449803654e-08, + "loss": 0.5046, + "step": 22160 + }, + { + "epoch": 1.8884533446953558, + "grad_norm": 13.875, + "learning_rate": 1.9302817537204672e-08, + "loss": 0.4293, + "step": 22161 + }, + { + "epoch": 1.8885385598636557, + "grad_norm": 17.75, + "learning_rate": 1.9273412333979667e-08, + "loss": 0.8356, + "step": 22162 + }, + { + "epoch": 1.8886237750319557, + "grad_norm": 14.5625, + "learning_rate": 1.9244029371221635e-08, + "loss": 0.5116, + "step": 22163 + }, + { + "epoch": 1.8887089902002556, + "grad_norm": 10.125, + "learning_rate": 1.9214668649461808e-08, + "loss": 0.4074, + "step": 22164 + }, + { + "epoch": 1.8887942053685556, + "grad_norm": 12.0625, + "learning_rate": 1.9185330169230326e-08, + "loss": 0.3136, + "step": 22165 + }, + { + "epoch": 1.8888794205368555, + "grad_norm": 15.4375, + "learning_rate": 1.9156013931057733e-08, + "loss": 0.8767, + "step": 22166 + }, + { + "epoch": 1.8889646357051555, + "grad_norm": 15.8125, + "learning_rate": 1.912671993547374e-08, + "loss": 0.6942, + "step": 22167 + }, + { + "epoch": 1.8890498508734554, + "grad_norm": 12.1875, + "learning_rate": 1.9097448183007516e-08, + "loss": 0.4247, + "step": 22168 + }, + { + "epoch": 1.8891350660417554, + "grad_norm": 17.125, + "learning_rate": 1.9068198674188214e-08, + "loss": 0.6581, + "step": 22169 + }, + { + "epoch": 1.8892202812100554, + "grad_norm": 15.25, + "learning_rate": 1.9038971409544304e-08, + "loss": 0.7606, + "step": 22170 + }, + { + "epoch": 1.8893054963783553, + "grad_norm": 19.125, + "learning_rate": 1.900976638960411e-08, + "loss": 0.7506, + "step": 22171 + }, + { + "epoch": 1.8893907115466553, + "grad_norm": 16.625, + "learning_rate": 1.898058361489527e-08, + "loss": 0.7247, + "step": 22172 + }, + { + "epoch": 1.8894759267149552, + "grad_norm": 20.875, + "learning_rate": 1.8951423085945136e-08, + "loss": 0.7494, + "step": 22173 + }, + { + "epoch": 1.8895611418832552, + "grad_norm": 18.25, + "learning_rate": 1.8922284803280787e-08, + "loss": 0.4926, + "step": 22174 + }, + { + "epoch": 1.8896463570515551, + "grad_norm": 15.4375, + "learning_rate": 1.889316876742861e-08, + "loss": 0.5829, + "step": 22175 + }, + { + "epoch": 1.889731572219855, + "grad_norm": 52.25, + "learning_rate": 1.886407497891485e-08, + "loss": 0.5139, + "step": 22176 + }, + { + "epoch": 1.889816787388155, + "grad_norm": 12.3125, + "learning_rate": 1.8835003438265333e-08, + "loss": 0.5832, + "step": 22177 + }, + { + "epoch": 1.889902002556455, + "grad_norm": 18.75, + "learning_rate": 1.8805954146005334e-08, + "loss": 0.6383, + "step": 22178 + }, + { + "epoch": 1.889987217724755, + "grad_norm": 15.4375, + "learning_rate": 1.8776927102659715e-08, + "loss": 0.6285, + "step": 22179 + }, + { + "epoch": 1.890072432893055, + "grad_norm": 20.25, + "learning_rate": 1.8747922308753053e-08, + "loss": 0.9649, + "step": 22180 + }, + { + "epoch": 1.890157648061355, + "grad_norm": 12.9375, + "learning_rate": 1.871893976480965e-08, + "loss": 0.5265, + "step": 22181 + }, + { + "epoch": 1.8902428632296548, + "grad_norm": 13.5625, + "learning_rate": 1.868997947135312e-08, + "loss": 0.385, + "step": 22182 + }, + { + "epoch": 1.8903280783979548, + "grad_norm": 14.5625, + "learning_rate": 1.8661041428907068e-08, + "loss": 0.679, + "step": 22183 + }, + { + "epoch": 1.8904132935662548, + "grad_norm": 19.875, + "learning_rate": 1.8632125637993853e-08, + "loss": 1.0204, + "step": 22184 + }, + { + "epoch": 1.8904985087345547, + "grad_norm": 14.0, + "learning_rate": 1.860323209913653e-08, + "loss": 0.4018, + "step": 22185 + }, + { + "epoch": 1.8905837239028547, + "grad_norm": 17.125, + "learning_rate": 1.8574360812857183e-08, + "loss": 0.7517, + "step": 22186 + }, + { + "epoch": 1.8906689390711546, + "grad_norm": 17.0, + "learning_rate": 1.85455117796772e-08, + "loss": 0.8336, + "step": 22187 + }, + { + "epoch": 1.8907541542394546, + "grad_norm": 21.5, + "learning_rate": 1.8516685000118246e-08, + "loss": 0.6667, + "step": 22188 + }, + { + "epoch": 1.8908393694077545, + "grad_norm": 17.125, + "learning_rate": 1.8487880474701154e-08, + "loss": 0.524, + "step": 22189 + }, + { + "epoch": 1.8909245845760545, + "grad_norm": 13.5, + "learning_rate": 1.845909820394648e-08, + "loss": 0.434, + "step": 22190 + }, + { + "epoch": 1.8910097997443545, + "grad_norm": 16.375, + "learning_rate": 1.8430338188374224e-08, + "loss": 0.7357, + "step": 22191 + }, + { + "epoch": 1.8910950149126544, + "grad_norm": 15.0, + "learning_rate": 1.840160042850425e-08, + "loss": 0.7634, + "step": 22192 + }, + { + "epoch": 1.8911802300809544, + "grad_norm": 12.875, + "learning_rate": 1.8372884924855867e-08, + "loss": 0.3374, + "step": 22193 + }, + { + "epoch": 1.8912654452492543, + "grad_norm": 17.75, + "learning_rate": 1.834419167794796e-08, + "loss": 0.965, + "step": 22194 + }, + { + "epoch": 1.8913506604175543, + "grad_norm": 18.5, + "learning_rate": 1.8315520688299005e-08, + "loss": 0.3337, + "step": 22195 + }, + { + "epoch": 1.8914358755858542, + "grad_norm": 19.75, + "learning_rate": 1.828687195642706e-08, + "loss": 0.9393, + "step": 22196 + }, + { + "epoch": 1.8915210907541542, + "grad_norm": 13.875, + "learning_rate": 1.8258245482849906e-08, + "loss": 0.4949, + "step": 22197 + }, + { + "epoch": 1.8916063059224542, + "grad_norm": 28.25, + "learning_rate": 1.8229641268085048e-08, + "loss": 0.7875, + "step": 22198 + }, + { + "epoch": 1.8916915210907541, + "grad_norm": 37.0, + "learning_rate": 1.820105931264901e-08, + "loss": 0.6778, + "step": 22199 + }, + { + "epoch": 1.891776736259054, + "grad_norm": 12.4375, + "learning_rate": 1.8172499617058608e-08, + "loss": 0.428, + "step": 22200 + }, + { + "epoch": 1.891861951427354, + "grad_norm": 14.8125, + "learning_rate": 1.8143962181829815e-08, + "loss": 0.6965, + "step": 22201 + }, + { + "epoch": 1.891947166595654, + "grad_norm": 16.625, + "learning_rate": 1.8115447007478332e-08, + "loss": 0.5398, + "step": 22202 + }, + { + "epoch": 1.892032381763954, + "grad_norm": 14.8125, + "learning_rate": 1.8086954094519576e-08, + "loss": 0.5666, + "step": 22203 + }, + { + "epoch": 1.892117596932254, + "grad_norm": 19.875, + "learning_rate": 1.805848344346828e-08, + "loss": 0.3861, + "step": 22204 + }, + { + "epoch": 1.8922028121005539, + "grad_norm": 12.75, + "learning_rate": 1.8030035054839022e-08, + "loss": 0.4156, + "step": 22205 + }, + { + "epoch": 1.8922880272688538, + "grad_norm": 12.3125, + "learning_rate": 1.800160892914571e-08, + "loss": 0.2874, + "step": 22206 + }, + { + "epoch": 1.8923732424371538, + "grad_norm": 20.75, + "learning_rate": 1.7973205066902365e-08, + "loss": 1.0864, + "step": 22207 + }, + { + "epoch": 1.8924584576054537, + "grad_norm": 12.125, + "learning_rate": 1.7944823468622195e-08, + "loss": 0.5776, + "step": 22208 + }, + { + "epoch": 1.8925436727737537, + "grad_norm": 13.5625, + "learning_rate": 1.7916464134817846e-08, + "loss": 0.5831, + "step": 22209 + }, + { + "epoch": 1.8926288879420536, + "grad_norm": 15.5625, + "learning_rate": 1.788812706600196e-08, + "loss": 0.5283, + "step": 22210 + }, + { + "epoch": 1.8927141031103536, + "grad_norm": 14.375, + "learning_rate": 1.7859812262686628e-08, + "loss": 0.7142, + "step": 22211 + }, + { + "epoch": 1.8927993182786536, + "grad_norm": 15.875, + "learning_rate": 1.7831519725383385e-08, + "loss": 0.5516, + "step": 22212 + }, + { + "epoch": 1.8928845334469535, + "grad_norm": 14.0625, + "learning_rate": 1.7803249454603765e-08, + "loss": 0.4874, + "step": 22213 + }, + { + "epoch": 1.8929697486152535, + "grad_norm": 20.875, + "learning_rate": 1.7775001450858474e-08, + "loss": 0.5609, + "step": 22214 + }, + { + "epoch": 1.8930549637835534, + "grad_norm": 13.9375, + "learning_rate": 1.7746775714657938e-08, + "loss": 0.4931, + "step": 22215 + }, + { + "epoch": 1.8931401789518534, + "grad_norm": 15.9375, + "learning_rate": 1.771857224651244e-08, + "loss": 0.6245, + "step": 22216 + }, + { + "epoch": 1.8932253941201533, + "grad_norm": 16.25, + "learning_rate": 1.7690391046931294e-08, + "loss": 0.591, + "step": 22217 + }, + { + "epoch": 1.8933106092884533, + "grad_norm": 13.75, + "learning_rate": 1.7662232116424094e-08, + "loss": 0.7353, + "step": 22218 + }, + { + "epoch": 1.8933958244567533, + "grad_norm": 15.5, + "learning_rate": 1.7634095455499463e-08, + "loss": 0.4752, + "step": 22219 + }, + { + "epoch": 1.8934810396250532, + "grad_norm": 12.6875, + "learning_rate": 1.7605981064666154e-08, + "loss": 0.3946, + "step": 22220 + }, + { + "epoch": 1.8935662547933532, + "grad_norm": 10.875, + "learning_rate": 1.7577888944431824e-08, + "loss": 0.3866, + "step": 22221 + }, + { + "epoch": 1.8936514699616531, + "grad_norm": 13.1875, + "learning_rate": 1.75498190953044e-08, + "loss": 0.414, + "step": 22222 + }, + { + "epoch": 1.893736685129953, + "grad_norm": 14.3125, + "learning_rate": 1.7521771517791247e-08, + "loss": 0.6509, + "step": 22223 + }, + { + "epoch": 1.893821900298253, + "grad_norm": 15.3125, + "learning_rate": 1.749374621239891e-08, + "loss": 0.8166, + "step": 22224 + }, + { + "epoch": 1.893907115466553, + "grad_norm": 22.125, + "learning_rate": 1.7465743179633926e-08, + "loss": 0.5885, + "step": 22225 + }, + { + "epoch": 1.893992330634853, + "grad_norm": 16.375, + "learning_rate": 1.743776242000242e-08, + "loss": 0.4569, + "step": 22226 + }, + { + "epoch": 1.894077545803153, + "grad_norm": 13.5, + "learning_rate": 1.7409803934009816e-08, + "loss": 0.6827, + "step": 22227 + }, + { + "epoch": 1.8941627609714529, + "grad_norm": 14.25, + "learning_rate": 1.738186772216183e-08, + "loss": 0.5041, + "step": 22228 + }, + { + "epoch": 1.894247976139753, + "grad_norm": 15.8125, + "learning_rate": 1.735395378496277e-08, + "loss": 0.5296, + "step": 22229 + }, + { + "epoch": 1.894333191308053, + "grad_norm": 18.875, + "learning_rate": 1.7326062122917238e-08, + "loss": 0.731, + "step": 22230 + }, + { + "epoch": 1.894418406476353, + "grad_norm": 15.75, + "learning_rate": 1.729819273652941e-08, + "loss": 0.5395, + "step": 22231 + }, + { + "epoch": 1.894503621644653, + "grad_norm": 18.125, + "learning_rate": 1.727034562630292e-08, + "loss": 0.4331, + "step": 22232 + }, + { + "epoch": 1.8945888368129529, + "grad_norm": 16.875, + "learning_rate": 1.724252079274083e-08, + "loss": 0.5287, + "step": 22233 + }, + { + "epoch": 1.8946740519812528, + "grad_norm": 26.0, + "learning_rate": 1.7214718236345934e-08, + "loss": 0.813, + "step": 22234 + }, + { + "epoch": 1.8947592671495528, + "grad_norm": 16.25, + "learning_rate": 1.718693795762075e-08, + "loss": 0.2972, + "step": 22235 + }, + { + "epoch": 1.8948444823178527, + "grad_norm": 18.25, + "learning_rate": 1.7159179957067228e-08, + "loss": 0.5384, + "step": 22236 + }, + { + "epoch": 1.8949296974861527, + "grad_norm": 13.0625, + "learning_rate": 1.71314442351872e-08, + "loss": 0.434, + "step": 22237 + }, + { + "epoch": 1.8950149126544527, + "grad_norm": 15.1875, + "learning_rate": 1.710373079248165e-08, + "loss": 0.4771, + "step": 22238 + }, + { + "epoch": 1.8951001278227526, + "grad_norm": 19.125, + "learning_rate": 1.7076039629451424e-08, + "loss": 0.9232, + "step": 22239 + }, + { + "epoch": 1.8951853429910526, + "grad_norm": 19.875, + "learning_rate": 1.7048370746597098e-08, + "loss": 0.8364, + "step": 22240 + }, + { + "epoch": 1.8952705581593525, + "grad_norm": 14.0625, + "learning_rate": 1.702072414441841e-08, + "loss": 0.4483, + "step": 22241 + }, + { + "epoch": 1.8953557733276525, + "grad_norm": 15.1875, + "learning_rate": 1.6993099823415098e-08, + "loss": 0.5277, + "step": 22242 + }, + { + "epoch": 1.8954409884959524, + "grad_norm": 17.875, + "learning_rate": 1.696549778408635e-08, + "loss": 0.585, + "step": 22243 + }, + { + "epoch": 1.8955262036642524, + "grad_norm": 15.375, + "learning_rate": 1.6937918026931065e-08, + "loss": 0.7944, + "step": 22244 + }, + { + "epoch": 1.8956114188325524, + "grad_norm": 14.25, + "learning_rate": 1.6910360552447464e-08, + "loss": 0.5009, + "step": 22245 + }, + { + "epoch": 1.8956966340008523, + "grad_norm": 9.75, + "learning_rate": 1.6882825361133616e-08, + "loss": 0.2388, + "step": 22246 + }, + { + "epoch": 1.8957818491691523, + "grad_norm": 14.625, + "learning_rate": 1.6855312453487177e-08, + "loss": 0.701, + "step": 22247 + }, + { + "epoch": 1.8958670643374522, + "grad_norm": 13.8125, + "learning_rate": 1.6827821830005252e-08, + "loss": 0.4177, + "step": 22248 + }, + { + "epoch": 1.8959522795057522, + "grad_norm": 12.0625, + "learning_rate": 1.6800353491184528e-08, + "loss": 0.4362, + "step": 22249 + }, + { + "epoch": 1.8960374946740521, + "grad_norm": 21.5, + "learning_rate": 1.677290743752169e-08, + "loss": 1.029, + "step": 22250 + }, + { + "epoch": 1.896122709842352, + "grad_norm": 12.625, + "learning_rate": 1.6745483669512312e-08, + "loss": 0.3774, + "step": 22251 + }, + { + "epoch": 1.896207925010652, + "grad_norm": 14.125, + "learning_rate": 1.6718082187652252e-08, + "loss": 0.3903, + "step": 22252 + }, + { + "epoch": 1.896293140178952, + "grad_norm": 14.1875, + "learning_rate": 1.6690702992436662e-08, + "loss": 0.5067, + "step": 22253 + }, + { + "epoch": 1.896378355347252, + "grad_norm": 20.875, + "learning_rate": 1.6663346084360153e-08, + "loss": 0.6823, + "step": 22254 + }, + { + "epoch": 1.896463570515552, + "grad_norm": 37.5, + "learning_rate": 1.663601146391719e-08, + "loss": 1.1558, + "step": 22255 + }, + { + "epoch": 1.8965487856838519, + "grad_norm": 30.375, + "learning_rate": 1.6608699131601818e-08, + "loss": 0.81, + "step": 22256 + }, + { + "epoch": 1.8966340008521518, + "grad_norm": 25.125, + "learning_rate": 1.6581409087907534e-08, + "loss": 0.609, + "step": 22257 + }, + { + "epoch": 1.8967192160204518, + "grad_norm": 15.75, + "learning_rate": 1.6554141333327554e-08, + "loss": 0.733, + "step": 22258 + }, + { + "epoch": 1.8968044311887517, + "grad_norm": 19.125, + "learning_rate": 1.6526895868354404e-08, + "loss": 0.3938, + "step": 22259 + }, + { + "epoch": 1.8968896463570517, + "grad_norm": 14.9375, + "learning_rate": 1.649967269348074e-08, + "loss": 0.5497, + "step": 22260 + }, + { + "epoch": 1.8969748615253517, + "grad_norm": 17.75, + "learning_rate": 1.6472471809198253e-08, + "loss": 0.6176, + "step": 22261 + }, + { + "epoch": 1.8970600766936516, + "grad_norm": 14.125, + "learning_rate": 1.6445293215998636e-08, + "loss": 0.7186, + "step": 22262 + }, + { + "epoch": 1.8971452918619516, + "grad_norm": 19.125, + "learning_rate": 1.641813691437302e-08, + "loss": 0.6694, + "step": 22263 + }, + { + "epoch": 1.8972305070302515, + "grad_norm": 11.5625, + "learning_rate": 1.639100290481213e-08, + "loss": 0.436, + "step": 22264 + }, + { + "epoch": 1.8973157221985515, + "grad_norm": 19.25, + "learning_rate": 1.6363891187806403e-08, + "loss": 0.7687, + "step": 22265 + }, + { + "epoch": 1.8974009373668514, + "grad_norm": 20.875, + "learning_rate": 1.6336801763845582e-08, + "loss": 0.9266, + "step": 22266 + }, + { + "epoch": 1.8974861525351514, + "grad_norm": 18.125, + "learning_rate": 1.6309734633419284e-08, + "loss": 0.4169, + "step": 22267 + }, + { + "epoch": 1.8975713677034514, + "grad_norm": 13.625, + "learning_rate": 1.6282689797016692e-08, + "loss": 0.5117, + "step": 22268 + }, + { + "epoch": 1.8976565828717513, + "grad_norm": 9.75, + "learning_rate": 1.625566725512645e-08, + "loss": 0.2369, + "step": 22269 + }, + { + "epoch": 1.8977417980400513, + "grad_norm": 16.625, + "learning_rate": 1.6228667008237054e-08, + "loss": 0.5543, + "step": 22270 + }, + { + "epoch": 1.8978270132083512, + "grad_norm": 13.25, + "learning_rate": 1.6201689056836167e-08, + "loss": 0.713, + "step": 22271 + }, + { + "epoch": 1.8979122283766512, + "grad_norm": 18.375, + "learning_rate": 1.6174733401411457e-08, + "loss": 0.7061, + "step": 22272 + }, + { + "epoch": 1.8979974435449511, + "grad_norm": 16.625, + "learning_rate": 1.6147800042450035e-08, + "loss": 0.5003, + "step": 22273 + }, + { + "epoch": 1.898082658713251, + "grad_norm": 12.4375, + "learning_rate": 1.612088898043859e-08, + "loss": 0.373, + "step": 22274 + }, + { + "epoch": 1.898167873881551, + "grad_norm": 14.625, + "learning_rate": 1.609400021586355e-08, + "loss": 0.333, + "step": 22275 + }, + { + "epoch": 1.898253089049851, + "grad_norm": 11.75, + "learning_rate": 1.6067133749210486e-08, + "loss": 0.3897, + "step": 22276 + }, + { + "epoch": 1.898338304218151, + "grad_norm": 77.0, + "learning_rate": 1.6040289580965264e-08, + "loss": 0.6086, + "step": 22277 + }, + { + "epoch": 1.898423519386451, + "grad_norm": 15.5, + "learning_rate": 1.6013467711612773e-08, + "loss": 0.4753, + "step": 22278 + }, + { + "epoch": 1.8985087345547509, + "grad_norm": 14.75, + "learning_rate": 1.5986668141638046e-08, + "loss": 0.5092, + "step": 22279 + }, + { + "epoch": 1.8985939497230508, + "grad_norm": 15.0625, + "learning_rate": 1.595989087152486e-08, + "loss": 0.512, + "step": 22280 + }, + { + "epoch": 1.8986791648913508, + "grad_norm": 15.625, + "learning_rate": 1.5933135901757547e-08, + "loss": 0.5688, + "step": 22281 + }, + { + "epoch": 1.8987643800596508, + "grad_norm": 15.0625, + "learning_rate": 1.5906403232819195e-08, + "loss": 0.5521, + "step": 22282 + }, + { + "epoch": 1.8988495952279507, + "grad_norm": 30.875, + "learning_rate": 1.587969286519331e-08, + "loss": 0.7264, + "step": 22283 + }, + { + "epoch": 1.8989348103962507, + "grad_norm": 13.75, + "learning_rate": 1.5853004799362275e-08, + "loss": 0.3812, + "step": 22284 + }, + { + "epoch": 1.8990200255645506, + "grad_norm": 20.75, + "learning_rate": 1.582633903580849e-08, + "loss": 0.5939, + "step": 22285 + }, + { + "epoch": 1.8991052407328506, + "grad_norm": 13.375, + "learning_rate": 1.579969557501393e-08, + "loss": 0.3817, + "step": 22286 + }, + { + "epoch": 1.8991904559011505, + "grad_norm": 16.25, + "learning_rate": 1.5773074417459873e-08, + "loss": 0.6632, + "step": 22287 + }, + { + "epoch": 1.8992756710694505, + "grad_norm": 18.75, + "learning_rate": 1.5746475563627463e-08, + "loss": 1.0349, + "step": 22288 + }, + { + "epoch": 1.8993608862377505, + "grad_norm": 15.3125, + "learning_rate": 1.5719899013997427e-08, + "loss": 0.5587, + "step": 22289 + }, + { + "epoch": 1.8994461014060504, + "grad_norm": 14.125, + "learning_rate": 1.5693344769049794e-08, + "loss": 0.6674, + "step": 22290 + }, + { + "epoch": 1.8995313165743504, + "grad_norm": 13.375, + "learning_rate": 1.5666812829264744e-08, + "loss": 0.656, + "step": 22291 + }, + { + "epoch": 1.8996165317426503, + "grad_norm": 22.5, + "learning_rate": 1.564030319512175e-08, + "loss": 0.9626, + "step": 22292 + }, + { + "epoch": 1.8997017469109503, + "grad_norm": 17.125, + "learning_rate": 1.5613815867099453e-08, + "loss": 0.4243, + "step": 22293 + }, + { + "epoch": 1.8997869620792502, + "grad_norm": 20.875, + "learning_rate": 1.558735084567692e-08, + "loss": 0.6271, + "step": 22294 + }, + { + "epoch": 1.8998721772475502, + "grad_norm": 14.5625, + "learning_rate": 1.5560908131332237e-08, + "loss": 0.7042, + "step": 22295 + }, + { + "epoch": 1.8999573924158502, + "grad_norm": 16.625, + "learning_rate": 1.5534487724543356e-08, + "loss": 0.5418, + "step": 22296 + }, + { + "epoch": 1.9000426075841501, + "grad_norm": 20.25, + "learning_rate": 1.5508089625787535e-08, + "loss": 0.576, + "step": 22297 + }, + { + "epoch": 1.90012782275245, + "grad_norm": 20.5, + "learning_rate": 1.5481713835541894e-08, + "loss": 0.883, + "step": 22298 + }, + { + "epoch": 1.90021303792075, + "grad_norm": 17.0, + "learning_rate": 1.5455360354283135e-08, + "loss": 0.6079, + "step": 22299 + }, + { + "epoch": 1.90029825308905, + "grad_norm": 16.375, + "learning_rate": 1.5429029182487397e-08, + "loss": 0.6916, + "step": 22300 + }, + { + "epoch": 1.90038346825735, + "grad_norm": 14.9375, + "learning_rate": 1.540272032063056e-08, + "loss": 0.4265, + "step": 22301 + }, + { + "epoch": 1.90046868342565, + "grad_norm": 25.5, + "learning_rate": 1.537643376918807e-08, + "loss": 0.9618, + "step": 22302 + }, + { + "epoch": 1.9005538985939499, + "grad_norm": 11.1875, + "learning_rate": 1.5350169528634966e-08, + "loss": 0.4291, + "step": 22303 + }, + { + "epoch": 1.9006391137622498, + "grad_norm": 16.75, + "learning_rate": 1.532392759944587e-08, + "loss": 0.459, + "step": 22304 + }, + { + "epoch": 1.9007243289305498, + "grad_norm": 13.5, + "learning_rate": 1.5297707982094983e-08, + "loss": 0.3995, + "step": 22305 + }, + { + "epoch": 1.9008095440988497, + "grad_norm": 17.375, + "learning_rate": 1.5271510677055956e-08, + "loss": 0.9546, + "step": 22306 + }, + { + "epoch": 1.9008947592671497, + "grad_norm": 14.1875, + "learning_rate": 1.5245335684802437e-08, + "loss": 0.5611, + "step": 22307 + }, + { + "epoch": 1.9009799744354496, + "grad_norm": 11.5625, + "learning_rate": 1.5219183005807242e-08, + "loss": 0.3231, + "step": 22308 + }, + { + "epoch": 1.9010651896037496, + "grad_norm": 17.375, + "learning_rate": 1.519305264054319e-08, + "loss": 0.7909, + "step": 22309 + }, + { + "epoch": 1.9011504047720496, + "grad_norm": 17.375, + "learning_rate": 1.5166944589482258e-08, + "loss": 0.9723, + "step": 22310 + }, + { + "epoch": 1.9012356199403495, + "grad_norm": 20.375, + "learning_rate": 1.5140858853096437e-08, + "loss": 0.5954, + "step": 22311 + }, + { + "epoch": 1.9013208351086495, + "grad_norm": 19.375, + "learning_rate": 1.511479543185701e-08, + "loss": 0.8528, + "step": 22312 + }, + { + "epoch": 1.9014060502769494, + "grad_norm": 27.125, + "learning_rate": 1.5088754326235274e-08, + "loss": 0.5171, + "step": 22313 + }, + { + "epoch": 1.9014912654452494, + "grad_norm": 24.875, + "learning_rate": 1.5062735536701263e-08, + "loss": 0.484, + "step": 22314 + }, + { + "epoch": 1.9015764806135493, + "grad_norm": 20.25, + "learning_rate": 1.5036739063725436e-08, + "loss": 0.9198, + "step": 22315 + }, + { + "epoch": 1.9016616957818493, + "grad_norm": 17.0, + "learning_rate": 1.501076490777756e-08, + "loss": 0.6133, + "step": 22316 + }, + { + "epoch": 1.9017469109501492, + "grad_norm": 10.1875, + "learning_rate": 1.4984813069327253e-08, + "loss": 0.2487, + "step": 22317 + }, + { + "epoch": 1.9018321261184492, + "grad_norm": 14.1875, + "learning_rate": 1.4958883548843033e-08, + "loss": 0.4993, + "step": 22318 + }, + { + "epoch": 1.9019173412867492, + "grad_norm": 18.875, + "learning_rate": 1.4932976346793832e-08, + "loss": 0.5389, + "step": 22319 + }, + { + "epoch": 1.9020025564550491, + "grad_norm": 19.5, + "learning_rate": 1.4907091463647745e-08, + "loss": 0.781, + "step": 22320 + }, + { + "epoch": 1.902087771623349, + "grad_norm": 10.625, + "learning_rate": 1.4881228899872452e-08, + "loss": 0.2053, + "step": 22321 + }, + { + "epoch": 1.902172986791649, + "grad_norm": 11.4375, + "learning_rate": 1.48553886559355e-08, + "loss": 0.3877, + "step": 22322 + }, + { + "epoch": 1.902258201959949, + "grad_norm": 16.5, + "learning_rate": 1.4829570732303456e-08, + "loss": 0.743, + "step": 22323 + }, + { + "epoch": 1.902343417128249, + "grad_norm": 23.0, + "learning_rate": 1.480377512944331e-08, + "loss": 0.9175, + "step": 22324 + }, + { + "epoch": 1.902428632296549, + "grad_norm": 14.1875, + "learning_rate": 1.4778001847820938e-08, + "loss": 0.5835, + "step": 22325 + }, + { + "epoch": 1.9025138474648489, + "grad_norm": 23.0, + "learning_rate": 1.475225088790222e-08, + "loss": 1.0265, + "step": 22326 + }, + { + "epoch": 1.9025990626331488, + "grad_norm": 18.875, + "learning_rate": 1.4726522250152475e-08, + "loss": 0.5368, + "step": 22327 + }, + { + "epoch": 1.9026842778014488, + "grad_norm": 17.25, + "learning_rate": 1.4700815935036611e-08, + "loss": 0.717, + "step": 22328 + }, + { + "epoch": 1.9027694929697487, + "grad_norm": 13.8125, + "learning_rate": 1.4675131943019116e-08, + "loss": 0.4175, + "step": 22329 + }, + { + "epoch": 1.9028547081380487, + "grad_norm": 19.875, + "learning_rate": 1.4649470274564481e-08, + "loss": 0.7567, + "step": 22330 + }, + { + "epoch": 1.9029399233063486, + "grad_norm": 21.125, + "learning_rate": 1.4623830930135946e-08, + "loss": 0.7178, + "step": 22331 + }, + { + "epoch": 1.9030251384746486, + "grad_norm": 13.25, + "learning_rate": 1.4598213910197029e-08, + "loss": 0.6074, + "step": 22332 + }, + { + "epoch": 1.9031103536429486, + "grad_norm": 16.375, + "learning_rate": 1.4572619215210692e-08, + "loss": 0.7601, + "step": 22333 + }, + { + "epoch": 1.9031955688112485, + "grad_norm": 13.1875, + "learning_rate": 1.4547046845639485e-08, + "loss": 0.3952, + "step": 22334 + }, + { + "epoch": 1.9032807839795485, + "grad_norm": 19.5, + "learning_rate": 1.4521496801945533e-08, + "loss": 0.9361, + "step": 22335 + }, + { + "epoch": 1.9033659991478484, + "grad_norm": 14.5625, + "learning_rate": 1.4495969084590417e-08, + "loss": 0.7269, + "step": 22336 + }, + { + "epoch": 1.9034512143161484, + "grad_norm": 17.0, + "learning_rate": 1.4470463694035431e-08, + "loss": 0.6829, + "step": 22337 + }, + { + "epoch": 1.9035364294844483, + "grad_norm": 17.375, + "learning_rate": 1.4444980630741873e-08, + "loss": 0.5051, + "step": 22338 + }, + { + "epoch": 1.9036216446527483, + "grad_norm": 12.3125, + "learning_rate": 1.4419519895169792e-08, + "loss": 0.3317, + "step": 22339 + }, + { + "epoch": 1.9037068598210483, + "grad_norm": 12.6875, + "learning_rate": 1.4394081487779516e-08, + "loss": 0.5243, + "step": 22340 + }, + { + "epoch": 1.9037920749893482, + "grad_norm": 16.5, + "learning_rate": 1.4368665409030536e-08, + "loss": 0.5212, + "step": 22341 + }, + { + "epoch": 1.9038772901576482, + "grad_norm": 20.125, + "learning_rate": 1.4343271659382484e-08, + "loss": 1.0732, + "step": 22342 + }, + { + "epoch": 1.9039625053259481, + "grad_norm": 11.5625, + "learning_rate": 1.4317900239294025e-08, + "loss": 0.3886, + "step": 22343 + }, + { + "epoch": 1.904047720494248, + "grad_norm": 12.25, + "learning_rate": 1.4292551149223538e-08, + "loss": 0.4434, + "step": 22344 + }, + { + "epoch": 1.904132935662548, + "grad_norm": 13.6875, + "learning_rate": 1.4267224389629408e-08, + "loss": 0.4923, + "step": 22345 + }, + { + "epoch": 1.904218150830848, + "grad_norm": 17.0, + "learning_rate": 1.4241919960969185e-08, + "loss": 0.6493, + "step": 22346 + }, + { + "epoch": 1.904303365999148, + "grad_norm": 10.9375, + "learning_rate": 1.4216637863700006e-08, + "loss": 0.3618, + "step": 22347 + }, + { + "epoch": 1.904388581167448, + "grad_norm": 14.75, + "learning_rate": 1.4191378098278863e-08, + "loss": 0.2869, + "step": 22348 + }, + { + "epoch": 1.9044737963357479, + "grad_norm": 15.375, + "learning_rate": 1.4166140665162198e-08, + "loss": 0.5116, + "step": 22349 + }, + { + "epoch": 1.9045590115040478, + "grad_norm": 20.0, + "learning_rate": 1.4140925564806173e-08, + "loss": 0.8809, + "step": 22350 + }, + { + "epoch": 1.9046442266723478, + "grad_norm": 14.4375, + "learning_rate": 1.4115732797666399e-08, + "loss": 0.525, + "step": 22351 + }, + { + "epoch": 1.9047294418406477, + "grad_norm": 12.75, + "learning_rate": 1.4090562364198063e-08, + "loss": 0.5143, + "step": 22352 + }, + { + "epoch": 1.9048146570089477, + "grad_norm": 11.8125, + "learning_rate": 1.406541426485608e-08, + "loss": 0.3479, + "step": 22353 + }, + { + "epoch": 1.9048998721772477, + "grad_norm": 27.125, + "learning_rate": 1.404028850009495e-08, + "loss": 0.6341, + "step": 22354 + }, + { + "epoch": 1.9049850873455476, + "grad_norm": 12.6875, + "learning_rate": 1.4015185070368476e-08, + "loss": 0.442, + "step": 22355 + }, + { + "epoch": 1.9050703025138476, + "grad_norm": 16.75, + "learning_rate": 1.3990103976130737e-08, + "loss": 0.649, + "step": 22356 + }, + { + "epoch": 1.9051555176821475, + "grad_norm": 15.0625, + "learning_rate": 1.3965045217834427e-08, + "loss": 0.528, + "step": 22357 + }, + { + "epoch": 1.9052407328504475, + "grad_norm": 15.1875, + "learning_rate": 1.3940008795932935e-08, + "loss": 0.5905, + "step": 22358 + }, + { + "epoch": 1.9053259480187474, + "grad_norm": 18.5, + "learning_rate": 1.391499471087826e-08, + "loss": 0.5399, + "step": 22359 + }, + { + "epoch": 1.9054111631870474, + "grad_norm": 14.5625, + "learning_rate": 1.3890002963122679e-08, + "loss": 0.5119, + "step": 22360 + }, + { + "epoch": 1.9054963783553474, + "grad_norm": 11.875, + "learning_rate": 1.3865033553117773e-08, + "loss": 0.4171, + "step": 22361 + }, + { + "epoch": 1.9055815935236473, + "grad_norm": 16.375, + "learning_rate": 1.3840086481314852e-08, + "loss": 0.9304, + "step": 22362 + }, + { + "epoch": 1.9056668086919473, + "grad_norm": 17.375, + "learning_rate": 1.3815161748164385e-08, + "loss": 0.7465, + "step": 22363 + }, + { + "epoch": 1.9057520238602472, + "grad_norm": 13.375, + "learning_rate": 1.3790259354117264e-08, + "loss": 0.4436, + "step": 22364 + }, + { + "epoch": 1.9058372390285472, + "grad_norm": 22.0, + "learning_rate": 1.3765379299623128e-08, + "loss": 0.5752, + "step": 22365 + }, + { + "epoch": 1.9059224541968471, + "grad_norm": 15.5625, + "learning_rate": 1.3740521585131617e-08, + "loss": 0.7235, + "step": 22366 + }, + { + "epoch": 1.906007669365147, + "grad_norm": 13.5625, + "learning_rate": 1.3715686211092233e-08, + "loss": 0.4148, + "step": 22367 + }, + { + "epoch": 1.906092884533447, + "grad_norm": 17.125, + "learning_rate": 1.3690873177953507e-08, + "loss": 0.854, + "step": 22368 + }, + { + "epoch": 1.906178099701747, + "grad_norm": 17.0, + "learning_rate": 1.3666082486163829e-08, + "loss": 0.6283, + "step": 22369 + }, + { + "epoch": 1.906263314870047, + "grad_norm": 14.1875, + "learning_rate": 1.3641314136171452e-08, + "loss": 0.3706, + "step": 22370 + }, + { + "epoch": 1.906348530038347, + "grad_norm": 21.125, + "learning_rate": 1.3616568128423658e-08, + "loss": 0.7561, + "step": 22371 + }, + { + "epoch": 1.9064337452066469, + "grad_norm": 15.8125, + "learning_rate": 1.3591844463367726e-08, + "loss": 0.7676, + "step": 22372 + }, + { + "epoch": 1.9065189603749468, + "grad_norm": 15.625, + "learning_rate": 1.3567143141450523e-08, + "loss": 0.6314, + "step": 22373 + }, + { + "epoch": 1.9066041755432468, + "grad_norm": 15.9375, + "learning_rate": 1.3542464163118358e-08, + "loss": 0.4293, + "step": 22374 + }, + { + "epoch": 1.9066893907115467, + "grad_norm": 15.1875, + "learning_rate": 1.3517807528817123e-08, + "loss": 0.7607, + "step": 22375 + }, + { + "epoch": 1.9067746058798467, + "grad_norm": 16.875, + "learning_rate": 1.3493173238992574e-08, + "loss": 0.853, + "step": 22376 + }, + { + "epoch": 1.9068598210481467, + "grad_norm": 16.5, + "learning_rate": 1.346856129408991e-08, + "loss": 0.7179, + "step": 22377 + }, + { + "epoch": 1.9069450362164466, + "grad_norm": 15.3125, + "learning_rate": 1.3443971694553636e-08, + "loss": 0.356, + "step": 22378 + }, + { + "epoch": 1.9070302513847466, + "grad_norm": 18.0, + "learning_rate": 1.341940444082826e-08, + "loss": 0.9654, + "step": 22379 + }, + { + "epoch": 1.9071154665530465, + "grad_norm": 12.0625, + "learning_rate": 1.3394859533357729e-08, + "loss": 0.5506, + "step": 22380 + }, + { + "epoch": 1.9072006817213465, + "grad_norm": 16.875, + "learning_rate": 1.337033697258558e-08, + "loss": 0.5317, + "step": 22381 + }, + { + "epoch": 1.9072858968896464, + "grad_norm": 18.75, + "learning_rate": 1.3345836758955067e-08, + "loss": 0.7054, + "step": 22382 + }, + { + "epoch": 1.9073711120579464, + "grad_norm": 12.4375, + "learning_rate": 1.3321358892908615e-08, + "loss": 0.2618, + "step": 22383 + }, + { + "epoch": 1.9074563272262464, + "grad_norm": 15.0, + "learning_rate": 1.3296903374888925e-08, + "loss": 0.5632, + "step": 22384 + }, + { + "epoch": 1.9075415423945463, + "grad_norm": 15.625, + "learning_rate": 1.3272470205337867e-08, + "loss": 0.6366, + "step": 22385 + }, + { + "epoch": 1.9076267575628463, + "grad_norm": 14.4375, + "learning_rate": 1.3248059384696894e-08, + "loss": 0.5884, + "step": 22386 + }, + { + "epoch": 1.9077119727311462, + "grad_norm": 15.8125, + "learning_rate": 1.3223670913407038e-08, + "loss": 0.4866, + "step": 22387 + }, + { + "epoch": 1.9077971878994462, + "grad_norm": 17.25, + "learning_rate": 1.3199304791909063e-08, + "loss": 0.598, + "step": 22388 + }, + { + "epoch": 1.9078824030677461, + "grad_norm": 14.9375, + "learning_rate": 1.3174961020643445e-08, + "loss": 0.7598, + "step": 22389 + }, + { + "epoch": 1.907967618236046, + "grad_norm": 17.75, + "learning_rate": 1.3150639600050113e-08, + "loss": 0.7255, + "step": 22390 + }, + { + "epoch": 1.908052833404346, + "grad_norm": 22.75, + "learning_rate": 1.31263405305683e-08, + "loss": 0.5201, + "step": 22391 + }, + { + "epoch": 1.908138048572646, + "grad_norm": 19.5, + "learning_rate": 1.3102063812637233e-08, + "loss": 0.7663, + "step": 22392 + }, + { + "epoch": 1.908223263740946, + "grad_norm": 12.75, + "learning_rate": 1.3077809446695871e-08, + "loss": 0.4035, + "step": 22393 + }, + { + "epoch": 1.908308478909246, + "grad_norm": 16.125, + "learning_rate": 1.3053577433182196e-08, + "loss": 0.5902, + "step": 22394 + }, + { + "epoch": 1.9083936940775459, + "grad_norm": 20.25, + "learning_rate": 1.302936777253433e-08, + "loss": 0.8054, + "step": 22395 + }, + { + "epoch": 1.9084789092458458, + "grad_norm": 17.5, + "learning_rate": 1.3005180465189559e-08, + "loss": 0.8141, + "step": 22396 + }, + { + "epoch": 1.9085641244141458, + "grad_norm": 20.25, + "learning_rate": 1.2981015511585037e-08, + "loss": 0.8522, + "step": 22397 + }, + { + "epoch": 1.9086493395824458, + "grad_norm": 23.625, + "learning_rate": 1.2956872912157637e-08, + "loss": 0.6387, + "step": 22398 + }, + { + "epoch": 1.9087345547507457, + "grad_norm": 14.9375, + "learning_rate": 1.2932752667343256e-08, + "loss": 0.6417, + "step": 22399 + }, + { + "epoch": 1.9088197699190457, + "grad_norm": 17.5, + "learning_rate": 1.2908654777578077e-08, + "loss": 0.6902, + "step": 22400 + }, + { + "epoch": 1.9089049850873456, + "grad_norm": 16.125, + "learning_rate": 1.2884579243297585e-08, + "loss": 0.7657, + "step": 22401 + }, + { + "epoch": 1.9089902002556456, + "grad_norm": 15.25, + "learning_rate": 1.2860526064936707e-08, + "loss": 0.324, + "step": 22402 + }, + { + "epoch": 1.9090754154239455, + "grad_norm": 15.0625, + "learning_rate": 1.2836495242930236e-08, + "loss": 0.5822, + "step": 22403 + }, + { + "epoch": 1.9091606305922455, + "grad_norm": 16.375, + "learning_rate": 1.2812486777712408e-08, + "loss": 0.2957, + "step": 22404 + }, + { + "epoch": 1.9092458457605455, + "grad_norm": 18.125, + "learning_rate": 1.2788500669716902e-08, + "loss": 0.6466, + "step": 22405 + }, + { + "epoch": 1.9093310609288454, + "grad_norm": 16.375, + "learning_rate": 1.2764536919377263e-08, + "loss": 0.699, + "step": 22406 + }, + { + "epoch": 1.9094162760971454, + "grad_norm": 18.875, + "learning_rate": 1.2740595527126615e-08, + "loss": 0.7627, + "step": 22407 + }, + { + "epoch": 1.9095014912654453, + "grad_norm": 13.8125, + "learning_rate": 1.2716676493397668e-08, + "loss": 0.4097, + "step": 22408 + }, + { + "epoch": 1.9095867064337453, + "grad_norm": 38.5, + "learning_rate": 1.2692779818622435e-08, + "loss": 1.0276, + "step": 22409 + }, + { + "epoch": 1.9096719216020452, + "grad_norm": 15.8125, + "learning_rate": 1.2668905503232937e-08, + "loss": 0.7804, + "step": 22410 + }, + { + "epoch": 1.9097571367703452, + "grad_norm": 13.4375, + "learning_rate": 1.2645053547660629e-08, + "loss": 0.4428, + "step": 22411 + }, + { + "epoch": 1.9098423519386452, + "grad_norm": 14.3125, + "learning_rate": 1.262122395233642e-08, + "loss": 0.5864, + "step": 22412 + }, + { + "epoch": 1.9099275671069451, + "grad_norm": 14.0, + "learning_rate": 1.2597416717690797e-08, + "loss": 0.47, + "step": 22413 + }, + { + "epoch": 1.910012782275245, + "grad_norm": 12.9375, + "learning_rate": 1.2573631844154387e-08, + "loss": 0.5949, + "step": 22414 + }, + { + "epoch": 1.910097997443545, + "grad_norm": 14.4375, + "learning_rate": 1.254986933215671e-08, + "loss": 0.69, + "step": 22415 + }, + { + "epoch": 1.910183212611845, + "grad_norm": 65.5, + "learning_rate": 1.2526129182127144e-08, + "loss": 0.7393, + "step": 22416 + }, + { + "epoch": 1.910268427780145, + "grad_norm": 11.0, + "learning_rate": 1.2502411394494928e-08, + "loss": 0.3549, + "step": 22417 + }, + { + "epoch": 1.910353642948445, + "grad_norm": 17.75, + "learning_rate": 1.2478715969688331e-08, + "loss": 0.7321, + "step": 22418 + }, + { + "epoch": 1.9104388581167449, + "grad_norm": 13.6875, + "learning_rate": 1.245504290813604e-08, + "loss": 0.5578, + "step": 22419 + }, + { + "epoch": 1.9105240732850448, + "grad_norm": 11.5, + "learning_rate": 1.2431392210265487e-08, + "loss": 0.5024, + "step": 22420 + }, + { + "epoch": 1.9106092884533448, + "grad_norm": 19.375, + "learning_rate": 1.2407763876504109e-08, + "loss": 0.629, + "step": 22421 + }, + { + "epoch": 1.9106945036216447, + "grad_norm": 19.625, + "learning_rate": 1.2384157907278926e-08, + "loss": 0.7235, + "step": 22422 + }, + { + "epoch": 1.9107797187899447, + "grad_norm": 14.0, + "learning_rate": 1.2360574303016543e-08, + "loss": 0.3473, + "step": 22423 + }, + { + "epoch": 1.9108649339582446, + "grad_norm": 24.875, + "learning_rate": 1.2337013064143144e-08, + "loss": 0.6643, + "step": 22424 + }, + { + "epoch": 1.9109501491265446, + "grad_norm": 13.9375, + "learning_rate": 1.2313474191084363e-08, + "loss": 0.3494, + "step": 22425 + }, + { + "epoch": 1.9110353642948446, + "grad_norm": 14.9375, + "learning_rate": 1.228995768426583e-08, + "loss": 0.6096, + "step": 22426 + }, + { + "epoch": 1.9111205794631445, + "grad_norm": 23.125, + "learning_rate": 1.2266463544112345e-08, + "loss": 0.8526, + "step": 22427 + }, + { + "epoch": 1.9112057946314445, + "grad_norm": 13.3125, + "learning_rate": 1.2242991771048568e-08, + "loss": 0.5948, + "step": 22428 + }, + { + "epoch": 1.9112910097997444, + "grad_norm": 25.75, + "learning_rate": 1.2219542365498466e-08, + "loss": 0.6469, + "step": 22429 + }, + { + "epoch": 1.9113762249680444, + "grad_norm": 20.875, + "learning_rate": 1.2196115327885866e-08, + "loss": 0.943, + "step": 22430 + }, + { + "epoch": 1.9114614401363443, + "grad_norm": 14.6875, + "learning_rate": 1.2172710658634318e-08, + "loss": 0.6619, + "step": 22431 + }, + { + "epoch": 1.9115466553046443, + "grad_norm": 17.375, + "learning_rate": 1.214932835816668e-08, + "loss": 0.7032, + "step": 22432 + }, + { + "epoch": 1.9116318704729443, + "grad_norm": 18.625, + "learning_rate": 1.212596842690525e-08, + "loss": 0.6572, + "step": 22433 + }, + { + "epoch": 1.9117170856412442, + "grad_norm": 16.625, + "learning_rate": 1.210263086527233e-08, + "loss": 0.588, + "step": 22434 + }, + { + "epoch": 1.9118023008095442, + "grad_norm": 19.375, + "learning_rate": 1.2079315673689807e-08, + "loss": 0.8764, + "step": 22435 + }, + { + "epoch": 1.9118875159778441, + "grad_norm": 16.25, + "learning_rate": 1.205602285257887e-08, + "loss": 0.5839, + "step": 22436 + }, + { + "epoch": 1.911972731146144, + "grad_norm": 19.5, + "learning_rate": 1.203275240236043e-08, + "loss": 1.0433, + "step": 22437 + }, + { + "epoch": 1.912057946314444, + "grad_norm": 23.0, + "learning_rate": 1.200950432345499e-08, + "loss": 0.8281, + "step": 22438 + }, + { + "epoch": 1.912143161482744, + "grad_norm": 13.4375, + "learning_rate": 1.1986278616282764e-08, + "loss": 0.4818, + "step": 22439 + }, + { + "epoch": 1.912228376651044, + "grad_norm": 17.375, + "learning_rate": 1.196307528126342e-08, + "loss": 0.6498, + "step": 22440 + }, + { + "epoch": 1.912313591819344, + "grad_norm": 13.5625, + "learning_rate": 1.1939894318816065e-08, + "loss": 0.5491, + "step": 22441 + }, + { + "epoch": 1.9123988069876439, + "grad_norm": 23.125, + "learning_rate": 1.1916735729359946e-08, + "loss": 0.7516, + "step": 22442 + }, + { + "epoch": 1.9124840221559438, + "grad_norm": 11.3125, + "learning_rate": 1.1893599513313341e-08, + "loss": 0.397, + "step": 22443 + }, + { + "epoch": 1.9125692373242438, + "grad_norm": 20.25, + "learning_rate": 1.1870485671094528e-08, + "loss": 0.6726, + "step": 22444 + }, + { + "epoch": 1.9126544524925437, + "grad_norm": 13.125, + "learning_rate": 1.184739420312095e-08, + "loss": 0.4048, + "step": 22445 + }, + { + "epoch": 1.9127396676608437, + "grad_norm": 15.125, + "learning_rate": 1.1824325109810048e-08, + "loss": 0.799, + "step": 22446 + }, + { + "epoch": 1.9128248828291436, + "grad_norm": 14.125, + "learning_rate": 1.1801278391578574e-08, + "loss": 0.4129, + "step": 22447 + }, + { + "epoch": 1.9129100979974436, + "grad_norm": 15.875, + "learning_rate": 1.177825404884328e-08, + "loss": 0.8818, + "step": 22448 + }, + { + "epoch": 1.9129953131657436, + "grad_norm": 17.375, + "learning_rate": 1.1755252082019942e-08, + "loss": 0.5926, + "step": 22449 + }, + { + "epoch": 1.9130805283340435, + "grad_norm": 13.9375, + "learning_rate": 1.1732272491524338e-08, + "loss": 0.4745, + "step": 22450 + }, + { + "epoch": 1.9131657435023435, + "grad_norm": 22.5, + "learning_rate": 1.1709315277771972e-08, + "loss": 0.6819, + "step": 22451 + }, + { + "epoch": 1.9132509586706434, + "grad_norm": 17.375, + "learning_rate": 1.1686380441177231e-08, + "loss": 1.0185, + "step": 22452 + }, + { + "epoch": 1.9133361738389434, + "grad_norm": 13.0, + "learning_rate": 1.1663467982154925e-08, + "loss": 0.4636, + "step": 22453 + }, + { + "epoch": 1.9134213890072433, + "grad_norm": 19.0, + "learning_rate": 1.1640577901118888e-08, + "loss": 0.533, + "step": 22454 + }, + { + "epoch": 1.9135066041755433, + "grad_norm": 13.4375, + "learning_rate": 1.1617710198482956e-08, + "loss": 0.5736, + "step": 22455 + }, + { + "epoch": 1.9135918193438433, + "grad_norm": 22.375, + "learning_rate": 1.1594864874660272e-08, + "loss": 0.6739, + "step": 22456 + }, + { + "epoch": 1.9136770345121432, + "grad_norm": 19.25, + "learning_rate": 1.1572041930063698e-08, + "loss": 0.629, + "step": 22457 + }, + { + "epoch": 1.9137622496804432, + "grad_norm": 13.6875, + "learning_rate": 1.1549241365105684e-08, + "loss": 0.5069, + "step": 22458 + }, + { + "epoch": 1.9138474648487431, + "grad_norm": 16.75, + "learning_rate": 1.1526463180198122e-08, + "loss": 0.5691, + "step": 22459 + }, + { + "epoch": 1.913932680017043, + "grad_norm": 18.5, + "learning_rate": 1.1503707375752904e-08, + "loss": 0.4273, + "step": 22460 + }, + { + "epoch": 1.914017895185343, + "grad_norm": 17.375, + "learning_rate": 1.1480973952181091e-08, + "loss": 0.6816, + "step": 22461 + }, + { + "epoch": 1.914103110353643, + "grad_norm": 13.8125, + "learning_rate": 1.1458262909893464e-08, + "loss": 0.5591, + "step": 22462 + }, + { + "epoch": 1.914188325521943, + "grad_norm": 14.6875, + "learning_rate": 1.143557424930053e-08, + "loss": 0.8277, + "step": 22463 + }, + { + "epoch": 1.914273540690243, + "grad_norm": 16.375, + "learning_rate": 1.1412907970812098e-08, + "loss": 0.7017, + "step": 22464 + }, + { + "epoch": 1.9143587558585429, + "grad_norm": 18.125, + "learning_rate": 1.1390264074837981e-08, + "loss": 1.0599, + "step": 22465 + }, + { + "epoch": 1.9144439710268428, + "grad_norm": 16.5, + "learning_rate": 1.1367642561787295e-08, + "loss": 0.5307, + "step": 22466 + }, + { + "epoch": 1.9145291861951428, + "grad_norm": 12.4375, + "learning_rate": 1.1345043432068881e-08, + "loss": 0.6101, + "step": 22467 + }, + { + "epoch": 1.9146144013634427, + "grad_norm": 13.5, + "learning_rate": 1.132246668609116e-08, + "loss": 0.5551, + "step": 22468 + }, + { + "epoch": 1.9146996165317427, + "grad_norm": 13.8125, + "learning_rate": 1.1299912324262002e-08, + "loss": 0.6443, + "step": 22469 + }, + { + "epoch": 1.9147848317000427, + "grad_norm": 16.875, + "learning_rate": 1.1277380346988997e-08, + "loss": 0.9283, + "step": 22470 + }, + { + "epoch": 1.9148700468683426, + "grad_norm": 18.375, + "learning_rate": 1.1254870754679458e-08, + "loss": 0.6707, + "step": 22471 + }, + { + "epoch": 1.9149552620366426, + "grad_norm": 10.4375, + "learning_rate": 1.1232383547739866e-08, + "loss": 0.3311, + "step": 22472 + }, + { + "epoch": 1.9150404772049425, + "grad_norm": 16.125, + "learning_rate": 1.1209918726576979e-08, + "loss": 0.3583, + "step": 22473 + }, + { + "epoch": 1.9151256923732425, + "grad_norm": 37.75, + "learning_rate": 1.1187476291596445e-08, + "loss": 0.6923, + "step": 22474 + }, + { + "epoch": 1.9152109075415424, + "grad_norm": 14.5, + "learning_rate": 1.1165056243204048e-08, + "loss": 0.5156, + "step": 22475 + }, + { + "epoch": 1.9152961227098424, + "grad_norm": 13.875, + "learning_rate": 1.1142658581804744e-08, + "loss": 0.5436, + "step": 22476 + }, + { + "epoch": 1.9153813378781424, + "grad_norm": 14.75, + "learning_rate": 1.1120283307803348e-08, + "loss": 0.5007, + "step": 22477 + }, + { + "epoch": 1.9154665530464423, + "grad_norm": 13.125, + "learning_rate": 1.1097930421604259e-08, + "loss": 0.4662, + "step": 22478 + }, + { + "epoch": 1.9155517682147423, + "grad_norm": 14.4375, + "learning_rate": 1.1075599923611458e-08, + "loss": 0.5888, + "step": 22479 + }, + { + "epoch": 1.9156369833830422, + "grad_norm": 14.0625, + "learning_rate": 1.1053291814228235e-08, + "loss": 0.5537, + "step": 22480 + }, + { + "epoch": 1.9157221985513422, + "grad_norm": 16.25, + "learning_rate": 1.1031006093858015e-08, + "loss": 0.3978, + "step": 22481 + }, + { + "epoch": 1.9158074137196421, + "grad_norm": 36.0, + "learning_rate": 1.1008742762903258e-08, + "loss": 0.7182, + "step": 22482 + }, + { + "epoch": 1.915892628887942, + "grad_norm": 13.9375, + "learning_rate": 1.0986501821766554e-08, + "loss": 0.4365, + "step": 22483 + }, + { + "epoch": 1.915977844056242, + "grad_norm": 21.5, + "learning_rate": 1.0964283270849668e-08, + "loss": 0.6692, + "step": 22484 + }, + { + "epoch": 1.916063059224542, + "grad_norm": 14.9375, + "learning_rate": 1.0942087110554084e-08, + "loss": 0.4515, + "step": 22485 + }, + { + "epoch": 1.916148274392842, + "grad_norm": 13.875, + "learning_rate": 1.0919913341281008e-08, + "loss": 0.5465, + "step": 22486 + }, + { + "epoch": 1.916233489561142, + "grad_norm": 18.75, + "learning_rate": 1.0897761963431091e-08, + "loss": 0.6176, + "step": 22487 + }, + { + "epoch": 1.9163187047294419, + "grad_norm": 16.625, + "learning_rate": 1.087563297740457e-08, + "loss": 0.7088, + "step": 22488 + }, + { + "epoch": 1.9164039198977418, + "grad_norm": 14.8125, + "learning_rate": 1.0853526383601404e-08, + "loss": 0.6015, + "step": 22489 + }, + { + "epoch": 1.9164891350660418, + "grad_norm": 15.625, + "learning_rate": 1.0831442182420992e-08, + "loss": 0.7489, + "step": 22490 + }, + { + "epoch": 1.9165743502343418, + "grad_norm": 16.375, + "learning_rate": 1.080938037426274e-08, + "loss": 0.6621, + "step": 22491 + }, + { + "epoch": 1.9166595654026417, + "grad_norm": 20.625, + "learning_rate": 1.078734095952494e-08, + "loss": 0.395, + "step": 22492 + }, + { + "epoch": 1.9167447805709417, + "grad_norm": 13.0625, + "learning_rate": 1.0765323938606021e-08, + "loss": 0.4263, + "step": 22493 + }, + { + "epoch": 1.9168299957392416, + "grad_norm": 14.9375, + "learning_rate": 1.074332931190386e-08, + "loss": 0.5519, + "step": 22494 + }, + { + "epoch": 1.9169152109075416, + "grad_norm": 14.875, + "learning_rate": 1.072135707981578e-08, + "loss": 0.4628, + "step": 22495 + }, + { + "epoch": 1.9170004260758415, + "grad_norm": 20.0, + "learning_rate": 1.069940724273924e-08, + "loss": 0.5568, + "step": 22496 + }, + { + "epoch": 1.9170856412441415, + "grad_norm": 16.75, + "learning_rate": 1.067747980107045e-08, + "loss": 0.3768, + "step": 22497 + }, + { + "epoch": 1.9171708564124414, + "grad_norm": 20.5, + "learning_rate": 1.065557475520576e-08, + "loss": 0.6064, + "step": 22498 + }, + { + "epoch": 1.9172560715807414, + "grad_norm": 14.125, + "learning_rate": 1.0633692105541244e-08, + "loss": 0.6848, + "step": 22499 + }, + { + "epoch": 1.9173412867490414, + "grad_norm": 13.75, + "learning_rate": 1.0611831852472138e-08, + "loss": 0.8272, + "step": 22500 + }, + { + "epoch": 1.9174265019173413, + "grad_norm": 17.875, + "learning_rate": 1.0589993996393544e-08, + "loss": 0.8754, + "step": 22501 + }, + { + "epoch": 1.9175117170856413, + "grad_norm": 12.4375, + "learning_rate": 1.0568178537700008e-08, + "loss": 0.4858, + "step": 22502 + }, + { + "epoch": 1.9175969322539412, + "grad_norm": 19.0, + "learning_rate": 1.0546385476785937e-08, + "loss": 0.5525, + "step": 22503 + }, + { + "epoch": 1.9176821474222412, + "grad_norm": 16.875, + "learning_rate": 1.0524614814045042e-08, + "loss": 0.5528, + "step": 22504 + }, + { + "epoch": 1.9177673625905411, + "grad_norm": 20.75, + "learning_rate": 1.0502866549870622e-08, + "loss": 0.7828, + "step": 22505 + }, + { + "epoch": 1.917852577758841, + "grad_norm": 12.625, + "learning_rate": 1.0481140684655833e-08, + "loss": 0.4273, + "step": 22506 + }, + { + "epoch": 1.917937792927141, + "grad_norm": 16.375, + "learning_rate": 1.0459437218793279e-08, + "loss": 0.5491, + "step": 22507 + }, + { + "epoch": 1.918023008095441, + "grad_norm": 15.0625, + "learning_rate": 1.0437756152675282e-08, + "loss": 0.8989, + "step": 22508 + }, + { + "epoch": 1.918108223263741, + "grad_norm": 18.375, + "learning_rate": 1.0416097486693476e-08, + "loss": 0.5715, + "step": 22509 + }, + { + "epoch": 1.918193438432041, + "grad_norm": 17.5, + "learning_rate": 1.0394461221239216e-08, + "loss": 0.7097, + "step": 22510 + }, + { + "epoch": 1.9182786536003409, + "grad_norm": 22.75, + "learning_rate": 1.0372847356703574e-08, + "loss": 0.8584, + "step": 22511 + }, + { + "epoch": 1.9183638687686408, + "grad_norm": 9.75, + "learning_rate": 1.0351255893477075e-08, + "loss": 0.235, + "step": 22512 + }, + { + "epoch": 1.9184490839369408, + "grad_norm": 23.375, + "learning_rate": 1.0329686831950098e-08, + "loss": 1.3183, + "step": 22513 + }, + { + "epoch": 1.9185342991052408, + "grad_norm": 11.9375, + "learning_rate": 1.0308140172512054e-08, + "loss": 0.3388, + "step": 22514 + }, + { + "epoch": 1.9186195142735407, + "grad_norm": 24.375, + "learning_rate": 1.0286615915552633e-08, + "loss": 0.7583, + "step": 22515 + }, + { + "epoch": 1.9187047294418407, + "grad_norm": 16.375, + "learning_rate": 1.026511406146069e-08, + "loss": 0.6557, + "step": 22516 + }, + { + "epoch": 1.9187899446101406, + "grad_norm": 11.6875, + "learning_rate": 1.0243634610624802e-08, + "loss": 0.3889, + "step": 22517 + }, + { + "epoch": 1.9188751597784406, + "grad_norm": 17.0, + "learning_rate": 1.0222177563433134e-08, + "loss": 0.7879, + "step": 22518 + }, + { + "epoch": 1.9189603749467405, + "grad_norm": 13.0625, + "learning_rate": 1.0200742920273288e-08, + "loss": 0.5139, + "step": 22519 + }, + { + "epoch": 1.9190455901150405, + "grad_norm": 14.5, + "learning_rate": 1.0179330681532873e-08, + "loss": 0.5789, + "step": 22520 + }, + { + "epoch": 1.9191308052833405, + "grad_norm": 13.4375, + "learning_rate": 1.0157940847598662e-08, + "loss": 0.6491, + "step": 22521 + }, + { + "epoch": 1.9192160204516404, + "grad_norm": 13.8125, + "learning_rate": 1.0136573418857153e-08, + "loss": 0.4337, + "step": 22522 + }, + { + "epoch": 1.9193012356199404, + "grad_norm": 16.5, + "learning_rate": 1.0115228395694565e-08, + "loss": 0.6612, + "step": 22523 + }, + { + "epoch": 1.9193864507882403, + "grad_norm": 15.25, + "learning_rate": 1.009390577849656e-08, + "loss": 0.9463, + "step": 22524 + }, + { + "epoch": 1.9194716659565403, + "grad_norm": 41.25, + "learning_rate": 1.0072605567648525e-08, + "loss": 1.2677, + "step": 22525 + }, + { + "epoch": 1.9195568811248402, + "grad_norm": 37.25, + "learning_rate": 1.0051327763535429e-08, + "loss": 0.7296, + "step": 22526 + }, + { + "epoch": 1.9196420962931402, + "grad_norm": 13.25, + "learning_rate": 1.0030072366541688e-08, + "loss": 0.405, + "step": 22527 + }, + { + "epoch": 1.9197273114614402, + "grad_norm": 13.625, + "learning_rate": 1.0008839377051299e-08, + "loss": 0.3942, + "step": 22528 + }, + { + "epoch": 1.9198125266297401, + "grad_norm": 18.25, + "learning_rate": 9.987628795448123e-09, + "loss": 0.7356, + "step": 22529 + }, + { + "epoch": 1.91989774179804, + "grad_norm": 16.125, + "learning_rate": 9.966440622115603e-09, + "loss": 0.3869, + "step": 22530 + }, + { + "epoch": 1.91998295696634, + "grad_norm": 16.25, + "learning_rate": 9.94527485743621e-09, + "loss": 0.6267, + "step": 22531 + }, + { + "epoch": 1.92006817213464, + "grad_norm": 17.125, + "learning_rate": 9.924131501792834e-09, + "loss": 0.7686, + "step": 22532 + }, + { + "epoch": 1.92015338730294, + "grad_norm": 12.6875, + "learning_rate": 9.903010555567389e-09, + "loss": 0.2718, + "step": 22533 + }, + { + "epoch": 1.92023860247124, + "grad_norm": 14.375, + "learning_rate": 9.881912019141516e-09, + "loss": 0.6447, + "step": 22534 + }, + { + "epoch": 1.9203238176395399, + "grad_norm": 12.6875, + "learning_rate": 9.860835892896575e-09, + "loss": 0.3742, + "step": 22535 + }, + { + "epoch": 1.9204090328078398, + "grad_norm": 14.375, + "learning_rate": 9.839782177213236e-09, + "loss": 0.6716, + "step": 22536 + }, + { + "epoch": 1.9204942479761398, + "grad_norm": 18.25, + "learning_rate": 9.818750872472166e-09, + "loss": 0.4224, + "step": 22537 + }, + { + "epoch": 1.9205794631444397, + "grad_norm": 12.1875, + "learning_rate": 9.797741979053477e-09, + "loss": 0.4535, + "step": 22538 + }, + { + "epoch": 1.9206646783127397, + "grad_norm": 10.5, + "learning_rate": 9.776755497336588e-09, + "loss": 0.3764, + "step": 22539 + }, + { + "epoch": 1.9207498934810396, + "grad_norm": 24.25, + "learning_rate": 9.755791427700917e-09, + "loss": 0.7252, + "step": 22540 + }, + { + "epoch": 1.9208351086493396, + "grad_norm": 32.75, + "learning_rate": 9.734849770525329e-09, + "loss": 0.7087, + "step": 22541 + }, + { + "epoch": 1.9209203238176396, + "grad_norm": 10.25, + "learning_rate": 9.71393052618813e-09, + "loss": 0.38, + "step": 22542 + }, + { + "epoch": 1.9210055389859395, + "grad_norm": 13.25, + "learning_rate": 9.693033695067494e-09, + "loss": 0.3871, + "step": 22543 + }, + { + "epoch": 1.9210907541542395, + "grad_norm": 11.8125, + "learning_rate": 9.672159277540754e-09, + "loss": 0.3912, + "step": 22544 + }, + { + "epoch": 1.9211759693225394, + "grad_norm": 11.875, + "learning_rate": 9.651307273985527e-09, + "loss": 0.2991, + "step": 22545 + }, + { + "epoch": 1.9212611844908394, + "grad_norm": 17.75, + "learning_rate": 9.630477684778316e-09, + "loss": 0.6307, + "step": 22546 + }, + { + "epoch": 1.9213463996591393, + "grad_norm": 16.25, + "learning_rate": 9.609670510295765e-09, + "loss": 0.7497, + "step": 22547 + }, + { + "epoch": 1.9214316148274393, + "grad_norm": 15.75, + "learning_rate": 9.588885750913685e-09, + "loss": 0.5276, + "step": 22548 + }, + { + "epoch": 1.9215168299957393, + "grad_norm": 18.625, + "learning_rate": 9.568123407007884e-09, + "loss": 0.5032, + "step": 22549 + }, + { + "epoch": 1.9216020451640392, + "grad_norm": 15.0, + "learning_rate": 9.547383478953343e-09, + "loss": 0.6313, + "step": 22550 + }, + { + "epoch": 1.9216872603323392, + "grad_norm": 22.0, + "learning_rate": 9.526665967125038e-09, + "loss": 0.8362, + "step": 22551 + }, + { + "epoch": 1.9217724755006391, + "grad_norm": 12.875, + "learning_rate": 9.505970871897252e-09, + "loss": 0.5725, + "step": 22552 + }, + { + "epoch": 1.921857690668939, + "grad_norm": 15.3125, + "learning_rate": 9.485298193643854e-09, + "loss": 0.5713, + "step": 22553 + }, + { + "epoch": 1.921942905837239, + "grad_norm": 23.5, + "learning_rate": 9.464647932738713e-09, + "loss": 1.1379, + "step": 22554 + }, + { + "epoch": 1.922028121005539, + "grad_norm": 16.0, + "learning_rate": 9.444020089554861e-09, + "loss": 0.6007, + "step": 22555 + }, + { + "epoch": 1.922113336173839, + "grad_norm": 20.125, + "learning_rate": 9.423414664464919e-09, + "loss": 0.6331, + "step": 22556 + }, + { + "epoch": 1.922198551342139, + "grad_norm": 16.25, + "learning_rate": 9.402831657841366e-09, + "loss": 0.9929, + "step": 22557 + }, + { + "epoch": 1.9222837665104389, + "grad_norm": 13.5625, + "learning_rate": 9.382271070056265e-09, + "loss": 0.391, + "step": 22558 + }, + { + "epoch": 1.9223689816787388, + "grad_norm": 10.625, + "learning_rate": 9.361732901480985e-09, + "loss": 0.2398, + "step": 22559 + }, + { + "epoch": 1.9224541968470388, + "grad_norm": 13.75, + "learning_rate": 9.34121715248676e-09, + "loss": 0.6736, + "step": 22560 + }, + { + "epoch": 1.9225394120153387, + "grad_norm": 11.0625, + "learning_rate": 9.3207238234444e-09, + "loss": 0.3752, + "step": 22561 + }, + { + "epoch": 1.9226246271836387, + "grad_norm": 17.375, + "learning_rate": 9.300252914724028e-09, + "loss": 0.4634, + "step": 22562 + }, + { + "epoch": 1.9227098423519386, + "grad_norm": 14.625, + "learning_rate": 9.279804426695765e-09, + "loss": 0.5634, + "step": 22563 + }, + { + "epoch": 1.9227950575202386, + "grad_norm": 13.5, + "learning_rate": 9.259378359729038e-09, + "loss": 0.4986, + "step": 22564 + }, + { + "epoch": 1.9228802726885386, + "grad_norm": 23.75, + "learning_rate": 9.238974714193133e-09, + "loss": 1.2789, + "step": 22565 + }, + { + "epoch": 1.9229654878568385, + "grad_norm": 19.75, + "learning_rate": 9.218593490456507e-09, + "loss": 0.7566, + "step": 22566 + }, + { + "epoch": 1.9230507030251385, + "grad_norm": 16.0, + "learning_rate": 9.198234688887614e-09, + "loss": 0.7058, + "step": 22567 + }, + { + "epoch": 1.9231359181934384, + "grad_norm": 20.75, + "learning_rate": 9.177898309854494e-09, + "loss": 0.8693, + "step": 22568 + }, + { + "epoch": 1.9232211333617384, + "grad_norm": 15.8125, + "learning_rate": 9.157584353724353e-09, + "loss": 0.8775, + "step": 22569 + }, + { + "epoch": 1.9233063485300383, + "grad_norm": 13.4375, + "learning_rate": 9.137292820864397e-09, + "loss": 0.5003, + "step": 22570 + }, + { + "epoch": 1.9233915636983383, + "grad_norm": 16.75, + "learning_rate": 9.117023711641415e-09, + "loss": 0.5035, + "step": 22571 + }, + { + "epoch": 1.9234767788666383, + "grad_norm": 12.4375, + "learning_rate": 9.096777026421643e-09, + "loss": 0.2213, + "step": 22572 + }, + { + "epoch": 1.9235619940349382, + "grad_norm": 12.3125, + "learning_rate": 9.0765527655709e-09, + "loss": 0.4109, + "step": 22573 + }, + { + "epoch": 1.9236472092032382, + "grad_norm": 14.4375, + "learning_rate": 9.056350929454588e-09, + "loss": 0.5318, + "step": 22574 + }, + { + "epoch": 1.9237324243715381, + "grad_norm": 18.625, + "learning_rate": 9.036171518437969e-09, + "loss": 0.5912, + "step": 22575 + }, + { + "epoch": 1.923817639539838, + "grad_norm": 15.125, + "learning_rate": 9.016014532885752e-09, + "loss": 0.4754, + "step": 22576 + }, + { + "epoch": 1.923902854708138, + "grad_norm": 13.0625, + "learning_rate": 8.995879973161814e-09, + "loss": 0.5072, + "step": 22577 + }, + { + "epoch": 1.923988069876438, + "grad_norm": 18.125, + "learning_rate": 8.975767839630444e-09, + "loss": 0.4691, + "step": 22578 + }, + { + "epoch": 1.924073285044738, + "grad_norm": 16.375, + "learning_rate": 8.955678132654688e-09, + "loss": 0.7053, + "step": 22579 + }, + { + "epoch": 1.924158500213038, + "grad_norm": 17.375, + "learning_rate": 8.935610852598003e-09, + "loss": 0.6823, + "step": 22580 + }, + { + "epoch": 1.9242437153813379, + "grad_norm": 12.125, + "learning_rate": 8.91556599982274e-09, + "loss": 0.3331, + "step": 22581 + }, + { + "epoch": 1.9243289305496378, + "grad_norm": 15.125, + "learning_rate": 8.895543574691245e-09, + "loss": 0.6116, + "step": 22582 + }, + { + "epoch": 1.9244141457179378, + "grad_norm": 14.5, + "learning_rate": 8.875543577565177e-09, + "loss": 0.5752, + "step": 22583 + }, + { + "epoch": 1.9244993608862377, + "grad_norm": 14.375, + "learning_rate": 8.855566008806189e-09, + "loss": 0.4723, + "step": 22584 + }, + { + "epoch": 1.9245845760545377, + "grad_norm": 24.5, + "learning_rate": 8.835610868775241e-09, + "loss": 0.6455, + "step": 22585 + }, + { + "epoch": 1.9246697912228377, + "grad_norm": 18.375, + "learning_rate": 8.81567815783274e-09, + "loss": 0.6921, + "step": 22586 + }, + { + "epoch": 1.9247550063911376, + "grad_norm": 15.25, + "learning_rate": 8.795767876339094e-09, + "loss": 0.5211, + "step": 22587 + }, + { + "epoch": 1.9248402215594376, + "grad_norm": 16.375, + "learning_rate": 8.775880024654155e-09, + "loss": 0.7015, + "step": 22588 + }, + { + "epoch": 1.9249254367277375, + "grad_norm": 12.0625, + "learning_rate": 8.756014603137075e-09, + "loss": 0.3585, + "step": 22589 + }, + { + "epoch": 1.9250106518960375, + "grad_norm": 12.75, + "learning_rate": 8.736171612147154e-09, + "loss": 0.6216, + "step": 22590 + }, + { + "epoch": 1.9250958670643374, + "grad_norm": 15.125, + "learning_rate": 8.716351052042577e-09, + "loss": 0.6348, + "step": 22591 + }, + { + "epoch": 1.9251810822326374, + "grad_norm": 31.625, + "learning_rate": 8.696552923181945e-09, + "loss": 0.9524, + "step": 22592 + }, + { + "epoch": 1.9252662974009374, + "grad_norm": 11.375, + "learning_rate": 8.676777225922751e-09, + "loss": 0.4501, + "step": 22593 + }, + { + "epoch": 1.9253515125692373, + "grad_norm": 14.1875, + "learning_rate": 8.657023960622345e-09, + "loss": 0.5972, + "step": 22594 + }, + { + "epoch": 1.9254367277375373, + "grad_norm": 16.625, + "learning_rate": 8.637293127637946e-09, + "loss": 0.4909, + "step": 22595 + }, + { + "epoch": 1.9255219429058372, + "grad_norm": 12.6875, + "learning_rate": 8.617584727325796e-09, + "loss": 0.3016, + "step": 22596 + }, + { + "epoch": 1.9256071580741372, + "grad_norm": 16.625, + "learning_rate": 8.597898760042272e-09, + "loss": 0.5652, + "step": 22597 + }, + { + "epoch": 1.9256923732424371, + "grad_norm": 15.5625, + "learning_rate": 8.578235226143066e-09, + "loss": 0.5353, + "step": 22598 + }, + { + "epoch": 1.925777588410737, + "grad_norm": 12.625, + "learning_rate": 8.558594125983449e-09, + "loss": 0.3686, + "step": 22599 + }, + { + "epoch": 1.925862803579037, + "grad_norm": 14.5625, + "learning_rate": 8.538975459918414e-09, + "loss": 0.623, + "step": 22600 + }, + { + "epoch": 1.925948018747337, + "grad_norm": 10.875, + "learning_rate": 8.519379228302398e-09, + "loss": 0.2783, + "step": 22601 + }, + { + "epoch": 1.926033233915637, + "grad_norm": 15.5, + "learning_rate": 8.499805431489844e-09, + "loss": 0.6176, + "step": 22602 + }, + { + "epoch": 1.926118449083937, + "grad_norm": 11.875, + "learning_rate": 8.480254069833938e-09, + "loss": 0.3249, + "step": 22603 + }, + { + "epoch": 1.9262036642522369, + "grad_norm": 19.375, + "learning_rate": 8.460725143688425e-09, + "loss": 0.7907, + "step": 22604 + }, + { + "epoch": 1.9262888794205368, + "grad_norm": 13.9375, + "learning_rate": 8.44121865340608e-09, + "loss": 0.4327, + "step": 22605 + }, + { + "epoch": 1.9263740945888368, + "grad_norm": 14.625, + "learning_rate": 8.421734599339537e-09, + "loss": 0.4961, + "step": 22606 + }, + { + "epoch": 1.9264593097571368, + "grad_norm": 15.375, + "learning_rate": 8.402272981840599e-09, + "loss": 0.6472, + "step": 22607 + }, + { + "epoch": 1.9265445249254367, + "grad_norm": 16.0, + "learning_rate": 8.382833801261065e-09, + "loss": 0.6273, + "step": 22608 + }, + { + "epoch": 1.9266297400937367, + "grad_norm": 14.5625, + "learning_rate": 8.363417057952322e-09, + "loss": 0.4786, + "step": 22609 + }, + { + "epoch": 1.9267149552620366, + "grad_norm": 14.25, + "learning_rate": 8.344022752265201e-09, + "loss": 0.5719, + "step": 22610 + }, + { + "epoch": 1.9268001704303366, + "grad_norm": 13.875, + "learning_rate": 8.324650884550117e-09, + "loss": 0.408, + "step": 22611 + }, + { + "epoch": 1.9268853855986365, + "grad_norm": 12.875, + "learning_rate": 8.305301455157205e-09, + "loss": 0.5297, + "step": 22612 + }, + { + "epoch": 1.9269706007669365, + "grad_norm": 14.6875, + "learning_rate": 8.285974464436186e-09, + "loss": 0.5254, + "step": 22613 + }, + { + "epoch": 1.9270558159352364, + "grad_norm": 16.5, + "learning_rate": 8.266669912736225e-09, + "loss": 0.7244, + "step": 22614 + }, + { + "epoch": 1.9271410311035364, + "grad_norm": 10.6875, + "learning_rate": 8.247387800406209e-09, + "loss": 0.2749, + "step": 22615 + }, + { + "epoch": 1.9272262462718364, + "grad_norm": 13.0625, + "learning_rate": 8.228128127794472e-09, + "loss": 0.4528, + "step": 22616 + }, + { + "epoch": 1.9273114614401363, + "grad_norm": 14.25, + "learning_rate": 8.208890895249344e-09, + "loss": 0.5463, + "step": 22617 + }, + { + "epoch": 1.9273966766084363, + "grad_norm": 13.0625, + "learning_rate": 8.18967610311805e-09, + "loss": 0.4294, + "step": 22618 + }, + { + "epoch": 1.9274818917767362, + "grad_norm": 11.6875, + "learning_rate": 8.170483751748226e-09, + "loss": 0.394, + "step": 22619 + }, + { + "epoch": 1.9275671069450362, + "grad_norm": 17.75, + "learning_rate": 8.151313841486402e-09, + "loss": 0.7623, + "step": 22620 + }, + { + "epoch": 1.9276523221133361, + "grad_norm": 20.125, + "learning_rate": 8.132166372679245e-09, + "loss": 0.9883, + "step": 22621 + }, + { + "epoch": 1.927737537281636, + "grad_norm": 14.125, + "learning_rate": 8.11304134567259e-09, + "loss": 0.4393, + "step": 22622 + }, + { + "epoch": 1.927822752449936, + "grad_norm": 17.5, + "learning_rate": 8.093938760811992e-09, + "loss": 0.6267, + "step": 22623 + }, + { + "epoch": 1.927907967618236, + "grad_norm": 19.25, + "learning_rate": 8.074858618442871e-09, + "loss": 1.0013, + "step": 22624 + }, + { + "epoch": 1.927993182786536, + "grad_norm": 14.0625, + "learning_rate": 8.055800918909813e-09, + "loss": 0.5435, + "step": 22625 + }, + { + "epoch": 1.928078397954836, + "grad_norm": 16.625, + "learning_rate": 8.036765662557262e-09, + "loss": 0.775, + "step": 22626 + }, + { + "epoch": 1.9281636131231359, + "grad_norm": 11.625, + "learning_rate": 8.017752849729387e-09, + "loss": 0.3296, + "step": 22627 + }, + { + "epoch": 1.9282488282914358, + "grad_norm": 15.0, + "learning_rate": 7.998762480769528e-09, + "loss": 0.4901, + "step": 22628 + }, + { + "epoch": 1.9283340434597358, + "grad_norm": 10.1875, + "learning_rate": 7.97979455602102e-09, + "loss": 0.2073, + "step": 22629 + }, + { + "epoch": 1.9284192586280358, + "grad_norm": 20.625, + "learning_rate": 7.960849075826643e-09, + "loss": 0.7853, + "step": 22630 + }, + { + "epoch": 1.9285044737963357, + "grad_norm": 15.125, + "learning_rate": 7.941926040528486e-09, + "loss": 0.4122, + "step": 22631 + }, + { + "epoch": 1.9285896889646357, + "grad_norm": 24.0, + "learning_rate": 7.923025450468914e-09, + "loss": 0.9443, + "step": 22632 + }, + { + "epoch": 1.9286749041329356, + "grad_norm": 14.5625, + "learning_rate": 7.90414730598918e-09, + "loss": 0.7375, + "step": 22633 + }, + { + "epoch": 1.9287601193012356, + "grad_norm": 14.0, + "learning_rate": 7.88529160743054e-09, + "loss": 0.364, + "step": 22634 + }, + { + "epoch": 1.9288453344695355, + "grad_norm": 14.75, + "learning_rate": 7.866458355133693e-09, + "loss": 0.4319, + "step": 22635 + }, + { + "epoch": 1.9289305496378355, + "grad_norm": 15.5, + "learning_rate": 7.847647549439064e-09, + "loss": 0.5604, + "step": 22636 + }, + { + "epoch": 1.9290157648061355, + "grad_norm": 20.5, + "learning_rate": 7.828859190686517e-09, + "loss": 1.0356, + "step": 22637 + }, + { + "epoch": 1.9291009799744354, + "grad_norm": 14.25, + "learning_rate": 7.810093279215502e-09, + "loss": 0.5609, + "step": 22638 + }, + { + "epoch": 1.9291861951427354, + "grad_norm": 22.25, + "learning_rate": 7.791349815365334e-09, + "loss": 0.7215, + "step": 22639 + }, + { + "epoch": 1.9292714103110353, + "grad_norm": 10.375, + "learning_rate": 7.77262879947463e-09, + "loss": 0.266, + "step": 22640 + }, + { + "epoch": 1.9293566254793353, + "grad_norm": 12.4375, + "learning_rate": 7.753930231881452e-09, + "loss": 0.433, + "step": 22641 + }, + { + "epoch": 1.9294418406476352, + "grad_norm": 13.0, + "learning_rate": 7.73525411292414e-09, + "loss": 0.4533, + "step": 22642 + }, + { + "epoch": 1.9295270558159352, + "grad_norm": 24.875, + "learning_rate": 7.716600442939926e-09, + "loss": 0.9608, + "step": 22643 + }, + { + "epoch": 1.9296122709842352, + "grad_norm": 16.875, + "learning_rate": 7.6979692222659e-09, + "loss": 0.5017, + "step": 22644 + }, + { + "epoch": 1.9296974861525351, + "grad_norm": 14.875, + "learning_rate": 7.679360451238877e-09, + "loss": 0.5903, + "step": 22645 + }, + { + "epoch": 1.929782701320835, + "grad_norm": 14.25, + "learning_rate": 7.660774130195114e-09, + "loss": 0.3448, + "step": 22646 + }, + { + "epoch": 1.929867916489135, + "grad_norm": 16.5, + "learning_rate": 7.642210259470318e-09, + "loss": 0.7383, + "step": 22647 + }, + { + "epoch": 1.929953131657435, + "grad_norm": 13.6875, + "learning_rate": 7.62366883940005e-09, + "loss": 0.5643, + "step": 22648 + }, + { + "epoch": 1.930038346825735, + "grad_norm": 17.625, + "learning_rate": 7.605149870319462e-09, + "loss": 0.9396, + "step": 22649 + }, + { + "epoch": 1.930123561994035, + "grad_norm": 18.625, + "learning_rate": 7.586653352563145e-09, + "loss": 0.3825, + "step": 22650 + }, + { + "epoch": 1.9302087771623349, + "grad_norm": 35.0, + "learning_rate": 7.568179286465277e-09, + "loss": 1.088, + "step": 22651 + }, + { + "epoch": 1.9302939923306348, + "grad_norm": 17.875, + "learning_rate": 7.549727672359896e-09, + "loss": 0.9372, + "step": 22652 + }, + { + "epoch": 1.9303792074989348, + "grad_norm": 15.1875, + "learning_rate": 7.53129851058021e-09, + "loss": 0.3839, + "step": 22653 + }, + { + "epoch": 1.9304644226672347, + "grad_norm": 16.875, + "learning_rate": 7.51289180145942e-09, + "loss": 0.8535, + "step": 22654 + }, + { + "epoch": 1.9305496378355347, + "grad_norm": 15.1875, + "learning_rate": 7.494507545330042e-09, + "loss": 0.44, + "step": 22655 + }, + { + "epoch": 1.9306348530038346, + "grad_norm": 16.25, + "learning_rate": 7.476145742524449e-09, + "loss": 0.7547, + "step": 22656 + }, + { + "epoch": 1.9307200681721346, + "grad_norm": 11.625, + "learning_rate": 7.457806393374317e-09, + "loss": 0.2323, + "step": 22657 + }, + { + "epoch": 1.9308052833404346, + "grad_norm": 15.75, + "learning_rate": 7.43948949821105e-09, + "loss": 0.5486, + "step": 22658 + }, + { + "epoch": 1.9308904985087345, + "grad_norm": 17.125, + "learning_rate": 7.421195057365632e-09, + "loss": 0.8541, + "step": 22659 + }, + { + "epoch": 1.9309757136770345, + "grad_norm": 14.75, + "learning_rate": 7.402923071168772e-09, + "loss": 0.6784, + "step": 22660 + }, + { + "epoch": 1.9310609288453344, + "grad_norm": 16.25, + "learning_rate": 7.3846735399506206e-09, + "loss": 0.6166, + "step": 22661 + }, + { + "epoch": 1.9311461440136344, + "grad_norm": 15.0625, + "learning_rate": 7.366446464041055e-09, + "loss": 0.5266, + "step": 22662 + }, + { + "epoch": 1.9312313591819343, + "grad_norm": 16.625, + "learning_rate": 7.3482418437693936e-09, + "loss": 0.7341, + "step": 22663 + }, + { + "epoch": 1.9313165743502343, + "grad_norm": 12.75, + "learning_rate": 7.330059679464402e-09, + "loss": 0.4716, + "step": 22664 + }, + { + "epoch": 1.9314017895185343, + "grad_norm": 12.0, + "learning_rate": 7.311899971454983e-09, + "loss": 0.6512, + "step": 22665 + }, + { + "epoch": 1.9314870046868342, + "grad_norm": 15.25, + "learning_rate": 7.2937627200690695e-09, + "loss": 0.2825, + "step": 22666 + }, + { + "epoch": 1.9315722198551342, + "grad_norm": 14.9375, + "learning_rate": 7.275647925634455e-09, + "loss": 0.6686, + "step": 22667 + }, + { + "epoch": 1.9316574350234341, + "grad_norm": 19.375, + "learning_rate": 7.257555588478654e-09, + "loss": 0.7956, + "step": 22668 + }, + { + "epoch": 1.931742650191734, + "grad_norm": 20.0, + "learning_rate": 7.2394857089284896e-09, + "loss": 0.6906, + "step": 22669 + }, + { + "epoch": 1.931827865360034, + "grad_norm": 15.25, + "learning_rate": 7.221438287310506e-09, + "loss": 0.7041, + "step": 22670 + }, + { + "epoch": 1.931913080528334, + "grad_norm": 17.125, + "learning_rate": 7.2034133239508304e-09, + "loss": 0.6499, + "step": 22671 + }, + { + "epoch": 1.931998295696634, + "grad_norm": 16.5, + "learning_rate": 7.185410819175176e-09, + "loss": 0.6882, + "step": 22672 + }, + { + "epoch": 1.932083510864934, + "grad_norm": 12.0625, + "learning_rate": 7.167430773308837e-09, + "loss": 0.3566, + "step": 22673 + }, + { + "epoch": 1.9321687260332339, + "grad_norm": 16.875, + "learning_rate": 7.1494731866769695e-09, + "loss": 0.2835, + "step": 22674 + }, + { + "epoch": 1.9322539412015338, + "grad_norm": 19.875, + "learning_rate": 7.131538059603759e-09, + "loss": 0.7884, + "step": 22675 + }, + { + "epoch": 1.9323391563698338, + "grad_norm": 11.9375, + "learning_rate": 7.113625392413393e-09, + "loss": 0.4734, + "step": 22676 + }, + { + "epoch": 1.9324243715381337, + "grad_norm": 14.125, + "learning_rate": 7.0957351854297754e-09, + "loss": 0.3725, + "step": 22677 + }, + { + "epoch": 1.9325095867064337, + "grad_norm": 15.5625, + "learning_rate": 7.0778674389759825e-09, + "loss": 0.3687, + "step": 22678 + }, + { + "epoch": 1.9325948018747336, + "grad_norm": 18.375, + "learning_rate": 7.06002215337509e-09, + "loss": 0.5184, + "step": 22679 + }, + { + "epoch": 1.9326800170430336, + "grad_norm": 15.0625, + "learning_rate": 7.04219932894934e-09, + "loss": 0.5969, + "step": 22680 + }, + { + "epoch": 1.9327652322113336, + "grad_norm": 15.5, + "learning_rate": 7.0243989660209734e-09, + "loss": 0.7348, + "step": 22681 + }, + { + "epoch": 1.9328504473796335, + "grad_norm": 12.8125, + "learning_rate": 7.006621064911679e-09, + "loss": 0.5079, + "step": 22682 + }, + { + "epoch": 1.9329356625479335, + "grad_norm": 20.625, + "learning_rate": 6.988865625942726e-09, + "loss": 0.6684, + "step": 22683 + }, + { + "epoch": 1.9330208777162334, + "grad_norm": 20.375, + "learning_rate": 6.971132649434692e-09, + "loss": 1.1204, + "step": 22684 + }, + { + "epoch": 1.9331060928845334, + "grad_norm": 18.375, + "learning_rate": 6.953422135708432e-09, + "loss": 0.4149, + "step": 22685 + }, + { + "epoch": 1.9331913080528333, + "grad_norm": 18.375, + "learning_rate": 6.935734085083828e-09, + "loss": 0.6352, + "step": 22686 + }, + { + "epoch": 1.9332765232211333, + "grad_norm": 32.25, + "learning_rate": 6.918068497880487e-09, + "loss": 0.8614, + "step": 22687 + }, + { + "epoch": 1.9333617383894333, + "grad_norm": 20.25, + "learning_rate": 6.900425374417596e-09, + "loss": 0.8512, + "step": 22688 + }, + { + "epoch": 1.9334469535577332, + "grad_norm": 18.75, + "learning_rate": 6.8828047150140686e-09, + "loss": 0.8382, + "step": 22689 + }, + { + "epoch": 1.9335321687260332, + "grad_norm": 28.875, + "learning_rate": 6.86520651998826e-09, + "loss": 0.6198, + "step": 22690 + }, + { + "epoch": 1.9336173838943331, + "grad_norm": 14.125, + "learning_rate": 6.847630789658388e-09, + "loss": 0.686, + "step": 22691 + }, + { + "epoch": 1.933702599062633, + "grad_norm": 12.625, + "learning_rate": 6.830077524341838e-09, + "loss": 0.4367, + "step": 22692 + }, + { + "epoch": 1.933787814230933, + "grad_norm": 16.125, + "learning_rate": 6.8125467243558565e-09, + "loss": 0.6743, + "step": 22693 + }, + { + "epoch": 1.933873029399233, + "grad_norm": 15.5625, + "learning_rate": 6.795038390017272e-09, + "loss": 0.6767, + "step": 22694 + }, + { + "epoch": 1.933958244567533, + "grad_norm": 19.25, + "learning_rate": 6.7775525216425e-09, + "loss": 0.7401, + "step": 22695 + }, + { + "epoch": 1.934043459735833, + "grad_norm": 21.25, + "learning_rate": 6.760089119547536e-09, + "loss": 0.8233, + "step": 22696 + }, + { + "epoch": 1.9341286749041329, + "grad_norm": 12.0625, + "learning_rate": 6.742648184047962e-09, + "loss": 0.7685, + "step": 22697 + }, + { + "epoch": 1.9342138900724328, + "grad_norm": 20.875, + "learning_rate": 6.725229715458803e-09, + "loss": 0.8169, + "step": 22698 + }, + { + "epoch": 1.9342991052407328, + "grad_norm": 12.3125, + "learning_rate": 6.707833714095085e-09, + "loss": 0.5704, + "step": 22699 + }, + { + "epoch": 1.9343843204090327, + "grad_norm": 15.6875, + "learning_rate": 6.690460180271002e-09, + "loss": 0.5682, + "step": 22700 + }, + { + "epoch": 1.9344695355773327, + "grad_norm": 17.0, + "learning_rate": 6.673109114300469e-09, + "loss": 0.7426, + "step": 22701 + }, + { + "epoch": 1.9345547507456327, + "grad_norm": 13.375, + "learning_rate": 6.6557805164972625e-09, + "loss": 0.4975, + "step": 22702 + }, + { + "epoch": 1.9346399659139326, + "grad_norm": 21.0, + "learning_rate": 6.638474387174326e-09, + "loss": 0.8428, + "step": 22703 + }, + { + "epoch": 1.9347251810822326, + "grad_norm": 12.375, + "learning_rate": 6.621190726644467e-09, + "loss": 0.499, + "step": 22704 + }, + { + "epoch": 1.9348103962505325, + "grad_norm": 15.8125, + "learning_rate": 6.6039295352199325e-09, + "loss": 0.6846, + "step": 22705 + }, + { + "epoch": 1.9348956114188325, + "grad_norm": 13.4375, + "learning_rate": 6.586690813212699e-09, + "loss": 0.5637, + "step": 22706 + }, + { + "epoch": 1.9349808265871324, + "grad_norm": 12.875, + "learning_rate": 6.569474560934319e-09, + "loss": 0.4794, + "step": 22707 + }, + { + "epoch": 1.9350660417554324, + "grad_norm": 14.25, + "learning_rate": 6.552280778696074e-09, + "loss": 0.6628, + "step": 22708 + }, + { + "epoch": 1.9351512569237324, + "grad_norm": 28.125, + "learning_rate": 6.535109466808271e-09, + "loss": 0.7128, + "step": 22709 + }, + { + "epoch": 1.9352364720920323, + "grad_norm": 16.75, + "learning_rate": 6.5179606255816315e-09, + "loss": 0.7488, + "step": 22710 + }, + { + "epoch": 1.9353216872603323, + "grad_norm": 15.8125, + "learning_rate": 6.500834255325772e-09, + "loss": 0.6221, + "step": 22711 + }, + { + "epoch": 1.9354069024286322, + "grad_norm": 9.9375, + "learning_rate": 6.483730356350304e-09, + "loss": 0.3922, + "step": 22712 + }, + { + "epoch": 1.9354921175969322, + "grad_norm": 17.25, + "learning_rate": 6.4666489289642865e-09, + "loss": 0.5282, + "step": 22713 + }, + { + "epoch": 1.9355773327652321, + "grad_norm": 15.0625, + "learning_rate": 6.449589973476361e-09, + "loss": 0.458, + "step": 22714 + }, + { + "epoch": 1.935662547933532, + "grad_norm": 17.5, + "learning_rate": 6.432553490194754e-09, + "loss": 0.6653, + "step": 22715 + }, + { + "epoch": 1.935747763101832, + "grad_norm": 18.5, + "learning_rate": 6.4155394794275514e-09, + "loss": 0.7565, + "step": 22716 + }, + { + "epoch": 1.935832978270132, + "grad_norm": 13.3125, + "learning_rate": 6.398547941482008e-09, + "loss": 0.4604, + "step": 22717 + }, + { + "epoch": 1.935918193438432, + "grad_norm": 18.0, + "learning_rate": 6.38157887666524e-09, + "loss": 0.5182, + "step": 22718 + }, + { + "epoch": 1.936003408606732, + "grad_norm": 21.375, + "learning_rate": 6.364632285283806e-09, + "loss": 0.8154, + "step": 22719 + }, + { + "epoch": 1.9360886237750319, + "grad_norm": 14.5625, + "learning_rate": 6.3477081676441285e-09, + "loss": 0.6668, + "step": 22720 + }, + { + "epoch": 1.9361738389433318, + "grad_norm": 12.4375, + "learning_rate": 6.330806524052075e-09, + "loss": 0.4308, + "step": 22721 + }, + { + "epoch": 1.9362590541116318, + "grad_norm": 15.625, + "learning_rate": 6.313927354812816e-09, + "loss": 0.564, + "step": 22722 + }, + { + "epoch": 1.9363442692799318, + "grad_norm": 14.75, + "learning_rate": 6.297070660231525e-09, + "loss": 0.7012, + "step": 22723 + }, + { + "epoch": 1.9364294844482317, + "grad_norm": 19.125, + "learning_rate": 6.28023644061268e-09, + "loss": 0.5094, + "step": 22724 + }, + { + "epoch": 1.9365146996165317, + "grad_norm": 19.0, + "learning_rate": 6.2634246962609006e-09, + "loss": 0.5025, + "step": 22725 + }, + { + "epoch": 1.9365999147848316, + "grad_norm": 13.0, + "learning_rate": 6.246635427479414e-09, + "loss": 0.5157, + "step": 22726 + }, + { + "epoch": 1.9366851299531316, + "grad_norm": 21.0, + "learning_rate": 6.229868634572145e-09, + "loss": 0.7772, + "step": 22727 + }, + { + "epoch": 1.9367703451214315, + "grad_norm": 18.5, + "learning_rate": 6.213124317841768e-09, + "loss": 0.6077, + "step": 22728 + }, + { + "epoch": 1.9368555602897315, + "grad_norm": 25.375, + "learning_rate": 6.196402477590957e-09, + "loss": 0.7908, + "step": 22729 + }, + { + "epoch": 1.9369407754580314, + "grad_norm": 14.0, + "learning_rate": 6.179703114121971e-09, + "loss": 0.4777, + "step": 22730 + }, + { + "epoch": 1.9370259906263314, + "grad_norm": 20.875, + "learning_rate": 6.163026227736374e-09, + "loss": 0.6595, + "step": 22731 + }, + { + "epoch": 1.9371112057946314, + "grad_norm": 14.125, + "learning_rate": 6.1463718187357305e-09, + "loss": 0.4099, + "step": 22732 + }, + { + "epoch": 1.9371964209629313, + "grad_norm": 24.375, + "learning_rate": 6.129739887421049e-09, + "loss": 1.1084, + "step": 22733 + }, + { + "epoch": 1.9372816361312313, + "grad_norm": 10.8125, + "learning_rate": 6.113130434092645e-09, + "loss": 0.3347, + "step": 22734 + }, + { + "epoch": 1.9373668512995312, + "grad_norm": 14.9375, + "learning_rate": 6.096543459050697e-09, + "loss": 0.8335, + "step": 22735 + }, + { + "epoch": 1.9374520664678312, + "grad_norm": 17.0, + "learning_rate": 6.07997896259524e-09, + "loss": 0.526, + "step": 22736 + }, + { + "epoch": 1.9375372816361311, + "grad_norm": 12.25, + "learning_rate": 6.063436945025341e-09, + "loss": 0.3248, + "step": 22737 + }, + { + "epoch": 1.937622496804431, + "grad_norm": 23.0, + "learning_rate": 6.046917406640068e-09, + "loss": 0.5618, + "step": 22738 + }, + { + "epoch": 1.937707711972731, + "grad_norm": 17.625, + "learning_rate": 6.030420347737792e-09, + "loss": 0.8767, + "step": 22739 + }, + { + "epoch": 1.937792927141031, + "grad_norm": 16.875, + "learning_rate": 6.013945768616885e-09, + "loss": 0.9016, + "step": 22740 + }, + { + "epoch": 1.937878142309331, + "grad_norm": 21.125, + "learning_rate": 5.9974936695747495e-09, + "loss": 1.1139, + "step": 22741 + }, + { + "epoch": 1.937963357477631, + "grad_norm": 15.625, + "learning_rate": 5.981064050908925e-09, + "loss": 0.4342, + "step": 22742 + }, + { + "epoch": 1.938048572645931, + "grad_norm": 14.875, + "learning_rate": 5.9646569129161165e-09, + "loss": 0.8037, + "step": 22743 + }, + { + "epoch": 1.9381337878142308, + "grad_norm": 15.1875, + "learning_rate": 5.948272255893173e-09, + "loss": 0.7155, + "step": 22744 + }, + { + "epoch": 1.9382190029825308, + "grad_norm": 23.875, + "learning_rate": 5.931910080135689e-09, + "loss": 0.8513, + "step": 22745 + }, + { + "epoch": 1.9383042181508308, + "grad_norm": 32.75, + "learning_rate": 5.9155703859398174e-09, + "loss": 0.8244, + "step": 22746 + }, + { + "epoch": 1.9383894333191307, + "grad_norm": 13.875, + "learning_rate": 5.899253173600461e-09, + "loss": 0.3749, + "step": 22747 + }, + { + "epoch": 1.9384746484874307, + "grad_norm": 18.375, + "learning_rate": 5.882958443412662e-09, + "loss": 0.7539, + "step": 22748 + }, + { + "epoch": 1.9385598636557306, + "grad_norm": 13.625, + "learning_rate": 5.866686195670768e-09, + "loss": 0.4786, + "step": 22749 + }, + { + "epoch": 1.9386450788240306, + "grad_norm": 17.5, + "learning_rate": 5.850436430669126e-09, + "loss": 0.6354, + "step": 22750 + }, + { + "epoch": 1.9387302939923305, + "grad_norm": 15.3125, + "learning_rate": 5.834209148700976e-09, + "loss": 0.4538, + "step": 22751 + }, + { + "epoch": 1.9388155091606305, + "grad_norm": 19.375, + "learning_rate": 5.818004350059969e-09, + "loss": 0.7187, + "step": 22752 + }, + { + "epoch": 1.9389007243289305, + "grad_norm": 24.5, + "learning_rate": 5.801822035038651e-09, + "loss": 0.9284, + "step": 22753 + }, + { + "epoch": 1.9389859394972304, + "grad_norm": 20.375, + "learning_rate": 5.785662203929565e-09, + "loss": 0.7277, + "step": 22754 + }, + { + "epoch": 1.9390711546655304, + "grad_norm": 20.625, + "learning_rate": 5.7695248570246995e-09, + "loss": 0.7639, + "step": 22755 + }, + { + "epoch": 1.9391563698338303, + "grad_norm": 12.6875, + "learning_rate": 5.753409994615628e-09, + "loss": 0.5018, + "step": 22756 + }, + { + "epoch": 1.9392415850021303, + "grad_norm": 17.125, + "learning_rate": 5.737317616993782e-09, + "loss": 0.5999, + "step": 22757 + }, + { + "epoch": 1.9393268001704302, + "grad_norm": 23.0, + "learning_rate": 5.721247724449624e-09, + "loss": 1.0947, + "step": 22758 + }, + { + "epoch": 1.9394120153387302, + "grad_norm": 13.8125, + "learning_rate": 5.705200317273896e-09, + "loss": 0.3761, + "step": 22759 + }, + { + "epoch": 1.9394972305070302, + "grad_norm": 13.9375, + "learning_rate": 5.6891753957565024e-09, + "loss": 0.5121, + "step": 22760 + }, + { + "epoch": 1.9395824456753301, + "grad_norm": 14.8125, + "learning_rate": 5.673172960186796e-09, + "loss": 0.6826, + "step": 22761 + }, + { + "epoch": 1.93966766084363, + "grad_norm": 23.5, + "learning_rate": 5.6571930108541285e-09, + "loss": 1.0396, + "step": 22762 + }, + { + "epoch": 1.93975287601193, + "grad_norm": 21.5, + "learning_rate": 5.641235548047297e-09, + "loss": 1.1264, + "step": 22763 + }, + { + "epoch": 1.93983809118023, + "grad_norm": 12.25, + "learning_rate": 5.625300572054681e-09, + "loss": 0.35, + "step": 22764 + }, + { + "epoch": 1.93992330634853, + "grad_norm": 18.0, + "learning_rate": 5.609388083164108e-09, + "loss": 1.0244, + "step": 22765 + }, + { + "epoch": 1.94000852151683, + "grad_norm": 16.375, + "learning_rate": 5.593498081663262e-09, + "loss": 0.5349, + "step": 22766 + }, + { + "epoch": 1.9400937366851299, + "grad_norm": 23.125, + "learning_rate": 5.577630567839276e-09, + "loss": 0.874, + "step": 22767 + }, + { + "epoch": 1.9401789518534298, + "grad_norm": 14.8125, + "learning_rate": 5.561785541978865e-09, + "loss": 0.5889, + "step": 22768 + }, + { + "epoch": 1.9402641670217298, + "grad_norm": 15.25, + "learning_rate": 5.545963004368326e-09, + "loss": 0.3875, + "step": 22769 + }, + { + "epoch": 1.9403493821900297, + "grad_norm": 14.6875, + "learning_rate": 5.530162955293683e-09, + "loss": 0.5834, + "step": 22770 + }, + { + "epoch": 1.9404345973583297, + "grad_norm": 22.25, + "learning_rate": 5.514385395040401e-09, + "loss": 0.9507, + "step": 22771 + }, + { + "epoch": 1.9405198125266296, + "grad_norm": 11.125, + "learning_rate": 5.498630323893667e-09, + "loss": 0.2734, + "step": 22772 + }, + { + "epoch": 1.9406050276949296, + "grad_norm": 19.75, + "learning_rate": 5.482897742137977e-09, + "loss": 0.6552, + "step": 22773 + }, + { + "epoch": 1.9406902428632296, + "grad_norm": 14.625, + "learning_rate": 5.467187650057826e-09, + "loss": 0.5363, + "step": 22774 + }, + { + "epoch": 1.9407754580315295, + "grad_norm": 16.125, + "learning_rate": 5.451500047937153e-09, + "loss": 0.6275, + "step": 22775 + }, + { + "epoch": 1.9408606731998295, + "grad_norm": 18.75, + "learning_rate": 5.4358349360593425e-09, + "loss": 0.7074, + "step": 22776 + }, + { + "epoch": 1.9409458883681294, + "grad_norm": 14.1875, + "learning_rate": 5.420192314707501e-09, + "loss": 0.369, + "step": 22777 + }, + { + "epoch": 1.9410311035364294, + "grad_norm": 13.625, + "learning_rate": 5.404572184164181e-09, + "loss": 0.558, + "step": 22778 + }, + { + "epoch": 1.9411163187047293, + "grad_norm": 18.375, + "learning_rate": 5.388974544711933e-09, + "loss": 0.7846, + "step": 22779 + }, + { + "epoch": 1.9412015338730293, + "grad_norm": 24.125, + "learning_rate": 5.3733993966323396e-09, + "loss": 0.8802, + "step": 22780 + }, + { + "epoch": 1.9412867490413293, + "grad_norm": 11.8125, + "learning_rate": 5.357846740207118e-09, + "loss": 0.5686, + "step": 22781 + }, + { + "epoch": 1.9413719642096292, + "grad_norm": 15.25, + "learning_rate": 5.342316575717016e-09, + "loss": 0.6471, + "step": 22782 + }, + { + "epoch": 1.9414571793779292, + "grad_norm": 14.75, + "learning_rate": 5.326808903443059e-09, + "loss": 0.5192, + "step": 22783 + }, + { + "epoch": 1.9415423945462291, + "grad_norm": 13.5625, + "learning_rate": 5.311323723665163e-09, + "loss": 0.4873, + "step": 22784 + }, + { + "epoch": 1.941627609714529, + "grad_norm": 15.8125, + "learning_rate": 5.29586103666338e-09, + "loss": 0.6752, + "step": 22785 + }, + { + "epoch": 1.941712824882829, + "grad_norm": 14.375, + "learning_rate": 5.280420842716932e-09, + "loss": 0.6036, + "step": 22786 + }, + { + "epoch": 1.941798040051129, + "grad_norm": 14.5, + "learning_rate": 5.2650031421049015e-09, + "loss": 0.5152, + "step": 22787 + }, + { + "epoch": 1.941883255219429, + "grad_norm": 17.5, + "learning_rate": 5.2496079351059544e-09, + "loss": 0.691, + "step": 22788 + }, + { + "epoch": 1.941968470387729, + "grad_norm": 12.25, + "learning_rate": 5.234235221998479e-09, + "loss": 0.4721, + "step": 22789 + }, + { + "epoch": 1.9420536855560289, + "grad_norm": 12.8125, + "learning_rate": 5.2188850030598925e-09, + "loss": 0.374, + "step": 22790 + }, + { + "epoch": 1.9421389007243288, + "grad_norm": 14.5625, + "learning_rate": 5.2035572785677505e-09, + "loss": 0.4807, + "step": 22791 + }, + { + "epoch": 1.9422241158926288, + "grad_norm": 25.0, + "learning_rate": 5.188252048799192e-09, + "loss": 0.554, + "step": 22792 + }, + { + "epoch": 1.9423093310609287, + "grad_norm": 13.1875, + "learning_rate": 5.172969314030663e-09, + "loss": 0.4729, + "step": 22793 + }, + { + "epoch": 1.9423945462292287, + "grad_norm": 20.125, + "learning_rate": 5.157709074538331e-09, + "loss": 0.4669, + "step": 22794 + }, + { + "epoch": 1.9424797613975286, + "grad_norm": 10.8125, + "learning_rate": 5.142471330597948e-09, + "loss": 0.5353, + "step": 22795 + }, + { + "epoch": 1.9425649765658286, + "grad_norm": 10.5625, + "learning_rate": 5.127256082484849e-09, + "loss": 0.3437, + "step": 22796 + }, + { + "epoch": 1.9426501917341286, + "grad_norm": 16.25, + "learning_rate": 5.112063330474232e-09, + "loss": 0.5875, + "step": 22797 + }, + { + "epoch": 1.9427354069024285, + "grad_norm": 13.8125, + "learning_rate": 5.09689307484032e-09, + "loss": 0.5612, + "step": 22798 + }, + { + "epoch": 1.9428206220707285, + "grad_norm": 14.9375, + "learning_rate": 5.08174531585734e-09, + "loss": 0.644, + "step": 22799 + }, + { + "epoch": 1.9429058372390284, + "grad_norm": 12.1875, + "learning_rate": 5.066620053799237e-09, + "loss": 0.5222, + "step": 22800 + }, + { + "epoch": 1.9429910524073284, + "grad_norm": 22.125, + "learning_rate": 5.051517288939128e-09, + "loss": 0.7805, + "step": 22801 + }, + { + "epoch": 1.9430762675756283, + "grad_norm": 15.375, + "learning_rate": 5.036437021549989e-09, + "loss": 0.4743, + "step": 22802 + }, + { + "epoch": 1.9431614827439283, + "grad_norm": 12.8125, + "learning_rate": 5.021379251904379e-09, + "loss": 0.508, + "step": 22803 + }, + { + "epoch": 1.9432466979122283, + "grad_norm": 15.6875, + "learning_rate": 5.006343980274303e-09, + "loss": 0.6215, + "step": 22804 + }, + { + "epoch": 1.9433319130805282, + "grad_norm": 11.3125, + "learning_rate": 4.991331206931488e-09, + "loss": 0.392, + "step": 22805 + }, + { + "epoch": 1.9434171282488282, + "grad_norm": 19.125, + "learning_rate": 4.976340932147383e-09, + "loss": 0.6619, + "step": 22806 + }, + { + "epoch": 1.9435023434171281, + "grad_norm": 15.4375, + "learning_rate": 4.961373156192745e-09, + "loss": 0.6131, + "step": 22807 + }, + { + "epoch": 1.943587558585428, + "grad_norm": 25.125, + "learning_rate": 4.946427879338051e-09, + "loss": 0.8166, + "step": 22808 + }, + { + "epoch": 1.943672773753728, + "grad_norm": 11.625, + "learning_rate": 4.931505101853501e-09, + "loss": 0.415, + "step": 22809 + }, + { + "epoch": 1.943757988922028, + "grad_norm": 16.5, + "learning_rate": 4.916604824008603e-09, + "loss": 0.7277, + "step": 22810 + }, + { + "epoch": 1.943843204090328, + "grad_norm": 13.8125, + "learning_rate": 4.901727046072585e-09, + "loss": 0.822, + "step": 22811 + }, + { + "epoch": 1.943928419258628, + "grad_norm": 13.625, + "learning_rate": 4.8868717683144e-09, + "loss": 0.3708, + "step": 22812 + }, + { + "epoch": 1.9440136344269279, + "grad_norm": 12.4375, + "learning_rate": 4.872038991002581e-09, + "loss": 0.4393, + "step": 22813 + }, + { + "epoch": 1.9440988495952278, + "grad_norm": 19.75, + "learning_rate": 4.85722871440511e-09, + "loss": 0.7644, + "step": 22814 + }, + { + "epoch": 1.9441840647635278, + "grad_norm": 15.1875, + "learning_rate": 4.84244093878955e-09, + "loss": 0.5745, + "step": 22815 + }, + { + "epoch": 1.9442692799318277, + "grad_norm": 14.3125, + "learning_rate": 4.827675664423187e-09, + "loss": 0.3719, + "step": 22816 + }, + { + "epoch": 1.9443544951001277, + "grad_norm": 18.625, + "learning_rate": 4.8129328915728915e-09, + "loss": 0.6716, + "step": 22817 + }, + { + "epoch": 1.9444397102684277, + "grad_norm": 24.0, + "learning_rate": 4.798212620505116e-09, + "loss": 0.6655, + "step": 22818 + }, + { + "epoch": 1.9445249254367276, + "grad_norm": 18.0, + "learning_rate": 4.783514851485621e-09, + "loss": 0.3232, + "step": 22819 + }, + { + "epoch": 1.9446101406050276, + "grad_norm": 16.0, + "learning_rate": 4.768839584780305e-09, + "loss": 0.7177, + "step": 22820 + }, + { + "epoch": 1.9446953557733275, + "grad_norm": 13.875, + "learning_rate": 4.754186820654094e-09, + "loss": 0.6823, + "step": 22821 + }, + { + "epoch": 1.9447805709416275, + "grad_norm": 21.125, + "learning_rate": 4.7395565593719165e-09, + "loss": 0.948, + "step": 22822 + }, + { + "epoch": 1.9448657861099274, + "grad_norm": 21.75, + "learning_rate": 4.7249488011982815e-09, + "loss": 0.479, + "step": 22823 + }, + { + "epoch": 1.9449510012782274, + "grad_norm": 24.375, + "learning_rate": 4.710363546397006e-09, + "loss": 0.4929, + "step": 22824 + }, + { + "epoch": 1.9450362164465274, + "grad_norm": 15.1875, + "learning_rate": 4.695800795231631e-09, + "loss": 0.414, + "step": 22825 + }, + { + "epoch": 1.9451214316148273, + "grad_norm": 26.0, + "learning_rate": 4.681260547965416e-09, + "loss": 0.9006, + "step": 22826 + }, + { + "epoch": 1.9452066467831273, + "grad_norm": 14.8125, + "learning_rate": 4.666742804861208e-09, + "loss": 0.3901, + "step": 22827 + }, + { + "epoch": 1.9452918619514272, + "grad_norm": 14.375, + "learning_rate": 4.652247566181156e-09, + "loss": 0.4674, + "step": 22828 + }, + { + "epoch": 1.9453770771197272, + "grad_norm": 19.625, + "learning_rate": 4.637774832187136e-09, + "loss": 0.4595, + "step": 22829 + }, + { + "epoch": 1.9454622922880271, + "grad_norm": 13.8125, + "learning_rate": 4.623324603141022e-09, + "loss": 0.5765, + "step": 22830 + }, + { + "epoch": 1.945547507456327, + "grad_norm": 14.0, + "learning_rate": 4.608896879303715e-09, + "loss": 0.4858, + "step": 22831 + }, + { + "epoch": 1.945632722624627, + "grad_norm": 15.8125, + "learning_rate": 4.594491660935979e-09, + "loss": 0.7352, + "step": 22832 + }, + { + "epoch": 1.945717937792927, + "grad_norm": 13.0625, + "learning_rate": 4.580108948298162e-09, + "loss": 0.4173, + "step": 22833 + }, + { + "epoch": 1.945803152961227, + "grad_norm": 10.8125, + "learning_rate": 4.565748741650055e-09, + "loss": 0.3293, + "step": 22834 + }, + { + "epoch": 1.945888368129527, + "grad_norm": 23.375, + "learning_rate": 4.5514110412511746e-09, + "loss": 0.5904, + "step": 22835 + }, + { + "epoch": 1.9459735832978269, + "grad_norm": 12.5, + "learning_rate": 4.537095847360895e-09, + "loss": 0.4584, + "step": 22836 + }, + { + "epoch": 1.9460587984661268, + "grad_norm": 17.25, + "learning_rate": 4.522803160237621e-09, + "loss": 0.627, + "step": 22837 + }, + { + "epoch": 1.9461440136344268, + "grad_norm": 16.125, + "learning_rate": 4.5085329801396186e-09, + "loss": 0.7121, + "step": 22838 + }, + { + "epoch": 1.9462292288027268, + "grad_norm": 14.375, + "learning_rate": 4.494285307325014e-09, + "loss": 0.5242, + "step": 22839 + }, + { + "epoch": 1.9463144439710267, + "grad_norm": 12.75, + "learning_rate": 4.480060142051102e-09, + "loss": 0.5207, + "step": 22840 + }, + { + "epoch": 1.9463996591393267, + "grad_norm": 17.125, + "learning_rate": 4.465857484574898e-09, + "loss": 0.6064, + "step": 22841 + }, + { + "epoch": 1.9464848743076266, + "grad_norm": 14.5, + "learning_rate": 4.4516773351531425e-09, + "loss": 0.5086, + "step": 22842 + }, + { + "epoch": 1.9465700894759266, + "grad_norm": 20.5, + "learning_rate": 4.4375196940421585e-09, + "loss": 0.4294, + "step": 22843 + }, + { + "epoch": 1.9466553046442265, + "grad_norm": 15.375, + "learning_rate": 4.423384561497573e-09, + "loss": 0.5493, + "step": 22844 + }, + { + "epoch": 1.9467405198125265, + "grad_norm": 16.125, + "learning_rate": 4.409271937775017e-09, + "loss": 0.3998, + "step": 22845 + }, + { + "epoch": 1.9468257349808265, + "grad_norm": 20.0, + "learning_rate": 4.395181823129424e-09, + "loss": 0.9498, + "step": 22846 + }, + { + "epoch": 1.9469109501491264, + "grad_norm": 13.9375, + "learning_rate": 4.381114217815452e-09, + "loss": 0.633, + "step": 22847 + }, + { + "epoch": 1.9469961653174264, + "grad_norm": 16.5, + "learning_rate": 4.367069122087481e-09, + "loss": 0.4816, + "step": 22848 + }, + { + "epoch": 1.9470813804857263, + "grad_norm": 23.375, + "learning_rate": 4.353046536199057e-09, + "loss": 0.871, + "step": 22849 + }, + { + "epoch": 1.9471665956540263, + "grad_norm": 17.25, + "learning_rate": 4.3390464604035895e-09, + "loss": 0.7952, + "step": 22850 + }, + { + "epoch": 1.9472518108223262, + "grad_norm": 12.5625, + "learning_rate": 4.325068894954209e-09, + "loss": 0.4503, + "step": 22851 + }, + { + "epoch": 1.9473370259906262, + "grad_norm": 13.75, + "learning_rate": 4.3111138401033535e-09, + "loss": 0.7788, + "step": 22852 + }, + { + "epoch": 1.9474222411589261, + "grad_norm": 18.125, + "learning_rate": 4.2971812961034584e-09, + "loss": 0.6312, + "step": 22853 + }, + { + "epoch": 1.947507456327226, + "grad_norm": 17.125, + "learning_rate": 4.283271263206129e-09, + "loss": 0.4665, + "step": 22854 + }, + { + "epoch": 1.947592671495526, + "grad_norm": 15.5625, + "learning_rate": 4.269383741662692e-09, + "loss": 0.4937, + "step": 22855 + }, + { + "epoch": 1.947677886663826, + "grad_norm": 13.3125, + "learning_rate": 4.255518731724195e-09, + "loss": 0.4926, + "step": 22856 + }, + { + "epoch": 1.947763101832126, + "grad_norm": 14.875, + "learning_rate": 4.241676233641273e-09, + "loss": 0.5833, + "step": 22857 + }, + { + "epoch": 1.947848317000426, + "grad_norm": 17.75, + "learning_rate": 4.227856247663864e-09, + "loss": 0.7203, + "step": 22858 + }, + { + "epoch": 1.947933532168726, + "grad_norm": 14.3125, + "learning_rate": 4.214058774041768e-09, + "loss": 0.5708, + "step": 22859 + }, + { + "epoch": 1.9480187473370258, + "grad_norm": 11.5, + "learning_rate": 4.20028381302437e-09, + "loss": 0.6724, + "step": 22860 + }, + { + "epoch": 1.9481039625053258, + "grad_norm": 14.3125, + "learning_rate": 4.1865313648606356e-09, + "loss": 0.5599, + "step": 22861 + }, + { + "epoch": 1.9481891776736258, + "grad_norm": 13.5, + "learning_rate": 4.172801429799117e-09, + "loss": 0.586, + "step": 22862 + }, + { + "epoch": 1.9482743928419257, + "grad_norm": 15.0625, + "learning_rate": 4.159094008087672e-09, + "loss": 0.5967, + "step": 22863 + }, + { + "epoch": 1.9483596080102257, + "grad_norm": 14.875, + "learning_rate": 4.145409099974296e-09, + "loss": 0.4505, + "step": 22864 + }, + { + "epoch": 1.9484448231785256, + "grad_norm": 15.6875, + "learning_rate": 4.131746705706291e-09, + "loss": 0.5463, + "step": 22865 + }, + { + "epoch": 1.9485300383468256, + "grad_norm": 13.0625, + "learning_rate": 4.1181068255302656e-09, + "loss": 0.317, + "step": 22866 + }, + { + "epoch": 1.9486152535151255, + "grad_norm": 15.3125, + "learning_rate": 4.104489459692967e-09, + "loss": 0.7845, + "step": 22867 + }, + { + "epoch": 1.9487004686834255, + "grad_norm": 15.6875, + "learning_rate": 4.09089460844031e-09, + "loss": 0.7015, + "step": 22868 + }, + { + "epoch": 1.9487856838517255, + "grad_norm": 17.0, + "learning_rate": 4.077322272018208e-09, + "loss": 0.5473, + "step": 22869 + }, + { + "epoch": 1.9488708990200254, + "grad_norm": 14.75, + "learning_rate": 4.063772450671605e-09, + "loss": 0.373, + "step": 22870 + }, + { + "epoch": 1.9489561141883254, + "grad_norm": 15.1875, + "learning_rate": 4.0502451446455816e-09, + "loss": 0.4812, + "step": 22871 + }, + { + "epoch": 1.9490413293566253, + "grad_norm": 15.6875, + "learning_rate": 4.036740354184526e-09, + "loss": 0.6789, + "step": 22872 + }, + { + "epoch": 1.9491265445249253, + "grad_norm": 17.625, + "learning_rate": 4.02325807953241e-09, + "loss": 0.7824, + "step": 22873 + }, + { + "epoch": 1.9492117596932252, + "grad_norm": 14.75, + "learning_rate": 4.009798320933067e-09, + "loss": 0.504, + "step": 22874 + }, + { + "epoch": 1.9492969748615252, + "grad_norm": 15.6875, + "learning_rate": 3.996361078629496e-09, + "loss": 0.5327, + "step": 22875 + }, + { + "epoch": 1.9493821900298252, + "grad_norm": 13.125, + "learning_rate": 3.982946352864697e-09, + "loss": 0.3957, + "step": 22876 + }, + { + "epoch": 1.9494674051981251, + "grad_norm": 15.625, + "learning_rate": 3.969554143881116e-09, + "loss": 0.5723, + "step": 22877 + }, + { + "epoch": 1.949552620366425, + "grad_norm": 15.625, + "learning_rate": 3.9561844519205025e-09, + "loss": 0.9371, + "step": 22878 + }, + { + "epoch": 1.949637835534725, + "grad_norm": 17.75, + "learning_rate": 3.942837277224748e-09, + "loss": 0.7208, + "step": 22879 + }, + { + "epoch": 1.949723050703025, + "grad_norm": 13.75, + "learning_rate": 3.929512620034909e-09, + "loss": 0.4366, + "step": 22880 + }, + { + "epoch": 1.949808265871325, + "grad_norm": 20.125, + "learning_rate": 3.916210480591765e-09, + "loss": 0.5498, + "step": 22881 + }, + { + "epoch": 1.949893481039625, + "grad_norm": 17.125, + "learning_rate": 3.9029308591356786e-09, + "loss": 0.82, + "step": 22882 + }, + { + "epoch": 1.9499786962079249, + "grad_norm": 14.125, + "learning_rate": 3.889673755906875e-09, + "loss": 0.6988, + "step": 22883 + }, + { + "epoch": 1.9500639113762248, + "grad_norm": 21.5, + "learning_rate": 3.876439171144469e-09, + "loss": 0.8951, + "step": 22884 + }, + { + "epoch": 1.9501491265445248, + "grad_norm": 19.625, + "learning_rate": 3.863227105087991e-09, + "loss": 0.7829, + "step": 22885 + }, + { + "epoch": 1.9502343417128247, + "grad_norm": 12.0625, + "learning_rate": 3.850037557976139e-09, + "loss": 0.499, + "step": 22886 + }, + { + "epoch": 1.9503195568811247, + "grad_norm": 16.25, + "learning_rate": 3.8368705300471934e-09, + "loss": 0.6692, + "step": 22887 + }, + { + "epoch": 1.9504047720494246, + "grad_norm": 13.875, + "learning_rate": 3.82372602153902e-09, + "loss": 0.5806, + "step": 22888 + }, + { + "epoch": 1.9504899872177246, + "grad_norm": 15.3125, + "learning_rate": 3.810604032689347e-09, + "loss": 0.7795, + "step": 22889 + }, + { + "epoch": 1.9505752023860246, + "grad_norm": 18.0, + "learning_rate": 3.797504563735205e-09, + "loss": 0.6544, + "step": 22890 + }, + { + "epoch": 1.9506604175543245, + "grad_norm": 22.75, + "learning_rate": 3.784427614913349e-09, + "loss": 0.8243, + "step": 22891 + }, + { + "epoch": 1.9507456327226245, + "grad_norm": 24.0, + "learning_rate": 3.771373186459981e-09, + "loss": 0.8327, + "step": 22892 + }, + { + "epoch": 1.9508308478909244, + "grad_norm": 14.1875, + "learning_rate": 3.758341278611022e-09, + "loss": 0.8261, + "step": 22893 + }, + { + "epoch": 1.9509160630592244, + "grad_norm": 16.5, + "learning_rate": 3.745331891602116e-09, + "loss": 0.6354, + "step": 22894 + }, + { + "epoch": 1.9510012782275243, + "grad_norm": 10.5625, + "learning_rate": 3.732345025668216e-09, + "loss": 0.3424, + "step": 22895 + }, + { + "epoch": 1.9510864933958243, + "grad_norm": 16.375, + "learning_rate": 3.7193806810439935e-09, + "loss": 0.5841, + "step": 22896 + }, + { + "epoch": 1.9511717085641243, + "grad_norm": 11.75, + "learning_rate": 3.7064388579639844e-09, + "loss": 0.3738, + "step": 22897 + }, + { + "epoch": 1.9512569237324242, + "grad_norm": 33.0, + "learning_rate": 3.693519556661751e-09, + "loss": 0.5823, + "step": 22898 + }, + { + "epoch": 1.9513421389007242, + "grad_norm": 19.375, + "learning_rate": 3.680622777370857e-09, + "loss": 0.67, + "step": 22899 + }, + { + "epoch": 1.9514273540690241, + "grad_norm": 12.3125, + "learning_rate": 3.667748520324449e-09, + "loss": 0.3177, + "step": 22900 + }, + { + "epoch": 1.951512569237324, + "grad_norm": 16.875, + "learning_rate": 3.6548967857551186e-09, + "loss": 0.8116, + "step": 22901 + }, + { + "epoch": 1.951597784405624, + "grad_norm": 15.0, + "learning_rate": 3.642067573895042e-09, + "loss": 0.8315, + "step": 22902 + }, + { + "epoch": 1.951682999573924, + "grad_norm": 10.1875, + "learning_rate": 3.6292608849761156e-09, + "loss": 0.2401, + "step": 22903 + }, + { + "epoch": 1.951768214742224, + "grad_norm": 19.75, + "learning_rate": 3.6164767192298222e-09, + "loss": 0.7365, + "step": 22904 + }, + { + "epoch": 1.951853429910524, + "grad_norm": 17.0, + "learning_rate": 3.603715076886949e-09, + "loss": 0.5623, + "step": 22905 + }, + { + "epoch": 1.951938645078824, + "grad_norm": 13.0, + "learning_rate": 3.590975958178561e-09, + "loss": 0.5289, + "step": 22906 + }, + { + "epoch": 1.952023860247124, + "grad_norm": 14.25, + "learning_rate": 3.5782593633344743e-09, + "loss": 0.344, + "step": 22907 + }, + { + "epoch": 1.952109075415424, + "grad_norm": 11.75, + "learning_rate": 3.565565292584644e-09, + "loss": 0.4612, + "step": 22908 + }, + { + "epoch": 1.952194290583724, + "grad_norm": 18.625, + "learning_rate": 3.5528937461583314e-09, + "loss": 0.6618, + "step": 22909 + }, + { + "epoch": 1.952279505752024, + "grad_norm": 24.75, + "learning_rate": 3.5402447242846584e-09, + "loss": 0.3983, + "step": 22910 + }, + { + "epoch": 1.9523647209203239, + "grad_norm": 12.375, + "learning_rate": 3.527618227192192e-09, + "loss": 0.5091, + "step": 22911 + }, + { + "epoch": 1.9524499360886238, + "grad_norm": 11.0625, + "learning_rate": 3.5150142551092225e-09, + "loss": 0.3522, + "step": 22912 + }, + { + "epoch": 1.9525351512569238, + "grad_norm": 15.75, + "learning_rate": 3.5024328082632063e-09, + "loss": 0.8253, + "step": 22913 + }, + { + "epoch": 1.9526203664252237, + "grad_norm": 16.25, + "learning_rate": 3.4898738868817394e-09, + "loss": 0.5758, + "step": 22914 + }, + { + "epoch": 1.9527055815935237, + "grad_norm": 16.125, + "learning_rate": 3.4773374911917233e-09, + "loss": 0.6468, + "step": 22915 + }, + { + "epoch": 1.9527907967618237, + "grad_norm": 13.5625, + "learning_rate": 3.4648236214197828e-09, + "loss": 0.4465, + "step": 22916 + }, + { + "epoch": 1.9528760119301236, + "grad_norm": 11.4375, + "learning_rate": 3.4523322777918476e-09, + "loss": 0.4035, + "step": 22917 + }, + { + "epoch": 1.9529612270984236, + "grad_norm": 17.125, + "learning_rate": 3.43986346053371e-09, + "loss": 0.7947, + "step": 22918 + }, + { + "epoch": 1.9530464422667235, + "grad_norm": 14.6875, + "learning_rate": 3.4274171698708835e-09, + "loss": 0.5772, + "step": 22919 + }, + { + "epoch": 1.9531316574350235, + "grad_norm": 17.125, + "learning_rate": 3.4149934060281886e-09, + "loss": 0.7187, + "step": 22920 + }, + { + "epoch": 1.9532168726033234, + "grad_norm": 20.375, + "learning_rate": 3.402592169230029e-09, + "loss": 0.6684, + "step": 22921 + }, + { + "epoch": 1.9533020877716234, + "grad_norm": 19.375, + "learning_rate": 3.390213459700531e-09, + "loss": 0.6214, + "step": 22922 + }, + { + "epoch": 1.9533873029399234, + "grad_norm": 21.25, + "learning_rate": 3.377857277663543e-09, + "loss": 1.0383, + "step": 22923 + }, + { + "epoch": 1.9534725181082233, + "grad_norm": 13.1875, + "learning_rate": 3.365523623342082e-09, + "loss": 0.6702, + "step": 22924 + }, + { + "epoch": 1.9535577332765233, + "grad_norm": 13.25, + "learning_rate": 3.3532124969593017e-09, + "loss": 0.4363, + "step": 22925 + }, + { + "epoch": 1.9536429484448232, + "grad_norm": 27.5, + "learning_rate": 3.3409238987376637e-09, + "loss": 1.3392, + "step": 22926 + }, + { + "epoch": 1.9537281636131232, + "grad_norm": 32.0, + "learning_rate": 3.3286578288989345e-09, + "loss": 1.1013, + "step": 22927 + }, + { + "epoch": 1.9538133787814231, + "grad_norm": 15.625, + "learning_rate": 3.3164142876651595e-09, + "loss": 0.6894, + "step": 22928 + }, + { + "epoch": 1.953898593949723, + "grad_norm": 13.75, + "learning_rate": 3.3041932752574113e-09, + "loss": 0.3678, + "step": 22929 + }, + { + "epoch": 1.953983809118023, + "grad_norm": 13.5, + "learning_rate": 3.2919947918964856e-09, + "loss": 0.5912, + "step": 22930 + }, + { + "epoch": 1.954069024286323, + "grad_norm": 11.625, + "learning_rate": 3.2798188378029005e-09, + "loss": 0.3738, + "step": 22931 + }, + { + "epoch": 1.954154239454623, + "grad_norm": 15.0, + "learning_rate": 3.267665413196619e-09, + "loss": 0.5846, + "step": 22932 + }, + { + "epoch": 1.954239454622923, + "grad_norm": 12.875, + "learning_rate": 3.2555345182973266e-09, + "loss": 0.4001, + "step": 22933 + }, + { + "epoch": 1.9543246697912229, + "grad_norm": 13.5, + "learning_rate": 3.2434261533242915e-09, + "loss": 0.4544, + "step": 22934 + }, + { + "epoch": 1.9544098849595228, + "grad_norm": 13.875, + "learning_rate": 3.23134031849609e-09, + "loss": 0.4502, + "step": 22935 + }, + { + "epoch": 1.9544951001278228, + "grad_norm": 21.25, + "learning_rate": 3.2192770140314357e-09, + "loss": 1.0115, + "step": 22936 + }, + { + "epoch": 1.9545803152961228, + "grad_norm": 10.875, + "learning_rate": 3.2072362401482092e-09, + "loss": 0.3436, + "step": 22937 + }, + { + "epoch": 1.9546655304644227, + "grad_norm": 11.375, + "learning_rate": 3.1952179970638763e-09, + "loss": 0.292, + "step": 22938 + }, + { + "epoch": 1.9547507456327227, + "grad_norm": 17.5, + "learning_rate": 3.1832222849957626e-09, + "loss": 0.6814, + "step": 22939 + }, + { + "epoch": 1.9548359608010226, + "grad_norm": 14.9375, + "learning_rate": 3.171249104160501e-09, + "loss": 0.7186, + "step": 22940 + }, + { + "epoch": 1.9549211759693226, + "grad_norm": 13.875, + "learning_rate": 3.159298454774584e-09, + "loss": 0.3969, + "step": 22941 + }, + { + "epoch": 1.9550063911376225, + "grad_norm": 19.5, + "learning_rate": 3.1473703370540896e-09, + "loss": 0.5557, + "step": 22942 + }, + { + "epoch": 1.9550916063059225, + "grad_norm": 15.9375, + "learning_rate": 3.1354647512142624e-09, + "loss": 0.7334, + "step": 22943 + }, + { + "epoch": 1.9551768214742224, + "grad_norm": 37.0, + "learning_rate": 3.123581697470346e-09, + "loss": 1.1417, + "step": 22944 + }, + { + "epoch": 1.9552620366425224, + "grad_norm": 17.625, + "learning_rate": 3.1117211760373077e-09, + "loss": 0.6734, + "step": 22945 + }, + { + "epoch": 1.9553472518108224, + "grad_norm": 13.375, + "learning_rate": 3.0998831871292823e-09, + "loss": 0.5857, + "step": 22946 + }, + { + "epoch": 1.9554324669791223, + "grad_norm": 17.0, + "learning_rate": 3.0880677309602646e-09, + "loss": 0.5576, + "step": 22947 + }, + { + "epoch": 1.9555176821474223, + "grad_norm": 14.3125, + "learning_rate": 3.0762748077435566e-09, + "loss": 0.7198, + "step": 22948 + }, + { + "epoch": 1.9556028973157222, + "grad_norm": 13.1875, + "learning_rate": 3.0645044176925985e-09, + "loss": 0.5303, + "step": 22949 + }, + { + "epoch": 1.9556881124840222, + "grad_norm": 14.5, + "learning_rate": 3.0527565610198596e-09, + "loss": 0.5773, + "step": 22950 + }, + { + "epoch": 1.9557733276523221, + "grad_norm": 12.0625, + "learning_rate": 3.0410312379376695e-09, + "loss": 0.3299, + "step": 22951 + }, + { + "epoch": 1.955858542820622, + "grad_norm": 14.125, + "learning_rate": 3.029328448658081e-09, + "loss": 0.576, + "step": 22952 + }, + { + "epoch": 1.955943757988922, + "grad_norm": 16.25, + "learning_rate": 3.0176481933923142e-09, + "loss": 0.8553, + "step": 22953 + }, + { + "epoch": 1.956028973157222, + "grad_norm": 17.25, + "learning_rate": 3.0059904723517275e-09, + "loss": 0.698, + "step": 22954 + }, + { + "epoch": 1.956114188325522, + "grad_norm": 14.9375, + "learning_rate": 2.9943552857468473e-09, + "loss": 0.6108, + "step": 22955 + }, + { + "epoch": 1.956199403493822, + "grad_norm": 15.75, + "learning_rate": 2.9827426337877828e-09, + "loss": 0.5805, + "step": 22956 + }, + { + "epoch": 1.9562846186621219, + "grad_norm": 14.375, + "learning_rate": 2.9711525166846445e-09, + "loss": 0.6658, + "step": 22957 + }, + { + "epoch": 1.9563698338304218, + "grad_norm": 17.125, + "learning_rate": 2.9595849346467087e-09, + "loss": 0.4852, + "step": 22958 + }, + { + "epoch": 1.9564550489987218, + "grad_norm": 17.625, + "learning_rate": 2.948039887883114e-09, + "loss": 0.7131, + "step": 22959 + }, + { + "epoch": 1.9565402641670218, + "grad_norm": 19.25, + "learning_rate": 2.9365173766023048e-09, + "loss": 0.8092, + "step": 22960 + }, + { + "epoch": 1.9566254793353217, + "grad_norm": 21.375, + "learning_rate": 2.9250174010127254e-09, + "loss": 0.8703, + "step": 22961 + }, + { + "epoch": 1.9567106945036217, + "grad_norm": 13.5, + "learning_rate": 2.913539961322126e-09, + "loss": 0.596, + "step": 22962 + }, + { + "epoch": 1.9567959096719216, + "grad_norm": 14.1875, + "learning_rate": 2.90208505773798e-09, + "loss": 0.465, + "step": 22963 + }, + { + "epoch": 1.9568811248402216, + "grad_norm": 14.75, + "learning_rate": 2.890652690467205e-09, + "loss": 0.5611, + "step": 22964 + }, + { + "epoch": 1.9569663400085215, + "grad_norm": 17.625, + "learning_rate": 2.8792428597163023e-09, + "loss": 0.7974, + "step": 22965 + }, + { + "epoch": 1.9570515551768215, + "grad_norm": 14.875, + "learning_rate": 2.867855565691635e-09, + "loss": 0.7436, + "step": 22966 + }, + { + "epoch": 1.9571367703451215, + "grad_norm": 13.375, + "learning_rate": 2.856490808598872e-09, + "loss": 0.4579, + "step": 22967 + }, + { + "epoch": 1.9572219855134214, + "grad_norm": 19.375, + "learning_rate": 2.845148588643404e-09, + "loss": 0.9404, + "step": 22968 + }, + { + "epoch": 1.9573072006817214, + "grad_norm": 10.5, + "learning_rate": 2.833828906030206e-09, + "loss": 0.3313, + "step": 22969 + }, + { + "epoch": 1.9573924158500213, + "grad_norm": 14.25, + "learning_rate": 2.822531760963837e-09, + "loss": 0.578, + "step": 22970 + }, + { + "epoch": 1.9574776310183213, + "grad_norm": 14.375, + "learning_rate": 2.811257153648578e-09, + "loss": 0.5709, + "step": 22971 + }, + { + "epoch": 1.9575628461866212, + "grad_norm": 16.625, + "learning_rate": 2.8000050842880155e-09, + "loss": 0.639, + "step": 22972 + }, + { + "epoch": 1.9576480613549212, + "grad_norm": 11.75, + "learning_rate": 2.788775553085321e-09, + "loss": 0.3511, + "step": 22973 + }, + { + "epoch": 1.9577332765232212, + "grad_norm": 19.0, + "learning_rate": 2.777568560243804e-09, + "loss": 1.2777, + "step": 22974 + }, + { + "epoch": 1.9578184916915211, + "grad_norm": 15.25, + "learning_rate": 2.7663841059658025e-09, + "loss": 0.5507, + "step": 22975 + }, + { + "epoch": 1.957903706859821, + "grad_norm": 18.0, + "learning_rate": 2.7552221904533772e-09, + "loss": 0.5241, + "step": 22976 + }, + { + "epoch": 1.957988922028121, + "grad_norm": 10.0, + "learning_rate": 2.7440828139083107e-09, + "loss": 0.24, + "step": 22977 + }, + { + "epoch": 1.958074137196421, + "grad_norm": 17.375, + "learning_rate": 2.7329659765318317e-09, + "loss": 0.5549, + "step": 22978 + }, + { + "epoch": 1.958159352364721, + "grad_norm": 11.5625, + "learning_rate": 2.72187167852489e-09, + "loss": 0.4907, + "step": 22979 + }, + { + "epoch": 1.958244567533021, + "grad_norm": 15.8125, + "learning_rate": 2.7107999200880197e-09, + "loss": 0.561, + "step": 22980 + }, + { + "epoch": 1.9583297827013209, + "grad_norm": 15.6875, + "learning_rate": 2.6997507014212e-09, + "loss": 0.7476, + "step": 22981 + }, + { + "epoch": 1.9584149978696208, + "grad_norm": 15.5, + "learning_rate": 2.688724022724132e-09, + "loss": 0.6087, + "step": 22982 + }, + { + "epoch": 1.9585002130379208, + "grad_norm": 19.875, + "learning_rate": 2.6777198841961003e-09, + "loss": 0.774, + "step": 22983 + }, + { + "epoch": 1.9585854282062207, + "grad_norm": 13.25, + "learning_rate": 2.6667382860359746e-09, + "loss": 0.582, + "step": 22984 + }, + { + "epoch": 1.9586706433745207, + "grad_norm": 16.25, + "learning_rate": 2.655779228442068e-09, + "loss": 0.5083, + "step": 22985 + }, + { + "epoch": 1.9587558585428206, + "grad_norm": 13.75, + "learning_rate": 2.644842711612694e-09, + "loss": 0.5607, + "step": 22986 + }, + { + "epoch": 1.9588410737111206, + "grad_norm": 14.0625, + "learning_rate": 2.6339287357451947e-09, + "loss": 0.4884, + "step": 22987 + }, + { + "epoch": 1.9589262888794206, + "grad_norm": 15.3125, + "learning_rate": 2.6230373010370513e-09, + "loss": 0.3584, + "step": 22988 + }, + { + "epoch": 1.9590115040477205, + "grad_norm": 26.25, + "learning_rate": 2.612168407684912e-09, + "loss": 0.6188, + "step": 22989 + }, + { + "epoch": 1.9590967192160205, + "grad_norm": 16.125, + "learning_rate": 2.601322055885147e-09, + "loss": 0.633, + "step": 22990 + }, + { + "epoch": 1.9591819343843204, + "grad_norm": 14.75, + "learning_rate": 2.5904982458339888e-09, + "loss": 0.604, + "step": 22991 + }, + { + "epoch": 1.9592671495526204, + "grad_norm": 22.625, + "learning_rate": 2.579696977726698e-09, + "loss": 0.7801, + "step": 22992 + }, + { + "epoch": 1.9593523647209203, + "grad_norm": 15.875, + "learning_rate": 2.568918251758812e-09, + "loss": 0.5949, + "step": 22993 + }, + { + "epoch": 1.9594375798892203, + "grad_norm": 26.375, + "learning_rate": 2.5581620681250364e-09, + "loss": 0.7303, + "step": 22994 + }, + { + "epoch": 1.9595227950575203, + "grad_norm": 13.875, + "learning_rate": 2.5474284270195214e-09, + "loss": 0.4028, + "step": 22995 + }, + { + "epoch": 1.9596080102258202, + "grad_norm": 18.75, + "learning_rate": 2.5367173286364178e-09, + "loss": 1.0238, + "step": 22996 + }, + { + "epoch": 1.9596932253941202, + "grad_norm": 11.8125, + "learning_rate": 2.52602877316932e-09, + "loss": 0.2802, + "step": 22997 + }, + { + "epoch": 1.9597784405624201, + "grad_norm": 14.6875, + "learning_rate": 2.5153627608112683e-09, + "loss": 0.6173, + "step": 22998 + }, + { + "epoch": 1.95986365573072, + "grad_norm": 19.25, + "learning_rate": 2.5047192917550257e-09, + "loss": 0.6868, + "step": 22999 + }, + { + "epoch": 1.95994887089902, + "grad_norm": 13.3125, + "learning_rate": 2.4940983661929375e-09, + "loss": 0.4554, + "step": 23000 + }, + { + "epoch": 1.96003408606732, + "grad_norm": 20.25, + "learning_rate": 2.4834999843170724e-09, + "loss": 0.5386, + "step": 23001 + }, + { + "epoch": 1.96011930123562, + "grad_norm": 12.375, + "learning_rate": 2.472924146318806e-09, + "loss": 0.3945, + "step": 23002 + }, + { + "epoch": 1.96020451640392, + "grad_norm": 15.1875, + "learning_rate": 2.462370852389234e-09, + "loss": 0.6153, + "step": 23003 + }, + { + "epoch": 1.9602897315722199, + "grad_norm": 25.25, + "learning_rate": 2.4518401027190387e-09, + "loss": 0.8293, + "step": 23004 + }, + { + "epoch": 1.9603749467405198, + "grad_norm": 37.25, + "learning_rate": 2.441331897498761e-09, + "loss": 0.7087, + "step": 23005 + }, + { + "epoch": 1.9604601619088198, + "grad_norm": 16.125, + "learning_rate": 2.4308462369181107e-09, + "loss": 0.7311, + "step": 23006 + }, + { + "epoch": 1.9605453770771197, + "grad_norm": 20.375, + "learning_rate": 2.4203831211665196e-09, + "loss": 0.7735, + "step": 23007 + }, + { + "epoch": 1.9606305922454197, + "grad_norm": 11.8125, + "learning_rate": 2.409942550433142e-09, + "loss": 0.3098, + "step": 23008 + }, + { + "epoch": 1.9607158074137196, + "grad_norm": 15.9375, + "learning_rate": 2.3995245249067157e-09, + "loss": 0.4936, + "step": 23009 + }, + { + "epoch": 1.9608010225820196, + "grad_norm": 27.375, + "learning_rate": 2.3891290447754236e-09, + "loss": 0.7621, + "step": 23010 + }, + { + "epoch": 1.9608862377503196, + "grad_norm": 19.125, + "learning_rate": 2.3787561102270317e-09, + "loss": 0.8545, + "step": 23011 + }, + { + "epoch": 1.9609714529186195, + "grad_norm": 27.625, + "learning_rate": 2.3684057214493073e-09, + "loss": 0.895, + "step": 23012 + }, + { + "epoch": 1.9610566680869195, + "grad_norm": 13.0, + "learning_rate": 2.3580778786289062e-09, + "loss": 0.5552, + "step": 23013 + }, + { + "epoch": 1.9611418832552194, + "grad_norm": 17.125, + "learning_rate": 2.347772581952762e-09, + "loss": 0.8315, + "step": 23014 + }, + { + "epoch": 1.9612270984235194, + "grad_norm": 16.75, + "learning_rate": 2.3374898316068375e-09, + "loss": 0.6713, + "step": 23015 + }, + { + "epoch": 1.9613123135918193, + "grad_norm": 19.375, + "learning_rate": 2.327229627777233e-09, + "loss": 0.727, + "step": 23016 + }, + { + "epoch": 1.9613975287601193, + "grad_norm": 17.0, + "learning_rate": 2.3169919706490797e-09, + "loss": 0.6789, + "step": 23017 + }, + { + "epoch": 1.9614827439284193, + "grad_norm": 17.625, + "learning_rate": 2.3067768604076448e-09, + "loss": 0.5146, + "step": 23018 + }, + { + "epoch": 1.9615679590967192, + "grad_norm": 14.125, + "learning_rate": 2.296584297237364e-09, + "loss": 0.4511, + "step": 23019 + }, + { + "epoch": 1.9616531742650192, + "grad_norm": 15.75, + "learning_rate": 2.2864142813223965e-09, + "loss": 0.7398, + "step": 23020 + }, + { + "epoch": 1.9617383894333191, + "grad_norm": 14.125, + "learning_rate": 2.276266812846761e-09, + "loss": 0.517, + "step": 23021 + }, + { + "epoch": 1.961823604601619, + "grad_norm": 17.0, + "learning_rate": 2.2661418919935062e-09, + "loss": 0.6405, + "step": 23022 + }, + { + "epoch": 1.961908819769919, + "grad_norm": 17.875, + "learning_rate": 2.2560395189458182e-09, + "loss": 0.6748, + "step": 23023 + }, + { + "epoch": 1.961994034938219, + "grad_norm": 13.875, + "learning_rate": 2.2459596938861905e-09, + "loss": 0.416, + "step": 23024 + }, + { + "epoch": 1.962079250106519, + "grad_norm": 21.375, + "learning_rate": 2.2359024169968378e-09, + "loss": 0.6156, + "step": 23025 + }, + { + "epoch": 1.962164465274819, + "grad_norm": 15.875, + "learning_rate": 2.225867688459421e-09, + "loss": 1.0962, + "step": 23026 + }, + { + "epoch": 1.9622496804431189, + "grad_norm": 12.1875, + "learning_rate": 2.2158555084553224e-09, + "loss": 0.417, + "step": 23027 + }, + { + "epoch": 1.9623348956114188, + "grad_norm": 14.375, + "learning_rate": 2.2058658771653697e-09, + "loss": 0.4612, + "step": 23028 + }, + { + "epoch": 1.9624201107797188, + "grad_norm": 31.375, + "learning_rate": 2.1958987947702514e-09, + "loss": 0.7578, + "step": 23029 + }, + { + "epoch": 1.9625053259480187, + "grad_norm": 30.0, + "learning_rate": 2.1859542614501016e-09, + "loss": 0.9946, + "step": 23030 + }, + { + "epoch": 1.9625905411163187, + "grad_norm": 14.9375, + "learning_rate": 2.17603227738436e-09, + "loss": 0.5833, + "step": 23031 + }, + { + "epoch": 1.9626757562846187, + "grad_norm": 14.3125, + "learning_rate": 2.1661328427526053e-09, + "loss": 0.7809, + "step": 23032 + }, + { + "epoch": 1.9627609714529186, + "grad_norm": 13.5, + "learning_rate": 2.156255957733583e-09, + "loss": 0.3277, + "step": 23033 + }, + { + "epoch": 1.9628461866212186, + "grad_norm": 17.125, + "learning_rate": 2.1464016225057614e-09, + "loss": 0.5945, + "step": 23034 + }, + { + "epoch": 1.9629314017895185, + "grad_norm": 12.25, + "learning_rate": 2.136569837247332e-09, + "loss": 0.5264, + "step": 23035 + }, + { + "epoch": 1.9630166169578185, + "grad_norm": 15.75, + "learning_rate": 2.12676060213593e-09, + "loss": 0.6614, + "step": 23036 + }, + { + "epoch": 1.9631018321261184, + "grad_norm": 14.0, + "learning_rate": 2.1169739173487745e-09, + "loss": 0.5344, + "step": 23037 + }, + { + "epoch": 1.9631870472944184, + "grad_norm": 16.625, + "learning_rate": 2.107209783062808e-09, + "loss": 0.7302, + "step": 23038 + }, + { + "epoch": 1.9632722624627184, + "grad_norm": 28.25, + "learning_rate": 2.0974681994544165e-09, + "loss": 0.5176, + "step": 23039 + }, + { + "epoch": 1.9633574776310183, + "grad_norm": 14.875, + "learning_rate": 2.087749166699571e-09, + "loss": 0.5706, + "step": 23040 + }, + { + "epoch": 1.9634426927993183, + "grad_norm": 13.25, + "learning_rate": 2.0780526849739634e-09, + "loss": 0.4429, + "step": 23041 + }, + { + "epoch": 1.9635279079676182, + "grad_norm": 17.0, + "learning_rate": 2.068378754452732e-09, + "loss": 0.7627, + "step": 23042 + }, + { + "epoch": 1.9636131231359182, + "grad_norm": 21.875, + "learning_rate": 2.058727375311015e-09, + "loss": 1.0042, + "step": 23043 + }, + { + "epoch": 1.9636983383042181, + "grad_norm": 18.0, + "learning_rate": 2.0490985477228386e-09, + "loss": 0.6375, + "step": 23044 + }, + { + "epoch": 1.963783553472518, + "grad_norm": 10.8125, + "learning_rate": 2.0394922718623698e-09, + "loss": 0.2024, + "step": 23045 + }, + { + "epoch": 1.963868768640818, + "grad_norm": 13.0, + "learning_rate": 2.02990854790322e-09, + "loss": 0.5834, + "step": 23046 + }, + { + "epoch": 1.963953983809118, + "grad_norm": 47.5, + "learning_rate": 2.020347376018583e-09, + "loss": 0.8603, + "step": 23047 + }, + { + "epoch": 1.964039198977418, + "grad_norm": 11.0, + "learning_rate": 2.0108087563812385e-09, + "loss": 0.2898, + "step": 23048 + }, + { + "epoch": 1.964124414145718, + "grad_norm": 15.25, + "learning_rate": 2.0012926891635475e-09, + "loss": 0.4462, + "step": 23049 + }, + { + "epoch": 1.9642096293140179, + "grad_norm": 11.5625, + "learning_rate": 1.9917991745373176e-09, + "loss": 0.3421, + "step": 23050 + }, + { + "epoch": 1.9642948444823178, + "grad_norm": 15.375, + "learning_rate": 1.9823282126743558e-09, + "loss": 0.5738, + "step": 23051 + }, + { + "epoch": 1.9643800596506178, + "grad_norm": 25.875, + "learning_rate": 1.9728798037456366e-09, + "loss": 0.8628, + "step": 23052 + }, + { + "epoch": 1.9644652748189178, + "grad_norm": 20.375, + "learning_rate": 1.963453947921995e-09, + "loss": 0.9686, + "step": 23053 + }, + { + "epoch": 1.9645504899872177, + "grad_norm": 19.125, + "learning_rate": 1.9540506453737117e-09, + "loss": 0.77, + "step": 23054 + }, + { + "epoch": 1.9646357051555177, + "grad_norm": 33.0, + "learning_rate": 1.94466989627079e-09, + "loss": 0.5928, + "step": 23055 + }, + { + "epoch": 1.9647209203238176, + "grad_norm": 10.625, + "learning_rate": 1.935311700782677e-09, + "loss": 0.425, + "step": 23056 + }, + { + "epoch": 1.9648061354921176, + "grad_norm": 15.75, + "learning_rate": 1.9259760590785438e-09, + "loss": 0.4426, + "step": 23057 + }, + { + "epoch": 1.9648913506604175, + "grad_norm": 14.25, + "learning_rate": 1.9166629713270046e-09, + "loss": 0.4274, + "step": 23058 + }, + { + "epoch": 1.9649765658287175, + "grad_norm": 10.875, + "learning_rate": 1.9073724376963974e-09, + "loss": 0.2492, + "step": 23059 + }, + { + "epoch": 1.9650617809970174, + "grad_norm": 18.625, + "learning_rate": 1.8981044583547823e-09, + "loss": 0.6669, + "step": 23060 + }, + { + "epoch": 1.9651469961653174, + "grad_norm": 38.25, + "learning_rate": 1.8888590334693866e-09, + "loss": 0.8172, + "step": 23061 + }, + { + "epoch": 1.9652322113336174, + "grad_norm": 12.0, + "learning_rate": 1.879636163207299e-09, + "loss": 0.3306, + "step": 23062 + }, + { + "epoch": 1.9653174265019173, + "grad_norm": 17.375, + "learning_rate": 1.87043584773533e-09, + "loss": 0.7051, + "step": 23063 + }, + { + "epoch": 1.9654026416702173, + "grad_norm": 16.75, + "learning_rate": 1.8612580872195974e-09, + "loss": 0.5711, + "step": 23064 + }, + { + "epoch": 1.9654878568385172, + "grad_norm": 16.75, + "learning_rate": 1.8521028818262178e-09, + "loss": 0.5496, + "step": 23065 + }, + { + "epoch": 1.9655730720068172, + "grad_norm": 11.0625, + "learning_rate": 1.842970231720198e-09, + "loss": 0.3839, + "step": 23066 + }, + { + "epoch": 1.9656582871751171, + "grad_norm": 19.875, + "learning_rate": 1.8338601370669618e-09, + "loss": 0.9758, + "step": 23067 + }, + { + "epoch": 1.965743502343417, + "grad_norm": 15.0, + "learning_rate": 1.8247725980310993e-09, + "loss": 0.643, + "step": 23068 + }, + { + "epoch": 1.965828717511717, + "grad_norm": 15.0625, + "learning_rate": 1.815707614776646e-09, + "loss": 0.4918, + "step": 23069 + }, + { + "epoch": 1.965913932680017, + "grad_norm": 13.5625, + "learning_rate": 1.8066651874674989e-09, + "loss": 0.4265, + "step": 23070 + }, + { + "epoch": 1.965999147848317, + "grad_norm": 23.0, + "learning_rate": 1.7976453162669994e-09, + "loss": 0.9298, + "step": 23071 + }, + { + "epoch": 1.966084363016617, + "grad_norm": 14.9375, + "learning_rate": 1.7886480013383501e-09, + "loss": 0.4361, + "step": 23072 + }, + { + "epoch": 1.9661695781849169, + "grad_norm": 15.0, + "learning_rate": 1.7796732428439213e-09, + "loss": 0.7417, + "step": 23073 + }, + { + "epoch": 1.9662547933532168, + "grad_norm": 14.1875, + "learning_rate": 1.770721040945944e-09, + "loss": 0.5085, + "step": 23074 + }, + { + "epoch": 1.9663400085215168, + "grad_norm": 17.875, + "learning_rate": 1.7617913958062338e-09, + "loss": 0.8252, + "step": 23075 + }, + { + "epoch": 1.9664252236898168, + "grad_norm": 15.5, + "learning_rate": 1.752884307586189e-09, + "loss": 0.716, + "step": 23076 + }, + { + "epoch": 1.9665104388581167, + "grad_norm": 10.5625, + "learning_rate": 1.743999776446792e-09, + "loss": 0.5298, + "step": 23077 + }, + { + "epoch": 1.9665956540264167, + "grad_norm": 13.75, + "learning_rate": 1.73513780254847e-09, + "loss": 0.625, + "step": 23078 + }, + { + "epoch": 1.9666808691947166, + "grad_norm": 11.375, + "learning_rate": 1.7262983860515114e-09, + "loss": 0.237, + "step": 23079 + }, + { + "epoch": 1.9667660843630166, + "grad_norm": 12.0625, + "learning_rate": 1.717481527115511e-09, + "loss": 0.3264, + "step": 23080 + }, + { + "epoch": 1.9668512995313165, + "grad_norm": 19.625, + "learning_rate": 1.7086872258997855e-09, + "loss": 0.7508, + "step": 23081 + }, + { + "epoch": 1.9669365146996165, + "grad_norm": 10.8125, + "learning_rate": 1.6999154825635134e-09, + "loss": 0.2095, + "step": 23082 + }, + { + "epoch": 1.9670217298679165, + "grad_norm": 23.125, + "learning_rate": 1.6911662972649012e-09, + "loss": 0.5202, + "step": 23083 + }, + { + "epoch": 1.9671069450362164, + "grad_norm": 14.4375, + "learning_rate": 1.682439670162156e-09, + "loss": 0.5985, + "step": 23084 + }, + { + "epoch": 1.9671921602045164, + "grad_norm": 18.0, + "learning_rate": 1.673735601413068e-09, + "loss": 0.5777, + "step": 23085 + }, + { + "epoch": 1.9672773753728163, + "grad_norm": 17.375, + "learning_rate": 1.6650540911750113e-09, + "loss": 0.6757, + "step": 23086 + }, + { + "epoch": 1.9673625905411163, + "grad_norm": 16.875, + "learning_rate": 1.6563951396045275e-09, + "loss": 1.2129, + "step": 23087 + }, + { + "epoch": 1.9674478057094162, + "grad_norm": 21.375, + "learning_rate": 1.6477587468582968e-09, + "loss": 0.7593, + "step": 23088 + }, + { + "epoch": 1.9675330208777162, + "grad_norm": 19.5, + "learning_rate": 1.6391449130924442e-09, + "loss": 0.7467, + "step": 23089 + }, + { + "epoch": 1.9676182360460162, + "grad_norm": 16.5, + "learning_rate": 1.6305536384625399e-09, + "loss": 0.6727, + "step": 23090 + }, + { + "epoch": 1.9677034512143161, + "grad_norm": 17.875, + "learning_rate": 1.621984923123876e-09, + "loss": 0.7189, + "step": 23091 + }, + { + "epoch": 1.967788666382616, + "grad_norm": 15.8125, + "learning_rate": 1.613438767231329e-09, + "loss": 0.7864, + "step": 23092 + }, + { + "epoch": 1.967873881550916, + "grad_norm": 14.9375, + "learning_rate": 1.6049151709392196e-09, + "loss": 0.602, + "step": 23093 + }, + { + "epoch": 1.967959096719216, + "grad_norm": 26.75, + "learning_rate": 1.5964141344017302e-09, + "loss": 0.7621, + "step": 23094 + }, + { + "epoch": 1.968044311887516, + "grad_norm": 18.0, + "learning_rate": 1.5879356577723494e-09, + "loss": 1.0914, + "step": 23095 + }, + { + "epoch": 1.968129527055816, + "grad_norm": 23.625, + "learning_rate": 1.5794797412042872e-09, + "loss": 0.8505, + "step": 23096 + }, + { + "epoch": 1.9682147422241159, + "grad_norm": 12.3125, + "learning_rate": 1.5710463848504776e-09, + "loss": 0.5152, + "step": 23097 + }, + { + "epoch": 1.9682999573924158, + "grad_norm": 13.6875, + "learning_rate": 1.5626355888632982e-09, + "loss": 1.0984, + "step": 23098 + }, + { + "epoch": 1.9683851725607158, + "grad_norm": 18.5, + "learning_rate": 1.554247353394711e-09, + "loss": 0.7624, + "step": 23099 + }, + { + "epoch": 1.9684703877290157, + "grad_norm": 15.3125, + "learning_rate": 1.5458816785961228e-09, + "loss": 0.6482, + "step": 23100 + }, + { + "epoch": 1.9685556028973157, + "grad_norm": 51.5, + "learning_rate": 1.5375385646190787e-09, + "loss": 0.8037, + "step": 23101 + }, + { + "epoch": 1.9686408180656156, + "grad_norm": 24.0, + "learning_rate": 1.5292180116140143e-09, + "loss": 1.2655, + "step": 23102 + }, + { + "epoch": 1.9687260332339156, + "grad_norm": 13.25, + "learning_rate": 1.5209200197313645e-09, + "loss": 0.4831, + "step": 23103 + }, + { + "epoch": 1.9688112484022156, + "grad_norm": 14.75, + "learning_rate": 1.5126445891212871e-09, + "loss": 0.6919, + "step": 23104 + }, + { + "epoch": 1.9688964635705155, + "grad_norm": 19.25, + "learning_rate": 1.5043917199329682e-09, + "loss": 0.7322, + "step": 23105 + }, + { + "epoch": 1.9689816787388155, + "grad_norm": 19.0, + "learning_rate": 1.4961614123158718e-09, + "loss": 0.6141, + "step": 23106 + }, + { + "epoch": 1.9690668939071154, + "grad_norm": 14.625, + "learning_rate": 1.487953666418629e-09, + "loss": 0.4339, + "step": 23107 + }, + { + "epoch": 1.9691521090754154, + "grad_norm": 24.75, + "learning_rate": 1.4797684823894543e-09, + "loss": 0.3565, + "step": 23108 + }, + { + "epoch": 1.9692373242437153, + "grad_norm": 11.0, + "learning_rate": 1.4716058603764239e-09, + "loss": 0.3067, + "step": 23109 + }, + { + "epoch": 1.9693225394120153, + "grad_norm": 15.8125, + "learning_rate": 1.4634658005270586e-09, + "loss": 0.8576, + "step": 23110 + }, + { + "epoch": 1.9694077545803153, + "grad_norm": 14.5, + "learning_rate": 1.4553483029881854e-09, + "loss": 0.6379, + "step": 23111 + }, + { + "epoch": 1.9694929697486152, + "grad_norm": 19.0, + "learning_rate": 1.44725336790677e-09, + "loss": 0.8173, + "step": 23112 + }, + { + "epoch": 1.9695781849169152, + "grad_norm": 16.875, + "learning_rate": 1.439180995428946e-09, + "loss": 1.0586, + "step": 23113 + }, + { + "epoch": 1.9696634000852151, + "grad_norm": 17.125, + "learning_rate": 1.4311311857007072e-09, + "loss": 0.7584, + "step": 23114 + }, + { + "epoch": 1.969748615253515, + "grad_norm": 15.8125, + "learning_rate": 1.423103938867354e-09, + "loss": 0.7925, + "step": 23115 + }, + { + "epoch": 1.969833830421815, + "grad_norm": 14.625, + "learning_rate": 1.415099255074187e-09, + "loss": 0.592, + "step": 23116 + }, + { + "epoch": 1.969919045590115, + "grad_norm": 19.75, + "learning_rate": 1.4071171344656742e-09, + "loss": 0.5554, + "step": 23117 + }, + { + "epoch": 1.970004260758415, + "grad_norm": 15.3125, + "learning_rate": 1.3991575771861443e-09, + "loss": 0.3557, + "step": 23118 + }, + { + "epoch": 1.970089475926715, + "grad_norm": 12.0625, + "learning_rate": 1.3912205833793713e-09, + "loss": 0.3957, + "step": 23119 + }, + { + "epoch": 1.9701746910950149, + "grad_norm": 16.5, + "learning_rate": 1.3833061531888514e-09, + "loss": 0.5071, + "step": 23120 + }, + { + "epoch": 1.9702599062633148, + "grad_norm": 18.125, + "learning_rate": 1.375414286757526e-09, + "loss": 0.7778, + "step": 23121 + }, + { + "epoch": 1.9703451214316148, + "grad_norm": 17.875, + "learning_rate": 1.3675449842280585e-09, + "loss": 0.8993, + "step": 23122 + }, + { + "epoch": 1.9704303365999147, + "grad_norm": 37.25, + "learning_rate": 1.3596982457425579e-09, + "loss": 0.5876, + "step": 23123 + }, + { + "epoch": 1.9705155517682147, + "grad_norm": 14.0, + "learning_rate": 1.3518740714431321e-09, + "loss": 0.3453, + "step": 23124 + }, + { + "epoch": 1.9706007669365146, + "grad_norm": 15.375, + "learning_rate": 1.3440724614707802e-09, + "loss": 0.4762, + "step": 23125 + }, + { + "epoch": 1.9706859821048146, + "grad_norm": 17.75, + "learning_rate": 1.3362934159666385e-09, + "loss": 0.5011, + "step": 23126 + }, + { + "epoch": 1.9707711972731146, + "grad_norm": 14.6875, + "learning_rate": 1.3285369350714284e-09, + "loss": 0.776, + "step": 23127 + }, + { + "epoch": 1.9708564124414145, + "grad_norm": 18.875, + "learning_rate": 1.3208030189251764e-09, + "loss": 0.7474, + "step": 23128 + }, + { + "epoch": 1.9709416276097145, + "grad_norm": 13.75, + "learning_rate": 1.313091667667632e-09, + "loss": 0.5405, + "step": 23129 + }, + { + "epoch": 1.9710268427780144, + "grad_norm": 27.75, + "learning_rate": 1.305402881438128e-09, + "loss": 0.4619, + "step": 23130 + }, + { + "epoch": 1.9711120579463144, + "grad_norm": 12.625, + "learning_rate": 1.2977366603757202e-09, + "loss": 0.4542, + "step": 23131 + }, + { + "epoch": 1.9711972731146146, + "grad_norm": 16.25, + "learning_rate": 1.29009300461877e-09, + "loss": 0.6752, + "step": 23132 + }, + { + "epoch": 1.9712824882829145, + "grad_norm": 18.375, + "learning_rate": 1.2824719143056385e-09, + "loss": 0.5236, + "step": 23133 + }, + { + "epoch": 1.9713677034512145, + "grad_norm": 13.375, + "learning_rate": 1.2748733895738551e-09, + "loss": 0.4807, + "step": 23134 + }, + { + "epoch": 1.9714529186195144, + "grad_norm": 13.5, + "learning_rate": 1.26729743056081e-09, + "loss": 0.5434, + "step": 23135 + }, + { + "epoch": 1.9715381337878144, + "grad_norm": 16.375, + "learning_rate": 1.2597440374034764e-09, + "loss": 0.3572, + "step": 23136 + }, + { + "epoch": 1.9716233489561144, + "grad_norm": 17.125, + "learning_rate": 1.2522132102381345e-09, + "loss": 0.725, + "step": 23137 + }, + { + "epoch": 1.9717085641244143, + "grad_norm": 19.5, + "learning_rate": 1.2447049492010643e-09, + "loss": 0.529, + "step": 23138 + }, + { + "epoch": 1.9717937792927143, + "grad_norm": 13.375, + "learning_rate": 1.2372192544279904e-09, + "loss": 0.5407, + "step": 23139 + }, + { + "epoch": 1.9718789944610142, + "grad_norm": 42.25, + "learning_rate": 1.2297561260539436e-09, + "loss": 0.7963, + "step": 23140 + }, + { + "epoch": 1.9719642096293142, + "grad_norm": 10.5, + "learning_rate": 1.2223155642140938e-09, + "loss": 0.3503, + "step": 23141 + }, + { + "epoch": 1.9720494247976141, + "grad_norm": 17.25, + "learning_rate": 1.214897569042639e-09, + "loss": 0.4181, + "step": 23142 + }, + { + "epoch": 1.972134639965914, + "grad_norm": 19.875, + "learning_rate": 1.2075021406737775e-09, + "loss": 0.6367, + "step": 23143 + }, + { + "epoch": 1.972219855134214, + "grad_norm": 18.875, + "learning_rate": 1.2001292792410136e-09, + "loss": 0.8345, + "step": 23144 + }, + { + "epoch": 1.972305070302514, + "grad_norm": 18.5, + "learning_rate": 1.1927789848777127e-09, + "loss": 0.535, + "step": 23145 + }, + { + "epoch": 1.972390285470814, + "grad_norm": 18.625, + "learning_rate": 1.1854512577168242e-09, + "loss": 0.6726, + "step": 23146 + }, + { + "epoch": 1.972475500639114, + "grad_norm": 18.875, + "learning_rate": 1.1781460978904646e-09, + "loss": 0.8215, + "step": 23147 + }, + { + "epoch": 1.9725607158074139, + "grad_norm": 14.25, + "learning_rate": 1.1708635055307505e-09, + "loss": 0.6118, + "step": 23148 + }, + { + "epoch": 1.9726459309757138, + "grad_norm": 17.0, + "learning_rate": 1.1636034807695207e-09, + "loss": 0.3633, + "step": 23149 + }, + { + "epoch": 1.9727311461440138, + "grad_norm": 11.375, + "learning_rate": 1.1563660237376428e-09, + "loss": 0.3144, + "step": 23150 + }, + { + "epoch": 1.9728163613123137, + "grad_norm": 12.5, + "learning_rate": 1.1491511345659844e-09, + "loss": 0.4339, + "step": 23151 + }, + { + "epoch": 1.9729015764806137, + "grad_norm": 12.875, + "learning_rate": 1.1419588133851357e-09, + "loss": 0.4516, + "step": 23152 + }, + { + "epoch": 1.9729867916489137, + "grad_norm": 14.9375, + "learning_rate": 1.1347890603247146e-09, + "loss": 0.4728, + "step": 23153 + }, + { + "epoch": 1.9730720068172136, + "grad_norm": 13.1875, + "learning_rate": 1.1276418755147566e-09, + "loss": 0.5482, + "step": 23154 + }, + { + "epoch": 1.9731572219855136, + "grad_norm": 13.3125, + "learning_rate": 1.1205172590839086e-09, + "loss": 0.6259, + "step": 23155 + }, + { + "epoch": 1.9732424371538135, + "grad_norm": 13.875, + "learning_rate": 1.113415211161234e-09, + "loss": 0.5267, + "step": 23156 + }, + { + "epoch": 1.9733276523221135, + "grad_norm": 14.75, + "learning_rate": 1.1063357318749634e-09, + "loss": 0.4483, + "step": 23157 + }, + { + "epoch": 1.9734128674904134, + "grad_norm": 14.1875, + "learning_rate": 1.0992788213531892e-09, + "loss": 0.5168, + "step": 23158 + }, + { + "epoch": 1.9734980826587134, + "grad_norm": 20.875, + "learning_rate": 1.092244479723309e-09, + "loss": 0.6854, + "step": 23159 + }, + { + "epoch": 1.9735832978270134, + "grad_norm": 20.625, + "learning_rate": 1.0852327071123047e-09, + "loss": 1.0474, + "step": 23160 + }, + { + "epoch": 1.9736685129953133, + "grad_norm": 25.625, + "learning_rate": 1.0782435036471583e-09, + "loss": 0.395, + "step": 23161 + }, + { + "epoch": 1.9737537281636133, + "grad_norm": 19.0, + "learning_rate": 1.0712768694540188e-09, + "loss": 0.588, + "step": 23162 + }, + { + "epoch": 1.9738389433319132, + "grad_norm": 12.125, + "learning_rate": 1.0643328046586188e-09, + "loss": 0.2899, + "step": 23163 + }, + { + "epoch": 1.9739241585002132, + "grad_norm": 16.125, + "learning_rate": 1.0574113093868298e-09, + "loss": 0.6833, + "step": 23164 + }, + { + "epoch": 1.9740093736685131, + "grad_norm": 12.75, + "learning_rate": 1.0505123837634135e-09, + "loss": 0.4787, + "step": 23165 + }, + { + "epoch": 1.974094588836813, + "grad_norm": 17.125, + "learning_rate": 1.0436360279132696e-09, + "loss": 0.6312, + "step": 23166 + }, + { + "epoch": 1.974179804005113, + "grad_norm": 17.5, + "learning_rate": 1.0367822419604657e-09, + "loss": 0.553, + "step": 23167 + }, + { + "epoch": 1.974265019173413, + "grad_norm": 16.25, + "learning_rate": 1.0299510260289302e-09, + "loss": 0.8832, + "step": 23168 + }, + { + "epoch": 1.974350234341713, + "grad_norm": 17.0, + "learning_rate": 1.0231423802421758e-09, + "loss": 0.8022, + "step": 23169 + }, + { + "epoch": 1.974435449510013, + "grad_norm": 26.0, + "learning_rate": 1.0163563047231596e-09, + "loss": 0.5269, + "step": 23170 + }, + { + "epoch": 1.9745206646783129, + "grad_norm": 19.625, + "learning_rate": 1.0095927995945608e-09, + "loss": 0.7514, + "step": 23171 + }, + { + "epoch": 1.9746058798466128, + "grad_norm": 11.75, + "learning_rate": 1.0028518649785047e-09, + "loss": 0.3114, + "step": 23172 + }, + { + "epoch": 1.9746910950149128, + "grad_norm": 15.4375, + "learning_rate": 9.961335009968377e-10, + "loss": 0.9102, + "step": 23173 + }, + { + "epoch": 1.9747763101832128, + "grad_norm": 14.4375, + "learning_rate": 9.894377077711291e-10, + "loss": 0.5267, + "step": 23174 + }, + { + "epoch": 1.9748615253515127, + "grad_norm": 19.0, + "learning_rate": 9.82764485422255e-10, + "loss": 0.8224, + "step": 23175 + }, + { + "epoch": 1.9749467405198127, + "grad_norm": 43.75, + "learning_rate": 9.761138340708132e-10, + "loss": 0.6936, + "step": 23176 + }, + { + "epoch": 1.9750319556881126, + "grad_norm": 13.5, + "learning_rate": 9.694857538368463e-10, + "loss": 0.5346, + "step": 23177 + }, + { + "epoch": 1.9751171708564126, + "grad_norm": 12.8125, + "learning_rate": 9.628802448403974e-10, + "loss": 0.4519, + "step": 23178 + }, + { + "epoch": 1.9752023860247125, + "grad_norm": 15.6875, + "learning_rate": 9.562973072006766e-10, + "loss": 0.9345, + "step": 23179 + }, + { + "epoch": 1.9752876011930125, + "grad_norm": 14.625, + "learning_rate": 9.497369410366164e-10, + "loss": 0.4527, + "step": 23180 + }, + { + "epoch": 1.9753728163613125, + "grad_norm": 13.125, + "learning_rate": 9.43199146466872e-10, + "loss": 0.4334, + "step": 23181 + }, + { + "epoch": 1.9754580315296124, + "grad_norm": 12.5, + "learning_rate": 9.366839236096826e-10, + "loss": 0.3288, + "step": 23182 + }, + { + "epoch": 1.9755432466979124, + "grad_norm": 11.5625, + "learning_rate": 9.301912725824536e-10, + "loss": 0.7561, + "step": 23183 + }, + { + "epoch": 1.9756284618662123, + "grad_norm": 16.25, + "learning_rate": 9.237211935027301e-10, + "loss": 0.6324, + "step": 23184 + }, + { + "epoch": 1.9757136770345123, + "grad_norm": 15.6875, + "learning_rate": 9.17273686487502e-10, + "loss": 0.663, + "step": 23185 + }, + { + "epoch": 1.9757988922028122, + "grad_norm": 30.625, + "learning_rate": 9.108487516530651e-10, + "loss": 0.5023, + "step": 23186 + }, + { + "epoch": 1.9758841073711122, + "grad_norm": 15.5625, + "learning_rate": 9.044463891157151e-10, + "loss": 0.742, + "step": 23187 + }, + { + "epoch": 1.9759693225394122, + "grad_norm": 12.9375, + "learning_rate": 8.980665989910542e-10, + "loss": 0.5283, + "step": 23188 + }, + { + "epoch": 1.976054537707712, + "grad_norm": 15.0625, + "learning_rate": 8.917093813945454e-10, + "loss": 0.5337, + "step": 23189 + }, + { + "epoch": 1.976139752876012, + "grad_norm": 15.5625, + "learning_rate": 8.853747364408194e-10, + "loss": 0.4634, + "step": 23190 + }, + { + "epoch": 1.976224968044312, + "grad_norm": 16.375, + "learning_rate": 8.790626642443678e-10, + "loss": 0.6058, + "step": 23191 + }, + { + "epoch": 1.976310183212612, + "grad_norm": 16.875, + "learning_rate": 8.727731649194049e-10, + "loss": 0.3757, + "step": 23192 + }, + { + "epoch": 1.976395398380912, + "grad_norm": 22.0, + "learning_rate": 8.665062385795897e-10, + "loss": 0.5928, + "step": 23193 + }, + { + "epoch": 1.976480613549212, + "grad_norm": 21.125, + "learning_rate": 8.602618853380262e-10, + "loss": 0.7695, + "step": 23194 + }, + { + "epoch": 1.9765658287175119, + "grad_norm": 19.25, + "learning_rate": 8.540401053076797e-10, + "loss": 0.5815, + "step": 23195 + }, + { + "epoch": 1.9766510438858118, + "grad_norm": 17.625, + "learning_rate": 8.478408986008213e-10, + "loss": 0.4575, + "step": 23196 + }, + { + "epoch": 1.9767362590541118, + "grad_norm": 22.75, + "learning_rate": 8.416642653297224e-10, + "loss": 1.0323, + "step": 23197 + }, + { + "epoch": 1.9768214742224117, + "grad_norm": 12.5, + "learning_rate": 8.355102056058217e-10, + "loss": 0.3818, + "step": 23198 + }, + { + "epoch": 1.9769066893907117, + "grad_norm": 12.5625, + "learning_rate": 8.29378719540419e-10, + "loss": 0.3655, + "step": 23199 + }, + { + "epoch": 1.9769919045590116, + "grad_norm": 15.0625, + "learning_rate": 8.232698072442591e-10, + "loss": 0.4592, + "step": 23200 + }, + { + "epoch": 1.9770771197273116, + "grad_norm": 14.6875, + "learning_rate": 8.171834688278091e-10, + "loss": 0.4865, + "step": 23201 + }, + { + "epoch": 1.9771623348956116, + "grad_norm": 17.75, + "learning_rate": 8.11119704400981e-10, + "loss": 0.6239, + "step": 23202 + }, + { + "epoch": 1.9772475500639115, + "grad_norm": 15.5625, + "learning_rate": 8.050785140734096e-10, + "loss": 0.653, + "step": 23203 + }, + { + "epoch": 1.9773327652322115, + "grad_norm": 16.75, + "learning_rate": 7.990598979541742e-10, + "loss": 0.4711, + "step": 23204 + }, + { + "epoch": 1.9774179804005114, + "grad_norm": 30.125, + "learning_rate": 7.930638561520764e-10, + "loss": 1.1897, + "step": 23205 + }, + { + "epoch": 1.9775031955688114, + "grad_norm": 14.5, + "learning_rate": 7.870903887755021e-10, + "loss": 0.6074, + "step": 23206 + }, + { + "epoch": 1.9775884107371113, + "grad_norm": 14.375, + "learning_rate": 7.811394959324203e-10, + "loss": 0.4635, + "step": 23207 + }, + { + "epoch": 1.9776736259054113, + "grad_norm": 16.0, + "learning_rate": 7.752111777303839e-10, + "loss": 0.3942, + "step": 23208 + }, + { + "epoch": 1.9777588410737112, + "grad_norm": 18.625, + "learning_rate": 7.693054342763906e-10, + "loss": 0.6744, + "step": 23209 + }, + { + "epoch": 1.9778440562420112, + "grad_norm": 19.875, + "learning_rate": 7.634222656772994e-10, + "loss": 0.8584, + "step": 23210 + }, + { + "epoch": 1.9779292714103112, + "grad_norm": 13.9375, + "learning_rate": 7.575616720392754e-10, + "loss": 0.335, + "step": 23211 + }, + { + "epoch": 1.9780144865786111, + "grad_norm": 10.625, + "learning_rate": 7.517236534683447e-10, + "loss": 0.3457, + "step": 23212 + }, + { + "epoch": 1.978099701746911, + "grad_norm": 21.75, + "learning_rate": 7.459082100701176e-10, + "loss": 0.7888, + "step": 23213 + }, + { + "epoch": 1.978184916915211, + "grad_norm": 18.75, + "learning_rate": 7.401153419493712e-10, + "loss": 0.6673, + "step": 23214 + }, + { + "epoch": 1.978270132083511, + "grad_norm": 21.75, + "learning_rate": 7.343450492111604e-10, + "loss": 1.1788, + "step": 23215 + }, + { + "epoch": 1.978355347251811, + "grad_norm": 15.4375, + "learning_rate": 7.285973319594297e-10, + "loss": 0.596, + "step": 23216 + }, + { + "epoch": 1.978440562420111, + "grad_norm": 17.75, + "learning_rate": 7.228721902982628e-10, + "loss": 0.688, + "step": 23217 + }, + { + "epoch": 1.9785257775884109, + "grad_norm": 18.375, + "learning_rate": 7.1716962433091e-10, + "loss": 0.5774, + "step": 23218 + }, + { + "epoch": 1.9786109927567108, + "grad_norm": 12.625, + "learning_rate": 7.114896341606225e-10, + "loss": 0.4801, + "step": 23219 + }, + { + "epoch": 1.9786962079250108, + "grad_norm": 15.5, + "learning_rate": 7.058322198900957e-10, + "loss": 0.3919, + "step": 23220 + }, + { + "epoch": 1.9787814230933107, + "grad_norm": 17.0, + "learning_rate": 7.001973816213314e-10, + "loss": 0.7137, + "step": 23221 + }, + { + "epoch": 1.9788666382616107, + "grad_norm": 19.0, + "learning_rate": 6.945851194561926e-10, + "loss": 0.5701, + "step": 23222 + }, + { + "epoch": 1.9789518534299106, + "grad_norm": 14.0625, + "learning_rate": 6.889954334962646e-10, + "loss": 0.4946, + "step": 23223 + }, + { + "epoch": 1.9790370685982106, + "grad_norm": 20.0, + "learning_rate": 6.834283238424389e-10, + "loss": 0.8554, + "step": 23224 + }, + { + "epoch": 1.9791222837665106, + "grad_norm": 21.875, + "learning_rate": 6.778837905954683e-10, + "loss": 0.7566, + "step": 23225 + }, + { + "epoch": 1.9792074989348105, + "grad_norm": 19.25, + "learning_rate": 6.723618338552729e-10, + "loss": 0.4918, + "step": 23226 + }, + { + "epoch": 1.9792927141031105, + "grad_norm": 23.125, + "learning_rate": 6.668624537219115e-10, + "loss": 0.6919, + "step": 23227 + }, + { + "epoch": 1.9793779292714104, + "grad_norm": 15.0, + "learning_rate": 6.613856502944715e-10, + "loss": 0.1969, + "step": 23228 + }, + { + "epoch": 1.9794631444397104, + "grad_norm": 18.25, + "learning_rate": 6.55931423672318e-10, + "loss": 0.7214, + "step": 23229 + }, + { + "epoch": 1.9795483596080103, + "grad_norm": 18.75, + "learning_rate": 6.504997739537056e-10, + "loss": 0.5708, + "step": 23230 + }, + { + "epoch": 1.9796335747763103, + "grad_norm": 12.0625, + "learning_rate": 6.45090701236889e-10, + "loss": 0.4487, + "step": 23231 + }, + { + "epoch": 1.9797187899446103, + "grad_norm": 12.625, + "learning_rate": 6.397042056195679e-10, + "loss": 0.2648, + "step": 23232 + }, + { + "epoch": 1.9798040051129102, + "grad_norm": 28.625, + "learning_rate": 6.343402871991644e-10, + "loss": 1.1565, + "step": 23233 + }, + { + "epoch": 1.9798892202812102, + "grad_norm": 25.375, + "learning_rate": 6.289989460724066e-10, + "loss": 0.6484, + "step": 23234 + }, + { + "epoch": 1.9799744354495101, + "grad_norm": 12.75, + "learning_rate": 6.236801823361616e-10, + "loss": 0.5695, + "step": 23235 + }, + { + "epoch": 1.98005965061781, + "grad_norm": 14.9375, + "learning_rate": 6.183839960863247e-10, + "loss": 0.4795, + "step": 23236 + }, + { + "epoch": 1.98014486578611, + "grad_norm": 14.1875, + "learning_rate": 6.131103874186528e-10, + "loss": 0.4457, + "step": 23237 + }, + { + "epoch": 1.98023008095441, + "grad_norm": 12.1875, + "learning_rate": 6.078593564283475e-10, + "loss": 0.6531, + "step": 23238 + }, + { + "epoch": 1.98031529612271, + "grad_norm": 15.8125, + "learning_rate": 6.026309032104716e-10, + "loss": 0.4453, + "step": 23239 + }, + { + "epoch": 1.98040051129101, + "grad_norm": 66.0, + "learning_rate": 5.974250278593941e-10, + "loss": 0.9572, + "step": 23240 + }, + { + "epoch": 1.9804857264593099, + "grad_norm": 21.5, + "learning_rate": 5.922417304692063e-10, + "loss": 0.6631, + "step": 23241 + }, + { + "epoch": 1.9805709416276098, + "grad_norm": 15.75, + "learning_rate": 5.870810111335834e-10, + "loss": 0.6596, + "step": 23242 + }, + { + "epoch": 1.9806561567959098, + "grad_norm": 18.375, + "learning_rate": 5.819428699459229e-10, + "loss": 0.4828, + "step": 23243 + }, + { + "epoch": 1.9807413719642097, + "grad_norm": 20.375, + "learning_rate": 5.768273069987896e-10, + "loss": 0.6896, + "step": 23244 + }, + { + "epoch": 1.9808265871325097, + "grad_norm": 18.75, + "learning_rate": 5.717343223847482e-10, + "loss": 0.5104, + "step": 23245 + }, + { + "epoch": 1.9809118023008097, + "grad_norm": 12.8125, + "learning_rate": 5.666639161959475e-10, + "loss": 0.6057, + "step": 23246 + }, + { + "epoch": 1.9809970174691096, + "grad_norm": 21.375, + "learning_rate": 5.616160885239808e-10, + "loss": 0.5526, + "step": 23247 + }, + { + "epoch": 1.9810822326374096, + "grad_norm": 16.5, + "learning_rate": 5.565908394598863e-10, + "loss": 0.5616, + "step": 23248 + }, + { + "epoch": 1.9811674478057095, + "grad_norm": 15.25, + "learning_rate": 5.515881690947023e-10, + "loss": 0.3627, + "step": 23249 + }, + { + "epoch": 1.9812526629740095, + "grad_norm": 17.375, + "learning_rate": 5.466080775186344e-10, + "loss": 0.9317, + "step": 23250 + }, + { + "epoch": 1.9813378781423094, + "grad_norm": 13.8125, + "learning_rate": 5.416505648218884e-10, + "loss": 0.5098, + "step": 23251 + }, + { + "epoch": 1.9814230933106094, + "grad_norm": 19.875, + "learning_rate": 5.367156310939759e-10, + "loss": 0.7808, + "step": 23252 + }, + { + "epoch": 1.9815083084789094, + "grad_norm": 11.375, + "learning_rate": 5.318032764238535e-10, + "loss": 0.3538, + "step": 23253 + }, + { + "epoch": 1.9815935236472093, + "grad_norm": 17.875, + "learning_rate": 5.269135009006165e-10, + "loss": 0.7669, + "step": 23254 + }, + { + "epoch": 1.9816787388155093, + "grad_norm": 18.375, + "learning_rate": 5.220463046125279e-10, + "loss": 0.4303, + "step": 23255 + }, + { + "epoch": 1.9817639539838092, + "grad_norm": 14.75, + "learning_rate": 5.172016876474339e-10, + "loss": 0.6543, + "step": 23256 + }, + { + "epoch": 1.9818491691521092, + "grad_norm": 20.625, + "learning_rate": 5.123796500930422e-10, + "loss": 0.707, + "step": 23257 + }, + { + "epoch": 1.9819343843204091, + "grad_norm": 14.5, + "learning_rate": 5.075801920363665e-10, + "loss": 0.5756, + "step": 23258 + }, + { + "epoch": 1.982019599488709, + "grad_norm": 16.5, + "learning_rate": 5.02803313564143e-10, + "loss": 1.2815, + "step": 23259 + }, + { + "epoch": 1.982104814657009, + "grad_norm": 13.6875, + "learning_rate": 4.980490147628303e-10, + "loss": 0.3331, + "step": 23260 + }, + { + "epoch": 1.982190029825309, + "grad_norm": 17.125, + "learning_rate": 4.933172957180543e-10, + "loss": 0.5887, + "step": 23261 + }, + { + "epoch": 1.982275244993609, + "grad_norm": 18.5, + "learning_rate": 4.886081565157186e-10, + "loss": 0.5691, + "step": 23262 + }, + { + "epoch": 1.982360460161909, + "grad_norm": 16.125, + "learning_rate": 4.839215972406164e-10, + "loss": 0.6155, + "step": 23263 + }, + { + "epoch": 1.9824456753302089, + "grad_norm": 20.875, + "learning_rate": 4.79257617977541e-10, + "loss": 0.928, + "step": 23264 + }, + { + "epoch": 1.9825308904985088, + "grad_norm": 18.5, + "learning_rate": 4.746162188107306e-10, + "loss": 1.0592, + "step": 23265 + }, + { + "epoch": 1.9826161056668088, + "grad_norm": 16.125, + "learning_rate": 4.699973998241459e-10, + "loss": 0.561, + "step": 23266 + }, + { + "epoch": 1.9827013208351087, + "grad_norm": 15.5625, + "learning_rate": 4.6540116110119217e-10, + "loss": 0.5419, + "step": 23267 + }, + { + "epoch": 1.9827865360034087, + "grad_norm": 33.0, + "learning_rate": 4.6082750272499756e-10, + "loss": 0.8275, + "step": 23268 + }, + { + "epoch": 1.9828717511717087, + "grad_norm": 11.625, + "learning_rate": 4.5627642477813485e-10, + "loss": 0.345, + "step": 23269 + }, + { + "epoch": 1.9829569663400086, + "grad_norm": 14.3125, + "learning_rate": 4.517479273427605e-10, + "loss": 0.3628, + "step": 23270 + }, + { + "epoch": 1.9830421815083086, + "grad_norm": 15.0625, + "learning_rate": 4.4724201050089235e-10, + "loss": 0.7604, + "step": 23271 + }, + { + "epoch": 1.9831273966766085, + "grad_norm": 12.875, + "learning_rate": 4.4275867433399286e-10, + "loss": 0.3542, + "step": 23272 + }, + { + "epoch": 1.9832126118449085, + "grad_norm": 13.5625, + "learning_rate": 4.382979189228309e-10, + "loss": 0.5714, + "step": 23273 + }, + { + "epoch": 1.9832978270132084, + "grad_norm": 15.125, + "learning_rate": 4.338597443481751e-10, + "loss": 0.655, + "step": 23274 + }, + { + "epoch": 1.9833830421815084, + "grad_norm": 18.375, + "learning_rate": 4.2944415069023915e-10, + "loss": 0.4823, + "step": 23275 + }, + { + "epoch": 1.9834682573498084, + "grad_norm": 17.25, + "learning_rate": 4.2505113802882023e-10, + "loss": 0.8536, + "step": 23276 + }, + { + "epoch": 1.9835534725181083, + "grad_norm": 26.125, + "learning_rate": 4.206807064432994e-10, + "loss": 0.8074, + "step": 23277 + }, + { + "epoch": 1.9836386876864083, + "grad_norm": 13.4375, + "learning_rate": 4.163328560125024e-10, + "loss": 0.5872, + "step": 23278 + }, + { + "epoch": 1.9837239028547082, + "grad_norm": 15.3125, + "learning_rate": 4.120075868152551e-10, + "loss": 0.6299, + "step": 23279 + }, + { + "epoch": 1.9838091180230082, + "grad_norm": 17.0, + "learning_rate": 4.077048989295507e-10, + "loss": 0.6026, + "step": 23280 + }, + { + "epoch": 1.9838943331913081, + "grad_norm": 18.5, + "learning_rate": 4.0342479243310475e-10, + "loss": 0.3762, + "step": 23281 + }, + { + "epoch": 1.983979548359608, + "grad_norm": 19.25, + "learning_rate": 3.991672674033553e-10, + "loss": 0.5901, + "step": 23282 + }, + { + "epoch": 1.984064763527908, + "grad_norm": 15.375, + "learning_rate": 3.9493232391732415e-10, + "loss": 0.6387, + "step": 23283 + }, + { + "epoch": 1.984149978696208, + "grad_norm": 15.0, + "learning_rate": 3.907199620513391e-10, + "loss": 0.5621, + "step": 23284 + }, + { + "epoch": 1.984235193864508, + "grad_norm": 14.125, + "learning_rate": 3.8653018188158917e-10, + "loss": 0.3461, + "step": 23285 + }, + { + "epoch": 1.984320409032808, + "grad_norm": 14.625, + "learning_rate": 3.8236298348398594e-10, + "loss": 0.5143, + "step": 23286 + }, + { + "epoch": 1.9844056242011079, + "grad_norm": 14.1875, + "learning_rate": 3.7821836693346935e-10, + "loss": 0.5395, + "step": 23287 + }, + { + "epoch": 1.9844908393694078, + "grad_norm": 13.8125, + "learning_rate": 3.740963323052571e-10, + "loss": 0.4294, + "step": 23288 + }, + { + "epoch": 1.9845760545377078, + "grad_norm": 15.0, + "learning_rate": 3.699968796737341e-10, + "loss": 0.6105, + "step": 23289 + }, + { + "epoch": 1.9846612697060078, + "grad_norm": 16.875, + "learning_rate": 3.659200091130077e-10, + "loss": 0.6301, + "step": 23290 + }, + { + "epoch": 1.9847464848743077, + "grad_norm": 13.3125, + "learning_rate": 3.618657206966303e-10, + "loss": 0.3805, + "step": 23291 + }, + { + "epoch": 1.9848317000426077, + "grad_norm": 13.75, + "learning_rate": 3.5783401449801525e-10, + "loss": 0.7186, + "step": 23292 + }, + { + "epoch": 1.9849169152109076, + "grad_norm": 17.5, + "learning_rate": 3.5382489058988224e-10, + "loss": 0.4876, + "step": 23293 + }, + { + "epoch": 1.9850021303792076, + "grad_norm": 13.1875, + "learning_rate": 3.498383490446733e-10, + "loss": 0.5962, + "step": 23294 + }, + { + "epoch": 1.9850873455475075, + "grad_norm": 11.875, + "learning_rate": 3.458743899346917e-10, + "loss": 0.3941, + "step": 23295 + }, + { + "epoch": 1.9851725607158075, + "grad_norm": 14.5625, + "learning_rate": 3.4193301333126927e-10, + "loss": 0.6987, + "step": 23296 + }, + { + "epoch": 1.9852577758841075, + "grad_norm": 16.5, + "learning_rate": 3.3801421930573784e-10, + "loss": 0.536, + "step": 23297 + }, + { + "epoch": 1.9853429910524074, + "grad_norm": 26.875, + "learning_rate": 3.3411800792901293e-10, + "loss": 0.9491, + "step": 23298 + }, + { + "epoch": 1.9854282062207074, + "grad_norm": 14.9375, + "learning_rate": 3.3024437927131614e-10, + "loss": 0.5384, + "step": 23299 + }, + { + "epoch": 1.9855134213890073, + "grad_norm": 12.4375, + "learning_rate": 3.263933334027303e-10, + "loss": 0.4388, + "step": 23300 + }, + { + "epoch": 1.9855986365573073, + "grad_norm": 19.875, + "learning_rate": 3.2256487039292186e-10, + "loss": 0.8038, + "step": 23301 + }, + { + "epoch": 1.9856838517256072, + "grad_norm": 16.875, + "learning_rate": 3.187589903110022e-10, + "loss": 0.7514, + "step": 23302 + }, + { + "epoch": 1.9857690668939072, + "grad_norm": 13.375, + "learning_rate": 3.1497569322580526e-10, + "loss": 0.5579, + "step": 23303 + }, + { + "epoch": 1.9858542820622072, + "grad_norm": 24.25, + "learning_rate": 3.112149792056096e-10, + "loss": 0.8181, + "step": 23304 + }, + { + "epoch": 1.9859394972305071, + "grad_norm": 10.9375, + "learning_rate": 3.074768483184165e-10, + "loss": 0.4085, + "step": 23305 + }, + { + "epoch": 1.986024712398807, + "grad_norm": 17.75, + "learning_rate": 3.0376130063181073e-10, + "loss": 0.5633, + "step": 23306 + }, + { + "epoch": 1.986109927567107, + "grad_norm": 23.75, + "learning_rate": 3.0006833621282204e-10, + "loss": 0.5607, + "step": 23307 + }, + { + "epoch": 1.986195142735407, + "grad_norm": 29.375, + "learning_rate": 2.9639795512834133e-10, + "loss": 1.0904, + "step": 23308 + }, + { + "epoch": 1.986280357903707, + "grad_norm": 17.125, + "learning_rate": 2.927501574445657e-10, + "loss": 1.0816, + "step": 23309 + }, + { + "epoch": 1.986365573072007, + "grad_norm": 16.375, + "learning_rate": 2.8912494322755335e-10, + "loss": 0.431, + "step": 23310 + }, + { + "epoch": 1.9864507882403069, + "grad_norm": 19.125, + "learning_rate": 2.855223125426687e-10, + "loss": 0.6658, + "step": 23311 + }, + { + "epoch": 1.9865360034086068, + "grad_norm": 14.8125, + "learning_rate": 2.819422654551374e-10, + "loss": 0.6147, + "step": 23312 + }, + { + "epoch": 1.9866212185769068, + "grad_norm": 21.375, + "learning_rate": 2.7838480202949105e-10, + "loss": 0.7043, + "step": 23313 + }, + { + "epoch": 1.9867064337452067, + "grad_norm": 21.125, + "learning_rate": 2.748499223302614e-10, + "loss": 0.7865, + "step": 23314 + }, + { + "epoch": 1.9867916489135067, + "grad_norm": 12.1875, + "learning_rate": 2.7133762642114756e-10, + "loss": 0.7474, + "step": 23315 + }, + { + "epoch": 1.9868768640818066, + "grad_norm": 15.4375, + "learning_rate": 2.67847914365571e-10, + "loss": 0.6573, + "step": 23316 + }, + { + "epoch": 1.9869620792501066, + "grad_norm": 14.875, + "learning_rate": 2.643807862268144e-10, + "loss": 0.5092, + "step": 23317 + }, + { + "epoch": 1.9870472944184066, + "grad_norm": 17.5, + "learning_rate": 2.609362420673278e-10, + "loss": 0.5382, + "step": 23318 + }, + { + "epoch": 1.9871325095867065, + "grad_norm": 16.375, + "learning_rate": 2.575142819494225e-10, + "loss": 0.5776, + "step": 23319 + }, + { + "epoch": 1.9872177247550065, + "grad_norm": 20.5, + "learning_rate": 2.541149059349934e-10, + "loss": 0.8098, + "step": 23320 + }, + { + "epoch": 1.9873029399233064, + "grad_norm": 16.75, + "learning_rate": 2.507381140853804e-10, + "loss": 0.7249, + "step": 23321 + }, + { + "epoch": 1.9873881550916064, + "grad_norm": 18.25, + "learning_rate": 2.4738390646164567e-10, + "loss": 0.6222, + "step": 23322 + }, + { + "epoch": 1.9874733702599063, + "grad_norm": 16.375, + "learning_rate": 2.4405228312443516e-10, + "loss": 0.4627, + "step": 23323 + }, + { + "epoch": 1.9875585854282063, + "grad_norm": 16.75, + "learning_rate": 2.407432441339785e-10, + "loss": 0.4503, + "step": 23324 + }, + { + "epoch": 1.9876438005965062, + "grad_norm": 19.125, + "learning_rate": 2.374567895499502e-10, + "loss": 0.6984, + "step": 23325 + }, + { + "epoch": 1.9877290157648062, + "grad_norm": 19.75, + "learning_rate": 2.3419291943174716e-10, + "loss": 0.8107, + "step": 23326 + }, + { + "epoch": 1.9878142309331062, + "grad_norm": 15.6875, + "learning_rate": 2.3095163383848873e-10, + "loss": 1.0178, + "step": 23327 + }, + { + "epoch": 1.9878994461014061, + "grad_norm": 13.75, + "learning_rate": 2.277329328287392e-10, + "loss": 0.4053, + "step": 23328 + }, + { + "epoch": 1.987984661269706, + "grad_norm": 13.625, + "learning_rate": 2.2453681646050774e-10, + "loss": 0.5675, + "step": 23329 + }, + { + "epoch": 1.988069876438006, + "grad_norm": 18.0, + "learning_rate": 2.213632847916647e-10, + "loss": 0.7549, + "step": 23330 + }, + { + "epoch": 1.988155091606306, + "grad_norm": 12.4375, + "learning_rate": 2.1821233787966412e-10, + "loss": 0.5089, + "step": 23331 + }, + { + "epoch": 1.988240306774606, + "grad_norm": 17.5, + "learning_rate": 2.1508397578112738e-10, + "loss": 0.7466, + "step": 23332 + }, + { + "epoch": 1.988325521942906, + "grad_norm": 13.875, + "learning_rate": 2.1197819855295342e-10, + "loss": 0.5205, + "step": 23333 + }, + { + "epoch": 1.9884107371112059, + "grad_norm": 18.25, + "learning_rate": 2.0889500625093095e-10, + "loss": 0.8504, + "step": 23334 + }, + { + "epoch": 1.9884959522795058, + "grad_norm": 15.0, + "learning_rate": 2.0583439893098746e-10, + "loss": 0.6026, + "step": 23335 + }, + { + "epoch": 1.9885811674478058, + "grad_norm": 14.3125, + "learning_rate": 2.027963766483565e-10, + "loss": 0.5881, + "step": 23336 + }, + { + "epoch": 1.9886663826161057, + "grad_norm": 14.4375, + "learning_rate": 1.9978093945799415e-10, + "loss": 0.6151, + "step": 23337 + }, + { + "epoch": 1.9887515977844057, + "grad_norm": 14.4375, + "learning_rate": 1.9678808741444012e-10, + "loss": 0.5574, + "step": 23338 + }, + { + "epoch": 1.9888368129527056, + "grad_norm": 24.125, + "learning_rate": 1.9381782057167896e-10, + "loss": 0.7941, + "step": 23339 + }, + { + "epoch": 1.9889220281210056, + "grad_norm": 15.25, + "learning_rate": 1.90870138983279e-10, + "loss": 0.6016, + "step": 23340 + }, + { + "epoch": 1.9890072432893056, + "grad_norm": 17.625, + "learning_rate": 1.8794504270266967e-10, + "loss": 0.6921, + "step": 23341 + }, + { + "epoch": 1.9890924584576055, + "grad_norm": 13.5625, + "learning_rate": 1.8504253178272535e-10, + "loss": 0.4144, + "step": 23342 + }, + { + "epoch": 1.9891776736259055, + "grad_norm": 13.5625, + "learning_rate": 1.821626062759041e-10, + "loss": 0.4146, + "step": 23343 + }, + { + "epoch": 1.9892628887942054, + "grad_norm": 16.75, + "learning_rate": 1.7930526623410883e-10, + "loss": 0.6635, + "step": 23344 + }, + { + "epoch": 1.9893481039625054, + "grad_norm": 12.875, + "learning_rate": 1.7647051170910369e-10, + "loss": 0.3788, + "step": 23345 + }, + { + "epoch": 1.9894333191308053, + "grad_norm": 21.375, + "learning_rate": 1.736583427520977e-10, + "loss": 1.052, + "step": 23346 + }, + { + "epoch": 1.9895185342991053, + "grad_norm": 15.75, + "learning_rate": 1.7086875941374482e-10, + "loss": 0.7536, + "step": 23347 + }, + { + "epoch": 1.9896037494674053, + "grad_norm": 13.6875, + "learning_rate": 1.6810176174469894e-10, + "loss": 0.4115, + "step": 23348 + }, + { + "epoch": 1.9896889646357052, + "grad_norm": 14.5625, + "learning_rate": 1.6535734979492013e-10, + "loss": 0.3688, + "step": 23349 + }, + { + "epoch": 1.9897741798040052, + "grad_norm": 15.5, + "learning_rate": 1.626355236138133e-10, + "loss": 0.5693, + "step": 23350 + }, + { + "epoch": 1.9898593949723051, + "grad_norm": 14.3125, + "learning_rate": 1.5993628325078338e-10, + "loss": 0.7412, + "step": 23351 + }, + { + "epoch": 1.989944610140605, + "grad_norm": 23.0, + "learning_rate": 1.572596287544026e-10, + "loss": 1.0815, + "step": 23352 + }, + { + "epoch": 1.990029825308905, + "grad_norm": 14.4375, + "learning_rate": 1.5460556017324325e-10, + "loss": 0.5194, + "step": 23353 + }, + { + "epoch": 1.990115040477205, + "grad_norm": 19.5, + "learning_rate": 1.5197407755518369e-10, + "loss": 0.5725, + "step": 23354 + }, + { + "epoch": 1.990200255645505, + "grad_norm": 10.9375, + "learning_rate": 1.493651809478247e-10, + "loss": 0.2269, + "step": 23355 + }, + { + "epoch": 1.990285470813805, + "grad_norm": 13.6875, + "learning_rate": 1.4677887039807327e-10, + "loss": 0.248, + "step": 23356 + }, + { + "epoch": 1.9903706859821049, + "grad_norm": 26.625, + "learning_rate": 1.4421514595297504e-10, + "loss": 0.7722, + "step": 23357 + }, + { + "epoch": 1.9904559011504048, + "grad_norm": 13.6875, + "learning_rate": 1.4167400765874307e-10, + "loss": 0.5324, + "step": 23358 + }, + { + "epoch": 1.9905411163187048, + "grad_norm": 13.875, + "learning_rate": 1.3915545556131283e-10, + "loss": 0.699, + "step": 23359 + }, + { + "epoch": 1.9906263314870047, + "grad_norm": 14.0625, + "learning_rate": 1.3665948970606468e-10, + "loss": 0.5638, + "step": 23360 + }, + { + "epoch": 1.9907115466553047, + "grad_norm": 16.75, + "learning_rate": 1.3418611013837901e-10, + "loss": 0.5347, + "step": 23361 + }, + { + "epoch": 1.9907967618236047, + "grad_norm": 19.5, + "learning_rate": 1.3173531690266473e-10, + "loss": 0.7225, + "step": 23362 + }, + { + "epoch": 1.9908819769919046, + "grad_norm": 15.125, + "learning_rate": 1.2930711004333073e-10, + "loss": 0.6781, + "step": 23363 + }, + { + "epoch": 1.9909671921602046, + "grad_norm": 18.0, + "learning_rate": 1.2690148960436966e-10, + "loss": 0.536, + "step": 23364 + }, + { + "epoch": 1.9910524073285045, + "grad_norm": 16.625, + "learning_rate": 1.2451845562894137e-10, + "loss": 0.4712, + "step": 23365 + }, + { + "epoch": 1.9911376224968045, + "grad_norm": 21.625, + "learning_rate": 1.2215800816048339e-10, + "loss": 0.7013, + "step": 23366 + }, + { + "epoch": 1.9912228376651044, + "grad_norm": 16.75, + "learning_rate": 1.1982014724146174e-10, + "loss": 0.8034, + "step": 23367 + }, + { + "epoch": 1.9913080528334044, + "grad_norm": 17.25, + "learning_rate": 1.1750487291406488e-10, + "loss": 0.5762, + "step": 23368 + }, + { + "epoch": 1.9913932680017044, + "grad_norm": 18.5, + "learning_rate": 1.152121852203425e-10, + "loss": 0.6992, + "step": 23369 + }, + { + "epoch": 1.9914784831700043, + "grad_norm": 14.625, + "learning_rate": 1.1294208420151165e-10, + "loss": 0.6394, + "step": 23370 + }, + { + "epoch": 1.9915636983383043, + "grad_norm": 15.75, + "learning_rate": 1.1069456989865057e-10, + "loss": 0.4918, + "step": 23371 + }, + { + "epoch": 1.9916489135066042, + "grad_norm": 18.5, + "learning_rate": 1.0846964235242119e-10, + "loss": 0.9906, + "step": 23372 + }, + { + "epoch": 1.9917341286749042, + "grad_norm": 15.625, + "learning_rate": 1.062673016029303e-10, + "loss": 0.8268, + "step": 23373 + }, + { + "epoch": 1.9918193438432041, + "grad_norm": 12.9375, + "learning_rate": 1.0408754769014595e-10, + "loss": 0.4655, + "step": 23374 + }, + { + "epoch": 1.991904559011504, + "grad_norm": 18.5, + "learning_rate": 1.0193038065334226e-10, + "loss": 0.4301, + "step": 23375 + }, + { + "epoch": 1.991989774179804, + "grad_norm": 12.6875, + "learning_rate": 9.979580053151582e-11, + "loss": 0.4808, + "step": 23376 + }, + { + "epoch": 1.992074989348104, + "grad_norm": 15.9375, + "learning_rate": 9.768380736338568e-11, + "loss": 0.8434, + "step": 23377 + }, + { + "epoch": 1.992160204516404, + "grad_norm": 14.0, + "learning_rate": 9.559440118683816e-11, + "loss": 0.3944, + "step": 23378 + }, + { + "epoch": 1.992245419684704, + "grad_norm": 12.5625, + "learning_rate": 9.352758203975965e-11, + "loss": 0.2396, + "step": 23379 + }, + { + "epoch": 1.9923306348530039, + "grad_norm": 18.125, + "learning_rate": 9.148334995962016e-11, + "loss": 0.6497, + "step": 23380 + }, + { + "epoch": 1.9924158500213038, + "grad_norm": 12.5625, + "learning_rate": 8.946170498319584e-11, + "loss": 0.5476, + "step": 23381 + }, + { + "epoch": 1.9925010651896038, + "grad_norm": 16.125, + "learning_rate": 8.746264714698527e-11, + "loss": 0.8115, + "step": 23382 + }, + { + "epoch": 1.9925862803579037, + "grad_norm": 16.25, + "learning_rate": 8.548617648734825e-11, + "loss": 0.6526, + "step": 23383 + }, + { + "epoch": 1.9926714955262037, + "grad_norm": 14.5625, + "learning_rate": 8.353229303981191e-11, + "loss": 0.6659, + "step": 23384 + }, + { + "epoch": 1.9927567106945037, + "grad_norm": 12.375, + "learning_rate": 8.160099683976463e-11, + "loss": 0.6611, + "step": 23385 + }, + { + "epoch": 1.9928419258628036, + "grad_norm": 19.125, + "learning_rate": 7.969228792203965e-11, + "loss": 0.8846, + "step": 23386 + }, + { + "epoch": 1.9929271410311036, + "grad_norm": 14.9375, + "learning_rate": 7.780616632119264e-11, + "loss": 0.4177, + "step": 23387 + }, + { + "epoch": 1.9930123561994035, + "grad_norm": 12.75, + "learning_rate": 7.594263207122421e-11, + "loss": 0.4071, + "step": 23388 + }, + { + "epoch": 1.9930975713677035, + "grad_norm": 14.3125, + "learning_rate": 7.410168520599614e-11, + "loss": 0.5367, + "step": 23389 + }, + { + "epoch": 1.9931827865360034, + "grad_norm": 14.25, + "learning_rate": 7.228332575853758e-11, + "loss": 0.528, + "step": 23390 + }, + { + "epoch": 1.9932680017043034, + "grad_norm": 13.0, + "learning_rate": 7.048755376187766e-11, + "loss": 0.4956, + "step": 23391 + }, + { + "epoch": 1.9933532168726034, + "grad_norm": 17.0, + "learning_rate": 6.871436924835162e-11, + "loss": 0.7185, + "step": 23392 + }, + { + "epoch": 1.9934384320409033, + "grad_norm": 19.0, + "learning_rate": 6.696377225015594e-11, + "loss": 0.9757, + "step": 23393 + }, + { + "epoch": 1.9935236472092033, + "grad_norm": 24.0, + "learning_rate": 6.52357627986544e-11, + "loss": 0.6906, + "step": 23394 + }, + { + "epoch": 1.9936088623775032, + "grad_norm": 18.25, + "learning_rate": 6.353034092548838e-11, + "loss": 0.5773, + "step": 23395 + }, + { + "epoch": 1.9936940775458032, + "grad_norm": 11.5, + "learning_rate": 6.184750666105022e-11, + "loss": 0.2112, + "step": 23396 + }, + { + "epoch": 1.9937792927141031, + "grad_norm": 18.375, + "learning_rate": 6.018726003600983e-11, + "loss": 0.6498, + "step": 23397 + }, + { + "epoch": 1.993864507882403, + "grad_norm": 15.3125, + "learning_rate": 5.854960108020447e-11, + "loss": 0.5303, + "step": 23398 + }, + { + "epoch": 1.993949723050703, + "grad_norm": 14.875, + "learning_rate": 5.693452982347136e-11, + "loss": 0.7258, + "step": 23399 + }, + { + "epoch": 1.994034938219003, + "grad_norm": 18.125, + "learning_rate": 5.53420462948151e-11, + "loss": 0.5227, + "step": 23400 + }, + { + "epoch": 1.994120153387303, + "grad_norm": 19.0, + "learning_rate": 5.37721505229627e-11, + "loss": 0.9858, + "step": 23401 + }, + { + "epoch": 1.994205368555603, + "grad_norm": 12.375, + "learning_rate": 5.222484253636362e-11, + "loss": 0.3322, + "step": 23402 + }, + { + "epoch": 1.9942905837239029, + "grad_norm": 14.9375, + "learning_rate": 5.0700122363050995e-11, + "loss": 0.6295, + "step": 23403 + }, + { + "epoch": 1.9943757988922028, + "grad_norm": 19.875, + "learning_rate": 4.919799003050285e-11, + "loss": 0.7383, + "step": 23404 + }, + { + "epoch": 1.9944610140605028, + "grad_norm": 9.625, + "learning_rate": 4.771844556578087e-11, + "loss": 0.2615, + "step": 23405 + }, + { + "epoch": 1.9945462292288028, + "grad_norm": 13.125, + "learning_rate": 4.6261488995807956e-11, + "loss": 0.3842, + "step": 23406 + }, + { + "epoch": 1.9946314443971027, + "grad_norm": 14.0, + "learning_rate": 4.4827120346813134e-11, + "loss": 0.435, + "step": 23407 + }, + { + "epoch": 1.9947166595654027, + "grad_norm": 18.625, + "learning_rate": 4.341533964474787e-11, + "loss": 0.3524, + "step": 23408 + }, + { + "epoch": 1.9948018747337026, + "grad_norm": 17.5, + "learning_rate": 4.2026146915008505e-11, + "loss": 0.5361, + "step": 23409 + }, + { + "epoch": 1.9948870899020026, + "grad_norm": 15.4375, + "learning_rate": 4.0659542182852616e-11, + "loss": 0.6589, + "step": 23410 + }, + { + "epoch": 1.9949723050703025, + "grad_norm": 15.75, + "learning_rate": 3.93155254728439e-11, + "loss": 0.6018, + "step": 23411 + }, + { + "epoch": 1.9950575202386025, + "grad_norm": 16.5, + "learning_rate": 3.799409680940724e-11, + "loss": 0.4634, + "step": 23412 + }, + { + "epoch": 1.9951427354069025, + "grad_norm": 20.125, + "learning_rate": 3.6695256216273675e-11, + "loss": 0.5748, + "step": 23413 + }, + { + "epoch": 1.9952279505752024, + "grad_norm": 14.875, + "learning_rate": 3.541900371703544e-11, + "loss": 0.6229, + "step": 23414 + }, + { + "epoch": 1.9953131657435024, + "grad_norm": 15.5625, + "learning_rate": 3.4165339334729654e-11, + "loss": 0.3247, + "step": 23415 + }, + { + "epoch": 1.9953983809118023, + "grad_norm": 30.75, + "learning_rate": 3.2934263091977115e-11, + "loss": 0.4226, + "step": 23416 + }, + { + "epoch": 1.9954835960801023, + "grad_norm": 33.5, + "learning_rate": 3.1725775010982284e-11, + "loss": 1.1166, + "step": 23417 + }, + { + "epoch": 1.9955688112484022, + "grad_norm": 14.3125, + "learning_rate": 3.053987511367207e-11, + "loss": 0.4443, + "step": 23418 + }, + { + "epoch": 1.9956540264167022, + "grad_norm": 13.375, + "learning_rate": 2.937656342155704e-11, + "loss": 0.6238, + "step": 23419 + }, + { + "epoch": 1.9957392415850022, + "grad_norm": 10.5, + "learning_rate": 2.823583995545387e-11, + "loss": 0.2475, + "step": 23420 + }, + { + "epoch": 1.9958244567533021, + "grad_norm": 19.125, + "learning_rate": 2.7117704736040475e-11, + "loss": 0.5291, + "step": 23421 + }, + { + "epoch": 1.995909671921602, + "grad_norm": 25.0, + "learning_rate": 2.6022157783578416e-11, + "loss": 0.5864, + "step": 23422 + }, + { + "epoch": 1.995994887089902, + "grad_norm": 11.5625, + "learning_rate": 2.4949199117912937e-11, + "loss": 0.2874, + "step": 23423 + }, + { + "epoch": 1.996080102258202, + "grad_norm": 25.625, + "learning_rate": 2.389882875833416e-11, + "loss": 0.5458, + "step": 23424 + }, + { + "epoch": 1.996165317426502, + "grad_norm": 13.25, + "learning_rate": 2.2871046723854652e-11, + "loss": 0.3514, + "step": 23425 + }, + { + "epoch": 1.996250532594802, + "grad_norm": 21.75, + "learning_rate": 2.186585303293187e-11, + "loss": 0.9411, + "step": 23426 + }, + { + "epoch": 1.9963357477631019, + "grad_norm": 14.8125, + "learning_rate": 2.088324770402328e-11, + "loss": 0.694, + "step": 23427 + }, + { + "epoch": 1.9964209629314018, + "grad_norm": 15.625, + "learning_rate": 1.9923230754614886e-11, + "loss": 0.5039, + "step": 23428 + }, + { + "epoch": 1.9965061780997018, + "grad_norm": 12.875, + "learning_rate": 1.8985802202053926e-11, + "loss": 0.5889, + "step": 23429 + }, + { + "epoch": 1.9965913932680017, + "grad_norm": 16.75, + "learning_rate": 1.8070962063548856e-11, + "loss": 0.6589, + "step": 23430 + }, + { + "epoch": 1.9966766084363017, + "grad_norm": 13.375, + "learning_rate": 1.717871035533669e-11, + "loss": 0.3435, + "step": 23431 + }, + { + "epoch": 1.9967618236046016, + "grad_norm": 13.5, + "learning_rate": 1.630904709365444e-11, + "loss": 0.7092, + "step": 23432 + }, + { + "epoch": 1.9968470387729016, + "grad_norm": 22.125, + "learning_rate": 1.546197229432278e-11, + "loss": 0.9153, + "step": 23433 + }, + { + "epoch": 1.9969322539412016, + "grad_norm": 18.875, + "learning_rate": 1.4637485972468502e-11, + "loss": 0.6093, + "step": 23434 + }, + { + "epoch": 1.9970174691095015, + "grad_norm": 13.5625, + "learning_rate": 1.3835588143079615e-11, + "loss": 0.3844, + "step": 23435 + }, + { + "epoch": 1.9971026842778015, + "grad_norm": 12.8125, + "learning_rate": 1.3056278820727797e-11, + "loss": 0.2483, + "step": 23436 + }, + { + "epoch": 1.9971878994461014, + "grad_norm": 15.625, + "learning_rate": 1.2299558019290836e-11, + "loss": 0.6519, + "step": 23437 + }, + { + "epoch": 1.9972731146144014, + "grad_norm": 19.125, + "learning_rate": 1.156542575264652e-11, + "loss": 0.5929, + "step": 23438 + }, + { + "epoch": 1.9973583297827013, + "grad_norm": 15.4375, + "learning_rate": 1.0853882033839969e-11, + "loss": 0.5593, + "step": 23439 + }, + { + "epoch": 1.9974435449510013, + "grad_norm": 14.8125, + "learning_rate": 1.0164926876055082e-11, + "loss": 0.4518, + "step": 23440 + }, + { + "epoch": 1.9975287601193012, + "grad_norm": 16.625, + "learning_rate": 9.498560291365532e-12, + "loss": 0.6395, + "step": 23441 + }, + { + "epoch": 1.9976139752876012, + "grad_norm": 17.0, + "learning_rate": 8.854782292122554e-12, + "loss": 0.8704, + "step": 23442 + }, + { + "epoch": 1.9976991904559012, + "grad_norm": 16.375, + "learning_rate": 8.23359288984471e-12, + "loss": 0.7305, + "step": 23443 + }, + { + "epoch": 1.9977844056242011, + "grad_norm": 12.0, + "learning_rate": 7.634992095634231e-12, + "loss": 0.3576, + "step": 23444 + }, + { + "epoch": 1.997869620792501, + "grad_norm": 15.75, + "learning_rate": 7.058979920593345e-12, + "loss": 0.4708, + "step": 23445 + }, + { + "epoch": 1.997954835960801, + "grad_norm": 16.125, + "learning_rate": 6.505556374852839e-12, + "loss": 0.7115, + "step": 23446 + }, + { + "epoch": 1.998040051129101, + "grad_norm": 16.375, + "learning_rate": 5.974721468543498e-12, + "loss": 0.4002, + "step": 23447 + }, + { + "epoch": 1.998125266297401, + "grad_norm": 19.5, + "learning_rate": 5.466475211240996e-12, + "loss": 0.958, + "step": 23448 + }, + { + "epoch": 1.998210481465701, + "grad_norm": 16.875, + "learning_rate": 4.9808176121046715e-12, + "loss": 0.9991, + "step": 23449 + }, + { + "epoch": 1.9982956966340009, + "grad_norm": 15.75, + "learning_rate": 4.517748679877531e-12, + "loss": 0.4989, + "step": 23450 + }, + { + "epoch": 1.9983809118023008, + "grad_norm": 13.0625, + "learning_rate": 4.077268423025027e-12, + "loss": 0.5058, + "step": 23451 + }, + { + "epoch": 1.9984661269706008, + "grad_norm": 15.1875, + "learning_rate": 3.659376849596274e-12, + "loss": 0.4856, + "step": 23452 + }, + { + "epoch": 1.9985513421389007, + "grad_norm": 16.625, + "learning_rate": 3.2640739668077237e-12, + "loss": 0.6734, + "step": 23453 + }, + { + "epoch": 1.9986365573072007, + "grad_norm": 20.75, + "learning_rate": 2.89135978215338e-12, + "loss": 0.6762, + "step": 23454 + }, + { + "epoch": 1.9987217724755006, + "grad_norm": 17.0, + "learning_rate": 2.541234302155804e-12, + "loss": 0.6329, + "step": 23455 + }, + { + "epoch": 1.9988069876438006, + "grad_norm": 16.375, + "learning_rate": 2.2136975331987775e-12, + "loss": 0.4668, + "step": 23456 + }, + { + "epoch": 1.9988922028121006, + "grad_norm": 18.125, + "learning_rate": 1.90874948124975e-12, + "loss": 0.5856, + "step": 23457 + }, + { + "epoch": 1.9989774179804005, + "grad_norm": 16.75, + "learning_rate": 1.6263901517210579e-12, + "loss": 0.9124, + "step": 23458 + }, + { + "epoch": 1.9990626331487005, + "grad_norm": 15.6875, + "learning_rate": 1.366619549886261e-12, + "loss": 0.7548, + "step": 23459 + }, + { + "epoch": 1.9991478483170004, + "grad_norm": 16.125, + "learning_rate": 1.1294376801862517e-12, + "loss": 0.7113, + "step": 23460 + }, + { + "epoch": 1.9992330634853004, + "grad_norm": 12.125, + "learning_rate": 9.148445472006994e-13, + "loss": 0.7011, + "step": 23461 + }, + { + "epoch": 1.9993182786536003, + "grad_norm": 14.875, + "learning_rate": 7.228401545378294e-13, + "loss": 0.5103, + "step": 23462 + }, + { + "epoch": 1.9994034938219003, + "grad_norm": 13.625, + "learning_rate": 5.534245058058663e-13, + "loss": 0.5976, + "step": 23463 + }, + { + "epoch": 1.9994887089902003, + "grad_norm": 15.625, + "learning_rate": 4.0659760405792335e-13, + "loss": 0.7625, + "step": 23464 + }, + { + "epoch": 1.9995739241585002, + "grad_norm": 14.5625, + "learning_rate": 2.823594519307804e-13, + "loss": 0.6131, + "step": 23465 + }, + { + "epoch": 1.9996591393268002, + "grad_norm": 13.0, + "learning_rate": 1.8071005164488343e-13, + "loss": 0.6435, + "step": 23466 + }, + { + "epoch": 1.9997443544951001, + "grad_norm": 13.9375, + "learning_rate": 1.0164940514312272e-13, + "loss": 0.5344, + "step": 23467 + }, + { + "epoch": 1.9998295696634, + "grad_norm": 14.8125, + "learning_rate": 4.517751367449919e-14, + "loss": 0.3828, + "step": 23468 + }, + { + "epoch": 1.9999147848317, + "grad_norm": 22.875, + "learning_rate": 1.1294378488013736e-14, + "loss": 0.5472, + "step": 23469 + }, + { + "epoch": 2.0, + "grad_norm": 16.75, + "learning_rate": 0.0, + "loss": 0.7522, + "step": 23470 + } + ], + "logging_steps": 1, + "max_steps": 23470, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 5868, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.3209448924525363e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}